The initial version of a maturity roadmap to help guide businesses when adopting AI technology into their workflow. IBM Watson Studio is referenced as an example of technology that can help in accelerating the adoption process.
How to Use a Semantic Layer to Deliver Actionable Insights at ScaleDATAVERSITY
Learn about using a semantic layer to enable actionable insights for everyone and streamline data and analytics access throughout your organization. This session will offer practical advice based on a decade of experience making semantic layers work for Enterprise customers.
Attend this session to learn about:
- Delivering critical business data to users faster than ever at scale using a semantic layer
- Enabling data teams to model and deliver a semantic layer on data in the cloud.
- Maintaining a single source of governed metrics and business data
- Achieving speed of thought query performance and consistent KPIs across any BI/AI tool like Excel, Power BI, Tableau, Looker, DataRobot, Databricks and more.
- Providing dimensional analysis capability that accelerates performance with no need to extract data from the cloud data warehouse
Who should attend this session?
Data & Analytics leaders and practitioners (e.g., Chief Data Officers, data scientists, data literacy, business intelligence, and analytics professionals).
As the adoption of AI technologies increases and matures, the focus will shift from exploration to time to market, productivity and integration with existing workflows. Governing Enterprise data, scaling AI model development, selecting a complete, collaborative hybrid platform and tools for rapid solution deployments are key focus areas for growing data scientist teams tasked to respond to business challenges. This talk will cover the challenges and innovations for AI at scale for the Enterprise focusing on the modernization of data analytics, the AI ladder and AI life cycle and infrastructure architecture considerations. We will conclude by viewing the benefits and innovation of running your modern AI and Data Analytics applications such as SAS Viya and SAP HANA on IBM Power Systems and IBM Storage in hybrid cloud environments.
MLOps and Data Quality: Deploying Reliable ML Models in ProductionProvectus
Looking to build a robust machine learning infrastructure to streamline MLOps? Learn from Provectus experts how to ensure the success of your MLOps initiative by implementing Data QA components in your ML infrastructure.
For most organizations, the development of multiple machine learning models, their deployment and maintenance in production are relatively new tasks. Join Provectus as we explain how to build an end-to-end infrastructure for machine learning, with a focus on data quality and metadata management, to standardize and streamline machine learning life cycle management (MLOps).
Agenda
- Data Quality and why it matters
- Challenges and solutions of Data Testing
- Challenges and solutions of Model Testing
- MLOps pipelines and why they matter
- How to expand validation pipelines for Data Quality
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...DATAVERSITY
Change is hard, especially in response to negative stimuli or what is perceived as negative stimuli. So organizations need to reframe how they think about data privacy, security and governance, treating them as value centers to 1) ensure enterprise data can flow where it needs to, 2) prevent – not just react – to internal and external threats, and 3) comply with data privacy and security regulations.
Working together, these roles can accelerate faster access to approved, relevant and higher quality data – and that means more successful use cases, faster speed to insights, and better business outcomes. However, both new information and tools are required to make the shift from defense to offense, reducing data drama while increasing its value.
Join us for this panel discussion with experts in these fields as they discuss:
- Recent research about where data privacy, security and governance stand
- The most valuable enterprise data use cases
- The common obstacles to data value creation
- New approaches to data privacy, security and governance
- Their advice on how to shift from a reactive to resilient mindset/culture/organization
You’ll be educated, entertained and inspired by this panel and their expertise in using the data trifecta to innovate more often, operate more efficiently, and differentiate more strategically.
Learn to Use Databricks for the Full ML LifecycleDatabricks
Machine learning development brings many new complexities beyond the traditional software development lifecycle. Unlike traditional software development, ML developers want to try multiple algorithms, tools and parameters to get the best results, and they need to track this information to reproduce work. In addition, developers need to use many distinct systems to productionize models. In this talk, learn how to operationalize ML across the full lifecycle with Databricks Machine Learning.
The quest for the insight-driven enterprise has spurned a mass exodus to the cloud. But cloud data ecosystems can be very complex with multiple data storage and processing options.
These slides-based on the webinar featuring leading IT analyst firm EMA, Amazon Web Services (AWS), and Trifacta--will help you: understand technology trends that simplify your analytics modernization journey; learn best practices to operationalize data management on AWS; establish operational excellence leveraging AWS data storage and processing; accelerate time-to-value for analytics projects with data preparation on AWS.
Introdution to Dataops and AIOps (or MLOps)Adrien Blind
This presentation introduces the audience to the DataOps and AIOps practices. It deals with organizational & tech aspects, and provide hints to start you data journey.
How to Use a Semantic Layer to Deliver Actionable Insights at ScaleDATAVERSITY
Learn about using a semantic layer to enable actionable insights for everyone and streamline data and analytics access throughout your organization. This session will offer practical advice based on a decade of experience making semantic layers work for Enterprise customers.
Attend this session to learn about:
- Delivering critical business data to users faster than ever at scale using a semantic layer
- Enabling data teams to model and deliver a semantic layer on data in the cloud.
- Maintaining a single source of governed metrics and business data
- Achieving speed of thought query performance and consistent KPIs across any BI/AI tool like Excel, Power BI, Tableau, Looker, DataRobot, Databricks and more.
- Providing dimensional analysis capability that accelerates performance with no need to extract data from the cloud data warehouse
Who should attend this session?
Data & Analytics leaders and practitioners (e.g., Chief Data Officers, data scientists, data literacy, business intelligence, and analytics professionals).
As the adoption of AI technologies increases and matures, the focus will shift from exploration to time to market, productivity and integration with existing workflows. Governing Enterprise data, scaling AI model development, selecting a complete, collaborative hybrid platform and tools for rapid solution deployments are key focus areas for growing data scientist teams tasked to respond to business challenges. This talk will cover the challenges and innovations for AI at scale for the Enterprise focusing on the modernization of data analytics, the AI ladder and AI life cycle and infrastructure architecture considerations. We will conclude by viewing the benefits and innovation of running your modern AI and Data Analytics applications such as SAS Viya and SAP HANA on IBM Power Systems and IBM Storage in hybrid cloud environments.
MLOps and Data Quality: Deploying Reliable ML Models in ProductionProvectus
Looking to build a robust machine learning infrastructure to streamline MLOps? Learn from Provectus experts how to ensure the success of your MLOps initiative by implementing Data QA components in your ML infrastructure.
For most organizations, the development of multiple machine learning models, their deployment and maintenance in production are relatively new tasks. Join Provectus as we explain how to build an end-to-end infrastructure for machine learning, with a focus on data quality and metadata management, to standardize and streamline machine learning life cycle management (MLOps).
Agenda
- Data Quality and why it matters
- Challenges and solutions of Data Testing
- Challenges and solutions of Model Testing
- MLOps pipelines and why they matter
- How to expand validation pipelines for Data Quality
The Data Trifecta – Privacy, Security & Governance Race from Reactivity to Re...DATAVERSITY
Change is hard, especially in response to negative stimuli or what is perceived as negative stimuli. So organizations need to reframe how they think about data privacy, security and governance, treating them as value centers to 1) ensure enterprise data can flow where it needs to, 2) prevent – not just react – to internal and external threats, and 3) comply with data privacy and security regulations.
Working together, these roles can accelerate faster access to approved, relevant and higher quality data – and that means more successful use cases, faster speed to insights, and better business outcomes. However, both new information and tools are required to make the shift from defense to offense, reducing data drama while increasing its value.
Join us for this panel discussion with experts in these fields as they discuss:
- Recent research about where data privacy, security and governance stand
- The most valuable enterprise data use cases
- The common obstacles to data value creation
- New approaches to data privacy, security and governance
- Their advice on how to shift from a reactive to resilient mindset/culture/organization
You’ll be educated, entertained and inspired by this panel and their expertise in using the data trifecta to innovate more often, operate more efficiently, and differentiate more strategically.
Learn to Use Databricks for the Full ML LifecycleDatabricks
Machine learning development brings many new complexities beyond the traditional software development lifecycle. Unlike traditional software development, ML developers want to try multiple algorithms, tools and parameters to get the best results, and they need to track this information to reproduce work. In addition, developers need to use many distinct systems to productionize models. In this talk, learn how to operationalize ML across the full lifecycle with Databricks Machine Learning.
The quest for the insight-driven enterprise has spurned a mass exodus to the cloud. But cloud data ecosystems can be very complex with multiple data storage and processing options.
These slides-based on the webinar featuring leading IT analyst firm EMA, Amazon Web Services (AWS), and Trifacta--will help you: understand technology trends that simplify your analytics modernization journey; learn best practices to operationalize data management on AWS; establish operational excellence leveraging AWS data storage and processing; accelerate time-to-value for analytics projects with data preparation on AWS.
Introdution to Dataops and AIOps (or MLOps)Adrien Blind
This presentation introduces the audience to the DataOps and AIOps practices. It deals with organizational & tech aspects, and provide hints to start you data journey.
Vertex AI - Unified ML Platform for the entire AI workflow on Google CloudMárton Kodok
Vertex AI is a managed ML platform for practitioners to accelerate experiments and deploy AI models.
Enhanced developer experience
- Build with the groundbreaking ML tools that power Google
- Approachable from the non-ML developer perspective (AutoML, managed models, training)
- Ease the life of a data scientist/ML (has feature store, managed datasets, endpoints, notebooks)
- Infrastructure management overhead have been almost completely eliminated
- Unified UI for the entire ML workflow
- End-to-end integration for data and AI with build pipelines that outperform and solve complex ML tasks
- Explainable AI and TensorBoard to visualize and track ML experiments
This describes a conceptual model approach to designing an enterprise data fabric. This is the set of hardware and software infrastructure, tools and facilities to implement, administer, manage and operate data operations across the entire span of the data within the enterprise across all data activities including data acquisition, transformation, storage, distribution, integration, replication, availability, security, protection, disaster recovery, presentation, analytics, preservation, retention, backup, retrieval, archival, recall, deletion, monitoring, capacity planning across all data storage platforms enabling use by applications to meet the data needs of the enterprise.
The conceptual data fabric model represents a rich picture of the enterprise’s data context. It embodies an idealised and target data view.
Designing a data fabric enables the enterprise respond to and take advantage of key related data trends:
• Internal and External Digital Expectations
• Cloud Offerings and Services
• Data Regulations
• Analytics Capabilities
It enables the IT function demonstrate positive data leadership. It shows the IT function is able and willing to respond to business data needs. It allows the enterprise to meet data challenges
• More and more data of many different types
• Increasingly distributed platform landscape
• Compliance and regulation
• Newer data technologies
• Shadow IT where the IT function cannot deliver IT change and new data facilities quickly
It is concerned with the design an open and flexible data fabric that improves the responsiveness of the IT function and reduces shadow IT.
The catalyst for the success of automobiles came not through the invention of the car but rather through the establishment of an innovative assembly line. History shows us that the ability to mass produce and distribute a product is the key to driving adoption of any innovation, and machine learning is no different. MLOps is the assembly line of Machine Learning and in this presentation we will discuss the core capabilities your organization should be focused on to implement a successful MLOps system.
The last year has put a new lens on what speed to insights actually mean - day-old data became useless, and only in-the-moment-insights became relevant, pushing data and analytics teams to their breaking point. The results, everyone has fast forwarded in their transformation and modernization plans, and it's also made us look differently at dashboards and the type of information that we're getting the business. Join this live event and hear about the data teams ditching their dashboards to embrace modern cloud analytics.
A Data Driven Roadmap to Enterprise AI Strategy (Sponsored by Contino) - AWS ...Amazon Web Services
AI is transforming every aspect of our daily lives and the data landscape is becoming increasing open and transparent, thanks to the Consumer Data Right, most notably Open Banking. Between the high level academia and low level algorithms, where should the modern business leader start on their AI journey and harness true value from their data? Let us show you a step by step, data-driven approach towards enterprise-wide AI adoption.
Using a Semantic and Graph-based Data Catalog in a Modern Data FabricCambridge Semantics
Watch this webinar to learn about the benefits of using semantic and graph database technology to create a Data Catalog of all of an enterprise's data, regardless of source or format, as part of a modern IT or data management stack and an important step toward building an Enterprise Data Fabric.
MLOps – Applying DevOps to Competitive AdvantageDATAVERSITY
MLOps is a practice for collaboration between Data Science and operations to manage the production machine learning (ML) lifecycles. As an amalgamation of “machine learning” and “operations,” MLOps applies DevOps principles to ML delivery, enabling the delivery of ML-based innovation at scale to result in:
Faster time to market of ML-based solutions
More rapid rate of experimentation, driving innovation
Assurance of quality, trustworthiness, and ethical AI
MLOps is essential for scaling ML. Without it, enterprises risk struggling with costly overhead and stalled progress. Several vendors have emerged with offerings to support MLOps: the major offerings are Microsoft Azure ML and Google Vertex AI. We looked at these offerings from the perspective of enterprise features and time-to-value.
Databricks: A Tool That Empowers You To Do More With DataDatabricks
In this talk we will present how Databricks has enabled the author to achieve more with data, enabling one person to build a coherent data project with data engineering, analysis and science components, with better collaboration, better productionalization methods, with larger datasets and faster.
The talk will include a demo that will illustrate how the multiple functionalities of Databricks help to build a coherent data project with Databricks jobs, Delta Lake and auto-loader for data engineering, SQL Analytics for Data Analysis, Spark ML and MLFlow for data science, and Projects for collaboration.
What you need to know about Generative AI and Data Management?Denodo
Watch full webinar here: https://buff.ly/3UXy0A2
It should be no surprise that Generative AI will have a profound impact to data management in years to come. Much like other areas of the technology sector, the opportunities presented by GenAI will accelerate our efforts around all aspects of data management, including self-service, automation, data governance and security. On the other hand, it is also becoming clearer that to unleash the true potential of AI assistants powered by GenAI, we need novel implementation strategies and a reimagined data architecture. This presents an exhilarating yet challenging future, demanding innovative thinking and methodologies in data management.
Join us on this webinar to learn about:
- The opportunities and challenges presented by GenAI today.
- Exploiting GenAI to democratize data management.
- How to augment GenAI applications with corporate data and knowledge.
- How to get started.
MLOps Bridging the gap between Data Scientists and Ops.Knoldus Inc.
Through this session we're going to introduce the MLOps lifecycle and discuss the hidden loopholes that can affect the MLProject. Then we are going to discuss the ML Model lifecycle and discuss the problem with training. We're going to introduce the MLFlow Tracking module in order to track the experiments.
You had a strategy. You were executing it. You were then side-swiped by COVID, spending countless cycles blocking and tackling. It is now time to step back onto your path.
CCG is holding a workshop to help you update your roadmap and get your team back on track and review how Microsoft Azure Solutions can be leveraged to build a strong foundation for governed data insights.
AI Maturity Levels and the Analytics TranslatorGoDataDriven
Buzzwords like Big Data, Cloud, and AI have been out there now for a couple of years. But today, businesses have a clear focus on the application of data use cases and the challenges around that such as metadata management, governance, security, and maintainability in general. Everybody seems to have some version of a data lake and wants to consolidate it into something (more) useful, or move from an on-premise version to the cloud. There is a general need to streamline current practices while also attempting to give multiple segments of users (data scientists, analysts, marketeers, business people, and HR) access in a way that is tailored to their needs and skills. In other words: businesses today are heavily invested in data and AI, but many have a hard time knowing how to mature it to the next level.
This is exactly where a "maturity model" comes into play. The goal of a maturity model is to help businesses in understanding their current and target competencies. This helps organisations in defining a roadmap for improving their competency. A maturity model is therefore one way of structuring progression, whether the company already embraces data science as a core competency, or, if it is just getting started.
In this presentation on maturity models, we answer the following questions:
1. What exactly is a maturity model and why would you need it? We address this by sharing GoDataDriven's maturity model and describing the different phases we have identified based on our experience in the field.
2. How can you use a maturity model to advance your organisation? Having a maturity model alone is not enough, in order for it to be valuable you need to act upon it. This paper provides concrete examples on how to do act based on practical stories and experiences from our clients and ourselves.
Building a Data Strategy Your C-Suite Will SupportReid Colson
Being a data leader in any industry is an advantage that creates measurable financial benefits. Many studies have shown this – I’ve seen them from Bain, McKinsey, MIT and more. Since most firms are measured on profit, getting good at making data driven decisions is a key to being competitive. You can't get there without a plan. That is where a data strategy comes in.
In speaking with ~300 firms who indicated that their organizations were effective in using data and analytics, McKinsey found that construction of a data strategy was the number one contributing factor to their success. Being good at using data to drive decisions creates a meaningful profit advantage and those who are leaders indicated that the number one driver of their success was their data strategy.
This presentation will cover what a data strategy is, how to construct one, and how to get buy in from your executive team. The author is a former Fortune 500 Chief Data Officer and has held senior data roles at Capital One and Markel.
Here are a few helpful links for your data journey:
Free Data Investment ROI Template:
https://www.udig.com/digging-in/roi-calculator-for-it-projects/
Real world data use cases:
https://www.udig.com/our-work/?category=data
Contact Me:
https://www.udig.com/contact/
Building the Artificially Intelligent EnterpriseDatabricks
This session looks at where we are today with data and analytics and what is needed to transition to the Artificially Intelligent Enterprise.
How do you mobilise developers to exploit what data scientists and business analysts have built? How do you align it all with business strategy to maximise business outcomes? How do you combine BI, predictive and prescriptive analytics, automation and reinforcement learning to get maximum value across the enterprise? What is the blueprint for building the artificially intelligent enterprise?
•Data and analytics – Where are we?
•Why is the journey only half-way done?
•2021 and beyond – The new era of AI usage and not just build
•The requirement – event-driven, on-demand and automated analytics
•Operationalising what you build – DataOps, MLOps and RPA
•Mobilising the masses to integrate AI into processes – what needs to be done?
•Business strategy alignment – the guiding light to AI utilisation for high reward
•Agility step change – the shift to no-code integration of AI by citizen developers
•Recording decisions, and analysing business impact
•Reinforcement-learning – transitioning to continuous reward
Vertex AI - Unified ML Platform for the entire AI workflow on Google CloudMárton Kodok
Vertex AI is a managed ML platform for practitioners to accelerate experiments and deploy AI models.
Enhanced developer experience
- Build with the groundbreaking ML tools that power Google
- Approachable from the non-ML developer perspective (AutoML, managed models, training)
- Ease the life of a data scientist/ML (has feature store, managed datasets, endpoints, notebooks)
- Infrastructure management overhead have been almost completely eliminated
- Unified UI for the entire ML workflow
- End-to-end integration for data and AI with build pipelines that outperform and solve complex ML tasks
- Explainable AI and TensorBoard to visualize and track ML experiments
This describes a conceptual model approach to designing an enterprise data fabric. This is the set of hardware and software infrastructure, tools and facilities to implement, administer, manage and operate data operations across the entire span of the data within the enterprise across all data activities including data acquisition, transformation, storage, distribution, integration, replication, availability, security, protection, disaster recovery, presentation, analytics, preservation, retention, backup, retrieval, archival, recall, deletion, monitoring, capacity planning across all data storage platforms enabling use by applications to meet the data needs of the enterprise.
The conceptual data fabric model represents a rich picture of the enterprise’s data context. It embodies an idealised and target data view.
Designing a data fabric enables the enterprise respond to and take advantage of key related data trends:
• Internal and External Digital Expectations
• Cloud Offerings and Services
• Data Regulations
• Analytics Capabilities
It enables the IT function demonstrate positive data leadership. It shows the IT function is able and willing to respond to business data needs. It allows the enterprise to meet data challenges
• More and more data of many different types
• Increasingly distributed platform landscape
• Compliance and regulation
• Newer data technologies
• Shadow IT where the IT function cannot deliver IT change and new data facilities quickly
It is concerned with the design an open and flexible data fabric that improves the responsiveness of the IT function and reduces shadow IT.
The catalyst for the success of automobiles came not through the invention of the car but rather through the establishment of an innovative assembly line. History shows us that the ability to mass produce and distribute a product is the key to driving adoption of any innovation, and machine learning is no different. MLOps is the assembly line of Machine Learning and in this presentation we will discuss the core capabilities your organization should be focused on to implement a successful MLOps system.
The last year has put a new lens on what speed to insights actually mean - day-old data became useless, and only in-the-moment-insights became relevant, pushing data and analytics teams to their breaking point. The results, everyone has fast forwarded in their transformation and modernization plans, and it's also made us look differently at dashboards and the type of information that we're getting the business. Join this live event and hear about the data teams ditching their dashboards to embrace modern cloud analytics.
A Data Driven Roadmap to Enterprise AI Strategy (Sponsored by Contino) - AWS ...Amazon Web Services
AI is transforming every aspect of our daily lives and the data landscape is becoming increasing open and transparent, thanks to the Consumer Data Right, most notably Open Banking. Between the high level academia and low level algorithms, where should the modern business leader start on their AI journey and harness true value from their data? Let us show you a step by step, data-driven approach towards enterprise-wide AI adoption.
Using a Semantic and Graph-based Data Catalog in a Modern Data FabricCambridge Semantics
Watch this webinar to learn about the benefits of using semantic and graph database technology to create a Data Catalog of all of an enterprise's data, regardless of source or format, as part of a modern IT or data management stack and an important step toward building an Enterprise Data Fabric.
MLOps – Applying DevOps to Competitive AdvantageDATAVERSITY
MLOps is a practice for collaboration between Data Science and operations to manage the production machine learning (ML) lifecycles. As an amalgamation of “machine learning” and “operations,” MLOps applies DevOps principles to ML delivery, enabling the delivery of ML-based innovation at scale to result in:
Faster time to market of ML-based solutions
More rapid rate of experimentation, driving innovation
Assurance of quality, trustworthiness, and ethical AI
MLOps is essential for scaling ML. Without it, enterprises risk struggling with costly overhead and stalled progress. Several vendors have emerged with offerings to support MLOps: the major offerings are Microsoft Azure ML and Google Vertex AI. We looked at these offerings from the perspective of enterprise features and time-to-value.
Databricks: A Tool That Empowers You To Do More With DataDatabricks
In this talk we will present how Databricks has enabled the author to achieve more with data, enabling one person to build a coherent data project with data engineering, analysis and science components, with better collaboration, better productionalization methods, with larger datasets and faster.
The talk will include a demo that will illustrate how the multiple functionalities of Databricks help to build a coherent data project with Databricks jobs, Delta Lake and auto-loader for data engineering, SQL Analytics for Data Analysis, Spark ML and MLFlow for data science, and Projects for collaboration.
What you need to know about Generative AI and Data Management?Denodo
Watch full webinar here: https://buff.ly/3UXy0A2
It should be no surprise that Generative AI will have a profound impact to data management in years to come. Much like other areas of the technology sector, the opportunities presented by GenAI will accelerate our efforts around all aspects of data management, including self-service, automation, data governance and security. On the other hand, it is also becoming clearer that to unleash the true potential of AI assistants powered by GenAI, we need novel implementation strategies and a reimagined data architecture. This presents an exhilarating yet challenging future, demanding innovative thinking and methodologies in data management.
Join us on this webinar to learn about:
- The opportunities and challenges presented by GenAI today.
- Exploiting GenAI to democratize data management.
- How to augment GenAI applications with corporate data and knowledge.
- How to get started.
MLOps Bridging the gap between Data Scientists and Ops.Knoldus Inc.
Through this session we're going to introduce the MLOps lifecycle and discuss the hidden loopholes that can affect the MLProject. Then we are going to discuss the ML Model lifecycle and discuss the problem with training. We're going to introduce the MLFlow Tracking module in order to track the experiments.
You had a strategy. You were executing it. You were then side-swiped by COVID, spending countless cycles blocking and tackling. It is now time to step back onto your path.
CCG is holding a workshop to help you update your roadmap and get your team back on track and review how Microsoft Azure Solutions can be leveraged to build a strong foundation for governed data insights.
AI Maturity Levels and the Analytics TranslatorGoDataDriven
Buzzwords like Big Data, Cloud, and AI have been out there now for a couple of years. But today, businesses have a clear focus on the application of data use cases and the challenges around that such as metadata management, governance, security, and maintainability in general. Everybody seems to have some version of a data lake and wants to consolidate it into something (more) useful, or move from an on-premise version to the cloud. There is a general need to streamline current practices while also attempting to give multiple segments of users (data scientists, analysts, marketeers, business people, and HR) access in a way that is tailored to their needs and skills. In other words: businesses today are heavily invested in data and AI, but many have a hard time knowing how to mature it to the next level.
This is exactly where a "maturity model" comes into play. The goal of a maturity model is to help businesses in understanding their current and target competencies. This helps organisations in defining a roadmap for improving their competency. A maturity model is therefore one way of structuring progression, whether the company already embraces data science as a core competency, or, if it is just getting started.
In this presentation on maturity models, we answer the following questions:
1. What exactly is a maturity model and why would you need it? We address this by sharing GoDataDriven's maturity model and describing the different phases we have identified based on our experience in the field.
2. How can you use a maturity model to advance your organisation? Having a maturity model alone is not enough, in order for it to be valuable you need to act upon it. This paper provides concrete examples on how to do act based on practical stories and experiences from our clients and ourselves.
Building a Data Strategy Your C-Suite Will SupportReid Colson
Being a data leader in any industry is an advantage that creates measurable financial benefits. Many studies have shown this – I’ve seen them from Bain, McKinsey, MIT and more. Since most firms are measured on profit, getting good at making data driven decisions is a key to being competitive. You can't get there without a plan. That is where a data strategy comes in.
In speaking with ~300 firms who indicated that their organizations were effective in using data and analytics, McKinsey found that construction of a data strategy was the number one contributing factor to their success. Being good at using data to drive decisions creates a meaningful profit advantage and those who are leaders indicated that the number one driver of their success was their data strategy.
This presentation will cover what a data strategy is, how to construct one, and how to get buy in from your executive team. The author is a former Fortune 500 Chief Data Officer and has held senior data roles at Capital One and Markel.
Here are a few helpful links for your data journey:
Free Data Investment ROI Template:
https://www.udig.com/digging-in/roi-calculator-for-it-projects/
Real world data use cases:
https://www.udig.com/our-work/?category=data
Contact Me:
https://www.udig.com/contact/
Building the Artificially Intelligent EnterpriseDatabricks
This session looks at where we are today with data and analytics and what is needed to transition to the Artificially Intelligent Enterprise.
How do you mobilise developers to exploit what data scientists and business analysts have built? How do you align it all with business strategy to maximise business outcomes? How do you combine BI, predictive and prescriptive analytics, automation and reinforcement learning to get maximum value across the enterprise? What is the blueprint for building the artificially intelligent enterprise?
•Data and analytics – Where are we?
•Why is the journey only half-way done?
•2021 and beyond – The new era of AI usage and not just build
•The requirement – event-driven, on-demand and automated analytics
•Operationalising what you build – DataOps, MLOps and RPA
•Mobilising the masses to integrate AI into processes – what needs to be done?
•Business strategy alignment – the guiding light to AI utilisation for high reward
•Agility step change – the shift to no-code integration of AI by citizen developers
•Recording decisions, and analysing business impact
•Reinforcement-learning – transitioning to continuous reward
Translating AI from Concept to Reality: Five Keys to Implementing AI for Know...Enterprise Knowledge
Lulit Tesfaye explains how foundational knowledge management and knowledge engineering approaches can play a key role in ensuring enterprise Artificial Intelligence (AI) initiatives start right, quickly demonstrate business value, and “stick” within the organization. The presentation includes real world case studies and examples of how organizations are approaching their data and AI transformations through knowledge maturity models to translate organizational information and data into actionable and clickable solutions. Originally delivered at data.world Summit, Spring 2022.
Successful artificial intelligence enables organizations to capture the thought process of top performers and deploy it as a virtual coach. Combining artificial intelligence with expert knowledge, metadata generation, auto-classification, and taxonomy management delivers great knowledge transfer.
In this webinar Discovery Machine and Concept Searching will demonstrate how their combined offering enables enterprises to establish an effective information framework by enhancing access to corporate knowledge sources with artificial intelligence.
Join us to find out more about how the solution can save your organization both time and money, while increasing accuracy and consistency of corporate knowledge access.
What you will learn about during this session:
• Capturing enterprise knowledge and deploying subject matter expertise as a virtual coach
• Effective content identification and classification, regardless of content location in the enterprise
• Eliminating the error and cost burdens of identification and management of records
• Documenting knowledge in the context of business process to create tangible knowledge assets
• Increasing the quality of information for decision making
• Automatic migration of content driven by classification of metadata
Speakers:
Todd Griffith, CTO and Co-Founder at Discovery Machine
Ken Lemons, Vice President Federal Programs at Concept Searching
John Challis, Founder and Chief Executive Officer at Concept Searching
Smarter businesses apply AI to learn and continuously evolve the way they work. To extract full value from AI, companies need data strategy that gives them access to all their data – no matter where it lives – in an environment that easily scales and applies the latest discovery technology including advanced analytics, visualization and AI. Learn how IBM Watson and Data provides all the tools companies need to embed AI, machine learning and deep learning in their business, while enabling professionals to gain the most from their data to drive smarter business and lead industry-changing transformations.
How to classify documents automatically using NLPSkyl.ai
About the webinar
Documents come in different shapes and sizes - From technical documents, customer support chat, emails, reviews to news articles - all of them contain information that is valuable to the business.
Managing these large volume data documents in a traditional manual way has been a complex and time-consuming task that requires enormous human efforts.
In this webinar, we will discuss how Machine learning can be used to identify and automatically label news articles into categories like business, politics, music, etc. This can be applied in another context like categorizing emails, reviews, and processing text documents, etc.
What you will learn
- How businesses are leveraging document classification to their advantage
- Best practice to automate machine learning models in hours not months
- Demo: Classify news articles into the right category using convolution neural network
ICP for Data- Enterprise platform for AI, ML and Data ScienceKaran Sachdeva
IBM Cloud Private for Data, an ultimate platform for all AI, ML and Data Science workloads. Integrated analytics platform based on Containers and micro services. Works with Kubernetes and dockers, even with Redhat openshift. Delivers the variety of business use cases in all industries- FS, Telco, Retail, Manufacturing etc
IBM i & digital transformation - Presentation & basic demo
IBM Watson Studio, IBM DSX Local w/ Open Source (Spark) & IBM Technology (OpenPower, CAPI, NVLINK)
With the rapid growth in data and move towards data commercialisation there are multiple aspects to focus on and prioritize the steps being taken across an enterprise. Enterprises face many challenges when it comes to truly becoming a data driven organization and realize the full potential of data. Some of those challenges include data availability, capacity to process, store and analyze this data, sharing the models and data artefacts across different teams etc. Most of these challenges could be handled through a platform which is Cloud based, scalable, and offers different capabilities for Governance, security, reusability and their likes. In this talk, I will talk about how IBM Cloud Pak serves as a framework for implementing your AI Strategy and how it could be used to build different artefacts while adhering to above listed requirements and being future ready. We will further illustrate how Cloud Pak for Data fastens and shortens the route to data commercialisation?
Once you’ve made the decision to leverage AI and/or machine learning, now you need to figure out how you will source the training data that is necessary for a fully functioning algorithm. Depending on your use case, you might need a significant amount of training data, and you’ll want to consider how that is labeled and annotated too.
View Applause's webinar with Cognilytica principal analysts Ronald Schmelzer and Kathleen Walch, alongside Kristin Simonini, Applause’s Vice President of Product, as they tackle the modern challenges that today’s companies face with sourcing training data.
Data science is the critical element in exploiting data, but several problems prevent organisations from maximising its value. Data scientists often find it hard to work efficiently, with delays in getting access to needed data and resources. Enterprise developers find it hard to incorporate machine learning models into their applications, and IT spends too much time supporting complex environments. Business users rarely are directly involved in the process and don’t have the means to build and consume their own predictive models. All of this means that business executives are not seeing the full ROI they expect from their data science and analytics investments. In this session, we will introduce some cloud based solutions designed to address these challenges.
Speaker: Stephen Weingartner, Solution Engineer, Oracle
Every business is looking for a game-changer in data science, machine learning, and AI. Most organizations are also looking for ways to tap into open-source and commercial data science tools such as Python, RStudio, Apache Spark, Jupyter, and Zeppelin notebooks, to accelerate predictive and machine learning model building and deployment while leveraging the scale, security and governance of the Hortonworks Data Platform and other commercial platforms.
Ana Maria Echeverri will demonstrate how to accelerate data science, machine learning, and deep learning workflows by using IBM Watson Studio, an integrated environment for data scientists, application developers, and subject matter experts. This suite of tools allows to collaboratively connect to data, wrangle that data and use it to build, train and deploy models at scale while using Open Source skills (i.e.: Python) and expanding into cognitive capabilities through access to Watson APIs to build AI-powered applications. If you love Python and want to tap into the power of IBM Watson, this is the session for you.
Bridging the Gap: Analyzing Data in and Below the CloudInside Analysis
The Briefing Room with Dean Abbott and Tableau Software
Live Webcast July 23, 2013
http://www.insideanalysis.com
Today’s desire for analytics extends well beyond the traditional domain of Business Intelligence. That’s partly because business users are realizing the value of mixing and matching all kinds of data, from all kinds of sources. One emerging market driver is Cloud-based data, and the desire companies have to analyze this data cohesively with their on-premise data sets.
Register for this episode of The Briefing Room to learn from Analyst Dean Abbott, who will explain how the ability to access data in the cloud can play a critical role for generating business value from analytics. He’ll be briefed by Ellie Fields of Tableau Software who will tout Tableau’s latest release, which includes native connectors to cloud-based applications like Salesforce.com, Amazon Redshift, Google Analytics and BigQuery. She’ll also demonstrate how Tableau can combine cloud data with other data sources, including spreadsheets, databases, cubes and even Big Data.
Accelerate Self-Service Analytics with Data Virtualization and VisualizationDenodo
Watch full webinar here: https://bit.ly/3fpitC3
Enterprise organizations are shifting to self-service analytics as business users need real-time access to holistic and consistent views of data regardless of its location, source or type for arriving at critical decisions.
Data Virtualization and Data Visualization work together through a universal semantic layer. Learn how they enable self-service data discovery and improve performance of your reports and dashboards.
In this session, you will learn:
- Challenges faced by business users
- How data virtualization enables self-service analytics
- Use case and lessons from customer success
- Overview of the highlight features in Tableau
IBM's Watson is a machine-learning platform that’s been built to mirror the same learning process that humans have: Observe, Interpret, Evaluate and Decide. Through the use of this cognitive framework, Watson can search through a database of information and pull out key insights to bridge gaps in human knowledge. It’s expertise scaling for enterprise.
Watson has already helped businesses across a variety of industries increase their customer engagement, data discovery and informed decision making abilities. Is your business next?
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
2. About Me
Focus / Passion
• AI, Cognitive, Emerging Technology
• Analytics
• Data (Architecture, Modeling, Integration)
• Cloud Service Architecture
• Applying the above to real-world business problems
Education & Certification
• M.S. Software Engineering
• B.S, Physics
• Data Mgmt, AI, Cloud, Docker, DevOps, …
Proud Member of
the IBM WolfPack
David Solomon
Technical
Evangelist, IBM
dsdlsolomo
@dlsolomo
Team-wolfpack
8. Gain value
from your data,
without limits
Access your data
All sources and all types
Flexibility
Support all data
types, all workloads,
all consumption models
Machine Learning
Make better decisions,
provide smarter capabilities
Democratize access
Provide data-driven decisions
to everyone
Simplicity
A unified experience in
managing your data
landscape
Cloud journey
Support your data regardless
of location
Essential elements of a hybrid data management strategy
11. 11
Introducing an AI Readiness Maturity Model
Insight
Hindered
Hindsight-
Driven
Data-
Driven
Insight-
Driven
AI-Driven
• Minimal data mgmt.
• Spreadsheets are
primary data tool
• Minimal standards
• Minimal Governance
• Centralized DBs for
critical data
• Some governance
• Siloed use of
unstructured data
• Data integration and
governance practice
• Organized use of
unstructured data
• Siloed Data Science
practices
• Data Science
practices in place
• Hybrid-data mgmt.
practice in place
• Leverage both cloud
and on-prem data
• Fully data-driven
business
• Access to all
required AI training
data
Data
Readiness
• Spreadsheet
analysis
• Desktop BI tools
• Minimal standards
• Soiled practices
• Focus on descriptive
analytics (What
Happened?)
• Standardized
reporting formats
• Diagnostic analytics
• Siloed use of
predictive analytics
• Siloed use of
Machine Learning
models
• Standard use of
Machine Learning
• Predictive analytics
• Siloed use of
prescriptive
analytics
• Prescriptive
analytics
• Fully insight-driven
business
Analytics
Readiness
Hindered
Business
Outcomes
Operational
Efficiency and
Cost Savings
Competitiveness Competitive
Advantage
Market Leader
• None
• Siloed
Experimentation
• Limited use for
siloed applications
• Initial production AI
applications
• Some alignment of
AI with business
strategy
• Standard AI practice
• Full alignment of AI
with business
strategy
AI
Capability
13. …and has grown to an entire portfolio of cognitive technologies
Retrieve and Rank
Language
• Conversation
• Document
Conversion
• Language
Translator
• Natural Language
Classifier
• Natural Language
Understanding
• Personality Insights
• Retrieve and Rank
• Tone Analyzer
Speech
• Speech to Text
• Text to Speech
Vision
• Visual Recognition
Data Insights
• Discovery
• Discovery News
• Watson Knowledge
Studio
Natural Language
Classifier
Tone Analyzer
14. Tools & Infrastructure
• Need an environment
that enables a “fail
fast” approach
• Discrete tools present
barriers to productivity
Governance
• If the data isn’t secure,
self-service isn’t a
reality
• Challenge
understanding data
lineage and getting to
a system of truth
Skills
• Data Science skills are
in low supply and high
demand
• Nurturing new data
professionals is
challenging
Data
• Data resides in silos &
difficult to access
• Unstructured and
external data wasn’t
considered
14
Why are enterprises struggling to
capture the value of AI?
How can these challenges be tackled in a timely manner?
15. Watson Studio
Supporting the end-to-end AI workflow
Prepare Data
for Analysis
Build and Train
ML/DL Models
Deploy Models
Monitor, Analyze
and Manage
Search and Find
Relevant Data
Connect &
Access Data
• Connect and
discover content
from multiple data
sources in the
cloud or on
premises.
• Bring structured
and unstructured
data to one toolkit.
• Clean and prepare
your data with Data
Refinery, a tool to
create data
preparation
pipelines visually.
• Use popular open
source libraries to
prepare
unstructured data.
• Democratize the
creation of ML and DL
models. Design your
AI models
programmatically or
visually with the most
popular open source
and IBM ML/DL
frameworks
• Leverage transfer
learning on pre-
trained models using
Watson tools to
adapt to your business
domain.
• Train at scale on
GPUs and
distributed compute
• Deploy your models
easily and have
them scale
automatically for
online, batch or
streaming use
cases
• Monitor the
performance of the
models in
production and
trigger automatic
retraining and
redeployment of
models.
• Find data
(structured,
unstructured) and
AI assets (e.g.,
ML/DL models,
notebooks, Watson
Data Kits) in the
Knowledge
Catalog
15
16. Her Job:
Builds AI application that meet the
requirements of the business.
What she does:
• Starts PoCs which includes
gathering content, dialog
building and model training
• Focus is on app building for the
team or company to use. Will
handle ML Ops as needed
Sometimes known as:
Front-end, back-end, full stack,
mobile or low-code developer
Tanya
Domain Expert
Her Job:
To transfer knowledge to Watson for
a successful user experience.
What she does:
• Range of domain knowledge and
uses that to teach Watson and
develop a custom models
• As Tanya gains more experience
she optimizes her knowledge to
teach Watson to design better
end-user experiences.
Sometimes known as:
Subject matter expert, content
strategist.
His Job:
Transform data into knowledge for
solving business problems.
What he does:
•Runs experiments to build custom
models that solve business problems.
•Use techniques such as Machine
Learning or Deep Learning and
works with Tanya to validate success
of trained models.
Watson Studio
Built for AI teams – enabling team productivity and collaboration
Sometimes known as:
ML/DL engineer, Modeler, Data Miner
Ed
Data Engineer
His Job:
Architects how data is organized
and ensures operability
What he does:
• Builds data infrastructure and ETL
pipelines. Works with Spark,
Hadoop, and HDFS.
• Works with data scientist to
transform research models into
production quality systems.
Sometimes known as:
Data infrastructure engineer
Mike
Data Scientist
Deb
The Developer
16
17. Watson Studio
Comprehensive set of tools for the end-to-end AI workflow
Model Lifecycle Management
Machine Learning Runtimes Deep Learning Runtimes
Authoring Tools
Cloud Infrastructure as a Service
Watson
API
Tools
Model
Builder
• Most popular open source frameworks
• IBM best-in-class frameworks
• Create, collaborate, deploy, and monitor
• Best of breed open source & IBM tools
• Code (R, Python or Scala) and no-code/visual
modeling tools
• Fully managed service
• Container-based resource management
• Elastic pay as you go CPU/GPU power
Data
Refinery
17
18. Watson Studio
Differentiating Capabilities
• Data Scientists, Subject Matter experts,
Business Analysts & Developers all in one
environment to accelerate innovation,
collaboration and productivity
• Built-in learning to get started or go the
distance with advanced tutorials
Integrated Collaboration Environment
• Best in-breed open source and IBM tools that
support the end-to-end AI lifecycle
• Choice of code or no-code tools to build and
train your own ML/DL models or easily train
and customize pre-trained Watson APIs
Choice of Tools for the full AI lifecycle
• Use Watson smarts and recommendations
for the best algorithms to use given your
data, OR
• Use the rich capabilities and controls to fine
tune your models
Support for all levels of expertise
• Monitor batch training experiments then
compare cross-model performance without
worrying about log transfers and scripts to
visualize results.
• You focus on designing your neural networks.
We’ll manage and track your assets.
Experiment centric DL workflow
• Deploy models into production then monitor
them to evaluate performance.
• Capture new data for continuous learning and
retrain models so they continually adapt to
changing conditions.
Model lifecycle & management
• Intelligent discovery of data and AI assets
that enables reuse & improves productivity
• Seamlessly integrated for productive use with
Machine Learning and Data science
• Powerful governance tools to control and
protect access to data
Integrated with Knowledge Catalog
18
For too long – data has been held captive within our systems of record. Isolated by the rigidity of platform/application/workload choices, segregated by business line, business function, and data type or initial usage.
The result is splintered views of segmented data that’s difficult to access on the whole, and impossible to attempt to gain true analytical insight from…..
And even this only speaks to the snapshot today and current models. The challenges are compounded as businesses look to change, grow, iterate practices, innovate, or disrupt markets.
Attempts at data science, machine learning, and deep learning are made moot by the fact that insights are only as good as the access to supporting data – which again is too fragmented to provide full value.
We believe, that in order to change this paradigm, a hybrid data management strategy should contain the elements here:
Access to all data regardless of source or type
The flexibility to support changing workloads and consumption cases
Possess intelligent analytics such as machine learning AT the data source
And…
Provide access to insights across the business, its functions, and to all users for better decision making
# # #
You need three essential elements on your journey to digital transformation.
You need to know your data. Typically this means building a 360-degree view of your focus area—for example, a 360-degree view of your customer. You need to gather your internal data and may also need to include external data from social media, click stream, census or other relevant sources.
This data must also be accessible by all users and/or applications that need it. This could mean making data globally accessible or running applications in the cloud. Consider that an application may need to access data from multiple data sources, so providing a common access layer is important to reduce application coding.
2. You need to be able to trust your data. Well-governed data provides confidence in not just the data itself, but the outcomes from analytics, reports and other tasks based on that data. There are two key points to data governance: First, you must have the ability to ensure the data is secure and adheres to compliance regulations. And second, you must have the ability to govern the data so your users can find and access information themselves, at the exact time they need it.
3. You must be able to use your data as a source for insights and intelligence. This means having not only the right skills and tools in place to surface insights, but also the right technology to learn from the data and improve accuracy each time that data is analyzed.
Three years ago Watson made it’s debut on the US Quiz show, Jeopardy, in a very public proof point of radical new technology. Jeopardy was the result of an IBM Grand Challenge – putting top scientists to work on a seemingly impossible task. IBM undertakes Grand Challenges every decade or so. The last grand challenge was “Deep Blue” in 1997 - a chess-playing computer that won the second six-game match against world champion Garry Kasparov by two wins to one with three draws.
Whether you attended one of the many IBM watch parties, watched the show at home, viewed it on YouTube later, or just read the newspapers, you witnessed history. Watson bested the 2 top champions, including Ken Jennings, who won 74 games and over $3M – the longest winning streak in J history.
Not only did Watson win, but in doing so it ushered in a whole new era of computing.
Additional Background:
What fascinated the IBM researchers was how Jeopardy was the ultimate test of IT capabilities because it relied on many human cognitive abilities traditionally seen beyond the capability of computers, such as:
The ability to discern double meanings of words, puns, rhymes, and inferred hints.
Extremely rapid responses (sifting through 200 million pages of information - in the span of seconds)
The ability to process vast amounts of information to make complex and subtle logical connections
A team of 15 IBM researchers working in collaboration with a pool of top universities as a “Deep QA” project. For the Watson team, replicating the human capabilities was an enormous challenge, moving beyond keyword searches and queries of structured data to asking questions and accessing and assessing a vast amount of unstructured data to find the best answer. But IBM that knew the solution to this challenge had the potential to change the way businesses use information and make decisions.