Explainable AI makes the algorithms to be transparent where they interpret, visualize, explain and integrate for fair, secure and trustworthy AI applications.
Introductory presentation to Explainable AI, defending its main motivations and importance. We describe briefly the main techniques available in March 2020 and share many references to allow the reader to continue his/her studies.
Explainable AI (XAI) is becoming Must-Have NFR for most AI enabled product or solution deployments. Keen to know viewpoints and collaboration opportunities.
An Introduction to XAI! Towards Trusting Your ML Models!Mansour Saffar
Machine learning (ML) is currently disrupting almost every industry and is being used as the core component in many systems. The decisions made by these systems may have a great impact on society and specific individuals and thus the decision-making process has to be clear and explainable so humans can trust it. Explainable AI (XAI) is a rather new field in ML in which researchers try to develop models that are able to explain the decision-making process behind ML models. In this talk, we'll learn about the fundamentals of XAI and discuss why we need to start to integrate XAI with our ML models!
Presented in Edmonton DataScience Meetup on October 2nd, 2019. Learn more: https://youtu.be/gEkPXOsDt_w
[Video recording available at https://www.youtube.com/playlist?list=PLewjn-vrZ7d3x0M4Uu_57oaJPRXkiS221]
Artificial Intelligence is increasingly playing an integral role in determining our day-to-day experiences. Moreover, with proliferation of AI based solutions in areas such as hiring, lending, criminal justice, healthcare, and education, the resulting personal and professional implications of AI are far-reaching. The dominant role played by AI models in these domains has led to a growing concern regarding potential bias in these models, and a demand for model transparency and interpretability. In addition, model explainability is a prerequisite for building trust and adoption of AI systems in high stakes domains requiring reliability and safety such as healthcare and automated transportation, and critical industrial applications with significant economic implications such as predictive maintenance, exploration of natural resources, and climate change modeling.
As a consequence, AI researchers and practitioners have focused their attention on explainable AI to help them better trust and understand models at scale. The challenges for the research community include (i) defining model explainability, (ii) formulating explainability tasks for understanding model behavior and developing solutions for these tasks, and finally (iii) designing measures for evaluating the performance of models in explainability tasks.
In this tutorial, we present an overview of model interpretability and explainability in AI, key regulations / laws, and techniques / tools for providing explainability as part of AI/ML systems. Then, we focus on the application of explainability techniques in industry, wherein we present practical challenges / guidelines for effectively using explainability techniques and lessons learned from deploying explainable models for several web-scale machine learning and data mining applications. We present case studies across different companies, spanning application domains such as search & recommendation systems, hiring, sales, and lending. Finally, based on our experiences in industry, we identify open problems and research directions for the data mining / machine learning community.
Artificial Intelligence is increasingly playing an integral role in determining our day-to-day experiences. Moreover, with proliferation of AI based solutions in areas such as hiring, lending, criminal justice, healthcare, and education, the resulting personal and professional implications of AI are far-reaching. The dominant role played by AI models in these domains has led to a growing concern regarding potential bias in these models, and a demand for model transparency and interpretability. In addition, model explainability is a prerequisite for building trust and adoption of AI systems in high stakes domains requiring reliability and safety such as healthcare and automated transportation, and critical industrial applications with significant economic implications such as predictive maintenance, exploration of natural resources, and climate change modeling.
As a consequence, AI researchers and practitioners have focused their attention on explainable AI to help them better trust and understand models at scale. The challenges for the research community include (i) defining model explainability, (ii) formulating explainability tasks for understanding model behavior and developing solutions for these tasks, and finally (iii) designing measures for evaluating the performance of models in explainability tasks.
In this tutorial, we present an overview of model interpretability and explainability in AI, key regulations / laws, and techniques / tools for providing explainability as part of AI/ML systems. Then, we focus on the application of explainability techniques in industry, wherein we present practical challenges / guidelines for effectively using explainability techniques and lessons learned from deploying explainable models for several web-scale machine learning and data mining applications. We present case studies across different companies, spanning application domains such as search & recommendation systems, sales, lending, and fraud detection. Finally, based on our experiences in industry, we identify open problems and research directions for the data mining / machine learning community.
Slide for Arithmer Seminar given by Dr. Daisuke Sato (Arithmer) at Arithmer inc.
The topic is on "explainable AI".
"Arithmer Seminar" is weekly held, where professionals from within and outside our company give lectures on their respective expertise.
The slides are made by the lecturer from outside our company, and shared here with his/her permission.
Arithmer株式会社は東京大学大学院数理科学研究科発の数学の会社です。私達は現代数学を応用して、様々な分野のソリューションに、新しい高度AIシステムを導入しています。AIをいかに上手に使って仕事を効率化するか、そして人々の役に立つ結果を生み出すのか、それを考えるのが私たちの仕事です。
Arithmer began at the University of Tokyo Graduate School of Mathematical Sciences. Today, our research of modern mathematics and AI systems has the capability of providing solutions when dealing with tough complex issues. At Arithmer we believe it is our job to realize the functions of AI through improving work efficiency and producing more useful results for society.
Introductory presentation to Explainable AI, defending its main motivations and importance. We describe briefly the main techniques available in March 2020 and share many references to allow the reader to continue his/her studies.
Explainable AI (XAI) is becoming Must-Have NFR for most AI enabled product or solution deployments. Keen to know viewpoints and collaboration opportunities.
An Introduction to XAI! Towards Trusting Your ML Models!Mansour Saffar
Machine learning (ML) is currently disrupting almost every industry and is being used as the core component in many systems. The decisions made by these systems may have a great impact on society and specific individuals and thus the decision-making process has to be clear and explainable so humans can trust it. Explainable AI (XAI) is a rather new field in ML in which researchers try to develop models that are able to explain the decision-making process behind ML models. In this talk, we'll learn about the fundamentals of XAI and discuss why we need to start to integrate XAI with our ML models!
Presented in Edmonton DataScience Meetup on October 2nd, 2019. Learn more: https://youtu.be/gEkPXOsDt_w
[Video recording available at https://www.youtube.com/playlist?list=PLewjn-vrZ7d3x0M4Uu_57oaJPRXkiS221]
Artificial Intelligence is increasingly playing an integral role in determining our day-to-day experiences. Moreover, with proliferation of AI based solutions in areas such as hiring, lending, criminal justice, healthcare, and education, the resulting personal and professional implications of AI are far-reaching. The dominant role played by AI models in these domains has led to a growing concern regarding potential bias in these models, and a demand for model transparency and interpretability. In addition, model explainability is a prerequisite for building trust and adoption of AI systems in high stakes domains requiring reliability and safety such as healthcare and automated transportation, and critical industrial applications with significant economic implications such as predictive maintenance, exploration of natural resources, and climate change modeling.
As a consequence, AI researchers and practitioners have focused their attention on explainable AI to help them better trust and understand models at scale. The challenges for the research community include (i) defining model explainability, (ii) formulating explainability tasks for understanding model behavior and developing solutions for these tasks, and finally (iii) designing measures for evaluating the performance of models in explainability tasks.
In this tutorial, we present an overview of model interpretability and explainability in AI, key regulations / laws, and techniques / tools for providing explainability as part of AI/ML systems. Then, we focus on the application of explainability techniques in industry, wherein we present practical challenges / guidelines for effectively using explainability techniques and lessons learned from deploying explainable models for several web-scale machine learning and data mining applications. We present case studies across different companies, spanning application domains such as search & recommendation systems, hiring, sales, and lending. Finally, based on our experiences in industry, we identify open problems and research directions for the data mining / machine learning community.
Artificial Intelligence is increasingly playing an integral role in determining our day-to-day experiences. Moreover, with proliferation of AI based solutions in areas such as hiring, lending, criminal justice, healthcare, and education, the resulting personal and professional implications of AI are far-reaching. The dominant role played by AI models in these domains has led to a growing concern regarding potential bias in these models, and a demand for model transparency and interpretability. In addition, model explainability is a prerequisite for building trust and adoption of AI systems in high stakes domains requiring reliability and safety such as healthcare and automated transportation, and critical industrial applications with significant economic implications such as predictive maintenance, exploration of natural resources, and climate change modeling.
As a consequence, AI researchers and practitioners have focused their attention on explainable AI to help them better trust and understand models at scale. The challenges for the research community include (i) defining model explainability, (ii) formulating explainability tasks for understanding model behavior and developing solutions for these tasks, and finally (iii) designing measures for evaluating the performance of models in explainability tasks.
In this tutorial, we present an overview of model interpretability and explainability in AI, key regulations / laws, and techniques / tools for providing explainability as part of AI/ML systems. Then, we focus on the application of explainability techniques in industry, wherein we present practical challenges / guidelines for effectively using explainability techniques and lessons learned from deploying explainable models for several web-scale machine learning and data mining applications. We present case studies across different companies, spanning application domains such as search & recommendation systems, sales, lending, and fraud detection. Finally, based on our experiences in industry, we identify open problems and research directions for the data mining / machine learning community.
Slide for Arithmer Seminar given by Dr. Daisuke Sato (Arithmer) at Arithmer inc.
The topic is on "explainable AI".
"Arithmer Seminar" is weekly held, where professionals from within and outside our company give lectures on their respective expertise.
The slides are made by the lecturer from outside our company, and shared here with his/her permission.
Arithmer株式会社は東京大学大学院数理科学研究科発の数学の会社です。私達は現代数学を応用して、様々な分野のソリューションに、新しい高度AIシステムを導入しています。AIをいかに上手に使って仕事を効率化するか、そして人々の役に立つ結果を生み出すのか、それを考えるのが私たちの仕事です。
Arithmer began at the University of Tokyo Graduate School of Mathematical Sciences. Today, our research of modern mathematics and AI systems has the capability of providing solutions when dealing with tough complex issues. At Arithmer we believe it is our job to realize the functions of AI through improving work efficiency and producing more useful results for society.
Explainable Artificial Intelligence (XAI)
Presented at Lightning Talk session at ICACCI'18 on 20th September 208
An Explainable AI (XAI) or Transparent AI is an artificial intelligence (AI) whose actions can be easily understood by humans. It contrasts with the concept of the "black box" in machine learning, meaning the "interpretability" of the workings of complex algorithms, where even their designers cannot explain why the AI arrived at a specific decision.
https://en.wikipedia.org/wiki/Explainable_Artificial_Intelligence
Unified Approach to Interpret Machine Learning Model: SHAP + LIMEDatabricks
For companies that solve real-world problems and generate revenue from the data science products, being able to understand why a model makes a certain prediction can be as crucial as achieving high prediction accuracy in many applications. However, as data scientists pursuing higher accuracy by implementing complex algorithms such as ensemble or deep learning models, the algorithm itself becomes a blackbox and it creates the trade-off between accuracy and interpretability of a model’s output.
To address this problem, a unified framework SHAP (SHapley Additive exPlanations) was developed to help users interpret the predictions of complex models. In this session, we will talk about how to apply SHAP to various modeling approaches (GLM, XGBoost, CNN) to explain how each feature contributes and extract intuitive insights from a particular prediction. This talk is intended to introduce the concept of general purpose model explainer, as well as help practitioners understand SHAP and its applications.
[Video recording available at https://www.youtube.com/playlist?list=PLewjn-vrZ7d3x0M4Uu_57oaJPRXkiS221]
Artificial Intelligence is increasingly playing an integral role in determining our day-to-day experiences. Moreover, with proliferation of AI based solutions in areas such as hiring, lending, criminal justice, healthcare, and education, the resulting personal and professional implications of AI are far-reaching. The dominant role played by AI models in these domains has led to a growing concern regarding potential bias in these models, and a demand for model transparency and interpretability. In addition, model explainability is a prerequisite for building trust and adoption of AI systems in high stakes domains requiring reliability and safety such as healthcare and automated transportation, and critical industrial applications with significant economic implications such as predictive maintenance, exploration of natural resources, and climate change modeling.
As a consequence, AI researchers and practitioners have focused their attention on explainable AI to help them better trust and understand models at scale. The challenges for the research community include (i) defining model explainability, (ii) formulating explainability tasks for understanding model behavior and developing solutions for these tasks, and finally (iii) designing measures for evaluating the performance of models in explainability tasks.
In this tutorial, we present an overview of model interpretability and explainability in AI, key regulations / laws, and techniques / tools for providing explainability as part of AI/ML systems. Then, we focus on the application of explainability techniques in industry, wherein we present practical challenges / guidelines for effectively using explainability techniques and lessons learned from deploying explainable models for several web-scale machine learning and data mining applications. We present case studies across different companies, spanning application domains such as search & recommendation systems, hiring, sales, and lending. Finally, based on our experiences in industry, we identify open problems and research directions for the data mining / machine learning community.
Artificial Intelligence, Machine Learning and Deep LearningSujit Pal
Slides for talk Abhishek Sharma and I gave at the Gennovation tech talks (https://gennovationtalks.com/) at Genesis. The talk was part of outreach for the Deep Learning Enthusiasts meetup group at San Francisco. My part of the talk is covered from slides 19-34.
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Krishnaram Kenthapadi
[Video available at https://sites.google.com/view/ResponsibleAITutorial]
Artificial Intelligence is increasingly being used in decisions and processes that are critical for individuals, businesses, and society, especially in areas such as hiring, lending, criminal justice, healthcare, and education. Recent ethical challenges and undesirable outcomes associated with AI systems have highlighted the need for regulations, best practices, and practical tools to help data scientists and ML developers build AI systems that are secure, privacy-preserving, transparent, explainable, fair, and accountable – to avoid unintended and potentially harmful consequences and compliance challenges.
In this tutorial, we will present an overview of responsible AI, highlighting model explainability, fairness, and privacy in AI, key regulations/laws, and techniques/tools for providing understanding around AI/ML systems. Then, we will focus on the application of explainability, fairness assessment/unfairness mitigation, and privacy techniques in industry, wherein we present practical challenges/guidelines for using such techniques effectively and lessons learned from deploying models for several web-scale machine learning and data mining applications. We will present case studies across different companies, spanning many industries and application domains. Finally, based on our experiences in industry, we will identify open problems and research directions for the AI community.
Spark 2019: Equifax's SVP Data & Analytics, Peter Maynard, discusses the notion (and importance) of explainable AI in the financial services sector. He looks at the work Equifax have done to crack open the black box by creating patented AI technology that helps companies make smarter, explainable decisions using AI.
Artificial Intelligence is increasingly playing an integral role in determining our day-to-day experiences. Moreover, with proliferation of AI based solutions in areas such as hiring, lending, criminal justice, healthcare, and education, the resulting personal and professional implications of AI are far-reaching. The dominant role played by AI models in these domains has led to a growing concern regarding potential bias in these models, and a demand for model transparency and interpretability. In addition, model explainability is a prerequisite for building trust and adoption of AI systems in high stakes domains requiring reliability and safety such as healthcare and automated transportation, as well as critical industrial applications with significant economic implications such as predictive maintenance, exploration of natural resources, and climate change modeling.
As a consequence, AI researchers and practitioners have focused their attention on explainable AI to help them better trust and understand models at scale. The challenges for the research community include (i) defining model explainability, (ii) formulating explainability tasks for understanding model behavior and developing solutions for these tasks, and finally (iii) designing measures for evaluating the performance of models in explainability tasks.
In this tutorial, we will first motivate the need for model interpretability and explainability in AI from societal, legal, customer/end-user, and model developer perspectives. [Note: Due to time constraints, we will not focus on techniques/tools for providing explainability as part of AI/ML systems.] Then, we will focus on the real-world application of explainability techniques in industry, wherein we present practical challenges / implications for using explainability techniques effectively and lessons learned from deploying explainable models for several web-scale machine learning and data mining applications. We will present case studies across different companies, spanning application domains such as search and recommendation systems, sales, lending, and fraud detection. Finally, based on our experiences in industry, we will identify open problems and research directions for the research community.
Differences Between Machine Learning Ml Artificial Intelligence Ai And Deep L...SlideTeam
"You can download this product from SlideTeam.net"
Differences between Machine Learning ML Artificial Intelligence AI and Deep Learning DL is for the mid level managers to give information about what is AI, what is Machine Learning, what is deep learning, Machine learning process. You can also know the difference between Machine learning and Deep learning to understand AI, ML, and DL in a better way for business growth. https://bit.ly/325zI9o
A tremendous backlog of predictive modeling problems in the industry and short supply of trained data scientists have spiked interest in automation over the last few years. A new academic field, AutoML, has emerged. However, there is a significant gap between the topics that are academically interesting and automation capabilities that are necessary to solve real-world industrial problems end-to-end. An even greater challenge is enabling a non-expert to build a robust and trustworthy AI solution for their company. In this talk, we’ll discuss what an industry-grade AutoML system consists of and the scientific and engineering challenges of building it.
This was presented at the London Artificial Intelligence & Deep Learning Meetup.
https://www.meetup.com/London-Artificial-Intelligence-Deep-Learning/events/245251725/
Enjoy the recording: https://youtu.be/CY3t11vuuOM.
- - -
Kasia discussed complexities of interpreting black-box algorithms and how these may affect some industries. She presented the most popular methods of interpreting Machine Learning classifiers, for example, feature importance or partial dependence plots and Bayesian networks. Finally, she introduced Local Interpretable Model-Agnostic Explanations (LIME) framework for explaining predictions of black-box learners – including text- and image-based models - using breast cancer data as a specific case scenario.
Kasia Kulma is a Data Scientist at Aviva with a soft spot for R. She obtained a PhD (Uppsala University, Sweden) in evolutionary biology in 2013 and has been working on all things data ever since. For example, she has built recommender systems, customer segmentations, predictive models and now she is leading an NLP project at the UK’s leading insurer. In spare time she tries to relax by hiking & camping, but if that doesn’t work ;) she co-organizes R-Ladies meetups and writes a data science blog R-tastic (https://kkulma.github.io/).
https://www.linkedin.com/in/kasia-kulma-phd-7695b923/
Artificial Intelligence And Machine Learning PowerPoint Presentation Slides C...SlideTeam
Artificial Intelligence And Machine Learning PowerPoint Presentation Slides arrange insightful data using industry-best design practices. Highlight the differences between machine intelligence, machine learning, and deep learning through our PPT format. Utilize this PowerPoint slideshow to present advantages, disadvantages, learning techniques, and types of supervised machine learning. Further, cover the merits, demerits, and types of unsupervised machine learning. Communicate important details concerning reinforcement learning. Familiarize your viewers with the expert system in artificial intelligence. Outline examples, characteristics, constituents, uses, advantages, drawbacks, and other aspects of the expert system. Compile the deep learning process, recurrent neural networks, and convolutional neural networks through this PowerPoint theme. Present an impactful introduction to artificial intelligence. Introduce kinds, algorithms, trends, and use cases of artificial intelligence. This presentation is not only easy-to-follow but also very convenient to edit, even if you have no prior design experience. Smash the download button and start instant personalization. Our Artificial Intelligence And Machine Learning PowerPoint Presentation Slides Complete Deck are explicit and effective. They combine clarity and concise expression. https://bit.ly/3hKg7PV
Feature Engineering in Machine LearningKnoldus Inc.
In this Knolx we are going to explore Data Preprocessing and Feature Engineering Techniques. We will also understand what is Feature Engineering and its importance in Machine Learning. How Feature Engineering can help in getting the best results from the algorithms.
Machine Learning. What is machine learning. Normal computer vs ML. Types of Machine Learning. Some ML Object detection methods. Faster CNN, RCNN, YOLO, SSD. Real Life ML Applications. Best Programming Languages for ML. Difference Between Machine Learning And Artificial Intelligence. Advantages of Machine Learning. Disadvantages of Machine Learning
Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...Analytics India Magazine
Most organizations understand the predictive power and the potential gains from AIML, but AI and ML are still now a black box technology for them. While deep learning and neural networks can provide excellent inputs to businesses, leaders are challenged to use them because of the complete blind faith required to ‘trust’ AI. In this talk we will use the latest technological developments from researchers, the US defense department, and the industry to unbox the black box and provide businesses a clear understanding of the policy levers that they can pull, why, and by how much, to make effective decisions?
Explainable Artificial Intelligence (XAI)
Presented at Lightning Talk session at ICACCI'18 on 20th September 208
An Explainable AI (XAI) or Transparent AI is an artificial intelligence (AI) whose actions can be easily understood by humans. It contrasts with the concept of the "black box" in machine learning, meaning the "interpretability" of the workings of complex algorithms, where even their designers cannot explain why the AI arrived at a specific decision.
https://en.wikipedia.org/wiki/Explainable_Artificial_Intelligence
Unified Approach to Interpret Machine Learning Model: SHAP + LIMEDatabricks
For companies that solve real-world problems and generate revenue from the data science products, being able to understand why a model makes a certain prediction can be as crucial as achieving high prediction accuracy in many applications. However, as data scientists pursuing higher accuracy by implementing complex algorithms such as ensemble or deep learning models, the algorithm itself becomes a blackbox and it creates the trade-off between accuracy and interpretability of a model’s output.
To address this problem, a unified framework SHAP (SHapley Additive exPlanations) was developed to help users interpret the predictions of complex models. In this session, we will talk about how to apply SHAP to various modeling approaches (GLM, XGBoost, CNN) to explain how each feature contributes and extract intuitive insights from a particular prediction. This talk is intended to introduce the concept of general purpose model explainer, as well as help practitioners understand SHAP and its applications.
[Video recording available at https://www.youtube.com/playlist?list=PLewjn-vrZ7d3x0M4Uu_57oaJPRXkiS221]
Artificial Intelligence is increasingly playing an integral role in determining our day-to-day experiences. Moreover, with proliferation of AI based solutions in areas such as hiring, lending, criminal justice, healthcare, and education, the resulting personal and professional implications of AI are far-reaching. The dominant role played by AI models in these domains has led to a growing concern regarding potential bias in these models, and a demand for model transparency and interpretability. In addition, model explainability is a prerequisite for building trust and adoption of AI systems in high stakes domains requiring reliability and safety such as healthcare and automated transportation, and critical industrial applications with significant economic implications such as predictive maintenance, exploration of natural resources, and climate change modeling.
As a consequence, AI researchers and practitioners have focused their attention on explainable AI to help them better trust and understand models at scale. The challenges for the research community include (i) defining model explainability, (ii) formulating explainability tasks for understanding model behavior and developing solutions for these tasks, and finally (iii) designing measures for evaluating the performance of models in explainability tasks.
In this tutorial, we present an overview of model interpretability and explainability in AI, key regulations / laws, and techniques / tools for providing explainability as part of AI/ML systems. Then, we focus on the application of explainability techniques in industry, wherein we present practical challenges / guidelines for effectively using explainability techniques and lessons learned from deploying explainable models for several web-scale machine learning and data mining applications. We present case studies across different companies, spanning application domains such as search & recommendation systems, hiring, sales, and lending. Finally, based on our experiences in industry, we identify open problems and research directions for the data mining / machine learning community.
Artificial Intelligence, Machine Learning and Deep LearningSujit Pal
Slides for talk Abhishek Sharma and I gave at the Gennovation tech talks (https://gennovationtalks.com/) at Genesis. The talk was part of outreach for the Deep Learning Enthusiasts meetup group at San Francisco. My part of the talk is covered from slides 19-34.
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Krishnaram Kenthapadi
[Video available at https://sites.google.com/view/ResponsibleAITutorial]
Artificial Intelligence is increasingly being used in decisions and processes that are critical for individuals, businesses, and society, especially in areas such as hiring, lending, criminal justice, healthcare, and education. Recent ethical challenges and undesirable outcomes associated with AI systems have highlighted the need for regulations, best practices, and practical tools to help data scientists and ML developers build AI systems that are secure, privacy-preserving, transparent, explainable, fair, and accountable – to avoid unintended and potentially harmful consequences and compliance challenges.
In this tutorial, we will present an overview of responsible AI, highlighting model explainability, fairness, and privacy in AI, key regulations/laws, and techniques/tools for providing understanding around AI/ML systems. Then, we will focus on the application of explainability, fairness assessment/unfairness mitigation, and privacy techniques in industry, wherein we present practical challenges/guidelines for using such techniques effectively and lessons learned from deploying models for several web-scale machine learning and data mining applications. We will present case studies across different companies, spanning many industries and application domains. Finally, based on our experiences in industry, we will identify open problems and research directions for the AI community.
Spark 2019: Equifax's SVP Data & Analytics, Peter Maynard, discusses the notion (and importance) of explainable AI in the financial services sector. He looks at the work Equifax have done to crack open the black box by creating patented AI technology that helps companies make smarter, explainable decisions using AI.
Artificial Intelligence is increasingly playing an integral role in determining our day-to-day experiences. Moreover, with proliferation of AI based solutions in areas such as hiring, lending, criminal justice, healthcare, and education, the resulting personal and professional implications of AI are far-reaching. The dominant role played by AI models in these domains has led to a growing concern regarding potential bias in these models, and a demand for model transparency and interpretability. In addition, model explainability is a prerequisite for building trust and adoption of AI systems in high stakes domains requiring reliability and safety such as healthcare and automated transportation, as well as critical industrial applications with significant economic implications such as predictive maintenance, exploration of natural resources, and climate change modeling.
As a consequence, AI researchers and practitioners have focused their attention on explainable AI to help them better trust and understand models at scale. The challenges for the research community include (i) defining model explainability, (ii) formulating explainability tasks for understanding model behavior and developing solutions for these tasks, and finally (iii) designing measures for evaluating the performance of models in explainability tasks.
In this tutorial, we will first motivate the need for model interpretability and explainability in AI from societal, legal, customer/end-user, and model developer perspectives. [Note: Due to time constraints, we will not focus on techniques/tools for providing explainability as part of AI/ML systems.] Then, we will focus on the real-world application of explainability techniques in industry, wherein we present practical challenges / implications for using explainability techniques effectively and lessons learned from deploying explainable models for several web-scale machine learning and data mining applications. We will present case studies across different companies, spanning application domains such as search and recommendation systems, sales, lending, and fraud detection. Finally, based on our experiences in industry, we will identify open problems and research directions for the research community.
Differences Between Machine Learning Ml Artificial Intelligence Ai And Deep L...SlideTeam
"You can download this product from SlideTeam.net"
Differences between Machine Learning ML Artificial Intelligence AI and Deep Learning DL is for the mid level managers to give information about what is AI, what is Machine Learning, what is deep learning, Machine learning process. You can also know the difference between Machine learning and Deep learning to understand AI, ML, and DL in a better way for business growth. https://bit.ly/325zI9o
A tremendous backlog of predictive modeling problems in the industry and short supply of trained data scientists have spiked interest in automation over the last few years. A new academic field, AutoML, has emerged. However, there is a significant gap between the topics that are academically interesting and automation capabilities that are necessary to solve real-world industrial problems end-to-end. An even greater challenge is enabling a non-expert to build a robust and trustworthy AI solution for their company. In this talk, we’ll discuss what an industry-grade AutoML system consists of and the scientific and engineering challenges of building it.
This was presented at the London Artificial Intelligence & Deep Learning Meetup.
https://www.meetup.com/London-Artificial-Intelligence-Deep-Learning/events/245251725/
Enjoy the recording: https://youtu.be/CY3t11vuuOM.
- - -
Kasia discussed complexities of interpreting black-box algorithms and how these may affect some industries. She presented the most popular methods of interpreting Machine Learning classifiers, for example, feature importance or partial dependence plots and Bayesian networks. Finally, she introduced Local Interpretable Model-Agnostic Explanations (LIME) framework for explaining predictions of black-box learners – including text- and image-based models - using breast cancer data as a specific case scenario.
Kasia Kulma is a Data Scientist at Aviva with a soft spot for R. She obtained a PhD (Uppsala University, Sweden) in evolutionary biology in 2013 and has been working on all things data ever since. For example, she has built recommender systems, customer segmentations, predictive models and now she is leading an NLP project at the UK’s leading insurer. In spare time she tries to relax by hiking & camping, but if that doesn’t work ;) she co-organizes R-Ladies meetups and writes a data science blog R-tastic (https://kkulma.github.io/).
https://www.linkedin.com/in/kasia-kulma-phd-7695b923/
Artificial Intelligence And Machine Learning PowerPoint Presentation Slides C...SlideTeam
Artificial Intelligence And Machine Learning PowerPoint Presentation Slides arrange insightful data using industry-best design practices. Highlight the differences between machine intelligence, machine learning, and deep learning through our PPT format. Utilize this PowerPoint slideshow to present advantages, disadvantages, learning techniques, and types of supervised machine learning. Further, cover the merits, demerits, and types of unsupervised machine learning. Communicate important details concerning reinforcement learning. Familiarize your viewers with the expert system in artificial intelligence. Outline examples, characteristics, constituents, uses, advantages, drawbacks, and other aspects of the expert system. Compile the deep learning process, recurrent neural networks, and convolutional neural networks through this PowerPoint theme. Present an impactful introduction to artificial intelligence. Introduce kinds, algorithms, trends, and use cases of artificial intelligence. This presentation is not only easy-to-follow but also very convenient to edit, even if you have no prior design experience. Smash the download button and start instant personalization. Our Artificial Intelligence And Machine Learning PowerPoint Presentation Slides Complete Deck are explicit and effective. They combine clarity and concise expression. https://bit.ly/3hKg7PV
Feature Engineering in Machine LearningKnoldus Inc.
In this Knolx we are going to explore Data Preprocessing and Feature Engineering Techniques. We will also understand what is Feature Engineering and its importance in Machine Learning. How Feature Engineering can help in getting the best results from the algorithms.
Machine Learning. What is machine learning. Normal computer vs ML. Types of Machine Learning. Some ML Object detection methods. Faster CNN, RCNN, YOLO, SSD. Real Life ML Applications. Best Programming Languages for ML. Difference Between Machine Learning And Artificial Intelligence. Advantages of Machine Learning. Disadvantages of Machine Learning
Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...Analytics India Magazine
Most organizations understand the predictive power and the potential gains from AIML, but AI and ML are still now a black box technology for them. While deep learning and neural networks can provide excellent inputs to businesses, leaders are challenged to use them because of the complete blind faith required to ‘trust’ AI. In this talk we will use the latest technological developments from researchers, the US defense department, and the industry to unbox the black box and provide businesses a clear understanding of the policy levers that they can pull, why, and by how much, to make effective decisions?
Ethical AI: Establish an AI/ML Governance framework addressing Reproducibility, Explainability, Bias & Accountability for Enterprise AI use-cases.
Presentation on “Open Source Enterprise AI/ML Governance” at Linux Foundation’s Open Compliance Summit, Dec 2020 (https://events.linuxfoundation.org/open-compliance-summit/)
Full article: https://towardsdatascience.com/ethical-ai-its-implications-for-enterprise-ai-use-cases-and-governance-81602078f5db
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for EveryoneJames Anderson
If Artificial Intelligence (AI) is a black-box, how can a human comprehend and trust the results of Machine Learning (ML) alogrithms? Explainable AI (XAI) tries to shed light into that AI black-box so humans can trust what is going on. Our speaker Meg Dickey-Kurdziolek is currently a UX Researcher for Google Cloud AI and Industry Solutions, where she focuses her research on Explainable AI and Model Understanding. Recording of the presentation: https://youtu.be/6N2DNN_HDWU
Interpretable Machine Learning_ Techniques for Model Explainability.Tyrion Lannister
In this article, we will explore the importance of interpretable machine learning, its techniques, and its significance in the ever-evolving field of artificial intelligence.
Explainable AI (XAI) stands to address all these challenges and focuses on developing methods and techniques that bring transparency and comprehensibility to AI systems. Its primary objective is to empower users with a clear understanding of the reasoning and logic behind AI algorithms’ decisions.
invited talk in the ExUM workshop in the UMAP 2022 conference
abstract:
Explainability has become an important topic both in Data Science and AI in general and in recommender systems in particular, as algorithms have become much less inherently explainable. However, explainability has different interpretations and goals in different fields. For example, interpretability and explanainability tools in machine learning are predominantly developed for Data Scientists to understand and scrutinize their models. Current tools are therefore often quite technical and not very ‘user-friendly’. I will illustrate this with our recent work on improving the explainability of model-agnostic tools such as LIME and SHAP. Another stream of research on explainability in the HCI and XAI fields focuses more on users’ needs for explainability, such as contrastive and selective explanations and explanations that fit with the mental models and beliefs of the user. However, how to satisfy those needs is still an open question. Based on recent work in interactive AI and machine learning, I will propose that explainability goes together with interactivity, and will illustrate this with examples from our own work in music genre exploration, that combines visualizations and interactive tools to help users understand and tune our exploration model.
Improved Interpretability and Explainability of Deep Learning Models.pdfNarinder Singh Punn
This file aims to give a thorough overview of the current state and future prospects of interpretability and explainability in deep learning, making it a valuable resource for students, researchers, and professionals in the field. The post will comprehensively cover the following aspects:
Introduction to Interpretability and Explainability: Explaining what these concepts mean in the context of deep learning and why they are critical.
The Need for Transparency: Discussing the importance of interpretability and explainability in AI, focusing on ethical considerations, trust in AI systems, and regulatory compliance.
Key Concepts and Definitions: Clarifying terms like “black-box” models, interpretability, explainability, and their relevance in deep learning.
Methods and Techniques:
Visualization Techniques: Detailing methods like feature visualization, attention mechanisms, and tools like Grad-CAM.
Feature Importance Analysis: Exploring techniques like SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-agnostic Explanations) for understanding feature contributions.
Decision Boundary Analysis: Discussing methods to analyze and visualize the decision boundaries of models.
Practical Implementations and Code Examples: Providing examples of how these techniques can be implemented using popular deep learning frameworks like TensorFlow or PyTorch.
Case Studies and Real-World Applications: Presenting real-world scenarios where interpretability and explainability have played a vital role, especially in fields like healthcare, finance, and autonomous systems.
Challenges and Limitations: Addressing the challenges in achieving interpretability and the trade-offs with model complexity and performance.
Future Directions and Research Trends: Discussing ongoing research, emerging trends, and potential future advancements in making deep learning models more interpretable and explainable.
Conclusion: Summarizing the key takeaways and the importance of continued efforts in this area.
References and Further Reading: Providing a list of academic papers, articles, and resources for readers who wish to delve deeper into the topic.
Section 1: Introduction to Interpretability and Explainability
The field of deep learning has witnessed exponential growth in recent years, leading to significant advancements in various applications such as image recognition, natural language processing, and autonomous systems. However, as these neural network models become increasingly complex, they often resemble “black boxes”, where the decision-making process is not transparent or understandable to users. This obscurity raises concerns, especially in critical applications, and underscores the need for interpretability and explainability in deep learning models.
What are Interpretability and Explainability?
Interpretability: This refers to the degree to which a human can understand the cause of a decision made by a machine learning model. It’s about answering the questio
The Ultimate Guide On Difference Between AI And Machine LearningGayathriB80
Discover the difference between AI and Machine Learning in this concise guide. Explore the fundamental dissimilarities, applications, and key features of each.
Data scientists have a duty to ensure they analyze data and train machine learning models responsibly; respecting individual privacy, mitigating bias, and ensuring transparency. This module explores some considerations and techniques for applying responsible machine learning principles.
Practical Explainable AI: How to build trustworthy, transparent and unbiased ...Raheel Ahmad
This presentation is from the Federated & Distributed Machine Learning Conference. This talk focuses on why we need explainable AI and how can we build models that are trustworthy, transparency and unbiased.
The slide has details on below points:
1. Introduction to Machine Learning
2. What are the challenges in acceptance of Machine Learning in Banks
3. How to overcome the challenges in adoption of Machine Learning in Banks
4. How to find new use cases of Machine Learning
5. Few current interesting use cases of Machine Learning
Please contact me (shekup@gmail.com) or connect with me on LinkedIn (https://www.linkedin.com/in/shekup/) for more explanation on ML and how it may help your business.
The slides are inspired by:
Survey & interviews done by me with Bankers & Technology Professionals
Presentation from Google NEXT 2017
Presentation by DATUM on Youtube
Royal Society Machine Learning
Big Data & Social Analytics Course from MIT & GetSmarter
Artificial Intelligence and Machine learning basics.pptxshoeb62
Artificial Intelligence (AI) and Machine Learning (ML) are transformative technologies reshaping industries and revolutionizing our lives. AI simulates human intelligence, while ML enables machines to learn from data without explicit programming. These technologies analyze vast amounts of data to identify patterns and insights, enhancing decision-making in healthcare, finance, transportation, and entertainment.
In healthcare, AI-powered diagnostic tools assist doctors in disease identification and treatment recommendations. In finance, algorithms detect fraud, predict market trends, and optimize investment portfolios. Transportation benefits from AI-driven autonomous vehicles and traffic management systems, improving safety and efficiency. Entertainment platforms use AI to personalize content recommendations, increasing user engagement and retention.
E-commerce businesses leverage AI and ML for inventory management, demand forecasting, and personalized marketing, enhancing customer experiences. Education platforms offer personalized learning materials and resources using ML algorithms, improving educational outcomes and student engagement.
While AI and ML offer immense potential, they also raise ethical and societal concerns, such as data privacy, algorithmic bias, and job displacement. Addressing these challenges requires collaborative efforts from policymakers, industry leaders, and society to ensure responsible and ethical use of these technologies.
In conclusion, AI and ML are driving innovation across various sectors, offering opportunities for advancement and improvement. Embracing these technologies requires a balanced approach that fosters innovation while safeguarding ethical principles, promoting equitable access, and ensuring responsible stewardship for the betterment of humanity.
Data Science Deep Roots in Healthcare IndustryDinesh V
Data Science transforms the healthcare industry with impeccable solutions that can improve patient care through EHRs, medical imaging, drug discovery, predictive medicines and genetics and genomics.
Healthcare evolves with Data InteroperabilityDinesh V
Data Interoperability is the ability to share and work within and across organizational boundaries to advance the effective delivery of healthcare for individuals and communities.
Mastering Customers Moments in Retail Realm Dinesh V
Customer experience can be enhanced in retail industry by implementing advanced innovative yet useful technologies of Artificial Intelligence, Blockchain, etc.
Customers express their opinions in complex ways for businesses. From analyzing user reviews to enhancing the businesses, Sentiment analysis plays a significant role.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
2. Factors driving rapid advancement of AI
Third Wave of AI
SymbolicAI
Logic rules represent
knowledge
No learning capability
and poor handling of
uncertainty
StatisticalAI
Statistical models for
specific domains training
on big data.
No contextual capability
and minimal explainability.
Explainable AI
Systems construct
explanatory models.
Systems learn and
reason with new
tests and situations.
GPUs, On-chip
Neural Network
Data
Availability
Cloud
Infrastructure
New
Algorithms
2
3. There are two ways to provide explainable AI:
• Use Machine learning approaches that are inherently explainable such as decision
trees, knowledge graphs and similarity models.
• Develop new approaches to explain complicated neural networks.
What is XAI?
It is a AI system that explains their decision making which is referred as Explainable AI or
XAI. The goal of XAI is to provide verifiable explanations of how machine learning systems
makes decisions and let humans to be in the loop.
3
4. What is Explainable AI?
Black BoxAI
Data
Black-Box
AI
AI
product
Explainable AI
Explainable
AI
Explainable
AI Product
Explanation
Decision
Feedback
Decision,
Recommendation
Clear & transparent
Predictions
I understand why
I understand why not
I know why you succeed or fail
I understand, so I trust you
Confusion with today’s AI
Black Box
Why did you do that?
Why did you not do that?
When do you succeed or fail?
How do I correct an error?
4
Data
5. Black-box AI Creates Confusion
and Doubt
Can I trust our AI
decisions?
How do I answer this
customer complaint?
Is this the best model
that can be built?
How doI monitor and
debug thismodel?
Why I am getting this
decision?
How can I get a better
decision?
Poor Decision
Black-box
AI
</>
</>
Are these AI system
decisions fair?
Internal Audit,
Regulators
Data Scientists
IT & Operations
Business Owner
5
Customer Support
6. Why Do We Need It?
6
• Artificial Intelligence are increasingly implemented in our everyday lives to assist humans
in making decisions.
• These trivial decisions can vary from a lifestyle choices to more complex decisions such as
loan approvals, investments, court decisions and selection of job candidates.
• Many AI algorithms are a Blackbox that is not transparent. This leads to trustability
concerns. In order to trust these systems, humans want accountability and explanation.
7. Why Do We Need It?
7
• While the machine learning systems deployed in 2008 were mostly within the products
of tech-first companies (i.e. Google, YouTube), the false prediction would result in the
wrong recommendation to the application user.
• But, when it is being deployed in other industries such as military,healthcare, finance, it
would lead to adverse consequences affecting many lives.
• Thus, we create AI systems that explain their decision making.
8. 8
• Why did you do that?
• Why not something else?
• When do you succeed?
• When do you fail?
• When can I trust you?
• How do I correct an error?
We are entering a new
age of AI applications.
Machine learning is the
core technology.
Machine learning
models are opaque,
non-intuitive, and
difficult for people to
understand.
AI System DoD and non-DoD
Application
Transportation
Security
Medicine
Finance
Legal
Military
User
9. Process of XAI
AI is Interpretability. It collaborates between• The significant enabler of explainable
human and artificial intelligence.
• Interpretability is a degree to which a human can understand the cause of a decision.
• It strengthens trust and transparency,explains decisions, fulfil regulatory requirements,
and improve models.
• The stages of AI explainability is categorized into pre-modelling, explainable modelling
and post-modelling. They focus on explainability at the dataset stage and during model
development.
9
10. Explainable
AI
Feedback Loop
Train
Deploy
Monitor
A/B Test
Predict
“Explainability By Design" For AI Products
Model Diagnostics
Root Cause Analytics
Debug
Performance Monitoring
Fairness Monitoring
Model Comparison
Cohort Analysis
Model Debugging
Model Visualization
Model Evaluation
Compliance Testing
QA
Model Launch Signoff
Model Release Mgmt
Explainable Decisions
API Support
10
11. Explainability Approaches
11
• The popular Local Interpretable Model-agnostic Explanations (LIME) approach provides
explanation for an instance prediction of a model in terms of input features, the explanation
family, etc.
• Post-hoc Explainability approach of AI Model creates.
• Individual prediction explanations with input features, influential concepts, local
decision rules.
• The global prediction explanations with partial dependence plots, global
feature importance, global decision rules.
• The build an interpretability model approach creates
• Logistic regressions, decision trees, generalized additive models(GAMs).
12. Why Explainability: Improve ML Model
Standard ML Interpretable ML
Data
ML
model
Model/data
Improvement
Data
ML
model
Predictions
Generalization error Generalization error + Human experience
Verified predictions
Interpretability
Humaninspection
12
13. Explanation Targets
13
• The target specifies the object of an explainability method which varies in type, scope, and
complexity.
• The type of explanation target is often determined according to the role-specific goals of
end users.
• There are two types of targets: inside vs outside, which can also be referred as mechanistic
vs Functional.
• AI experts require a mechanistic explanation of some component inside a model to
understand how layers of a deep network respond to input data in order to debug or
validate the model.
14. Explanation Targets
14
• In contrast, non-experts often require a functional explanation to understand how some
output outside a model is produced.
• In addition, targets can vary in their scope. The outside-type targets are typically some
form of model prediction. They can be either local or global explanations.
• The inside-type targets also vary depending on the architecture of the underlying
model.
• They can either be a single neuron, or layers in a neural network.
15. Explanation Drivers
15
• The most common type of drivers are input features to an AI model.
• Explaining an image classifier predictions in terms of individual input pixels can result in
explanations that are too noisy, too expensive to compute, and more importantly, difficult to
interpret.
• Alternatively, we can rely on a more interpretable representation of input features knownas
super-pixels in the case of image classifier prediction.
• All factors that have an impact on the development of an AI model can be termed as
explanation drivers.
Explainer
(LIME)
16. Explanation Families
16
• A post-hoc explanation aims at communicating some information about how a target is
caused by drivers for a given AI model.
• An explanation family must be chosen such that its information content is easily
interpretable by the user.
• Importance Scores - The individual importance scores are meant to communicate
the relative contribution made by each explanation driver to a given target.
• Decision Rules - Decision trees is where outcome represents prediction of an AI
model and condition is a simple function defined over input features.
17. Explanation Families
17
• Decision Trees - Unlike decision rules, they are structured as a graph where internal
nodes represent conditional tests on input features and leaf nodes represent model out-
comes. In a decision tree each input example can satisfy only one path from the root node
to a leaf node.
• Dependency Plots - They aim at communicating how a target’s value varies as a given
explanation drivers’ value varies, in other words, how a target’s value depends on a
driver’s value.
18. • To explain pre-developed AI models, multiple methods have been proposed.
• They vary in terms of their Explanation target, Explanation drivers, Explanation family
and Extraction mechanism.
• XAI is an active research area with new, improved methods being developed consistently.
• Such diversity of choices can make it challenging of XAI experts to adopt the
most suitable approach for a given application.
• This challenge is addressed by presenting a snapshot of the most notable post-
modelling explainability methods.
Conclusion
18
19. To assist you with our services,
please reach us at
hello@mitosistech.com
www.mitosistech.com
IND: +91-78240 35173
US: +1-(415) 251-2064