Part of the ongoing effort with Skater for enabling better Model Interpretation for Deep Neural Network models presented at the AI Conference.
https://conferences.oreilly.com/artificial-intelligence/ai-ny/public/schedule/detail/65118
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Sri Ambati
This meetup took place in Mountain View on January 24th, 2019.
Description:
With the effort and contributions from researchers and practitioners from academia and industry, Machine Learning Interpretation has become a young sub-field of ML. However, the norms around its definition and understanding is still in its infancy and there are numerous different approaches emerging rapidly. However, there seems to be a lack of a consistent explanation framework to evaluate and consistently benchmark different algorithms - evaluating against interpretation, completeness and consistency of the algorithms.
The idea with the gym is to provide a controlled interactive environment for all forms of Machine Learning algorithms, - initially focusing on supervised predictive modeling problems, to allow analysts and data-scientists to explore, debug and generate insightful understanding of the models by
1.Model Validation: Ways to explore and validate black box ML systems enabling model comparison both globally and locally - identifying biases in the training data through interpretation.
2.What-if Analysis: An interactive environment where communication can happen i.e. enable learning through interactions. User having the ability to conduct "What-If" analysis - effect of single or multiple features and their interactions
3.Model Debugging: Ways to analyze the misbehavior of the model by exploring counterfactual examples(adversarial examples and training)
4. Interpretable Models: Ability to build natively interpretable models - with the goal to simplify complex models to enable better understanding.
The central concept with MLI gym is to have an interactive environment where one could explore and simulate variations in the world(a world post a model is operationalized) beyond the defined model metrics point estimates - e.g. ROC-AUC, confusion matrix, RMSE, R2 score and others.
Speaker's Bio:
Pramit is a Lead Data Scientist/ at H2O.ai. His area of interests is building Statistical/Machine Learning models(Bayesian and Frequentist Modeling techniques) to help the business realize their data-driven goals.
Currently, he is exploring "Model Interpretation" as means to efficiently understand the true nature of predictive models to enable model robustness and security. He believes effective Model Inference coupled with Adversarial training could lead to building trustworthy models with known blind spots. He has started an open source project Skater: https://github.com/datascienceinc/Skater to solve the need for Model Inference(The project is still in its early stages of development but check it out, always eager for feedback)
The importance of model fairness and interpretability in AI systemsFrancesca Lazzeri, PhD
Machine learning model fairness and interpretability are critical for data scientists, researchers and developers to explain their models and understand the value and accuracy of their findings. Interpretability is also important to debug machine learning models and make informed decisions about how to improve them.
In this session, Francesca will go over a few methods and tools that enable you to "unpack” machine learning models, gain insights into how and why they produce specific results, assess your AI systems fairness and mitigate any observed fairness issues.
Using open-source fairness and interpretability packages, attendees will learn how to:
- Explain model prediction by generating feature importance values for the entire model and/or individual data points.
- Achieve model interpretability on real-world datasets at scale, during training and inference.
- Use an interactive visualization dashboard to discover patterns in data and explanations at training time.
- Leverage additional interactive visualizations to assess which groups of users might be negatively impacted by a model and compare multiple models in terms of their fairness and performance.
In this Lunch & Learn session, Chirag Jain gives us a friendly & gentle introduction to Machine Learning & walks through High-Level Learning frameworks using Linear Classifiers.
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Sri Ambati
This meetup took place in Mountain View on January 24th, 2019.
Description:
With the effort and contributions from researchers and practitioners from academia and industry, Machine Learning Interpretation has become a young sub-field of ML. However, the norms around its definition and understanding is still in its infancy and there are numerous different approaches emerging rapidly. However, there seems to be a lack of a consistent explanation framework to evaluate and consistently benchmark different algorithms - evaluating against interpretation, completeness and consistency of the algorithms.
The idea with the gym is to provide a controlled interactive environment for all forms of Machine Learning algorithms, - initially focusing on supervised predictive modeling problems, to allow analysts and data-scientists to explore, debug and generate insightful understanding of the models by
1.Model Validation: Ways to explore and validate black box ML systems enabling model comparison both globally and locally - identifying biases in the training data through interpretation.
2.What-if Analysis: An interactive environment where communication can happen i.e. enable learning through interactions. User having the ability to conduct "What-If" analysis - effect of single or multiple features and their interactions
3.Model Debugging: Ways to analyze the misbehavior of the model by exploring counterfactual examples(adversarial examples and training)
4. Interpretable Models: Ability to build natively interpretable models - with the goal to simplify complex models to enable better understanding.
The central concept with MLI gym is to have an interactive environment where one could explore and simulate variations in the world(a world post a model is operationalized) beyond the defined model metrics point estimates - e.g. ROC-AUC, confusion matrix, RMSE, R2 score and others.
Speaker's Bio:
Pramit is a Lead Data Scientist/ at H2O.ai. His area of interests is building Statistical/Machine Learning models(Bayesian and Frequentist Modeling techniques) to help the business realize their data-driven goals.
Currently, he is exploring "Model Interpretation" as means to efficiently understand the true nature of predictive models to enable model robustness and security. He believes effective Model Inference coupled with Adversarial training could lead to building trustworthy models with known blind spots. He has started an open source project Skater: https://github.com/datascienceinc/Skater to solve the need for Model Inference(The project is still in its early stages of development but check it out, always eager for feedback)
The importance of model fairness and interpretability in AI systemsFrancesca Lazzeri, PhD
Machine learning model fairness and interpretability are critical for data scientists, researchers and developers to explain their models and understand the value and accuracy of their findings. Interpretability is also important to debug machine learning models and make informed decisions about how to improve them.
In this session, Francesca will go over a few methods and tools that enable you to "unpack” machine learning models, gain insights into how and why they produce specific results, assess your AI systems fairness and mitigate any observed fairness issues.
Using open-source fairness and interpretability packages, attendees will learn how to:
- Explain model prediction by generating feature importance values for the entire model and/or individual data points.
- Achieve model interpretability on real-world datasets at scale, during training and inference.
- Use an interactive visualization dashboard to discover patterns in data and explanations at training time.
- Leverage additional interactive visualizations to assess which groups of users might be negatively impacted by a model and compare multiple models in terms of their fairness and performance.
In this Lunch & Learn session, Chirag Jain gives us a friendly & gentle introduction to Machine Learning & walks through High-Level Learning frameworks using Linear Classifiers.
Lecture1 introduction to machine learningUmmeSalmaM1
Machine Learning is a field of computer science which deals with the study of computer algorithms that improve automatically through experience. In this PPT we discuss the following concepts - Prerequisite, Definition, Introduction to Machine Learning (ML), Fields associated with ML, Need for ML, Difference between Artificial Intelligence, Machine Learning, Deep Learning, Types of learning in ML, Applications of ML, Limitations of Machine Learning.
Continual Learning with Deep Architectures - Tutorial ICML 2021Vincenzo Lomonaco
Humans have the extraordinary ability to learn continually from experience. Not only we can apply previously learned knowledge and skills to new situations, we can also use these as the foundation for later learning. One of the grand goals of Artificial Intelligence (AI) is building an artificial “continual learning” agent that constructs a sophisticated understanding of the world from its own experience through the autonomous incremental development of ever more complex knowledge and skills (Parisi, 2019). However, despite early speculations and few pioneering works (Ring, 1998; Thrun, 1998; Carlson, 2010), very little research and effort has been devoted to address this vision. Current AI systems greatly suffer from the exposure to new data or environments which even slightly differ from the ones for which they have been trained for (Goodfellow, 2013). Moreover, the learning process is usually constrained on fixed datasets within narrow and isolated tasks which may hardly lead to the emergence of more complex and autonomous intelligent behaviors. In essence, continual learning and adaptation capabilities, while more than often thought as fundamental pillars of every intelligent agent, have been mostly left out of the main AI research focus.
In this tutorial, we propose to summarize the application of these ideas in light of the more recent advances in machine learning research and in the context of deep architectures for AI (Lomonaco, 2019). Starting from a motivation and a brief history, we link recent Continual Learning advances to previous research endeavours on related topics and we summarize the state-of-the-art in terms of major approaches, benchmarks and key results. In the second part of the tutorial we plan to cover more exploratory studies about Continual Learning with low supervised signals and the relationships with other paradigms such as Unsupervised, Semi-Supervised and Reinforcement Learning. We will also highlight the impact of recent Neuroscience discoveries in the design of original continual learning algorithms as well as their deployment in real-world applications. Finally, we will underline the notion of continual learning as a key technological enabler for Sustainable Machine Learning and its societal impact, as well as recap interesting research questions and directions worth addressing in the future.
Authors: Vincenzo Lomonaco, Irina Rish
Official Website: https://sites.google.com/view/cltutorial-icml2021
My presentation gives a brief overview about soft computing and it's concepts. Such as..Neural networks, Machine learning, Artificial Intelligence etc...
Computational Rationality I - a Lecture at Aalto University by Antti OulasvirtaAalto University
This 2-hour lecture looks at the emerging field of Computational Rationality. Lecture given March 12, 2018, for the Aalto University Master's level course on "Probabilistic Programming and Reinforcement Learning for Cognition and Interaction." Based on: Gershman et al 2015 Science, Lewis et al 2014 Topics in Cog Sci, and Gershman & Daw 2017 Annu Rev Psych
Interpreting complex machine learning models can be difficult. Given an interpretation, its meaningfulness and reliability are hard to evaluate. Even more, depending on the purpose (debugging, ...), a technique in the literature may be more appropriate than others. How to choose the best approach in the landscape of the existing techniques?
This talk is organized as a virtual "walk" through different techniques for interpreting machine learning, and particularly deep learning. Moving from the inside out, we will first cover techniques (such as gradient ascent and deconvolution) for interpreting the internal state of the model, namely its neurons, channels and layer activations. We will then focus on the model behavior from the outside. The model output, for instance, can be explained by attributing the final decisions to subsets of input pixels (as in saliency, occlusion and class activation maps) or to higher-level concepts, such as object size, scale and texture. Concept-based attribution, in particular, has been our research focus over the last years, allowing us to explain deep learning in simple terms to clinicians. For this, digital pathology and retinopathy were our main application domains. In addition, concept-based interpretability helped us explain internal CNN mechanisms such as the encoding of scale and memorization of input-label pairs.
Lecture 2 Basic Concepts in Machine Learning for Language TechnologyMarina Santini
Definition of Machine Learning
Type of Machine Learning:
Classification
Regression
Supervised Learning
Unsupervised Learning
Reinforcement Learning
Supervised Learning:
Supervised Classification
Training set
Hypothesis class
Empirical error
Margin
Noise
Inductive bias
Generalization
Model assessment
Cross-Validation
Classification in NLP
Types of Classification
Introductory presentation to Explainable AI, defending its main motivations and importance. We describe briefly the main techniques available in March 2020 and share many references to allow the reader to continue his/her studies.
Machine Learning for Dummies (without mathematics)ActiveEon
It presents an introduction and the basic concepts of machine learning without mathematics. This is a short presentation for beginners in machine learning.
Deep Learning has taken the digital world by storm. As a general purpose technology, it is now present in all walks of life. Although the fundamental developments in methodology have been slowing down in the past few years, applications are flourishing with major breakthroughs in Computer Vision, NLP and Biomedical Sciences. The primary successes can be attributed to the availability of large labelled data, powerful GPU servers and programming frameworks, and advances in neural architecture engineering. This combination enables rapid construction of large, efficient neural networks that scale to the real world. But the fundamental questions of unsupervised learning, deep reasoning, and rapid contextual adaptation remain unsolved. We shall call what we currently have Deep Learning 1.0, and the next possible breakthroughs as Deep Learning 2.0.
This is part 2 of the Tutorial delivered at IEEE SSCI 2020, Canberra, December 1st (Virtual).
Lecture1 introduction to machine learningUmmeSalmaM1
Machine Learning is a field of computer science which deals with the study of computer algorithms that improve automatically through experience. In this PPT we discuss the following concepts - Prerequisite, Definition, Introduction to Machine Learning (ML), Fields associated with ML, Need for ML, Difference between Artificial Intelligence, Machine Learning, Deep Learning, Types of learning in ML, Applications of ML, Limitations of Machine Learning.
Continual Learning with Deep Architectures - Tutorial ICML 2021Vincenzo Lomonaco
Humans have the extraordinary ability to learn continually from experience. Not only we can apply previously learned knowledge and skills to new situations, we can also use these as the foundation for later learning. One of the grand goals of Artificial Intelligence (AI) is building an artificial “continual learning” agent that constructs a sophisticated understanding of the world from its own experience through the autonomous incremental development of ever more complex knowledge and skills (Parisi, 2019). However, despite early speculations and few pioneering works (Ring, 1998; Thrun, 1998; Carlson, 2010), very little research and effort has been devoted to address this vision. Current AI systems greatly suffer from the exposure to new data or environments which even slightly differ from the ones for which they have been trained for (Goodfellow, 2013). Moreover, the learning process is usually constrained on fixed datasets within narrow and isolated tasks which may hardly lead to the emergence of more complex and autonomous intelligent behaviors. In essence, continual learning and adaptation capabilities, while more than often thought as fundamental pillars of every intelligent agent, have been mostly left out of the main AI research focus.
In this tutorial, we propose to summarize the application of these ideas in light of the more recent advances in machine learning research and in the context of deep architectures for AI (Lomonaco, 2019). Starting from a motivation and a brief history, we link recent Continual Learning advances to previous research endeavours on related topics and we summarize the state-of-the-art in terms of major approaches, benchmarks and key results. In the second part of the tutorial we plan to cover more exploratory studies about Continual Learning with low supervised signals and the relationships with other paradigms such as Unsupervised, Semi-Supervised and Reinforcement Learning. We will also highlight the impact of recent Neuroscience discoveries in the design of original continual learning algorithms as well as their deployment in real-world applications. Finally, we will underline the notion of continual learning as a key technological enabler for Sustainable Machine Learning and its societal impact, as well as recap interesting research questions and directions worth addressing in the future.
Authors: Vincenzo Lomonaco, Irina Rish
Official Website: https://sites.google.com/view/cltutorial-icml2021
My presentation gives a brief overview about soft computing and it's concepts. Such as..Neural networks, Machine learning, Artificial Intelligence etc...
Computational Rationality I - a Lecture at Aalto University by Antti OulasvirtaAalto University
This 2-hour lecture looks at the emerging field of Computational Rationality. Lecture given March 12, 2018, for the Aalto University Master's level course on "Probabilistic Programming and Reinforcement Learning for Cognition and Interaction." Based on: Gershman et al 2015 Science, Lewis et al 2014 Topics in Cog Sci, and Gershman & Daw 2017 Annu Rev Psych
Interpreting complex machine learning models can be difficult. Given an interpretation, its meaningfulness and reliability are hard to evaluate. Even more, depending on the purpose (debugging, ...), a technique in the literature may be more appropriate than others. How to choose the best approach in the landscape of the existing techniques?
This talk is organized as a virtual "walk" through different techniques for interpreting machine learning, and particularly deep learning. Moving from the inside out, we will first cover techniques (such as gradient ascent and deconvolution) for interpreting the internal state of the model, namely its neurons, channels and layer activations. We will then focus on the model behavior from the outside. The model output, for instance, can be explained by attributing the final decisions to subsets of input pixels (as in saliency, occlusion and class activation maps) or to higher-level concepts, such as object size, scale and texture. Concept-based attribution, in particular, has been our research focus over the last years, allowing us to explain deep learning in simple terms to clinicians. For this, digital pathology and retinopathy were our main application domains. In addition, concept-based interpretability helped us explain internal CNN mechanisms such as the encoding of scale and memorization of input-label pairs.
Lecture 2 Basic Concepts in Machine Learning for Language TechnologyMarina Santini
Definition of Machine Learning
Type of Machine Learning:
Classification
Regression
Supervised Learning
Unsupervised Learning
Reinforcement Learning
Supervised Learning:
Supervised Classification
Training set
Hypothesis class
Empirical error
Margin
Noise
Inductive bias
Generalization
Model assessment
Cross-Validation
Classification in NLP
Types of Classification
Introductory presentation to Explainable AI, defending its main motivations and importance. We describe briefly the main techniques available in March 2020 and share many references to allow the reader to continue his/her studies.
Machine Learning for Dummies (without mathematics)ActiveEon
It presents an introduction and the basic concepts of machine learning without mathematics. This is a short presentation for beginners in machine learning.
Deep Learning has taken the digital world by storm. As a general purpose technology, it is now present in all walks of life. Although the fundamental developments in methodology have been slowing down in the past few years, applications are flourishing with major breakthroughs in Computer Vision, NLP and Biomedical Sciences. The primary successes can be attributed to the availability of large labelled data, powerful GPU servers and programming frameworks, and advances in neural architecture engineering. This combination enables rapid construction of large, efficient neural networks that scale to the real world. But the fundamental questions of unsupervised learning, deep reasoning, and rapid contextual adaptation remain unsolved. We shall call what we currently have Deep Learning 1.0, and the next possible breakthroughs as Deep Learning 2.0.
This is part 2 of the Tutorial delivered at IEEE SSCI 2020, Canberra, December 1st (Virtual).
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for EveryoneJames Anderson
If Artificial Intelligence (AI) is a black-box, how can a human comprehend and trust the results of Machine Learning (ML) alogrithms? Explainable AI (XAI) tries to shed light into that AI black-box so humans can trust what is going on. Our speaker Meg Dickey-Kurdziolek is currently a UX Researcher for Google Cloud AI and Industry Solutions, where she focuses her research on Explainable AI and Model Understanding. Recording of the presentation: https://youtu.be/6N2DNN_HDWU
A tremendous backlog of predictive modeling problems in the industry and short supply of trained data scientists have spiked interest in automation over the last few years. A new academic field, AutoML, has emerged. However, there is a significant gap between the topics that are academically interesting and automation capabilities that are necessary to solve real-world industrial problems end-to-end. An even greater challenge is enabling a non-expert to build a robust and trustworthy AI solution for their company. In this talk, we’ll discuss what an industry-grade AutoML system consists of and the scientific and engineering challenges of building it.
Highlights and summary of long-running programmatic research on data science; practices, roles, tools, skills, organization models, workflow, outlook, etc. Profiles and persona definition for data scientist model. Landscape of org models for data science and drivers for capability planning. Secondary research materials.
Data pipelines are the heart and soul of data science. Are you a beginner looking to understand data pipelines? A glimpse into what they are and how they work.
Data Science & AI Road Map by Python & Computer science tutor in MalaysiaAhmed Elmalla
The slides were used in a trial session for a student aiming to learn python to do Data science projects .
The session video can be watched from the link below
https://youtu.be/CwCe1pKOVI8
I have over 20 years of experience in both teaching & in completing computer science projects with certificates from Stanford, Alberta, Pennsylvania, California Irvine universities.
I teach the following subjects:
1) IGCSE A-level 9618 / AS-Level
2) AP Computer Science exam A
3) Python (basics, automating staff, Data Analysis, AI & Flask)
4) Java (using Duke University syllabus)
5) Descriptive statistics using SQL
6) PHP, SQL, MYSQL & Codeigniter framework (using University of Michigan syllabus)
7) Android Apps development using Java
8) C / C++ (using University of Colorado syllabus)
Check Trial Classes:
1) A-Level Trial Class : https://youtu.be/v3k7A0nNb9Q
2) AS level trial Class : https://youtu.be/wj14KpfbaPo
3) 0478 IGCSE class : https://youtu.be/sG7PrqagAes
4) AI & Data Science class: https://youtu.be/CwCe1pKOVI8
https://elmalla.info/blog/68-tutor-profile-slide-share
You can get your trial Class now by booking : https://calendly.com/ahmed-elmalla/30min
And you can contact me on
https://wa.me/0060167074241
by Python & Computer science tutor in Malaysia
MLOps and Data Quality: Deploying Reliable ML Models in ProductionProvectus
Looking to build a robust machine learning infrastructure to streamline MLOps? Learn from Provectus experts how to ensure the success of your MLOps initiative by implementing Data QA components in your ML infrastructure.
For most organizations, the development of multiple machine learning models, their deployment and maintenance in production are relatively new tasks. Join Provectus as we explain how to build an end-to-end infrastructure for machine learning, with a focus on data quality and metadata management, to standardize and streamline machine learning life cycle management (MLOps).
Agenda
- Data Quality and why it matters
- Challenges and solutions of Data Testing
- Challenges and solutions of Model Testing
- MLOps pipelines and why they matter
- How to expand validation pipelines for Data Quality
Transform your career with our Data Science course in Hyderabad. Master machine learning, Python, big data analysis, and data visualization. Our training and expert mentors prepare you for high-demand roles, making you a sought-after data scientist in Hyderabad's tech scene.
Transform your career with our Data Science course in Hyderabad. Master machine learning, Python, big data analysis, and data visualization. Our training and expert mentors prepare you for high-demand roles, making you a sought-after data scientist in Hyderabad's tech scene.
Transform your career with our Data Science course in Hyderabad. Master machine learning, Python, big data analysis, and data visualization. Our training and expert mentors prepare you for high-demand roles, making you a sought-after data scientist in Hyderabad's tech scene.
Similar to Model evaluation in the land of deep learning (20)
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
StarCompliance is a leading firm specializing in the recovery of stolen cryptocurrency. Our comprehensive services are designed to assist individuals and organizations in navigating the complex process of fraud reporting, investigation, and fund recovery. We combine cutting-edge technology with expert legal support to provide a robust solution for victims of crypto theft.
Our Services Include:
Reporting to Tracking Authorities:
We immediately notify all relevant centralized exchanges (CEX), decentralized exchanges (DEX), and wallet providers about the stolen cryptocurrency. This ensures that the stolen assets are flagged as scam transactions, making it impossible for the thief to use them.
Assistance with Filing Police Reports:
We guide you through the process of filing a valid police report. Our support team provides detailed instructions on which police department to contact and helps you complete the necessary paperwork within the critical 72-hour window.
Launching the Refund Process:
Our team of experienced lawyers can initiate lawsuits on your behalf and represent you in various jurisdictions around the world. They work diligently to recover your stolen funds and ensure that justice is served.
At StarCompliance, we understand the urgency and stress involved in dealing with cryptocurrency theft. Our dedicated team works quickly and efficiently to provide you with the support and expertise needed to recover your assets. Trust us to be your partner in navigating the complexities of the crypto world and safeguarding your investments.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
1. Learn more at datascience.com | Empower Your Data Scientists
Model evaluation in the land of deep learning
May 2, 2018
2. Learn more at datascience.com | Empower Your Data Scientists 2
Host
● Lead data scientist at DataScience.com.
● Currently, exploring better ways to extract, evaluate, and explain the learned decision policies of
predictive models. Recently started an open source project, Skater, to improve the process of model
interpretation to enable better model evaluation and model security.
● Before joining DataScience.com, used machine learning algorithms to find love for eHarmony
customers.
Pramit Choudhary
@MaverickPramit
www.linkedin.com/in/pramitc/
pramit@datascience.com
3. Learn more at datascience.com | Empower Your Data Scientists
Agenda
● Understand the problem of model opacity
● Define the “what” and “why” of model interpretation
● Define the scope of model interpretation
● Introduce Skater
● How to interpreting Deep Neural Networks(DNNs) model to improve Model
Performance with Skater?
● Demo
● Ability to generate and prevent adversarial attacks with Skater
● Q&A
● References
4. Learn more at datascience.com | Empower Your Data Scientists
The Problem of Model Opacity
(Y|X)
Training Set
{xi
∈ X, yi
∈ Y}
Why I am getting
weird predictions?
Was my model
biased?
I am not 100%
sure what’s in
the box; I didn’t
build the model.
“By 2018, half of business ethics violations will occur through improper use of
big data analytics.” —
Source: https://www.gartner.com/newsroom/id/3144217
PredictorBlack Box Model
Holdout/Test Set
{xi
∈ X, yi
∈ Y}
5. Learn more at datascience.com | Empower Your Data Scientists 5
A collection of visual and/or interactive artifacts that provide
a user with sufficient description of the model behavior to
accurately perform tasks like evaluation, trusting, predicting,
or improving the model.
Assistant Professor Sameer Singh,
University of California, Irvine
QUOTE
___
6. Learn more at datascience.com | Empower Your Data Scientists
What is Model Interpretation?
● Ability to explain and present a model in a way that is understandable to
humans.
● A Model’s result is self descriptive and needs no further explanation;
expressed in terms of input and output.
7. Empower Your Data Scientists
Why is Model Interpretation Important?
Producer / Model Maker:
● Data scientist/analyst building a model
● Consultants helping clients
Consumer / Model Breaker:
● Business owners or data engineers
● Risk/security assessment managers
● Humans being affected by the model
“Explain the model.”
8. Learn more at datascience.com | Empower Your Data Scientists
Ideas collapse.
9. Learn more at datascience.com | Empower Your Data Scientists 9
While model interpretation is a hard problem, it's within the
role of the data scientist to guide the other stakeholders
through different levels of interpretation, recognize the
caveats, highlight ambiguities, etc
Paco Nathan,
Director Learning Group, O’Reilly
QUOTE
___
10. Learn more at datascience.com | Empower Your Data Scientists 10
Motives for Model Interpretation
1. Debugging and improving an ML system.
2. Exploring and discovering latent or hidden feature
interactions (useful for feature engineering/selection
and resolving preconceptions).
3. Understanding model variability.
4. Helps in model comparison.
5. Building domain knowledge about a particular use
case.
6. Bring transparency to decision making to enable
trust.
1. Explain the model/algorithm.
2. Explain the key features driving the KPI.
3. Verify and validate the accountability of ML
learning systems, e.g., causes for false positives in
credit scoring, insurance claim frauds.
4. Identify blind spots to prevent adversarial attacks
or fix dataset errors.
5. Ability to share the explanations to consumers of
the predictive model.
6. Comply with data protection regulations, e.g., EU’s
GDPR.
● Data Scientist
● Machine Learning Engineer
● Data Analyst
● Statistician
● Data Science Manager
● Business Owner
● Data Engineer
● Auditors/Risk Managers
Producer/Model Maker Consumer/Model Breaker
11. Learn more at datascience.com | Empower Your Data Scientists
With model interpretation, we want to answer the following questions:
○ Why did the model behave in a certain way?
○ What was the reason for false positives? What are the relevant variables driving a model’s
outcome, e.g., customer lifetime value, fraud detection, image classification, spam
detection?
○ How can we trust the predictions of a “black box” model? Is the predictive model biased?
How can we guarantee model’s security against adversarial attacks?
What Do We Want to Achieve?
12. Learn more at datascience.com | Empower Your Data Scientists
Focusing on supervised learning problems.
13. Learn more at datascience.com | Empower Your Data Scientists
Scope of Interpretation
Global Interpretation
Being able to explain the conditional interaction
between dependent (response) variables and
independent (predictor or explanatory) variables
based on the complete dataset
Global
Interpretation
Local Interpretation
Local Interpretation
Being able to explain the conditional interaction
between dependent (response) variables and
independent (predictor or explanatory) variables
with respect to a single prediction
14. Learn more at datascience.com | Empower Your Data Scientists
How Do We Enable Model Interpretation?
Reference: Kim, Been & Doshi-Velez, Finale. (ICML 2017). Interpretable Machine
Learning:
The fuss, the concrete and the questions
( http://people.csail.mit.edu/beenkim/papers/BeenK_FinaleDV_ICML2017_tutorial.pdf )
Visual question answering (VQA): Is this a case of
distracted driving?
Relevance: Insurance fraud
Top predictions include p(seat-belt)=0.75,
p(limousine)=0.051, and p(golf cart) = 0.017
15. Learn more at datascience.com | Empower Your Data Scientists
Introducing Skater
https://github.com/datascienceinc/Skater
If you like the idea, follow
our journey!
Gitter Channel (join us here):
https://gitter.im/datascienceinc-skater
/Lobby
A unified framework to enable Model Interpretation both globally (on the basis of a complete data
set) and locally (in-regards to an individual prediction)
16. Learn more at datascience.com | Empower Your Data Scientists
Machine Learning Workflow
Define
Hypothesis
Use relevant key
performance
indicators
Handle Data
Handle Missing
Data
Data Partitioning
Engineer and
Select
Features
Transform data
Select relevant
features
Build Model
Build a predictive
model
Deploy Model
Operationalize
analytics as
scalable REST APIs
Test and Monitor
Model
1. Log and track behavior
2. Evaluate
3. Conduct A/B or
multi-armed bandit testing
1 2 3 4 5 6
Model Interpretation: In-Memory Models
● Model assessment
● Explain model at a global and local level
● Publish insights, make collaborative and
informed decisions
Model Interpretation: Deployed Models
● Explore and explain model behavior
● Debug and discover errors to improve
performance
RETRAIN
EVALUATE
Improve existing hypothesis or generate a new one
17. Learn more at datascience.com | Empower Your Data Scientists
An Interpretable Machine Learning System
Interpretability with Rule
Extraction
18. Learn more at datascience.com | Empower Your Data Scientists
Skater’s goal
Model Performance (Accuracy)
Interpretability/
Degree of
Opacity
Deep learning models
Support vector machines
Random forest
K-nearest neighbors
Linear/Logistic Regression
Decision trees
Bayesian Rule List
XGBoost
Note: The purpose of the chart is not to mirror any benchmark on model performance, but to articulate the opacity of predictive models
Enable trust and faithfulness by
ability to infer, debug and
understand
19. Learn more at datascience.com | Empower Your Data Scientists
Deep Neural Networks
● Helps build expressive and flexible models by
learning arbitrary non-linear and non-convex
functions easily
● Can be expressed in different architectures;
optimizing for accuracy and computation efficiency
often leads to complex designs
● Need for manual feature engineering is less; lower
layers can extract complex features
● With advancement in software -
Keras/Tensorflow/MXNet and hardware-better
integration with GPU, it’s easier to train DNN’s over
billions of data points optimizing over large
number of parameters.
● But, models are often perceived as black boxes
because of lack of tools to infer them
neurons/
units
connectors
Input Layer
Output Layer
Hidden
Layer 1
Hidden
Layer 2
20. Learn more at datascience.com | Empower Your Data Scientists
Modern DNNs with complex designs
Optimizing on accuracy, has made Modern DNN architectures complex and often difficult to interpret
and understand as humans
Image Source: Canziani, Alfredo, Paszke, Adam & Culurciello, Eugenio (2016). An Analysis of Deep Neural Network Models for Practical Applications
21. Learn more at datascience.com | Empower Your Data Scientists
Layer-wise Relevance Propagation (LRP)
● Decomposes the predictions of a deep neural network to pixel level relevance scores using first
order approximation
● Initially proposed by Bach S. et. al (2015) https://doi.org/10.1371/journal.pone.0130140
● Computed as a backward pass using a modified gradient from the output layer to the input layer
● Skater supports e-LRP (a version of LRP) as proposed by Ancona M., Ceolini E., Cengiz Ö., Gross
M. (2018) in Towards better understanding of gradient-based attribution methods for Deep Neural
Networks using chain rule with a modified gradient
● Scope of Interpretation: Local Interpretation
● Framework supported by Skater:
22. Learn more at datascience.com | Empower Your Data Scientists
Integrated Gradient
● Computes relevance score for Deep Networks for Image and Text using first order approximation
● Proposed by Sundararajan, Mukund, Taly, Ankur & Yan, Qibi (2017) in Axiomatic Attribution for
Deep Networks
● Implementation adopted as suggested by Ancona M., Ceolini E., Cengiz Ö., Gross M. (2018) in
Towards better understanding of gradient-based attribution methods for Deep Neural Networks
● Determines relevance (contribution) of an input X = {x1
, x2
, … xn
} ∈ Rn
relative to baseline input X’
● Compute the average gradient while the input varies along a linear path from a baseline ’ to
● Baseline ’: for Image: ; for Text: zero embedding vector
● Statisfies sensitivity and implementation Invariance
● Scope of Interpretation: Local Interpretation
● Framework supported by Skater:
23. Learn more at datascience.com | Empower Your Data Scientists
Demo
Evaluating Deep Neural Networks with Skater
1. CNN on MNIST dataset
2. Imagenet with pre-trained Inception-V3
3. CNN/LSTM sentiment analysis with IMDB dataset
24. Learn more at datascience.com | Empower Your Data Scientists
Evaluate Model Stability
Figure: An MNIST experiment with CNN model with 98.8% train and 98.6 test
‘Accuracy’. Interpretation CNN model with ‘ReLU activation using e-LRP. Image-6 on the
left is in-correctly classified as 0. Skater provides the ability to infer the cause of
mis-classification ( Pixels colored in Red have a positive influence and Blue negative
influence ). Images share a semantic properties globally. In the above example we can
see 6 and 0 sharing semantic properties around the lower curvy round ‘O’, probably the
reason for misclassification.
Train/Test set “Accuracy”:
98.8/98.6
25. Learn more at datascience.com | Empower Your Data Scientists
Identifying blind spots
Distortion(D)
X + r ∈ [0,
1]m
X ∈ Rm
, where R is a image vector, mapped to a discrete label set, L ∈ {1 … k}. Interpreting CNN model
on MNIST dataset.
Relevant Image pixels
are retained and
correctly identified
26. Learn more at datascience.com | Empower Your Data Scientists
Creating Adversarial Examples
27. Learn more at datascience.com | Empower Your Data Scientists
Conditional adversarial tested against pre-trained Inception-V3
28. Learn more at datascience.com | Empower Your Data Scientists
More Examples
sports car: 0.54%
grille: 0.21%
washbasin: 0.43%
ping-pong ball:
0.14%
29. Learn more at datascience.com | Empower Your Data Scientists
What about text classification?
Trained Model: CNN for sentiment analysis
Dataset: IMDB
For a trained DNN classifier model(F),
● Craft an adversarial attack by adding
perturbation ΔX
X* = X + ΔX
● ΔX = <Insert, delete, replace>
● F(X) ≄ F(X*)
X =
excellent tale of two boys that do whatever they can to get away from there abusive father lord of the rings star elijah wood is
outstanding in this unforgettable role this movie is one of the main reasons i haven't touched a single beer and never will as
long as i live that might make me sound like a nerd but that's what i have to say it is a wonder why this isn't as a classic
american tale
30. Learn more at datascience.com | Empower Your Data Scientists
● Replace [“outstanding”, “excellent”, “classic”, “unforgettable”, “touched”, “never”] with
“<UNK>”
<UNK> tale of two boys that do whatever they can to get away from there abusive father lord of the rings star elijah wood is
<UNK> in this <UNK> role this movie is one of the main reasons i haven't <UNK> a single beer and <UNK> will as long as i live
that might make me sound like a nerd but that's what i have to say it is a wonder why this isn't as a <UNK> american tale
Negative
signals positive
signals
31. Learn more at datascience.com | Empower Your Data Scientists
Evaluate
(Y|X)
Data
Data
Unboxed model
Evaluate
Partial dependence plot
Relative variable importance
Local Interpretable Model
Explanation (LIME)
R or Python model (linear, nonlinear, ensemble, neural networks)
Scikit-learn, caret and rpart packages for CRAN
H20.ai, Algorithmia, etc.
WITHOUT INTERPRETATION ...
WITH SKATER ...
Black box model
How do I understand my
models?
Bayesian rule list (BRL)
DNNs - Integrated Gradient
DNNs - e-Layerwise Relevance
Propagation (e-LRP)
32. Learn more at datascience.com | Empower Your Data Scientists
Special Thanks
● Marco Ancona, Researcher at ETH Zurich-Department of Computer Science, for helping us in
enabling the journey of supporting interpretation for DNNs in Skater
● O’Reilly and AI Conference for allowing us to share our thoughts with all you guys
33. Learn more at datascience.com | Empower Your Data Scientists
Q&A
info@datascience.com
@DataScienceInc
@MaverickPramit
34. Learn more at datascience.com | Empower Your Data Scientists
Future Work and Improvement
● Add better support for generating adversarial examples
● Define and support ways to quantitatively evaluate Model Interpretation
● Implement feature perturbation using Occlusion.
● Add support for other frameworks - PyTorch/MXNet/CNTK
● Explore concept recognition. Xie, Sarker, Doran, Hitzler, & Raymer. (2017). Relating Input
Concepts to Convolutional Neural Network Decisions. arXiv:1711.08006
● More experimentation and notebook examples on inferring DNNs
35. Learn more at datascience.com | Empower Your Data Scientists
References
● Szegedy C. et al. (2014). Intriguing properties of neural networks. arXiv:1312.6199
● Basch S. et al. (2015). On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise
Relevance Propagation. PloS one, 10(7):e0130140
● Canziani A., Paszke A., Culurciello E. (2016). An Analysis of Deep Neural Network Models for
Practical Applications. arXiv:1605.07678
● Jain S., Ghosh R. ( 2017). Visualizing Deep Learning Networks. arXiv:1605.07678
● Liang B. et al. (2017) Deep Text Classification Can be Fooled. arXiv:1704.08006
● Sundararajan M., Taly A., Yan Q. (2017) Axiomatic Attribution for Deep Networks. arXiv:1703.01365
● Ancona M., Ceolini E., Cengiz Ö., Gross M. (2018). Towards better understanding of gradient-based
attribution methods for Deep Neural Networks. arXiv: 1711.06104
● Olah, C. et al. (2018). The Building Blocks of Interpretability.
36. Learn more at datascience.com | Empower Your Data Scientists
END