SlideShare a Scribd company logo
1 of 60
Download to read offline
HOW DO WE
TRAIN AI TO BE
ETHICAL AND UNBIASED?
MARK BORG
AI MALTA SUMMIT – 13 JULY 2018
RECENT ACHIEVEMENTS IN AI
2
Word
Error
rate
Improvements in word error rate over time on the Switchboard
conversational speech recognition benchmark.
Credit: Awni Hannun
Automated Speech Recognition results
Credit: Business Insider/Yu Han
RECENT ACHIEVEMENTS IN AI
3
Credit: H. Fang et al. (2015), “From Captions to Visual Concepts and Back”
#1 A woman holding a
camera in a crowd.
Image Captioning
RECENT ACHIEVEMENTS IN AI
4
0 days
AlphaGo Zero has no prior knowledge of the
game and only the basic rules as an input.
3 days
AlphaGo Zero surpasses the abilities of AlphaGo
Lee, the version that beat world champion Lee
Sedol in 4 out of 5 games in 2016.
21 days
AlphaGo Zero reaches the level of AlphaGo
Master, the version that defeated 60 top
professionals online and world champion Ke Jie in
3 out of 3 games in 2017.
40 days
AlphaGo Zero surpasses
all other versions of
AlphaGo and, arguably,
becomes the best Go
player in the world. It
does this entirely from
self-play, with no human
intervention and using no
historical data.
Credit: DeepMind
AlphaGo Zero
WIDESPREAD USE OF AI
• AI has now wide and deep societal influences, permeating every sphere of our lives
• No longer single applications operating in standalone mode
• ML Pipelines, more complex AI systems, operating at Internet Scale
• AI as a Service (AIaaS), Machine Learning as a Service (MLaaS)
• Running “under the hood”, as well as in “human-facing technology”
• High-stake applications, sometimes involving life-and-death decisions
➢ AI-enabled Future
➢ Benefits and Implications
5
BENEFITS AND CONCERNS OF AI
6
• What if an AI algorithm could predict death better than doctors?
• The “dying algorithm” (NY Times)
• Stanford's AI Predicts Death for Better End-of-Life Care (IEEE Spectrum)
• What are the benefits and implications of such a system?
CONCERNS
• A Predictive Policing algorithm unfairly targeted
certain neighbourhoods – Chicago 2013/2014
• Idea: to stop crime before it occurs
• Unintended consequences due to systematic bias in the
data used by these systems
• Saunders et al. (2016), “Predictions put into practice: a quasi-
experimental evaluation of Chicago’s predictive policing project”
• COMPAS assesses a defendant’s risk of re-offending
• used for bail determination by judges
• Issues of reliability and racial bias
• Dressel & Farid (2018), “The Accuracy, Fairness, and Limits of Predicting
Recividism”
7Credit: ProPublica
CONCERNS
• YouTube Recommender system
• The algorithm appears to have concluded that
people are drawn to content that is more
extreme than what they started with — or to
incendiary content in general
• Accusations that YouTube is acting as a
“radicalisation agent”
8
Credit: Covington
Recommendations drive 70%
of YouTube’s viewing time
(~200 million
recommendations per day)
YouTube tops a cumulative of 1
billion hours of video per day in
2017
CONCERNS
• Adversarial AI
9Credit: IBM
Credit: Biggio & Roli
CONCERNS
• Ethical and moral issues
• Self driving cars
10
The Trolley Problem
Credit: Waymo
(Philippa Foot, 1967)
LONG-TERM CONCERNS
• GAI, Superintelligence, existential threat, need for Benevolent AI
• The Sorcerer’s Apprentice problem
• Eliezer Yudkowsky: The Paperclip Maximiser Scenario
11
Credit: Disney
If a machine can think,
it might think more
intelligently than we do,
and then where should we
be? …
This new danger … is
certainly something which
can give us anxiety
Alan Turing, 1951
“
“
IMPLICATIONS & CONSEQUENCES OF AI
• To maximise the benefits of AI: saving lives, raising the quality of life, …
Need also to address issues and consequences
• the “rough edges of AI” – Eric Horvitz (Microsoft Research)
• Robustness, Ethics, Benevolent AI
• Short-term implications (need solving now)
• Longer term implications (prepare the groundwork…)
• Spans multiple fields: engineering, cognitive science, philosophy, etc.
12
13
AIES
ICAILEP
Conference on Artificial Intelligence:
Law, Ethics, and Policy
7008 - Standard for Ethically Driven Nudging for Robotic, Intelligent & Autonomous Systems
7009 - Standard for Fail-Safe Design of Autonomous & Semi-Autonomous Systems
7010 - Wellbeing Metrics Standard for Ethical Artificial Intelligence & Autonomous Systems
IMPLICATIONS & CONSEQUENCES OF AI
14
Benevolent AI
AI Safety
Robust AI Beneficial AI
Value Alignment
AI Ethics
Roboethics
Machine Ethics
Adversarial AI
Increasedcomplexity
AI transparency
ANI
Artificial Narrow Intelligence
AGI
Artificial General Intelligence
ASI
Artificial Super Intelligence
15
LADDER OF
CAUSATION
Credit: Judea Pearl (2018),
“The Book of Why: The New
Science of Cause and Effect”
AI SAFETY
• Data Bias (Algorithmic Bias)
• Fairness
• AI Robustness & Reliability
• AI Transparency
16
DATA BIAS
• Algorithmic Bias is NOT model bias (bias-variance trade-off, generalisation problem)
• Algorithmic Bias (or Data Bias) – will always be present; need to minimise the impact
• E.g. predictive policing algorithm
• Police-recorded datasets suffer from systematic bias:
• Non-complete census
• Not a representative random sample
• Crime databases do not measure crime; they measure some complex interaction
between criminality, policing strategy, and community-police relationships
17
DATA BIAS
• Data bias is prevalent throughout the whole field of AI
• Unintentional bias vs. intentional bias
• Addressing data bias has particular significance in ML pipelines, complex AI systems,
AIaaS, etc.
• E.g.
• Howard (2017), “Addressing Bias in Machine Learning Algorithms: A Pilot Study on
Emotion Recognition for Intelligent Systems”
• did not perform well for children
• original training dataset had few such cases
18
DATA BIAS
• Unintentional self-created bias (“poisoning your own data”)
• E.g. Google Flu Trends
• began suggesting flu-related queries to people who did not have the flu, and
thus Google Flu Trends began itself corrupting the dataset by seeding it with
excess flu-related queries, thus creating a feedback loop
• Despite good intentions, biased data can lead to a far worse result
• E.g. beauty.ai
• a startup organising the world's first AI-driven beauty contest in 2016
• The concept is to remove social biases of human judges
• problem: image samples used to train the algorithms weren’t balanced in terms
of race and ethnicity.
• so-called 'white guy problem’
19
DATA BIAS
• Naive application of algorithms to everyday problems
could amplify structural discrimination and reproduce
biases present in the data
• Detecting (automatically?) such bias and addressing it?
• Quite difficult!
• Since AI is data-driven, it’s difficult
20
Credit: Buolamwini & Gebru
• Some very recent work on two fronts:
• More balanced datasets, e.g., new facial image dataset released in February 2018 (Pilot Parliaments
Benchmark dataset)
• Buolamwini & Gebru (2018), “Gender Shades: Intersectional Accuracy Disparities in Commercial Gender
Classification”
• Measuring bias and fairness:
• Shaikh et al. (2017), “An End-to-End Machine Learning Pipeline that Ensures Fairness Policies”
• Srivastava & Rossi (2018), “Towards Composable Bias Rating of AI Services”
DATA BIAS
• creating more balanced (heterogeneous) datasets
• One solution would be to create shared and regulated databases that are in
possession of no single entity, thus preventing any party from unilaterally manipulating
the data to their own favour.
• public datasets curated to be bias free
• One concern is that even when machine-learning systems are programmed to
be blind to race or gender, for example, they may use other signals in data
such as the location of a person’s home as a proxy for it
• E.g. In COMPAS bail system, geographic neighbourhood highly correlated to ethnicity,
thus still suffering from racial discrimination
21
AI ROBUSTNESS & RELIABILITY
• Making AI systems more robust, so that they work as intended, without failing
or getting misused?
• Reliable prediction of performance
• Avoiding overconfidence in AI systems
• How much it knows that it does not know?
• Make strong predictions that are just inaccurate
• Classification label accuracy + ROC curve
• Learning to predict confidence
• Current statistical models “tend to assume that the data that they’ll see in the future will
look a lot like the data they’ve seen in the past”
22
AI ROBUSTNESS & RELIABILITY
• Blindspots of algorithms (Eric Horvitz, Microsoft Research)
• The “unknown unknowns” (Tom Dietterich , Oregon State University)
• AI algorithms for learning and acting safely in the presence of unknown unknowns.
• learning about blindspots of algorithms:
• Lakkaraju (2016), “Identifying Unknown Unknowns in the Open World: Representations
and Policies for Guided Exploration”
• Ramakrishnan (2018), “Discovering Blind Spots in Reinforcement Learning”
• Human supervision
• human correction to prevent AI failure
23
AI ROBUSTNESS & RELIABILITY
• Watch out for Anomalies?
• Robust anomaly detection
• BugID system by Tom Dietterich,
• A system that learn when there is another unknown class out there
• automated counting of freshwater macro-invertebrates
• trained on 29 insect classes, detection of novel classes
• monitoring of performance of the system, especially for self-learning systems or local-
based learning
• E.g. Microsoft’s chatbot Tay
• research into adding a “reflection layer” into systems (introspection?)
• Failsafe designs
• Auto-pilot of self-driving cars disengaging suddenly
24
MISUSE OF AI
• Privacy Challenges
• Exclusion – denying services
• Persuasion, and Manipulation of Attention / Behaviour / Beliefs
• Harms
• Hacking of AI systems
• Adversarial AI
25
MISUSE OF AI
• Tay
• Microsoft’s Chatbot
• "The more you talk the smarter Tay gets!"
• March 2016, suspended after 16 hours
• Tay’s conversation extended to racist, inflammatory and political statements
• A main problem was Tay’s “repeat after me” feature
• Intentional misuse of AI (coordinated attack)
• Neff and Nagy (2016), “Talking to Bots: Symbiotic Agency and the Case of Tay”
26
Credit: Microsoft
MISUSE OF AI
• Harnessing AI to increase attention & engagement for a particular application or service
• large-scale personalised targeting
• Persuasion, and Manipulation of Attention / Behaviour / Beliefs
• auto twitter feed generation that persuades a user to click on links
• data-driven behaviour change
• Intentional / Unintentional
27
YOUTUBE RECOMMENDER SYSTEM
• The recommender system’s goal is to
maximise attention and engagement via
personalised targeting
• Eric Horvitz (Microsoft Research) calls this:
"Adversarial Attacks on Attention"
28
Credit: Eric Horvitz
Recommendations drive
70% of YouTube’s viewing
time (~200 million
recommendations per day)
YouTube tops a cumulative of
1 billion hours of video per
day in 2017
Recommendation system architecture demonstrating the “funnel”
where candidate videos are retrieved and ranked before
presenting only a few to the user
Covington et al. (2016), “Deep Neural Networks for YouTube
Recommendations”
YOUTUBE RECOMMENDER SYSTEM
• Its algorithm seems to have concluded that people are drawn to content that is more
extreme than what they started with — or to incendiary content in general
• a bias toward extreme/divisive/inflammatory/fringe/sensational content
• WSJ investigation (Feb 2018):
• amplifies human bias, fake news, isolate users in "filter bubbles"
• AlgoTransparency.org
• Zeynep Tufekci (sociologist, Univ. of North Carolina):
• Calls YouTube the “Great Radicaliser”
• AI exploiting a natural human desire to "look behind the curtain", to dig deeper into something that
engages us. As we click and click, we are carried along by the exciting sensation of uncovering more
secrets and deeper truths. YouTube leads viewers down a rabbit hole of extremism, while Google racks
up the ad sales.
29
YOUTUBE RECOMMENDER SYSTEM
• But is the algorithm really to blame?
• Main issue due to scale
• Also simplified human behaviour modelling:
• Watching more nuanced content or videos which diverge from the
established viewing pattern could be rooted out as noise, thus contributing
to simplification and generalization of interests towards the more extreme
ends of a spectrum, instead of complex content catering to views which
are harder to define.
• Possible solutions?
• YouTube has been applying changes to their algorithm
• Improved human behaviour models
• Changes to the exploration-exploitation strategy adopted by the
recommender system
• Value policies, encoding notion of “time well spent”
30
ADVERSARIAL AI
• Goodfellow et al. (2015), “Explaining and Harnessing Adversarial Examples”
• Szegedy (2013), “Traversing the manifold to find blind spots in the input space”
• DNN can be easily fooled by adversaries
• No need for hand-crafting the adversarial attack
• Can exploit AI to perform an adversarial attack
• One AI deceiving another AI
31
“panda”
57.7% confidence
“gibbon”
99.3% confidence
Adversarial Noise
(exaggerated)
ADVERSARIAL AI
• Adversarial systems subtly alter normal inputs such that humans doing the same task can
easily recognize what the intended input is, but mislead the AI into giving a predictable
and very different false output.
• Performed by stealth (humans won’t spot the difference)
• Potential Attacks
• Adversarial examples can be printed out on standard paper, then photographed with a
standard smartphone, and it will still fool AI systems.
• Kurakin et al. (2017), “Adversarial examples in the physical world”
32
Credit: Biggio & Roli
ADVERSARIAL AI
• The famous 3D printed Turtle that fooled
Google’s AI
• Adversarial attacks without perturbing the
whole image
33
Athalye et al. (2017), “Synthesizing Robust Adversarial Examples”
• Sharif et al. (2016), “Accessorize to a Crime: Real and
Stealthy Attacks on State-of-the-Art Face Recognition”
• Impersonation attacks
• Invisibility attacks
Credit: Sharif et al. (2016)
ADVERSARIAL AI
• Audio Adversarial attacks
• Carlini and Wagner (2018), “Audio Adversarial Examples: Targeted Attacks on Speech-to-Text”
• Given any speech audio, can produce another that is 99.9% similar to the original, but contains any text
one wants.
• Fools DeepSpeech with 100% accuracy
34Credit: IBM
ADVERSARIAL AI
• Not limited to Deep Neural Networks
• Papernot et al. (2016), “Transferability in machine learning: from phenomena to black-box
attacks using adversarial samples”
• DNNs, logistic regression, support vector machines, decision trees, nearest neighbour
classifiers, ensembles – all vulnerable to adversarial AI!
• any machine learning classifier can be tricked to give incorrect predictions, and with a little
bit of work, one can get them to give pretty much any result one wants
35
ADVERSARIAL AI
36
Adversarial AI
panda57.7%
Defended AI model
+
gibbon100%
ADVERSARIAL AI
• White-box adversarial attack
37
Adversarial AI
panda57.7%
Defended AI model
gradient
+
gibbon100%
Score-based
attack
Substitute AI model
ADVERSARIAL AI
• Black-box adversarial attack
38
Adversarial AI
panda
Defended AI model
+
gibbon
Transfer-based
attack
Decision-based
attack
scores & gradients
Some countermeasures:
DEFENDING AGAINST ADVERSARIAL AI
39
• Smoothing and hiding the gradients
• Randomisation techniques
• image compression
• Image blurring
• random image resizing
• employ dropout in neural networks
• Defensive distillation
• Use of ensembles
• Evaluate model’s adversarial resilience
• Metrics available
• Pre-emptive hardening of AI models
• Enhance robustness to tampering
Some defensive techniques:
IBM Adversarial Robustness Toolbox (ART)
https://github.com/IBM/adversarial-robustness-toolbox
Cleverhans library
https://github.com/openai/cleverhans
DeepFool
https://github.com/LTS4/DeepFool
• Fredrikson et al. (2015), “Model Inversion Attacks that Exploit Confidence Information and Basic
Countermeasures”
• Violating privacy of subjects in the training set
ADVERSARIAL AI – MODEL INVERSION ATTACKS
40
Adversarial AI
Tom
2.3%
70%
Training Set
Tom
Tom
Defended AI model
Face Recognition
• Leveraging adversarial AI to make a generative model, consisting of two neural networks
competing with each other
• The discriminator tries to distinguish genuine data from forgeries created by the generator
• The generator turns random noise into imitations of the data, in an attempt to fool the
discriminator
GENERATIVE ADVERSARIAL NETWORKS (GANS)
41
real
DiscriminatorGenerator
random
noise
fake
real fake
AI ETHICS & VALUE ALIGNMENT
• Codification of Ethics
• Values, Utility Functions
• Teaching AI to be Ethical
• Reinforcement Learning
• Inverse Reinforcement Learning and beyond
▪ Ethics – comprehending “right” from “wrong”, and behaving in a right way
▪ Value Alignment – ensuring that the goals, behaviours, values and ethics of autonomous
AI systems align with those of humans
42
CODIFICATION OF ETHICS
• Rule-based ethics (deontological ethics)
• Isaac Asimov’s “Three Laws of Robotics”
• And similar sets of rules
43
• Challenges:
• Too rigid
• Asimov’s literature addresses many of these issues:
• conflicts between the 3 laws, conflicts within a law by itself, conflicting orders, etc.
• How to codify the rules?
• How to program the notion of "harm"?
• Often human ethics and values are implicit
• Process of elicitation is very challenging
Isaac Asimov (1942)
CODIFICATION OF ETHICS
• Pre-programming ethical rules:
• Impossible to program for every scenario
• Fail to address uncertainty and randomness
• Fail to address ambiguous cases, ethical and moral dilemmas
• Rules on their own not enough
• Must be accompanied by very strong accountability mechanisms
• Need moral conflict resolution mechanism
• Values and ethics are dependent on the socio-cultural context
• Difficult to standardise
• Need to account for changes in the values of society, shifts in beliefs, attitudes, etc.
44
CODIFICATION OF ETHICS
45
• Rule-based ethics example:
• specifically & explicitly programme ethical values into self-driving
cars to prioritize the protection of human life above all else
• In the event of an unavoidable accident, the car should be
“prohibited to offset victims against one another”
• A car must not choose whether to kill a person based on individual
features, when a fatal crash is inescapable
Credit: BMVI (www.bmvi.de) The Trolley Problem
VALUES, UTILITY FUNCTIONS
• Ethics as Utility Functions
• Any system or person who acts or gives advice is using some value system of what is important and
what is not
• Utility-based Agent
• Agent’s Actions
• Agent’s Beliefs
• Agent’s Preferences
• The agent chooses actions based on their outcomes
• Outcomes are what the agent has preference on
• Preferences → Utility → Utility Function
• A policy specifies what an agent should do under all contingencies
• An agent wants to find an optimal policy – one that maximises its expected utility
46
TEACHING AI TO BE ETHICAL
• Teaching AI ethics, social rules and norms
• Adopt a “blank slate” approach
• Similar to how a human child learns ethics from those around him/her
• Basic values are learnt, and the AI will, in time, be able to apply those principles in unforeseen scenarios
• What machine learning method to use?
47Credit: GoodAI
TEACHING AI TO BE ETHICAL
48
• Reinforcement Learning
• Has shown promise in learning policies that can solve complex problems
• An agent explores its environment, performing action after action and receiving rewards and punishments
according to the reward function (i.e. utility function)
• As it repeats this, the agent will gradually learn to perform the right actions in the right states so as to
maximise its reward
• Reward = total sum of the actions’ rewards over
time, where future rewards are discounted (treated
as less valuable than present rewards)
• When learning ethics, the reward function will
reward/punish the agent depending on the choice of
action performed, whether “right” or “wrong”
Environment
Model
Reward
Function(s)
Reinforcement
Learning
Reward-
maximising
behaviour
Kose (2017), “Ethical Artificial Intelligence – An Open Question”
TEACHING AI TO BE ETHICAL
• Reinforcement Learning (RL) challenges:
• Difficulty in setting up ethical scenarios in the environment model of RL
• May take a very long time till the agent manages to fully cover all ethical scenarios, ambiguous cases, etc.
49
• Potential solution:
• Using stories as a way of short-circuiting the
reinforcement learning process
• Employ more complex stories as time goes by
• Riedl et al. (2016), “Using Stories to Teach Human
Values to Artificial Agents”
Environment
Model
Reward
Function(s)
Reinforcement
Learning
Kose (2017), “Ethical Artificial Intelligence – An Open Question”
Reward-
maximising
behaviour
TEACHING AI TO BE ETHICAL
• Another solution:
• Curriculum-based approach to improve the learning process
• The learning process in humans and animals is enhanced when scenarios are not randomly presented, but
organized in a meaningful order – gradual exposure to an increasing number of concepts, and to more
complex ones
• For teaching ethics, simpler scenarios are presented before more complex and ambiguous cases
• GoodAI’s “School for AI” project is employing a curriculum based approach for enhancing the teaching of
ethics via reinforcement learning
• www.goodai.com/school-for-ai
• Bengio et al. (2009), “Curriculum Learning”
• Weinshall et al. (2018), “Curriculum Learning by Transfer Learning: Theory and Experiments with Deep
Networks”
50
TEACHING AI TO BE ETHICAL
• Crowd-Sourcing Ethics and Morality
• Crowdsourced stories simplify the manually-intensive process of creating stories
• Can capture consensus for ambiguous and moral dilemmas (“wisdom of the crowds”)
• Example:
• An AI agent is given several hundred stories about stealing versus not stealing, explores different actions in a
reinforcement learning setting, and learns the consequences and optimal policy based on the rewards /
punishments given. (Mark Riedl, Georgia Tech)
51
TEACHING AI TO BE ETHICAL
• MIT’s “Moral Machine”:
• Crowdsourcing to aid self-driving cars
make better moral decisions in cases
of moral dilemmas (variations of the
Trolley problem)
• http://moralmachine.mit.edu
52
TEACHING AI TO BE ETHICAL
• Reinforcement Learning (RL) requires the manual specification of the reward function
• “Reward engineering” is hard (especially for ethics)
• May be susceptible to “reward cheating” by the AI agent
53
In RL, the reward function is specified by the
user, and then the agent does the acting
What if the agent could instead watch
someone else do the acting, and try to come
up with the reward function by itself?
Environment
Model
Reward
Function(s)
Reinforcement
Learning
Provided by the user
Reward-
maximising
behaviour
TEACHING AI TO BE ETHICAL
• Inverse Reinforcement Learning (IRL)
• IRL is able to learn the underlying reward function
(what is ethical?) from expert demonstrations
(humans solving ethical problems)
• IRL is also called “imitation-based learning”
• Learn from watching good behaviour
54
Reward-
maximising
behaviour
Environment
Model
Reward
Function(s)
Reinforcement
Learning
Provided by the user
Environment
Model
Reward
Function(s)
Inverse
Reinforcement
Learning
Provided by the user
Observed
Behaviour
Reward-
maximising
behaviour
Reinforcement
Learning
TEACHING AI TO BE ETHICAL
• Inverse Reinforcement Learning (IRL)
• Very promising results for AI ethics (value alignment)
• No need to explicitly model rules, reward function
• Recent works advocating IRL:
• Russell et al. (2016), “Research Priorities for Robust and
Beneficial Artificial Intelligence”
• Abel (2016), “Reinforcement Learning as a Framework for
Ethical Decision Making”
• Challenges of IRL:
• Interpretability of the auto-learnt reward function
• Human bias can creep into the observed behaviour
• Difficulty of learnt ethics to be domain independent
• Arnold (2017), “Value Alignment or Misalignment – What Will
Keep Systems Accountable?”
55
Environment
Model
Reward
Function(s)
Inverse
Reinforcement
Learning
Provided by the user
Observed
Behaviour
Reward-
maximising
behaviour
Reinforcement
Learning
BEYOND IRL…
• Cooperative IRL
• What if we reward both the “good behaviour” of the AI while
learning ethics, as well as reward the “good teaching” of the
human?
• Cooperation between AI and humans to accomplish a shared
goal – value alignment
• Generative Adversarial Networks (GANs)
• Hadfield-Menell (2016), “Cooperative Inverse Reinforcement
Learning”
56
Environment
Model
Reward
Function(s)
Inverse
Reinforcement
Learning
Provided by the user
Observed
Behaviour
Reward-
maximising
behaviour
Reinforcement
Learning
BEYOND IRL…
• Harnessing Counterfactuals
• … “imagination” rung on the ladder of causation
• As perfect knowledge of the world is unavailable, counterfactuals
allow for the revision of one’s belief system and the relying solely
on past (data driven) experience
• It is also through counterfactuals that one ultimately enters into
social appraisals of blame and praise
• Might prove to be one of the key technologies needed both for
the advancement of AI itself on the trajectory towards AGN, as
well as for aligning as much as possible the values of machines
with our values to achieve benevolent AI
57
BENEVOLENT AI
58
Value Alignment
AI values
our values
BENEVOLENT AI
59
our values
AI values
mutually
beneficial
values
Value Alignment
Everything we love about civilisation is a product of intelligence, so amplifying our
human intelligence with artificial intelligence has the potential of helping civilisation
flourish like never before – as long as we manage to keep the technology beneficial.
Max Tegmark, Cosmologist & President of the Future of Life Institute
“
“
Thank you
60

More Related Content

What's hot

Fairness in Machine Learning and AI
Fairness in Machine Learning and AIFairness in Machine Learning and AI
Fairness in Machine Learning and AISeth Grimes
 
AI Governance and Ethics - Industry Standards
AI Governance and Ethics - Industry StandardsAI Governance and Ethics - Industry Standards
AI Governance and Ethics - Industry StandardsAnsgar Koene
 
Fairness and Bias in Machine Learning
Fairness and Bias in Machine LearningFairness and Bias in Machine Learning
Fairness and Bias in Machine LearningSurya Dutta
 
Generative AI Risks & Concerns
Generative AI Risks & ConcernsGenerative AI Risks & Concerns
Generative AI Risks & ConcernsAjitesh Kumar
 
Bias in AI-systems: A multi-step approach
Bias in AI-systems: A multi-step approachBias in AI-systems: A multi-step approach
Bias in AI-systems: A multi-step approachEirini Ntoutsi
 
Measures and mismeasures of algorithmic fairness
Measures and mismeasures of algorithmic fairnessMeasures and mismeasures of algorithmic fairness
Measures and mismeasures of algorithmic fairnessManojit Nandi
 
Algorithmic Bias - What is it? Why should we care? What can we do about it?
Algorithmic Bias - What is it? Why should we care? What can we do about it? Algorithmic Bias - What is it? Why should we care? What can we do about it?
Algorithmic Bias - What is it? Why should we care? What can we do about it? University of Minnesota, Duluth
 
Fairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsFairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsKrishnaram Kenthapadi
 
Technology for everyone - AI ethics and Bias
Technology for everyone - AI ethics and BiasTechnology for everyone - AI ethics and Bias
Technology for everyone - AI ethics and BiasMarion Mulder
 
The Ethics of Artificial Intelligence
The Ethics of Artificial IntelligenceThe Ethics of Artificial Intelligence
The Ethics of Artificial IntelligenceKarl Seiler
 
Ethical issues facing Artificial Intelligence
Ethical issues facing Artificial IntelligenceEthical issues facing Artificial Intelligence
Ethical issues facing Artificial IntelligenceRah Abdelhak
 
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Krishnaram Kenthapadi
 
Responsible AI
Responsible AIResponsible AI
Responsible AINeo4j
 
AIF360 - Trusted and Fair AI
AIF360 - Trusted and Fair AIAIF360 - Trusted and Fair AI
AIF360 - Trusted and Fair AIAnimesh Singh
 

What's hot (20)

Fairness in Machine Learning and AI
Fairness in Machine Learning and AIFairness in Machine Learning and AI
Fairness in Machine Learning and AI
 
AI Governance and Ethics - Industry Standards
AI Governance and Ethics - Industry StandardsAI Governance and Ethics - Industry Standards
AI Governance and Ethics - Industry Standards
 
Fairness and Bias in Machine Learning
Fairness and Bias in Machine LearningFairness and Bias in Machine Learning
Fairness and Bias in Machine Learning
 
Bias in AI
Bias in AIBias in AI
Bias in AI
 
Generative AI Risks & Concerns
Generative AI Risks & ConcernsGenerative AI Risks & Concerns
Generative AI Risks & Concerns
 
Ethics and AI
Ethics and AIEthics and AI
Ethics and AI
 
Bias in AI-systems: A multi-step approach
Bias in AI-systems: A multi-step approachBias in AI-systems: A multi-step approach
Bias in AI-systems: A multi-step approach
 
Measures and mismeasures of algorithmic fairness
Measures and mismeasures of algorithmic fairnessMeasures and mismeasures of algorithmic fairness
Measures and mismeasures of algorithmic fairness
 
Algorithmic Bias - What is it? Why should we care? What can we do about it?
Algorithmic Bias - What is it? Why should we care? What can we do about it? Algorithmic Bias - What is it? Why should we care? What can we do about it?
Algorithmic Bias - What is it? Why should we care? What can we do about it?
 
Fairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsFairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML Systems
 
Implementing Ethics in AI
Implementing Ethics in AIImplementing Ethics in AI
Implementing Ethics in AI
 
AI
AIAI
AI
 
Technology for everyone - AI ethics and Bias
Technology for everyone - AI ethics and BiasTechnology for everyone - AI ethics and Bias
Technology for everyone - AI ethics and Bias
 
The Ethics of Artificial Intelligence
The Ethics of Artificial IntelligenceThe Ethics of Artificial Intelligence
The Ethics of Artificial Intelligence
 
Ethical issues facing Artificial Intelligence
Ethical issues facing Artificial IntelligenceEthical issues facing Artificial Intelligence
Ethical issues facing Artificial Intelligence
 
Algorithmic fairness
Algorithmic fairnessAlgorithmic fairness
Algorithmic fairness
 
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
 
Responsible AI
Responsible AIResponsible AI
Responsible AI
 
Responsible AI
Responsible AIResponsible AI
Responsible AI
 
AIF360 - Trusted and Fair AI
AIF360 - Trusted and Fair AIAIF360 - Trusted and Fair AI
AIF360 - Trusted and Fair AI
 

Similar to How do we train AI to be Ethical and Unbiased?

Ethics for Conversational AI
Ethics for Conversational AIEthics for Conversational AI
Ethics for Conversational AIVerena Rieser
 
Designing AI for Humanity at dmi:Design Leadership Conference in Boston
Designing AI for Humanity at dmi:Design Leadership Conference in BostonDesigning AI for Humanity at dmi:Design Leadership Conference in Boston
Designing AI for Humanity at dmi:Design Leadership Conference in BostonCarol Smith
 
iConference 2018 BIAS workshop keynote
iConference 2018 BIAS workshop keynoteiConference 2018 BIAS workshop keynote
iConference 2018 BIAS workshop keynoteAnsgar Koene
 
AI Ethics and Implications For Developing Societies.pptx
AI Ethics and Implications For Developing Societies.pptxAI Ethics and Implications For Developing Societies.pptx
AI Ethics and Implications For Developing Societies.pptxIshaku Gayus Bwala
 
A Glimpse Into the Future of Data Science - What's Next for AI, Big Data & Ma...
A Glimpse Into the Future of Data Science - What's Next for AI, Big Data & Ma...A Glimpse Into the Future of Data Science - What's Next for AI, Big Data & Ma...
A Glimpse Into the Future of Data Science - What's Next for AI, Big Data & Ma...Pangea.ai
 
AI Ethical Framework.pptx
AI Ethical Framework.pptxAI Ethical Framework.pptx
AI Ethical Framework.pptxDavid Atkinson
 
20240104 HICSS Panel on AI and Legal Ethical 20240103 v7.pptx
20240104 HICSS  Panel on AI and Legal Ethical 20240103 v7.pptx20240104 HICSS  Panel on AI and Legal Ethical 20240103 v7.pptx
20240104 HICSS Panel on AI and Legal Ethical 20240103 v7.pptxISSIP
 
Trusted, Transparent and Fair AI using Open Source
Trusted, Transparent and Fair AI using Open SourceTrusted, Transparent and Fair AI using Open Source
Trusted, Transparent and Fair AI using Open SourceAnimesh Singh
 
Ethical Dimensions of Artificial Intelligence (AI) by Rinshad Choorappara
Ethical Dimensions of Artificial Intelligence (AI) by Rinshad ChoorapparaEthical Dimensions of Artificial Intelligence (AI) by Rinshad Choorappara
Ethical Dimensions of Artificial Intelligence (AI) by Rinshad ChoorapparaRinshad Choorappara
 
The Ethics of Artificial Intelligence in Digital Ecosystems
The Ethics of Artificial Intelligence in Digital EcosystemsThe Ethics of Artificial Intelligence in Digital Ecosystems
The Ethics of Artificial Intelligence in Digital Ecosystemswashikmaryam
 
KIVI Innovation Drinks - Presentation Philip Brey.pdf
KIVI Innovation Drinks - Presentation Philip Brey.pdfKIVI Innovation Drinks - Presentation Philip Brey.pdf
KIVI Innovation Drinks - Presentation Philip Brey.pdfSreyaseeDasBhattacha1
 
Ethical AI - Open Compliance Summit 2020
Ethical AI - Open Compliance Summit 2020Ethical AI - Open Compliance Summit 2020
Ethical AI - Open Compliance Summit 2020Debmalya Biswas
 
Artificial Intelligence: The Next 5(0) Years
Artificial Intelligence: The Next 5(0) YearsArtificial Intelligence: The Next 5(0) Years
Artificial Intelligence: The Next 5(0) YearsMarlon Dumas
 
Algorithmic fairness
Algorithmic fairnessAlgorithmic fairness
Algorithmic fairnessAnthonyMelson
 

Similar to How do we train AI to be Ethical and Unbiased? (20)

Ethics for Conversational AI
Ethics for Conversational AIEthics for Conversational AI
Ethics for Conversational AI
 
Designing AI for Humanity at dmi:Design Leadership Conference in Boston
Designing AI for Humanity at dmi:Design Leadership Conference in BostonDesigning AI for Humanity at dmi:Design Leadership Conference in Boston
Designing AI for Humanity at dmi:Design Leadership Conference in Boston
 
iConference 2018 BIAS workshop keynote
iConference 2018 BIAS workshop keynoteiConference 2018 BIAS workshop keynote
iConference 2018 BIAS workshop keynote
 
AI Ethics and Implications For Developing Societies.pptx
AI Ethics and Implications For Developing Societies.pptxAI Ethics and Implications For Developing Societies.pptx
AI Ethics and Implications For Developing Societies.pptx
 
A Glimpse Into the Future of Data Science - What's Next for AI, Big Data & Ma...
A Glimpse Into the Future of Data Science - What's Next for AI, Big Data & Ma...A Glimpse Into the Future of Data Science - What's Next for AI, Big Data & Ma...
A Glimpse Into the Future of Data Science - What's Next for AI, Big Data & Ma...
 
AI Ethical Framework.pptx
AI Ethical Framework.pptxAI Ethical Framework.pptx
AI Ethical Framework.pptx
 
RAPIDE
RAPIDERAPIDE
RAPIDE
 
20240104 HICSS Panel on AI and Legal Ethical 20240103 v7.pptx
20240104 HICSS  Panel on AI and Legal Ethical 20240103 v7.pptx20240104 HICSS  Panel on AI and Legal Ethical 20240103 v7.pptx
20240104 HICSS Panel on AI and Legal Ethical 20240103 v7.pptx
 
Trusted, Transparent and Fair AI using Open Source
Trusted, Transparent and Fair AI using Open SourceTrusted, Transparent and Fair AI using Open Source
Trusted, Transparent and Fair AI using Open Source
 
Ethical Dimensions of Artificial Intelligence (AI) by Rinshad Choorappara
Ethical Dimensions of Artificial Intelligence (AI) by Rinshad ChoorapparaEthical Dimensions of Artificial Intelligence (AI) by Rinshad Choorappara
Ethical Dimensions of Artificial Intelligence (AI) by Rinshad Choorappara
 
inte
inteinte
inte
 
The Ethics of Artificial Intelligence in Digital Ecosystems
The Ethics of Artificial Intelligence in Digital EcosystemsThe Ethics of Artificial Intelligence in Digital Ecosystems
The Ethics of Artificial Intelligence in Digital Ecosystems
 
KIVI Innovation Drinks - Presentation Philip Brey.pdf
KIVI Innovation Drinks - Presentation Philip Brey.pdfKIVI Innovation Drinks - Presentation Philip Brey.pdf
KIVI Innovation Drinks - Presentation Philip Brey.pdf
 
Tecnologías emergentes: priorizando al ciudadano
Tecnologías emergentes: priorizando al ciudadanoTecnologías emergentes: priorizando al ciudadano
Tecnologías emergentes: priorizando al ciudadano
 
#AI: In Whose Image?
#AI: In Whose Image?#AI: In Whose Image?
#AI: In Whose Image?
 
Ethical AI - Open Compliance Summit 2020
Ethical AI - Open Compliance Summit 2020Ethical AI - Open Compliance Summit 2020
Ethical AI - Open Compliance Summit 2020
 
Artificial Intelligence: The Next 5(0) Years
Artificial Intelligence: The Next 5(0) YearsArtificial Intelligence: The Next 5(0) Years
Artificial Intelligence: The Next 5(0) Years
 
AI Forum-2019_Nakagawa
AI Forum-2019_NakagawaAI Forum-2019_Nakagawa
AI Forum-2019_Nakagawa
 
Ai titech-virach-20191026
Ai titech-virach-20191026Ai titech-virach-20191026
Ai titech-virach-20191026
 
Algorithmic fairness
Algorithmic fairnessAlgorithmic fairness
Algorithmic fairness
 

Recently uploaded

Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...apidays
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Jeffrey Haguewood
 
FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024The Digital Insurer
 
Six Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal OntologySix Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal Ontologyjohnbeverley2021
 
Platformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityPlatformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityWSO2
 
Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Zilliz
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingEdi Saputra
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...apidays
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoffsammart93
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherRemote DBA Services
 
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...Zilliz
 
ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesrafiqahmad00786416
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfsudhanshuwaghmare1
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyKhushali Kathiriya
 
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Angeliki Cooney
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FMESafe Software
 
[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdfSandro Moreira
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDropbox
 

Recently uploaded (20)

Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
 
FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024
 
Six Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal OntologySix Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal Ontology
 
Platformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityPlatformless Horizons for Digital Adaptability
Platformless Horizons for Digital Adaptability
 
Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
Emergent Methods: Multi-lingual narrative tracking in the news - real-time ex...
 
ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challenges
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor Presentation
 

How do we train AI to be Ethical and Unbiased?

  • 1. HOW DO WE TRAIN AI TO BE ETHICAL AND UNBIASED? MARK BORG AI MALTA SUMMIT – 13 JULY 2018
  • 2. RECENT ACHIEVEMENTS IN AI 2 Word Error rate Improvements in word error rate over time on the Switchboard conversational speech recognition benchmark. Credit: Awni Hannun Automated Speech Recognition results Credit: Business Insider/Yu Han
  • 3. RECENT ACHIEVEMENTS IN AI 3 Credit: H. Fang et al. (2015), “From Captions to Visual Concepts and Back” #1 A woman holding a camera in a crowd. Image Captioning
  • 4. RECENT ACHIEVEMENTS IN AI 4 0 days AlphaGo Zero has no prior knowledge of the game and only the basic rules as an input. 3 days AlphaGo Zero surpasses the abilities of AlphaGo Lee, the version that beat world champion Lee Sedol in 4 out of 5 games in 2016. 21 days AlphaGo Zero reaches the level of AlphaGo Master, the version that defeated 60 top professionals online and world champion Ke Jie in 3 out of 3 games in 2017. 40 days AlphaGo Zero surpasses all other versions of AlphaGo and, arguably, becomes the best Go player in the world. It does this entirely from self-play, with no human intervention and using no historical data. Credit: DeepMind AlphaGo Zero
  • 5. WIDESPREAD USE OF AI • AI has now wide and deep societal influences, permeating every sphere of our lives • No longer single applications operating in standalone mode • ML Pipelines, more complex AI systems, operating at Internet Scale • AI as a Service (AIaaS), Machine Learning as a Service (MLaaS) • Running “under the hood”, as well as in “human-facing technology” • High-stake applications, sometimes involving life-and-death decisions ➢ AI-enabled Future ➢ Benefits and Implications 5
  • 6. BENEFITS AND CONCERNS OF AI 6 • What if an AI algorithm could predict death better than doctors? • The “dying algorithm” (NY Times) • Stanford's AI Predicts Death for Better End-of-Life Care (IEEE Spectrum) • What are the benefits and implications of such a system?
  • 7. CONCERNS • A Predictive Policing algorithm unfairly targeted certain neighbourhoods – Chicago 2013/2014 • Idea: to stop crime before it occurs • Unintended consequences due to systematic bias in the data used by these systems • Saunders et al. (2016), “Predictions put into practice: a quasi- experimental evaluation of Chicago’s predictive policing project” • COMPAS assesses a defendant’s risk of re-offending • used for bail determination by judges • Issues of reliability and racial bias • Dressel & Farid (2018), “The Accuracy, Fairness, and Limits of Predicting Recividism” 7Credit: ProPublica
  • 8. CONCERNS • YouTube Recommender system • The algorithm appears to have concluded that people are drawn to content that is more extreme than what they started with — or to incendiary content in general • Accusations that YouTube is acting as a “radicalisation agent” 8 Credit: Covington Recommendations drive 70% of YouTube’s viewing time (~200 million recommendations per day) YouTube tops a cumulative of 1 billion hours of video per day in 2017
  • 9. CONCERNS • Adversarial AI 9Credit: IBM Credit: Biggio & Roli
  • 10. CONCERNS • Ethical and moral issues • Self driving cars 10 The Trolley Problem Credit: Waymo (Philippa Foot, 1967)
  • 11. LONG-TERM CONCERNS • GAI, Superintelligence, existential threat, need for Benevolent AI • The Sorcerer’s Apprentice problem • Eliezer Yudkowsky: The Paperclip Maximiser Scenario 11 Credit: Disney If a machine can think, it might think more intelligently than we do, and then where should we be? … This new danger … is certainly something which can give us anxiety Alan Turing, 1951 “ “
  • 12. IMPLICATIONS & CONSEQUENCES OF AI • To maximise the benefits of AI: saving lives, raising the quality of life, … Need also to address issues and consequences • the “rough edges of AI” – Eric Horvitz (Microsoft Research) • Robustness, Ethics, Benevolent AI • Short-term implications (need solving now) • Longer term implications (prepare the groundwork…) • Spans multiple fields: engineering, cognitive science, philosophy, etc. 12
  • 13. 13 AIES ICAILEP Conference on Artificial Intelligence: Law, Ethics, and Policy 7008 - Standard for Ethically Driven Nudging for Robotic, Intelligent & Autonomous Systems 7009 - Standard for Fail-Safe Design of Autonomous & Semi-Autonomous Systems 7010 - Wellbeing Metrics Standard for Ethical Artificial Intelligence & Autonomous Systems
  • 14. IMPLICATIONS & CONSEQUENCES OF AI 14 Benevolent AI AI Safety Robust AI Beneficial AI Value Alignment AI Ethics Roboethics Machine Ethics Adversarial AI Increasedcomplexity AI transparency ANI Artificial Narrow Intelligence AGI Artificial General Intelligence ASI Artificial Super Intelligence
  • 15. 15 LADDER OF CAUSATION Credit: Judea Pearl (2018), “The Book of Why: The New Science of Cause and Effect”
  • 16. AI SAFETY • Data Bias (Algorithmic Bias) • Fairness • AI Robustness & Reliability • AI Transparency 16
  • 17. DATA BIAS • Algorithmic Bias is NOT model bias (bias-variance trade-off, generalisation problem) • Algorithmic Bias (or Data Bias) – will always be present; need to minimise the impact • E.g. predictive policing algorithm • Police-recorded datasets suffer from systematic bias: • Non-complete census • Not a representative random sample • Crime databases do not measure crime; they measure some complex interaction between criminality, policing strategy, and community-police relationships 17
  • 18. DATA BIAS • Data bias is prevalent throughout the whole field of AI • Unintentional bias vs. intentional bias • Addressing data bias has particular significance in ML pipelines, complex AI systems, AIaaS, etc. • E.g. • Howard (2017), “Addressing Bias in Machine Learning Algorithms: A Pilot Study on Emotion Recognition for Intelligent Systems” • did not perform well for children • original training dataset had few such cases 18
  • 19. DATA BIAS • Unintentional self-created bias (“poisoning your own data”) • E.g. Google Flu Trends • began suggesting flu-related queries to people who did not have the flu, and thus Google Flu Trends began itself corrupting the dataset by seeding it with excess flu-related queries, thus creating a feedback loop • Despite good intentions, biased data can lead to a far worse result • E.g. beauty.ai • a startup organising the world's first AI-driven beauty contest in 2016 • The concept is to remove social biases of human judges • problem: image samples used to train the algorithms weren’t balanced in terms of race and ethnicity. • so-called 'white guy problem’ 19
  • 20. DATA BIAS • Naive application of algorithms to everyday problems could amplify structural discrimination and reproduce biases present in the data • Detecting (automatically?) such bias and addressing it? • Quite difficult! • Since AI is data-driven, it’s difficult 20 Credit: Buolamwini & Gebru • Some very recent work on two fronts: • More balanced datasets, e.g., new facial image dataset released in February 2018 (Pilot Parliaments Benchmark dataset) • Buolamwini & Gebru (2018), “Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification” • Measuring bias and fairness: • Shaikh et al. (2017), “An End-to-End Machine Learning Pipeline that Ensures Fairness Policies” • Srivastava & Rossi (2018), “Towards Composable Bias Rating of AI Services”
  • 21. DATA BIAS • creating more balanced (heterogeneous) datasets • One solution would be to create shared and regulated databases that are in possession of no single entity, thus preventing any party from unilaterally manipulating the data to their own favour. • public datasets curated to be bias free • One concern is that even when machine-learning systems are programmed to be blind to race or gender, for example, they may use other signals in data such as the location of a person’s home as a proxy for it • E.g. In COMPAS bail system, geographic neighbourhood highly correlated to ethnicity, thus still suffering from racial discrimination 21
  • 22. AI ROBUSTNESS & RELIABILITY • Making AI systems more robust, so that they work as intended, without failing or getting misused? • Reliable prediction of performance • Avoiding overconfidence in AI systems • How much it knows that it does not know? • Make strong predictions that are just inaccurate • Classification label accuracy + ROC curve • Learning to predict confidence • Current statistical models “tend to assume that the data that they’ll see in the future will look a lot like the data they’ve seen in the past” 22
  • 23. AI ROBUSTNESS & RELIABILITY • Blindspots of algorithms (Eric Horvitz, Microsoft Research) • The “unknown unknowns” (Tom Dietterich , Oregon State University) • AI algorithms for learning and acting safely in the presence of unknown unknowns. • learning about blindspots of algorithms: • Lakkaraju (2016), “Identifying Unknown Unknowns in the Open World: Representations and Policies for Guided Exploration” • Ramakrishnan (2018), “Discovering Blind Spots in Reinforcement Learning” • Human supervision • human correction to prevent AI failure 23
  • 24. AI ROBUSTNESS & RELIABILITY • Watch out for Anomalies? • Robust anomaly detection • BugID system by Tom Dietterich, • A system that learn when there is another unknown class out there • automated counting of freshwater macro-invertebrates • trained on 29 insect classes, detection of novel classes • monitoring of performance of the system, especially for self-learning systems or local- based learning • E.g. Microsoft’s chatbot Tay • research into adding a “reflection layer” into systems (introspection?) • Failsafe designs • Auto-pilot of self-driving cars disengaging suddenly 24
  • 25. MISUSE OF AI • Privacy Challenges • Exclusion – denying services • Persuasion, and Manipulation of Attention / Behaviour / Beliefs • Harms • Hacking of AI systems • Adversarial AI 25
  • 26. MISUSE OF AI • Tay • Microsoft’s Chatbot • "The more you talk the smarter Tay gets!" • March 2016, suspended after 16 hours • Tay’s conversation extended to racist, inflammatory and political statements • A main problem was Tay’s “repeat after me” feature • Intentional misuse of AI (coordinated attack) • Neff and Nagy (2016), “Talking to Bots: Symbiotic Agency and the Case of Tay” 26 Credit: Microsoft
  • 27. MISUSE OF AI • Harnessing AI to increase attention & engagement for a particular application or service • large-scale personalised targeting • Persuasion, and Manipulation of Attention / Behaviour / Beliefs • auto twitter feed generation that persuades a user to click on links • data-driven behaviour change • Intentional / Unintentional 27
  • 28. YOUTUBE RECOMMENDER SYSTEM • The recommender system’s goal is to maximise attention and engagement via personalised targeting • Eric Horvitz (Microsoft Research) calls this: "Adversarial Attacks on Attention" 28 Credit: Eric Horvitz Recommendations drive 70% of YouTube’s viewing time (~200 million recommendations per day) YouTube tops a cumulative of 1 billion hours of video per day in 2017 Recommendation system architecture demonstrating the “funnel” where candidate videos are retrieved and ranked before presenting only a few to the user Covington et al. (2016), “Deep Neural Networks for YouTube Recommendations”
  • 29. YOUTUBE RECOMMENDER SYSTEM • Its algorithm seems to have concluded that people are drawn to content that is more extreme than what they started with — or to incendiary content in general • a bias toward extreme/divisive/inflammatory/fringe/sensational content • WSJ investigation (Feb 2018): • amplifies human bias, fake news, isolate users in "filter bubbles" • AlgoTransparency.org • Zeynep Tufekci (sociologist, Univ. of North Carolina): • Calls YouTube the “Great Radicaliser” • AI exploiting a natural human desire to "look behind the curtain", to dig deeper into something that engages us. As we click and click, we are carried along by the exciting sensation of uncovering more secrets and deeper truths. YouTube leads viewers down a rabbit hole of extremism, while Google racks up the ad sales. 29
  • 30. YOUTUBE RECOMMENDER SYSTEM • But is the algorithm really to blame? • Main issue due to scale • Also simplified human behaviour modelling: • Watching more nuanced content or videos which diverge from the established viewing pattern could be rooted out as noise, thus contributing to simplification and generalization of interests towards the more extreme ends of a spectrum, instead of complex content catering to views which are harder to define. • Possible solutions? • YouTube has been applying changes to their algorithm • Improved human behaviour models • Changes to the exploration-exploitation strategy adopted by the recommender system • Value policies, encoding notion of “time well spent” 30
  • 31. ADVERSARIAL AI • Goodfellow et al. (2015), “Explaining and Harnessing Adversarial Examples” • Szegedy (2013), “Traversing the manifold to find blind spots in the input space” • DNN can be easily fooled by adversaries • No need for hand-crafting the adversarial attack • Can exploit AI to perform an adversarial attack • One AI deceiving another AI 31 “panda” 57.7% confidence “gibbon” 99.3% confidence Adversarial Noise (exaggerated)
  • 32. ADVERSARIAL AI • Adversarial systems subtly alter normal inputs such that humans doing the same task can easily recognize what the intended input is, but mislead the AI into giving a predictable and very different false output. • Performed by stealth (humans won’t spot the difference) • Potential Attacks • Adversarial examples can be printed out on standard paper, then photographed with a standard smartphone, and it will still fool AI systems. • Kurakin et al. (2017), “Adversarial examples in the physical world” 32 Credit: Biggio & Roli
  • 33. ADVERSARIAL AI • The famous 3D printed Turtle that fooled Google’s AI • Adversarial attacks without perturbing the whole image 33 Athalye et al. (2017), “Synthesizing Robust Adversarial Examples” • Sharif et al. (2016), “Accessorize to a Crime: Real and Stealthy Attacks on State-of-the-Art Face Recognition” • Impersonation attacks • Invisibility attacks Credit: Sharif et al. (2016)
  • 34. ADVERSARIAL AI • Audio Adversarial attacks • Carlini and Wagner (2018), “Audio Adversarial Examples: Targeted Attacks on Speech-to-Text” • Given any speech audio, can produce another that is 99.9% similar to the original, but contains any text one wants. • Fools DeepSpeech with 100% accuracy 34Credit: IBM
  • 35. ADVERSARIAL AI • Not limited to Deep Neural Networks • Papernot et al. (2016), “Transferability in machine learning: from phenomena to black-box attacks using adversarial samples” • DNNs, logistic regression, support vector machines, decision trees, nearest neighbour classifiers, ensembles – all vulnerable to adversarial AI! • any machine learning classifier can be tricked to give incorrect predictions, and with a little bit of work, one can get them to give pretty much any result one wants 35
  • 37. ADVERSARIAL AI • White-box adversarial attack 37 Adversarial AI panda57.7% Defended AI model gradient + gibbon100% Score-based attack
  • 38. Substitute AI model ADVERSARIAL AI • Black-box adversarial attack 38 Adversarial AI panda Defended AI model + gibbon Transfer-based attack Decision-based attack scores & gradients
  • 39. Some countermeasures: DEFENDING AGAINST ADVERSARIAL AI 39 • Smoothing and hiding the gradients • Randomisation techniques • image compression • Image blurring • random image resizing • employ dropout in neural networks • Defensive distillation • Use of ensembles • Evaluate model’s adversarial resilience • Metrics available • Pre-emptive hardening of AI models • Enhance robustness to tampering Some defensive techniques: IBM Adversarial Robustness Toolbox (ART) https://github.com/IBM/adversarial-robustness-toolbox Cleverhans library https://github.com/openai/cleverhans DeepFool https://github.com/LTS4/DeepFool
  • 40. • Fredrikson et al. (2015), “Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures” • Violating privacy of subjects in the training set ADVERSARIAL AI – MODEL INVERSION ATTACKS 40 Adversarial AI Tom 2.3% 70% Training Set Tom Tom Defended AI model Face Recognition
  • 41. • Leveraging adversarial AI to make a generative model, consisting of two neural networks competing with each other • The discriminator tries to distinguish genuine data from forgeries created by the generator • The generator turns random noise into imitations of the data, in an attempt to fool the discriminator GENERATIVE ADVERSARIAL NETWORKS (GANS) 41 real DiscriminatorGenerator random noise fake real fake
  • 42. AI ETHICS & VALUE ALIGNMENT • Codification of Ethics • Values, Utility Functions • Teaching AI to be Ethical • Reinforcement Learning • Inverse Reinforcement Learning and beyond ▪ Ethics – comprehending “right” from “wrong”, and behaving in a right way ▪ Value Alignment – ensuring that the goals, behaviours, values and ethics of autonomous AI systems align with those of humans 42
  • 43. CODIFICATION OF ETHICS • Rule-based ethics (deontological ethics) • Isaac Asimov’s “Three Laws of Robotics” • And similar sets of rules 43 • Challenges: • Too rigid • Asimov’s literature addresses many of these issues: • conflicts between the 3 laws, conflicts within a law by itself, conflicting orders, etc. • How to codify the rules? • How to program the notion of "harm"? • Often human ethics and values are implicit • Process of elicitation is very challenging Isaac Asimov (1942)
  • 44. CODIFICATION OF ETHICS • Pre-programming ethical rules: • Impossible to program for every scenario • Fail to address uncertainty and randomness • Fail to address ambiguous cases, ethical and moral dilemmas • Rules on their own not enough • Must be accompanied by very strong accountability mechanisms • Need moral conflict resolution mechanism • Values and ethics are dependent on the socio-cultural context • Difficult to standardise • Need to account for changes in the values of society, shifts in beliefs, attitudes, etc. 44
  • 45. CODIFICATION OF ETHICS 45 • Rule-based ethics example: • specifically & explicitly programme ethical values into self-driving cars to prioritize the protection of human life above all else • In the event of an unavoidable accident, the car should be “prohibited to offset victims against one another” • A car must not choose whether to kill a person based on individual features, when a fatal crash is inescapable Credit: BMVI (www.bmvi.de) The Trolley Problem
  • 46. VALUES, UTILITY FUNCTIONS • Ethics as Utility Functions • Any system or person who acts or gives advice is using some value system of what is important and what is not • Utility-based Agent • Agent’s Actions • Agent’s Beliefs • Agent’s Preferences • The agent chooses actions based on their outcomes • Outcomes are what the agent has preference on • Preferences → Utility → Utility Function • A policy specifies what an agent should do under all contingencies • An agent wants to find an optimal policy – one that maximises its expected utility 46
  • 47. TEACHING AI TO BE ETHICAL • Teaching AI ethics, social rules and norms • Adopt a “blank slate” approach • Similar to how a human child learns ethics from those around him/her • Basic values are learnt, and the AI will, in time, be able to apply those principles in unforeseen scenarios • What machine learning method to use? 47Credit: GoodAI
  • 48. TEACHING AI TO BE ETHICAL 48 • Reinforcement Learning • Has shown promise in learning policies that can solve complex problems • An agent explores its environment, performing action after action and receiving rewards and punishments according to the reward function (i.e. utility function) • As it repeats this, the agent will gradually learn to perform the right actions in the right states so as to maximise its reward • Reward = total sum of the actions’ rewards over time, where future rewards are discounted (treated as less valuable than present rewards) • When learning ethics, the reward function will reward/punish the agent depending on the choice of action performed, whether “right” or “wrong” Environment Model Reward Function(s) Reinforcement Learning Reward- maximising behaviour Kose (2017), “Ethical Artificial Intelligence – An Open Question”
  • 49. TEACHING AI TO BE ETHICAL • Reinforcement Learning (RL) challenges: • Difficulty in setting up ethical scenarios in the environment model of RL • May take a very long time till the agent manages to fully cover all ethical scenarios, ambiguous cases, etc. 49 • Potential solution: • Using stories as a way of short-circuiting the reinforcement learning process • Employ more complex stories as time goes by • Riedl et al. (2016), “Using Stories to Teach Human Values to Artificial Agents” Environment Model Reward Function(s) Reinforcement Learning Kose (2017), “Ethical Artificial Intelligence – An Open Question” Reward- maximising behaviour
  • 50. TEACHING AI TO BE ETHICAL • Another solution: • Curriculum-based approach to improve the learning process • The learning process in humans and animals is enhanced when scenarios are not randomly presented, but organized in a meaningful order – gradual exposure to an increasing number of concepts, and to more complex ones • For teaching ethics, simpler scenarios are presented before more complex and ambiguous cases • GoodAI’s “School for AI” project is employing a curriculum based approach for enhancing the teaching of ethics via reinforcement learning • www.goodai.com/school-for-ai • Bengio et al. (2009), “Curriculum Learning” • Weinshall et al. (2018), “Curriculum Learning by Transfer Learning: Theory and Experiments with Deep Networks” 50
  • 51. TEACHING AI TO BE ETHICAL • Crowd-Sourcing Ethics and Morality • Crowdsourced stories simplify the manually-intensive process of creating stories • Can capture consensus for ambiguous and moral dilemmas (“wisdom of the crowds”) • Example: • An AI agent is given several hundred stories about stealing versus not stealing, explores different actions in a reinforcement learning setting, and learns the consequences and optimal policy based on the rewards / punishments given. (Mark Riedl, Georgia Tech) 51
  • 52. TEACHING AI TO BE ETHICAL • MIT’s “Moral Machine”: • Crowdsourcing to aid self-driving cars make better moral decisions in cases of moral dilemmas (variations of the Trolley problem) • http://moralmachine.mit.edu 52
  • 53. TEACHING AI TO BE ETHICAL • Reinforcement Learning (RL) requires the manual specification of the reward function • “Reward engineering” is hard (especially for ethics) • May be susceptible to “reward cheating” by the AI agent 53 In RL, the reward function is specified by the user, and then the agent does the acting What if the agent could instead watch someone else do the acting, and try to come up with the reward function by itself? Environment Model Reward Function(s) Reinforcement Learning Provided by the user Reward- maximising behaviour
  • 54. TEACHING AI TO BE ETHICAL • Inverse Reinforcement Learning (IRL) • IRL is able to learn the underlying reward function (what is ethical?) from expert demonstrations (humans solving ethical problems) • IRL is also called “imitation-based learning” • Learn from watching good behaviour 54 Reward- maximising behaviour Environment Model Reward Function(s) Reinforcement Learning Provided by the user Environment Model Reward Function(s) Inverse Reinforcement Learning Provided by the user Observed Behaviour Reward- maximising behaviour Reinforcement Learning
  • 55. TEACHING AI TO BE ETHICAL • Inverse Reinforcement Learning (IRL) • Very promising results for AI ethics (value alignment) • No need to explicitly model rules, reward function • Recent works advocating IRL: • Russell et al. (2016), “Research Priorities for Robust and Beneficial Artificial Intelligence” • Abel (2016), “Reinforcement Learning as a Framework for Ethical Decision Making” • Challenges of IRL: • Interpretability of the auto-learnt reward function • Human bias can creep into the observed behaviour • Difficulty of learnt ethics to be domain independent • Arnold (2017), “Value Alignment or Misalignment – What Will Keep Systems Accountable?” 55 Environment Model Reward Function(s) Inverse Reinforcement Learning Provided by the user Observed Behaviour Reward- maximising behaviour Reinforcement Learning
  • 56. BEYOND IRL… • Cooperative IRL • What if we reward both the “good behaviour” of the AI while learning ethics, as well as reward the “good teaching” of the human? • Cooperation between AI and humans to accomplish a shared goal – value alignment • Generative Adversarial Networks (GANs) • Hadfield-Menell (2016), “Cooperative Inverse Reinforcement Learning” 56 Environment Model Reward Function(s) Inverse Reinforcement Learning Provided by the user Observed Behaviour Reward- maximising behaviour Reinforcement Learning
  • 57. BEYOND IRL… • Harnessing Counterfactuals • … “imagination” rung on the ladder of causation • As perfect knowledge of the world is unavailable, counterfactuals allow for the revision of one’s belief system and the relying solely on past (data driven) experience • It is also through counterfactuals that one ultimately enters into social appraisals of blame and praise • Might prove to be one of the key technologies needed both for the advancement of AI itself on the trajectory towards AGN, as well as for aligning as much as possible the values of machines with our values to achieve benevolent AI 57
  • 59. BENEVOLENT AI 59 our values AI values mutually beneficial values Value Alignment Everything we love about civilisation is a product of intelligence, so amplifying our human intelligence with artificial intelligence has the potential of helping civilisation flourish like never before – as long as we manage to keep the technology beneficial. Max Tegmark, Cosmologist & President of the Future of Life Institute “ “