SlideShare a Scribd company logo
1 of 1
Download to read offline
Do Nothing
• AGI is impossible
- Humans are unique, a soul exists
- Existence of upper limits on an entity’s effective intelli-
gence around the level of human geniuses
- Individual intelligence is always outmatched by the dis-
tributed intelligence found in social
systems of many minds
• AGI is too distant to be worth our
attention
- not in this century
- not on our level of understanding of AI theory
- the discussion has negative utility
• Little risk, no action needed
- AGIs will not have any particular
motivation to act against us
- AGI cannot do much
- High intelligence implies high morality
• Let it kill us
- Humans are bad
- Higher intelligence is more important than
humanity
- Quantum immortality will make only Safe AGI observ-
able
- AI is our child – we must step aside
- AI will share some of our values and will construct an-
cestor simulations which contain humans
- We should not constrain AGI because it
deserves full freedom to act
• Safety measures are dangerous
themselves
- The more we try to confine it, the more AGI
will try to escape
AGI Safety Solutions Map
Created by Alexey Turchin based on the article “Responses to Catastrophic AGI Risk: A Survey” by Kaj Sotala and Roman V.
Yampolskiy, 2013
Ideas from the article are inside red border line.
Ideas not derived from the article are marked to their authors in [...]
“No” means that a method definitely will not work (it is point of view of Alexey Turchin and other people may have different
opinions)
“May be” means that small chances that a method may work (like 1 per cent)
“Marginal” that a method may be used for short time or as additionally precaution measure but don’t change the situation
much.
“Promising” means that a method seems to be good enough to invest more in its research.
© Alexey Turchin, 2015, GNU-like license. Free copying and design editing, discuss major updates with Alexey Turchin.
All other roadmaps in last versions: http://immortalityroadmap.com/
and here: https://www.scribd.com/avturchin , https://www.scribd.com/doc/265853723/AGI-Safety-Solutions
mailto: alexei.turchin@gmail.com
Proofreading and advise: Michael Anissimov
Societal
Proposals
External
Constraints
Internal
Constraints
AI used to
create Safe
AI
Integrate into Society
Marginal
• Legal and Economic Controls:
- AGIs will be law-abiding
- Share resources
- Taxation and regulation
- Integrate into government
• Foster Positive Values
• State AGI: State transforms into AGI through
gradual replacement of existing functions (computer
tax regulation and law implementation)
Regulate Research
Marginal
• International regulation
- international treaty
- boards
- licensing research
• Encourage research into safe AGI
- funding for research
- open-source approach
• Differential Technological Progress
- curtail unsafe AI architectures
- prevent AGI arms race
- one center of AGI development
• International Mass Surveillance
- narrow AIs as controllers of AGI research
Enhance Human
Capabilities
Marginal
• Create AGI by human upgrade
- cognitive enhancement
- moral enhancement
• Keep humans on the same level
as AGI
AGI Confinement
Marginal
• Total isolation
• Remote place like Mars or Pluto
• Only text communication
• Internet firewall
• Virtual reality:
many levels of isolation
• “Safe Questions” (not implying dan-
gerous behaviour)
• Resetting AGI memory
• Checks and balances
- baits
- check for unusual pattern in the AGI
- circuit breakers
- off switch
• Cryptobox: homomorphic encryption
[Christiano]
Goal Checking
Marginal
• Automatically search for failure modes
of goal system
• List of signs of failure modes
• Committee for goal checking
• Prevent failure modes
Philosophical
Constrains
Marginal
• AGI fooled to be in simulation which
could switch him off
• Persuade AGI that humans have soul
or other incomputable quality and it
needs them
• Prove that we really maybe in
the simulation and AGI should keep peo-
ple alive to not risk punishment from
high-level simulators
Prevent
Self-improvement
Promising
• Limit number of self-improvement steps
• Secure permission from human operator
for each step of self-improvement
• Slow rate of self-improvement
• Limit areas of self-improvement
• Limit access to own source code
• Limit IQ of the AGI
• Limit available hardware for the AGI
• Limit working speed of the AGI
• Prevent large architecture changes
• Create as many AGIs as needed to
prevent any of them from total power
• Differential self-improvement: slow on
high levels and quick on low level
Oracle AI
May be
• Answer questions
• No goals
• No memory
• Not one AI but many narrow Tool
AIs with different goals
• Slave AI – only obey commands
Short rulesets
May be
• “Three laws by Asimov”
• Categorical imperative
• “Principle of voluntary joyous growth”
• Utilitarianism
• AGI obeys existing criminal laws
• Short rulesets of other types
Bottom-Up and
Hybrid Safe AGI
May be
• Evolutionary invariants of ethics
• Evolved morality
• Reinforced learning of ethics
• Learning values by pattern recognition
• Human-like architecture
Formal
Verification
May be
• “Goal-oriented Learning Meta-
Architecture” – search for better way to
known goal [Goertzel]
• Provable self-modifying algorithms
AGI tasked to
create Safe AGI
May be
• Oracle AI gives friendliness theory
• Oracle AI designs safe AI
• Oracle AI tests different AI designs
• Oracle AI checks “AI constitution” for
failure modes and prevents them
• Group of uploads works on creating and
proving “AI constitution”
• Two-level AI: emulation for goals
control, Tool AI for problem solving
Timeless
decision
theory deal
Marginal
• Humans invest in creating AGI, AGI
invests in being safe to humans
• AGI is safe to humans because it
expects that its own next versions
will be safe to him
AGI
Enforcement
Marginal
• AGIs to watch other AGIs
- Multilevel system
- Community of AGIs
• AI runs simulation with AGI inside
• AGI actions are tested in
simulation for unintended
consequences
• First AI checks safety of decisions
of another AGI
AGI Nanny
May be
• Singleton
• No self-improvement
• Prevent other AIs from becoming
too powerful
• Minimum rulesets: prevent
death, x-risks and suffering,
prohibition of forced brain
reprogramming [Turchin]
Relinquish
Technology
• Outlaw AGI
- outlaw research on AI
- nuclear war against AGI rivals
[Turchin]
• Restrict Hardware
- limit processing power
- outlaw supercomputers
- control microchip fabs
- switch off electricity or internet
Control room
and
Gatekeeper
Marginal
• Monitors of internal processes
• AI is constantly showing what it is think-
ing about
• Power switch
• Explicit goals and subgoals
• Firewall
• Upgrade control
• Mirror AIs to check for strange
behaviour (3 copies of the AGI)
• Seed AI testing for unintended behaviour
Motivational
Weaknesses
Marginal
• High Discount Rates for time
or distance (don’t try to save
Universe)
• Easily satisfiable goals
• Promised safety for the AGI
• Calculated indifference to some
threats
• Legal machine language
Coherent
Extrapolated
Volition
May be
• AI predicts what our values
would be if we had more time to
think about them (CEV)
AI Сonstitution
Promising
AI is regulated by proven large set of
rules; some possible rules:
Meta rules:
• Obey commands of its creator
• Report and explain value changes
• No secrets
• Limit self-improvement
• Stop after 10 years
Content:
• Do not kill people
• Cure death
• Prevent severe suffering
• Prevent x-risks
• Don’t rewrite human brains against
their will
Decision Theory
Based
Friendliness
Marginal
• Cooperate in prisoner dilemma
• Don’t take actions which are
irreversible
• Anti-Pascaline agent [Armstrong]
• Corrigible reasoning [MIRI]
AGI without one
Supergoal
Marginal
• Any one-goal system may be
dangerous like “maniac”
• Democracy of goals
• Values not goals
• AI capable on reflection on its
goals and values
• All possible goals in different
proportions (like in humans)
Situation
Improvement
Marginal
• Funding of safe AGI research
• Slowing development of AGI
• Promotion of the idea of AGI
safety
• Ban military AGI
• Friendly AGI training for AGI
researchers
• One main international AGI
project, not many rivals
• Safe AGI architectures: e.g. no
genetic algorithms
• Hire existing ethical researchers
and use legal and ethical literature
• Hire existing safe AI researchers
AI as Next Step
in Human Evolution
Promising
• AI based on emulations
- Society of EM’s working at high speed
- Upload a specially trained
person with math skills and solid ethic as Seed AI
- EM as value and awareness core of AGI
• Use human-like value architecture: cultural
upgrade of natural drives, reward system,
complex values, many different architectures
• AGI use humans (“me”) as a source of qualia (actuality)
and preserve them
• AI is based on many human brains connected together
(neouronet)
• Psychogenic singularity: use raw power of human brain
and its unconsciousness to create AI (like savants)
• AI is created by as gradual upgrade of one or several
human beings (posthumans)
• AI is created by genetic experiments like a child or
superbrain
• AI pretends to be best human (Armstrong)
Simple
solutions
Complex
solutions
Value learning
Promising
• AGIs cooperate with humans in value
learning
• AI learns values by studying uploaded
brain
• Education of childlike Seed AI
Multilevel
Safe AGI
Promising
1) Unchangeable core with
constitution and values
2) ЕМ as a cloud around this core
3) Tool AIs as extension of EM
4) External confinement and
control
5) Self-improvement is limited
6) Test its actions in virtual
reality
7) Virtual jury for action
judgment
8) Works as Nanny AI
9) Use it to predict failure modes
of its own constitution
Integrated
Solutions
Military
Nanny AI
Marginal
Wins war for world domination,
then become Nanny AI
AGI Democracy
May be
• AGIs with different architectures
suggest different proposals for
AGI safety and vote on them
[I.Artuhov]
Other useful
ideas
• Value of human values
• Secret way of testing right
behaviour will result in obeying
spirit of law [Armstrong]
• Break AI into modules
[Armstrong]
• Prevent orthogonality of means
and values (like in humans)
[Phil Torres]
• Approval-directed agents
[Christiano]
Logical or philosophical
“landmines”
which will stop too complex AI
Marginal
• Idea of absurdity of any goal system
• AI halting problem
• Lobian obstacle [Yudkowsky]
• AGI is given false believes, which is known to
it but it still obeys them [Armstromg]
General
requirements
to the
Safe AI theory
• It must be easy to understand
and implement so all researchers
will prefer it
• It must help AI to win first and to
safely establish its power into
Singleton
• It should be computationally
simple
• It should have good chance to be
really safe
• Choose the ways to the safe AGI
which are not mutually exclusive
Meta
Virtual jury
• Virtual jury based on highly
moral uploads checks AI’s
decisions, goal system and CEV
[Yudkowsky]
AI Testing
May be
• Test AI inside multilevel simulation: if it
breaks 1st level, terminate.
• Test AI with difficult moral choices
• Test AI decisions inside simulations for
inundated consequences
• Give AI idea that its decisions may be
tested by unknown method
• Create parallel AI which only test deci-
sions of first AI, but is build on different
principles (may be it is human upload or
virtual jury)
• AI self-improve in different branches,
which later test each other for core values
consistency
• AI present full model of the future which
would result from its decisions, and human
operators (or uploads) validate them
International
AI agency
• Choose the best AI safety theory
and recommend its implementation
Catch the moment
of beginning of
malignant AI self-
improvement
Marginal
After AI acquires malignant goal system
and starts self-improvement, it is still
stupid and could be stopped
Give AI to everybody
Suggested by Elon Musk in OpenAI project
Knowledge
constrains
• Separete source code of AI, ex-
ecutable code, data and knowl-
edge about how to build AI.
• AI can’t phisically read its own
source code or its attempts are
recorded
Signs
constrains
Identify all signs of possible
wrongdoings by AI and make in-
depended circut breakers.
Signs:
- Reading source code
- Modifing own code
- Creating AI models
- Power consumption changes
- Behavior patterns changes
- Internal work pattern changes
- Access to Internet on AI tipics
- Attempts to cheat and lie
- Goal system changes
Opacity
constrains
Use opaque architecture which
is difficult to understand and
improve (some kinds of neural
nets)
Open goal
system
The actual goal system of an
AI must be presented to exter-
nal observer in the form of an
open tree (text, graph) and any
changes in it should be saved re-
motely in unchangeble way.
If the AI will try to create anoth-
er goal tale it will be presented
in first.
Children
constrain
In order to selfupgrade, an AI
must first create new version of
itself and test it.
Some AI-architectures may make
it very complex task with many
fails and resource consuming
“education period” (neural nets)
Technical level solution
One can’t make safe rocket, without actually knowing
how to build a rocket. The same may be true about AI.
So we may need to wait until more will be known
about most promissing AI architectures and use more
help from actual developers.
Record AI’s
thought stream
Build AI in the way that its high-
level resaonong are presented in
the open form and record them
and analyze, may be using an-
other AI.
Intelligence
distribution
Create sitiation in which distributed collectibe intelli-
gence of humanity is always greater than of any com-
pany, or computer, so it can’t get desisive advantage.
Distributed science
Neuronet

More Related Content

What's hot

Artificial general intelligence research project at Keen Software House (3/2015)
Artificial general intelligence research project at Keen Software House (3/2015)Artificial general intelligence research project at Keen Software House (3/2015)
Artificial general intelligence research project at Keen Software House (3/2015)Marek Rosa
 
Ethics in the use of Data & AI
Ethics in the use of Data & AI Ethics in the use of Data & AI
Ethics in the use of Data & AI Kalilur Rahman
 
Patterson Consulting: What is Artificial Intelligence?
Patterson Consulting: What is Artificial Intelligence?Patterson Consulting: What is Artificial Intelligence?
Patterson Consulting: What is Artificial Intelligence?Josh Patterson
 
What is Artificial Intelligence
What is Artificial IntelligenceWhat is Artificial Intelligence
What is Artificial IntelligenceJosh Patterson
 
Business Outsourcing to Asia
Business Outsourcing to AsiaBusiness Outsourcing to Asia
Business Outsourcing to AsiaConferencias FIST
 
Bringing AI to Business Intelligence
Bringing AI to Business IntelligenceBringing AI to Business Intelligence
Bringing AI to Business IntelligenceSi Krishan
 
Artificial Intelligence Applications, Research, and Economics
Artificial Intelligence Applications, Research, and EconomicsArtificial Intelligence Applications, Research, and Economics
Artificial Intelligence Applications, Research, and EconomicsIkhlaq Sidhu
 
Designing Trustable AI Experiences at IxDA Pittsburgh, Jan 2019
Designing Trustable AI Experiences at IxDA Pittsburgh, Jan 2019Designing Trustable AI Experiences at IxDA Pittsburgh, Jan 2019
Designing Trustable AI Experiences at IxDA Pittsburgh, Jan 2019Carol Smith
 
Introduction to Artificial Intelligence and Machine Learning: Ecosystem and T...
Introduction to Artificial Intelligence and Machine Learning: Ecosystem and T...Introduction to Artificial Intelligence and Machine Learning: Ecosystem and T...
Introduction to Artificial Intelligence and Machine Learning: Ecosystem and T...Raffaele Mauro
 

What's hot (13)

Artificial general intelligence research project at Keen Software House (3/2015)
Artificial general intelligence research project at Keen Software House (3/2015)Artificial general intelligence research project at Keen Software House (3/2015)
Artificial general intelligence research project at Keen Software House (3/2015)
 
Ethics in the use of Data & AI
Ethics in the use of Data & AI Ethics in the use of Data & AI
Ethics in the use of Data & AI
 
Patterson Consulting: What is Artificial Intelligence?
Patterson Consulting: What is Artificial Intelligence?Patterson Consulting: What is Artificial Intelligence?
Patterson Consulting: What is Artificial Intelligence?
 
What is Artificial Intelligence
What is Artificial IntelligenceWhat is Artificial Intelligence
What is Artificial Intelligence
 
Unit 1
Unit 1Unit 1
Unit 1
 
Business Outsourcing to Asia
Business Outsourcing to AsiaBusiness Outsourcing to Asia
Business Outsourcing to Asia
 
Sins2016
Sins2016Sins2016
Sins2016
 
Bringing AI to Business Intelligence
Bringing AI to Business IntelligenceBringing AI to Business Intelligence
Bringing AI to Business Intelligence
 
Artificial Intelligence Applications, Research, and Economics
Artificial Intelligence Applications, Research, and EconomicsArtificial Intelligence Applications, Research, and Economics
Artificial Intelligence Applications, Research, and Economics
 
Designing Trustable AI Experiences at IxDA Pittsburgh, Jan 2019
Designing Trustable AI Experiences at IxDA Pittsburgh, Jan 2019Designing Trustable AI Experiences at IxDA Pittsburgh, Jan 2019
Designing Trustable AI Experiences at IxDA Pittsburgh, Jan 2019
 
The Ethics of AI
The Ethics of AIThe Ethics of AI
The Ethics of AI
 
Introduction to Artificial Intelligence and Machine Learning: Ecosystem and T...
Introduction to Artificial Intelligence and Machine Learning: Ecosystem and T...Introduction to Artificial Intelligence and Machine Learning: Ecosystem and T...
Introduction to Artificial Intelligence and Machine Learning: Ecosystem and T...
 
What should students learn in an age of AI?
What should students learn in an age of AI?What should students learn in an age of AI?
What should students learn in an age of AI?
 

Viewers also liked

Ibm csoc response_future _ai 20160907 v5
Ibm csoc response_future _ai 20160907 v5Ibm csoc response_future _ai 20160907 v5
Ibm csoc response_future _ai 20160907 v5ISSIP
 
150810 Artificial Intelligence Policies
150810 Artificial Intelligence Policies150810 Artificial Intelligence Policies
150810 Artificial Intelligence PoliciesSamantha Hautea
 
Security_issues of AI
Security_issues of AISecurity_issues of AI
Security_issues of AI대선 최
 
Korea day1 keynote 20161013 v6
Korea day1 keynote 20161013 v6Korea day1 keynote 20161013 v6
Korea day1 keynote 20161013 v6ISSIP
 
Deep Learning - The Force of AI Awakens
Deep Learning - The Force of AI AwakensDeep Learning - The Force of AI Awakens
Deep Learning - The Force of AI AwakensPieter Buteneers
 
The Impact of Artificial Intelligence on the Built Environment
The Impact of Artificial Intelligence on the Built EnvironmentThe Impact of Artificial Intelligence on the Built Environment
The Impact of Artificial Intelligence on the Built EnvironmentJames Dearsley
 
Deep Learning for Public Safety in Chicago and San Francisco
Deep Learning for Public Safety in Chicago and San FranciscoDeep Learning for Public Safety in Chicago and San Francisco
Deep Learning for Public Safety in Chicago and San FranciscoSri Ambati
 
What Do You Do with a Problem Like AI?
What Do You Do with a Problem Like AI?What Do You Do with a Problem Like AI?
What Do You Do with a Problem Like AI?Lilian Edwards
 
How ai work, Abstract for dummies based on excellent artilce from The Economist
How ai work, Abstract for dummies based on excellent artilce from The EconomistHow ai work, Abstract for dummies based on excellent artilce from The Economist
How ai work, Abstract for dummies based on excellent artilce from The EconomistBertrand Petit
 
Chatbot AI Aeromexico (public)
Chatbot AI Aeromexico (public)Chatbot AI Aeromexico (public)
Chatbot AI Aeromexico (public)Brian Gross
 
AI & IoT in the development of smart cities
AI & IoT in the development of smart citiesAI & IoT in the development of smart cities
AI & IoT in the development of smart citiesRaunak Mundada
 
AWS re:Invent 2016: Machine Learning State of the Union Mini Con (MAC206)
AWS re:Invent 2016: Machine Learning State of the Union Mini Con (MAC206)AWS re:Invent 2016: Machine Learning State of the Union Mini Con (MAC206)
AWS re:Invent 2016: Machine Learning State of the Union Mini Con (MAC206)Amazon Web Services
 
Smart Cities in the AI Era
Smart Cities in the AI EraSmart Cities in the AI Era
Smart Cities in the AI EraNVIDIA
 
Design Ethics for Artificial Intelligence
Design Ethics for Artificial IntelligenceDesign Ethics for Artificial Intelligence
Design Ethics for Artificial IntelligenceCharbel Zeaiter
 

Viewers also liked (16)

Ibm csoc response_future _ai 20160907 v5
Ibm csoc response_future _ai 20160907 v5Ibm csoc response_future _ai 20160907 v5
Ibm csoc response_future _ai 20160907 v5
 
150810 Artificial Intelligence Policies
150810 Artificial Intelligence Policies150810 Artificial Intelligence Policies
150810 Artificial Intelligence Policies
 
Security_issues of AI
Security_issues of AISecurity_issues of AI
Security_issues of AI
 
Korea day1 keynote 20161013 v6
Korea day1 keynote 20161013 v6Korea day1 keynote 20161013 v6
Korea day1 keynote 20161013 v6
 
STOP AI training manual
STOP AI training manualSTOP AI training manual
STOP AI training manual
 
Deep Learning - The Force of AI Awakens
Deep Learning - The Force of AI AwakensDeep Learning - The Force of AI Awakens
Deep Learning - The Force of AI Awakens
 
The Impact of Artificial Intelligence on the Built Environment
The Impact of Artificial Intelligence on the Built EnvironmentThe Impact of Artificial Intelligence on the Built Environment
The Impact of Artificial Intelligence on the Built Environment
 
Deep Learning for Public Safety in Chicago and San Francisco
Deep Learning for Public Safety in Chicago and San FranciscoDeep Learning for Public Safety in Chicago and San Francisco
Deep Learning for Public Safety in Chicago and San Francisco
 
What Do You Do with a Problem Like AI?
What Do You Do with a Problem Like AI?What Do You Do with a Problem Like AI?
What Do You Do with a Problem Like AI?
 
How ai work, Abstract for dummies based on excellent artilce from The Economist
How ai work, Abstract for dummies based on excellent artilce from The EconomistHow ai work, Abstract for dummies based on excellent artilce from The Economist
How ai work, Abstract for dummies based on excellent artilce from The Economist
 
Chatbot AI Aeromexico (public)
Chatbot AI Aeromexico (public)Chatbot AI Aeromexico (public)
Chatbot AI Aeromexico (public)
 
AI & IoT in the development of smart cities
AI & IoT in the development of smart citiesAI & IoT in the development of smart cities
AI & IoT in the development of smart cities
 
AWS re:Invent 2016: Machine Learning State of the Union Mini Con (MAC206)
AWS re:Invent 2016: Machine Learning State of the Union Mini Con (MAC206)AWS re:Invent 2016: Machine Learning State of the Union Mini Con (MAC206)
AWS re:Invent 2016: Machine Learning State of the Union Mini Con (MAC206)
 
Smart Cities in the AI Era
Smart Cities in the AI EraSmart Cities in the AI Era
Smart Cities in the AI Era
 
Artificial Intelligence
Artificial IntelligenceArtificial Intelligence
Artificial Intelligence
 
Design Ethics for Artificial Intelligence
Design Ethics for Artificial IntelligenceDesign Ethics for Artificial Intelligence
Design Ethics for Artificial Intelligence
 

Similar to AI safety solutions

The Future is in Responsible Generative AI
The Future is in Responsible Generative AIThe Future is in Responsible Generative AI
The Future is in Responsible Generative AISaeed Al Dhaheri
 
What is AI ( Arificial Intelligence)?
What is AI ( Arificial Intelligence)?What is AI ( Arificial Intelligence)?
What is AI ( Arificial Intelligence)?MyAssignmenthelp.com
 
Generative AI - Responsible Path Forward.pdf
Generative AI - Responsible Path Forward.pdfGenerative AI - Responsible Path Forward.pdf
Generative AI - Responsible Path Forward.pdfSaeed Al Dhaheri
 
Global Governance of Generative AI: The Right Way Forward
Global Governance of Generative AI: The Right Way ForwardGlobal Governance of Generative AI: The Right Way Forward
Global Governance of Generative AI: The Right Way ForwardLilian Edwards
 
BB_54028_20160212 SVTLS AI_FINAL_shown
BB_54028_20160212 SVTLS AI_FINAL_shownBB_54028_20160212 SVTLS AI_FINAL_shown
BB_54028_20160212 SVTLS AI_FINAL_shownMark Holman
 
Managing Next Generation Threats to Cyber Security
Managing Next Generation Threats to Cyber SecurityManaging Next Generation Threats to Cyber Security
Managing Next Generation Threats to Cyber SecurityPriyanka Aash
 
AI for Everyone: Master the Basics
AI for Everyone: Master the BasicsAI for Everyone: Master the Basics
AI for Everyone: Master the BasicsStutty Srivastava
 
chapter 3 - Artificial Intelligence.pptx
chapter 3 - Artificial Intelligence.pptxchapter 3 - Artificial Intelligence.pptx
chapter 3 - Artificial Intelligence.pptxSisayNegash4
 
Intellectual property revisited
Intellectual property revisitedIntellectual property revisited
Intellectual property revisitedSerendipity Seraph
 
Intellectual property revisited
Intellectual property revisitedIntellectual property revisited
Intellectual property revisitedSerendipity Seraph
 
Regulating Artificial Intelligence
Regulating Artificial IntelligenceRegulating Artificial Intelligence
Regulating Artificial Intelligenceorrenprunckun
 
Marek Rosa - Inventing General Artificial Intelligence: A Vision and Methodology
Marek Rosa - Inventing General Artificial Intelligence: A Vision and MethodologyMarek Rosa - Inventing General Artificial Intelligence: A Vision and Methodology
Marek Rosa - Inventing General Artificial Intelligence: A Vision and MethodologyMachine Learning Prague
 
CIS 2015 The Ethics of Personal Data - Robin Wilton
CIS 2015 The Ethics of Personal Data - Robin WiltonCIS 2015 The Ethics of Personal Data - Robin Wilton
CIS 2015 The Ethics of Personal Data - Robin WiltonCloudIDSummit
 
Artificial intelligence- The science of intelligent programs
Artificial intelligence- The science of intelligent programsArtificial intelligence- The science of intelligent programs
Artificial intelligence- The science of intelligent programsDerak Davis
 
Deciphering AI: Human Expertise in the Age of Evolving AI
Deciphering AI: Human Expertise in the Age of Evolving AIDeciphering AI: Human Expertise in the Age of Evolving AI
Deciphering AI: Human Expertise in the Age of Evolving AILiming Zhu
 
AI - summary of focus groups.pdf
AI - summary of focus groups.pdfAI - summary of focus groups.pdf
AI - summary of focus groups.pdfDavid Wood
 
How to undermine AI dystopia
How to undermine AI dystopiaHow to undermine AI dystopia
How to undermine AI dystopiaBianca Ximenes
 
An introduction to the ethics of AI in education
An introduction to the ethics of AI in educationAn introduction to the ethics of AI in education
An introduction to the ethics of AI in educationJisc
 

Similar to AI safety solutions (20)

The Future is in Responsible Generative AI
The Future is in Responsible Generative AIThe Future is in Responsible Generative AI
The Future is in Responsible Generative AI
 
What is AI ( Arificial Intelligence)?
What is AI ( Arificial Intelligence)?What is AI ( Arificial Intelligence)?
What is AI ( Arificial Intelligence)?
 
Generative AI - Responsible Path Forward.pdf
Generative AI - Responsible Path Forward.pdfGenerative AI - Responsible Path Forward.pdf
Generative AI - Responsible Path Forward.pdf
 
Global Governance of Generative AI: The Right Way Forward
Global Governance of Generative AI: The Right Way ForwardGlobal Governance of Generative AI: The Right Way Forward
Global Governance of Generative AI: The Right Way Forward
 
BB_54028_20160212 SVTLS AI_FINAL_shown
BB_54028_20160212 SVTLS AI_FINAL_shownBB_54028_20160212 SVTLS AI_FINAL_shown
BB_54028_20160212 SVTLS AI_FINAL_shown
 
Managing Next Generation Threats to Cyber Security
Managing Next Generation Threats to Cyber SecurityManaging Next Generation Threats to Cyber Security
Managing Next Generation Threats to Cyber Security
 
Artificial Intelligence
Artificial IntelligenceArtificial Intelligence
Artificial Intelligence
 
AI for Everyone: Master the Basics
AI for Everyone: Master the BasicsAI for Everyone: Master the Basics
AI for Everyone: Master the Basics
 
chapter 3 - Artificial Intelligence.pptx
chapter 3 - Artificial Intelligence.pptxchapter 3 - Artificial Intelligence.pptx
chapter 3 - Artificial Intelligence.pptx
 
Intellectual property revisited
Intellectual property revisitedIntellectual property revisited
Intellectual property revisited
 
Intellectual property revisited
Intellectual property revisitedIntellectual property revisited
Intellectual property revisited
 
Regulating Artificial Intelligence
Regulating Artificial IntelligenceRegulating Artificial Intelligence
Regulating Artificial Intelligence
 
Marek Rosa - Inventing General Artificial Intelligence: A Vision and Methodology
Marek Rosa - Inventing General Artificial Intelligence: A Vision and MethodologyMarek Rosa - Inventing General Artificial Intelligence: A Vision and Methodology
Marek Rosa - Inventing General Artificial Intelligence: A Vision and Methodology
 
CIS 2015 The Ethics of Personal Data - Robin Wilton
CIS 2015 The Ethics of Personal Data - Robin WiltonCIS 2015 The Ethics of Personal Data - Robin Wilton
CIS 2015 The Ethics of Personal Data - Robin Wilton
 
Artificial intelligence- The science of intelligent programs
Artificial intelligence- The science of intelligent programsArtificial intelligence- The science of intelligent programs
Artificial intelligence- The science of intelligent programs
 
Deciphering AI: Human Expertise in the Age of Evolving AI
Deciphering AI: Human Expertise in the Age of Evolving AIDeciphering AI: Human Expertise in the Age of Evolving AI
Deciphering AI: Human Expertise in the Age of Evolving AI
 
AI - summary of focus groups.pdf
AI - summary of focus groups.pdfAI - summary of focus groups.pdf
AI - summary of focus groups.pdf
 
Unit1_AI&ML (2).pptx
Unit1_AI&ML (2).pptxUnit1_AI&ML (2).pptx
Unit1_AI&ML (2).pptx
 
How to undermine AI dystopia
How to undermine AI dystopiaHow to undermine AI dystopia
How to undermine AI dystopia
 
An introduction to the ethics of AI in education
An introduction to the ethics of AI in educationAn introduction to the ethics of AI in education
An introduction to the ethics of AI in education
 

More from avturchin

Fighting aging as effective altruism
Fighting aging as effective altruismFighting aging as effective altruism
Fighting aging as effective altruismavturchin
 
А.В.Турчин. Технологическое воскрешение умерших
А.В.Турчин. Технологическое воскрешение умершихА.В.Турчин. Технологическое воскрешение умерших
А.В.Турчин. Технологическое воскрешение умершихavturchin
 
Technological resurrection
Technological resurrectionTechnological resurrection
Technological resurrectionavturchin
 
Messaging future AI
Messaging future AIMessaging future AI
Messaging future AIavturchin
 
Future of sex
Future of sexFuture of sex
Future of sexavturchin
 
Backup on the Moon
Backup on the MoonBackup on the Moon
Backup on the Moonavturchin
 
Near term AI safety
Near term AI safetyNear term AI safety
Near term AI safetyavturchin
 
цифровое бессмертие и искусство
цифровое бессмертие и искусствоцифровое бессмертие и искусство
цифровое бессмертие и искусствоavturchin
 
Digital immortality and art
Digital immortality and artDigital immortality and art
Digital immortality and artavturchin
 
Nuclear submarines as global risk shelters
Nuclear submarines  as global risk  sheltersNuclear submarines  as global risk  shelters
Nuclear submarines as global risk sheltersavturchin
 
Искусственный интеллект в искусстве
Искусственный интеллект в искусствеИскусственный интеллект в искусстве
Искусственный интеллект в искусствеavturchin
 
ИИ как новая форма жизни
ИИ как новая форма жизниИИ как новая форма жизни
ИИ как новая форма жизниavturchin
 
Космос нужен для бессмертия
Космос нужен для бессмертияКосмос нужен для бессмертия
Космос нужен для бессмертияavturchin
 
AI in life extension
AI in life extensionAI in life extension
AI in life extensionavturchin
 
Levels of the self-improvement of the AI
Levels of the self-improvement of the AILevels of the self-improvement of the AI
Levels of the self-improvement of the AIavturchin
 
The map of asteroids risks and defence
The map of asteroids risks and defenceThe map of asteroids risks and defence
The map of asteroids risks and defenceavturchin
 
Herman Khan. About cobalt bomb and nuclear weapons.
Herman Khan. About cobalt bomb and nuclear weapons.Herman Khan. About cobalt bomb and nuclear weapons.
Herman Khan. About cobalt bomb and nuclear weapons.avturchin
 
The map of the methods of optimisation
The map of the methods of optimisationThe map of the methods of optimisation
The map of the methods of optimisationavturchin
 
Как достичь осознанных сновидений
Как достичь осознанных сновиденийКак достичь осознанных сновидений
Как достичь осознанных сновиденийavturchin
 
The map of natural global catastrophic risks
The map of natural global catastrophic risksThe map of natural global catastrophic risks
The map of natural global catastrophic risksavturchin
 

More from avturchin (20)

Fighting aging as effective altruism
Fighting aging as effective altruismFighting aging as effective altruism
Fighting aging as effective altruism
 
А.В.Турчин. Технологическое воскрешение умерших
А.В.Турчин. Технологическое воскрешение умершихА.В.Турчин. Технологическое воскрешение умерших
А.В.Турчин. Технологическое воскрешение умерших
 
Technological resurrection
Technological resurrectionTechnological resurrection
Technological resurrection
 
Messaging future AI
Messaging future AIMessaging future AI
Messaging future AI
 
Future of sex
Future of sexFuture of sex
Future of sex
 
Backup on the Moon
Backup on the MoonBackup on the Moon
Backup on the Moon
 
Near term AI safety
Near term AI safetyNear term AI safety
Near term AI safety
 
цифровое бессмертие и искусство
цифровое бессмертие и искусствоцифровое бессмертие и искусство
цифровое бессмертие и искусство
 
Digital immortality and art
Digital immortality and artDigital immortality and art
Digital immortality and art
 
Nuclear submarines as global risk shelters
Nuclear submarines  as global risk  sheltersNuclear submarines  as global risk  shelters
Nuclear submarines as global risk shelters
 
Искусственный интеллект в искусстве
Искусственный интеллект в искусствеИскусственный интеллект в искусстве
Искусственный интеллект в искусстве
 
ИИ как новая форма жизни
ИИ как новая форма жизниИИ как новая форма жизни
ИИ как новая форма жизни
 
Космос нужен для бессмертия
Космос нужен для бессмертияКосмос нужен для бессмертия
Космос нужен для бессмертия
 
AI in life extension
AI in life extensionAI in life extension
AI in life extension
 
Levels of the self-improvement of the AI
Levels of the self-improvement of the AILevels of the self-improvement of the AI
Levels of the self-improvement of the AI
 
The map of asteroids risks and defence
The map of asteroids risks and defenceThe map of asteroids risks and defence
The map of asteroids risks and defence
 
Herman Khan. About cobalt bomb and nuclear weapons.
Herman Khan. About cobalt bomb and nuclear weapons.Herman Khan. About cobalt bomb and nuclear weapons.
Herman Khan. About cobalt bomb and nuclear weapons.
 
The map of the methods of optimisation
The map of the methods of optimisationThe map of the methods of optimisation
The map of the methods of optimisation
 
Как достичь осознанных сновидений
Как достичь осознанных сновиденийКак достичь осознанных сновидений
Как достичь осознанных сновидений
 
The map of natural global catastrophic risks
The map of natural global catastrophic risksThe map of natural global catastrophic risks
The map of natural global catastrophic risks
 

Recently uploaded

Cultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxCultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxpradhanghanshyam7136
 
Animal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxAnimal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxUmerFayaz5
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...RohitNehra6
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​kaibalyasahoo82800
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSarthak Sekhar Mondal
 
VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PPRINCE C P
 
Unlocking the Potential: Deep dive into ocean of Ceramic Magnets.pptx
Unlocking  the Potential: Deep dive into ocean of Ceramic Magnets.pptxUnlocking  the Potential: Deep dive into ocean of Ceramic Magnets.pptx
Unlocking the Potential: Deep dive into ocean of Ceramic Magnets.pptxanandsmhk
 
Work, Energy and Power for class 10 ICSE Physics
Work, Energy and Power for class 10 ICSE PhysicsWork, Energy and Power for class 10 ICSE Physics
Work, Energy and Power for class 10 ICSE Physicsvishikhakeshava1
 
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...Sérgio Sacani
 
Behavioral Disorder: Schizophrenia & it's Case Study.pdf
Behavioral Disorder: Schizophrenia & it's Case Study.pdfBehavioral Disorder: Schizophrenia & it's Case Study.pdf
Behavioral Disorder: Schizophrenia & it's Case Study.pdfSELF-EXPLANATORY
 
Analytical Profile of Coleus Forskohlii | Forskolin .pdf
Analytical Profile of Coleus Forskohlii | Forskolin .pdfAnalytical Profile of Coleus Forskohlii | Forskolin .pdf
Analytical Profile of Coleus Forskohlii | Forskolin .pdfSwapnil Therkar
 
Grafana in space: Monitoring Japan's SLIM moon lander in real time
Grafana in space: Monitoring Japan's SLIM moon lander  in real timeGrafana in space: Monitoring Japan's SLIM moon lander  in real time
Grafana in space: Monitoring Japan's SLIM moon lander in real timeSatoshi NAKAHIRA
 
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.aasikanpl
 
Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )aarthirajkumar25
 
Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Patrick Diehl
 
Natural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsNatural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsAArockiyaNisha
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bSérgio Sacani
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)PraveenaKalaiselvan1
 

Recently uploaded (20)

Cultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxCultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptx
 
Animal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxAnimal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptx
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
 
9953056974 Young Call Girls In Mahavir enclave Indian Quality Escort service
9953056974 Young Call Girls In Mahavir enclave Indian Quality Escort service9953056974 Young Call Girls In Mahavir enclave Indian Quality Escort service
9953056974 Young Call Girls In Mahavir enclave Indian Quality Escort service
 
The Philosophy of Science
The Philosophy of ScienceThe Philosophy of Science
The Philosophy of Science
 
VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C P
 
Unlocking the Potential: Deep dive into ocean of Ceramic Magnets.pptx
Unlocking  the Potential: Deep dive into ocean of Ceramic Magnets.pptxUnlocking  the Potential: Deep dive into ocean of Ceramic Magnets.pptx
Unlocking the Potential: Deep dive into ocean of Ceramic Magnets.pptx
 
Work, Energy and Power for class 10 ICSE Physics
Work, Energy and Power for class 10 ICSE PhysicsWork, Energy and Power for class 10 ICSE Physics
Work, Energy and Power for class 10 ICSE Physics
 
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
All-domain Anomaly Resolution Office U.S. Department of Defense (U) Case: “Eg...
 
Behavioral Disorder: Schizophrenia & it's Case Study.pdf
Behavioral Disorder: Schizophrenia & it's Case Study.pdfBehavioral Disorder: Schizophrenia & it's Case Study.pdf
Behavioral Disorder: Schizophrenia & it's Case Study.pdf
 
Analytical Profile of Coleus Forskohlii | Forskolin .pdf
Analytical Profile of Coleus Forskohlii | Forskolin .pdfAnalytical Profile of Coleus Forskohlii | Forskolin .pdf
Analytical Profile of Coleus Forskohlii | Forskolin .pdf
 
Grafana in space: Monitoring Japan's SLIM moon lander in real time
Grafana in space: Monitoring Japan's SLIM moon lander  in real timeGrafana in space: Monitoring Japan's SLIM moon lander  in real time
Grafana in space: Monitoring Japan's SLIM moon lander in real time
 
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
Call Girls in Mayapuri Delhi 💯Call Us 🔝9953322196🔝 💯Escort.
 
Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )
 
Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?
 
Natural Polymer Based Nanomaterials
Natural Polymer Based NanomaterialsNatural Polymer Based Nanomaterials
Natural Polymer Based Nanomaterials
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)
 

AI safety solutions

  • 1. Do Nothing • AGI is impossible - Humans are unique, a soul exists - Existence of upper limits on an entity’s effective intelli- gence around the level of human geniuses - Individual intelligence is always outmatched by the dis- tributed intelligence found in social systems of many minds • AGI is too distant to be worth our attention - not in this century - not on our level of understanding of AI theory - the discussion has negative utility • Little risk, no action needed - AGIs will not have any particular motivation to act against us - AGI cannot do much - High intelligence implies high morality • Let it kill us - Humans are bad - Higher intelligence is more important than humanity - Quantum immortality will make only Safe AGI observ- able - AI is our child – we must step aside - AI will share some of our values and will construct an- cestor simulations which contain humans - We should not constrain AGI because it deserves full freedom to act • Safety measures are dangerous themselves - The more we try to confine it, the more AGI will try to escape AGI Safety Solutions Map Created by Alexey Turchin based on the article “Responses to Catastrophic AGI Risk: A Survey” by Kaj Sotala and Roman V. Yampolskiy, 2013 Ideas from the article are inside red border line. Ideas not derived from the article are marked to their authors in [...] “No” means that a method definitely will not work (it is point of view of Alexey Turchin and other people may have different opinions) “May be” means that small chances that a method may work (like 1 per cent) “Marginal” that a method may be used for short time or as additionally precaution measure but don’t change the situation much. “Promising” means that a method seems to be good enough to invest more in its research. © Alexey Turchin, 2015, GNU-like license. Free copying and design editing, discuss major updates with Alexey Turchin. All other roadmaps in last versions: http://immortalityroadmap.com/ and here: https://www.scribd.com/avturchin , https://www.scribd.com/doc/265853723/AGI-Safety-Solutions mailto: alexei.turchin@gmail.com Proofreading and advise: Michael Anissimov Societal Proposals External Constraints Internal Constraints AI used to create Safe AI Integrate into Society Marginal • Legal and Economic Controls: - AGIs will be law-abiding - Share resources - Taxation and regulation - Integrate into government • Foster Positive Values • State AGI: State transforms into AGI through gradual replacement of existing functions (computer tax regulation and law implementation) Regulate Research Marginal • International regulation - international treaty - boards - licensing research • Encourage research into safe AGI - funding for research - open-source approach • Differential Technological Progress - curtail unsafe AI architectures - prevent AGI arms race - one center of AGI development • International Mass Surveillance - narrow AIs as controllers of AGI research Enhance Human Capabilities Marginal • Create AGI by human upgrade - cognitive enhancement - moral enhancement • Keep humans on the same level as AGI AGI Confinement Marginal • Total isolation • Remote place like Mars or Pluto • Only text communication • Internet firewall • Virtual reality: many levels of isolation • “Safe Questions” (not implying dan- gerous behaviour) • Resetting AGI memory • Checks and balances - baits - check for unusual pattern in the AGI - circuit breakers - off switch • Cryptobox: homomorphic encryption [Christiano] Goal Checking Marginal • Automatically search for failure modes of goal system • List of signs of failure modes • Committee for goal checking • Prevent failure modes Philosophical Constrains Marginal • AGI fooled to be in simulation which could switch him off • Persuade AGI that humans have soul or other incomputable quality and it needs them • Prove that we really maybe in the simulation and AGI should keep peo- ple alive to not risk punishment from high-level simulators Prevent Self-improvement Promising • Limit number of self-improvement steps • Secure permission from human operator for each step of self-improvement • Slow rate of self-improvement • Limit areas of self-improvement • Limit access to own source code • Limit IQ of the AGI • Limit available hardware for the AGI • Limit working speed of the AGI • Prevent large architecture changes • Create as many AGIs as needed to prevent any of them from total power • Differential self-improvement: slow on high levels and quick on low level Oracle AI May be • Answer questions • No goals • No memory • Not one AI but many narrow Tool AIs with different goals • Slave AI – only obey commands Short rulesets May be • “Three laws by Asimov” • Categorical imperative • “Principle of voluntary joyous growth” • Utilitarianism • AGI obeys existing criminal laws • Short rulesets of other types Bottom-Up and Hybrid Safe AGI May be • Evolutionary invariants of ethics • Evolved morality • Reinforced learning of ethics • Learning values by pattern recognition • Human-like architecture Formal Verification May be • “Goal-oriented Learning Meta- Architecture” – search for better way to known goal [Goertzel] • Provable self-modifying algorithms AGI tasked to create Safe AGI May be • Oracle AI gives friendliness theory • Oracle AI designs safe AI • Oracle AI tests different AI designs • Oracle AI checks “AI constitution” for failure modes and prevents them • Group of uploads works on creating and proving “AI constitution” • Two-level AI: emulation for goals control, Tool AI for problem solving Timeless decision theory deal Marginal • Humans invest in creating AGI, AGI invests in being safe to humans • AGI is safe to humans because it expects that its own next versions will be safe to him AGI Enforcement Marginal • AGIs to watch other AGIs - Multilevel system - Community of AGIs • AI runs simulation with AGI inside • AGI actions are tested in simulation for unintended consequences • First AI checks safety of decisions of another AGI AGI Nanny May be • Singleton • No self-improvement • Prevent other AIs from becoming too powerful • Minimum rulesets: prevent death, x-risks and suffering, prohibition of forced brain reprogramming [Turchin] Relinquish Technology • Outlaw AGI - outlaw research on AI - nuclear war against AGI rivals [Turchin] • Restrict Hardware - limit processing power - outlaw supercomputers - control microchip fabs - switch off electricity or internet Control room and Gatekeeper Marginal • Monitors of internal processes • AI is constantly showing what it is think- ing about • Power switch • Explicit goals and subgoals • Firewall • Upgrade control • Mirror AIs to check for strange behaviour (3 copies of the AGI) • Seed AI testing for unintended behaviour Motivational Weaknesses Marginal • High Discount Rates for time or distance (don’t try to save Universe) • Easily satisfiable goals • Promised safety for the AGI • Calculated indifference to some threats • Legal machine language Coherent Extrapolated Volition May be • AI predicts what our values would be if we had more time to think about them (CEV) AI Сonstitution Promising AI is regulated by proven large set of rules; some possible rules: Meta rules: • Obey commands of its creator • Report and explain value changes • No secrets • Limit self-improvement • Stop after 10 years Content: • Do not kill people • Cure death • Prevent severe suffering • Prevent x-risks • Don’t rewrite human brains against their will Decision Theory Based Friendliness Marginal • Cooperate in prisoner dilemma • Don’t take actions which are irreversible • Anti-Pascaline agent [Armstrong] • Corrigible reasoning [MIRI] AGI without one Supergoal Marginal • Any one-goal system may be dangerous like “maniac” • Democracy of goals • Values not goals • AI capable on reflection on its goals and values • All possible goals in different proportions (like in humans) Situation Improvement Marginal • Funding of safe AGI research • Slowing development of AGI • Promotion of the idea of AGI safety • Ban military AGI • Friendly AGI training for AGI researchers • One main international AGI project, not many rivals • Safe AGI architectures: e.g. no genetic algorithms • Hire existing ethical researchers and use legal and ethical literature • Hire existing safe AI researchers AI as Next Step in Human Evolution Promising • AI based on emulations - Society of EM’s working at high speed - Upload a specially trained person with math skills and solid ethic as Seed AI - EM as value and awareness core of AGI • Use human-like value architecture: cultural upgrade of natural drives, reward system, complex values, many different architectures • AGI use humans (“me”) as a source of qualia (actuality) and preserve them • AI is based on many human brains connected together (neouronet) • Psychogenic singularity: use raw power of human brain and its unconsciousness to create AI (like savants) • AI is created by as gradual upgrade of one or several human beings (posthumans) • AI is created by genetic experiments like a child or superbrain • AI pretends to be best human (Armstrong) Simple solutions Complex solutions Value learning Promising • AGIs cooperate with humans in value learning • AI learns values by studying uploaded brain • Education of childlike Seed AI Multilevel Safe AGI Promising 1) Unchangeable core with constitution and values 2) ЕМ as a cloud around this core 3) Tool AIs as extension of EM 4) External confinement and control 5) Self-improvement is limited 6) Test its actions in virtual reality 7) Virtual jury for action judgment 8) Works as Nanny AI 9) Use it to predict failure modes of its own constitution Integrated Solutions Military Nanny AI Marginal Wins war for world domination, then become Nanny AI AGI Democracy May be • AGIs with different architectures suggest different proposals for AGI safety and vote on them [I.Artuhov] Other useful ideas • Value of human values • Secret way of testing right behaviour will result in obeying spirit of law [Armstrong] • Break AI into modules [Armstrong] • Prevent orthogonality of means and values (like in humans) [Phil Torres] • Approval-directed agents [Christiano] Logical or philosophical “landmines” which will stop too complex AI Marginal • Idea of absurdity of any goal system • AI halting problem • Lobian obstacle [Yudkowsky] • AGI is given false believes, which is known to it but it still obeys them [Armstromg] General requirements to the Safe AI theory • It must be easy to understand and implement so all researchers will prefer it • It must help AI to win first and to safely establish its power into Singleton • It should be computationally simple • It should have good chance to be really safe • Choose the ways to the safe AGI which are not mutually exclusive Meta Virtual jury • Virtual jury based on highly moral uploads checks AI’s decisions, goal system and CEV [Yudkowsky] AI Testing May be • Test AI inside multilevel simulation: if it breaks 1st level, terminate. • Test AI with difficult moral choices • Test AI decisions inside simulations for inundated consequences • Give AI idea that its decisions may be tested by unknown method • Create parallel AI which only test deci- sions of first AI, but is build on different principles (may be it is human upload or virtual jury) • AI self-improve in different branches, which later test each other for core values consistency • AI present full model of the future which would result from its decisions, and human operators (or uploads) validate them International AI agency • Choose the best AI safety theory and recommend its implementation Catch the moment of beginning of malignant AI self- improvement Marginal After AI acquires malignant goal system and starts self-improvement, it is still stupid and could be stopped Give AI to everybody Suggested by Elon Musk in OpenAI project Knowledge constrains • Separete source code of AI, ex- ecutable code, data and knowl- edge about how to build AI. • AI can’t phisically read its own source code or its attempts are recorded Signs constrains Identify all signs of possible wrongdoings by AI and make in- depended circut breakers. Signs: - Reading source code - Modifing own code - Creating AI models - Power consumption changes - Behavior patterns changes - Internal work pattern changes - Access to Internet on AI tipics - Attempts to cheat and lie - Goal system changes Opacity constrains Use opaque architecture which is difficult to understand and improve (some kinds of neural nets) Open goal system The actual goal system of an AI must be presented to exter- nal observer in the form of an open tree (text, graph) and any changes in it should be saved re- motely in unchangeble way. If the AI will try to create anoth- er goal tale it will be presented in first. Children constrain In order to selfupgrade, an AI must first create new version of itself and test it. Some AI-architectures may make it very complex task with many fails and resource consuming “education period” (neural nets) Technical level solution One can’t make safe rocket, without actually knowing how to build a rocket. The same may be true about AI. So we may need to wait until more will be known about most promissing AI architectures and use more help from actual developers. Record AI’s thought stream Build AI in the way that its high- level resaonong are presented in the open form and record them and analyze, may be using an- other AI. Intelligence distribution Create sitiation in which distributed collectibe intelli- gence of humanity is always greater than of any com- pany, or computer, so it can’t get desisive advantage. Distributed science Neuronet