Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.

The map of x-risk-preventing organizations, people and internet resources


Published on

The map of x-risk-preventing organizations, people and internet resources

Three known attempts to make a map of x-risks prevention in the field of science exist.
First is the list from the Global Catastrophic Risks Institute in 2012-2013, and many links there are already not working:

The second was done by S. Armstrong in 2014:

And the most beautiful and useful map was created by Andrew Critch: But its ecosystem ignores organizations which have a different view of the nature of global risks (that is, they share the value of x-risks prevention, but have another world view).

In my map I have tried to add all currently active organizations which share the value of global risks prevention.

It also regards some active independent people as organizations, if they have an important blog or field of research, but not all people are mentioned in the map. If you think that you (or someone) should be in it, please write to me at
I used only open sources and public statements to learn about people and organizations, so I can’t provide information on the underlying net of relations.

I tried to give all organizations a short description based on its public statement and also my opinion about its activity.

In general it seems that all small organizations are focused on their collaboration with larger ones, that is MIRI and FHI, and small organizations tend to ignore each other; this is easily explainable from the social singnaling theory. Another explanation is that larger organizations have a great ability to make contacts.

It also appears that there are several organizations with similar goal statements.

It looks like the most cooperation exists in the field of AI safety, but most of the structure of this cooperation is not visible to the external viewer, in contrast to Wikipedia, where contributions of all individuals are visible.

It seems that the community in general lacks three things: a united internet forum for public discussion, an x-risks wikipedia and an x-risks related scientific journal.

Ideally, a forum should be used to brainstorm ideas, a scientific journal to publish the best ideas, peer review them and present them to the outer scientific community, and a wiki to collect results.

Currently it seems more like each organization is interested in creating its own research and hoping that someone will read it. Each small organization seems to want to be the only one to present the solutions to global problems and gain full attention from the UN and governments. It raises the problem of noise and rivalry; and also raises the problem of possible incompatible solutions, especially in AI safety.
The pdf is here:

Published in: Science
  • Be the first to comment

  • Be the first to like this

The map of x-risk-preventing organizations, people and internet resources

  1. 1. Organizations, sites and people involved in x-risks prevention AI risks General x-risks Nuclear Global warming Bio-risksSize and level of influence MIRI (Former Singularity insti- tute) E.Yudkowsky link FHI Future of humanity institute Oxford, link Nick Bostrom CSER Cambridge center of existential risks Martin Rees, link FLI Future of life institute Elon Musk link GCRI Global catastrophic risks institute, Seth Baum Interesting newsletter, many articles in scientific journals link X-risks institute Phil Torrres Interesting articles by its main author, focus on exis- tential terrorism and religion link Convergence Justin Shovelain Collective think tank concentrat- ed on mathematical modeling of x-risks link Lifeboat foundation Very large scientific boards which don’t actually do anything, but some useful discussion is going in its mail list site mail list - good one X-risks on Reddit Existentiarisk Control problem Willard Wells Author of “Apoc- alypses when” and prevention plan Bulletin of atomic scientists Famous doomsday clock link Arctic news Sam Carana Irreversible global warming because of methane hydrates eruption link World Health Organization (WHO) includes a division called the Global Alert and Response (GAR) which mon- itors and responds to global epidemic crisis. GAR helps member states with training and coordination of response to epidemics. link Global challenges Laszlo Szombatfalvy The Global Challenges Foundation works to raise awareness of the Global Catastrophic Risks. Primari- ly focused on climate change, other environmental degradation and po- litically motivated violence as well as how these threats are linked to poverty and rapid population growth, link Known very well and large amount of work is done Large and interest- ing research has been cre- ated, but not much people knows them Open places for discussion Longecity subforum link Foundational research institute “Currently, our research focus- es on reducing risks of dysto- pian futures in the context of emerging technologies”. Interesting work on AI safety link Club of Rome Still exist! Were famous in 1970s when they pro- duced “Limits of growth” Link CISAC “The Center for Internation- al Security and Cooperation is Stanford University’s hub for re- searchers tackling some of the world’s most pressing securi- ty and international cooperation problems” Nuclear, cybersecurity, bio, anti- terrorism, link EA Effective altruism EA forum Facebook • Existential risks (Adam Ford) • Global Catastrophic Risks Re- search and Discussion (Evan Gaensbauer) • Global catastrophic risks • Stop existential risks Nuclear threat initia- tive link Saving Humanity from Homo Sapiens Small one person organisation without any actual work link Jaan Tallinn investor in x-re- lated projects, wiki Skoll Global Threats Fund “To safeguard humanity from global threats” Climate, water security, pandem- ics, nuclear proliferation, link Stimson Center “The Stimson Center is a nonpartisan policy research center working to solve the world’s greatest threats to security and prosperity”. non-prolifiration link Investors and important figures Peter Thiel Invested in MIRI Elon Musk Want AI safety through Open AI and human on Mars as a backup plan Scientists and re- searchers Less- wrong Adrian Kent LHC risks Tobi Ord site “existential hope” Impact risks Robin Hanson Blog Societal collapse risks Katja Grace Fermi paradox and DA blog AI impacts A. Sandberg Participated in FHI and co-au- thored papers R. Freitas Nanotech risks Ploughshares Fund “Supports the smartest minds and most effective or- ganizations to reduce nucle- ar stockpiles, prevent new nuclear states, and increase global security” Link David Denkenberger agricultural risks Alexander Kononov Coined term ”indestructibility of civilization” Milan Circovic Stevenson probe, Anthropic shadow Fermi paradox Site Bruce Tonn Editor and writer link Large group of people working on AI safety, including, but not limited to: Steve Omohundro Luke Muehlhauser Stuart Armstrong Roman Yampolsky Nate Soares Vladimir Nesov Kaj Sotala Benja Fallenstein Riva Melissa Tez Jason Gaverick Matheny, wiki Andrew Critch, blog Paul Christiano Karl Shulman Anna Salamon Bill Gates Has its own foun- dation and vision of global risks Forum about risks of flu pandemic OpenAI Elon Musk wiki Stephen Hawking Warned about risks of aliens and AI Dennis Medows Resource depletion risks David Brin writer, “Existence” Vernor Vinge writer, created Singularity idea Writers IPCC International panel of climate change Zoltvan Istavn Presidential candidate from transhumanist party Wrote about x-risks Greg Igen writer “Permutation city” Leveraged research the site is almost empty now link Global Priori- ties Project Created Global catastrophic risk report-2016 Collaborate with UK govern- ment Dr. Toby Ord is member Connected with EA movement link X-risks net Alexei Turchin Creating full database on x-risks and prevention plan link Wiki- resources LW-wiki Nano risks Foresight institute link Holocen impact working group Estimate risks of recent impacts link Diffusing nuclear threat link Norvegian transhumanists Oxford Martin Programme on the Impacts of Future Technology, link NASAlink The United States Agency for Interna- tional Development (USAID) has its Emerging Pandemic Threats Program which aims to prevent and contain naturally generated pandemics at their source.[129] The Lawrence Livermore National Laboratory has a division called the Global Secu- rity Principal Directorate which re- searches on behalf of the government issues such as bio-security, counter-ter- rorism, etc. Link Public figures John Barnes “Mother of strorms” Aaron Dar Risks of supernovas Bill Napier Risks of dark comets Sam Altman Y combinator, Confounded Open AI R.Carrigan Risks of SETI Bill Joy Wrote famous arti- cle but now seems to lost interest Max Tegmark Wrote articles together with Bostrom EA forum link Intelligent agents forum Technical discussion on AI safety link Discussion in comments IEET Futureoflife