Taxonomy of Pathways to Dangerous Artificial Intelligence
2016
Abstract
In order to properly handle a dangerous Artificially Intelligent (AI) system it is important to understand how the system came to be in such a state. In popular culture (science fiction movies/books) AIs/Robots became self-aware and as a result rebel against humanity and decide to destroy it. While it is one possible scenario, it is probably the least likely path to appearance of dangerous AI. In this work, we survey, classify and analyze a number of circumstances, which might lead to arrival of malicious AI. To the best of our knowledge, this is the first attempt to systematically classify types of pathways leading to malevolent AI. Previous relevant work either surveyed specific goals/meta-rules which might lead to malevolent behavior in AIs (Ozkural 2014) or reviewed specific undesirable behaviors AGIs can exhibit at different stages of its development (Turchin July 10 2015a, Turchin July 10, 2015b).
References (44)
- Anonymous (2013). The Scientists' Call… To Ban Autonomous Lethal Robots. ICRAC International Committee for Robot Arms Control. http://icrac.net/call
- Bostrom, N. (2006 ). "What is a Singleton?" Linguistic and Philosophical Investigations 5(2): 48-54.
- Bostrom, N. (2011). "Information Hazards: A Typology of Potential Harms From Knowledge." Review of Contemporary Philosophy 10: 44-79.
- Bostrom, N. (2014). Superintelligence: Paths, dangers, strategies, Oxford University Press.
- Carrigan Jr, R. A. (2004). The Ultimate Hacker: SETI signals may need to be decontaminated. Bioastronomy 2002: Life Among the Stars.
- Castelfranchi, C. (2000). "Artificial liars: Why computers will (necessarily) deceive us and each other." Ethics and Information Technology 2(2): 113-119.
- Ćirković, M. M. (2015). "Linking simulation argument to the AI risk." Futures 72: 27-31.
- Clark, M. H. (2010). Cognitive illusions and the lying machine: a blueprint for sophistic mendacity, Rensselaer Polytechnic Institute.
- Cush, A. (January 22, 2015). Swiss Authorities Arrest Bot for Buying Drugs and Fake Passport. Gawker. http://internet.gawker.com/swiss-authorities-arrest-bot-for- buying-drugs-and-a-fak-1681098991.
- Dewey, D., S. Russell, M. Tegmark and e. al (2015). A Survey of Research Questions for Robust and Beneficial AI. Future of Life Institute. Available at: http://futureoflife.org/static/data/documents/research_survey.pdf.
- Eshelman, R. and D. Derrick (2015). "Relying on Kindness of Machines? The Security Threat of Artificial Agents." JFQ 77 2nd Quarter.
- Floridi, L. and J. W. Sanders (2001). "Artificial evil and the foundation of computer ethics." Ethics and Information Technology 3(1): 55-66.
- Frey, T. (June 2015). The Black Hat Robots are Coming. Futurist Speaker. http://www.futuristspeaker.com/2015/06/the-black-hat- robots-are-coming/.
- Goertzel, B. (2012). "Should Humanity Build a Global AI Nanny to Delay the Singularity Until It's Better Understood?" Journal of consciousness studies 19(1-2): 96-111.
- Goldhill, O. (May 12, 2015). Artificial intelligence experts are building the world's angriest robot. Should you be scared? The Telegraph. http://www.telegraph.co.uk/men/the- filter/11600593/Artificial-intelligence-should-you-be-scared-of- angry-robots.html.
- Johnson, N., G. Zhao, E. Hunsader, H. Qi, N. Johnson, J. Meng and B. Tivnan (2013). "Abrupt rise of new machine ecology beyond human response time." Scientific reports 3.
- Joy, B. (April 2000). "Why the Future Doesn't Need Us." Wired Magazine 8(4).
- Lieberman, H., A. Faaborg, W. Daher and J. Espinosa (2005). How to wreck a nice beach you sing calm incense. Proceedings of the 10th international conference on Intelligent user interfaces, ACM.
- Loosemore, R. P. (2014). The Maverick Nanny with a Dopamine Drip: Debunking Fallacies in the Theory of AI Motivation. 2014 AAAI Spring Symposium Series.
- Majot, A. M. and R. V. Yampolskiy (2014). AI safety engineering through introduction of self-reference into felicific calculus via artificial pain and pleasure. 2014 IEEE International Symposium on Ethics in Science, Technology and Engineering, IEEE.
- Nijholt, A. (2011). "No grice: computers that lie, deceive and conceal."
- Omohundro, S. (2012). Rational artificial intelligence for the greater good. Singularity Hypotheses, Springer: 161-179.
- Özkural, E. (2014). "Godseed: Benevolent or Malevolent?" arXiv preprint arXiv:1402.5380.
- Picard, R. W. and R. Picard (1997). Affective computing, MIT press Cambridge.
- Russell, S., D. Dewey, M. Tegmark and e. al (January 23, 2015). Research Priorities for Robust and Benecial Artificial Intelligence. Future of Life Institute. http://futureoflife.org/static/data/documents/research_priorities.pd f. Searle, J. R. (October 9, 2014). What Your Computer Can't Know. The New York Review of Books. http://www.nybooks.com/articles/archives/2014/oct/09/what- your-computer-cant-know.
- Simonite, T. (March 7, 2008). Should every computer chip have a cosmic ray detector? . New Scientist. https://www.newscientist.com/blog/technology/2008/03/do-we- need-cosmic-ray-alerts-for.html.
- Smith, D. (January 10, 2013). IBM's Watson Gets A 'Swear Filter' After Learning The Urban Dictionary. International Business Times. http://www.ibtimes.com/ibms-watson-gets-swear-filter- after-learning-urban-dictionary-1007734.
- Soares, N. and B. Fallenstein (2014). Aligning Superintelligence with Human Interests: A Technical Research Agenda, Tech. rep. Machine Intelligence Research Institute, 2014. URL: http://intelligence. org/files/TechnicalAgenda.pdf.
- Soares, N., B. Fallenstein, S. Armstrong and E. Yudkowsky (January 25-30, 2015). Corrigibility. Workshops at the Twenty- Ninth AAAI Conference on Artificial Intelligence. Austin, Texas, USA. Sotala, K. and R. V. Yampolskiy (2015). "Responses to catastrophic AGI risk: a survey." Physica Scripta 90(1): 018001.
- Turchin, A. (July 10 2015a). A Map: AGI Failures Modes and Levels. LessWrong. http://lesswrong.com/lw/mgf/a_map_agi_failures_modes_and_le vels/.
- Turchin, A. (July 10, 2015b). Human Extinction Risks due to Artificial Intelligence Development -55 ways we can be obliterated. IEET. http://ieet.org/index.php/IEET/more/turchin20150610.
- Turchin, A. (March 15, 2013). Risks of downloading alien AI via SETI search. LessWrong. http://lesswrong.com/lw/gzv/risks_of_downloading_alien_ai_via _seti_search/.
- Waser, M. (2011). Rational Universal Benevolence: Simpler, Safer, and Wiser Than "Friendly AI". Artificial General Intelligence, Springer: 153-162.
- Yampolskiy, R. (2012). "Leakproofing the Singularity Artificial Intelligence Confinement Problem." Journal of Consciousness Studies 19(1-2): 1-2.
- Yampolskiy, R. and J. Fox (2012). "Safety Engineering for Artificial General Intelligence." Topoi: 1-10.
- Yampolskiy, R. V. (2013). Artificial intelligence safety engineering: Why machine ethics is a wrong approach. Philosophy and Theory of Artificial Intelligence, Springer Berlin Heidelberg: 389-396.
- Yampolskiy, R. V. (2014). "Utility Function Security in Artificially Intelligent Agents." Journal of Experimental and Theoretical Artificial Intelligence (JETAI): 1-17.
- Yampolskiy, R. V. (2015). Analysis of types of self-improving software. Artificial General Intelligence, Springer: 384-393.
- Yampolskiy, R. V. (2015). Artificial Superintelligence: a Futuristic Approach, Chapman and Hall/CRC.
- Yampolskiy, R. V. (2015). On the limits of recursively self- improving AGI. Artificial General Intelligence, Springer: 394-403.
- Yampolskiy, R. V. (2015). The Space of Possible Mind Designs. Artificial General Intelligence, Springer: 218-227.
- Yampolskiy, R. V. (October 3-4, 2011). What to Do with the Singularity Paradox? Philosophy and Theory of Artificial Intelligence (PT-AI2011). Thessaloniki, Greece.
- Yudkowsky, E. (2011). Complex Value Systems in Friendly AI. Artificial General Intelligence. J. Schmidhuber, K. Thórisson and M. Looks, Springer Berlin / Heidelberg. 6830: 388-393.
- Yudkowsky, E. S. (May 2004 ). Coherent Extrapolated Volition. Available at: http://singinst.org/upload/CEV.html, Singularity Institute for Artificial Intelligence.