This article may contain wording that promotes the subject through exaggeration of unnoteworthy facts. Please help improve it by removing or replacing such wording. (March 2023) (Learn how and when to remove this template message)
Eliezer Yudkowsky
Eliezer Yudkowsky, Stanford 2006 (square crop).jpg
Yudkowsky at Stanford University in 2006
Eliezer Shlomo Yudkowsky

(1979-09-11) September 11, 1979 (age 43)
OrganizationMachine Intelligence Research Institute
Known forCoining the term friendly artificial intelligence
Research on AI safety
Rationality writing
Founder of LessWrong

Eliezer Shlomo Yudkowsky (born September 11, 1979) is an American decision theory and artificial intelligence (AI) researcher and writer, best known for popularizing the idea of friendly artificial intelligence.[1][2] He is a co-founder[3] and research fellow at the Machine Intelligence Research Institute (MIRI), a private research nonprofit based in Berkeley, California.[4] His work on the prospect of a runaway intelligence explosion was an influence on Nick Bostrom's Superintelligence: Paths, Dangers, Strategies.[5]

Work in artificial intelligence safety

See also: Machine Intelligence Research Institute

Goal learning and incentives in software systems

Yudkowsky's views on the safety challenges posed by future generations of AI systems are discussed in the undergraduate textbook in AI, Stuart Russell and Peter Norvig's Artificial Intelligence: A Modern Approach. Noting the difficulty of formally specifying general-purpose goals by hand, Russell and Norvig cite Yudkowsky's proposal that autonomous and adaptive systems be designed to learn correct behavior over time:

Yudkowsky (2008)[6] goes into more detail about how to design a Friendly AI. He asserts that friendliness (a desire not to harm humans) should be designed in from the start, but that the designers should recognize both that their own designs may be flawed, and that the robot will learn and evolve over time. Thus the challenge is one of mechanism design—to design a mechanism for evolving AI under a system of checks and balances, and to give the systems utility functions that will remain friendly in the face of such changes.[1]

In response to the instrumental convergence concern, where autonomous decision-making systems with poorly designed goals would have default incentives to mistreat humans, Yudkowsky and other MIRI researchers have recommended that work be done to specify software agents that converge on safe default behaviors even when their goals are misspecified.[7][2]

Capabilities forecasting

In the intelligence explosion scenario hypothesized by I. J. Good, recursively self-improving AI systems quickly transition from subhuman general intelligence to superintelligent. Nick Bostrom's 2014 book Superintelligence: Paths, Dangers, Strategies sketches out Good's argument in detail, while citing writing by Yudkowsky on the risk that anthropomorphizing advanced AI systems will cause people to misunderstand the nature of an intelligence explosion. "AI might make an apparently sharp jump in intelligence purely as the result of anthropomorphism, the human tendency to think of 'village idiot' and 'Einstein' as the extreme ends of the intelligence scale, instead of nearly indistinguishable points on the scale of minds-in-general."[1][3][6][8]

In Artificial Intelligence: A Modern Approach, authors Stuart Russell and Peter Norvig raise the objection that there are known limits to intelligent problem-solving from computational complexity theory; if there are strong limits on how efficiently algorithms can solve various computer science tasks, then intelligence explosion may not be possible.[1]

Rationality writing

Between 2006 and 2009, Yudkowsky and Robin Hanson were the principal contributors to Overcoming Bias, a cognitive and social science blog sponsored by the Future of Humanity Institute of Oxford University. In February 2009, Yudkowsky founded LessWrong, a "community blog devoted to refining the art of human rationality".[9][10] Overcoming Bias has since functioned as Hanson's personal blog.

Over 300 blogposts by Yudkowsky on philosophy and science (originally written on LessWrong and Overcoming Bias) were released as an ebook entitled Rationality: From AI to Zombies by the Machine Intelligence Research Institute (MIRI) in 2015.[11] MIRI has also published Inadequate Equilibria, Yudkowsky's 2017 ebook on the subject of societal inefficiencies.[12]

Yudkowsky has also written several works of fiction. His fanfiction novel, Harry Potter and the Methods of Rationality, uses plot elements from J. K. Rowling's Harry Potter series to illustrate topics in science.[9][13] The New Yorker described Harry Potter and the Methods of Rationality as a retelling of Rowling's original "in an attempt to explain Harry's wizardry through the scientific method".[14]

Personal life

Yudkowsky is an autodidact[15] and did not attend high school or college.[16] He was raised as a Modern Orthodox Jew.[17]

Academic publications

See also


  1. ^ a b c d Russell, Stuart; Norvig, Peter (2009). Artificial Intelligence: A Modern Approach. Prentice Hall. ISBN 978-0-13-604259-4.
  2. ^ a b Leighton, Jonathan (2011). The Battle for Compassion: Ethics in an Apathetic Universe. Algora. ISBN 978-0-87586-870-7.
  3. ^ a b Dowd, Maureen (March 26, 2017). "Elon Musk's Billion-Dollar Crusade to Stop the A.I. Apocalypse". Vanity Fair. Archived from the original on July 26, 2018. Retrieved July 28, 2018.
  4. ^ Kurzweil, Ray (2005). The Singularity Is Near. New York City: Viking Penguin. ISBN 978-0-670-03384-3.
  5. ^ Ford, Paul (February 11, 2015). "Our Fear of Artificial Intelligence". MIT Technology Review. Archived from the original on March 30, 2019. Retrieved April 9, 2019.
  6. ^ a b Yudkowsky, Eliezer (2008). "Artificial Intelligence as a Positive and Negative Factor in Global Risk" (PDF). In Bostrom, Nick; Ćirković, Milan (eds.). Global Catastrophic Risks. Oxford University Press. ISBN 978-0199606504. Archived (PDF) from the original on March 2, 2013. Retrieved October 16, 2015.
  7. ^ Soares, Nate; Fallenstein, Benja; Yudkowsky, Eliezer (2015). "Corrigibility". AAAI Workshops: Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, January 25–26, 2015. AAAI Publications. Archived from the original on January 15, 2016. Retrieved October 16, 2015.
  8. ^ Bostrom, Nick (2014). Superintelligence: Paths, Dangers, Strategies. ISBN 978-0199678112.
  9. ^ a b Miller, James (2012). Singularity Rising. BenBella Books, Inc. ISBN 978-1936661657.
  10. ^ Miller, James (July 28, 2011). "You Can Learn How To Become More Rational". Business Insider. Archived from the original on August 10, 2018. Retrieved March 25, 2014.
  11. ^ Miller, James D. "Rifts in Rationality – New Rambler Review". Archived from the original on July 28, 2018. Retrieved July 28, 2018.
  12. ^ Machine Intelligence Research Institute. "Inadequate Equilibria: Where and How Civilizations Get Stuck". Archived from the original on September 21, 2020. Retrieved May 13, 2020.
  13. ^ Snyder, Daniel D. (July 18, 2011). "'Harry Potter' and the Key to Immortality". The Atlantic. Archived from the original on December 23, 2015. Retrieved June 13, 2022.
  14. ^ Packer, George (2011). "No Death, No Taxes: The Libertarian Futurism of a Silicon Valley Billionaire". The New Yorker. p. 54. Archived from the original on December 14, 2016. Retrieved October 12, 2015.
  15. ^ Matthews, Dylan; Pinkerton, Byrd (June 19, 2019). "He co-founded Skype. Now he's spending his fortune on stopping dangerous AI". Vox. Archived from the original on March 6, 2020. Retrieved March 22, 2020.
  16. ^ Saperstein, Gregory (August 9, 2012). "5 Minutes With a Visionary: Eliezer Yudkowsky". CNBC. Archived from the original on August 1, 2017. Retrieved September 9, 2017.
  17. ^ Yudkowsky, Eliezer (October 4, 2007). "Avoiding your belief's real weak points". LessWrong. Archived from the original on May 2, 2021. Retrieved April 30, 2021.