Eliezer Yudkowsky

Eliezer S.

Yudkowsky (/ˌɛliˈɛzər ˌjʌdˈkski/ EH-lee-EH-zər YUD-KOW-skee; born September 11, 1979) is an American artificial intelligence researcher and writer on decision theory and ethics, best known for popularizing ideas related to friendly artificial intelligence, including the idea that there might not be a "fire alarm" for AI. He is the founder of and a research fellow at the Machine Intelligence Research Institute (MIRI), a private research nonprofit based in Berkeley, California. His work on the prospect of a runaway intelligence explosion influenced philosopher Nick Bostrom's 2014 book Superintelligence: Paths, Dangers, Strategies.

Eliezer Yudkowsky
Eliezer Yudkowsky
Yudkowsky at Stanford University in 2006
Born
Eliezer Shlomo Yudkowsky

(1979-09-11) September 11, 1979 (age 44)
OrganizationMachine Intelligence Research Institute
Known forCoining the term friendly artificial intelligence
Research on AI safety
Rationality writing
Founder of LessWrong
Websitewww.yudkowsky.net

Work in artificial intelligence safety

Goal learning and incentives in software systems

Yudkowsky's views on the safety challenges future generations of AI systems pose are discussed in Stuart Russell's and Peter Norvig's undergraduate textbook Artificial Intelligence: A Modern Approach. Noting the difficulty of formally specifying general-purpose goals by hand, Russell and Norvig cite Yudkowsky's proposal that autonomous and adaptive systems be designed to learn correct behavior over time:

Yudkowsky (2008) goes into more detail about how to design a Friendly AI. He asserts that friendliness (a desire not to harm humans) should be designed in from the start, but that the designers should recognize both that their own designs may be flawed, and that the robot will learn and evolve over time. Thus the challenge is one of mechanism design—to design a mechanism for evolving AI under a system of checks and balances, and to give the systems utility functions that will remain friendly in the face of such changes.

In response to the instrumental convergence concern, that autonomous decision-making systems with poorly designed goals would have default incentives to mistreat humans, Yudkowsky and other MIRI researchers have recommended that work be done to specify software agents that converge on safe default behaviors even when their goals are misspecified.

Capabilities forecasting

In the intelligence explosion scenario hypothesized by I. J. Good, recursively self-improving AI systems quickly transition from subhuman general intelligence to superintelligent. Nick Bostrom's 2014 book Superintelligence: Paths, Dangers, Strategies sketches out Good's argument in detail, while citing Yudkowsky on the risk that anthropomorphizing advanced AI systems will cause people to misunderstand the nature of an intelligence explosion. "AI might make an apparently sharp jump in intelligence purely as the result of anthropomorphism, the human tendency to think of 'village idiot' and 'Einstein' as the extreme ends of the intelligence scale, instead of nearly indistinguishable points on the scale of minds-in-general."

Eliezer Yudkowsky 
Eliezer debating Destiny at Manifest 2023

In Artificial Intelligence: A Modern Approach, Russell and Norvig raise the objection that there are known limits to intelligent problem-solving from computational complexity theory; if there are strong limits on how efficiently algorithms can solve various tasks, an intelligence explosion may not be possible.

Time op-ed

In a 2023 op-ed for Time magazine, Yudkowsky discussed the risk of artificial intelligence and proposed action that could be taken to limit it, including a total halt on the development of AI, or even "destroy[ing] a rogue datacenter by airstrike". The article helped introduce the debate about AI alignment to the mainstream, leading a reporter to ask President Joe Biden a question about AI safety at a press briefing.

Rationality writing

Between 2006 and 2009, Yudkowsky and Robin Hanson were the principal contributors to Overcoming Bias, a cognitive and social science blog sponsored by the Future of Humanity Institute of Oxford University. In February 2009, Yudkowsky founded LessWrong, a "community blog devoted to refining the art of human rationality". Overcoming Bias has since functioned as Hanson's personal blog.

Over 300 blog posts by Yudkowsky on philosophy and science (originally written on LessWrong and Overcoming Bias) were released as an ebook, Rationality: From AI to Zombies, by MIRI in 2015. MIRI has also published Inadequate Equilibria, Yudkowsky's 2017 ebook on societal inefficiencies.

Yudkowsky has also written several works of fiction. His fanfiction novel Harry Potter and the Methods of Rationality uses plot elements from J. K. Rowling's Harry Potter series to illustrate topics in science. The New Yorker described Harry Potter and the Methods of Rationality as a retelling of Rowling's original "in an attempt to explain Harry's wizardry through the scientific method".

Personal life

Yudkowsky is an autodidact and did not attend high school or college. He was raised as a Modern Orthodox Jew, but does not identify religiously as a Jew.

Academic publications

  • Yudkowsky, Eliezer (2007). "Levels of Organization in General Intelligence" (PDF). Artificial General Intelligence. Berlin: Springer.
  • Yudkowsky, Eliezer (2008). "Cognitive Biases Potentially Affecting Judgement of Global Risks" (PDF). In Bostrom, Nick; Ćirković, Milan (eds.). Global Catastrophic Risks. Oxford University Press. ISBN 978-0199606504.
  • Yudkowsky, Eliezer (2008). "Artificial Intelligence as a Positive and Negative Factor in Global Risk" (PDF). In Bostrom, Nick; Ćirković, Milan (eds.). Global Catastrophic Risks. Oxford University Press. ISBN 978-0199606504.
  • Yudkowsky, Eliezer (2011). "Complex Value Systems in Friendly AI" (PDF). Artificial General Intelligence: 4th International Conference, AGI 2011, Mountain View, CA, USA, August 3–6, 2011. Berlin: Springer.
  • Yudkowsky, Eliezer (2012). "Friendly Artificial Intelligence". In Eden, Ammon; Moor, James; Søraker, John; et al. (eds.). Singularity Hypotheses: A Scientific and Philosophical Assessment. The Frontiers Collection. Berlin: Springer. pp. 181–195. doi:10.1007/978-3-642-32560-1_10. ISBN 978-3-642-32559-5.
  • Bostrom, Nick; Yudkowsky, Eliezer (2014). "The Ethics of Artificial Intelligence" (PDF). In Frankish, Keith; Ramsey, William (eds.). The Cambridge Handbook of Artificial Intelligence. New York: Cambridge University Press. ISBN 978-0-521-87142-6.
  • LaVictoire, Patrick; Fallenstein, Benja; Yudkowsky, Eliezer; Bárász, Mihály; Christiano, Paul; Herreshoff, Marcello (2014). "Program Equilibrium in the Prisoner's Dilemma via Löb's Theorem". Multiagent Interaction without Prior Coordination: Papers from the AAAI-14 Workshop. AAAI Publications. Archived from the original on April 15, 2021. Retrieved October 16, 2015.
  • Soares, Nate; Fallenstein, Benja; Yudkowsky, Eliezer (2015). "Corrigibility" (PDF). AAAI Workshops: Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, January 25–26, 2015. AAAI Publications.

See also

Notes

References

Tags:

Eliezer Yudkowsky Work in artificial intelligence safetyEliezer Yudkowsky Rationality writingEliezer Yudkowsky Personal lifeEliezer Yudkowsky Academic publicationsEliezer Yudkowsky

🔥 Trending searches on Wiki English:

Monk (TV series)Holy WeekScarlett JohanssonQueen VictoriaEmma StoneBenji Madden2024 in filmArgentina national football teamBeslan school siegeMarylandNATOAlyson HanniganMarch 2024 lunar eclipseMarilyn MonroeCape Verde2024 NCAA Division I men's basketball tournamentBreaking BadRashida JonesKYURMia KhalifaBrittany SnowRise of the RōninVietnamFermi paradoxKerry Von ErichStranger ThingsGary Clark Jr.List of solar eclipses in the 21st centurySuge KnightMalik ScottChelsea ManningSameer RizviMisa HyltonGoldie HawnBanksyDune (2021 film)Mao ZedongJake PaulBoris YeltsinShaquille O'NealJeffrey EpsteinBramayugamKaty PerryProject 2025Rosalind ChaoCity (artwork)UnthinkableArnold SchwarzeneggerAadhaarLiam CunninghamAngelina JolieCaitlin ClarkZionismBenjamin NetanyahuJennifer LopezEuropeRoad House (1989 film)Halo (TV series)2020 United States presidential electionInstagramShyneMonsterVerseI-35W Mississippi River bridgeKriti SanonDamsel (2024 film)Malaysia Airlines Flight 3702024 Russian presidential electionJack BlackShivam DubeBrazilSean CombsViral videoAndy WarholRobin WilliamsThe Goat LifeNarendra ModiRussell Simmons🡆 More