Nate Soares (born Nathaniel Soares) is an American artificial intelligence researcher known for his work on existential risk from AI. [1] [2] In 2014, Soares co-authored a paper that introduced the term AI alignment , the challenge of making increasingly capable AIs behave as intended. [3] [4] Soares is the president of the Machine Intelligence Research Institute (MIRI), [5] a research nonprofit based in Berkeley, California.
In a 2025 book authored with Eliezer Yudkowsky, If Anyone Builds It, Everyone Dies , Soares argues that the creation of vastly smarter-than-human AI (or superintelligence) “using anything remotely like current techniques” would very likely result in human extinction. [6] [7] [8] Soares argues that the field of AI alignment is in a nascent state, and that international regulatory intervention is likely necessary to prevent AI developers from racing to build catastrophically dangerous systems. [9]
Soares received his Bachelor of Science degree (in computer science and economics) from George Washington University in 2011. [10] Soares worked as a research associate at the National Institute of Standards and Technology and as a contractor for the United States Department of Defense, creating software tools for the National Defense University, before spending time at Microsoft and Google. [11] [12]
Soares left Google in 2014 to become a research fellow at the Machine Intelligence Research Institute. [13] He served as the lead author on MIRI's research agenda, which in January 2015 was cited heavily in Research Priorities for Robust and Beneficial Artificial Intelligence, an open letter calling for AI scientists to prioritize technical research “not only on making AI more capable, but also on maximizing the societal benefit of AI.” [14] [15] This included “research on the possibility of superintelligent machines or rapid, sustained self-improvement (intelligence explosion).” [14]
Shortly after joining MIRI, Soares became the institute's executive director. [13] [16] In 2017, he gave a talk at Google outlining open research problems in AI alignment, and arguing that the alignment problem looks especially difficult. [17]
In 2023, MIRI shifted from a focus on alignment research to a focus on warning policymakers and the public about the risks posed by potential future developments in AI. [18] Coinciding with this change, Soares transitioned from the role of executive director to president, with Malo Bourgon serving as MIRI's new CEO. [18] [19]
We named this broad research problem 'corrigibility,' in the 2014 paper that also introduced the term 'AI alignment problem' (which had previously been called the 'friendly AI problem' by us and the 'control problem' by others).
In his view, companies like OpenAI and Anthropic are racing toward the profoundly dangerous goal of artificial superintelligence — an AI that far outstrips the capabilities of any one human, or even all humans combined. Once that comes into existence, Soares argues, it will be impossible to control[....] Soares sketched out a policy vision to stave off humanity's demise. His ask is in one sense simple: a global ban on advanced AI research.