Leopold Aschenbrenner

Last updated

Leopold Aschenbrenner is an artificial intelligence (AI) researcher. He was part of OpenAI's "Superalignment" team, before he was fired in April 2024 over an alleged information leak. He has published a popular essay called "Situational Awareness" about the emergence of artificial general intelligence and related security risks.

Contents

Biography

Aschenbrenner was born in Germany. He did research for the Global Priorities Initiative at Oxford University and co-authored a 2024 working paper with Philip Trammell of Oxford. He also has some experience with the effective altruism movement. [1]

OpenAI

Aschenbrenner joined OpenAI in 2023, on a project called "Superalignment" that researches how potential future superintelligences could be aligned with human values. [2]

In April 2023, a hacker gained access to OpenAI's internal messaging system and stole information, an event that OpenAI kept private. [3] Subsequently, Aschenbrenner wrote a memo to OpenAI's board of directors about the possibility of industrial espionage by Chinese and other foreign entities, arguing that OpenAI's security was insufficient. According to Aschenbrenner, this memo led to tensions between the board and the leadership about security, and he received a warning from human resources. OpenAI later fired him in April 2024 over an alleged information leak, which Aschenbrenner said was about a benign brainstorming document shared to three external researchers for feedback. OpenAI stated that the firing is unrelated to the security memo, whereas Aschenbrenner said that it was made explicit to him at the time that it was a major reason. [4] [5] The "Superalignment" team was dissolved one month later, with the departure from OpenAI of other researchers such as Ilya Sutskever and Jan Leike. [6]

Investment

Aschenbrenner said that he started an investment firm with investors Patrick and John Collison, Daniel Gross, and Nat Friedman. [7] [8]

Situational Awareness essay

Aschenbrenner wrote a 165-page essay named "Situational Awareness: The Decade Ahead". It contains sections that predict the emergence of AGI, imagines a path from AGI to superintelligence describes four risks to humanity, outlines a way for humans to deal with superintelligent machines, and articulates the principles of an "AGI realism". He specifically warns that the United States needs to defend against the use AI technologies by countries such as Russia and China. [8] His analysis is based on future capacity for AI systems to conduct AI research, what a Forbes writer referred to as "recursive self-improvement and runaway superintelligence." [9]

Related Research Articles

<span class="mw-page-title-main">Eliezer Yudkowsky</span> American AI researcher and writer (born 1979)

Eliezer S. Yudkowsky is an American artificial intelligence researcher and writer on decision theory and ethics, best known for popularizing ideas related to friendly artificial intelligence. He is the founder of and a research fellow at the Machine Intelligence Research Institute (MIRI), a private research nonprofit based in Berkeley, California. His work on the prospect of a runaway intelligence explosion influenced philosopher Nick Bostrom's 2014 book Superintelligence: Paths, Dangers, Strategies.

Friendly artificial intelligence is hypothetical artificial general intelligence (AGI) that would have a positive (benign) effect on humanity or at least align with human interests such as fostering the improvement of the human species. It is a part of the ethics of artificial intelligence and is closely related to machine ethics. While machine ethics is concerned with how an artificially intelligent agent should behave, friendly artificial intelligence research is focused on how to practically bring about this behavior and ensuring it is adequately constrained.

<span class="mw-page-title-main">Nick Bostrom</span> Philosopher and writer (born 1973)

Nick Bostrom is a philosopher known for his work on existential risk, the anthropic principle, human enhancement ethics, whole brain emulation, superintelligence risks, and the reversal test. He was the founding director of the now dissolved Future of Humanity Institute at the University of Oxford and is now Principal Researcher at the Macrostrategy Research Initiative.

Veteran Intelligence Professionals for Sanity (VIPS) is a group of former officers of the United States Intelligence Community formed in January 2003. In February 2003, the group issued a statement accusing the Bush administration of misrepresenting U.S. national intelligence information in order to push the US and its allies toward that year's US-led invasion of Iraq. The group issued a letter stating that intelligence analysts were not being heeded by policy makers. The group initially numbered 25, mostly retired analysts.

Artificial general intelligence (AGI) is a type of artificial intelligence (AI) that matches or surpasses human cognitive capabilities across a wide range of cognitive tasks. This contrasts with narrow AI, which is limited to specific tasks. Artificial superintelligence (ASI), on the other hand, refers to AGI that greatly exceeds human cognitive capabilities. AGI is considered one of the definitions of strong AI.

A superintelligence is a hypothetical agent that possesses intelligence surpassing that of the brightest and most gifted human minds. "Superintelligence" may also refer to a property of problem-solving systems whether or not these high-level intellectual competencies are embodied in agents that act in the world. A superintelligence may or may not be created by an intelligence explosion and associated with a technological singularity.

The Machine Intelligence Research Institute (MIRI), formerly the Singularity Institute for Artificial Intelligence (SIAI), is a non-profit research institute focused since 2005 on identifying and managing potential existential risks from artificial general intelligence. MIRI's work has focused on a friendly AI approach to system design and on predicting the rate of technology development.

<span class="mw-page-title-main">AI takeover</span> Hypothetical outcome of artificial intelligence

An AI takeover is an imagined scenario in which artificial intelligence (AI) emerges as the dominant form of intelligence on Earth and computer programs or robots effectively take control of the planet away from the human species, which relies on human intelligence. Possible scenarios include replacement of the entire human workforce due to automation, takeover by an artificial superintelligence (ASI), and the notion of a robot uprising. Stories of AI takeovers have been popular throughout science fiction, but recent advancements have made the threat more real. Some public figures, such as Stephen Hawking and Elon Musk, have advocated research into precautionary measures to ensure future superintelligent machines remain under human control.

Recursive self-improvement (RSI) is a process in which an early or weak artificial general intelligence (AGI) system enhances its own capabilities and intelligence without human intervention, leading to a superintelligence or intelligence explosion.

The government of the People's Republic of China is engaged in espionage overseas, directed through diverse methods via the Ministry of State Security (MSS), the Ministry of Public Security (MPS), the United Front Work Department (UFWD), People's Liberation Army (PLA) via its Intelligence Bureau of the Joint Staff Department, and numerous front organizations and state-owned enterprises. It employs a variety of tactics including cyber espionage to gain access to sensitive information remotely, signals intelligence, human intelligence as well as influence operations through united front activity targeting overseas Chinese communities and associations. The Chinese government is also engaged in industrial espionage aimed at gathering information and technology to bolster its economy, as well as transnational repression of dissidents abroad such as supporters of the Tibetan independence movement and Uyghurs as well as the Taiwan independence movement, the Hong Kong independence movement, Falun Gong, pro-democracy activists, and other critics of the Chinese Communist Party (CCP). The United States alleges that the degree of intelligence activity is unprecedented in its assertiveness and engagement in multiple host countries, particularly the United States, with economic damages estimated to run into the hundreds of billions according to the Center for Strategic and International Studies.

Cyberwarfare by China is the aggregate of cyberattacks attributed to the organs of the People's Republic of China and various related advanced persistent threat (APT) groups.

In the field of artificial intelligence (AI) design, AI capability control proposals, also referred to as AI confinement, aim to increase our ability to monitor and control the behavior of AI systems, including proposed artificial general intelligences (AGIs), in order to reduce the danger they might pose if misaligned. However, capability control becomes less effective as agents become more intelligent and their ability to exploit flaws in human control systems increases, potentially resulting in an existential risk from AGI. Therefore, the Oxford philosopher Nick Bostrom and others recommend capability control methods only as a supplement to alignment methods.

<span class="mw-page-title-main">Roman Yampolskiy</span> Latvian computer scientist (born 1979)

Roman Vladimirovich Yampolskiy is a Latvian computer scientist at the University of Louisville, mostly known for his work on AI safety and cybersecurity. He holds a PhD from the University at Buffalo (2008). He is the founder and current director of Cyber Security Lab, in the department of Computer Engineering and Computer Science at the Speed School of Engineering of the University of Louisville.

Existential risk from artificial intelligence refers to the idea that substantial progress in artificial general intelligence (AGI) could lead to human extinction or an irreversible global catastrophe.

OpenAI is an American artificial intelligence (AI) research organization founded in December 2015 and headquartered in San Francisco, California. Its stated mission is to develop "safe and beneficial" artificial general intelligence (AGI), which it defines as "highly autonomous systems that outperform humans at most economically valuable work". As a leading organization in the ongoing AI boom, OpenAI is known for the GPT family of large language models, the DALL-E series of text-to-image models, and a text-to-video model named Sora. Its release of ChatGPT in November 2022 has been credited with catalyzing widespread interest in generative AI.

Ilya Sutskever is a Canadian-Israeli-Russian computer scientist who specializes in machine learning.

Regulation of artificial intelligence is the development of public sector policies and laws for promoting and regulating artificial intelligence (AI). It is part of the broader regulation of algorithms. The regulatory and policy landscape for AI is an emerging issue in jurisdictions worldwide, including for international organizations without direct enforcement power like the IEEE or the OECD.

Jan Leike is an AI alignment researcher who has worked at DeepMind and OpenAI. He joined Anthropic in May 2024.

PauseAI is a global political movement founded in the Netherlands with the stated aim of achieving global coordination to stop the development of artificial intelligence systems more powerful than GPT-4, at least until it is known how to build them safely, and keep them under democratic control. The movement was established in Utrecht in May 2023 by software entrepreneur Joep Meindertsma.

References

  1. Allen, Mike (2024-06-23). "10 takeaways: AI from now to 2034". Axios. Retrieved 2024-12-27.
  2. "Ex-OpenAI employee writes AI essay: War with China, resources and robots". heise online. 2024-07-02. Retrieved 2024-12-28.
  3. Metz, Cade (2024-07-04). "A Hacker Stole OpenAI Secrets, Raising Fears That China Could, Too". The New York Times . Archived from the original on 2024-12-26. Retrieved 2024-12-27.
  4. Altchek, Ana. "Ex-OpenAI employee speaks out about why he was fired: 'I ruffled some feathers'". Business Insider. Retrieved 2024-12-28.
  5. "Ex-OpenAI Employee Reveals Reason For Getting Fired, "Security Memo Was..."". NDTV. June 6, 2024. Retrieved 2024-12-28.
  6. Field, Hayden (2024-05-17). "OpenAI dissolves team focused on long-term AI risks, less than one year after announcing it". CNBC. Retrieved 2024-12-28.
  7. "Post Script: Patrick Collison the Swiss dictator; McKillen-Bono whiskey lands ex-Bank of Ireland governor". www.businesspost.ie. Retrieved 2024-12-27.
  8. 1 2 Naughton, John (2024-06-15). "How's this for a bombshell – the US must make AI its next Manhattan Project". The Observer. ISSN   0029-7712 . Retrieved 2024-12-27.
  9. Toews, Rob (November 5, 2024). "AI That Can Invent AI Is Coming. Buckle Up". Forbes. Retrieved 2024-12-27.