Jan Leike

Last updated

Jan Leike (born 1986or1987) [1] is an AI alignment researcher who has worked at DeepMind and OpenAI. He joined Anthropic in May 2024.

Contents

Education

Jan Leike obtained his undergraduate degree from the University of Freiburg in Germany. After earning a master's degree in computer science, he pursued a PhD in machine learning at the Australian National University under the supervision of Marcus Hutter. [2]

Career

Leike made a six-month postdoctoral fellowship at the Future of Humanity Institute before joining DeepMind to focus on empirical AI safety research, [2] where he collaborated with Shane Legg. [1]

OpenAI

In 2021, Leike joined OpenAI. [1] In June 2023, he and Ilya Sutskever became the co-leaders of the newly introduced "superalignment" project, which aimed to determine how to align future artificial superintelligences within four years to ensure their safety. This project involved automating AI alignment research using relatively advanced AI systems. At the time, Sutskever was OpenAI's Chief Scientist, and Leike was the Head of Alignment. [3] [1] Leike was featured in Time's list of the 100 most influential personalities in AI, both in 2023 [1] and in 2024. [4] In May 2024, Leike announced his resignation from OpenAI, following the departure of Ilya Sutskever, Daniel Kokotajlo and several other AI safety employees from the company. Leike wrote that "Over the past years, safety culture and processes have taken a backseat to shiny products", and that he "gradually lost trust" in OpenAI's leadership. [5] [6] [7]

In May 2024, Leike joined Anthropic, an AI company founded by former OpenAI employees. [8]

Related Research Articles

<span class="mw-page-title-main">Nick Bostrom</span> Philosopher and writer (born 1973)

Nick Bostrom is a philosopher known for his work on existential risk, the anthropic principle, human enhancement ethics, whole brain emulation, superintelligence risks, and the reversal test. He was the founding director of the now dissolved Future of Humanity Institute at the University of Oxford and is now Principal Researcher at the Macrostrategy Research Initiative.

<span class="mw-page-title-main">Geoffrey Hinton</span> British-Canadian computer scientist and psychologist (born 1947)

Geoffrey Everest Hinton is a British-Canadian computer scientist and cognitive psychologist, most noted for his work on artificial neural networks. From 2013 to 2023, he divided his time working for Google and the University of Toronto, before publicly announcing his departure from Google in May 2023, citing concerns about the risks of artificial intelligence (AI) technology. In 2017, he co-founded and became the chief scientific advisor of the Vector Institute in Toronto.

Artificial general intelligence (AGI) is a type of artificial intelligence (AI) that falls within the lower and upper limits of human cognitive capabilities across a wide range of cognitive tasks. This contrasts with narrow AI, which is limited to specific tasks. Artificial superintelligence (ASI), refers to types of intelligence that range from being only marginally smarter than the upper limits of human intelligence to greatly exceeding human cognitive capabilities by orders of magnitude. AGI is considered one of the definitions of strong AI.

A superintelligence is a hypothetical agent that possesses intelligence surpassing that of the brightest and most gifted human minds. "Superintelligence" may also refer to a property of problem-solving systems whether or not these high-level intellectual competencies are embodied in agents that act in the world. A superintelligence may or may not be created by an intelligence explosion and associated with a technological singularity.

<span class="mw-page-title-main">Jaan Tallinn</span> Estonian programmer and investor

Jaan Tallinn is an Estonian billionaire computer programmer and investor known for his participation in the development of Skype and file-sharing application FastTrack/Kazaa.

Anthropic PBC is a U.S.-based artificial intelligence (AI) startup public-benefit company, founded in 2021. It researches and develops AI to "study their safety properties at the technological frontier" and use this research to deploy safe, reliable models for the public. Anthropic has developed a family of large language models (LLMs) named Claude as a competitor to OpenAI's ChatGPT and Google's Gemini.

<span class="mw-page-title-main">Future of Humanity Institute</span> Defunct Oxford interdisciplinary research centre

The Future of Humanity Institute (FHI) was an interdisciplinary research centre at the University of Oxford investigating big-picture questions about humanity and its prospects. It was founded in 2005 as part of the Faculty of Philosophy and the Oxford Martin School. Its director was philosopher Nick Bostrom, and its research staff included futurist Anders Sandberg and Giving What We Can founder Toby Ord.

<span class="mw-page-title-main">Sam Altman</span> American entrepreneur and investor (born 1985)

Samuel Harris Altman is an American entrepreneur and investor best known as the CEO of OpenAI since 2019. He is also the chairman of clean energy companies Oklo Inc. and Helion Energy. Altman is considered to be one of the leading figures of the AI boom. He dropped out of Stanford University after two years and founded Loopt, a mobile social networking service, raising more than $30 million in venture capital. In 2011, Altman joined Y Combinator, a startup accelerator, and was its president from 2014 to 2019.

Daniel Gross is an American entrepreneur who co-founded Cue, led artificial intelligence efforts at Apple, served as a partner at Y Combinator, and is a notable technology investor in companies like Uber, Instacart, Figma, GitHub, Airtable, Rippling, CoreWeave, Character.ai, Perplexity.ai, and others.

Existential risk from artificial general intelligence refers to the idea that substantial progress in artificial general intelligence (AGI) could lead to human extinction or an irreversible global catastrophe.

OpenAI is an American artificial intelligence (AI) research organization founded in December 2015 and headquartered in San Francisco, California. Its mission is to develop "safe and beneficial" artificial general intelligence (AGI), which it defines as "highly autonomous systems that outperform humans at most economically valuable work". As a leading organization in the ongoing AI boom, OpenAI is known for the GPT family of large language models, the DALL-E series of text-to-image models, and a text-to-video model named Sora. Its release of ChatGPT in November 2022 has been credited with catalyzing widespread interest in generative AI.

In the field of artificial intelligence (AI), AI alignment aims to steer AI systems toward a person's or group's intended goals, preferences, and ethical principles. An AI system is considered aligned if it advances the intended objectives. A misaligned AI system pursues unintended objectives.

Ilya Sutskever is a Russian-Israeli computer scientist who specializes in machine learning.

Paul Christiano is an American researcher in the field of artificial intelligence (AI), with a specific focus on AI alignment, which is the subfield of AI safety research that aims to steer AI systems toward human interests. He formerly led the language model alignment team at OpenAI and became founder and head of the non-profit Alignment Research Center (ARC), which works on theoretical AI alignment and evaluations of machine learning models. In 2023, Christiano was named as one of the TIME 100 Most Influential People in AI.

<span class="mw-page-title-main">Dario Amodei</span> Italian-American entrepreneur

Dario Amodei is an Italian-American artificial intelligence researcher and entrepreneur. He is the co-founder and CEO of Anthropic, the company behind the large language model series Claude AI. He was previously the vice president of research at OpenAI.

<span class="mw-page-title-main">Removal of Sam Altman from OpenAI</span> 2023 business action

On November 17, 2023, OpenAI's board of directors ousted co-founder and chief executive Sam Altman after the board had no confidence in his leadership. The removal was caused by concerns about his handling of artificial intelligence safety, and allegations of abusive behavior. Altman was reinstated on November 22 after pressure from employees and investors.

Daniela Amodei is an Italian-American artificial intelligence researcher and entrepreneur. She is the President and co-founder of the artificial intelligence company Anthropic.

PauseAI is a global political movement founded in the Netherlands with the stated aim of achieving global coordination to stop the development of artificial intelligence systems more powerful than GPT-4, at least until it is known how to build them safely, and keep them under democratic control. The movement was established in Utrecht in May 2023 by software entrepreneur Joep Meindertsma.

The Safe and Secure Innovation for Frontier Artificial Intelligence Models Act, or SB 1047, is a 2024 California bill intended to "mitigate the risk of catastrophic harms from AI models so advanced that they are not yet known to exist". Specifically, the bill would apply to models which cost more than $100 million to train and were trained using a quantity of computing power greater than 1026 integer or floating-point operations. SB 1047 would apply to all AI companies doing business in California—the location of the company does not matter. The bill creates protections for whistleblowers and requires developers to perform risk assessments of their models prior to release, under the supervision of the Government Operations Agency. It would also establish CalCompute, a University of California public cloud computing cluster for startups, researchers and community groups.

Safe Superintelligence Inc. or SSI Inc. is an American artificial intelligence company founded by Open AI's former chief scientist Ilya Sutskever, Daniel Gross and Daniel Levy. The company's mission is to focus on safely developing a superintelligence, an agent capable of surpassing human intelligence.

References

  1. 1 2 3 4 5 "TIME100 AI 2023: Jan Leike". Time. 7 September 2023. Archived from the original on 19 May 2024. Retrieved 19 May 2024.
  2. 1 2 "An AI safety researcher on how to become an AI safety researcher". 80,000 Hours. Archived from the original on 19 May 2024. Retrieved 19 May 2024.
  3. Leike, Jan; Sutskever, Ilya (5 July 2023). "Introducing Superalignment". OpenAI. Archived from the original on 25 May 2024. Retrieved 20 May 2024.
  4. Booth, Harry (5 September 2024). "TIME100 AI 2024: Jan Leike". TIME. Archived from the original on 8 September 2024. Retrieved 8 September 2024.
  5. Samuel, Sigal (17 May 2024). ""I lost trust": Why the OpenAI team in charge of safeguarding humanity imploded". Vox. Archived from the original on 18 May 2024. Retrieved 20 May 2024.
  6. Bastian, Matthias (18 May 2024). "OpenAI's AI safety teams lost at least seven researchers in recent months". the decoder. Archived from the original on 20 May 2024. Retrieved 20 May 2024.
  7. Milmo, Dan (18 May 2024). "OpenAI putting 'shiny products' above safety, says departing researcher". The Observer. ISSN   0029-7712 . Retrieved 20 May 2024.
  8. "OpenAI researcher who resigned over safety concerns joins Anthropic". 28 May 2024. Archived from the original on 28 May 2024. Retrieved 28 May 2024.