Aidan Gomez

Last updated
Aidan Gomez
Aidan Gomez at Collision 2023.jpg
Born
Education University of Toronto
University of Oxford
Occupation(s)CEO, Cohere
Website aidangomez.ca

Aidan Gomez is a British-Canadian [1] computer scientist working in the field of artificial intelligence, with a focus on natural language processing. [2] He is the co-founder and CEO of the technology company Cohere. [3] [4]

Contents

Early life and education

Gomez grew up in Brighton, Ontario. [5] He graduated from the University of Toronto with a bachelor's degree in computer science and mathematics. [6] He was pursuing a PhD in computer science from the University of Oxford. [7] He paused his studies to launch Cohere; however, he ultimately was granted the PhD in 2024. [8]

Career

In 2017, as a 20 year-old intern at Google Brain, Gomez was one of eight authors of the research paper "Attention Is All You Need", [9] which is credited with changing the AI industry and helping lead to the creation of ChatGPT. [3] [4] [10] [11] The paper proposed a novel deep learning architecture called the transformer, that enables machine learning models to analyze large amounts of data for patterns, and then use those patterns to make predictions. It has been commonly adopted for training large language models and in the development of generative AI. [3] [12] In 2017, Gomez founded FOR.ai, [7] a program to help researchers learn machine learning techniques in a collaborative format, which later became Cohere For AI. [13]

As a PhD student, Gomez worked as a machine learning researcher at Google Brain. [7] At that time, he co-authored the paper "One Model to Learn Them All" about multi-task learning by a single neural network. [6]

In 2019, Gomez left Google Brain to launch Cohere, an enterprise-focused company that helps businesses implement AI into chatbots, search engines, and other products. [3] [4] [14] He is CEO of the company, which has been valued at over $2 billion. [4]

Gomez was named to the 2023 Time 100/AI list of the most influential people in the field of artificial intelligence. [3] He and his fellow Cohere founders Ivan Zhang and Nick Frosst were named number 1 on 2023 Maclean's AI Trailblazers Power List. [12]

Related Research Articles

Artificial intelligence (AI), in its broadest sense, is intelligence exhibited by machines, particularly computer systems. It is a field of research in computer science that develops and studies methods and software that enable machines to perceive their environment and uses learning and intelligence to take actions that maximize their chances of achieving defined goals. Such machines may be called AIs.

<span class="mw-page-title-main">Chatbot</span> Program that simulates conversation

A chatbot is a software application or web interface that is designed to mimic human conversation through text or voice interactions. Modern chatbots are typically online and use generative artificial intelligence systems that are capable of maintaining a conversation with a user in natural language and simulating the way a human would behave as a conversational partner. Such chatbots often use deep learning and natural language processing, but simpler chatbots have existed for decades.

<span class="mw-page-title-main">Jürgen Schmidhuber</span> German computer scientist

Jürgen Schmidhuber is a German computer scientist noted for his work in the field of artificial intelligence, specifically artificial neural networks. He is a scientific director of the Dalle Molle Institute for Artificial Intelligence Research in Switzerland. He is also director of the Artificial Intelligence Initiative and professor of the Computer Science program in the Computer, Electrical, and Mathematical Sciences and Engineering (CEMSE) division at the King Abdullah University of Science and Technology (KAUST) in Saudi Arabia.

<span class="mw-page-title-main">Geoffrey Hinton</span> British-Canadian computer scientist and psychologist (born 1947)

Geoffrey Everest Hinton is a British-Canadian computer scientist and cognitive psychologist, most noted for his work on artificial neural networks. From 2013 to 2023, he divided his time working for Google and the University of Toronto, before publicly announcing his departure from Google in May 2023, citing concerns about the risks of artificial intelligence (AI) technology. In 2017, he co-founded and became the chief scientific advisor of the Vector Institute in Toronto.

<span class="mw-page-title-main">Andrew Ng</span> American artificial intelligence researcher

Andrew Yan-Tak Ng is a British-American computer scientist and technology entrepreneur focusing on machine learning and artificial intelligence (AI). Ng was a cofounder and head of Google Brain and was the former Chief Scientist at Baidu, building the company's Artificial Intelligence Group into a team of several thousand people.

Google Brain was a deep learning artificial intelligence research team under the umbrella of Google AI, a research division at Google dedicated to artificial intelligence. Formed in 2011, it combined open-ended machine learning research with information systems and large-scale computing resources. It created tools such as TensorFlow, which allow neural networks to be used by the public, and multiple internal AI research projects, and aimed to create research opportunities in machine learning and natural language processing. It was merged into former Google sister company DeepMind to form Google DeepMind in April 2023.

<span class="mw-page-title-main">Google DeepMind</span> Artificial intelligence division

DeepMind Technologies Limited, doing business as Google DeepMind, is a British-American artificial intelligence research laboratory which serves as a subsidiary of Google. Founded in the UK in 2010, it was acquired by Google in 2014 and merged with Google AI's Google Brain division to become Google DeepMind in April 2023. The company is based in London, with research centres in Canada, France, Germany, and the United States.

Google AI is a division of Google dedicated to artificial intelligence. It was announced at Google I/O 2017 by CEO Sundar Pichai.

Synthetic media is a catch-all term for the artificial production, manipulation, and modification of data and media by automated means, especially through the use of artificial intelligence algorithms, such as for the purpose of misleading people or changing an original meaning. Synthetic media as a field has grown rapidly since the creation of generative adversarial networks, primarily through the rise of deepfakes as well as music synthesis, text generation, human image synthesis, speech synthesis, and more. Though experts use the term "synthetic media," individual methods such as deepfakes and text synthesis are sometimes not referred to as such by the media but instead by their respective terminology Significant attention arose towards the field of synthetic media starting in 2017 when Motherboard reported on the emergence of AI altered pornographic videos to insert the faces of famous actresses. Potential hazards of synthetic media include the spread of misinformation, further loss of trust in institutions such as media and government, the mass automation of creative and journalistic jobs and a retreat into AI-generated fantasy worlds. Synthetic media is an applied form of artificial imagination.

Generative Pre-trained Transformer 3 (GPT-3) is a large language model released by OpenAI in 2020.

Cohere is a Canadian multinational technology company focused on artificial intelligence for the enterprise, specializing in large language models. Cohere was founded in 2019 by Aidan Gomez, Ivan Zhang, and Nick Frosst, and is headquartered in Toronto and San Francisco, with offices in Palo Alto and London.

LaMDA is a family of conversational large language models developed by Google. Originally developed and introduced as Meena in 2020, the first-generation LaMDA was announced during the 2021 Google I/O keynote, while the second generation was announced the following year. In June 2022, LaMDA gained widespread attention when Google engineer Blake Lemoine made claims that the chatbot had become sentient. The scientific community has largely rejected Lemoine's claims, though it has led to conversations about the efficacy of the Turing test, which measures whether a computer can pass for a human. In February 2023, Google announced Bard, a conversational artificial intelligence chatbot powered by LaMDA, to counter the rise of OpenAI's ChatGPT.

A text-to-video model is a machine learning model which takes a natural language description as input and producing a video or multiples videos from the input.

Sébastien Bubeck is a French-American computer scientist and mathematician. He is currently Microsoft's Vice President of GenAI and leads the Machine Learning Foundations group at Microsoft Research Redmond. Bubeck was formerly professor at Princeton University and a researcher at the University of California, Berkeley. He is known for his contributions to online learning, optimization and more recently studying deep neural networks, and in particular transformer models.

Sparrow is a chatbot developed by the artificial intelligence research lab DeepMind, a subsidiary of Alphabet Inc. It is designed to answer users' questions correctly, while reducing the risk of unsafe and inappropriate answers. One motivation behind Sparrow is to address the problem of language models producing incorrect, biased or potentially harmful outputs. Sparrow is trained using human judgements, in order to be more “Helpful, Correct and Harmless” compared to baseline pre-trained language models. The development of Sparrow involved asking paid study participants to interact with Sparrow, and collecting their preferences to train a model of how useful an answer is.

<span class="mw-page-title-main">Generative pre-trained transformer</span> Type of large language model

Generative pre-trained transformers (GPT) are a type of large language model (LLM) and a prominent framework for generative artificial intelligence. They are artificial neural networks that are used in natural language processing tasks. GPTs are based on the transformer architecture, pre-trained on large data sets of unlabelled text, and able to generate novel human-like content. As of 2023, most LLMs have these characteristics and are sometimes referred to broadly as GPTs.

<span class="mw-page-title-main">Generative artificial intelligence</span> AI system capable of generating content in response to prompts

Generative artificial intelligence is artificial intelligence capable of generating text, images, videos, or other data using generative models, often in response to prompts. Generative AI models learn the patterns and structure of their input training data and then generate new data that has similar characteristics.

Ashish Vaswani is a computer scientist working in deep learning, who is known for his significant contributions to the field of artificial intelligence (AI) and natural language processing (NLP). He is one of the co-authors of the seminal paper "Attention Is All You Need" which introduced the Transformer model, a novel architecture that uses a self-attention mechanism and has since become foundational to many state-of-the-art models in NLP. Transformer architecture is the core of language models that power applications such as ChatGPT. He was a co-founder of Adept AI Labs and a former staff research scientist at Google Brain.

<span class="mw-page-title-main">Attention Is All You Need</span> 2017 research paper by Google

"Attention Is All You Need" is a landmark 2017 research paper authored by eight scientists working at Google, responsible for expanding 2014 attention mechanisms proposed by Bahdanau et al. into a new deep learning architecture known as the transformer. The paper is considered by some to be a founding document for modern artificial intelligence, as transformers became the main architecture of large language models. At the time, the focus of the research was on improving Seq2seq techniques for machine translation, but even in their paper the authors saw the potential for other tasks like question answering and for what is now called multimodal Generative AI.

References

  1. Kokalitcheva, Kia (2 July 2023). "What they're saying: Cohere CEO Aidan Gomez". Axios. Retrieved 3 November 2023.
  2. Watson, H.G. (19 July 2022). "Toronto Start-up Cohere Is Teaching Machines to Speak Our Language". Canadian Business. Retrieved 3 November 2023.
  3. 1 2 3 4 5 Chow, Andrew (7 September 2023). "Time100 AI". Time. Retrieved 3 November 2023.
  4. 1 2 3 4 Metz, Cade (2 May 2023). "Generative A.I. Start-Up Cohere Valued at About $2 Billion in Funding Round". New York Times. Retrieved 3 November 2023.
  5. Castaldo, Joe (26 October 2022). "Lost in translation". Globe and Mail. Retrieved 8 November 2023.
  6. 1 2 Haikara, Nina (21 June 2017). "Making an impact: U of T undergrad co-authors important machine learning study at Google". U of T News. Retrieved 8 November 2023.
  7. 1 2 3 Harris, Jeremie (12 June 2019). "Oxford's Aidan Gomez on working at Google and the quest for general intelligence". Towards Data Science. Retrieved 8 November 2023.
  8. https://twitter.com/aidangomez/status/1789569124688896022
  9. Vaswani, Ashish; Shazeer, Noam; Parmar, Niki; Uszkoreit, Jakob; Jones, Llion; Gomez, Aidan N; Kaiser, Łukasz; Polosukhin, Illia (2017). "Attention is All you Need" (PDF). Advances in Neural Information Processing Systems. 30. Curran Associates, Inc.
  10. Hammond, George (15 June 2023). "Aidan Gomez: AI threat to human existence is 'absurd' distraction from real risks". Financial Times. Retrieved 6 November 2023.
  11. Olson, Parmy (12 July 2023). "Meet the $4 Billion AI Superstars That Google Lost". Bloomberg. Retrieved 6 November 2023.
  12. 1 2 Watson, H.G. (16 March 2023). "The Power List: Ivan Zhang, Aidan Gomez & Nick Frosst are creating a smarter, friendlier chatbot". Maclean's. Retrieved 6 November 2023.
  13. Goldman, Sharon (14 June 2022). "Google Brain alum to helm new nonprofit AI research lab". VentureBeat. Retrieved 14 November 2023.
  14. "Aidan Gomez on Building an AI Business". Radical Ventures. Retrieved 6 November 2023.