Oriol Vinyals

Last updated
Oriol Vinyals
Born1983 (age 3940)
Education Universitat Politècnica de Catalunya
University of California, San Diego
University of California, Berkeley
Known for seq2seq
AlphaStar
Scientific career
Institutions Google
DeepMind
Thesis Beyond Deep Learning: Scalable Methods and Models for Learning  (2013)
Doctoral advisor Nelson Morgan

Oriol Vinyals (born 1983) is a Spanish machine learning researcher at DeepMind, [1] [2] where he is the principal research scientist. [3] His research in DeepMind is regularly featured in the mainstream media [4] [5] [6] especially after being acquired by Google. [7]

Contents

Education and career

Vinyals was born in Barcelona, Catalonia, Spain. [8] He studied mathematics and telecommunication engineering at the Universitat Politècnica de Catalunya. He then moved to the US and studied for a Master's degree in computer science at University of California, San Diego, and at University of California, Berkeley, where he received his PhD in 2013 under Nelson Morgan in the Department of Electrical Engineering and Computer Science.

Vinyals co-invented the seq2seq model for machine translation along with Ilya Sutskever and Quoc Viet Le. [9] He led AlphaStar research group at DeepMind, which applies artificial intelligence to computer games such as StarCraft II. [10]

In 2016, he was chosen by the magazine MIT Technology Review as one of the 35 most innovative young people under 35. [11]

See also

Related Research Articles

<span class="mw-page-title-main">Geoffrey Hinton</span> British-Canadian computer scientist and psychologist (born 1947)

Geoffrey Everest Hinton is a British-Canadian cognitive psychologist and computer scientist, most noted for his work on artificial neural networks. From 2013 to 2023, he divided his time working for Google and the University of Toronto, before publicly announcing his departure from Google in May 2023 citing concerns about the risks of artificial intelligence (AI) technology. In 2017, he co-founded and became the chief scientific advisor of the Vector Institute in Toronto.

<span class="mw-page-title-main">Deep learning</span> Branch of machine learning

Deep learning is part of a broader family of machine learning methods, which is based on artificial neural networks with representation learning. The adjective "deep" in deep learning refers to the use of multiple layers in the network. Methods used can be either supervised, semi-supervised or unsupervised.

Google Brain was a deep learning artificial intelligence research team under the umbrella of Google AI, a research division at Google dedicated to artificial intelligence. Formed in 2011, Google Brain combined open-ended machine learning research with information systems and large-scale computing resources. The team has created tools such as TensorFlow, which allow for neural networks to be used by the public, with multiple internal AI research projects. The team aims to create research opportunities in machine learning and natural language processing. The team was merged into former Google sister company DeepMind to form Google DeepMind in April 2023.

<span class="mw-page-title-main">Google DeepMind</span> Artificial intelligence company owned by Google

DeepMind Technologies Limited, doing business as Google DeepMind, is a British-American artificial intelligence research laboratory which serves as a subsidiary of Google. Founded in the UK in 2010, it was acquired by Google in 2014, becoming a wholly owned subsidiary of Google parent company Alphabet Inc. after Google's corporate restructuring in 2015. The company is based in London, with research centres in Canada, France, and the United States.

Neural machine translation (NMT) is an approach to machine translation that uses an artificial neural network to predict the likelihood of a sequence of words, typically modeling and then translating entire sentences in a single integrated model.

David Silver is a principal research scientist at Google DeepMind and a professor at University College London. He has led research on reinforcement learning with AlphaGo, AlphaZero and co-lead on AlphaStar.

<span class="mw-page-title-main">Ilya Sutskever</span> Canadian computer scientist (born 1985/86)

Ilya Sutskever is a Russian-born Israeli-Canadian computer scientist working in machine learning, who co-founded and serves as Chief Scientist of OpenAI.

Wojciech Zaremba is a Polish computer scientist, a founding team member of OpenAI (2016–present), where he leads both the Codex research and language teams. The teams actively work on AI that writes computer code and creating successors to GPT-3 respectively. The mission of OpenAI is to build safe artificial intelligence (AI), and ensure that its benefits are as evenly distributed as possible.

<span class="mw-page-title-main">Differentiable neural computer</span> Artificial neural network architecture

In artificial intelligence, a differentiable neural computer (DNC) is a memory augmented neural network architecture (MANN), which is typically recurrent in its implementation. The model was published in 2016 by Alex Graves et al. of DeepMind.

<span class="mw-page-title-main">WaveNet</span> Deep neural network for generating raw audio

WaveNet is a deep neural network for generating raw audio. It was created by researchers at London-based AI firm DeepMind. The technique, outlined in a paper in September 2016, is able to generate relatively realistic-sounding human-like voices by directly modelling waveforms using a neural network method trained with recordings of real speech. Tests with US English and Mandarin reportedly showed that the system outperforms Google's best existing text-to-speech (TTS) systems, although as of 2016 its text-to-speech synthesis still was less convincing than actual human speech. WaveNet's ability to generate raw waveforms means that it can model any kind of audio, including music.

AlphaStar is a computer program by DeepMind that plays the video game StarCraft II. It was unveiled to the public by name in January 2019. In a significant milestone for artificial intelligence, AlphaStar attained Grandmaster status in August 2019.

In video games, various artificial intelligence techniques have been used in a variety of ways, ranging from non-player character (NPC) control to procedural content generation (PCG). Machine learning is a subset of artificial intelligence that focuses on using algorithms and statistical models to make machines act without specific programming. This is in sharp contrast to traditional methods of artificial intelligence such as search trees and expert systems.

Seq2seq is a family of machine learning approaches used for natural language processing. Applications include language translation, image captioning, conversational models, and text summarization.

Samy Bengio is a Canadian computer scientist, Senior Director of AI and Machine Learning Research at Apple, and a former long-time scientist at Google known for leading a large group of researchers working in machine learning including adversarial settings. Bengio left Google shortly after the company fired his report, Timnit Gebru, without first notifying him. At the time, Bengio said that he had been "stunned" by what happened to Gebru. He is also among the three authors who developed Torch in 2002, the ancestor of PyTorch, one of today's two largest machine learning frameworks.

Perceiver is a transformer adapted to be able to process non-textual data, such as images, sounds and video, and spatial data. Transformers underlie other notable systems such as BERT and GPT-3, which preceded Perceiver. It adopts an asymmetric attention mechanism to distill inputs into a latent bottleneck, allowing it to learn from large amounts of heterogeneous data. Perceiver matches or outperforms specialized models on classification tasks.

Gato is a deep neural network for a range of complex tasks that exhibits multimodality. It can perform tasks such as engaging in a dialogue, playing video games, controlling a robot arm to stack blocks, and more. It was created by researchers at London-based AI firm DeepMind. It is a transformer, like GPT-3. According to MIT Technology Review, the system "learns multiple different tasks at the same time, which means it can switch between them without having to forget one skill before learning another" whereas "[t]he AI systems of today are called “narrow,” meaning they can only do a specific, restricted set of tasks such as generate text", and according to The Independent, it is a "'generalist agent' that can carry out a huge range of complex tasks, from stacking blocks to writing poetry". It uses supervised learning with 1.2B parameters. The technology has been described as "general purpose" artificial intelligence and a "step toward" artificial general intelligence.

NSynth is a WaveNet-based autoencoder for synthesizing audio, outlined in a paper in April 2017.

Lê Viết Quốc, or in romanized form Quoc Viet Le, is a Vietnamese-American computer scientist and a machine learning pioneer at Google Brain, which he established with others from Google. He co-invented doc2vec and seq2seq models in natural language processing. Le also initiated and lead the AutoML initiative at Google Brain, including the proposal of neural architecture search.

<span class="mw-page-title-main">AlphaDev</span> AI model that developer a super-human sorting algorithm

AlphaDev is an artificial intelligence system developed by Google DeepMind to discover enhanced computer science algorithms using reinforcement learning. AlphaDev is based on AlphaZero, a system that mastered the games of chess, shogi and go by self-play. AlphaDev applies the same approach to finding faster algorithms for fundamental tasks such as sorting and hashing.

Whisper is a weakly-supervised deep learning acoustic model for speech recognition made by the company OpenAI. It is capable of transcribing in English and several other languages. It is also capable of translating several non-English languages into English, having been trained with 125,000 hours of data for this purpose. The model was released as open-source software in September 2022. It uses an encoder-decoder transformer architecture. OpenAI has trained Whisper on 680,000 hours of multilingual and multitask data from the web, of which 117,000 hours cover non-English languages; OpenAI claims this has led to improved recognition of accents, background noise and jargon.

References

  1. Pascual, Manuel G. (2022-12-03). "Oriol Vinyals: "Nuestra generación verá una inteligencia artificial que iguale o supere a la del ser humano"". El País (in Spanish). Retrieved 2023-04-04.
  2. "Oriol Vinyals – Google Research". Google Research. Retrieved 2022-08-01.
  3. "DeepMind AI rivals average human competitive coder". BBC News. 2022-02-02. Retrieved 2023-04-04.
  4. Vincent, James (2022-02-02). "DeepMind says its new AI coding engine is as good as an average human programmer". The Verge. Retrieved 2023-04-04.
  5. Staff, Ars (2016-11-25). "Google DeepMind could invent the next generation of AI by playing Starcraft 2". Ars Technica. Retrieved 2023-04-04.
  6. "Google brain connects his StarCraft past with AI future". Financial Times. 2017-11-23. Retrieved 2023-04-04.
  7. "Oriol Vinyals – Google Research". Google Research. Retrieved 2023-04-04.
  8. "9. Oriol Vinyals: Sequence-to-Sequence Machine Learning - The Future of Machine Intelligence [Book]". www.oreilly.com. Retrieved 2022-07-31.
  9. Sutskever, Ilya; Vinyals, Oriol; Le, Quoc V (2014). "Sequence to Sequence Learning with Neural Networks". Advances in Neural Information Processing Systems. Curran Associates, Inc. 27. arXiv: 1409.3215 .
  10. Vinyals, Oriol; Babuschkin, Igor; Czarnecki, Wojciech M.; Mathieu, Michaël; Dudzik, Andrew; Chung, Junyoung; Choi, David H.; Powell, Richard; Ewalds, Timo; Georgiev, Petko; Oh, Junhyuk (2019-11-14). "Grandmaster level in StarCraft II using multi-agent reinforcement learning". Nature. 575 (7782): 350–354. Bibcode:2019Natur.575..350V. doi:10.1038/s41586-019-1724-z. ISSN   0028-0836. PMID   31666705. S2CID   204972004.
  11. "Oriol Vinyals". MIT Technology Review. Retrieved 2022-07-31.