Chelsea Finn

Last updated
Chelsea Finn
Chelsea Finn and Vestri the robot, UC Berkeley.jpg
Finn as a graduate student at UC Berkeley in 2017
Alma mater University of California, Berkeley
Massachusetts Institute of Technology
Known for Deep reinforcement learning
Scientific career
Institutions Stanford University
Thesis Learning to Learn with Gradients  (2018)
Doctoral advisor Sergey Levine
Pieter Abbeel
Website IRIS LAB

Chelsea Finn is an American computer scientist and assistant professor at Stanford University. Her research investigates intelligence through the interactions of robots, with the hope to create robotic systems that can learn how to learn. She is part of the Google Brain group.

Contents

Early life and education

Finn was an undergraduate student in electrical engineering at Massachusetts Institute of Technology. She moved to the University of California, Berkeley, where she studied gradient based algorithms in the Berkeley Artificial Intelligence Lab (BAIR). [1] Such algorithms allow machines to 'learn to learn', more akin to human learning than traditional machine learning systems. [2] [3] These “meta-learning” techniques train machines to quickly adapt, such that when they encounter new scenarios they can learn quickly. [4] As a doctoral student she worked as an intern at Google Brain, where she worked on robot learning algorithms from deep predictive models. She delivered a massive open online course on deep reinforcement learning. [5] [6] She was the first woman to win the C.V. & Daulat Ramamoorthy Distinguished Research Award. [7]

Research and career

Finn investigates the capabilities of robots to develop intelligence through learning and interaction. [8] She has made use of deep learning algorithms to simultaneously learn visual perception and control robotic skills. [9]

She developed meta-learning approaches to train neural networks to take in student code and output useful feedback. [10] She showed that the system could quickly adapt without too much input from the instructor. [10] She trialled the programme on Code in Place, a 12,000 student course delivered by Stanford University every year. She found that 97.9% of the time the students agreed with the feedback being given. [10] [11]

Awards and honors

Select publications

Related Research Articles

<span class="mw-page-title-main">Reinforcement learning</span> Field of machine learning

Reinforcement learning (RL) is an area of machine learning concerned with how intelligent agents ought to take actions in an environment in order to maximize the notion of cumulative reward. Reinforcement learning is one of three basic machine learning paradigms, alongside supervised learning and unsupervised learning.

<span class="mw-page-title-main">Machine learning</span> Study of algorithms that improve automatically through experience

Machine learning (ML) is an umbrella term for solving problems for which development of algorithms by human programmers would be cost-prohibitive, and instead the problems are solved by helping machines 'discover' their 'own' algorithms, without needing to be explicitly told what to do by any human-developed algorithms. When there was a vast amount of potential answers, the correct ones needed to be labeled as valid by human labelers initially and human supervision was needed.

<span class="mw-page-title-main">Meta-learning (computer science)</span> Subfield of machine learning

Meta learning is a subfield of machine learning where automatic learning algorithms are applied to metadata about machine learning experiments. As of 2017, the term had not found a standard interpretation, however the main goal is to use such metadata to understand how automatic learning can become flexible in solving learning problems, hence to improve the performance of existing learning algorithms or to learn (induce) the learning algorithm itself, hence the alternative term learning to learn.

In artificial intelligence, apprenticeship learning is the process of learning by observing an expert. It can be viewed as a form of supervised learning, where the training dataset consists of task executions by a demonstration teacher.

<span class="mw-page-title-main">Yann LeCun</span> French computer scientist (born 1960)

Yann André LeCun is a Turing Award winning French computer scientist working primarily in the fields of machine learning, computer vision, mobile robotics and computational neuroscience. He is the Silver Professor of the Courant Institute of Mathematical Sciences at New York University and Vice-President, Chief AI Scientist at Meta.

<span class="mw-page-title-main">Andrew Ng</span> American artificial intelligence researcher

Andrew Yan-Tak Ng is a British-American computer scientist and technology entrepreneur focusing on machine learning and artificial intelligence (AI). Ng was a cofounder and head of Google Brain and was the former Chief Scientist at Baidu, building the company's Artificial Intelligence Group into a team of several thousand people.

Google Brain was a deep learning artificial intelligence research team under the umbrella of Google AI, a research division at Google dedicated to artificial intelligence. Formed in 2011, Google Brain combines open-ended machine learning research with information systems and large-scale computing resources. The team has created tools such as TensorFlow, which allow for neural networks to be used by the public, with multiple internal AI research projects. The team aims to create research opportunities in machine learning and natural language processing. The team was merged into former Google sister company DeepMind to form Google DeepMind in April 2023.

An AI accelerator is a class of specialized hardware accelerator or computer system designed to accelerate artificial intelligence and machine learning applications, including artificial neural networks and machine vision. Typical applications include algorithms for robotics, Internet of Things, and other data-intensive or sensor-driven tasks. They are often manycore designs and generally focus on low-precision arithmetic, novel dataflow architectures or in-memory computing capability. As of 2018, a typical AI integrated circuit chip contains billions of MOSFET transistors. A number of vendor-specific terms exist for devices in this category, and it is an emerging technology without a dominant design.

<span class="mw-page-title-main">Outline of machine learning</span> Overview of and topical guide to machine learning

The following outline is provided as an overview of and topical guide to machine learning. Machine learning is a subfield of soft computing within computer science that evolved from the study of pattern recognition and computational learning theory in artificial intelligence. In 1959, Arthur Samuel defined machine learning as a "field of study that gives computers the ability to learn without being explicitly programmed". Machine learning explores the study and construction of algorithms that can learn from and make predictions on data. Such algorithms operate by building a model from an example training set of input observations in order to make data-driven predictions or decisions expressed as outputs, rather than following strictly static program instructions.

In artificial intelligence, researchers can induce the evolution of language in multi-agent systems when sufficiently capable AI agents have an incentive to cooperate on a task and the ability to exchange a set of symbols capable of serving as tokens in a generated language. Such languages can be evolved starting from a natural (human) language, or can be created ab initio. In addition, a new "interlingua" language may evolve within an AI tasked with translating between known languages.

<span class="mw-page-title-main">Pieter Abbeel</span> Machine learning researcher at Berkeley

Pieter Abbeel is a professor of electrical engineering and computer sciences, Director of the Berkeley Robot Learning Lab, and co-director of the Berkeley AI Research (BAIR) Lab at the University of California, Berkeley. He is also the co-founder of covariant.ai, a venture-funded start-up that aims to teach robots new, complex skills, and co-founder of Gradescope, an online grading system that has been implemented in over 500 universities nationwide. He is best known for his cutting-edge research in robotics and machine learning, particularly in deep reinforcement learning. In 2021, he joined AIX Ventures as an Investment Partner. AIX Ventures is a venture capital fund that invests in artificial intelligence startups.

<span class="mw-page-title-main">Deep reinforcement learning</span> Machine learning that combines deep learning and reinforcement learning

Deep reinforcement learning is a subfield of machine learning that combines reinforcement learning (RL) and deep learning. RL considers the problem of a computational agent learning to make decisions by trial and error. Deep RL incorporates deep learning into the solution, allowing agents to make decisions from unstructured input data without manual engineering of the state space. Deep RL algorithms are able to take in very large inputs and decide what actions to perform to optimize an objective. Deep reinforcement learning has been used for a diverse set of applications including but not limited to robotics, video games, natural language processing, computer vision, education, transportation, finance and healthcare.

In video games, various artificial intelligence techniques have been used in a variety of ways, ranging from non-player character (NPC) control to procedural content generation (PCG). Machine learning is a subset of artificial intelligence that focuses on using algorithms and statistical models to make machines act without specific programming. This is in sharp contrast to traditional methods of artificial intelligence such as search trees and expert systems.

<span class="mw-page-title-main">Multi-agent reinforcement learning</span> Sub-field of reinforcement learning

Multi-agent reinforcement learning (MARL) is a sub-field of reinforcement learning. It focuses on studying the behavior of multiple learning agents that coexist in a shared environment. Each agent is motivated by its own rewards, and does actions to advance its own interests; in some environments these interests are opposed to the interests of other agents, resulting in complex group dynamics.

Vivian Chu is an American roboticist and entrepreneur, specializing in the field of human-robot interaction. She is Chief Technology Officer at Diligent Robotics, a company she co-founded in 2017 for creating autonomous, mobile, socially intelligent robots.

<span class="mw-page-title-main">MuZero</span> Game-playing artificial intelligence

MuZero is a computer program developed by artificial intelligence research company DeepMind to master games without knowing their rules. Its release in 2019 included benchmarks of its performance in go, chess, shogi, and a standard suite of Atari games. The algorithm uses an approach similar to AlphaZero. It matched AlphaZero's performance in chess and shogi, improved on its performance in Go, and improved on the state of the art in mastering a suite of 57 Atari games, a visually-complex domain.

Isabelle Guyon is a French-born researcher in machine learning known for her work on support-vector machines, artificial neural networks and bioinformatics. She is a Chair Professor at the University of Paris-Saclay.

<span class="mw-page-title-main">Self-play</span> Reinforcement learning technique

Self-play is a technique for improving the performance of reinforcement learning agents. Intuitively, agents learn to improve their performance by playing "against themselves".

Vladlen Koltun is an Israeli-American computer scientist and intelligent systems researcher. He currently serves as distinguished scientist at Apple Inc. His main areas of research are artificial intelligence, computer vision, machine learning, and pattern recognition. He also made a significant contribution to robotics and autonomous driving.

References

  1. "Chelsea Finn: Teaching robots to learn". Berkeley Engineering. 2018-05-08. Retrieved 2022-05-20.
  2. "An Interview with Chelsea Finn: AI for Robotics". Technovation. Retrieved 2022-05-20.
  3. Natarajan, Nikhila. "Chelsea Finn is teaching Brett the Robot how the world works". ORF. Retrieved 2022-05-20.
  4. Finn, Chelsea (2018). Learning to Learn with Gradients (PDF). OCLC   1083628768.
  5. "CS 294 Deep Reinforcement Learning, Fall 2017". rail.eecs.berkeley.edu. Retrieved 2022-05-20.
  6. Kurenkov, Andrey (2021-10-14). "Chelsea Finn on Meta Learning & Model Based Reinforcement Learning". The Gradient. Retrieved 2022-05-20.
  7. 1 2 "Student Award: C.V. & Daulat Ramamoorthy Distinguished Research Award | EECS at UC Berkeley". www2.eecs.berkeley.edu. Retrieved 2022-05-20.
  8. "Chelsea Finn". CIFAR. Retrieved 2022-05-20.
  9. Interview with Professor Chelsea Finn, Stanford , retrieved 2022-05-20
  10. 1 2 3 Metz, Cade (2021-07-20). "Can A.I. Grade Your Next Test?". The New York Times. ISSN   0362-4331 . Retrieved 2022-05-20.
  11. Wu, Mike; Goodman, Noah; Piech, Chris; Finn, Chelsea (2021-10-04). "ProtoTransformer: A Meta-Learning Approach to Providing Student Feedback". arXiv: 2107.14035 [cs.CY].
  12. "Chelsea Finn – Rising Stars in EECS 2017" . Retrieved 2022-05-20.
  13. "Chelsea Finn". MIT Technology Review. Retrieved 2022-05-20.
  14. "News". EECS at UC Berkeley. Retrieved 2022-05-20.
  15. "Chelsea Finn". Association for Computing Machinery. Retrieved 2022-05-28.
  16. "Samsung AI Researcher of the Year". Samsung Advanced Institute of Technology. Retrieved 2022-05-20.
  17. "Intel's 2020 Rising Stars Awards". Intel. Retrieved 2022-05-20.
  18. "2021 Young Investigators - Office of Naval Research". www.onr.navy.mil. Retrieved 2022-05-20.
  19. "RAS Early Career Award - Academic - IEEE Robotics and Automation Society". www.ieee-ras.org. Retrieved 2022-05-20.