This article is an orphan, as no other articles link to it . Please introduce links to this page from related articles ; try the Find link tool for suggestions. (April 2024) |
Mengdi Wang | |
---|---|
Alma mater | Massachusetts Institute of Technology Tsinghua University |
Scientific career | |
Institutions | Princeton University |
Thesis | Stochastic methods for large-scale linear problems, variational inequalities, and convex optimization (2013) |
Doctoral advisor | Dimitri Bertsekas |
Mengdi Wang is a theoretical computer scientist who is a professor at Princeton University. Her research considers the fundamental theory that underpins reinforcement and machine learning. She was named one of MIT Technology Review 's 35 Under 35 in 2018.
Wang was an undergraduate student at Tsinghua University, where she specialized in automation. At the age of 18, she joined Massachusetts Institute of Technology as a graduate student, where she worked alongside Dimitri Bertsekas. [1] Her doctoral research developed stochastic methods for large-scale linear systems. [2]
Wang specializes in the theoretical frameworks that underpin machine learning and reinforcement learning. [3] She joined Princeton University as an assistant professor in 2014. [4] She was the first person to propose stochastic gradient methods for composition optimisation. [1] Her early work used reinforcement to minimize risk in financial portfolios and help hospitals identify potential complications. [3]
Wang has studied Markov decision processes, a model for reinforcement learning. She uses state compression methods to use empirical data to sketch black box Markov processes. [4]
In 2020, Wang joined the C3.ai Digital Transformation Institute, a consortium of researchers who seek to accelerate the use of artificial intelligence in society. She proposed that reinforcement learning could be used to protect educational establishments from COVID-19. [5] She used system identification and adaptive control to develop strategies to understand the health status of students, and to deploy algorithms that recommend interventions to decision makers. [5] In 2024, she was awarded a United States Department of Defense Multidisciplinary University Research Initiative program to develop AI and reinforcement learning for biological systems. [6] She showed it was possible to use large language models with semantic representation to design MRNA vaccines. [7]
In machine learning, a neural network is a model inspired by the structure and function of biological neural networks in animal brains.
Reinforcement learning (RL) is an interdisciplinary area of machine learning and optimal control concerned with how an intelligent agent ought to take actions in a dynamic environment in order to maximize the cumulative reward. Reinforcement learning is one of three basic machine learning paradigms, alongside supervised learning and unsupervised learning.
The expression computational intelligence (CI) usually refers to the ability of a computer to learn a specific task from data or experimental observation. Even though it is commonly considered a synonym of soft computing, there is still no commonly accepted definition of computational intelligence.
In probability theory and machine learning, the multi-armed bandit problem is a problem in which a decision maker iteratively selects one of multiple fixed choices when the properties of each choice are only partially known at the time of allocation, and may become better understood as time passes. A fundamental aspect of bandit problems is that choosing an arm does not affect the properties of the arm or other arms.
Claire Jennifer Tomlin is a British researcher in hybrid systems, distributed and decentralized optimization and control theory and holds the Charles A. Desoer Chair at the University of California, at Berkeley.
Harold Joseph Kushner is an American applied mathematician and a Professor Emeritus of Applied Mathematics at Brown University. He is known for his work on the theory of stochastic stability, the theory of non-linear filtering, and for the development of numerical methods for stochastic control problems such as the Markov chain approximation method. He is commonly cited as the first person to study Bayesian optimization, based on work he published in 1964.
Dimitri Panteli Bertsekas is an applied mathematician, electrical engineer, and computer scientist, a McAfee Professor at the Department of Electrical Engineering and Computer Science in School of Engineering at the Massachusetts Institute of Technology (MIT), Cambridge, Massachusetts, and also a Fulton Professor of Computational Decision Making at Arizona State University, Tempe.
Mark Herbert Ainsworth Davis was Professor of Mathematics at Imperial College London. He made fundamental contributions to the theory of stochastic processes, stochastic control and mathematical finance.
Yasuo Matsuyama is a Japanese researcher in machine learning and human-aware information processing.
Bayesian optimization is a sequential design strategy for global optimization of black-box functions that does not assume any functional forms. It is usually employed to optimize expensive-to-evaluate functions.
Bruce Edward Hajek is a Professor in the Coordinated Science Laboratory, the head of the Department of Electrical and Computer Engineering, and the Leonard C. and Mary Lou Hoeft Chair in Engineering at the University of Illinois Urbana–Champaign. He does research in communication networking, auction theory, stochastic analysis, combinatorial optimization, machine learning, information theory, and bioinformatics.
Vivek Shripad Borkar is an Indian electrical engineer, mathematician and an Institute chair professor at the Indian Institute of Technology, Mumbai. He is known for introducing analytical paradigm in stochastic optimal control processes and is an elected fellow of all the three major Indian science academies viz. the Indian Academy of Sciences, Indian National Science Academy and the National Academy of Sciences, India. He also holds elected fellowships of The World Academy of Sciences, Institute of Electrical and Electronics Engineers, Indian National Academy of Engineering and the American Mathematical Society. The Council of Scientific and Industrial Research, the apex agency of the Government of India for scientific research, awarded him the Shanti Swarup Bhatnagar Prize for Science and Technology, one of the highest Indian science awards for his contributions to Engineering Sciences in 1992. He received the TWAS Prize of the World Academy of Sciences in 2009.
The following outline is provided as an overview of and topical guide to machine learning:
Stephen John Young is a British researcher, Professor of Information Engineering at the University of Cambridge and an entrepreneur. He is one of the pioneers of automated speech recognition and statistical spoken dialogue systems. He served as the Senior Pro-Vice-Chancellor of the University of Cambridge from 2009 to 2015, responsible for planning and resources. From 2015 to 2019, he held a joint appointment between his professorship at Cambridge and Apple, where he was a senior member of the Siri development team.
Munther A. Dahleh is the William Coolidge Professor of electrical engineering and computer science and director of the Massachusetts Institute of Technology (MIT) Institute for Data, Systems, and Society (IDSS).
Animashree (Anima) Anandkumar is the Bren Professor of Computing at California Institute of Technology. Previously, she was a senior director of Machine Learning research at NVIDIA and a principal scientist at Amazon Web Services. Her research considers tensor-algebraic methods, deep learning and non-convex problems.
Deep reinforcement learning is a subfield of machine learning that combines reinforcement learning (RL) and deep learning. RL considers the problem of a computational agent learning to make decisions by trial and error. Deep RL incorporates deep learning into the solution, allowing agents to make decisions from unstructured input data without manual engineering of the state space. Deep RL algorithms are able to take in very large inputs and decide what actions to perform to optimize an objective. Deep reinforcement learning has been used for a diverse set of applications including but not limited to robotics, video games, natural language processing, computer vision, education, transportation, finance and healthcare.
Éric Moulines is a French researcher in statistical learning and signal processing. He received the silver medal from the CNRS in 2010, the France Télécom prize awarded in collaboration with the French Academy of Sciences in 2011. He was appointed a Fellow of the European Association for Signal Processing in 2012 and of the Institute of Mathematical Statistics in 2016. He is General Engineer of the Corps des Mines (X81).
Elad Hazan is an Israeli-American computer scientist, academic, author and researcher. He is a Professor of Computer Science at Princeton University, and the co-founder and director of Google AI Princeton.
Chelsea Finn is an American computer scientist and assistant professor at Stanford University. Her research investigates intelligence through the interactions of robots, with the hope to create robotic systems that can learn how to learn. She is part of the Google Brain group.