Christopher David Manning (born September 18, 1965) is a computer scientist and applied linguist whose research in the areas of natural language processing, artificial intelligence and machine learning is considered highly influential. He is the current Director of the Stanford Artificial Intelligence Laboratory (SAIL).
Manning has been described as “the leading researcher in natural language processing”, [1] well known for co-developing GloVe word vectors; the bilinear or multiplicative form of attention, now widely used in artificial neural networks including the transformer; tree-structured recursive neural networks; and approaches to and systems for Textual entailment. His main educational contributions are his textbooks Foundations of Statistical Natural Language Processing (1999) and Introduction to Information Retrieval (2008), and his course CS224N Natural Language Processing with Deep Learning, which is available online. Manning also pioneered the development of well-maintained open source computational linguistics software packages, including CoreNLP, Stanza, and GloVe. [2] [3] [4] [5]
Manning is the Thomas M. Siebel Professor in Machine Learning and a professor of Linguistics and Computer Science at Stanford University. He received a BA (Hons) degree majoring in mathematics, computer science, and linguistics from the Australian National University (1989) and a PhD in linguistics from Stanford (1994), under the guidance of Joan Bresnan. [6] [7] He was an assistant professor at Carnegie Mellon University (1994–96) and a lecturer at the University of Sydney (1996–99) before returning to Stanford as an assistant professor. At Stanford, he was promoted to associate professor in 2006 and to full professor in 2012. He was elected an AAAI Fellow in 2010. [8] He was previously President of the Association for Computational Linguistics (2015) and he has received an honorary doctorate from the University of Amsterdam (2023). Manning was awarded the IEEE John von Neumann Medal “for advances in computational representation and analysis of natural language” in 2024. [9] [1]
Manning's linguistic work includes his dissertation Ergativity: Argument Structure and Grammatical Relations (1996), a monograph Complex Predicates and Information Spreading in LFG (1999), [10] and his work developing Universal Dependencies, [11] from which he is the namesake of Manning's Law.
Manning's PhD students include Dan Klein, Sepandar Kamvar, Richard Socher, and Danqi Chen. [7] In 2021, he joined AIX Ventures [12] as an Investing Partner. AIX Ventures is a venture capital fund that invests in artificial intelligence startups.
Natural language processing (NLP) is a subfield of computer science and especially artificial intelligence. It is primarily concerned with providing computers with the ability to process data encoded in natural language and is thus closely related to information retrieval, knowledge representation and computational linguistics, a subfield of linguistics. Typically data is collected in text corpora, using either rule-based, statistical or neural-based approaches in machine learning and deep learning.
Natural language understanding (NLU) or natural language interpretation (NLI) is a subset of natural language processing in artificial intelligence that deals with machine reading comprehension. NLU has been considered an AI-hard problem.
Lemmatization in linguistics is the process of grouping together the inflected forms of a word so they can be analysed as a single item, identified by the word's lemma, or dictionary form.
Document clustering is the application of cluster analysis to textual documents. It has applications in automatic document organization, topic extraction and fast information retrieval or filtering.
Daniel Klein is an American computer scientist and professor of computer science at the University of California, Berkeley. His research focuses on natural language processing and artificial intelligence.
Karen Ida Boalth Spärck Jones was a self-taught programmer and a pioneering British computer scientist responsible for the concept of inverse document frequency (IDF), a technology that underlies most modern search engines. She was an advocate for women in computer science, her slogan being, "Computing is too important to be left to men." In 2019, The New York Times published her belated obituary in its series Overlooked, calling her "a pioneer of computer science for work combining statistics and linguistics, and an advocate for women in the field." From 2008, to recognize her achievements in the fields of information retrieval (IR) and natural language processing (NLP), the Karen Spärck Jones Award is awarded annually to a recipient for outstanding research in one or both of her fields.
Language and Communication Technologies is the scientific study of technologies that explore language and communication. It is an interdisciplinary field that encompasses the fields of computer science, linguistics and cognitive science.
Dragomir R. Radev was an American computer scientist who was a professor at Yale University, working on natural language processing and information retrieval. He also served as a University of Michigan computer science professor and Columbia University computer science adjunct professor, as well as a Member of the Advisory Board of Lawyaw.
Deep linguistic processing is a natural language processing framework which draws on theoretical and descriptive linguistics. It models language predominantly by way of theoretical syntactic/semantic theory. Deep linguistic processing approaches differ from "shallower" methods in that they yield more expressive and structural representations which directly capture long-distance dependencies and underlying predicate-argument structures.
The knowledge-intensive approach of deep linguistic processing requires considerable computational power, and has in the past sometimes been judged as being intractable. However, research in the early 2000s had made considerable advancement in efficiency of deep processing. Today, efficiency is no longer a major problem for applications using deep linguistic processing.
The following outline is provided as an overview of and topical guide to natural-language processing:
In natural language processing, a word embedding is a representation of a word. The embedding is used in text analysis. Typically, the representation is a real-valued vector that encodes the meaning of the word in such a way that the words that are closer in the vector space are expected to be similar in meaning. Word embeddings can be obtained using language modeling and feature learning techniques, where words or phrases from the vocabulary are mapped to vectors of real numbers.
Dan Roth is the Eduardo D. Glandt Distinguished Professor of Computer and Information Science at the University of Pennsylvania and the Chief AI Scientist at Oracle. Until June 2024 Dan was a VP/Distinguished Scientist at AWS AI. In his role at AWS Roth led over the last three years the scientific effort behind the first-generation Generative AI products from AWS, including Titan Models, Amazon Q efforts, and Bedrock, from inception until they became generally available.
Semantic folding theory describes a procedure for encoding the semantics of natural language text in a semantically grounded binary representation. This approach provides a framework for modelling how language data is processed by the neocortex.
Pascale Fung (馮雁) is a professor in the Department of Electronic & Computer Engineering and the Department of Computer Science & Engineering at the Hong Kong University of Science & Technology(HKUST). She is the director of the Centre for AI Research (CAiRE) at HKUST. She is an elected Fellow of the Institute of Electrical and Electronics Engineers (IEEE) for her “contributions to human-machine interactions”, an elected Fellow of the International Speech Communication Association for “fundamental contributions to the interdisciplinary area of spoken language human-machine interactions” and an elected Fellow of the Association for Computational Linguistics (ACL) for her “significant contributions toward statistical NLP, comparable corpora, and building intelligent systems that can understand and empathize with humans”.
Semantic spaces in the natural language domain aim to create representations of natural language that are capable of capturing meaning. The original motivation for semantic spaces stems from two core challenges of natural language: Vocabulary mismatch and ambiguity of natural language.
Pushpak Bhattacharyya is a computer scientist and a professor at Computer Science and Engineering Department, IIT Bombay. He served as the director of Indian Institute of Technology Patna from 2015 to 2021. He is regarded as the Godfather of NLP in India, mentioned by Nandan Nilekani, Co-founder and non-executive chairman of Infosys at the Inaugural event of Nilekani Centre at AI4Bharat, IIT Madras. He is a past president of Association for Computational Linguistics (2016–17), and Ex-Vijay and Sita Vashee Chair Professor He currently heads the Natural language processing research group Center For Indian Language Technology (CFILT) lab at IIT Bombay.
Lillian Lee is a computer scientist whose research involves natural language processing, sentiment analysis, and computational social science. She is a professor of computer science and information science at Cornell University, and co-editor-in-chief of the journal Transactions of the Association for Computational Linguistics.
Martha (Stone) Palmer is an American computer scientist. She is best known for her work on verb semantics, and for the creation of ontological resources such as PropBank and VerbNet.
Danqi Chen is a Chinese computer scientist and assistant professor at Princeton University specializing in the AI field of natural language processing (NLP). In 2019, she joined the Princeton NLP group, alongside Sanjeev Arora, Christiane Fellbaum, and Karthik Narasimhan. She was previously a visiting scientist at Facebook AI Research (FAIR). She earned her Ph.D. at Stanford University and her BS from Tsinghua University.
Bonnie Jean Dorr is an American computer scientist specializing in natural language processing, machine translation, automatic summarization, social computing, and explainable artificial intelligence. She is a professor and director of the Natural Language Processing Research Laboratory in the Department of Computer & Information Science & Engineering at the University of Florida. Gainesville, Florida She is professor emerita of computer science and linguistics and former dean at the University of Maryland, College Park, former associate director at the Florida Institute for Human and Machine Cognition,, and former president of the Association for Computational Linguistics.