Sentence processing

Last updated

Sentence processing takes place whenever a reader or listener processes a language utterance, either in isolation or in the context of a conversation or a text. Many studies of the human language comprehension process have focused on reading of single utterances (sentences) without context. Extensive research has shown that language comprehension is affected by context preceding a given utterance as well as many other factors.

Contents

Ambiguity

Sentence comprehension has to deal with ambiguity [1] in spoken and written utterances, for example lexical, structural, and semantic ambiguities. Ambiguity is ubiquitous, but people usually resolve it so effortlessly that they do not even notice it. For example, the sentence Time flies like an arrow has (at least) the interpretations Time moves as quickly as an arrow, A special kind of fly, called time fly, likes arrows and Measure the speed of flies like you would measure the speed of an arrow. Usually, readers will be aware of only the first interpretation. Educated readers though, spontaneously think about the arrow of time but inhibit that interpretation because it deviates from the original phrase and the temporal lobe acts as a switch.

Instances of ambiguity can be classified as local or global ambiguities. A sentence is globally ambiguous if it has two distinct interpretations. Examples are sentences like Someone shot the servant of the actress who was on the balcony (was it the servant or the actress who was on the balcony?) or The cop chased the criminal with a fast car (did the cop or the criminal have a fast car?). Comprehenders may have a preferential interpretation for either of these cases, but syntactically and semantically, neither of the possible interpretations can be ruled out.

Local ambiguities persist only for a short amount of time as an utterance is heard or written and are resolved during the course of the utterance so the complete utterance has only one interpretation. Examples include sentences like The critic wrote the book was enlightening, which is ambiguous when The critic wrote the book has been encountered, but was enlightening remains to be processed. Then, the sentence could end, stating that the critic is the author of the book, or it could go on to clarify that the critic wrote something about a book. The ambiguity ends at was enlightening, which determines that the second alternative is correct.

When readers process a local ambiguity, they settle on one of the possible interpretations immediately without waiting to hear or read more words that might help decide which interpretation is correct (the behaviour is called incremental processing). If readers are surprised by the turn the sentence really takes, processing is slowed and is visible for example in reading times. Locally-ambiguous sentences have, therefore, been used as test cases to investigate the influence of a number of different factors on human sentence processing. If a factor helps readers to avoid difficulty, it is clear that the factor plays a factor in sentence processing.

Theories

Experimental research has spawned a large number of hypotheses about the architecture and mechanisms of sentence comprehension. Issues like modularity versus interactive processing and serial versus parallel computation of analyses have been theoretical divides in the field.

Architectural issues

Modular vs. interactive

A modular view of sentence processing assumes that each factor involved in sentence processing is computed in its own module, which has limited means of communication with the other modules. For example, syntactic analysis creation takes place without input from semantic analysis or context-dependent information, which are processed separately. A common assumption of modular accounts is a feed-forward architecture in which the output of one processing step is passed on to the next step without feedback mechanisms that would allow the output of the first module to be corrected. Syntactic processing is usually taken to be the most basic analysis step, which feeds into semantic processing and the inclusion of other information. A separate mental module parses sentences and lexical access happens first. Then, one syntactic hypothesis is considered at a time. There is no initial influence of meaning, or semantic. Sentence processing is supported by a temporo-frontal network. Within the network, temporal regions subserve aspects of identification and frontal regions the building of syntactic and semantic relations. Temporal analyses of brain activation within this network support syntax-first models because they reveal that building of syntactic structure precedes semantic processes and that these interact only during a later stage. [2] [3]

Interactive accounts assume that all available information is processed at the same time and can immediately influence the computation of the final analysis. In the interactive model of sentence processing, there is no separate module for parsing. Lexical access, syntactic structure assignment, and meaning assignment happen at the same time in parallel. Several syntactic hypotheses can be considered at a time. The interactive model demonstrates an on-line interaction between the structural and lexical and phonetic levels of sentence processing. Each word, as it is heard in the context of normal discourse, is immediately entered into the processing system at all levels of description, and is simultaneously analyzed at all these levels in the light of whatever information is available at each level at that point in the processing of the sentence. [4] Interactive models of language processing assume that information flows both bottom-up and top-down, so that the representations formed at each level may be influenced by higher as well as lower levels. A framework called the interactive activation framework that embeds this key assumption among others, including the assumption that influences from different sources are combined nonlinearly. The nonlinearity means that information that may be decisive under some circumstances may have little or no effect under other conditions. In the interactive activation framework, the knowledge that guides processing is stored in the connections between units on the same and adjacent levels. The processing units that they connect may receive input from a number of different sources, which allows the knowledge that guides processing to be completely local while, at the same time, allowing the results of processing at one level to influence processing at other levels, both above and below. A basic assumption of the framework is that processing interactions are always reciprocal; it is this bi-directional characteristic that makes the system interactive. Bi-directional excitatory interactions between levels allow mutual simultaneous constraint among adjacent levels, and bi-directional inhibitory interactions within a level allow for competition among mutually incompatible interpretations of a portion of an input. The between-level excitatory interactions are captured in the models in two-way excitatory connections between mutually compatible processing units. Syntactic ambiguities are in fact based at the lexical level. In addition, more recent studies with more sensitive eye tracking machines have shown early context effects. Frequency and contextual information will modulate the activation of alternatives even when they are resolved in favor of the simple interpretation. Structural simplicity is cofounded with frequency, which goes against the garden path theory [5]

Serial vs. parallel

Serial accounts assume that humans construct only one of the possible interpretations at first and try another only if the first one turns out to be wrong. Parallel accounts assume the construction of multiple interpretations at the same time. To explain why comprehenders are usually only aware of one possible analysis of what they hear, models can assume that all analyses ranked, and the highest-ranking one is entertained.

Models

There are a number of influential models of human sentence processing that draw on different combinations of architectural choices.

Garden path model

The garden path model ( Frazier 1987 ) is a serial modular parsing model. It proposes that a single parse is constructed by a syntactic module. Contextual and semantic factors influence processing at a later stage and can induce re-analysis of the syntactic parse. Re-analysis is costly and leads to an observable slowdown in reading. When the parser encounters an ambiguity, it is guided by two principles: late closure and minimal attachment. The model has been supported with research on the early left anterior negativity, an event-related potential often elicited as a response to phrase structure violations.

Late closure causes new words or phrases to be attached to the current clause. For example, "John said he would leave yesterday" would be parsed as John said (he would leave yesterday), and not as John said (he would leave) yesterday (i.e., he spoke yesterday).

Minimal attachment is a strategy of parsimony: The parser builds the simplest syntactic structure possible (that is, the one with the fewest phrasal nodes).

Constraint-based model

Constraint-based theories of language comprehension [6] emphasize how people make use of the vast amount of probabilistic information available in the linguistic signal. Through statistical learning, [7] the frequencies and distribution of events in linguistic environments can be picked upon, which inform language comprehension. As such, language users are said to arrive at a particular interpretation over another during the comprehension of an ambiguous sentence by rapidly integrating these probabilistic constraints.

Good enough theory

The good enough approach to language comprehension, developed by Fernanda Ferreira and others, assumes that listeners do not always engage in full detailed processing of linguistic input. Rather, the system has a tendency to develop shallow and superficial representations when confronted with some difficulty. The theory takes an approach that somewhat combines both the garden path model and the constraint based model. The theory focuses on two main issues. The first is that representations formed from complex or difficult material are often shallow and incomplete. The second is that limited information sources are often consulted in cases where the comprehension system encounters difficulty. The theory can be put to test using various experiments in psycholinguistics that involve garden path misinterpretation, etc. [8] [9]

Methods

Behavioral tasks

In behavioral studies, subjects are often presented with linguistic stimuli and asked to perform an action. For example, they may be asked to make a judgment about a word (lexical decision), reproduce the stimulus, or name a visually presented word aloud. Speed (often reaction time: time taken to respond to the stimulus) and accuracy (proportion of correct responses) are commonly employed measures of performance in behavioral tasks. Researchers infer that the nature of the underlying process(es) required by the task gives rise to differences; slower rates and lower accuracy on these tasks are taken as measures of increased difficulty. An important component of any behavioral task is that it stays relatively true to 'normal' language comprehension—the ability to generalize the results of any task is restricted when the task has little in common with how people actually encounter language.

A common behavioral paradigm involves priming effects, wherein participants are presented first with a prime and then with a target word. The response time for the target word is affected by the relationship between the prime and the target. For example, Fischler (1977) investigated word encoding using the lexical decision task. She asked participants to make decisions about whether two strings of letters were English words. Sometimes the strings would be actual English words requiring a "yes" response, and other times they would be nonwords requiring a "no" response. A subset of the licit words were related semantically (e.g., cat-dog) while others were unrelated (e.g., bread-stem). Fischler found that related word pairs were responded to faster when compared to unrelated word pairs, which suggests that semantic relatedness can facilitate word encoding. [10]

Eye-movements

Eye tracking has been used to study online language processing. This method has been influential in informing knowledge of reading. [11] Additionally, Tanenhaus et al. (1995) [12] established the visual world paradigm, which takes advantage of eye movements to study online spoken language processing. This area of research capitalizes on the linking hypothesis that eye movements are closely linked to the current focus of attention.

Neuroimaging and evoked potentials

The rise of non-invasive techniques provides myriad opportunities for examining the brain bases of language comprehension. Common examples include positron emission tomography (PET), functional magnetic resonance imaging (fMRI), event-related potentials (ERPs) in electroencephalography (EEG) and magnetoencephalography (MEG), and transcranial magnetic stimulation (TMS). These techniques vary in their spatial and temporal resolutions (fMRI has a resolution of a few thousand neurons per pixel, and ERP has millisecond accuracy), and each type of methodology presents a set of advantages and disadvantages for studying a particular problem in language comprehension.

Computational modeling

Computational modeling is another means by which to explore language comprehension. Models, such as those instantiated in neural networks, are particularly useful because they requires theorists to be explicit in their hypotheses and because they can be used to generate accurate predictions for theoretical models that are so complex that they render discursive analysis unreliable. A classic example of computational modeling in language research is McClelland and Elman's TRACE model of speech perception. [13] A model of sentence processing can be found in Hale (2011)'s 'rational' Generalized Left Corner parser. [14] This model derives garden path effects as well as local coherence phenomena. Computational modeling can also help to relate sentence processing to other functions of language. For example, one model of ERP effects in sentence processing (e.g., N400 and P600) argues that these phenomena arise out learning processes that support language acquisition and linguistic adaptation. [15]

See also

Notes

  1. Altmann, Gerry (April 1998). "Ambiguity in sentence processing". Trends in Cognitive Sciences. 2 (4): 146–151. doi:10.1016/s1364-6613(98)01153-x. PMID   21227111. S2CID   12113211.
  2. Hillert, D., ed. (1998). Sentence Processing: A Cross-Linguistic Perspective. Syntax and Semantics 31. San Diego: Academic Press. p. 464. ISBN   978-0126135312.
  3. Friederici, Angela (1 February 2002). "Towards a neural basis of auditory sentence processing". Trends in Cognitive Sciences. 6 (2): 78–84. doi: 10.1016/S1364-6613(00)01839-8 . hdl: 11858/00-001M-0000-0010-E573-8 . PMID   15866191.
  4. Abrahams, V. C.; Rose, P. K. (18 July 1975). "Sentence perception as an interactive parallel process". Science. 189 (4198): 226–228. Bibcode:1975Sci...189..226M. doi:10.1126/science.189.4198.226. PMID   17733889. S2CID   30410322.
  5. MacDonald, Pearlmutter and Seidenberg, 1994).
  6. MacDonald, M. C.; Pearlmutter, M.; Seidenberg, M. (1994). "The Lexical Nature of Ambiguity Resolution". Psychological Review. 101 (4): 676–703. doi:10.1037/0033-295x.101.4.676. PMID   7984711. S2CID   15560738.
  7. Seidenberg, Mark S.; J.L. McClelland (1989). "A distributed developmental model of word recognition and naming". Psychological Review. 96 (4): 523–568. CiteSeerX   10.1.1.127.3083 . doi:10.1037/0033-295X.96.4.523. PMID   2798649.
  8. Ferreira et al. (2009)
  9. Ferreira et al. (2002)
  10. Fischler I. (1977). "Semantic facilitation without association in a lexical decision task". Memory & Cognition. 5 (3): 335–339. doi: 10.3758/bf03197580 . PMID   24202904.
  11. Rayner K. (1978). "Eye movements in reading and information processing". Psychological Bulletin. 85 (3): 618–660. CiteSeerX   10.1.1.294.4262 . doi:10.1037/0033-2909.85.3.618. PMID   353867.
  12. Tanenhaus M. K.; Spivey-Knowlton M. J.; Eberhard K. M.; Sedivy J. E. (1995). "Integration of visual and linguistic information in spoken language comprehension". Science. 268 (5217): 1632–1634. Bibcode:1995Sci...268.1632T. doi:10.1126/science.7777863. PMID   7777863. S2CID   3073956.
  13. McClelland, J.L.; Elman, J.L. (1986). "The TRACE model of speech perception". Cognitive Psychology. 18 (1): 1–86. doi:10.1016/0010-0285(86)90015-0. PMID   3753912. S2CID   7428866.
  14. Hale, John T. (2011). "What a Rational Parser Would do". Cognitive Science. 35 (3): 399–443. doi: 10.1111/j.1551-6709.2010.01145.x .
  15. Fitz, Hartmut; Chang, Franklin (2019-06-01). "Language ERPs reflect learning through prediction error propagation". Cognitive Psychology. 111: 15–52. doi:10.1016/j.cogpsych.2019.03.002. hdl: 21.11116/0000-0003-474D-8 . ISSN   0010-0285. PMID   30921626. S2CID   85501792.

Related Research Articles

Natural language processing (NLP) is an interdisciplinary subfield of computer science and artificial intelligence. It is primarily concerned with providing computers with the ability to process data encoded in natural language and is thus closely related to information retrieval, knowledge representation and computational linguistics, a subfield of linguistics. Typically data is collected in text corpora, using either rule-based, statistical or neural-based approaches in machine learning and deep learning.

<span class="mw-page-title-main">Broca's area</span> Speech production region in the dominant hemisphere of the hominid brain

Broca's area, or the Broca area, is a region in the frontal lobe of the dominant hemisphere, usually the left, of the brain with functions linked to speech production.

Syntactic ambiguity, also known as structural ambiguity, amphiboly, or amphibology, is characterized by the potential for a sentence to yield multiple interpretations due to its ambiguous syntax. This form of ambiguity is not derived from the varied meanings of individual words but rather from the relationships among words and clauses within a sentence, concealing interpretations beneath the word order. Consequently, a sentence presents as syntactically ambiguous when it permits reasonable derivation of several possible grammatical structures by an observer.

Psycholinguistics or psychology of language is the study of the interrelation between linguistic factors and psychological aspects. The discipline is mainly concerned with the mechanisms by which language is processed and represented in the mind and brain; that is, the psychological and neurobiological factors that enable humans to acquire, use, comprehend, and produce language.

<span class="mw-page-title-main">Neurolinguistics</span> Neuroscience and linguistics-related studies

Neurolinguistics is the study of neural mechanisms in the human brain that control the comprehension, production, and acquisition of language. As an interdisciplinary field, neurolinguistics draws methods and theories from fields such as neuroscience, linguistics, cognitive science, communication disorders and neuropsychology. Researchers are drawn to the field from a variety of backgrounds, bringing along a variety of experimental techniques as well as widely varying theoretical perspectives. Much work in neurolinguistics is informed by models in psycholinguistics and theoretical linguistics, and is focused on investigating how the brain can implement the processes that theoretical and psycholinguistics propose are necessary in producing and comprehending language. Neurolinguists study the physiological mechanisms by which the brain processes information related to language, and evaluate linguistic and psycholinguistic theories, using aphasiology, brain imaging, electrophysiology, and computer modeling.

Parsing, syntax analysis, or syntactic analysis is the process of analyzing a string of symbols, either in natural language, computer languages or data structures, conforming to the rules of a formal grammar. The term parsing comes from Latin pars (orationis), meaning part.

A garden-path sentence is a grammatically correct sentence that starts in such a way that a reader's most likely interpretation will be incorrect; the reader is lured into a parse that turns out to be a dead end or yields a clearly unintended meaning. "Garden path" refers to the saying "to be led down [or up] the garden path", meaning to be deceived, tricked, or seduced. In A Dictionary of Modern English Usage (1926), Fowler describes such sentences as unwittingly laying a "false scent".

In semantics, mathematical logic and related disciplines, the principle of compositionality is the principle that the meaning of a complex expression is determined by the meanings of its constituent expressions and the rules used to combine them. The principle is also called Frege's principle, because Gottlob Frege is widely credited for the first modern formulation of it. However, the principle has never been explicitly stated by Frege, and arguably it was already assumed by George Boole decades before Frege's work.

In linguistics, prosody is the study of elements of speech that are not individual phonetic segments but which are properties of syllables and larger units of speech, including linguistic functions such as intonation, stress, and rhythm. Such elements are known as suprasegmentals.

A speech error, commonly referred to as a slip of the tongue or misspeaking, is a deviation from the apparently intended form of an utterance. They can be subdivided into spontaneously and inadvertently produced speech errors and intentionally produced word-plays or puns. Another distinction can be drawn between production and comprehension errors. Errors in speech production and perception are also called performance errors. Some examples of speech error include sound exchange or sound anticipation errors. In sound exchange errors, the order of two individual morphemes is reversed, while in sound anticipation errors a sound from a later syllable replaces one from an earlier syllable. Slips of the tongue are a normal and common occurrence. One study shows that most people can make up to as much as 22 slips of the tongue per day.

Language production is the production of spoken or written language. In psycholinguistics, it describes all of the stages between having a concept to express and translating that concept into linguistic forms. These stages have been described in two types of processing models: the lexical access models and the serial models. Through these models, psycholinguists can look into how speeches are produced in different ways, such as when the speaker is bilingual. Psycholinguists learn more about these models and different kinds of speech by using language production research methods that include collecting speech errors and elicited production tasks.

The term linguistic performance was used by Noam Chomsky in 1960 to describe "the actual use of language in concrete situations". It is used to describe both the production, sometimes called parole, as well as the comprehension of language. Performance is defined in opposition to "competence"; the latter describes the mental knowledge that a speaker or listener has of language.

David Swinney was a prominent psycholinguist. His research on language comprehension contributed to methodological advances in his field.

<span class="mw-page-title-main">Janet Dean Fodor</span> American linguist (1942–2023)

Janet Dean Fodor was distinguished professor emerita of linguistics at the Graduate Center of the City University of New York. Her primary field was psycholinguistics, and her research interests included human sentence processing, prosody, learnability theory and L1 (first-language) acquisition.

The P600 is an event-related potential (ERP) component, or peak in electrical brain activity measured by electroencephalography (EEG). It is a language-relevant ERP component and is thought to be elicited by hearing or reading grammatical errors and other syntactic anomalies. Therefore, it is a common topic of study in neurolinguistic experiments investigating sentence processing in the human brain.

The early left anterior negativity is an event-related potential in electroencephalography (EEG), or component of brain activity that occurs in response to a certain kind of stimulus. It is characterized by a negative-going wave that peaks around 200 milliseconds or less after the onset of a stimulus, and most often occurs in response to linguistic stimuli that violate word-category or phrase structure rules. As such, it is frequently a topic of study in neurolinguistics experiments, specifically in areas such as sentence processing. While it is frequently used in language research, there is no evidence yet that it is necessarily a language-specific phenomenon.

Linguistic prediction is a phenomenon in psycholinguistics occurring whenever information about a word or other linguistic unit is activated before that unit is actually encountered. Evidence from eyetracking, event-related potentials, and other experimental methods indicates that in addition to integrating each subsequent word into the context formed by previously encountered words, language users may, under certain conditions, try to predict upcoming words. In particular, prediction seems to occur regularly when the context of a sentence greatly limits the possible words that have not yet been revealed. For instance, a person listening to a sentence like, "In the summer it is hot, and in the winter it is..." would be highly likely to predict the sentence completion "cold" in advance of actually hearing it. A form of prediction is also thought to occur in some types of lexical priming, a phenomenon whereby a word becomes easier to process if it is preceded by a related word. Linguistic prediction is an active area of research in psycholinguistics and cognitive neuroscience.

Bilingual lexical access is an area of psycholinguistics that studies the activation or retrieval process of the mental lexicon for bilingual people.

<span class="mw-page-title-main">Semantic parsing</span>

Semantic parsing is the task of converting a natural language utterance to a logical form: a machine-understandable representation of its meaning. Semantic parsing can thus be understood as extracting the precise meaning of an utterance. Applications of semantic parsing include machine translation, question answering, ontology induction, automated reasoning, and code generation. The phrase was first used in the 1970s by Yorick Wilks as the basis for machine translation programs working with only semantic representations. Semantic parsing is one of the important tasks in computational linguistics and natural language processing.

Maria Fernanda Ferreira is a cognitive psychologist known for empirical investigations in psycholinguistics and language processing. Ferreira is Professor of Psychology and the Principal investigator of the Ferreira Lab at University of California, Davis.

References

Further reading