This biography of a living person relies too much on references to primary sources .(September 2021) |
Scott Andrew Crossley | |
---|---|
Born | Simi Valley, California, US | 1 August 1973
Alma mater | |
Known for | |
Scientific career | |
Fields | |
Institutions | |
Thesis | A computational approach to assessing second language reading texts. (2006) |
Website | shared |
Scott Andrew Crossley (born 1973) is an American linguist. He is a professor of applied linguistics at Vanderbilt University, United States. [1] His research focuses on natural language processing and the application of computational tools and machine learning algorithms in learning analytics including second language acquisition, second language writing, and readability. His main interest area is the development and use of natural language processing tools in assessing writing quality and text difficulty.
Along with Cumming, Hyland, Kormos, Matsuda, Manchón, Ortega, Polio, Storch and Verspoor he is considered one of the most prominent researchers on second language writing.
Crossley obtained his Bachelor of Arts degree in history at the California State University of Northridge in 1999. In the same year he obtained a Teaching English as a second or foreign language certificate at the University of Memphis. He got his Master of Arts degree in English language at the University of Memphis in 2002 and a Doctor of Philosophy degree in 2006.
Crossley is noted for his work on second language acquisition with a special focus on second language writing and second language vocabulary. In his research he focused on coherence, cohesion, latent semantic analysis, hypernymy and readability. He is noted for the use of computational tools such as Coh-Metrix.
Crossley has been involved in the development of numerous natural language processing tools such as Constructed Response Analysis Tool (CRAT), Sentiment Analysis and Cognition Engine (SEANCE), Simple NLP (SiNLP), Tool for the Automatic Analysis of Lexical Sophistication (TAALES), Tool for the Automatic Analysis of Text Cohesion (TAACO), Tool for the Automatic Analysis of Syntactic Sophistication and Complexity (TAASSC).
Crossley's work has been published in the Written Communication, [2] The Modern Language Journal, [3] TESOL Quarterly, [4] Journal of Second Language Writing, [5] Language Learning, [6] and Studies in Second Language Acquisition. [7]
Corpus linguistics is the study of a language as that language is expressed in its text corpus, its body of "real world" text. Corpus linguistics proposes that a reliable analysis of a language is more feasible with corpora collected in the field—the natural context ("realia") of that language—with minimal experimental interference. The large collections of text allow linguists to run quantitative analyses on linguistic concepts, otherwise harder to quantify.
Word-sense disambiguation (WSD) is the process of identifying which sense of a word is meant in a sentence or other segment of context. In human language processing and cognition, it is usually subconscious/automatic but can often come to conscious attention when ambiguity impairs clarity of communication, given the pervasive polysemy in natural language. In computational linguistics, it is an open problem that affects other computer-related writing, such as discourse, improving relevance of search engines, anaphora resolution, coherence, and inference.
Readability is the ease with which a reader can understand a written text. The concept exists in both in natural language and programming languages though in different forms. In natural language, the readability of text depends on its content and its presentation. In programming, things such as programmer comments, choice of loop structure, and choice of names can determine the ease with which humans can read computer program code.
Second language writing is the study of writing performed by non-native speakers/writers of a language as a second or foreign language. According to Oxford University, second language writing is the expression of one's actions and what one wants to say in writing in a language other than one's native language. Learning a new language and writing in it is the most challenging thing. Learning a new language first requires an understanding of the writing system and the grammar of the language. Because grammar is the basis of writing. Learning the grammar of a language is the only way to write in that language. The extent to which non-native speakers write in formal or specialized domains, and the requirements for grammatical accuracy and compositional coherence, will vary according to the specific context. The process of second language writing has been an area of research in applied linguistics and second language acquisition theory since the middle of the 20th century. The focus has been mainly on second-language writing in academic settings. In the last few years, there has been a great deal of interest in and research on informal writing. These informal writings include writing in online contexts. In terms of instructional practices, the focus of second language writing instruction has traditionally been on achieving grammatical accuracy. However, this changed under the influence of compositional studies, which focused on conceptual and structural properties. Another development in the teaching of second language writing is the increasing use of models and the emphasis on the properties of particular writing genres. Recent research has analyzed how second-language writing differs from native-language writing, emphasizing the cultural factors that influence second-language writers. In general, second language acquisition research has transitioned from a primary focus on cognitive factors to a sociocultural perspective in which writing is viewed not only as an acquired language skill and cognitive ability but also, more broadly, as a socially situated communicative act involving a target audience. Recently, particular attention has been paid to the integration of written texts with other media (multimodality) and to the mixing of languages in online media.
Stylometry is the application of the study of linguistic style, usually to written language. It has also been applied successfully to music, paintings, and chess.
The sequence between semantic related ordered words is classified as a lexical chain. A lexical chain is a sequence of related words in writing, spanning narrow or wide context window. A lexical chain is independent of the grammatical structure of the text and in effect it is a list of words that captures a portion of the cohesive structure of the text. A lexical chain can provide a context for the resolution of an ambiguous term and enable disambiguation of concepts that the term represents.
Beryl T. "Sue" Atkins was a British lexicographer, specialising in computational lexicography, who pioneered the creation of bilingual dictionaries from corpus data.
Distributional semantics is a research area that develops and studies theories and methods for quantifying and categorizing semantic similarities between linguistic items based on their distributional properties in large samples of language data. The basic idea of distributional semantics can be summed up in the so-called distributional hypothesis: linguistic items with similar distributions have similar meanings.
Kathleen Bardovi-Harlig is an American linguist. She is currently Provost Professor and ESL Coordinator at Indiana University (Bloomington).
Linguistic categories include
Computational lexicology is a branch of computational linguistics, which is concerned with the use of computers in the study of lexicon. It has been more narrowly described by some scholars as the use of computers in the study of machine-readable dictionaries. It is distinguished from computational lexicography, which more properly would be the use of computers in the construction of dictionaries, though some researchers have used computational lexicography as synonymous.
Coh-Metrix is a computational tool that produces indices of the linguistic and discourse representations of a text. Developed by Arthur C. Graesser and Danielle S. McNamara, Coh-Metrix analyzes texts on many different features.
Linguistics is the scientific study of language. Linguistics is based on a theoretical as well as a descriptive study of language and is also interlinked with the applied fields of language studies and language learning, which entails the study of specific languages. Before the 20th century, linguistics evolved in conjunction with literary study and did not employ scientific methods. Modern-day linguistics is considered a science because it entails a comprehensive, systematic, objective, and precise analysis of all aspects of language – i.e., the cognitive, the social, the cultural, the psychological, the environmental, the biological, the literary, the grammatical, the paleographical, and the structural.
Norbert Schmitt is an American applied linguist and Emeritus Professor of Applied Linguistics at the University of Nottingham in the United Kingdom. He is known for his work on second-language vocabulary acquisition and second-language vocabulary teaching. He has published numerous books and papers on vocabulary acquisition.
David Ian Hanauer is Professor of Applied Linguistics/English at Indiana University of Pennsylvania and the Lead Assessment Coordinator for the SEA-PHAGES program at the University of Pittsburgh. He is the editor of the Scientific Study of Literature journal, the official publication of IGEL. Hanauer is an applied linguist specializing in assessment and literacy practices in the sciences and poetic inquiry. He has authored or co-authored over 75 journal articles and book chapters as well as 8 books. Hanauer’s research agenda is typified by the combination of qualitative and quantitative methods, as well as arts-based approaches, and scientific measurement of concepts traditionally considered abstract, such as voice in written text, project ownership and poeticity.
Marjolijn Verspoor is a Dutch linguist. She is a professor of English language and English as a second language at the University of Groningen, Netherlands. She is known for her work on Complex Dynamic Systems Theory and the application of dynamical systems theory to study second language development. Her interest is also in second language writing.
L2 Syntactical Complexity Analyzer (L2SCA) developed by Xiaofei Lu at the Pennsylvania State University, is a computational tool which produces syntactic complexity indices of written English language texts. Along with Coh-Metrix, the L2SCA is one of the most extensively used computational tool to compute indices of second language writing development. The L2SCA is also widely utilised in the field of corpus linguistics. The L2SCA is available in a single and a batch mode. The first provides the possibility of analyzing a single written text for 14 syntactic complexity indices. The latter allows the user to analyze 30 written texts simultaneously.
Scott Jarvis is an American linguist. He is a Professor of Applied Linguistics at Northern Arizona University, United States. His research focuses on second language acquisition more broadly, with a special focus on lexical diversity.
Danielle S. McNamara is an educational researcher known for her theoretical and empirical work with reading comprehension and the development of game-based literacy technologies. She is professor of psychology and senior research scientist at Arizona State University. She has previously held positions at University of Memphis, Old Dominion University, and University of Colorado, Boulder.
In linguistics and language technology, a language resource is a "[composition] of linguistic material used in the construction, improvement and/or evaluation of language processing applications, (...) in language and language-mediated research studies and applications."