This article needs additional citations for verification .(September 2021) |
Part of a series on |
Translation |
---|
Types |
Theory |
Technologies |
Localization |
Institutional |
Related topics |
|
Transcription in the linguistic sense is the systematic representation of spoken language in written form. The source can either be utterances (speech or sign language ) or preexisting text in another writing system.
Transcription should not be confused with translation, which means representing the meaning of text from a source-language in a target language, (e.g. Los Angeles (from source-language Spanish) means The Angels in the target language English); or with transliteration, which means representing the spelling of a text from one script to another.
In the academic discipline of linguistics, transcription is an essential part of the methodologies of (among others) phonetics, conversation analysis, dialectology, and sociolinguistics. It also plays an important role for several subfields of speech technology. Common examples for transcriptions outside academia are the proceedings of a court hearing such as a criminal trial (by a court reporter) or a physician's recorded voice notes (medical transcription). This article focuses on transcription in linguistics.
There are two main types of linguistic transcription. Phonetic transcription focuses on phonetic and phonological properties of spoken language. Systems for phonetic transcription thus furnish rules for mapping individual sounds or phones to written symbols. Systems for orthographic transcription, by contrast, consist of rules for mapping spoken words onto written forms as prescribed by the orthography of a given language. Phonetic transcription operates with specially defined character sets, usually the International Phonetic Alphabet.
The type of transcription chosen depends mostly on the context of usage. Because phonetic transcription strictly foregrounds the phonetic nature of language, it is mostly used for phonetic or phonological analyses. Orthographic transcription, however, has a morphological and a lexical component alongside the phonetic component (which aspect is represented to which degree depends on the language and orthography in question). This form of transcription is thus more convenient wherever semantic aspects of spoken language are transcribed. Phonetic transcription is more systematic in a scientific sense, but it is also more difficult to learn, more time-consuming to carry out and less widely applicable than orthographic transcription.
Mapping spoken language onto written symbols is not as straightforward a process as may seem at first glance. Written language is an idealization, made up of a limited set of clearly distinct and discrete symbols. Spoken language, on the other hand, is a continuous (as opposed to discrete) phenomenon, made up of a potentially unlimited number of components. There is no predetermined system for distinguishing and classifying these components and, consequently, no preset way of mapping these components onto written symbols.
Literature is relatively consistent in pointing out the nonneutrality of transcription practices. There is not and cannot be a neutral transcription system. Knowledge of social culture enters directly into the making of a transcript. They are captured in the texture of the transcript (Baker, 2005).
Transcription systems are sets of rules which define how spoken language is to be represented in written symbols. Most phonetic transcription systems are based on the International Phonetic Alphabet or, especially in speech technology, on its derivative SAMPA.
Examples for orthographic transcription systems (all from the field of conversation analysis or related fields) are:
Arguably the first system of its kind, originally sketched in (Sacks et al. 1978), later adapted for the use in computer readable corpora as CA-CHAT by (MacWhinney 2000). The field of Conversation Analysis itself includes a number of distinct approaches to transcription and sets of transcription conventions. These include, among others, Jefferson Notation. To analyze conversation, recorded data is typically transcribed into a written form that is agreeable to analysts. There are two common approaches. The first, called narrow transcription, captures the details of conversational interaction such as which particular words are stressed, which words are spoken with increased loudness, points at which the turns-at-talk overlap, how particular words are articulated, and so on. If such detail is less important, perhaps because the analyst is more concerned with the overall gross structure of the conversation or the relative distribution of turns-at-talk amongst the participants, then a second type of transcription known as broad transcription may be sufficient (Williamson, 2009).
The Jefferson Transcription System is a set of symbols, developed by Gail Jefferson, which is used for transcribing talk. Having had some previous experience in transcribing when she was hired in 1963 as a clerk typist at the UCLA Department of Public Health to transcribe sensitivity-training sessions for prison guards, Jefferson began transcribing some of the recordings that served as the materials out of which Harvey Sacks' earliest lectures were developed. Over four decades, for the majority of which she held no university position and was unsalaried, Jefferson's research into talk-in-interaction has set the standard for what became known as conversation analysis (CA). Her work has greatly influenced the sociological study of interaction, but also disciplines beyond, especially linguistics, communication, and anthropology. [1] This system is employed universally by those working from the CA perspective and is regarded as having become a near-globalized set of instructions for transcription. [2]
A system described in (DuBois et al. 1992), used for transcription of the Santa Barbara Corpus of Spoken American English [3] (SBCSAE), later developed further into DT2.
A system described in (Selting et al. 1998), later developed further into GAT2 (Selting et al. 2009), widely used in German speaking countries for prosodically oriented conversation analysis and interactional linguistics. [4] [5]
Arguably the first system of its kind, originally described in (Ehlich and Rehbein 1976) – see (Ehlich 1992) for an English reference - adapted for the use in computer readable corpora as (Rehbein et al. 2004), and widely used in functional pragmatics. [6] [7] [8]
Transcription was originally a process carried out manually, i.e. with pencil and paper, using an analogue sound recording stored on, e.g., a Compact Cassette. Nowadays, most transcription is done on computers. Recordings are usually digital audio files or video files, and transcriptions are electronic documents. Specialized computer software exists to assist the transcriber in efficiently creating a digital transcription from a digital recording.
Two types of transcription software can be used to assist the process of transcription: one that facilitates manual transcription and the other automated transcription. For the former, the work is still very much done by a human transcriber who listens to a recording and types up what is heard in a computer, and this type of software is often a multimedia player with functionality such as playback or changing speed. For the latter, automated transcription is achieved by a speech-to-text engine which converts audio or video files into electronic text. Some of the software would also include the function of annotation. [9]
The International Phonetic Alphabet (IPA) is an alphabetic system of phonetic notation based primarily on the Latin script. It was devised by the International Phonetic Association in the late 19th century as a standard written representation for the sounds of speech. The IPA is used by lexicographers, foreign language students and teachers, linguists, speech–language pathologists, singers, actors, constructed language creators, and translators.
Morphophonology is the branch of linguistics that studies the interaction between morphological and phonological or phonetic processes. Its chief focus is the sound changes that take place in morphemes when they combine to form words.
An orthography is a set of conventions for writing a language, including norms of spelling, punctuation, word boundaries, capitalization, hyphenation, and emphasis.
A phoneme is any set of similar speech sounds that is perceptually regarded by the speakers of a language as a single basic sound—a smallest possible phonetic unit—that helps distinguish one word from another. All languages contains phonemes, and all spoken languages include both consonant and vowel phonemes. Phonemes are primarily studied under the branch of linguistics known as phonology.
In phonetics, a phone is any distinct speech sound or gesture, regardless of whether the exact sound is critical to the meanings of words.
Transcription refers to the process of converting sounds into letters or musical notes, or producing a copy of something in another medium, including:
Greeklish, a portmanteau of the words Greek and English, also known as Grenglish, Latinoellinika/Λατινοελληνικά or ASCII Greek, is the Greek language written using the Latin script. Unlike standardized systems of Romanization of Greek, as used internationally for purposes such as rendering Greek proper names or place names, or for bibliographic purposes, the term Greeklish mainly refers to informal, ad-hoc practices of writing Greek text in environments where the use of the Greek alphabet is technically impossible or cumbersome, especially in electronic media. Greeklish was commonly used on the Internet when Greek people communicate by forum, e-mail, IRC, instant messaging and occasionally on SMS, mainly because older operating systems did not support non-Latin writing systems, or in a unicode form like UTF-8. Nowadays most Greek language content appears in the Greek alphabet.
Phonetic transcription is the visual representation of speech sounds by means of symbols. The most common type of phonetic transcription uses a phonetic alphabet, such as the International Phonetic Alphabet.
Conversation analysis (CA) is an approach to the study of social interaction that investigates the methods members use to achieve mutual understanding through the transcription of naturally occurring conversations from audio or video. It focuses on both verbal and non-verbal conduct, especially in situations of everyday life. CA originated as a sociological method, but has since spread to other fields. CA began with a focus on casual conversation, but its methods were subsequently adapted to embrace more task- and institution-centered interactions, such as those occurring in doctors' offices, courts, law enforcement, helplines, educational settings, and the mass media, and focus on multimodal and nonverbal activity in interaction, including gaze, body movement and gesture. As a consequence, the term conversation analysis has become something of a misnomer, but it has continued as a term for a distinctive and successful approach to the analysis of interactions. CA and ethnomethodology are sometimes considered one field and referred to as EMCA.
Wylie transliteration is a method for transliterating Tibetan script using only the letters available on a typical English-language typewriter. The system is named for the American scholar Turrell V. Wylie, who created the system and published it in a 1959 Harvard Journal of Asiatic Studies article. It has subsequently become a standard transliteration scheme in Tibetan studies, especially in the United States.
Marshallese, also known as Ebon, is a Micronesian language spoken in the Marshall Islands. The language of the Marshallese people, it is spoken by nearly all of the country's population of 59,000, making it the principal language. There are also roughly 27,000 Marshallese citizens residing in the United States, nearly all of whom speak Marshallese, as well as residents in other countries such as Nauru and Kiribati.
Americanist phonetic notation, also known as the North American Phonetic Alphabet (NAPA), the Americanist Phonetic Alphabet or the American Phonetic Alphabet (APA), is a system of phonetic notation originally developed by European and American anthropologists and language scientists for the phonetic and phonemic transcription of indigenous languages of the Americas and for languages of Europe. It is still commonly used by linguists working on, among others, Slavic, Uralic, Semitic languages and for the languages of the Caucasus, of India, and of much of Africa; however, Uralicists commonly use a variant known as the Uralic Phonetic Alphabet.
A pronunciation respelling for English is a notation used to convey the pronunciation of words in the English language, which do not have a phonemic orthography.
The British National Corpus (BNC) is a 100-million-word text corpus of samples of written and spoken English from a wide range of sources. The corpus covers British English of the late 20th century from a wide variety of genres, with the intention that it be a representative sample of spoken and written British English of that time. It is used in corpus linguistics for analysis of corpora.
Gail Jefferson was an American sociologist with an emphasis in sociolinguistics. She was, along with Harvey Sacks and Emanuel Schegloff, one of the founders of the area of research known as conversation analysis (CA). She is remembered for the methods and notational conventions she developed for transcribing speech, the latter forming the Jefferson Transcription System. This is now used widely in CA research.
A speech corpus is a database of speech audio files and text transcriptions. In speech technology, speech corpora are used, among other things, to create acoustic models. In linguistics, spoken corpora are used to do research into phonetic, conversation analysis, dialectology and other fields.
In linguistics, a prosodic unit is a segment of speech that occurs with specific prosodic properties. These properties can be those of stress, intonation, or tonal patterns.
Interactional linguistics (IL) is an interdisciplinary approach to grammar and interaction in the field of linguistics, that applies the methods of Conversation Analysis to the study of linguistic structures, including syntax, phonetics, morphology, and so on. Interactional linguistics is based on the principle that linguistic structures and uses are formed through interaction and it aims at understanding how languages are shaped through interaction. The approach focuses on temporality, activity implication and embodiment in interaction. Interactional linguistics asks research questions such as "How are linguistic patterns shaped by interaction?" and "How do linguistic patterns themselves shape interaction?".
A writing system comprises a set of symbols, called a script, as well as the rules by which the script represents a particular language. The earliest writing was invented during the late 4th millennium BC. Throughout history, each writing system invented without prior knowledge of writing gradually evolved from a system of proto-writing that included a small number of ideographs, which were not fully capable of encoding spoken language, and lacked the ability to express a broad range of ideas.
The former State Administration of Surveying and Mapping, Geographical Names Committee and former Script Reform Committee of the People's Republic of China have adopted several romanizations for Chinese, Mongolian, Tibetan and Uyghur, officially known as pinyin, Regulation of Phonetic Transcription in Hanyu Pinyin Letters of Place Names in Minority Nationality Languages and Orthography of Chinese Personal Name in Hanyu Pinyin Letters. These systems may be referred to as SASM/GNC/SRC transcriptions or SASM/GNC romanizations.
{{cite web}}
: CS1 maint: unfit URL (link)