Phonetic transcription

Last updated

Phonetic transcription (also known as phonetic script or phonetic notation) is the visual representation of speech sounds (or phones) by means of symbols. The most common type of phonetic transcription uses a phonetic alphabet, such as the International Phonetic Alphabet.


Versus orthography

The pronunciation of words in all languages changes over time. [1] However, their written forms (orthography) are often not modified to take account of such changes, and do not accurately represent the pronunciation. Pronunciation can also vary greatly among dialects of a language. Standard orthography in some languages, such as English and Tibetan, is often irregular and makes it difficult to predict pronunciation from spelling. For example, the words bough and through do not rhyme in English even though their spellings might suggest otherwise. Other languages, such as Spanish and Italian have a more consistent (but still imperfect) relationship between orthography and pronunciation, while a few languages may claim to have a fully phonemic spelling system (a phonemic orthography).

For most languages, phonetic transcription makes it possible to show pronunciation with something much nearer to a one-to-one relationship between sound and symbol than is possible with the language's orthography. Phonetic transcription allows one to step outside orthography, examine differences in pronunciation between dialects within a given language and identify changes in pronunciation that may take place over time.

A basic principle of phonetic transcription is that it should be applicable to all languages, and its symbols should denote the same phonetic properties whatever the language being transcribed. [2] It follows that a transcription devised for one individual language or group of languages is not a phonetic transcription but an orthography.

Narrow versus broad transcription

Phonetic transcription may be used to transcribe the phonemes of a language, or it may go further and specify their precise phonetic realization. In all systems of transcription there is a distinction between broad transcription and narrow transcription. Broad transcription indicates only the most noticeable phonetic features of an utterance, whereas narrow transcription encodes more information about the phonetic characteristics of the allophones in the utterance. The difference between broad and narrow is a continuum, but the difference between phonemic and phonetic transcription is usually treated as a binary distinction. [3] Phonemic transcription is a particular form of broad transcription which disregards all allophonic difference; as the name implies, it is not really a phonetic transcription at all (though at times it may coincide with one), but a representation of phonemic structure. A transcription which includes some allophonic detail but is closely linked to the phonemic structure of an utterance is called an allophonic transcription.

The advantage of the narrow transcription is that it can help learners to produce exactly the right sound, and allows linguists to make detailed analyses of language variation. [4] The disadvantage is that a narrow transcription is rarely representative of all speakers of a language. While most Americans, Canadians and Australians would pronounce the /t/ of little as a tap [ ɾ ], many speakers in southern England would pronounce /t/ as [ ʔ ] (a glottal stop; t-glottalization) and/or the second /l/ as a vowel resembling [ ʊ ] (L-vocalization), possibly yielding [ˈlɪʔʊ].

A further disadvantage of narrow transcription is that it involves a larger number of symbols and diacritics that may be unfamiliar to non-specialists. The advantage of broad transcription is that it usually allows statements to be made which apply across a more diverse language community. It is thus more appropriate for the pronunciation data in foreign language dictionaries, which may discuss phonetic details in the preface but rarely give them for each entry. A rule of thumb in many linguistics contexts is therefore to use a narrow transcription when it is necessary for the point being made, but a broad transcription whenever possible.

Types of notational systems

Most phonetic transcription is based on the assumption that linguistic sounds are segmentable into discrete units that can be represented by symbols. Many different types of transcription, or "notation", have been tried out: these may be divided into Alphabetic (which are based on the same principle as that which governs ordinary alphabetic writing, namely that of using one single simple symbol to represent each sound), and Analphabetic (notations which are not alphabetic) which represent each sound by a composite symbol made up of a number of signs put together. [5]


IPA IPA chart 2020.svg

The International Phonetic Alphabet (IPA) is the most widely used and well-known of present-day phonetic alphabets, and has a long history. It was created in the nineteenth century by European language teachers and linguists. It soon developed beyond its original purpose as a tool of foreign language pedagogy and is now also used extensively as a practical alphabet of phoneticians and linguists. It is found in many dictionaries, where it is used to indicate the pronunciation of words, but most American dictionaries for native English-speakers, e.g., American Heritage Dictionary of the English Language, Random House Dictionary of the English Language, Webster's Third New International Dictionary , avoid phonetic transcription and instead employ respelling systems based on the English alphabet, with diacritical marks over the vowels and stress marks. [6] (See Pronunciation respelling for English for a generic version.)

Another commonly encountered alphabetic tradition was originally created by American linguists for the transcription of Native American and European languages, and is still commonly used [ citation needed ] by linguists of Slavic, Indic, Semitic, Uralic (here known as the Uralic Phonetic Alphabet) and Caucasian languages. This is often labeled the Americanist phonetic alphabet despite having been widely used for languages outside the Americas. The principal difference between these alphabets and the IPA is that the specially created characters of the IPA are abandoned in favour of already existing typewriter characters with diacritics (e.g. many characters are borrowed from Eastern European orthographies) or digraphs. Examples of this transcription may be seen in Pike's Phonemics [7] and in many of the papers reprinted in Joos's Readings in Linguistics 1. [8] In the days before it was possible to create phonetic fonts for computer printers and computerized typesetting, this system allowed material to be typed on existing typewriters to create printable material.

There are also extended versions of the IPA, for example: Ext-IPA, VoQS, and Luciano Canepari's canIPA.

Aspects of alphabetic transcription

The International Phonetic Association recommends that a phonetic transcription should be enclosed in square brackets "[ ]". A transcription that specifically denotes only phonological contrasts may be enclosed in slashes "/ /" instead. If one is unsure, it is best to use brackets since by setting off a transcription with slashes, one makes a theoretical claim that every symbol phonemically contrasts for the language being transcribed.

For phonetic transcriptions, there is flexibility in how closely sounds may be transcribed. A transcription that gives only a basic idea of the sounds of a language in the broadest terms is called a broad transcription; in some cases, it may be equivalent to a phonemic transcription (only without any theoretical claims). A close transcription, indicating precise details of the sounds, is called a narrow transcription. They are not binary choices but the ends of a continuum, with many possibilities in between. All are enclosed in brackets.

For example, in some dialects the English word pretzel in a narrow transcription would be [ˈpɹ̥ʷɛʔts.ɫ̩], which notes several phonetic features that may not be evident even to a native speaker. An example of a broad transcription is [ˈpɹ̥ɛts.ɫ̩], which indicates only some of the features that are easier to hear. A yet broader transcription would be [ˈpɹɛts.l] in which every symbol represents an unambiguous speech sound but without going into any unnecessary detail. None of those transcriptions makes any claims about the phonemic status of the sounds. Instead, they represent certain ways in which it is possible to produce the sounds that make up the word. [9]

There are also several possibilities in how to transcribe the word phonemically, but here, the differences are generally of not precision but analysis. For example, pretzel could be /ˈprɛts.l̩/ or /ˈprɛts.əl/. The latter transcription suggests that there are two vowels in the word even if they cannot both be heard, but the former suggests that there is only one. [10]

Strictly speaking, it is not possible to have a distinction between "broad" and "narrow" within phonemic transcription, since the symbols chosen represent only sounds that have been shown to be distinctive. However, the symbols themselves may be more or less explicit about their phonetic realization. [11] A frequently cited example is the symbol chosen for the English consonant at the beginning of the words 'rue', 'rye', 'red': this is frequently transcribed as /r/, despite the symbol suggesting an association with the IPA symbol [r] which is used for a tongue-tip trill. It is equally possible within a phonemic transcription to use the symbol /ɹ/, which in IPA usage refers to an alveolar approximant; this is the more common realization for English pronunciation in America and England. Phonemic symbols will frequently be chosen to avoid diacritics as much as possible, under a 'one sound one symbol' policy, or may even be restricted to the ASCII symbols of a typical keyboard, as in the SAMPA alphabet. For example, the English word church may be transcribed as /tʃɝːtʃ/, a close approximation of its actual pronunciation, or more abstractly as /crc/, which is easier to type. Phonemic symbols should always be backed up by an explanation of their use and meaning, especially when they are as divergent from actual pronunciation as /crc/. [12]

Occasionally a transcription will be enclosed in pipes ("| |"). This goes beyond phonology into morphological analysis. For example, the words pets and beds could be transcribed phonetically as [pʰɛʔts] and [b̥ɛd̥z̥] (in a fairly narrow transcription), and phonemically as /pɛts/ and /bɛdz/. Because /s/ and /z/ are separate phonemes in English, they receive separate symbols in the phonemic analysis. However, a native English speaker would recognize that underneath this, they represent the same plural ending. This can be indicated with the pipe notation. If the plural ending is thought to be essentially an s, as English spelling would suggest, the words can be transcribed |pɛts| and |bɛds|. If it is essentially a z, these would be |pɛtz| and |bɛdz|.

To avoid confusion with IPA symbols, it may be desirable to specify when native orthography is being used, so that, for example, the English word jet is not read as "yet". This is done with angle brackets or chevrons: jet. It is also common to italicize such words, but the chevrons indicate specifically that they are in the original language's orthography, and not in English transliteration.


Visible Speech VisibleSpeech-illustrations.jpg
Visible Speech

In iconic phonetic notation, the shapes of the phonetic characters are designed so that they visually represent the position of articulators in the vocal tract. This is unlike alphabetic notation, where the correspondence between character shape and articulator position is arbitrary. This notation is potentially more flexible than alphabetic notation in showing more shades of pronunciation (MacMahon 1996:838–841). An example of iconic phonetic notation is the Visible Speech system, created by Scottish phonetician Alexander Melville Bell (Ellis 1869:15).


Another type of phonetic notation that is more precise than alphabetic notation is analphabetic phonetic notation. Instead of both the alphabetic and iconic notational types' general principle of using one symbol per sound, analphabetic notation uses long sequences of symbols to precisely describe the component features of an articulatory gesture (MacMahon 1996:842–844). This type of notation is reminiscent of the notation used in chemical formulas to denote the composition of chemical compounds. Although more descriptive than alphabetic notation, analphabetic notation is less practical for many purposes (e.g. for descriptive linguists doing fieldwork or for speech pathologists impressionistically transcribing speech disorders). As a result, this type of notation is uncommon.

Two examples of this type were developed by the Danish Otto Jespersen (1889) and American Kenneth Pike (1943). Pike's system, which is part of a larger goal of scientific description of phonetics, is particularly interesting in its challenge against the descriptive method of the phoneticians who created alphabetic systems like the IPA. An example of Pike's system can be demonstrated by the following. A syllabic voiced alveolar nasal consonant (/n̩/ in IPA) is notated as


In Pike's notation there are 5 main components (which are indicated using the example above):

  1. M – manner of production (i.e., MaIlDe)
  2. C – manner of controlling (i.e., CVoeIpvnn)
  3. description of stricture (i.e., APpaatdtltnransnsfSpvavdtlvtnransssfTpgagdtlwvtitvransnsf)
  4. S – segment type (i.e., Srp)
  5. F – phonetic function (i.e., FSs)

The components of the notational hierarchy of this consonant are explained below:

M = productive mechanism
a = air-stream mechanism
I = initiator
l = for lung air
D = direction of the air stream
e = egressive
C = controlling mechanism
V = valvate stricture
o = oral stricture
e = subvalvate esophageal stricture
I = degree of air-stream interruption
p = partial (continuants)
v = nonfrictional
n = nasal
n = resonant nasal
(Rank of stricture)
A = acme
P = primary
(Features of stricture)
p = point of articulation
a = alveolar
a = articulator
t = tongue tip
d = degree of articulation
t = in time
l = long
t = type of articulation
n = normal
r = relative strength
a = of articulating movement
n = normal
s = of acoustic impression
n = normal
s = shape of articulator
f = flat
(Rank of stricture)
S = secondary
(Features of stricture)
p = point of articulation
v = velic
a = articulator
v = velic
d = degree of articulation
t = in time
l = long
v = with cavity friction
t = type of articulation
n = normal
r = relative strength
a = of articulating movement
n = normal
s = of acoustic impression
s = soft
s = shape of articulator
f = flat
(Rank of stricture)
T = tertiary
(Features of stricture)
p = point of articulation
g = glottal
a = articulator
g = vocal folds
d = degree of articulation
t = in time
l = long
w = wide
v = with cavity friction
t = type of articulation
i = iterative
t = trill
v = vibratory trill
r = relative strength
a = of articulating movement
n = normal
s = of acoustic impression
n = normal
s = shape of articulator
f = flat
S = segmental type
r = real
p = perceptual
F = function phonetically
S = of the segment in the syllable
s = syllabic contoid

See also

Notational systems


Related Research Articles

International Phonetic Alphabet Alphabetic system of phonetic notation

The International Phonetic Alphabet (IPA) is an alphabetic system of phonetic notation based primarily on the Latin script. It was devised by the International Phonetic Association in the late 19th century as a standardized representation of speech sounds in written form. The IPA is used by lexicographers, foreign language students and teachers, linguists, speech-language pathologists, singers, actors, constructed language creators and translators.

In phonology and linguistics, a phoneme is a unit of sound that distinguishes one word from another in a particular language.

In phonetics and linguistics, a phone is any distinct speech sound or gesture, regardless of whether the exact sound is critical to the meanings of words.

The Speech Assessment Methods Phonetic Alphabet (SAMPA) is a computer-readable phonetic script using 7-bit printable ASCII characters, based on the International Phonetic Alphabet (IPA).

Phonetic alphabet can mean:

Transcription in the linguistic sense is the systematic representation of spoken language in written form. The source can either be utterances or preexisting text in another writing system.

Glottal stop Sound made by stopping airflow in the glottis

The glottal plosive or stop is a type of consonantal sound used in many spoken languages, produced by obstructing airflow in the vocal tract or, more precisely, the glottis. The symbol in the International Phonetic Alphabet that represents this sound is ⟨ʔ⟩.

A phonemic orthography is an orthography in which the graphemes correspond to the phonemes of the language. Natural languages rarely have perfectly phonemic orthographies; a high degree of grapheme-phoneme correspondence can be expected in orthographies based on alphabetic writing systems, but they differ in how complete this correspondence is. English orthography, for example, is alphabetic but highly nonphonemic; it was once mostly phonemic during the Middle English stage, when the modern spellings originated, but spoken English changed rapidly while the orthography was much more stable, resulting in the modern nonphonemic situation. However, because of their relatively recent modernizations compared to English, the Romanian, Italian, Turkish, Spanish, Finnish, Czech, Latvian and Polish orthographic systems come much closer to being consistent phonemic representations.

Kirshenbaum, sometimes called ASCII-IPA or erkIPA, is a system used to represent the International Phonetic Alphabet (IPA) in ASCII. This way it allows typewriting IPA-symbols by regular keyboard. It was developed for Usenet, notably the newsgroups sci.lang and alt.usage.english. It is named after Evan Kirshenbaum, who led the collaboration that created it. The eSpeak open source software speech synthesizer uses the Kirshenbaum scheme.

The Extended Speech Assessment Methods Phonetic Alphabet (X-SAMPA) is a variant of SAMPA developed in 1995 by John C. Wells, professor of phonetics at University College London. It is designed to unify the individual language SAMPA alphabets, and extend SAMPA to cover the entire range of characters in the 1993 version of International Phonetic Alphabet (IPA). The result is a SAMPA-inspired remapping of the IPA into 7-bit ASCII.

Voiceless palatal fricative

The voiceless palatal fricative is a type of consonantal sound used in some spoken languages. The symbol in the International Phonetic Alphabet that represents this sound is ⟨ç⟩, and the equivalent X-SAMPA symbol is C. It is the non-sibilant equivalent of the voiceless alveolo-palatal fricative.

In phonetics, a trill is a consonantal sound produced by vibrations between the active articulator and passive articulator. Standard Spanish ⟨rr⟩ as in perro, for example, is an alveolar trill.

In phonetics, a flap or tap is a type of consonantal sound, which is produced with a single contraction of the muscles so that one articulator is thrown against another.

Americanist phonetic notation, also known as the North American Phonetic Alphabet (NAPA), the Americanist Phonetic Alphabet or the American Phonetic Alphabet (APA), is a system of phonetic notation originally developed by European and American anthropologists and language scientists for the phonetic and phonemic transcription of indigenous languages of the Americas and for languages of Europe. It is still commonly used by linguists working on, among others, Slavic, Uralic, Semitic languages and for the languages of the Caucasus and of India; however, Uralists commonly use a variant known as the Uralic Phonetic Alphabet. Despite its name, the term "Americanist phonetic alphabet" has always been widely used outside the Americas. For example, a version of it is the standard for the transcription of Arabic in articles published in the Zeitschrift der Deutschen Morgenländischen Gesellschaft, the journal of the German Oriental Society.

The Uralic Phonetic Alphabet (UPA) or Finno-Ugric transcription system is a phonetic transcription or notational system used predominantly for the transcription and reconstruction of Uralic languages. It was first published in 1901 by Eemil Nestor Setälä, a Finnish linguist.

Extensions to the International Phonetic Alphabet Set of symbols

The extensions to the International Phonetic Alphabet, also extIPA symbols for disordered speech or simply extIPA, are a set of letters and diacritics devised by the International Clinical Phonetics and Linguistics Association to augment the International Phonetic Alphabet for the phonetic transcription of disordered speech. Some of the symbols are used for transcribing features of normal speech in IPA transcription, and are accepted as such by the International Phonetic Association.

A pronunciation respelling for English is a notation used to convey the pronunciation of words in the English language, which does not have a phonemic orthography.

Orthographic transcription is a transcription method that employs the standard spelling system of each target language.

Lilias Armstrong British phonetician (1882-1937)

Lilias Eveline Armstrong was an English phonetician. She worked at University College London, where she attained the rank of reader. Armstrong is most known for her work on English intonation as well as the phonetics and tone of Somali and Kikuyu. Her book on English intonation, written with Ida C. Ward, was in print for 50 years. Armstrong also provided some of the first detailed descriptions of tone in Somali and Kikuyu.

John Henry Esling, is a Canadian linguist specializing in phonetics. He is a Professor Emeritus of Linguistics at the University of Victoria, where he taught from 1981 to 2014. Esling was President of the International Phonetic Association from 2011 to 2015 and a co-editor of the 1999 Handbook of the International Phonetic Association.


  1. Shariatmadari, David (2019). Don't Believe a Word. Weidenfeld & Nicolson. pp. 21–40. ISBN   978-1-4746-0843-5.
  2. Crystal, David (1997). The Cambridge Encyclopedia of Language (2nd ed.). Cambridge University Press. p. 160.
  3. Laver, John (1994). Principles of Phonetics. Cambridge. p. 550. ISBN   0-521-45655-X.
  4. Ball, Martin; Rahilly, Joan (1999). Phonetics: the Science of Speech. Arnold. pp. 142–3. ISBN   0-340-70010-6.
  5. Abercrombie, David (1967). Elements of General Phonetics. Edinburgh. pp. 111–2.
  6. Landau, Sidney (2001) Dictionaries: The Art and Craft of Lexicography, 2nd ed., p 118. Cambridge University Press. ISBN   0-521-78512-X.
  7. Pike, Kenneth (1947). Phonemics. University of Michigan.
  8. Joos, M. (ed) (1957). Readings in Linguistics 1. University of Chicago.CS1 maint: extra text: authors list (link)
  9. Abercrombie, David (1967). Elements of General Phonetics. Edinburgh. pp. 128–9.
  10. Roach, Peter (2009). English Phonetics and Phonology (4th ed.). Cambridge University Press. pp. 100–1. ISBN   978-0-521-71740-3.
  11. Jones, Daniel (1967). An Outline of English Phonetics (9th ed.). Heffer. pp. 335–6.
  12. Laver, John (1994). Principles of Phonetics. Cambridge University Press. p. 551.