General phonetics // was "Newbie"
From: | Morgan Palaeo Associates <morganpalaeo@...> |
Date: | Tuesday, January 13, 2004, 10:26 |
Tristan McLeay wrote:
> I think the term was popularised around the time I made the CXS chart.
Looking at the chart it's obvious not everyone agrees on stress
marking :-)
Personally, I think comma for secondary stress is by far the most
logical of the three options. It's very counterintuitive to have a
"heavier" character to mark secondary stress than to mark primary
stress ("heavier" = "takes more ink to draw").
Mind you, I've never understood why a retroflex trill is not shaded
impossible. If it's retroflex then the tip of your tongue points
backwards, and you can't trill like that.
***
Changing topic quite a bit, but still talking about phonetics, I
thought people might like to see how I went about explaining phonetics
in my research assignment last year.
You see, I find the process of how people choose to explain things and
why to be quite fascinating. Therefore I assume that many other people
find it interesting, too (it's definitely one reason why lots of
scientists enjoy reading good popular science books). In this case,
the relevant background is that I, a non-linguist, needed to explain
phonetics and phonology to a lecturer, also a non-linguist, with the
emphasis being on the dimensions within which the data exists (very
important in DMKD).
In practical terms, consider this a review of the fundamentals and an
opportunity to pester me for fun if there's some obscure and pedantic
error :-)
3 Phonetics and Phonology
The applications of data mining to phonetics and phonology cannot
be discussed without a brief overview of those fields. Phonetics
is a study of speech sounds (phones) that emphasises their means
of articulation and corresponding acoustic properties, while
phonology is a more abstract study of speech sounds as they are
utilised by language.
3.1 Phonetics
Consonants and vowels are categorised along different sets of
dimensions, and can be underestood from either an articulatory or
an acoustic perspective. The characters of the International
Phonetic Alphabet correspond to distinct phones, but with limited
resolution. For obvious reasons, there is a tradeoff between the
accuracy and resolution of acoustic analysis and the size of the
database that it is viable to constuct (see later), and since
improvement of these qualities is an ongoing research area, we can
expect the phonetic analysis database of the future to be larger
and more exact than what is possible today, increasing the
applicability of data mining.
3.1.1 Consonants
Consonants generally have much more complex acoustic properties
than vowels because many of them involve the production of
turbulent air flow. The author has little knowledge about acoustic
analysis of consonants, but papers such as de Mareuil (1999)
contain keywords that might aid an interesting reader in learning
more. The articulatory dimensions along which a consonant can
vary have multiple continuous and discrete components, but the
standard IPA transcriptions capture a simplified description that
suffices for most purposes (Catford 2001). For example the
description of 's' as an unvoiced aveolar fricative refers to three
discrete-valued dimensions ('unvoiced' meaning that the vocal
chords are not vibrated, 'aveolar' meaning that the sound is
produced near the ridge behind the upper teeth, and 'fricative'
meaning that turbulence is created by forcing air through a narrow
channel) but this description discretises some dimensions that are
really continuous (e.g. the distance between the tongue and the
teeth) and ignores other dimensions altogether (e.g. the part of
the tongue used to make the sound: the distinction between apico-
lamino- and dorso- articulations).
3.1.2 Vowels
Vowels can be described using either acoustic dimensions or
articulatory ones, corresponding to low-level or high-level
descriptions of the vowel respectively. Again, IPA symbols offer
only a rough approximation, but a higher resolution is attainable
in which continuous-valued dimensions are indeed treated as
continuous.
The articulatory description of a vowel includes two major
dimensions and several minor dimensions (Catford 2001). The major
dimensions concern the horizontal and vertical position of the
tongue, distinguishing between front vowels (as in 'head'), back
vowels (as in 'hoard'), closed vowels (as in 'hoot') and open
vowels (as in 'heart'). A map of these dimensions is often
referred to as the vowel diagram, and is well illustrated with
respect to several dialects in Mannell: Vowels. The most important
of the minor dimensions is the extent to which the lips are rounded
(roundedness), another is the extent to which air passes through
the nose as well as the mouth (nasalisation) and there are more
besides. The continuous nature of the major dimensions is very
important - which is why the vowel diagram exists as a standard
illustration - while the continuous nature of the minor dimensions
is less important, and they are conventionally described in
either/or terms (the lips are either rounded or not; the vowel is
either nasalised or not).
The low-level (acoustic) dimensions of vowels are defined by
formant frequencies, which are the frequencies in the acoustic
spectrum of the sound that are amplified by a particular
configuration of the mputh and throat (Catford 2001). Formants
are numbered in ascending order of frequency, each one being a
continuous-valued dimension. For example the first formant of an
'ee'-like sound is around 250 Hz, while that of an 'ah'-like sound
is around 750 Hz. Likewise the second formants are around 2300 and
1400 Hz respectively. The number of formants that must be recorded
depends depends upon the range of contrasts that may affect the
vowel, for example all else being equal the primary dimensions of a
vowel can be described using only the first two formants but if lip
rounding etc are not constant then two formants are insufficient.
3.1.3 Temporal Dimensions
The dimensions discussed above are defined in abstract articulatory
or acoustic spaces. The temporal dimension is also important, as it
allows us to capture information about the length for which a phone
is sustained, the way that its articulation is influenced by its
neighbours, and the frequency at which different phones appear
together. The importance of such informaiton is well established
and studied. For one thing, languages make direct use of it. The
words 'hut' and 'heart', for example, are distinguised only by the
duration of the 'ah' sound. Diphthongs and affricates are examples
where neighbouring phones are best understood as a unit - in the
'eye' diphthong an 'ah' sound glides toweards an 'ee' sound,, and
in the 'ch' affricate a 't' sound gives way to a 'sh' sound
(Catford 2001). For another thing, people have a tendency to modify
a sound in order to assist the flow of the word, for example the
'h' in 'huge' is objectively much closer to the 'ch' in German
'ich' than the 'h' in 'hunt'. The related phenomenon of
assimilation means that neighbouring phones often acquire the same
point of articulation, for example if there was a word 'sumda' then
it would quite likely mutate to 'sunda' or 'sumba' (and the word
'import' is derived from Latin 'in' + 'port'). For these reasons
the fact that phonetic data exists in a spatiotemporal context is
certainly important.
3.2 Phonology
A phone is a sound in language viewed from a low-level perspective,
while a phoneme is a sound viewed from a higher-level perspective.
The mechanics of spoken languages are not concerned with details of
articulation, but only with the fact that units of sound can be put
together to make words (Catford 2001). For example the difference
between the 'h' in 'huge' and 'hunt' was mentioned above, yet from
the point of view of how English works they are nevertheless the
same unit. For this reason, the articulations of 'h' are said to be
different /allphones/ of the same /phoneme/. Similarly the distinct
'oo' sounds in 'food' and 'fool' are allophones of another phoneme,
the latter being modified by the following 'l'. It should be
apparent that the 'ah' sounds in 'heart' and 'hut' are distinct
phonemes, whereas the 'oo' sounds in 'feud' and 'refute' are the
same phoneme, because English doesn't care that one is longer than
the other in that case (vowels are typically shortened in English
when followed by an unvoiced consnant). Diphthongs and affricates
etc are single phonemes, because they are single units in language.
3.3. Sound Change
It is well established that the dominant pronunciation of words
changes over time, and that this process eventually leads to new
dialects and even languages. A driving force behind this process is
the tendency people have to imitate the speech of people who are
socially important to them (peers and role models). Children
consistently adapt the accent of their peers and not the accent of
their parents. This process explains why pronunciation changes are
sometimes detectable from one generation to the next.
Sound change and similar processes lead eventually to language
diversion, for example Italian and Spanish are both descendents of
Latin with many similarities remaining in their phonology,
vocabulary and grammar. Whether data mining can be used to form
hypotheses about which languages are related to which will be
discussed later.
3.4 Size and Resolution of Phonetic Databases
<snip>
References
----------
J. C. Catford: /A Practical Introduction to Phonetics: Second Edition
(2001). Oxford University Press.
R. Mannell: /The Vowels of Australian English and Other English
Dialects/. Website: www.ling.mq.edu.au/units/ling210-901/phonetics/
ausenglish/auseng_vowels.html (part of linguistics website at
Macquarie University). Accessed 2003.
P. B. de Mareuil, C. Corredor-Ardoy & M. Adda-Decker: /Multi-Lingual
Automated Phoneme Clustering/ (1999). ICPh599 San Francisco.
Search Internet for title to find this third reference.
Adrian.
Replies