1. Method 1: Application of an ordered sequence of language-specific rules (e.g. Chomsky and Halle 1968)
Main stress rule:
(1) V -> [1 stress] / X - C0]NAV
Alternating stress rule:
(2) V -> [1 stress] / - C0 V C0 V1 C0]
Lexical entry: hurricAn => hurricA1n [Main stress rule] => hu1rricA1n [Alternating stress rule] => hu1rricA2n [Stress subordination]
Advantages: explicitness; it works.
Disadvantages: a) the rules are language-specific; b) the rule ordering
is language-specific; c) there are many possible rule orders for the learner
to consider; learning is hard. d) Also,
parsing - working out the
prosody and recovering the underlying lexical form from a surface form
- is hard (Bear 1990), but not
that hard (Maxwell 1994).
2. Method 2: Parameters and principles (e.g. Halle and Vergnaud 1987, Dresher and Kaye 1990)
a) Principles (e.g. "feet are made of syllables") are true for
all
(or almost all) languages; innate; no learning is necessary.
b) Parameters are like paired rules, of which the learner must select
one option. Again, parameters are held to be innately specified, and so
the learner has less work to do than learning a rule system.
c) However, the way in which parameters and principles are applied
in generating words still employs an explicit ordering. For example, assignment
of stress to syllables assumes that syllabification is performed first;
computation of quantity-sensitive stress requires that the affiliation
of consonants to one syllable or another has been determined.
2.1. Syllabification (Roca and Johnson ch. 10)
Universal bottom-up syllabification algorithm (after Kahn 1976):
1) Project syllabics (vowels, including long vowels and diphthongs,
and syllabic consonants) to nuclei, as far as the language permits;
2) Parse segments preceding the nucleus into the onset, as far as the
language permits;
3) Parse unsyllabified segments following the nucleus into the coda,
as far as the language permits;
N | N | O | N | O | N | O | N | C | O | N | C | |||||||||||||||||||
| | | | | | | | /\ | | | | | | | | | /\ | | | | | |||||||||||||||||||
e.g. /pamflIt/ | =>1 | p | a | m | f | l | I | t | =>2 | p | a | m | f | l | I | t | =>3 | p | a | m | f | l | I | t |
4) Join nucleus+coda sequences into rimes;
5) Project any remaining nuclei into rimes;
6) Join onset+rime into syllables.
![]() |
![]() |
|||||||||||||||||||||||||||
/\ | /\ | |||||||||||||||||||||||||||
O | R | O | R | O | R | O | R | O | R | O | R | |||||||||||||||||
| | /\ | /\ | /\ | | | /\ | /\ | /\ | | | /\ | /\ | /\ | |||||||||||||||||
=>4 | p | N | C | f | l | N | C | =>5 | p | N | C | f | l | N | C | =>6 | p | N | C | f | l | N | C | |||||
| | | | | | | | | | | | | | | | | | | | | | | | |||||||||||||||||
a | m | I | t | a | m | I | t | a | m | I | t |
p. 360 (24) English stress algorithm (no. 5).
0) Mark each syllable with an asterisk - assumes that syllabification
rules have already applied.
1) ENDSTRESS[Right/Left]:
Project the rightmost/leftmost asterisk.
2) Extrametricality[Right/Left]: Make the rightmost/leftmost asterisk
extrametrical.
Flavours of (2):
English extrametricality: "... in nouns and suffixed adjectives."
(10b) Compound Extrametricality [Right/Left]
(13a) French Extrametricality: "schwa" on the right edge (if there
is one)
p. 335: Footing is mentioned, but not defined.
p. 343: Line Conflation: delete line 1. (Uh oh ... no more feet!)
* | * | * | * | * | * | * | <*> | ||||||
[sIg | nI | fI | k+![]() |
=>Projection | [sIg | nI | fI | k+![]() |
=>Extrametricality | [sIg | nI | fI | k+![]() |
* | * | * | * | * | * | ||||||||
[t![]() |
pan | zii]N | =>Projection | [t![]() |
pan | zii]N | =>Extrametricality | [t![]() |
pan | zii]N |
* | ||||||||||||||
* | (*) | (* | .) | (*) | (* | .) | ||||||||
* | * | * | <*> | * | * | * | <*> | * | * | * | <*> | |||
=>Accent | [sIg | nI | fI | k+![]() |
=>Footing | [sIg | nI | fI | k+![]() |
=>EndStress[Right] | [sIg | nI | fI | k+![]() |
* | ||||||||||||||
* | * | * | (*) | (*) | (*) | (*) | (*) | (*) | ||||||
* | * | * | * | * | * | * | * | * | ||||||
=>Accent | [t![]() |
pan | zii]N | =>Footing | [t![]() |
pan | zii]N | =>EndStress[Right] | [t![]() |
pan | zii]N |
* | ||||
* | * | * | <*> | |
=>Conflation | [sIg | nI | fI | k+![]() |
* | ||||
* | * | * | ||
=>Conflation | [t![]() |
pan | zii]N |
The principle weakness of the bottom-up method is that it employs multiple passes through the word. Also, it requires that whole word to be present (it cannot work as it goes along from start to end of the word), making it psychologically implausible. Also, although bottom-up processing is intuitively appealing, it does not deal well with evidence of top-down processing, such as "tip-of-the-tongue" states.
3. Method 3: constraint filtering in Optimality Theory
GEN: a (random? exhaustive?) generator of possible
structures for lexical strings.
CON: a ranked/ordered set of constraints
EVAL: a mechanism for determining which of many
possible structures is the correct, preferred, or least objectionable analysis.
3.1. Syllabification (Roca and Johnson ch. 19)
Constraints:
ONS | syllables must have onsets | |
NOCODA | syllables must not have codas | |
*COMPLEX | clusters/diphthongs are prohibited | |
PARSE | lexical segments must be parsed into syllable positions | |
FILLNuc | nucleus positions must be filled (with a V) | |
FILLOns | onset positions must be filled (with a C) |
The most optimal satisfaction of these constraints is to parse all syllables as just CV. Obviously, not all languages do this. If they do not have onsets in every syllable, it may be because faithfulness or alignment constraints override ONS. If they allow codas or clusters, it may be for similar reasons, or because PARSE is ranked higher than NOCODA or *COMPLEX.
Example: syllabification of /pamflIt/
/pamflIt/ | ONS | FILLNuc | NOCODA | *COMPLEX |
pa.mflIt | * | ** | ||
pam.flIt | ** | * | ||
pamf.lIt | ** | * | ||
pamfl.It | * | ** | *** | |
pa.mf.lI.t | ** |
In this example, it remains unclear why /pam.flIt/ is preferred to /pamf.lIt/, since both syllables have a coda and one cluster.
3.2. A result: Infixation
Tagalog um+abot => umabot; um+tawag => tumawag; um+gradwet =>grumadwet.
Optimality theoretic analysis: the constraint requiring the prefix um to be aligned with the left edge of the word is weaker than the NOCODA constraint. Note that with a vowelinitial stem such as abot, the /m/ of um can be parsed as an onset.
3.3. Stress assignment
In many languages, feet are binary (FTBIN)
and all syllables are parsed into feet, if possible (PARSE).
In words with an even number of syllables these two constraints are easily
simultaneously satisfied. In words with an odd number of syllables, however,
they cannot both be satisfied: either a) at least one syllable will not
be parsed into a foot, i.e. will be extrametrical, or b) one foot must
have an extra syllable, or c) there may be a onesyllable foot.
a) Pintupi (mála)(wàna), (púli)(kàla)tju;
Warao (yàpu)(rùki)(tàne)(háse), e(nàho)(ròa)(hàku)(tái).
c) Maranungku (lángka)(ràte)(tì)
In Optimality Theory it is said that for a) FTBIN
has priority over PARSE (FTBIN
» PARSE
), whereas for
c), PARSE
» FTBIN.
3.4. Alignment
In Pintupi, the extra syllable is wordfinal; in Warao, wordinitial.
Garawa is similar to Pintupi, except that the extra syllable follows the
primary foot: e.g. (káma)la(rà
i),
(
ánki)r
i(kìrim)(pàyi).
An Optimality-Theoretic analysis of Garawa holds that:
a) The left edge of every word is aligned with the left edge of a foot: ALIGN(PRWDL, FTL).
b) The right edge of every foot is aligned with the right edge of the word: ALIGN(FTR, PRWDR).
(b) will be violated by every foot except the last, but nonfinal feet will be even further away from the end of the word if the extra syllable is anything other than leftmost. Thus, the extra syllable will tend to fall at or near to the beginning of the word. Specifically, if (b) is prioritized over (a), the extra syllable will fall in absolute wordinitial position, as in Warao. If (a) is prioritized over (b), the extra syllable will fall as close as possible to the wordinitial position, but there will be a foot in absolute wordinitial position. This is the Garawa pattern. (See McCarthy and Prince 1993 and Kenstowicz 1995 for other interesting examples.)
3.5. Example (Roca and Johnson 605-6)
"FT-TYPETROCHEE"
is shorthand for ALIGN(sL,
FTL).
"ALIGN-RIGHT" is shorthand
for ALIGN(FTR, PRWDR)
![]() |
|
|
|
|
|
|
|
|
||||||
|
||||||||||||||
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
![]() |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
||||||||||||||
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|||||||||||
|
|
|
|
|
![]() |
|
||||||||
|
||||||||||||||
|
|
|
||||||||||||
|
|
|
|
|||||||||||
|
|
|
|
|
![]() |
|
||||||||
|
||||||||||||||
|
|
|
||||||||||||
|
|
|
|
|||||||||||
|
|
|
|
|
![]() |
|
Strengths: strong universalism.
Weaknesses:
i) Inexplicitness: what's in GEN? If generation of structures is random
(as Roca and Johnson p. 594 states), how do we ensure that the right
analysis is in the candidate set? If generation is exhaustive, the
mechanism becomes psychologically implausible. (But no more than most other
phonological theories.)
ii) Is it plausible that absolutely all constraints are universal?
Optimality Theory has no interesting way to deal with some kinds
of languagespecific constraints. For examples, according to Dell and
Elmedlaoui's analysis of Tashlhit Berber, every vowel and consonant is
a possible nucleus, but /a/ is always a nucleus (FILLNuca).
To guarantee that this constraint is exceptionless, it must be ranked higher
than all the other syllable parsing constraints: FILLNuca
»
PARSE » ONS » FILL
» FILLNuc
»
NOCODA. This is a languagespecific
fact about Tashlhit Berber. There is a more general fact, however: FILLNuca
is
more specific than FILLNuc,
so naturally it takes priority; otherwise it would not do any work in the
grammar! This is a familiar instance of the Elsewhere Condition, a.k.a.
subsumption.
4. Method 4: phrase structure parsing in Declarative Phonology (e.g. Church 1983, Coleman 1990, 2000)
To define a set of legal structures, a grammar is needed. To define
hierarchical structures, context-free phrase structure grammar is a good
starting point.
Syllable structure rules | Metrical structure rules | |||
1) | ![]() |
12) | PrWdLat -> (![]() |
|
2) | ![]() |
13) | PrWdLat -> (![]() ![]() |
|
3) | RH -> N Co | 14) | PrWdGer -> Fs ![]() |
|
4) | RH -> NH | 15) | F -> ![]() ![]() |
|
5) | RL -> NL | 16) | F -> ![]() |
|
6) | NL -> V | |||
7) | NH -> V V | |||
8) | O -> /s/ (C) (G) | |||
9) | O -> C | |||
10) | Co -> C | |||
11) | Co -> C C |
Parsing: there are several proven algorithms for parsing any context-free grammar (see e.g. Jurafsky and Martin 2000, ch. 10). One of the most common approaches is a left-to-right, top-down, nondeterministic method that makes an interesting contrast to the usual bottom-up method found in phonology textbooks.
e.g. parsing /dIvelp/Lat
by generation:
PrWdLat =>12
(L)
(Fw) Fs
(C[+em])
=>2 (O RL)
(Fw) Fs
(C[+em])
=> d I (Fw) Fs
(C[+em])
=>15 d I
L
(C[+em]) ...ignoring Fs
=>2 d IO
RL
L
(C[+em])
=>5 d Iv
NL
L
(C[+em])
=>6 d Iv
V
L
(C[+em])
=> d Iv
e
L
(C[+em])
=>2 d Iv
e O RL (C[+em])
=> d Iv
e l RL (C[+em])
=>5 d Iv
e l NL (C[+em])
=>6 d Iv
e l V (C[+em])
=> d Iv
e l
(C[+em])
=> d Iv
e l
p
In Declarative Phonology, the metrical grammars of different languages
are just different, though there are certain similarities. For examples
(other analyses are possible):
Pintupi: | PrWd -> Fs Fw*
(![]() |
F -> ![]() ![]() |
||||
Warao: | PrWd -> (![]() |
F -> ![]() ![]() |
||||
Maranungku: | PrWd -> Fs Fw*
(![]() |
F -> ![]() ![]() |
||||
Garawa: | PrWd -> Fs Fw* | Fs -> ![]() ![]() ![]() |
Fw -> ![]() ![]() |
Since context-free grammars are nondeterministic, sometimes there is more than one possible analysis of a string: ambiguity. If only one analysis is required, many strategies may be used to pick one:
1) Employ the rules in a particular order: e.g. big onsets first
(cf. 8, 9 above); empty/small coda rules first (10, 11; rules 3 and 4 would
need to be reversed).
2) Add a trans-syllabic constraint preventing filled codas with
empty onsets.
3) Allow/accept variation!
4) Calculate the probabilities of rules and derivations, and pick the
most
likely analysis (Coleman 2000).
References
Bear, J. (1990) Backwards phonology. In Karlgren (1990), Vol. 3, 13-20.
Church, K. W. (1983) Phrase-Structure Parsing: a method for taking advantage of allophonic constraints. MIT PhD thesis, distributed by Indiana University Linguistics Club.
Coleman, J. S. (1990) Unification Phonology: Another look at "synthesis-by-rule". In Karlgren (1990) Vol. 2, 79-84.
Coleman, J. S. (2000) Candidate selection. The Linguistic Review 17, 167-179.
Dresher, B. E. and J. Kaye (1990) A computational learning model for metrical phonology. Cognition 34, 137-195.
Halle, M. and J.-R. Vergnaud (1987) An Essay on Stress. MIT Press.
Jurafsky, D. and J. H. Martin (2000) Speech and Language Processing: an introduction to natural language processing, computational linguistics, and speech recognition. Upper Saddle River, NJ: Prentice Hall.
Kahn, D. (1976) Syllable-based generalizations in English phonology. MIT PhD dissertation, distributed by Indiana University Linguistics Club.
Karlgren, H. (ed.) (1990) COLING-90. Papers Presented to the Thirteenth International Conference on Computational Linguistics. Helsinki: University of Helsinki.
Kenstowicz, M. (1995) Cyclic vs. noncyclic constraint evaluation. Phonology 12. 397-436.
Maxwell, M. (1994) Parsing using linearly ordered phonological rules. Computational Phonology: First Meeting of the ACL Special Interest Group in Computational Phonology. Proceedings of the Workshop. Bernardsville, NJ: Association for Computational Linguistics. 59-70.
McCarthy, J. and A. Prince (1993) Generalized alignment. Yearbook of Morphology 1993. 79-153.