wfsts in asr basics of speech productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16...

33
Instructor: Preethi Jyothi WFSTs in ASR & Basics of Speech Production Lecture 6 CS 753

Upload: others

Post on 19-May-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Instructor: Preethi Jyothi

WFSTs in ASR &

Basics of Speech Production

Lecture 6

CS 753

Page 2: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Determinization/Minimization: Recap

• A (W)FST is deterministic if:

• Unique start state• No two transitions from a state share the same input label• No epsilon input labels

• Minimization finds an equivalent deterministic FST with the least number of states (and transitions)

• For a deterministic weighted automaton, weight pushing + (unweighted) automata minimization leads to a minimal weighted automaton

• Guaranteed to yield a deterministic/minimized WFSA under some technical conditions characterising the automata (e.g. twins property) and the weight semiring (allowing for weight pushing)

Page 3: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

0

1b:bad

4c:cab

7b:bead

11

c:cede

15

d:decade

2a:eps

5a:eps

8e:eps

12e:eps

16e:eps

3d:eps

6b:eps

9a:eps 10d:eps

13d:eps 14e:eps

17c:eps18

a:eps19d:eps 20e:eps

Example: Dictionary WFST

Page 4: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

0

1b:eps

2c:eps

3

d:decade

4a:bad

5e:bead

6a:cab

7

e:cede

8

e:eps

9d:eps

10a:eps

11b:eps

12d:eps

13c:eps

14d:eps

15e:eps

16a:eps

17d:eps 18e:eps

Determinized Dictionary WFST

Page 5: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

0

1b:eps

2c:eps

3

d:decade

4a:bad

5e:bead

6a:cab

7

e:cede

8e:eps

9

d:epsa:eps

b:eps

10d:eps

11c:eps

e:eps

a:eps

Minimized Dictionary WFST

Page 6: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

AcousticIndices

WFST-based ASR SystemLanguage

Model Word Sequence

AcousticModels

Triphones

ContextTransducer

Monophones

Pronunciation Model

Words

Page 7: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

AcousticIndices

Language Model Word

Sequence

AcousticModels

Triphones

ContextTransducer

Monophones

Pronunciation Model

Words

H

a/a_b

b/a_b

.

.

.x/y_z

One 3-state HMM for

each triphone

f1:ε

} FST Union + Closure

ResultingFST

H

f2:ε

f3:ε f4:ε f5:ε

f4:ε f6:εf0:a+a+b

WFST-based ASR System

Page 8: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

M. Mohri: Weighted FSTs in Speech Recognition 12

ε,* x,ε

x:x/ ε_ε

x,x

x:x/ ε_x

x,y

x:x/ ε_y

y,ε

y:y/ ε_ε

y,x

y:y/ ε_x

y,y

y:y/ ε_y x:x/x_ε

x:x/x_x

x:x/x_y

y:y/x_ ε

y:y/x_x

y:y/x_y

x:x/y_εx:x/y_x

x:x/y_y

y:y/y_ε

y:y/y_xy:y/y_y

Figure 8: Context-dependent triphone transducer.

3.1. Transducer Combination

Consider the pronunciation lexicon in Figure 2b. Suppose we form the union ofthis transducer with the pronunciation transducers for the remaining words inthe grammar G of Figure 2a and then take its Kleene closure by connecting anϵ-transition from each final state to the initial state. The resulting pronuncia-tion lexicon L would pair any sequence of words from that vocabulary to theircorresponding pronunciations. Thus,

L ◦G

gives a transducer that maps from phones to word sequences restricted to G.We used composition here to implement a context-independent substitution.

However, a major advantage of transducers in speech recognition is that they gen-eralize naturally the notion of context-independent substitution of a label to thecontext-dependent case. The transducer of Figure 8 does not correspond to a sim-ple substitution, since it describes the mapping from context-independent phonesto context-dependent triphonic models, denoted by phone/left context right context.Just two hypothetical phones x and y are shown for simplicity. Each state en-codes the knowledge of the previous and next phones. State labels in the figureare pairs (a, b) of the past a and the future b, with ϵ representing the start or endof a phone sequence and ∗ an unspecified future. For instance, it is easy to see thatthe phone sequence xyx is mapped by the transducer to x/ϵ y y/x x x/y ϵvia the unique state sequence (ϵ, ∗)(x, y)(y, x)(x, ϵ). More generally, when thereare n context-independent phones, this triphonic construction gives a transducerwith O(n2) states and O(n3) transitions. A tetraphonic construction would give atransducer with O(n3) states and O(n4) transitions. In real applications, context-dependency transducers will benefit significantly from determinization and mini-

Figure reproduced from “Weighted Finite State Transducers in Speech Recognition”, Mohri et al., 2002

Arc labels: “monophone : phone / left-context_right-context”C-1:

C

AcousticIndices

Language Model Word

Sequence

AcousticModels

Triphones

ContextTransducer

Monophones

Pronunciation Model

Words

WFST-based ASR System

Page 9: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

AcousticIndices

Language Model Word

Sequence

AcousticModels

Triphones

ContextTransducer

Monophones

Pronunciation Model

Words

L

Figure reproduced from “Weighted Finite State Transducers in Speech Recognition”, Mohri et al., 2002

M. Mohri: Weighted FSTs in Speech Recognition 5

(a)

0 1using:using/1 2data:data/0.66

3intuition:intuition/0.33

4

is:is/0.5

are:are/0.5

is:is/1

5better:better/0.7

worse:worse/0.3

(b)

0

1d:data/1

5d:dew/1

2ey:ε/0.5

ae:ε/0.5

6uw:ε/1

3t:ε/0.3

dx:ε/0.74

ax: ε /1

Figure 2: Weighted finite-state transducer examples.

and path weight are those given earlier for acceptors. A path’s output label isthe concatenation of output labels of its transitions.

The examples in Figure 2 encode (a superset of) the information in the WFSAsof Figure 1a-b as WFSTs. Figure 2a represents the same language model asFigure 1a by giving each transition identical input and output labels. This addsno new information, but is a convenient way of interpreting any acceptor as atransducer that we will use often.

Figure 2b represents a toy pronunciation lexicon as a mapping from phonesequences to words in the lexicon, in this example data and dew, with proba-bilities representing the likelihoods of alternative pronunciations. Since a wordpronunciation may be a sequence of several phones, the path corresponding toeach pronunciation has ϵ-output labels on all but the word-initial transition. Thistransducer has more information than the WFSA in Figure 1b. Since words areencoded by the output label, it is possible to combine the pronunciation trans-ducers for more than one word without losing word identity. Similarly, HMMstructures of the form given in Figure 1c can can be combined into a singletransducer that preserves phone model identity while sharing distribution sub-sequences whenever possible.

2.3. Weighted Transducer Algorithms

Speech recognition architectures commonly give the run-time decoder the taskof combining and optimizing transducers such as those in Figure 1. The decoderfinds word pronunciations in its lexicon and substitutes them into the grammar.Phonetic tree representations may be used to improve search efficiency at thispoint [Ortmanns et al., 1996]. The decoder then identifies the correct context-dependent models to use for each phone in context, and finally substitutes themto create an HMM-level transducer. The software that performs these opera-

WFST-based ASR System

Page 10: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

AcousticIndices

Language Model Word

Sequence

AcousticModels

Triphones

ContextTransducer

Monophones

Pronunciation Model

Words

0the birds/0.404

animals/1.789

are/0.693

were/0.693

boy/1.789

is

walking

G

WFST-based ASR System

Page 11: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Constructing the Decoding Graph

AcousticIndices

Language Model Word

Sequence

AcousticModels

Triphones

ContextTransducer

Monophones

Pronunciation Model

Words

H C L G

“Weighted Finite State Transducers in Speech Recognition”, Mohri et al., Computer Speech & Language, 2002

Decoding graph, D = H ⚬ C ⚬ L ⚬ G

Construct decoding search graph using H ⚬ C ⚬ L ⚬ G that maps acoustic states to word sequencesCarefully construct D using optimization algorithms:

D = min(det(H ⚬ det(C ⚬ det(L ⚬ G))))Decode test utterance O by aligning acceptor X (corresponding to O) with H ⚬ C ⚬ L ⚬ G:

X ⚬ H ⚬ C ⚬ L ⚬ GW ⇤ = argminW=out[⇡]

where π is a path in the composed FST, out[π] is the output label sequence of π

Page 12: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

AcousticIndices

Language Model Word

Sequence

AcousticModels

Triphones

ContextTransducer

Monophones

Pronunciation Model

Words

H C L G

“Weighted Finite State Transducers in Speech Recognition”, Mohri et al., Computer Speech & Language, 2002

Structure of X (derived from O):f0:10.578

f1:14.221

f1000:5.678

f500:8.123⠇

f0:9.21

f1:5.645

f1000:15.638

f500:11.233⠇

f0:19.12

f1:13.45

f1000:11.11

f500:20.21⠇ …………

f0:18.52

f1:12.33

f1000:15.99

f500:10.21⠇

Decode test utterance O by aligning acceptor X (corresponding to O) with H ⚬ C ⚬ L ⚬ G:

X ⚬ H ⚬ C ⚬ L ⚬ GW ⇤ = argminW=out[⇡]

where π is a path in the composed FST, out[π] is the output label sequence of π

Constructing the Decoding Graph

Page 13: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

AcousticIndices

Language Model Word

Sequence

AcousticModels

Triphones

ContextTransducer

Monophones

Pronunciation Model

Words

H C L G

“Weighted Finite State Transducers in Speech Recognition”, Mohri et al., Computer Speech & Language, 2002

• Each fk maps to a distinct triphone HMM state j • Weights of arcs in the ith chain link correspond to observation probabilities bj(oi) • X is a very large FST which is never explicitly constructed! • H ⚬ C ⚬ L ⚬ G is typically traversed dynamically (search algorithms will be covered

later in the semester)

f0:10.578

f1:14.221

f1000:5.678

f500:8.123⠇

f0:9.21

f1:5.645

f1000:15.638

f500:11.233⠇

f0:19.12

f1:13.45

f1000:11.11

f500:20.21⠇ …………

f0:18.52

f1:12.33

f1000:15.99

f500:10.21⠇

X

Constructing the Decoding Graph

Page 14: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

1st-Pass Recognition Networks – 40K NAB Task

network states transitions

G 1,339,664 3,926,010

L � G 8,606,729 11,406,721

det(L � G) 7,082,404 9,836,629

C � det(L � G)) 7,273,035 10,201,269

det(H � C � L � G) 18,317,359 21,237,992

F 3,188,274 6,108,907

min(F ) 2,616,948 5,497,952

OpenFst Part III. Applications Integrated Context-Dependent Networks in VLVR 55

1st-Pass Recognition Speed - 40K NAB Eval ’95

network x real-time

C � L � G 12.5

C � det(L � G) 1.2

det(H � C � L � G) 1.0

push(min(F )) 0.7

Recognition speed of the first-pass networks in the NAB 40, 000-word vocabulary task at 83%

word accuracy.

OpenFst Part III. Applications Integrated Context-Dependent Networks in VLVR 56

Impact of WFST Optimizations40K NAB Evaluation Set ’95 (83% word accuracy)

Tables from http://www.openfst.org/twiki/pub/FST/FstHltTutorial/tutorial_part3.pdf

Page 15: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Toolkits to work with finite-state machines• AT&T FSM Library (no longer supported)

http://www3.cs.stonybrook.edu/~algorith/implement/fsm/implement.shtml

• RWTH FSA Toolkithttps://www-i6.informatik.rwth-aachen.de/~kanthak/fsa.html

• Carmel https://www.isi.edu/licensed-sw/carmel/

• MIT FST Toolkithttp://people.csail.mit.edu/ilh/fst/

• OpenFST Toolkit (actively supported)http://www.openfst.org/twiki/bin/view/FST/WebHome

Page 16: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Brief Introduction to the OpenFST Toolkit

Page 17: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

ε:n

a:a

an:a0

0 1 an a <eps> 0an 1a 2

<eps> 0a 1n 2

1 2 <eps> n0 2 a a12

Inputalphabet (in.txt)

Outputalphabet (out.txt)

“0”�label�is�reserved�for�epsilon

A.txt

Quick Intro to OpenFst (www.openfst.org)

Page 18: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

ε:n/1.0

a:a/0.5

2/0.an:a/0.5

0

0 1 an a 0.51 2 <eps> n 1.00 2 a a 0.512 0.1

Quick Intro to OpenFst (www.openfst.org)

Page 19: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Compiling & Printing FSTs

The text FSTs need to be “compiled” into binary objects before further use with OpenFst utilities

• Command used to compile:

fstcompile --isymbols=in.txt --osymbols=out.txt A.txt A.fst

• Get back the text FST using a print command with the binary file:

fstprint --isymbols=in.txt --osymbols=out.txt A.fst A.txt

Page 20: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Composing FSTs

The text FSTs need to be “compiled” into binary objects before further use with OpenFst utilities

• Command used to compose:

fstcompose A.fst B.fst AB.fst

• OpenFST requirement: One or both of the input FSTs should be appropriately sorted before composition

fstarcsort —-sort_type=olabel A.fst |\ fstcompose - B.fst AB.fst

Page 21: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Drawing FSTs

Small FSTs can be visualized easily using the draw tool:

fstdraw --isymbols=in.txt --osymbols=out.txt A.fst |\ dot -Tpdf > A.pdf

0

1an:a

2a:a

<eps>:n

Page 22: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

FSTs can get very large!

Page 23: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Basics of Speech Production

Page 24: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Speech Production

are more easily explained if a single feature value is allowed to span (what appearson the surface to be) more than one segment. Autosegmental phonology posits somerelationships (or associations) between segments in different tiers, which limit thetypes of transformations that can occur. We will not make use of the details ofthis theory, other than the motivation that features inherently lie in different tiers ofrepresentation.

2.4.3 Articulatory phonology

In the late 1980s, Browman and Goldstein proposed articulatory phonology [BG86,BG92], a theory that differs from previous ones in that the basic units in the lexiconare not abstract binary features but rather articulatory gestures. A gesture is essen-tially an instruction to the vocal tract to produce a certain degree of constriction ata given location with a given set of articulators. For example, one gesture might be“narrow lip opening”, an instruction to the lips and jaw to position themselves so asto effect a narrow opening at the lips. Figure 2-3 shows the main articulators of thevocal tract to which articulatory gestures refer. We are mainly concerned with thelips, tongue, glottis (controlling voicing), and velum (controlling nasality).

Figure 2-3: A midsagittal section showing the major articulators of the vocal tract,reproduced from [oL04].

39

Schematic representation of the vocal organs

Schematic from L.Rabiner and B.-H.Juang , Fundamentals of speech recognition, 1993Figure from http://www.phon.ucl.ac.uk/courses/spsci/iss/week6.php

Page 25: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Sound units

• Phones are acoustically distinct units of speech

• Phonemes are abstract linguistic units that impart different meanings in a given language

• Minimal pair: pan vs. ban

• Allophones are different acoustic realisations of the same phoneme

• Phonetics is the study of speech sounds and how they’re produced

• Phonology is the study of patterns of sounds in different languages

Page 26: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Vowels• Sounds produced with no obstruction to the flow of air

through the vocal tract

Image from https://en.wikipedia.org/wiki/File:IPA_vowel_chart_2005.png

VOWEL QUADRILATERAL

Page 27: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Formants of vowels• Formants are resonance frequencies of the vocal tract (denoted by F1, F2, etc.)

• F0 denotes the fundamental frequency of the periodic source (vibrating vocal folds)

• Formant locations specify certain vowel characteristics

Image from: https://www.phon.ucl.ac.uk/courses/spsci/iss/week5.php

Page 28: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Spectrogram

• Spectrogram is a sequence of spectra stacked together in time, with amplitude of the frequency components expressed as a heat map

• Spectrograms of certain vowels: http://www.phon.ucl.ac.uk/courses/spsci/iss/week5.php

• Praat (http://www.fon.hum.uva.nl/praat/) is a good toolkit to analyse speech signals (plot spectrograms, generate formants/pitch curves, etc.)

Page 29: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Consonants (voicing/place/manner)

• “Consonants are made by restricting or blocking the airflow in some way, and may be voiced or unvoiced.” (J&M, Ch. 7)

Page 30: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Voiced/Unvoiced Sounds

• Sounds made with vocal cords vibrating: voiced

• E.g. /g/, /d/, etc.

• All English vowel sounds are voiced

• Sounds made without vocal cord vibration: voiceless

• E.g. /k/, /t/, etc.

Page 31: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Consonants (voicing/place/manner)

• “Consonants are made by restricting or blocking the airflow in some way, and may be voiced or unvoiced.” (J&M, Ch. 7)

• Consonants can be labeled depending on

• where the constriction is made

• how the constriction is made

Page 32: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Place of articulation

are more easily explained if a single feature value is allowed to span (what appearson the surface to be) more than one segment. Autosegmental phonology posits somerelationships (or associations) between segments in different tiers, which limit thetypes of transformations that can occur. We will not make use of the details ofthis theory, other than the motivation that features inherently lie in different tiers ofrepresentation.

2.4.3 Articulatory phonology

In the late 1980s, Browman and Goldstein proposed articulatory phonology [BG86,BG92], a theory that differs from previous ones in that the basic units in the lexiconare not abstract binary features but rather articulatory gestures. A gesture is essen-tially an instruction to the vocal tract to produce a certain degree of constriction ata given location with a given set of articulators. For example, one gesture might be“narrow lip opening”, an instruction to the lips and jaw to position themselves so asto effect a narrow opening at the lips. Figure 2-3 shows the main articulators of thevocal tract to which articulatory gestures refer. We are mainly concerned with thelips, tongue, glottis (controlling voicing), and velum (controlling nasality).

Figure 2-3: A midsagittal section showing the major articulators of the vocal tract,reproduced from [oL04].

39

• Bilabial (both lips) [b],[p],[m], etc.

• Labiodental (with lower lip and upper teeth) [f], [v], etc.

• Interdental (tip of tongue between teeth) [ⲑ] (thought), [δ] (this)

Page 33: WFSTs in ASR Basics of Speech Productionpjyothi/cs753/slides/lecture6.pdf · 8 e:eps 12 e:eps 16 e:eps d:eps 3 b:eps 6 9 a:eps d:eps 10 13 d:eps 14 e:eps 17 c:eps 18 a:eps 19 d:eps

Manner of articulation

are more easily explained if a single feature value is allowed to span (what appearson the surface to be) more than one segment. Autosegmental phonology posits somerelationships (or associations) between segments in different tiers, which limit thetypes of transformations that can occur. We will not make use of the details ofthis theory, other than the motivation that features inherently lie in different tiers ofrepresentation.

2.4.3 Articulatory phonology

In the late 1980s, Browman and Goldstein proposed articulatory phonology [BG86,BG92], a theory that differs from previous ones in that the basic units in the lexiconare not abstract binary features but rather articulatory gestures. A gesture is essen-tially an instruction to the vocal tract to produce a certain degree of constriction ata given location with a given set of articulators. For example, one gesture might be“narrow lip opening”, an instruction to the lips and jaw to position themselves so asto effect a narrow opening at the lips. Figure 2-3 shows the main articulators of thevocal tract to which articulatory gestures refer. We are mainly concerned with thelips, tongue, glottis (controlling voicing), and velum (controlling nasality).

Figure 2-3: A midsagittal section showing the major articulators of the vocal tract,reproduced from [oL04].

39

• Plosive/Stop (airflow completely blocked followed by a release) [p],[g],[t],etc.

• Fricative (constricted airflow) [f], [s], [th], etc.

• Affricate (stop + fricative) [ch], [jh], etc.

• Nasal (lowering velum) [n], [m], etc.

See realtime MRI productions of vowels and consonants here: http://sail.usc.edu/span/rtmri_ipa/je_2015.html