HMM
HMM
HMM
1
= 0.4
2
= 0.2
3
= 0.4
0.3
0.6 0.4
S
1
S
2
1
= 1.0
2
= 0.0
3
= 0.0
4
= 0.0
S
3
0.1
0.4
S
4
1.0
0.9
0.1
0.2
1
0
HMM Topologies
The topology must be specified in advance by the
system
designer
Common use in speech is to have one HMM per
phoneme, and three states per phoneme. Then, the
phoneme-level
HMMs can be connected to form word-level HMMs
1
= 1.0
2
= 0.0
3
= 0.0
0.4
0.6 0.3
A
1
A
2
A
3
0.5
0.7
0.5
0.5 0.2
B
1
B
2
B
3
0.3
0.8 0.4
0.6 0.3
A
1
A
2
A
3
0.5
0.7 0.4
0.6 0.2
T
1
T
2
T
3
4.0
0.8
0.5
0.7
0.5 0.6
HMM generation of observation sequences
The three basic HMM problems
Forward and Backward procedures
The Forward procedure
The Backward procedure
The Viterbi algorithm
The problem with choosing the individually most likely
states is that the overall state sequence may not be valid
Consider a situation where the individually most likely
states are = and +1=, but the transition
probability =0
Instead, and to avoid this problem, it is common to look
for the single best state sequence, at the expense of
having sub-optimal individual states
This is accomplished with the Viterbi algorithm
The Viterbi algorithm
- To find the single best state sequence we define yet
another variable =max12112=,1
2|
which represents the highest probability along a single
path that accounts for the first observations and ends at
state
By induction, +1 can be computed as +1=max
+1
To retrieve the state sequence, we also need to keep
track of the state that maximizes at each time , which
is done by constructing an array +1=arg max1
+1 is the state at time from which a transition to
state maximizes the probability +1
Baum-Welsh re-estimation
Re-estimation procedure
2
5
L
SHOW ALL ALERTS
SH OW AA AX L ER TS
SH SH SH OW OW OW AA AA AA L L L AX AX AX L L L ER ER ER TS TS TS
Example