Automatic Speech Recognition (CS753) Automatic Speech Recognition (CS753) Lecture 20: Discriminative Training for HMMs Instructor: Preethi Jyothi Mar 30, 2017
Discriminative Training
Recall: MLE for HMMs Maximum likelihood estimation (MLE) sets HMM parameters so as to maximise the objective function: N X L = log P λ ( X i | M i ) i =1 where X 1 , …, X i , … X N are training u tu erances M i is the HMM corresponding to the word sequence of X i λ corresponds to the HMM parameters What are some conceptual problems with this approach?
Discriminative Learning Discriminative models directly model the class posterior • probability or learn the parameters of a joint probability model discriminatively so that classification errors are minimised As opposed to generative models that a tu empt to learn a • probability model of the data distribution [Vapnik] “ one should solve the (classification/recognition) • problem directly and never solve a more general problem as an intermediate step ” [Vapnik]: V. Vapnik, Statistical Learning Theory, 1998
Discriminative Learning Two central issues in developing discriminative learning • methods: 1. Constructing suitable objective functions for optimisation 2. Developing optimization techniques for these objective functions
Maximum mutual information (MMI) estimation: Discriminative Training MMI aims to directly maximise the posterior probability • (criterion also referred to as conditional maximum likelihood) N X F MMI = log P λ ( M i | X i ) i =1 N P λ ( X i | M i ) P ( W i ) X = log P W 0 P λ ( X i | M W 0 ) P ( W 0 ) i =1 P(W) is the language model probability •
Why is it called MMI? Mutual information I ( X , W ) between acoustic data X and word • labels W is defined as: Pr( X, W ) X I ( X, W ) = Pr( X, W ) log Pr( X ) Pr( W ) X,W Pr( X, W ) log Pr( W | X ) X = Pr( W ) X,W = H ( W ) − H ( W | X ) where H(W) is the entropy of W and H(W|X) is the conditional entropy
Why is it called MMI? Assume H(W) is given via the language model. Then, • maximizing mutual information becomes equivalent to minimising conditional entropy N H ( W | X ) = − 1 X log Pr( W i | X i ) N i =1 N = − 1 Pr( X i | W i ) Pr( W i ) X log P W 0 Pr( X i | W 0 ) Pr( W 0 ) N i =1 Thus, MMI is equivalent to maximizing: • N P λ ( X i | M i ) P ( W i ) X F MMI = log P W 0 P λ ( X i | M W 0 ) P ( W 0 ) i =1
MMI estimation How do we compute this? N P λ ( X i | M i ) P ( W i ) X F MMI = log P W 0 P λ ( X i | M W 0 ) P ( W 0 ) i =1 Numerator: Likelihood of data given correct word sequence • Denominator: Total likelihood of the data given all possible • word sequences
Recall: Word La tu ices A word la tu ice is a pruned version of the decoding graph for an • u tu erance Acyclic directed graph with arc costs computed from acoustic • model and language model scores La tu ice nodes implicitly capture information about time within • the u tu erance OFTEN VERY VERY HAVE OFTEN FINE SIL I MOVE I VEAL I HAVE SIL SIL HAVE FINE SIL FAST I T VERY SIL SIL VERY MOVE HAVE IT Time Image from [GY08]: Gales & Young, Application of HMMs in speech recognition, NOW book, 2008
MMI estimation How do we compute this? N P λ ( X i | M i ) P ( W i ) X F MMI = log P W 0 P λ ( X i | M W 0 ) P ( W 0 ) i =1 Numerator: Likelihood of data given correct word sequence • Denominator: Total likelihood of the data given all possible • word sequences Estimate by generating la tu ices, and summing over all • the word sequences in the la tu ice
MMI Training and La tu ices Computing the denominator: Estimate by generating la tu ices, • and summing over all the words in the la tu ice Numerator la tu ices: Restrict G to a linear chain acceptor • representing the words in the correct word sequence. La tu ices are usually only computed once for MMI training. HMM parameter estimation for MMI uses the extended Baum- • Welch algorithm [V96,WP00] Like HMMs, can DNNs also be trained with an MMI-type • objective function? Yes! (More about this next week.) [V96]:Valtchev et al., Lattice-based discriminative training for large vocabulary speech recognition, 1996 [WP00]: Woodland and Povey, Large scale discriminative training for speech recognition, 2000
MMI results on Switchboard Switchboard results on two eval sets (SWB, CHE). Trained on • 300 hours of speech. Comparing maximum likelihood (ML) against discriminatively trained GMM systems and MMI- trained DNNs. SWB CHE Total GMM ML 21.2 36.4 28.8 GMM MMI 18.6 33.0 25.8 DNN CE 14.2 25.7 20.0 DNN MMI 12.9 24.6 18.8 [V et al.]:Vesely et al., Sequence discriminative training of DNNs, Interspeech 2013
Another Discriminative Training Objective: Minimum Phone/Word Error (MPE/MWE) MMI is an optimisation criterion at the sentence-level. • Change the criterion so that it is directly related to sub- sentence (i.e. word or phone) error rate. MPE/MWE objective function is defined as: • N P W P λ ( X i | M W ) P ( W ) A ( W, W i ) X F MPE/MWE = log P W 0 P λ ( X i | M W 0 ) P ( W 0 ) i =1 where A( W , W i ) is phone/word accuracy of the sentence W given the reference sentence W i i.e. the total phone count in W i minus the sum of insertion/deletion/substitution errors of W
MPE/MWE training N P W P λ ( X i | M W ) P ( W ) A ( W, W i ) X F MPE/MWE = log P W 0 P λ ( X i | M W 0 ) P ( W 0 ) i =1 The MPE/MWE criterion is a weighted average of the phone/ • word accuracy over all the training instances A( W , W i ) can be computed either at the phone or word level for • the MPE or MWE criterion, respectively The weighting given by MPE/MWE depends on the number of • incorrect phones/words in the string while MMI looks at whether the entire sentence is correct or not
MPE results on Switchboard Switchboard results on eval set SWB. Trained on 68 hours of • speech. Comparing maximum likelihood (MLE) against discriminatively trained (MMI/MPE/MWE) GMM systems SWB %WER redn GMM MLE 46.6 - GMM MMI 44.3 2.3 GMM MPE 43.1 3.5 GMM MWE 43.3 3.3 [V et al.]:Vesely et al., Sequence discriminative training of DNNs, Interspeech 2013
How does this fit within an ASR system?
Estimating acoustic model parameters If A: speech u tu erance and O A : acoustic features corresponding • to the u tu erance A, W ∗ = arg max P λ ( O A | W ) P β ( W ) W ASR decoding: Return the word sequence that jointly assigns • the highest probability to O A How do we estimate λ in P λ ( O A | W )? • MLE estimation • MMI estimation • MPE/MWE estimation • Covered in this class
Another way to improve ASR performance: System Combination
System Combination Combining recognition outputs from multiple systems to produce a • hypothesis that is more accurate than any of the original systems Most widely used technique: ROVER [ROVER]. • 1-best word sequences from each system are aligned using a • greedy dynamic programming algorithm Voting-based decision made for words aligned together • Can we do be tu er than just looking at 1-best sequences? • Image from [ROVER]: Fiscus, Post-processing method to yield reduced word error rates, 1997
Recall: Word Confusion Networks Word confusion networks are normalised word la tu ices that provide alignments for a fraction of word sequences in the word la tu ice (a) Word Lattice OFTEN VERY VERY HAVE OFTEN FINE SIL I MOVE I VEAL I HAVE SIL SIL HAVE FINE SIL FAST I T VERY SIL SIL VERY MOVE HAVE IT Time (b) Confusion Network I HAVE IT VEAL FINE - OFTEN MOVE - VERY FAST Image from [GY08]: Gales & Young, Application of HMMs in speech recognition, NOW book, 2008
System Combination Combining recognition outputs from multiple systems to produce a • hypothesis that is more accurate than any of the original systems Most widely used technique: ROVER [ROVER]. • 1-best word sequences from each system are aligned using a • greedy dynamic programming algorithm Voting-based decision made for words aligned together • Could align confusion networks instead of 1-best sequences • Image from [ROVER]: Fiscus, Post-processing method to yield reduced word error rates, 1997
Recommend
More recommend