HMMs-Jelinek

HMMs-Jelinek - Speech Recognition Hidden Markov Models for...

Info iconThis preview shows pages 1–9. Sign up to view the full content.

View Full Document Right Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon
This is the end of the preview. Sign up to access the rest of the document.

Unformatted text preview: Speech Recognition Hidden Markov Models for Speech Recognition February 13, 2012 Veton Kpuska 2 Outline Introduction Information Theoretic Approach to Automatic Speech Recognition Problem formulation Discrete Markov Processes Forward-Backward algorithm Viterbi search Baum-Welch parameter estimation Other considerations Multiple observation sequences Phone-based models for continuous speech recognition Continuous density HMMs Implementation issues February 13, 2012 Veton Kpuska 3 Information Theoretic Approach to ASR Statistical Formulation of Speech Recognition A denotes the acoustic evidence (collection of feature vectors, or data in general) based on which recognizer will make its decision about which words were spoken. W denotes a string of words each belonging to a fixed and known vocabulary. Speech Producer Acoustic Processor Linguistic Decoder Speaker's Mind Speech Speaker Acoustic Channel Speech Recognizer A W February 13, 2012 Veton Kpuska 4 Information Theoretic Approach to ASR Assume that A is a sequence of symbols taken from some alphabet A . W denotes a string of n words each belonging to a fixed and known vocabulary V . V ,..., , 2 1 = i m w w w w W A ,..., , 2 1 = i m a a a a A February 13, 2012 Veton Kpuska 5 Information Theoretic Approach to ASR If P( W | A ) denotes the probability that the words W were spoken, given that the evidence A was observed, then the recognizer should decide in favor of a word string satisfying: The recognizer will pick the most likely word string given the observed acoustic evidence. ( 29 A W W W | max arg P = February 13, 2012 Veton Kpuska 6 Information Theoretic Approach to ASR From the well known Bayes rule of probability theory: P( W ) Probability that the word string W will be uttered P( A | W ) Probability that when W was uttered the acoustic evidence A will be observed P( A ) is the average probability that A will be observed: ( 29 ( 29 ( 29 ( 29 ( 29 ( 29 ( 29 ( 29 A W W A A W | W W A A A W | P P P P P P P P | | = = ( 29 ( 29 ( 29 = ' ' ' | W W W A A P P P February 13, 2012 Veton Kpuska 7 Information Theoretic Approach to ASR Since Maximization in: Is carried out with the variable A fixed (e.g., there is not other acoustic data save the one we are give), it follows from Bayes rule that the recognizers aim is to find the word string that maximizes the product P( A | W )P( W ) , that is ( 29 A W W W | max arg P = ( 29 ( 29 W A W W W P P | max arg = February 13, 2012 Veton Kpuska 8 Markov Processes About Markov Chains...
View Full Document

Page1 / 40

HMMs-Jelinek - Speech Recognition Hidden Markov Models for...

This preview shows document pages 1 - 9. Sign up to view the full document.

View Full Document Right Arrow Icon
Ask a homework question - tutors are online