The Noisy Channel Model. CS 294-5: Statistical Natural Language Processing. Speech Recognition Architecture. Digitizing Speech

Similar documents
The Noisy Channel Model. Statistical NLP Spring Mel Freq. Cepstral Coefficients. Frame Extraction ... Lecture 10: Acoustic Models

Statistical NLP Spring The Noisy Channel Model

Statistical NLP Spring Digitizing Speech

Digitizing Speech. Statistical NLP Spring Frame Extraction. Gaussian Emissions. Vector Quantization. HMMs for Continuous Observations? ...

The Noisy Channel Model. Statistical NLP Spring Mel Freq. Cepstral Coefficients. Frame Extraction ... Lecture 9: Acoustic Models

Lecture 5: GMM Acoustic Modeling and Feature Extraction

CS 136 Lecture 5 Acoustic modeling Phoneme modeling

Speech and Language Processing. Chapter 9 of SLP Automatic Speech Recognition (II)

Speech Recognition. CS 294-5: Statistical Natural Language Processing. State-of-the-Art: Recognition. ASR for Dialog Systems.

Lecture 3: ASR: HMMs, Forward, Viterbi

CS 136a Lecture 7 Speech Recognition Architecture: Training models with the Forward backward algorithm

Experiments with a Gaussian Merging-Splitting Algorithm for HMM Training for Speech Recognition

University of Cambridge. MPhil in Computer Speech Text & Internet Technology. Module: Speech Processing II. Lecture 2: Hidden Markov Models I

CS 188: Artificial Intelligence Fall 2011

Hidden Markov Models and Gaussian Mixture Models

Hidden Markov Modelling

Automatic Speech Recognition (CS753)

Automatic Speech Recognition (CS753)

Ngram Review. CS 136 Lecture 10 Language Modeling. Thanks to Dan Jurafsky for these slides. October13, 2017 Professor Meteer

Detection-Based Speech Recognition with Sparse Point Process Models

Hidden Markov Models and Gaussian Mixture Models

STA 4273H: Statistical Machine Learning

FEATURE PRUNING IN LIKELIHOOD EVALUATION OF HMM-BASED SPEECH RECOGNITION. Xiao Li and Jeff Bilmes

Hidden Markov Model and Speech Recognition

On the Influence of the Delta Coefficients in a HMM-based Speech Recognition System

CS 343: Artificial Intelligence

Temporal Modeling and Basic Speech Recognition

STA 414/2104: Machine Learning

Chapter 9 Automatic Speech Recognition DRAFT

Sparse Models for Speech Recognition

CS 5522: Artificial Intelligence II

CS 5522: Artificial Intelligence II

PHONEME CLASSIFICATION OVER THE RECONSTRUCTED PHASE SPACE USING PRINCIPAL COMPONENT ANALYSIS

Engineering Part IIB: Module 4F11 Speech and Language Processing Lectures 4/5 : Speech Recognition Basics

Doctoral Course in Speech Recognition. May 2007 Kjell Elenius

Augmented Statistical Models for Speech Recognition

CS 343: Artificial Intelligence

Automatic Speech Recognition (CS753)

10. Hidden Markov Models (HMM) for Speech Processing. (some slides taken from Glass and Zue course)

Lecture 10. Discriminative Training, ROVER, and Consensus. Michael Picheny, Bhuvana Ramabhadran, Stanley F. Chen

Genones: Generalized Mixture Tying in Continuous. V. Digalakis P. Monaco H. Murveit. December 6, 1995 EDICS SA and

Speech Recognition HMM

Jorge Silva and Shrikanth Narayanan, Senior Member, IEEE. 1 is the probability measure induced by the probability density function

Machine Learning for Data Science (CS4786) Lecture 12

Lecture 9: Speech Recognition. Recognizing Speech

Lecture 9: Speech Recognition

Reformulating the HMM as a trajectory model by imposing explicit relationship between static and dynamic features

Lecture 3: Pattern Classification

Hidden Markov Models in Language Processing

Lecture 3. Gaussian Mixture Models and Introduction to HMM s. Michael Picheny, Bhuvana Ramabhadran, Stanley F. Chen, Markus Nussbaum-Thom

A TWO-LAYER NON-NEGATIVE MATRIX FACTORIZATION MODEL FOR VOCABULARY DISCOVERY. MengSun,HugoVanhamme

Sparse Forward-Backward for Fast Training of Conditional Random Fields

Design and Implementation of Speech Recognition Systems

Expectation Maximization (EM)

Estimation of Cepstral Coefficients for Robust Speech Recognition

Automatic Speech Recognition (CS753)

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi

Discriminative training of GMM-HMM acoustic model by RPCL type Bayesian Ying-Yang harmony learning

Upper Bound Kullback-Leibler Divergence for Hidden Markov Models with Application as Discrimination Measure for Speech Recognition

Deep Learning for Speech Recognition. Hung-yi Lee

Automatic Phoneme Recognition. Segmental Hidden Markov Models

ACS Introduction to NLP Lecture 2: Part of Speech (POS) Tagging

Weighted Finite State Transducers in Automatic Speech Recognition

Matrix Updates for Perceptron Training of Continuous Density Hidden Markov Models

N-gram Language Modeling

HIDDEN MARKOV MODELS IN SPEECH RECOGNITION

Boundary Contraction Training for Acoustic Models based on Discrete Deep Neural Networks

Outline of Today s Lecture

Robust Speech Recognition in the Presence of Additive Noise. Svein Gunnar Storebakken Pettersen

Probabilistic Graphical Models: MRFs and CRFs. CSE628: Natural Language Processing Guest Lecturer: Veselin Stoyanov

Lecture 6. Michael Picheny, Bhuvana Ramabhadran, Stanley F. Chen, Markus Nussbaum-Thom

An Evolutionary Programming Based Algorithm for HMM training

N-gram Language Modeling Tutorial

Input to Baum-Welch. O unlabeled sequence of observadons Q vocabulary of hidden states. For ice-cream task O = {1,3,2,,} Q = {H,C}

Lecture 3: Pattern Classification. Pattern classification

Automatic Speech Recognition (CS753)

Investigate more robust features for Speech Recognition using Deep Learning

Hidden Markov Models

Support Vector Machines using GMM Supervectors for Speaker Verification

Robust Speaker Identification

Estimation of Relative Operating Characteristics of Text Independent Speaker Verification

Monaural speech separation using source-adapted models

Mixtures of Gaussians with Sparse Regression Matrices. Constantinos Boulis, Jeffrey Bilmes

The Noisy Channel Model and Markov Models

Covariance Matrix Enhancement Approach to Train Robust Gaussian Mixture Models of Speech Data

] Automatic Speech Recognition (CS753)

SPEECH RECOGNITION USING TIME DOMAIN FEATURES FROM PHASE SPACE RECONSTRUCTIONS

Comparing linear and non-linear transformation of speech

Factor analysed hidden Markov models for speech recognition

Using Sub-Phonemic Units for HMM Based Phone Recognition

L11: Pattern recognition principles

Statistical Methods for NLP

Discriminative models for speech recognition

Improving the Multi-Stack Decoding Algorithm in a Segment-based Speech Recognizer

Graphical Models. Mark Gales. Lent Machine Learning for Language Processing: Lecture 3. MPhil in Advanced Computer Science

Low development cost, high quality speech recognition for new languages and domains. Cheap ASR

Design and Implementation of Speech Recognition Systems

ON THE USE OF MLP-DISTANCE TO ESTIMATE POSTERIOR PROBABILITIES BY KNN FOR SPEECH RECOGNITION

Basic Text Analysis. Hidden Markov Models. Joakim Nivre. Uppsala University Department of Linguistics and Philology

Why DNN Works for Acoustic Modeling in Speech Recognition?

Transcription:

CS 294-5: Statistical Natural Language Processing The Noisy Channel Model Speech Recognition II Lecture 21: 11/29/05 Search through space of all possible sentences. Pick the one that is most probable given the waveform. Slides directly from Dan Jurafsky, indirectly many others Speech Recognition Architecture Digitizing Speech Frame Extraction Mel Freq. Cepstral Coefficients A frame (25 ms wide) extracted every 10 ms Do FFT to get spectral information Like the spectrogram/spectrum we saw earlier 25 ms 10ms... Apply Mel scaling Linear below 1kHz, log above, equal samples above and below 1kHz Models human ear; more sensitivity in lower freqs a 1 a 2 a 3 Figure from Simon Arnfield Plus Discrete Cosine Transformation 1

Final Feature Vector HMMs for Speech 39 (real) features per 10 ms frame: 12 MFCC features 12 Delta MFCC features 12 Delta- Delta MFCC features 1 (log) frame energy 1 Delta (log) frame energy 1 Delta- Delta (log frame energy) So each frame is represented by a 39D vector Phones Aren t Homogeneous Need to Use Subphones 5000 Frequency (Hz) 0 0.48152 ay k 0.937203 Time (s) A Word with Subphones Viterbi Decoding 2

ASR Lexicon: Markov Models HMMs for Continuous Observations? Before: discrete, finite set of observations Now: spectral feature vectors are real- valued! Solution 1: discretization Solution 2: continuous emissions models Gaussians Multivariate Gaussians Mixtures of Multivariate Gaussians A state is progressively: Context independent subphone (~3 per phone) Context dependent phone (=triphones) State-tying of CD phone Idea: discretization Map MFCC vectors onto discrete symbols Compute probabilities just by counting This is called Vector Quantization or VQ Not used for ASR anymore; too simple Useful to consider as a starting point Vector Quantization Gaussian Emissions VQ is insufficient for real ASR Instead: Assume the possible values of the observation vectors are normally distributed. Represent the observation likelihood function as a Gaussian with mean µ j and variance σ j 2 f (x µ,σ) = 1 (x µ)2 exp( ) σ 2π 2σ 2 Gaussians for Acoustic Modeling A Gaussian is parameterized by a mean and a variance: Different means P(o q): P(o q) o P(o q) is highest here at mean P(o q is low here, very far from mean) Multivariate Gaussians Instead of a single mean µ and variance σ: 1 (x µ)2 f (x µ,σ) = exp( ) σ 2π 2σ 2 Vector of means µ and covariance matrix Σ 1 f (x µ,σ) = (2π) n /2 Σ exp 1 1/2 2 (x µ)t Σ 1 (x µ) Usually assume diagonal covariance This isn t very true for FFT features, but is fine for MFCC features 3

Gaussian Intuitions: Size of Σ Gaussians: Off-Diagonal µ = [0 0] µ = [0 0] µ = [0 0] Σ = I Σ = 0.6I Σ = 2I As Σ becomes larger, Gaussian becomes more spread out; as Σ becomes smaller, Gaussian more compressed As we increase the off-diagonal entries, more correlation between value of x and value of y Text and figures from Andrew Ng s lecture notes for CS229 Text and figures from Andrew Ng s lecture notes for CS229 In two dimensions In two dimensions From Chen, Picheny et al lecture slides From Chen, Picheny et al lecture slides But we re not there yet Single Gaussian may do a bad job of modeling distribution in any dimension: Solution: Mixtures of Gaussians Figure from Chen, Picheney et al slides Mixtures of Gaussians M mixtures of Gaussians: For diagonal covariance: b j (o t ) = M k=1 f (x µ jk,σ jk ) = c jk N(x,µ jk,σ jk ) c jk D exp( 1 D (x jkd µ jkd ) 2 ) 2π D 2 2 2 2 d =1 σ σ jkd jkd d =1 M M k=1 b j (o t ) = c jk N(o t,µ jk,σ jk ) k=1 4

GMMs Summary: each state has a likelihood function parameterized by: M Mixture weights M Mean Vectors of dimensionality D Either M Covariance Matrices of DxD Or more likely M Diagonal Covariance Matrices of DxD which is equivalent to M Variance Vectors of dimensionality D Training Mixture Models Forced Alignment Computing the Viterbi path over the training data is called forced alignment We know which word string to assign to each observation sequence. We just don t know the state sequence. So we constrain the path to go through the correct words And otherwise do normal Viterbi Result: state sequence! Modeling phonetic context Need with triphone models W iy r iy m iy n iy Implications of Cross-Word Triphones Possible triphones: 50x50x50=125,000 How many triphone types actually occur? 20K word WSJ Task (from Bryan Pellom) Word-internal models: need 14,300 triphones Cross-word models: need 54,400 triphones But in training data only 22,800 triphones occur! Need to generalize models. State Tying / Clustering [Young, Odell, Woodland 1994] How do we decide which triphones to cluster together? Use phonetic features (or broad phonetic classes ) Stop Nasal Fricative Sibilant Vowel lateral 5

State Tying Creating CD phones: Start with monophone, do EM training Clone Gaussians into triphones Build decision tree and cluster Gaussians Clone and train mixtures (GMMs Viterbi with 2 Words + Unif. LM Null transition from the end- state of each word to start- state of all (both) words. Figure from Huang et al page 612 Search space for unigram LM Search space with bigrams Figure from Huang et al page 617 Figure from Huang et al page 618 Speeding things up Viterbi is O(N 2 T), where N is total number of HMM states, and T is length This is too large for real-time search A ton of work in ASR search is just to make search faster: Beam search (pruning) Fast match Tree- based lexicons Beam Search Most common strategy (still!) Just like earlier in the term Instead of retaining all candidates at every time frame Use a threshold T to keep subset At each time t Identify state with lowest cost D min Each state with cost > D min + T is discarded ( pruned ) before moving on to time t+1 Empirically, beam size of 5-10% of search space 90-95% of HMM states don t have to be considered Vast savings in time 6

A* Decoding (2) A* Decoding (cont.) Tree structured lexicon Multipass Search N-best list Word Graph From Huang et al, page 664 From Huang et al, page 665 7

One-pass vs. multipass Potential problems with multipass Can t use for real-time (need end of sentence) (But can keep successive passes really fast) Each pass can introduce inadmissible pruning (But one-pass does the same w/beam pruning and fastmatch) Why multipass Very expensive KSs. (NL parsing,higher-order n-gram, etc) Spoken language understanding: N-best perfect interface Research: N-best list very powerful offline tools for algorithm development N-best lists needed for discriminant training (MMIE, MCE) to get rival hypotheses 8