Full-covariance model compensation for

Size: px
Start display at page:

Download "Full-covariance model compensation for"

Transcription

1 compensation transms Presentation Toshiba, 12 Mar 2008

2 Outline compensation transms compensation transms

3 Outline compensation transms compensation transms

4 Noise model x clean speech; n additive ; h convolutional ; -currupted speech y is y[k] = x[k] h[k] + n[k] In the mel-cepstral domain, static coefficients ( s ), y s = x s + h + C log ( 1 + exp ( C 1 (n s x s h) )) = x s + f (x s, n s, h). compensation transms Noise model M n = {µ n, Σ n, h}: the additive n is Gaussian; the convolutional h is stationary; they are ML-estimated [1].

5 Vector Taylor series Approximate with a first-order vector Taylor series expansion [2, 3]: y s = x s + f (x s, n s, h) x s + f (x s 0, n s 0, h) + (x s x s 0) f x s + (ns n s 0) f n s compensate static µ s y : compensation transms µ s y = E{y s } µ s x + f (µ s x, µ s n, h) similar Σ.

6 DPMC Data-driven parallel model combination [4] A Monte Carlo method. Draw N samples from distributions x s, n s, h; apply y s = x s + f (x s, n s, h) compensation transms estimate y s N ( µ s y, Σ s y ). As N, DPMC compensation becomes optimal.

7 Joint uncertainty decoding Joint uncertainty decoding finds compensation p(y t m) N (A (r) y t + b (r) ; µ (m), Σ (m) + Σ (r) bias ) from the joint distribution of the clean speech x and -corrupted speech y regression class r [ ] ([ ] [ ]) x µ x (r) Σ (r) x Σ xy (r) N y µ y (r), Σ yx (r) Σ y (r). compensation transms Normally, VTS is used to find the joint distribution. But DPMC can also be used!

8 Outline compensation transms compensation transms

9 Outline compensation transms compensation transms

10 Dynamic parameter compensation with VTS VTS compensation uses the continuous time approximation [5] delta and delta-delta parameters: so that y ys t µ y = E{y } ys x s µ x. compensation transms Attractive because simple.

11 Dynamic parameter compensation with DPMC are linear combinations of consecutive observations. If we knew y s t w... y s t+w, compensation could be exact: [ y s t y t ] = [ = A y s t 1 y s t y s t+1. ] y s t 1 y s t y s t+1 compensation transms

12 Dynamic parameter compensation with DPMC Draw N samples from distributions x s t w n s t w h.,.,. ; h estimate convert to x s t+w y s y y 2 y s t w. y s t+w N n s t+w N (µ y, Σ y ) ; ( Aµ y, AΣ y A T ). compensation transms

13 Dynamic parameter compensation Average KL divergence per feature dimension of components to a single-pass retrained system: KL divergence Uncompensated VTS DPMC compensation transms Dimension DPMC is better at model-based compensation than VTS.

14 Outline compensation transms compensation transms

15 compensation transms Different correlations in different conditions. Thanks to Hank Liao the figure and Toshiba Research Ltd. the data.

16 Compensating correlations System: Resource Management with artificial ; single-component system robustness; known profile with diagonal Σ n. Word error rates: Compensation Diagonal Full Uncompensated 38.2 % 64.6 % VTS (diagonal statistics) 15.5 % 18.1 % VTS (block statistics) 14.4 % 15.5 % DPMC (full statistics) 13.0 % 10.8 % Single-pass retrained 12.4 % 7.5 % compensation transms DPMC can compensate correlations. VTS cannot.

17 Outline compensation transms compensation transms

18 Joint Uncertainty Decoding with DPMC Joint uncertainty decoding currently uses VTS to find joint distribution [ ] ([ ] [ ]) x µ x (r) Σ (r) x Σ xy (r) N y µ y (r), Σ yx (r) Σ y (r). But it can use DPMC. compensation transms

19 Joint uncertainty decoding with DPMC Draw N samples from distributions x s t w n s t w h.,.,. ; h estimate convert to [ x y x s t+w x s t w. x s t+w y s t w. y s t+w ] N n s t+w N (µ x, y, Σ x, y ) ; ([ µx µ y ] [ Σx Σ, xy Σ yx Σ y ]). compensation transms

20 semi-tied covariance matrices semi-tied covariance matrices [6] convert joint uncertainty decoding to: p(y t m) N (ỹ t ; µ (m), Σ (m) + Σ (r) bias ) Σ (r) bias is full; but Σ (m) diag is diagonal; so that decoding is faster. N (A (r) ỹ t ; A (r) µ (m), Σ (m) diag ) compensation transms

21 Putting it together Recognition hypothesis Noise estimate µ n, Σ n, µ h JUD transm A (r) jnt, b(r) jnt, Σ(r) bias compensation transms Semi-tied covariance matrices A (r), Σ (m) diag

22 Outline compensation transms compensation transms

23 Resource Management corpus (1000 word vocabulary); artificial at 20 db; 9.5K components; 12 MFCCs, 0th coefficient, delta, acceleration; 16 regression classes; estimated profile [1] per speaker (other than SPR). compensation transms

24 Joint uncertainty decoding Joint m Transm WER VTS Diagonal 9.5 % DPMC Diagonal 8.6 % DPMC Full 6.8 % DPMC-based joint outperms VTS-based joint. Full compensation is better than diagonal. compensation transms

25 Overview Compensation WER Clean 38.0 % VTS, diagonal 8.5 % DPMC-joint, full 6.8 % semi-tied 6.8 % SPR joint 7.4 % SPR semi-tied 6.7 % Speaker-specific transm except single-pass retrained. semi-tied is faster to decode than Joint without the loss of accuracy. compensation transms

26 Outline compensation transms compensation transms

27 DPMC compensation outperms VTS: models dynamic parameters better; can model correlations; but is slower: VTS m required. DPMC-joint outperms VTS-joint. With predictive semi-tied covariance matrices decoding is fast. Future work: Faster method than DPMC; more accurate m the model. compensation transms

28 References I H. Liao and M. J. F. Gales, Joint uncertainty decoding robust large vocabulary, Cambridge University Engineering Department, Tech. Rep. cued/f-infeng/tr.552, November Appendix References P. J. Moreno, Speech recognition in noisy environments, Ph.D. dissertation, Carnegie Mellon University, A. Acero, L. Deng, T. Kristjansson, and J. Zhang, Hmm adaptation using vector Taylor series noisy, in Proceedings of the International Conference on Spoken Language Processing, vol. 3, 2000, pp M. J. F. Gales, -based techniques robust, Ph.D. dissertation, Cambridge University, 1995.

29 References II R. A. Gopinath, M. J. F. Gales, P. S. Gopalakrishnan, S. Balakrishnan-Aiyer, and M. A. Picheny, Robust in - permance of the IBM continuous speech recognizer on the ARPA spoke task, in Proceedings of the ARPA Workshop on Spoken Language System Technology, 1999, pp Appendix References M. J. F. Gales and R. C. van Dalen, linear transms robust, in Proceedings of the Automatic Speech Recognition and Understanding Workshop, 2007, pp

Model-Based Approaches to Robust Speech Recognition

Model-Based Approaches to Robust Speech Recognition Model-Based Approaches to Robust Speech Recognition Mark Gales with Hank Liao, Rogier van Dalen, Chris Longworth (work partly funded by Toshiba Research Europe Ltd) 11 June 2008 King s College London Seminar

More information

Noise Compensation for Subspace Gaussian Mixture Models

Noise Compensation for Subspace Gaussian Mixture Models Noise ompensation for ubspace Gaussian Mixture Models Liang Lu University of Edinburgh Joint work with KK hin, A. Ghoshal and. enals Liang Lu, Interspeech, eptember, 2012 Outline Motivation ubspace GMM

More information

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 9, SEPTEMBER

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 9, SEPTEMBER IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 21, NO. 9, SEPTEMBER 2013 1791 Joint Uncertainty Decoding for Noise Robust Subspace Gaussian Mixture Models Liang Lu, Student Member, IEEE,

More information

Improving Reverberant VTS for Hands-free Robust Speech Recognition

Improving Reverberant VTS for Hands-free Robust Speech Recognition Improving Reverberant VTS for Hands-free Robust Speech Recognition Y.-Q. Wang, M. J. F. Gales Cambridge University Engineering Department Trumpington St., Cambridge CB2 1PZ, U.K. {yw293, mjfg}@eng.cam.ac.uk

More information

ALGONQUIN - Learning dynamic noise models from noisy speech for robust speech recognition

ALGONQUIN - Learning dynamic noise models from noisy speech for robust speech recognition ALGONQUIN - Learning dynamic noise models from noisy speech for robust speech recognition Brendan J. Freyl, Trausti T. Kristjanssonl, Li Deng 2, Alex Acero 2 1 Probabilistic and Statistical Inference Group,

More information

Robust Speech Recognition in the Presence of Additive Noise. Svein Gunnar Storebakken Pettersen

Robust Speech Recognition in the Presence of Additive Noise. Svein Gunnar Storebakken Pettersen Robust Speech Recognition in the Presence of Additive Noise Svein Gunnar Storebakken Pettersen A Dissertation Submitted in Partial Fulfillment of the Requirements for the Degree of PHILOSOPHIAE DOCTOR

More information

Sparse Models for Speech Recognition

Sparse Models for Speech Recognition Sparse Models for Speech Recognition Weibin Zhang and Pascale Fung Human Language Technology Center Hong Kong University of Science and Technology Outline Introduction to speech recognition Motivations

More information

CEPSTRAL analysis has been widely used in signal processing

CEPSTRAL analysis has been widely used in signal processing 162 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 7, NO. 2, MARCH 1999 On Second-Order Statistics and Linear Estimation of Cepstral Coefficients Yariv Ephraim, Fellow, IEEE, and Mazin Rahim, Senior

More information

ELEN E Topics in Signal Processing Topic: Speech Recognition Lecture 9

ELEN E Topics in Signal Processing Topic: Speech Recognition Lecture 9 ELEN E6884 - Topics in Signal Processing Topic: Speech Recognition Lecture 9 Stanley F. Chen, Ellen Eide, and Michael A. Picheny IBM T.J. Watson Research Center Yorktown Heights, NY, USA stanchen@us.ibm.com,

More information

Uncertainty Decoding for Noise Robust Speech Recognition

Uncertainty Decoding for Noise Robust Speech Recognition Uncertainty Decoding for Noise Robust Speech Recognition Hank Liao Sidney Sussex College University of Cambridge September 2007 This dissertation is submitted for the degree of Doctor of Philosophy to

More information

ELEN E Topics in Signal Processing Topic: Speech Recognition Lecture 9

ELEN E Topics in Signal Processing Topic: Speech Recognition Lecture 9 ELEN E6884 - Topics in Signal Processing Topic: Speech Recognition Lecture 9 Stanley F. Chen, Michael A. Picheny, and Bhuvana Ramabhadran IBM T.J. Watson Research Center Yorktown Heights, NY, USA stanchen@us.ibm.com,

More information

Gaussian Mixture Model Uncertainty Learning (GMMUL) Version 1.0 User Guide

Gaussian Mixture Model Uncertainty Learning (GMMUL) Version 1.0 User Guide Gaussian Mixture Model Uncertainty Learning (GMMUL) Version 1. User Guide Alexey Ozerov 1, Mathieu Lagrange and Emmanuel Vincent 1 1 INRIA, Centre de Rennes - Bretagne Atlantique Campus de Beaulieu, 3

More information

MODEL-BASED TECHNIQUES FOR NOISE ROBUST SPEECH RECOGNITION

MODEL-BASED TECHNIQUES FOR NOISE ROBUST SPEECH RECOGNITION MODEL-BASED TECHNIQUES FOR NOISE ROBUST SPEECH RECOGNITION Mark John Francis Gales Gonville and Caius College September 1995 Dissertation submitted to the University of Cambridge for the degree of Doctor

More information

DNN-based uncertainty estimation for weighted DNN-HMM ASR

DNN-based uncertainty estimation for weighted DNN-HMM ASR DNN-based uncertainty estimation for weighted DNN-HMM ASR José Novoa, Josué Fredes, Nestor Becerra Yoma Speech Processing and Transmission Lab., Universidad de Chile nbecerra@ing.uchile.cl Abstract In

More information

GMM-based classification from noisy features

GMM-based classification from noisy features GMM-based classification from noisy features Alexey Ozerov, Mathieu Lagrange and Emmanuel Vincent INRIA, Centre de Rennes - Bretagne Atlantique STMS Lab IRCAM - CNRS - UPMC alexey.ozerov@inria.fr, mathieu.lagrange@ircam.fr,

More information

An Evolutionary Programming Based Algorithm for HMM training

An Evolutionary Programming Based Algorithm for HMM training An Evolutionary Programming Based Algorithm for HMM training Ewa Figielska,Wlodzimierz Kasprzak Institute of Control and Computation Engineering, Warsaw University of Technology ul. Nowowiejska 15/19,

More information

Covariance Matrix Enhancement Approach to Train Robust Gaussian Mixture Models of Speech Data

Covariance Matrix Enhancement Approach to Train Robust Gaussian Mixture Models of Speech Data Covariance Matrix Enhancement Approach to Train Robust Gaussian Mixture Models of Speech Data Jan Vaněk, Lukáš Machlica, Josef V. Psutka, Josef Psutka University of West Bohemia in Pilsen, Univerzitní

More information

Experiments with a Gaussian Merging-Splitting Algorithm for HMM Training for Speech Recognition

Experiments with a Gaussian Merging-Splitting Algorithm for HMM Training for Speech Recognition Experiments with a Gaussian Merging-Splitting Algorithm for HMM Training for Speech Recognition ABSTRACT It is well known that the expectation-maximization (EM) algorithm, commonly used to estimate hidden

More information

Reformulating the HMM as a trajectory model by imposing explicit relationship between static and dynamic features

Reformulating the HMM as a trajectory model by imposing explicit relationship between static and dynamic features Reformulating the HMM as a trajectory model by imposing explicit relationship between static and dynamic features Heiga ZEN (Byung Ha CHUN) Nagoya Inst. of Tech., Japan Overview. Research backgrounds 2.

More information

Subspace Gaussian Mixture Models for Automatic Speech Recognition. Liang Lu

Subspace Gaussian Mixture Models for Automatic Speech Recognition. Liang Lu Subspace Gaussian Mixture Models for Automatic Speech Recognition Liang Lu Doctor of Philosophy Institute for Language, Cognition and Computation School of Informatics University of Edinburgh 2013 Abstract

More information

The Noisy Channel Model. CS 294-5: Statistical Natural Language Processing. Speech Recognition Architecture. Digitizing Speech

The Noisy Channel Model. CS 294-5: Statistical Natural Language Processing. Speech Recognition Architecture. Digitizing Speech CS 294-5: Statistical Natural Language Processing The Noisy Channel Model Speech Recognition II Lecture 21: 11/29/05 Search through space of all possible sentences. Pick the one that is most probable given

More information

Estimation of Cepstral Coefficients for Robust Speech Recognition

Estimation of Cepstral Coefficients for Robust Speech Recognition Estimation of Cepstral Coefficients for Robust Speech Recognition by Kevin M. Indrebo, B.S., M.S. A Dissertation submitted to the Faculty of the Graduate School, Marquette University, in Partial Fulfillment

More information

Diagonal Priors for Full Covariance Speech Recognition

Diagonal Priors for Full Covariance Speech Recognition Diagonal Priors for Full Covariance Speech Recognition Peter Bell 1, Simon King 2 Centre for Speech Technology Research, University of Edinburgh Informatics Forum, 10 Crichton St, Edinburgh, EH8 9AB, UK

More information

SPEECH recognition systems based on hidden Markov

SPEECH recognition systems based on hidden Markov IEEE SIGNAL PROCESSING LETTERS, VOL. X, NO. X, 2014 1 Probabilistic Linear Discriminant Analysis for Acoustic Modelling Liang Lu, Member, IEEE and Steve Renals, Fellow, IEEE Abstract In this letter, we

More information

Making Machines Understand Us in Reverberant Rooms [Robustness against reverberation for automatic speech recognition]

Making Machines Understand Us in Reverberant Rooms [Robustness against reverberation for automatic speech recognition] Making Machines Understand Us in Reverberant Rooms [Robustness against reverberation for automatic speech recognition] Yoshioka, T., Sehr A., Delcroix M., Kinoshita K., Maas R., Nakatani T., Kellermann

More information

SELECTIVE ANGLE MEASUREMENTS FOR A 3D-AOA INSTRUMENTAL VARIABLE TMA ALGORITHM

SELECTIVE ANGLE MEASUREMENTS FOR A 3D-AOA INSTRUMENTAL VARIABLE TMA ALGORITHM SELECTIVE ANGLE MEASUREMENTS FOR A 3D-AOA INSTRUMENTAL VARIABLE TMA ALGORITHM Kutluyıl Doğançay Reza Arablouei School of Engineering, University of South Australia, Mawson Lakes, SA 595, Australia ABSTRACT

More information

Support Vector Machines using GMM Supervectors for Speaker Verification

Support Vector Machines using GMM Supervectors for Speaker Verification 1 Support Vector Machines using GMM Supervectors for Speaker Verification W. M. Campbell, D. E. Sturim, D. A. Reynolds MIT Lincoln Laboratory 244 Wood Street Lexington, MA 02420 Corresponding author e-mail:

More information

Model-based Approaches to Robust Speech Recognition in Diverse Environments

Model-based Approaches to Robust Speech Recognition in Diverse Environments Model-based Approaches to Robust Speech Recognition in Diverse Environments Yongqiang Wang Darwin College Engineering Department Cambridge University October 2015 This dissertation is submitted to the

More information

Recognition of Convolutive Speech Mixtures by Missing Feature Techniques for ICA

Recognition of Convolutive Speech Mixtures by Missing Feature Techniques for ICA Recognition of Convolutive Speech Mixtures by Missing Feature Techniques for ICA Dorothea Kolossa, Hiroshi Sawada, Ramon Fernandez Astudillo, Reinhold Orglmeister and Shoji Makino Electronics and Medical

More information

A Generative Model Based Kernel for SVM Classification in Multimedia Applications

A Generative Model Based Kernel for SVM Classification in Multimedia Applications Appears in Neural Information Processing Systems, Vancouver, Canada, 2003. A Generative Model Based Kernel for SVM Classification in Multimedia Applications Pedro J. Moreno Purdy P. Ho Hewlett-Packard

More information

ASPEAKER independent speech recognition system has to

ASPEAKER independent speech recognition system has to 930 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 13, NO. 5, SEPTEMBER 2005 Vocal Tract Normalization Equals Linear Transformation in Cepstral Space Michael Pitz and Hermann Ney, Member, IEEE

More information

Infinite Support Vector Machines in Speech Recognition

Infinite Support Vector Machines in Speech Recognition Infinite Support Vector Machines in Speech Recognition Jingzhou Yang, Rogier C. van Dalen and Mark Gales Department of Engineering, University of Cambridge Trumpington Street, Cambridge, CB2 1PZ, UK {jy308,rcv25,mjfg}@eng.cam.ac.uk

More information

Discriminative models for speech recognition

Discriminative models for speech recognition Discriminative models for speech recognition Anton Ragni Peterhouse University of Cambridge A thesis submitted for the degree of Doctor of Philosophy 2013 Declaration This dissertation is the result of

More information

GAUSSIAN PROCESS REGRESSION

GAUSSIAN PROCESS REGRESSION GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The

More information

Automatic Speech Recognition (CS753)

Automatic Speech Recognition (CS753) Automatic Speech Recognition (CS753) Lecture 12: Acoustic Feature Extraction for ASR Instructor: Preethi Jyothi Feb 13, 2017 Speech Signal Analysis Generate discrete samples A frame Need to focus on short

More information

Robust Speaker Identification

Robust Speaker Identification Robust Speaker Identification by Smarajit Bose Interdisciplinary Statistical Research Unit Indian Statistical Institute, Kolkata Joint work with Amita Pal and Ayanendranath Basu Overview } } } } } } }

More information

Lecture 5: GMM Acoustic Modeling and Feature Extraction

Lecture 5: GMM Acoustic Modeling and Feature Extraction CS 224S / LINGUIST 285 Spoken Language Processing Andrew Maas Stanford University Spring 2017 Lecture 5: GMM Acoustic Modeling and Feature Extraction Original slides by Dan Jurafsky Outline for Today Acoustic

More information

Confidence Estimation Methods for Neural Networks: A Practical Comparison

Confidence Estimation Methods for Neural Networks: A Practical Comparison , 6-8 000, Confidence Estimation Methods for : A Practical Comparison G. Papadopoulos, P.J. Edwards, A.F. Murray Department of Electronics and Electrical Engineering, University of Edinburgh Abstract.

More information

Usually the estimation of the partition function is intractable and it becomes exponentially hard when the complexity of the model increases. However,

Usually the estimation of the partition function is intractable and it becomes exponentially hard when the complexity of the model increases. However, Odyssey 2012 The Speaker and Language Recognition Workshop 25-28 June 2012, Singapore First attempt of Boltzmann Machines for Speaker Verification Mohammed Senoussaoui 1,2, Najim Dehak 3, Patrick Kenny

More information

Uncertainty Modeling without Subspace Methods for Text-Dependent Speaker Recognition

Uncertainty Modeling without Subspace Methods for Text-Dependent Speaker Recognition Uncertainty Modeling without Subspace Methods for Text-Dependent Speaker Recognition Patrick Kenny, Themos Stafylakis, Md. Jahangir Alam and Marcel Kockmann Odyssey Speaker and Language Recognition Workshop

More information

Noise Classification based on PCA. Nattanun Thatphithakkul, Boontee Kruatrachue, Chai Wutiwiwatchai, Vataya Boonpiam

Noise Classification based on PCA. Nattanun Thatphithakkul, Boontee Kruatrachue, Chai Wutiwiwatchai, Vataya Boonpiam Noise Classification based on PCA Nattanun Thatphithakkul, Boontee Kruatrachue, Chai Wutiwiwatchai, Vataya Boonpiam 1 Outline Introduction Principle component analysis (PCA) Classification using PCA Experiment

More information

Speech Enhancement with Applications in Speech Recognition

Speech Enhancement with Applications in Speech Recognition Speech Enhancement with Applications in Speech Recognition A First Year Report Submitted to the School of Computer Engineering of the Nanyang Technological University by Xiao Xiong for the Confirmation

More information

GWAS V: Gaussian processes

GWAS V: Gaussian processes GWAS V: Gaussian processes Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS V: Gaussian processes Summer 2011

More information

A Direct Criterion Minimization based fmllr via Gradient Descend

A Direct Criterion Minimization based fmllr via Gradient Descend A Direct Criterion Minimization based fmllr via Gradient Descend Jan Vaněk and Zbyněk Zajíc University of West Bohemia in Pilsen, Univerzitní 22, 306 14 Pilsen Faculty of Applied Sciences, Department of

More information

Hidden Markov Models and Gaussian Mixture Models

Hidden Markov Models and Gaussian Mixture Models Hidden Markov Models and Gaussian Mixture Models Hiroshi Shimodaira and Steve Renals Automatic Speech Recognition ASR Lectures 4&5 23&27 January 2014 ASR Lectures 4&5 Hidden Markov Models and Gaussian

More information

A TWO-LAYER NON-NEGATIVE MATRIX FACTORIZATION MODEL FOR VOCABULARY DISCOVERY. MengSun,HugoVanhamme

A TWO-LAYER NON-NEGATIVE MATRIX FACTORIZATION MODEL FOR VOCABULARY DISCOVERY. MengSun,HugoVanhamme A TWO-LAYER NON-NEGATIVE MATRIX FACTORIZATION MODEL FOR VOCABULARY DISCOVERY MengSun,HugoVanhamme Department of Electrical Engineering-ESAT, Katholieke Universiteit Leuven, Kasteelpark Arenberg 10, Bus

More information

Detection-Based Speech Recognition with Sparse Point Process Models

Detection-Based Speech Recognition with Sparse Point Process Models Detection-Based Speech Recognition with Sparse Point Process Models Aren Jansen Partha Niyogi Human Language Technology Center of Excellence Departments of Computer Science and Statistics ICASSP 2010 Dallas,

More information

Pattern Classification

Pattern Classification Pattern Classification Introduction Parametric classifiers Semi-parametric classifiers Dimensionality reduction Significance testing 6345 Automatic Speech Recognition Semi-Parametric Classifiers 1 Semi-Parametric

More information

Dual Estimation and the Unscented Transformation

Dual Estimation and the Unscented Transformation Dual Estimation and the Unscented Transformation Eric A. Wan ericwan@ece.ogi.edu Rudolph van der Merwe rudmerwe@ece.ogi.edu Alex T. Nelson atnelson@ece.ogi.edu Oregon Graduate Institute of Science & Technology

More information

Single and Multi Channel Feature Enhancement for Distant Speech Recognition

Single and Multi Channel Feature Enhancement for Distant Speech Recognition Single and Multi Channel Feature Enhancement for Distant Speech Recognition John McDonough (1), Matthias Wölfel (2), Friedrich Faubel (3) (1) (2) (3) Saarland University Spoken Language Systems Overview

More information

Session Variability Compensation in Automatic Speaker Recognition

Session Variability Compensation in Automatic Speaker Recognition Session Variability Compensation in Automatic Speaker Recognition Javier González Domínguez VII Jornadas MAVIR Universidad Autónoma de Madrid November 2012 Outline 1. The Inter-session Variability Problem

More information

Hidden Markov Modelling

Hidden Markov Modelling Hidden Markov Modelling Introduction Problem formulation Forward-Backward algorithm Viterbi search Baum-Welch parameter estimation Other considerations Multiple observation sequences Phone-based models

More information

Comparing linear and non-linear transformation of speech

Comparing linear and non-linear transformation of speech Comparing linear and non-linear transformation of speech Larbi Mesbahi, Vincent Barreaud and Olivier Boeffard IRISA / ENSSAT - University of Rennes 1 6, rue de Kerampont, Lannion, France {lmesbahi, vincent.barreaud,

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression Group Prof. Daniel Cremers 9. Gaussian Processes - Regression Repetition: Regularized Regression Before, we solved for w using the pseudoinverse. But: we can kernelize this problem as well! First step:

More information

FEATURE PRUNING IN LIKELIHOOD EVALUATION OF HMM-BASED SPEECH RECOGNITION. Xiao Li and Jeff Bilmes

FEATURE PRUNING IN LIKELIHOOD EVALUATION OF HMM-BASED SPEECH RECOGNITION. Xiao Li and Jeff Bilmes FEATURE PRUNING IN LIKELIHOOD EVALUATION OF HMM-BASED SPEECH RECOGNITION Xiao Li and Jeff Bilmes Department of Electrical Engineering University. of Washington, Seattle {lixiao, bilmes}@ee.washington.edu

More information

EE 565: Position, Navigation, and Timing

EE 565: Position, Navigation, and Timing EE 565: Position, Navigation, and Timing Kalman Filtering Example Aly El-Osery Kevin Wedeward Electrical Engineering Department, New Mexico Tech Socorro, New Mexico, USA In Collaboration with Stephen Bruder

More information

Segmental Recurrent Neural Networks for End-to-end Speech Recognition

Segmental Recurrent Neural Networks for End-to-end Speech Recognition Segmental Recurrent Neural Networks for End-to-end Speech Recognition Liang Lu, Lingpeng Kong, Chris Dyer, Noah Smith and Steve Renals TTI-Chicago, UoE, CMU and UW 9 September 2016 Background A new wave

More information

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression Group Prof. Daniel Cremers 4. Gaussian Processes - Regression Definition (Rep.) Definition: A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution.

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning 12. Gaussian Processes Alex Smola Carnegie Mellon University http://alex.smola.org/teaching/cmu2013-10-701 10-701 The Normal Distribution http://www.gaussianprocess.org/gpml/chapters/

More information

Discriminant Feature Space Transformations for Automatic Speech Recognition

Discriminant Feature Space Transformations for Automatic Speech Recognition Discriminant Feature Space Transformations for Automatic Speech Recognition Vikrant Tomar McGill ID: 260394445 Department of Electrical & Computer Engineering McGill University Montreal, Canada February

More information

Speech and Language Processing. Chapter 9 of SLP Automatic Speech Recognition (II)

Speech and Language Processing. Chapter 9 of SLP Automatic Speech Recognition (II) Speech and Language Processing Chapter 9 of SLP Automatic Speech Recognition (II) Outline for ASR ASR Architecture The Noisy Channel Model Five easy pieces of an ASR system 1) Language Model 2) Lexicon/Pronunciation

More information

Automatic Speech Recognition (CS753)

Automatic Speech Recognition (CS753) Automatic Speech Recognition (CS753) Lecture 21: Speaker Adaptation Instructor: Preethi Jyothi Oct 23, 2017 Speaker variations Major cause of variability in speech is the differences between speakers Speaking

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

Expectation propagation for signal detection in flat-fading channels

Expectation propagation for signal detection in flat-fading channels Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA

More information

Graphical models for part of speech tagging

Graphical models for part of speech tagging Indian Institute of Technology, Bombay and Research Division, India Research Lab Graphical models for part of speech tagging Different Models for POS tagging HMM Maximum Entropy Markov Models Conditional

More information

Lecture 3: ASR: HMMs, Forward, Viterbi

Lecture 3: ASR: HMMs, Forward, Viterbi Original slides by Dan Jurafsky CS 224S / LINGUIST 285 Spoken Language Processing Andrew Maas Stanford University Spring 2017 Lecture 3: ASR: HMMs, Forward, Viterbi Fun informative read on phonetics The

More information

R E S E A R C H R E P O R T Entropy-based multi-stream combination Hemant Misra a Hervé Bourlard a b Vivek Tyagi a IDIAP RR 02-24 IDIAP Dalle Molle Institute for Perceptual Artificial Intelligence ffl

More information

Lecture 10. Discriminative Training, ROVER, and Consensus. Michael Picheny, Bhuvana Ramabhadran, Stanley F. Chen

Lecture 10. Discriminative Training, ROVER, and Consensus. Michael Picheny, Bhuvana Ramabhadran, Stanley F. Chen Lecture 10 Discriminative Training, ROVER, and Consensus Michael Picheny, Bhuvana Ramabhadran, Stanley F. Chen IBM T.J. Watson Research Center Yorktown Heights, New York, USA {picheny,bhuvana,stanchen}@us.ibm.com

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

Linear Regression and Discrimination

Linear Regression and Discrimination Linear Regression and Discrimination Kernel-based Learning Methods Christian Igel Institut für Neuroinformatik Ruhr-Universität Bochum, Germany http://www.neuroinformatik.rub.de July 16, 2009 Christian

More information

NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group

NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION M. Schwab, P. Noll, and T. Sikora Technical University Berlin, Germany Communication System Group Einsteinufer 17, 1557 Berlin (Germany) {schwab noll

More information

MVA Processing of Speech Features. Chia-Ping Chen, Jeff Bilmes

MVA Processing of Speech Features. Chia-Ping Chen, Jeff Bilmes MVA Processing of Speech Features Chia-Ping Chen, Jeff Bilmes {chiaping,bilmes}@ee.washington.edu SSLI Lab Dept of EE, University of Washington Seattle, WA - UW Electrical Engineering UWEE Technical Report

More information

A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement

A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement Simon Leglaive 1 Laurent Girin 1,2 Radu Horaud 1 1: Inria Grenoble Rhône-Alpes 2: Univ. Grenoble Alpes, Grenoble INP,

More information

Model-Based Margin Estimation for Hidden Markov Model Learning and Generalization

Model-Based Margin Estimation for Hidden Markov Model Learning and Generalization 1 2 3 4 5 6 7 8 Model-Based Margin Estimation for Hidden Markov Model Learning and Generalization Sabato Marco Siniscalchi a,, Jinyu Li b, Chin-Hui Lee c a Faculty of Engineering and Architecture, Kore

More information

Gaussian Processes. 1 What problems can be solved by Gaussian Processes?

Gaussian Processes. 1 What problems can be solved by Gaussian Processes? Statistical Techniques in Robotics (16-831, F1) Lecture#19 (Wednesday November 16) Gaussian Processes Lecturer: Drew Bagnell Scribe:Yamuna Krishnamurthy 1 1 What problems can be solved by Gaussian Processes?

More information

Biologically Plausible Speech Recognition with LSTM Neural Nets

Biologically Plausible Speech Recognition with LSTM Neural Nets Biologically Plausible Speech Recognition with LSTM Neural Nets Alex Graves, Douglas Eck, Nicole Beringer, Juergen Schmidhuber Istituto Dalle Molle di Studi sull Intelligenza Artificiale, Galleria 2, 6928

More information

Estimating Covariance Using Factorial Hidden Markov Models

Estimating Covariance Using Factorial Hidden Markov Models Estimating Covariance Using Factorial Hidden Markov Models João Sedoc 1,2 with: Jordan Rodu 3, Lyle Ungar 1, Dean Foster 1 and Jean Gallier 1 1 University of Pennsylvania Philadelphia, PA joao@cis.upenn.edu

More information

Genones: Generalized Mixture Tying in Continuous. V. Digalakis P. Monaco H. Murveit. December 6, 1995 EDICS SA and

Genones: Generalized Mixture Tying in Continuous. V. Digalakis P. Monaco H. Murveit. December 6, 1995 EDICS SA and Genones: Generalized Mixture Tying in Continuous Hidden Markov Model-Based Speech Recognizers V. Digalakis P. Monaco H. Murveit Technical University of Crete Nuance Communications December 6, 1995 EDICS

More information

Necessary Corrections in Intransitive Likelihood-Ratio Classifiers

Necessary Corrections in Intransitive Likelihood-Ratio Classifiers Necessary Corrections in Intransitive Likelihood-Ratio Classifiers Gang Ji and Jeff Bilmes SSLI-Lab, Department of Electrical Engineering University of Washington Seattle, WA 9895-500 {gang,bilmes}@ee.washington.edu

More information

Why DNN Works for Acoustic Modeling in Speech Recognition?

Why DNN Works for Acoustic Modeling in Speech Recognition? Why DNN Works for Acoustic Modeling in Speech Recognition? Prof. Hui Jiang Department of Computer Science and Engineering York University, Toronto, Ont. M3J 1P3, CANADA Joint work with Y. Bao, J. Pan,

More information

Environmental and Speaker Robustness in Automatic Speech Recognition with Limited Learning Data

Environmental and Speaker Robustness in Automatic Speech Recognition with Limited Learning Data University of California Los Angeles Environmental and Speaker Robustness in Automatic Speech Recognition with Limited Learning Data A dissertation submitted in partial satisfaction of the requirements

More information

SPEECH RECOGNITION USING TIME DOMAIN FEATURES FROM PHASE SPACE RECONSTRUCTIONS

SPEECH RECOGNITION USING TIME DOMAIN FEATURES FROM PHASE SPACE RECONSTRUCTIONS SPEECH RECOGNITION USING TIME DOMAIN FEATURES FROM PHASE SPACE RECONSTRUCTIONS by Jinjin Ye, B.S. A THESIS SUBMITTED TO THE FACULTY OF THE GRADUATE SCHOOL IN PARTIAL FULFILLMENT OF THE REQUIREMENTS for

More information

Environmental Sound Classification in Realistic Situations

Environmental Sound Classification in Realistic Situations Environmental Sound Classification in Realistic Situations K. Haddad, W. Song Brüel & Kjær Sound and Vibration Measurement A/S, Skodsborgvej 307, 2850 Nærum, Denmark. X. Valero La Salle, Universistat Ramon

More information

An Asynchronous Hidden Markov Model for Audio-Visual Speech Recognition

An Asynchronous Hidden Markov Model for Audio-Visual Speech Recognition An Asynchronous Hidden Markov Model for Audio-Visual Speech Recognition Samy Bengio Dalle Molle Institute for Perceptual Artificial Intelligence (IDIAP) CP 592, rue du Simplon 4, 1920 Martigny, Switzerland

More information

CLOSE-TO-CLEAN REGULARIZATION RELATES

CLOSE-TO-CLEAN REGULARIZATION RELATES Worshop trac - ICLR 016 CLOSE-TO-CLEAN REGULARIZATION RELATES VIRTUAL ADVERSARIAL TRAINING, LADDER NETWORKS AND OTHERS Mudassar Abbas, Jyri Kivinen, Tapani Raio Department of Computer Science, School of

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 89 Part II

More information

TinySR. Peter Schmidt-Nielsen. August 27, 2014

TinySR. Peter Schmidt-Nielsen. August 27, 2014 TinySR Peter Schmidt-Nielsen August 27, 2014 Abstract TinySR is a light weight real-time small vocabulary speech recognizer written entirely in portable C. The library fits in a single file (plus header),

More information

Eigenvoice Speaker Adaptation via Composite Kernel PCA

Eigenvoice Speaker Adaptation via Composite Kernel PCA Eigenvoice Speaker Adaptation via Composite Kernel PCA James T. Kwok, Brian Mak and Simon Ho Department of Computer Science Hong Kong University of Science and Technology Clear Water Bay, Hong Kong [jamesk,mak,csho]@cs.ust.hk

More information

The Noisy Channel Model. Statistical NLP Spring Mel Freq. Cepstral Coefficients. Frame Extraction ... Lecture 10: Acoustic Models

The Noisy Channel Model. Statistical NLP Spring Mel Freq. Cepstral Coefficients. Frame Extraction ... Lecture 10: Acoustic Models Statistical NLP Spring 2009 The Noisy Channel Model Lecture 10: Acoustic Models Dan Klein UC Berkeley Search through space of all possible sentences. Pick the one that is most probable given the waveform.

More information

Statistical NLP Spring The Noisy Channel Model

Statistical NLP Spring The Noisy Channel Model Statistical NLP Spring 2009 Lecture 10: Acoustic Models Dan Klein UC Berkeley The Noisy Channel Model Search through space of all possible sentences. Pick the one that is most probable given the waveform.

More information

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer. University of Cambridge Engineering Part IIB & EIST Part II Paper I0: Advanced Pattern Processing Handouts 4 & 5: Multi-Layer Perceptron: Introduction and Training x y (x) Inputs x 2 y (x) 2 Outputs x

More information

Feature-Space Structural MAPLR with Regression Tree-based Multiple Transformation Matrices for DNN

Feature-Space Structural MAPLR with Regression Tree-based Multiple Transformation Matrices for DNN MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Feature-Space Structural MAPLR with Regression Tree-based Multiple Transformation Matrices for DNN Kanagawa, H.; Tachioka, Y.; Watanabe, S.;

More information

Automatic Speech Recognition and Statistical Machine Translation under Uncertainty

Automatic Speech Recognition and Statistical Machine Translation under Uncertainty Outlines Automatic Speech Recognition and Statistical Machine Translation under Uncertainty Lambert Mathias Advisor: Prof. William Byrne Thesis Committee: Prof. Gerard Meyer, Prof. Trac Tran and Prof.

More information

Uncertainty training and decoding methods of deep neural networks based on stochastic representation of enhanced features

Uncertainty training and decoding methods of deep neural networks based on stochastic representation of enhanced features MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Uncertainty training and decoding methods of deep neural networks based on stochastic representation of enhanced features Tachioka, Y.; Watanabe,

More information

Maximum Likelihood and Maximum A Posteriori Adaptation for Distributed Speaker Recognition Systems

Maximum Likelihood and Maximum A Posteriori Adaptation for Distributed Speaker Recognition Systems Maximum Likelihood and Maximum A Posteriori Adaptation for Distributed Speaker Recognition Systems Chin-Hung Sit 1, Man-Wai Mak 1, and Sun-Yuan Kung 2 1 Center for Multimedia Signal Processing Dept. of

More information

Gaussian Process Approximations of Stochastic Differential Equations

Gaussian Process Approximations of Stochastic Differential Equations Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Dan Cawford Manfred Opper John Shawe-Taylor May, 2006 1 Introduction Some of the most complex models routinely run

More information

Improved noise power spectral density tracking by a MAP-based postprocessor

Improved noise power spectral density tracking by a MAP-based postprocessor Improved noise power spectral density tracking by a MAP-based postprocessor Aleksej Chinaev, Alexander Krueger, Dang Hai Tran Vu, Reinhold Haeb-Umbach University of Paderborn, Germany March 8th, 01 Computer

More information

Gaussian Processes (10/16/13)

Gaussian Processes (10/16/13) STA561: Probabilistic machine learning Gaussian Processes (10/16/13) Lecturer: Barbara Engelhardt Scribes: Changwei Hu, Di Jin, Mengdi Wang 1 Introduction In supervised learning, we observe some inputs

More information