Noise Classification based on PCA. Nattanun Thatphithakkul, Boontee Kruatrachue, Chai Wutiwiwatchai, Vataya Boonpiam

Similar documents
Heeyoul (Henry) Choi. Dept. of Computer Science Texas A&M University

Introduction to Machine Learning

Modifying Voice Activity Detection in Low SNR by correction factors

Holdout and Cross-Validation Methods Overfitting Avoidance

SPEECH ENHANCEMENT USING PCA AND VARIANCE OF THE RECONSTRUCTION ERROR IN DISTRIBUTED SPEECH RECOGNITION

Detection-Based Speech Recognition with Sparse Point Process Models

Kernel Principal Component Analysis

Robust Sound Event Detection in Continuous Audio Environments

Introduction to Support Vector Machines

Full-covariance model compensation for

Eigenvoice Speaker Adaptation via Composite Kernel PCA

Convergence of Eigenspaces in Kernel Principal Component Analysis

DYNAMIC TEXTURE RECOGNITION USING ENHANCED LBP FEATURES

Applied Statistics. Multivariate Analysis - part II. Troels C. Petersen (NBI) Statistics is merely a quantization of common sense 1

Gopalkrishna Veni. Project 4 (Active Shape Models)

Tone Analysis in Harmonic-Frequency Domain and Feature Reduction using KLT+LVQ for Thai Isolated Word Recognition

Bearing fault diagnosis based on EMD-KPCA and ELM

L5 Support Vector Classification

Time Series Classification

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014

Joint Factor Analysis for Speaker Verification

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet.

Analysis of audio intercepts: Can we identify and locate the speaker?

Kernel Methods. Machine Learning A W VO

Deriving Principal Component Analysis (PCA)


PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Eigenface-based facial recognition

Manifold Learning: Theory and Applications to HRI

Covariance and Correlation Matrix

PATTERN CLASSIFICATION

Session Variability Compensation in Automatic Speaker Recognition

CSC242: Intro to AI. Lecture 21

Learning SVM Classifiers with Indefinite Kernels

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

Introduction to SVM and RVM

Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Announcements (repeat) Principal Components Analysis

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH

The effect of speaking rate and vowel context on the perception of consonants. in babble noise

Support Vector Machine. Industrial AI Lab.

Learning Multiple Tasks with a Sparse Matrix-Normal Penalty

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Machine learning for pervasive systems Classification in high-dimensional spaces

Discriminative Learning and Big Data

Chemometrics: Classification of spectra

THE PROBLEMS OF ROBUST LPC PARAMETRIZATION FOR. Petr Pollak & Pavel Sovka. Czech Technical University of Prague

Analysis of Spectral Kernel Design based Semi-supervised Learning

What is semi-supervised learning?

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Source localization in an ocean waveguide using supervised machine learning

Subspace Analysis for Facial Image Recognition: A Comparative Study. Yongbin Zhang, Lixin Lang and Onur Hamsici

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Principal Component Analysis

A Coupled Helmholtz Machine for PCA

Noise Compensation for Subspace Gaussian Mixture Models

Recurrence Enhances the Spatial Encoding of Static Inputs in Reservoir Networks

Support Vector Machine (continued)

Improved noise power spectral density tracking by a MAP-based postprocessor

Principal Components Analysis (PCA)

PHONEME CLASSIFICATION OVER THE RECONSTRUCTED PHASE SPACE USING PRINCIPAL COMPONENT ANALYSIS

Face Recognition and Biometric Systems

CHAPTER 3 FEATURE EXTRACTION USING GENETIC ALGORITHM BASED PRINCIPAL COMPONENT ANALYSIS

Convex Optimization in Classification Problems

PCA, Kernel PCA, ICA

Correspondence. Pulse Doppler Radar Target Recognition using a Two-Stage SVM Procedure

Modeling Classes of Shapes Suppose you have a class of shapes with a range of variations: System 2 Overview

Lecture 4: Feed Forward Neural Networks

2D Spectrogram Filter for Single Channel Speech Enhancement

Unsupervised Learning of Hierarchical Models. in collaboration with Josh Susskind and Vlad Mnih

SYMMETRIC MATRIX PERTURBATION FOR DIFFERENTIALLY-PRIVATE PRINCIPAL COMPONENT ANALYSIS. Hafiz Imtiaz and Anand D. Sarwate

Advanced Introduction to Machine Learning CMU-10715

Model-Based Margin Estimation for Hidden Markov Model Learning and Generalization

Investigate more robust features for Speech Recognition using Deep Learning

Principal Component Analysis (PCA)

When Dictionary Learning Meets Classification

PREDICTING SOLAR GENERATION FROM WEATHER FORECASTS. Chenlin Wu Yuhan Lou

Model-Based Approaches to Robust Speech Recognition

Statistical Machine Learning

Principal Component Analysis

Predicting Graph Labels using Perceptron. Shuang Song

Support Vector Machine. Industrial AI Lab. Prof. Seungchul Lee

Modifying A Linear Support Vector Machine for Microarray Data Classification

WHILE most current speech recognizers give acceptable

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

A Kernel on Persistence Diagrams for Machine Learning

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

EPL442: Computational

Machine Learning 2nd Edition

Adversarial Label Flips Attack on Support Vector Machines

COMPLEX PRINCIPAL COMPONENT SPECTRA EXTRACTION

Kernel Methods. Barnabás Póczos

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection

Shankar Shivappa University of California, San Diego April 26, CSE 254 Seminar in learning algorithms

Overview of Statistical Tools. Statistical Inference. Bayesian Framework. Modeling. Very simple case. Things are usually more complicated

Introduction to Neural Networks

Principal Component Analysis

Notes on Latent Semantic Analysis

Gender Classification in Speech Processing. Baiju M Nair(2011CRF3637) Geetanjali Srivastava(2012EEZ8304) Group No. 22

Transcription:

Noise Classification based on PCA Nattanun Thatphithakkul, Boontee Kruatrachue, Chai Wutiwiwatchai, Vataya Boonpiam 1

Outline Introduction Principle component analysis (PCA) Classification using PCA Experiment Conclusion 2

Introduction Problem One way to make a robust speech recognizer is to train by noisy speech. However, with various kinds of noise, the recognizer cannot achieve an acceptable accuracy. Solution To classify noise and select an acoustic model generated for that noise. 3

Introduction Propose Method Principal Component Analysis (PCA) is used to reduce the dimension of analyzed feature. A new classification structure is introduced. 4

PCA Principal Component Analysis Step 1 Calculate the covariance matrix M μ= 1 M i =1 x i (1) φ j =x j μ (2) M C= 1 M j =1 φ j φ j T (3) 5

PCA Principal Component Analysis Step 2 Calculate the eigenvectors and eigenvalues of the covariance matrix and select N eigenvectors, which correspond to the largest eigenvalue. Step 3 Calculate the weight vector w k =v k T x μ k =1,...,N (4) W T =[w 1 w 2...w N ] (5) 6

Classification using PCA PCA used for reducing dimension of feature vector before classification. Feature Extraction Speech PCA2 Classification2 clean white pink babble car Structure of Classical Classification using PCA (baseline) 7

Classification using PCA Weight vector of principle components (Train -> 5 types) 8

Classification using PCA Weight vector of principle components (Train -> 4 types) 9

Classification using PCA Compare weights between train 5 types and train 4 types 10

Classification using PCA Speech Feature Extraction PCA1 Classification1 clean noise PCA2 Classification2 white pink babble car Structure of Propose Classification using PCA 11

Classification using PCA Weight vector of principle components (Train -> 2 types) 12

Experiment Speech Data Clean environment (from NECTEC-ATR thai corpus) Noise environment (from NOISEX-92 database) white pink babble car Classified Techniques MFCC/LPC/LPCEPSTRA+HMM Norm-Log-Spectrum+PCA (1 step)+nn (nn_baseline) Norm-Log-Spectrum+PCA (2 step)+nn (nn_propose) Norm-Log-Spectrum+PCA (1 step)+svm (svm_baseline) Norm-Log-Spectrum+PCA (2 step)+svm (svm_propose) 13

Experiment Experiment data Process Type Num of File Training set for PCA Training set for classification Testing set for classification Clean 1,000 Original Noise 4,000 Total 5,000 Clean 12,306 Original Noise 306 Addition Noise 48,918 Total 61,530 Clean 6,400 Original Noise 102,400 Total 108,800 Remark : Clean = Clean Speech Original Noise = Noise Speech Addition Noise = Clean + Noise Speech ( SNR = 15 db 0 db ) 14

Experiment Noise Classification with HMM Parameter Error rate(%) MFCC_E_D_A, GM=1 5.41 MFCC_E_D_A, GM=8 4.49 MFCC_E_D_A, GM=16 4.18 LPCEPSTRA_E_D_A, GM=1 5.80 LPCEPSTRA_E_D_A, GM=8 5.41 LPCEPSTRA_E_D_A, GM=16 5.29 LPC_E_D_A, GM=1 6.94 LPC_E_D_A, GM=8 7.01 LPC_E_D_A, GM=16 6.09 Remark : _E = Energy _D = Delta coeffcients _A = Acceleration coefficients 15

Experiment Noise Classification with NN Num of hidden node Type Error Rate (%) 30 40 50 100 Baseline 4.55 Propose 1.20 Baseline 4.37 Propose 1.44 Baseline 4.33 Propose 1.24 Baseline 4.60 Propose 1.42 Remark : Baseline = Classical Classification using PCA ( 1 step ) Propose = Propose Classification using PCA ( 2 step ) 16

Experiment Noise Classification with SVM (RBF kernel) x y 2 k x, y =exp a 2 Parameter Type Error Rate (%) a = 0.01 a = 0.1 a = 0.5 a = 1 Remark : Baseline = Classical Classification using PCA ( 1 step ) Propose = Propose Classification using PCA ( 2 step ) Baseline 3.99 Propose 1.68 Baseline 3.97 Propose 1.71 Baseline 3.86 Propose 1.57 Baseline 3.78 Propose 1.64 17

Experiment Noise Classification with SVM (Polynimial kernel) k x, y = x. y p Parameter Type Error Rate (%) p = 1 p = 2 p = 3 Baseline 4.01 Propose 1.90 Baseline 19.43 Propose 6.62 Baseline 38.55 Propose 27.69 Remark : Baseline = Classical Classification using PCA ( 1 step ) Propose = Propose Classification using PCA ( 2 step ) 18

Experiment Noise Classification with SVM (Sigmoid kernel) k x, y =tanhkx. y Parameter Type Error Rate (%) k = 0.25 k = 0.5 k = 1 k = 2 Remark : Baseline = Classical Classification using PCA ( 1 step ) Propose = Propose Classification using PCA ( 2 step ) Baseline 4.02 Propose 1.92 Baseline 4.05 Propose 1.91 Baseline 4.35 Propose 1.70 Baseline 3.98 Propose 2.22 19

Experiment Summary of Classification Result 4.5 4 4.18 4.33 3.78 3.5 3 2.5 2 1.5 1 1.20 1.57 0.5 0 HMM NN_baseline SVM_baseline NN_propose SVM_propose 20

Conclusion This paper present a new method for noise classification used by projection to each data of the principle components. The new classification structure that seperates clean speech from noisy speech and then classify the type of noise gives a lower error rate comparing to the classical structure that classifies in one step 21

Acknowledgments The authors would like to thank Dr.Sanparith Marukatat for his comments and suggestion. 22

Question & Answer 23