Markov-Switching GARCH Models and Applications to Digital Processing of Speech Signals
|
|
- Marvin Marcus Perkins
- 6 years ago
- Views:
Transcription
1 Markov-Switching GARCH Models and Applications to Digital Processing of Speech Signals Electrical Engineering Department Technion - Israel Institute of Technology Supervised by: Prof. Israel Cohen
2 Outline Introduction - Spectral Enhancement for Speech Enhancement Markov-Switching GARCH Model
3 Spectral Enhancement Existing Algorithms Speech Enhancement - Main Goal Given a noisy observation of speech signal, y = x + d find the best estimate for the speech signal ˆx
4 Spectral Enhancement Existing Algorithms Short-Time Fourier Transform (STFT) Since the signal is highly non-stationary, STFT is applied to yield a time-frequency representation: and we get Y tk = X tk + D tk clean speech noisy speech, SNR=5 db
5 Spectral Enhancement Spectral Enhancement Existing Algorithms The expansion coefficients are sparse in the STFT domain H1 tk Y tk = X tk + D tk H0 tk (speech absent) : Y tk = D tk and the spectral enhancement problem can be formulated as { min E d (X tk, ˆX ) tk ˆqtk, ˆλ tk, ˆλ } d,tk, Y tk ˆX tk
6 Spectral Enhancement Spectral Enhancement Existing Algorithms The expansion coefficients are sparse in the STFT domain H1 tk (speech present) : Y tk = X tk + D tk H0 tk Y tk = D tk and the spectral enhancement problem can be formulated as { ) } min E d (X tk, ˆX tk ˆqtk, ˆλ tk, ˆλ d,tk, Y tk ˆX tk d (X tk, ˆX ) tk = g(x tk ) g(ˆx tk ) 2 - distortion measure q tk = P ( H1 tk ) - speech presence probability λ tk = E { X tk 2 H1 tk } - speech spectral variance λ d,tk = E { D tk 2} - noise spectral variance
7 Spectral Enhancement (cont.) Spectral Enhancement Existing Algorithms A spectral enhancement system requires: Estimation / Detection method: and Fidelity criterion, i.e., g(x) = X, X, log X A detector for speech spectral coefficients / speech presence probability Spectral model: Statistical models for {X tk } and {D tk } Spectral variance estimators ˆλ tk and ˆλ d,tk
8 Spectral Models Spectral Enhancement Existing Algorithms Decision-directed spectral variance estimator { 2 ( ˆλ tk = max µ ˆX t 1,k + (1 µ) Y tk 2 ˆλ ) } d,tk, λ min 0 µ 1 [Ephraim & Malah 84] Hidden-Markov Model A vector model: Hidden Markov chain with mixtures of AR processes. [Ephraim et al. 89, 92]
9 Spectral Models Spectral Enhancement Existing Algorithms Decision-directed spectral variance estimator { 2 ( ˆλ tk = max µ ˆX t 1,k + (1 µ) Y tk 2 ˆλ ) } d,tk, λ min 0 µ 1 [Ephraim & Malah 84] Hidden-Markov Model A vector model: Hidden Markov chain with mixtures of AR processes. [Ephraim et al. 89, 92]
10 Existing Algorithms Spectral Enhancement Existing Algorithms Spectral subtraction: detection (power thresholding) and independent estimation (power subtraction). Subspace approach: detection followed by estimation. MMSE, STSA, OM-LSA (and other): estimation under uncertainty (e.g., ˆX tk = E{X tk Y tk, H1 tk}p(htk 1 Y tk)). Drawbacks: Estimation under uncertainty degrades speech quality and prevent sufficient attenuation of noise-only coefficients compared to using an ideal detector. Erroneous detection may remove desired speech components or result in an annoying musical noise.
11 Existing Algorithms Spectral Enhancement Existing Algorithms Spectral subtraction: detection (power thresholding) and independent estimation (power subtraction). Subspace approach: detection followed by estimation. MMSE, STSA, OM-LSA (and other): estimation under uncertainty (e.g., ˆX tk = E{X tk Y tk, H1 tk}p(htk 1 Y tk)). Drawbacks: Estimation under uncertainty degrades speech quality and prevent sufficient attenuation of noise-only coefficients compared to using an ideal detector. Erroneous detection may remove desired speech components or result in an annoying musical noise.
12 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing GARCH: Generalized Autoregressive Conditional Heteroscedasticity
13 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Generalized Autoregressive Conditional Heteroscedasticity (GARCH) Model Robert F. Engle Nobel price (2003) in economic sciences for methods of analyzing economic time series with time-varying volatility (ARCH)
14 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Generalized Autoregressive Conditional Heteroscedasticity (GARCH) Model GARCH models [Engle 82; Bollerslev 86] are widely used in various financial applications such as risk management, option pricing, and foreign exchange. Explicitly parameterize the time-varying volatility in terms of past conditional variances and past squared innovations, while taking into account excess kurtosis and volatility clustering. Modeling speech expansion coefficients as GARCH processes offers reasonable model on which to base the variance estimation [Cohen 04, 06]
15 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Generalized Autoregressive Conditional Heteroscedasticity (GARCH) Model GARCH models [Engle 82; Bollerslev 86] are widely used in various financial applications such as risk management, option pricing, and foreign exchange. Explicitly parameterize the time-varying volatility in terms of past conditional variances and past squared innovations, while taking into account excess kurtosis and volatility clustering. Modeling speech expansion coefficients as GARCH processes offers reasonable model on which to base the variance estimation [Cohen 04, 06]
16 GARCH Model (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing 4.2 Deutschmark/British Pound Foreign Exchange Rate Exchange Rate Jan 1984 Jan 1986 Jan 1988 Jan Deutschmark/British Pound Daily Returns 0.25 Real value of speech spectral coefficients at 3 khz Return Jan 1984 Jan 1986 Jan 1988 Jan
17 Speech Spectral Analysis GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing When observing a time series of successive expansion coefficients in a fixed frequency bin, successive magnitudes of the expansion coefficients are highly correlated, whereas successive phases are nearly uncorrelated. Speech signals in the STFT domain are characterized by volatility clustering and heavy-tailed distribution. Real value of speech spectral coefficients at 3 khz
18 GARCH Model (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Let {y t } denote a real-valued discrete-time stochastic process, and let ψ t denote the information set available at time t. The innovation (prediction error) ε t at time t in the MMSE sense is given by ε t = y t E {y t ψ t 1 } The conditional variance (volatility) of y t is defined by σ 2 t = Var {y t ψ t 1 } = E { ε 2 t ψ t 1}.
19 GARCH Model (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing A linear GARCH model of order (p,q), ε t GARCH(p,q): ε t = σ t v t q p σt 2 = ξ + α i ε 2 t i + β j σt j 2 i=1 j=1 {v t } is a zero-mean unit-variance white noise process. To ensure positive conditional variances: ξ > 0, α i 0, β j 0, i = 1,...,q, j = 1,...,p, and for the existence of a finite unconditional variance: q p α i + β j < 1 i=1 j=1
20 GARCH Model (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing A linear GARCH model of order (p,q), ε t GARCH(p,q): ε t = σ t v t q p σt 2 = ξ + α i ε 2 t i + β j σt j 2 i=1 j=1 {v t } is a zero-mean unit-variance white noise process. To ensure positive conditional variances: ξ > 0, α i 0, β j 0, i = 1,...,q, j = 1,...,p, and for the existence of a finite unconditional variance: q p α i + β j < 1 i=1 j=1
21 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Markov-Switching GARCH Model In a Markov-switching GARCH model, the parameters are allowed to change in time according to a Markovian state to enable tracking of any shocks. Multi-state models are used for speech recognition and for speech enhancement [Drucker 68, Ephraim et al. 89, 92]. In case of speech signals the parameters may change in time, e.g., for representing different speech phonemes or speakers. A special state may be defined for speech absence hypothesis. This may results in a soft voice activity detector.
22 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Markov-Switching GARCH Model S t - an m-state, first-order hidden Markov chain with transition probabilities matrix A. A general (p,q)-order MS-GARCH model follows: Given that S t = s t : ε t = σ t,st v t σt,s 2 t q p = ξ st + α i,st ε 2 t i + β j,st σt j,s 2 t j i=1 j=1
23 Asymptotic Stationarity GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing For a single-state model: i α i + j β j < 1 is necessary and sufficient to have an asymptotic stationary variance and to guarantee a finite second-order moment [Bollerslev 86]. Clearly, for a MS-GARCH model: i α i,s + j β j,s < 1 s is a sufficient condition. Sufficient and necessary conditions for stationarity are known only for some degenerated cases.
24 Asymptotic Stationarity (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing The unconditional variance follows: E [ ε 2 ] [ ( )] t = E ψt 1,S t E ε 2 t ψ t 1,s t [ ( )] m ( ) = E St Eψt 1 σ 2 t,st s t = π st E ψt 1 σ 2 t,st s t where π st p (S t = s t ) s t=1
25 Asymptotic Stationarity (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Define: Ψ K (1) K (2) K (r) I m 0 m 0 m 0 m I m where 0 m 0 I m 0 m { K (i)} (α i,s + β i,s ) π s { A i, s, s = 1,...,m s, s π } s,s s and r max{p,q}.
26 Asymptotic Stationarity (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Let Λ be an m m square matrix with {Λ} ij = i,j = 1,...,m. Then: { (I Ψ) 1} ij, Theorem The process is asymptotically wide-sense stationary with asymptotic variance lim t E ( ε 2 t) = π Λξ, if and only if ρ(ψ) < 1 [Abramson and Cohen, Econometric Theory 07] Some regimes may allow volatility to grow over time and still the second-order moment will be finite. Similar method was applied to other MS-GARCH formulations.
27 Asymptotic Stationarity (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Let Λ be an m m square matrix with {Λ} ij = i,j = 1,...,m. Then: { (I Ψ) 1} ij, Theorem The process is asymptotically wide-sense stationary with asymptotic variance lim t E ( ε 2 t) = π Λξ, if and only if ρ(ψ) < 1 [Abramson and Cohen, Econometric Theory 07] Some regimes may allow volatility to grow over time and still the second-order moment will be finite. Similar method was applied to other MS-GARCH formulations.
28 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Back to Speech: Spectral Modeling of Speech Signals Using MS-GARCH
29 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing MS-GARCH Model in the Time-Frequency Domain A frame (or subband) dependent Markov chain S t with realization s t 0,1,...,m X tk MS GARCH(1,1): X tk = λ tk t 1,st V tk λ tk t 1,st = λ min,st + α st X t 1,k 2 ( ) + β st λt 1,k t 2,st 1 λ min,st 1 with λ min,st > 0, α st, β st 0. {V tk } are iid complex random variables with zero mean and unit variance. {X tk } are statistically dependent but, { X tk λ tk t 1,st, s t } are zero-mean statistically independent random variables.
30 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing MS-GARCH Model in the Time-Frequency Domain A frame (or subband) dependent Markov chain S t with realization s t 0,1,...,m X tk MS GARCH(1,1): X tk = λ tk t 1,st V tk λ tk t 1,st = λ min,st + α st X t 1,k 2 ( ) + β st λt 1,k t 2,st 1 λ min,st 1 with λ min,st > 0, α st, β st 0. {V tk } are iid complex random variables with zero mean and unit variance. {X tk } are statistically dependent but, { X tk λ tk t 1,st, s t } are zero-mean statistically independent random variables.
31 Model Estimation GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Parameters are traditionally estimated from a training set using ML approach. Alternative, parameters may be chosen such that each state would represent a different level of the spectral energy. Setting α s = β s = 0 for a specific s results in a constant variance, hence, may be used for speech absence.
32 Model estimation (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing The conditional variances in each subband (indexed n) are limited within a dynamic range of η g db. For the speech absence state (namely, s t = 0): λ min,n,0 = 10 log 10 ζg ηg/10, α n,0 = β n,0 = 0. where ζ g max t,k X tk 2. Under speech presence, η t db is the local dynamic range of the conditional variances such that { λ min,n,1 = max λ min,n,0,10 log ζn ηt/10} 10, and λ min,n,s, s = 2,...,m are log-spaced between λ min,n,1 and ζ n max t,k κn X tk 2.
33 Model estimation (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing The parameters α n,s,β n,s for s > 0 set the volatility level of the conditional variance. Under an immutable state s, the stationary variance follows 1 β n,s λ,n,s lim λ tk t 1,s = λ min,n,s t,k κ n 1 α n,s β n,s assuming α n,s + β n,s < 1. Different states are related to different dynamic ranges in ascending order, thus λ,n,s λ min,n,s+1 and therefore 1 β n,s 1 α n,s β n,s λ min,n,s+1 λ min,n,s.
34 Relation to HMM GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing In a standard HMM: p (X t X t 1,X t 2,..., s t,s t 1,s t 2,...) = p (X t s t ) each state specifies a specific pdf. In a MS-GARCH process both past observations and the regime-path are required for the conditional density. each state formulates the evolution of the conditional variance.
35 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Reconstruction of the Conditional Variance In a noiseless environment (econometric) the conditional variance is recursively reconstructed. In our case, the conditional variance is estimated using two steps: [Abramson and Cohen, Trans. SP 07] Propagation } ˆλ tk t 1,st =λ min,st + α st E { X t 1,k 2 Y t 1, s t + β st E { λ t 1,k t 2,St 1 Y t 1 }, s t βst E { λ min,st 1 Y t 1 }, s t } E { X t 1,k 2 Y t 1, s t = p ( s t 1 s t, Y t 1) } E { X t 1,k 2 Y t 1, s t 1 s t 1
36 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Reconstruction of the Conditional Variance (cont.) and Update ˆλ tk t,st =E { X tk 2 st, ˆλ } tk t 1,st,Y t [ ˆλ tk t 1,st = ˆλ tk t 1,st + σtk 2 ˆλ tk t 1,st σtk 2 + ˆλ tk t 1,st + σtk 2 Y tk 2 ]
37 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Relation to decision directed estimation Recall the decision-directed estimation: ˆλ DD tk = { 2 ) } max µ ˆX t 1,k + (1 µ) ( Y tk 2 σtk 2, ξ min σtk 2,0 µ 1 For an ARCH(1) model (i.e., β = 0) with α = 1 the update step can be written as [Abramson & Cohen, Trans. SP] ˆλ tk t = } ( ) µ tk E { X t 1,k 2 Y t 1 + (1 µ tk ) Y tk 2 σtk 2 + µ tk λ min with µ tk 1 ˆλ 2 tk t 1 (ˆλ tk t 1 +σ 2 tk) 2.
38 GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Relation to decision directed estimation (cont.) } For λ min << E { X t 1,k 2 Y t 1, the degenerated ARCH-based variance estimation with α = 1 is similar to the decision-directed approach with µ µ tk 2 } ˆX t 1,k E { X t 1,k 2 Y t 1 GARCH modeling allows α < 1 and β > 0 and it refers to the spectral variances as a random process rather than a sequence of random parameters which are heuristically evaluated.
39 State Probability GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing The conditional state probability is derived by p ( s t Y t) p (Y t s t, ˆλ t t 1,st )p ( s t Y t 1) = s t p (Y t s t, ˆλ ), t t 1,st p (s t Y t 1 ) where p ( s t Y t 1) = s t 1 p ( s t 1 Y t 1) a st 1,s t. Accordingly, 1 p (s t = 0 Y t ) is a soft voice activity detector for the subband. The problem of state smoothing, p(s t Y t+l ) with L > 0, is also addressed [Abramson and Cohen, SPL 06].
40 State smoothing GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing State smoothing, i.e., p (s t Y τ ) for τ > t in HMP relies on [Chang and Hancock 66,Lindgren 78, Askar and Derin 81] f (X t ψ t 1,s t ) = f (X t s t ) In an AR-HMP a finite set of past observation is required [Kim 94]. In a path-dependent MS-GARCH process the density is a function of all past values and active states. We derive generalization of both the forward-backward recursions and the stable backward recursion to path dependent model [Abramson and Cohen, IEEE Signal processing letters 06].
41 State smoothing (cont.) GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Define } ˆΛ t {ˆλt t 1,St S t = 1,...,m the generalized forward density α ( s t, Y t) f (s t, ˆΛ t,y t ) and the generalized backward density ( β Y t+l t+l St t+l 1, Y t+l 1) ( ) f Y t+l t+l St t+l 1, ˆΛ t, Yt t+l 1 then the noncausal state probability can be obtained by ( ) ( p s t Y t+l) ( = p s t ˆΛ ) α (s t, Y t ) β Y t+l t, Yt t+l t+1 s t, Y t = ( ) s t α(s t, Y t )β Yt+1 t+l s t, Y t
42 Generalized forward recursion GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing The generalized forward density satisfies the following recursion: α ( { s t, Y t) πs0 ( f (Y 0 s 0 ) ) t = 0 = f Y t s t, ˆλ t t,st s t 1 α ( s t 1, Y t 1) a st 1s t t = 1, 2...
43 Generalized backward recursion GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing The generalized backward density requires two recursive steps: Step I: ˆλt+l t+l 1,S t+l t ˆλ t+l t+l,s t+l t for l = 1,..., L, for all St t+l : = ξ st+l 1 + α st+l ˆλt+l 1 t+l 1,S t+l 1 t = g ˆλt+l t+l 1,S t t+l,y t+l, σ 2 Step II: for l = L,..., 1, for all St t+l : f Y t+l t+l St t+l, ˆλ t t 1,st, Yt t+l 1 = β Y t+l β Y t+l t+l St t+l 1, Y t+l 1 = È s t+l f Y t+l t+l St t+l t+l+1 St+l t + β st+1ˆλt+l 1 t+l 2,S t+l 1 t, Y t+l f Y t+l St t+l, ˆλ t t 1,st, Yt t+l 1, ˆλ t t 1,st, Yt t+l 1 a st+l 1 s t+l
44 State smoothing - results GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing The generalized recursions capture the non-memoryless structure of the Markov-switching GARCH model. Each step of the generalized backward recursion is calculated for m L+1 regime sequences Error rate Delay L (samples) State smoothing error rate for 3-state MSTF-GARCH models with SNRs of 5 db (triangle), 10 db (asterisk) and 15 db (circle).
45 Spectral enhancement GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Minimizing yields E { g (X tk ) g(ˆx tk ) 2 Y t} ) g (ˆXtk = E { g (X tk ) Y t} where E { g (X tk ) Y t} = s t p ( S t = s t Y t) E { g (X tk ) s t, Y t}
46 Spectral Enhancement GARCH Formulation Markov-switching GARCH model Asymptotic Stationarity Spectral Modeling State smoothing Having the set {ˆλtk t,st } st=m s t=0, minimizing the mean-square error of the log-spectral amplitude (LSA) { ( ) } 2 E log X tk log ˆX tk Y t yields: ˆX tk = Y tk s t G LSA (ˆξtk,st, ˆγ tk,st ) p(st Y t ), where G LSA (ξ,ϑ) is the LSA gain function [Ephraim and Malah 85].
47 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals Simultaneous Detection and Estimation
48 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals Recall Find an estimate ˆX tk which minimizes { g ) } 2 E (Xtk ) g (ˆX tk ψt where g (X) and g (X) are specific functions, e.g., X, X or log X. Alternatively, Bayesian formulation: ) C (X argminˆxtk tk, ˆX tk p (Y tk X tk ) p(x tk )dx tk where C X, ˆX ˆX 2 = g (X) g.
49 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals Recall Find an estimate ˆX tk which minimizes { g ) } 2 E (Xtk ) g (ˆX tk ψt where g (X) and g (X) are specific functions, e.g., X, X or log X. Alternatively, Bayesian formulation: ) C (X argminˆxtk tk, ˆX tk p (Y tk X tk ) p(x tk )dx tk where C X, ˆX ˆX 2 = g (X) g.
50 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals (cont.) Independent detection and estimation: Simultaneous detection and estimation:
51 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals (cont.) Guidelines [Middleton et al. 68, 72]: Integrated detector with a decision space { η0 tk, } ηtk 1. Under ηj tk, Hj tk is accepted and ˆX tk = ˆX tk,j is considered. Cost for the hypothesis-decision pair {H i,η j }: C ij (X, ˆX ) ( = b ij d ij X, ˆX ) d ij (X, ˆX ) is an appropriate distortion measure between the signal and its estimate under η j. b ij is a trade-off parameter of the cost associated with the pair {H i, η j } against other pairs.
52 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals (cont.) For the signal-observation {X,Y} the cost associated with the decision η j is C j (X, Y ) = = 1 i=0 1 i=0 C ij (X, ˆX ) p (H i X) p (H i )C ij (X, ˆX ) p (X H i )/p (X) The Combined Bayes risk of simultaneous detection and estimation: R = 1 j=0 Y X C j (X, Y )p (η j Y )p (Y X)p (X) dxdy
53 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals (cont.) The goal: Find the optimal detector and estimator which minimize the combined Bayes risk R = 1 j=0 Y p (η j Y ) 1 i=0 X C ij (X, ˆX ) p (H i )p (Y X)p (X H i ) dxdy
54 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals (cont.) Combined solution: Optimal estimator under a decision η j : Optimal decision rule: ˆX j = arg min ˆX 1 p (H i ) r ij (Y ) i=0 q [r 10 (Y ) r 11 (Y )] η 1 (1 q)[r 01 (Y ) r 00 (Y )] η 0 r ij (Y ) C ij (X, ˆX ) p (Y X) p (X H i )dx q p(h 1 )
55 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals (cont.) Recall C ij (X, ˆX ) ( = b ij d ij X, ˆX ) Normalize b 00 = b 11 = 1. b 10 is associated with missed-detection, b 01 is associated with false-detection. d ii ( ) is not necessarily zero (estimation error). Under H 0, a natural background noise level is desired.
56 Distortion Functions: Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals Quadratic distortion measure: ( ) X d ij X, ˆX ˆX 2 j j = 2 G f Y ˆX j i = 1 i = 0 Quadratic spectral amplitude (QSA) distortion measure: ( ) ( ) 2 X ˆX j i = 1 d ij X, ˆX j = ( ) 2 G f Y ˆX j i = 0 G f << 1 is a constant attenuation floor.
57 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals (cont.) Definitions: a priori and a posteriori SNRs ξ λ λ d, γ Generalized likelihood ratio: Y 2 λ d, υ ξγ 1 + ξ. Λ (ξ,γ) = q p (Y H 1 ) (1 q) p (Y H 0 ). φ j (ξ,γ) b 0j + b 1j Λ (ξ,γ).
58 QSA distortion measure: Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals Under a Gaussian model: [Abramson and Cohen, Trans. ASLP 07] Estimator ˆX j = b 1jΛ (ξ, γ)g STSA (ξ, γ) + b 0j G f φ j (ξ, γ) Y G j (ξ, γ)y, j = 0, 1 Detector q e υ b 10G0 2 G ξ where G STSA = 2υ γ η 1 η 0 (1 q) exp υ ξ (1 + ξ) γ (1 + υ)(b10 1) + 2(G1 b10g0) G STSA b 01 (G 1 G f ) 2 (G 0 G f ) 2 2 (1 + υ υ)i0 2 + υ υi1 2 [Ephraim and Malah 84]
59 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals of Speech Signals: The estimator ˆX tk = G j Y tk is optimal for any given detector.
60 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals of Speech Signals: The estimator ˆX tk = G j Y tk is optimal for any given detector.
61 QSA distortion measure: ξ = 15 db Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals 6 7 ξ = 5 db Gain [db] G 1 Gain [db] Gain [db] G 0 28 G (detection & estimation) G STSA (p<1) Instantaneous SNR (γ 1) [db] ξ = 5 db Gain [db] Instantaneous SNR (γ 1) [db] ξ = 15 db Instantaneous SNR (γ 1) [db] q = 0.8, G f = 20 db and b 10 = 1.1, b 01 = Instantaneous SNR (γ 1) [db]
62 Asymptotic behavior Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals If the false alarm parameter b 01 << Λ(ξ,γ) we get G 1 (ξ,γ) G STSA (ξ,γ). If b 01 >> Λ(ξ,γ) (wrong decision might be expensive) = G 1 (ξ,γ) G f. If the missed-detection parameter b 10 << Λ(ξ,γ) 1 then G 0 (ξ,γ) G f. But b 10 >> Λ(ξ,γ) 1 = G 0 (ξ,γ) G STSA (ξ,γ) to overcome the high cost for missed-detection.
63 Relation to classical algorithms Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals Λ(ξ,γ) Recall that 1+Λ(ξ,γ) = p (H 1 Y ) is the a posteriori speech presence probability. For equal cost parameters b ij = 1 i,j we have ˆX 1 = ˆX 0 = [p (H 1 Y )G STSA (ξ,γ) + (1 p (H 1 Y )) G f ]Y this is the estimator proposed by [Cohen 06]. In case G f = 0 we get the STSA suppression rule [Ephraim and Malah 84].
64 Guidelines Spectral Enhancement - Reformulation Detection and Estimation of Speech Signals Design of coupled detector and estimator with cost parameters which control the tradeoff between musical residual noise and speech distortion. This method enables to incorporate an independent detector for speech or transient noise with the optimal estimator to yield a suboptimal solution. Low computational complexity.
65 Blind Source Separation
66 Problem formulation Given a single mixture x = s 1 + s 2 find estimates ŝ 1 and ŝ 2. Since the problem is il-posed, some a priori information is required such as statistical codebook. In the STFT domain, at some specific time-index we have x = s 1 + s 2
67 Problem formulation (cont.) q 1 {1,...,m 1 } and q 2 {1,...,m 2 } denote the active states of the codebooks with known a priori probabilities p 1 (i) p (q 1 = i) and p 2 (i) p (q 2 = j). ( Given that q 1 = i and q 2 = j we assume s 1 CN ( ) and s 2 CN. 0,Σ (j) 2 0,Σ (i) 1 Hence, we need to find the active states at each time-frame and subsequently estimate each signal. )
68 Existing solutions Sequential classification and estimation: MAP classification: {î,ĵ } MMSE estimation: ŝ 1 = E { } s 1 x,î,ĵ = Σ (î) 1 = arg max p (x i,j) p (i,j) i,j ( Σ (î) 1 + Σ (ĵ) ) 1 2 x Wîĵ x Classification error may significantly distort the signal or result in residual interference.
69 Existing solutions (cont.) Direct estimation: ŝ 1 = E {s 1 x} = i,j p (i,j x) W ij x.
70 Simultaneous classification and estimation Consider a classifier η ij and a combined cost function Cī j ij (s,ŝ) bī j ij s ŝ 2 2. bī j ij > 0 are cost parameters for a decision that {i,j} is the active pair while q 1 = ī and q 2 = j. The combined risk: R = Cī j ij (s 1,ŝ 1 )p ( x s 1,ī, j ) p ( s 1 ī, j ) p ( ī, j ) p (η ij x) ds 1 dx i,j ī, j
71 Simultaneous classification and estimation (cont.) Using rī j ij (x,ŝ 1 ) = Cī j ij (s 1,ŝ 1 )p ( x s 1, j ) p ( s 1 ī ) ds 1 we can write: R = i,j p (η ij x) ī, j p ( ī, j ) rī j ij (x,ŝ 1 )dx and the combined solution is obtained by minimizing: min {R} η ij,ŝ 1
72 Optimal solution The optimal combined solution: The estimator under the decision η ij ŝ 1,ij = arg min ŝ 1 = ī, j p ( ī, j ) rī j ij (x,ŝ 1 ) ī j bī j ij p ( x ī, j ) p ( ī, j ) Wī j x ī j bī j ij p ( x ī, j ) p ( ī, j ) G ij x (1)
73 Optimal solution (cont.) The optimal classifier is obtained by min ij ī, j p ( ī, j ) rī j ij (x,ŝ 1 ) where rī j ij (x,ŝ 1 ) = bī j ij p ( x ī, j ) [ ( ) x H W 2 ī j 2W ī j G ij x + 1 T Σ ( j) ] 2 W ī j 1
74 Parameter selection A specific state is defined for signal absence, e.g., q = 0. We may choose bī j ij = bīi b j j and bīi = b 1,m i = 0, ī 0 b 1,f i 0, ī = 0 1 o.w. with relation to missed and false detection.
75 Joint classification and estimation Now we consider a cascade classification and estimation: η x = s 1 + s ij 2 Classifier Estimator ŝ 1,ij Under signal-presence decision we wish to control the residual interference and under signal-absence decision we limit the distortion level.
76 Joint classification and estimation (cont.) Under signal-presence decision (η 1j ): { } ŝ 1,1j = arg min p (q 1 = 1 x) E s 1 ŝ q 1 = 1,x ŝ 1 s.t. ε2 r (x) σ 2 r where ε 2 r (x) p (q 1 = 0 x) E { } s 1 ŝ q 1 = 0,x
77 Joint classification and estimation (cont.) Under signal-absence decision (η 0j ): { } ŝ 1,0j = arg min p (q 1 = 0 x) E s 1 ŝ q 1 = 0,x ŝ 1 s.t. ε2 d (x) σd 2, where { } ε 2 d (x) p (q 1 = 1 x) E s 1 ŝ q 1 = 1,x
78 Joint classification and estimation (cont.) The Lagrangian is given by (for η 0j ): { } ( ) L d (ŝ 1,µ d ) = p (q 1 = 0 x) E s 1 ŝ q 1 = 0,x +µ d ε2 d (x) σd 2 and ( ) µ d ε2 d (x) σd 2 = 0 for µ d 0 Similarly, under η 1j we have L r (ŝ 1,µ r ) with µ r.
79 Joint classification and estimation (cont.) By joining the solutions for both cases we obtain: ŝ 1,ij = ī j µīi p ( ī, j x ) Wī j x ī j µīi p ( ī, j x ) with µīi = µ d i = 0, ī = 1 µ r i = 1, ī = 0 1 o.w. which is similar to the solution for the simultaneous classification and estimation with b j j = 1, b 1,m = µ d, and b 1,f = µ r.
80 GARCH-based codebook
81 Block diagram
82 Distortion vs. interference reduction
83
84 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation
85 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Signal estimation - synthetic signals 3-state model 5-state model SNR improvement (db) theoretical limit MSTF GARCH, true par. MSTF GARCH, m=1 MSTF GARCH, m=3 MSTF GARCH, m=5 constant variance SNR improvement (db) theoretical limit MSTF GARCH, true par. MSTF GARCH, m=1 MSTF GARCH, m=3 MSTF GARCH, m=5 MSTF GARCH, m= Input SNR (db) Input SNR (db)
86 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Estimated conditional variances at various frequencies: Magnitude (db) Magnitude (db) Magnitude (db) khz Time [sec] khz Time [sec] khz Time [sec] Solid line: Signal s squared absolute value Dashed-dotted line: single-state model Dotted line: 3-state model Dashed line: 5-state model
87 Variance estimation (cont.) GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Estimated squared absolute values at frequency of 2 khz: 20 SNR= 10 db Magnitude (db) Time [sec] 20 SNR= 0 db Solid line: Speech signal s squared absolute value. Dashed-dotted line: 5-state MSTF-GARCH model. Dotted line: Decision-directed approach. Magnitude (db) Time [sec]
88 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Speech Enhancement - MS-GARCH Modeling 8 7 Clean signal Speech corrupted by a factory noise with 5 db SNR (LSD= 6.68 db, SegSNR= 0.05 db) 8 7 Frequency [khz] Frequency [khz] Amplitude Time [Sec] Speech reconstructed by using a 4-state model (LSD= 3.14 db, SegSNR= 6.76 db) 8 Amplitude Time [Sec] Comparison with DD Reverberated speech with 20 db SNR 19 Frequency [khz] SegSNR [db] Decision directed MS GARCH Amplitude Time [Sec] LSD [db] T [sec] 60
89 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Keyboard-Typing Noise, Detection vs. Estimation 8 Clean signal 8 Noisy signal with 0.8 db SNR (LSD=7.69 db, SegSNR=-2.23 db) Frequency [khz] Frequency [khz] Amplitude Time [Sec] Amplitude Time [Sec] 8 OM-LSA (LSD=6.77 db, SegSNR=-1.31 db) 8 Detection and estimation (LSD=2.52 db, SegSNR=4.41 db) Frequency [khz] Frequency [khz] Amplitude Time [Sec] Amplitude Time [Sec]
90 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Car Environment with a Siren, Detection vs. Estimation 8 Noisy signal with 3 db SNR OM-LSA Clean signal (LSD=6.56 db, SegSNR=-5.95 db) (LSD=4.609 db, SegSNR=-0.81 db) Frequency [khz] Frequency [khz] Frequency [khz] Amplitude Time [Sec] Amplitude Time [Sec] Amplitude Time [Sec] STSA with siren noise detector (LSD=4.37 db, SegSNR= db) Sim. detection and estimation (LSD=3.14 db, SegSNR=6.50 db) Frequency [khz] Frequency [khz] Amplitude Time [Sec] Amplitude Time [Sec]
91 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Voice Activity Detection Using MS-GARCH Model 1 SNR= 15 db Speech presence probabillity MSTF GARCH DD Instantaneous SNR [db] Higher dynamic range for the speech presence probabilities. Much lower probabilities for low energy coefficients.
92 Speech Dereverberation GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Noisy and reverberant speech:
93 Speech Dereverberation GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation 4 Clean signal 4 Corrupted signal LSD=4.87 db SegSNR=5.85 db Frequency [khz] Frequency [khz] Amplitude Amplitude Time [Sec] DD-based algorithm LSD=1.99 db SegSNR=8.35 db Time [Sec] Proposed algorithm LSD=1.70 db SegSNR=9.01 db Frequency [khz] Frequency [khz] Amplitude Amplitude Time [Sec] Time [Sec] * Joint work with Habets and Gannot.
94 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Single-Channel Blind Source Separation GARCH modeling with Simultaneous Classification and Estimation: Clean and mixed signals Estimated signals GMM speech proposed music GMM mix proposed Time [sec] Time [sec]
95 Summary Introduction GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation A new formulation for the problem of speech enhancement based on a simultaneous detection and estimation approach. A novel statistical model for speech signals in the STFT domain. The statistical model and the detection and estimation approach may be applied to speech enhancement in highly nonstationary noise environments, speech dereverberation and single-sensor blind source separation. Future Research: Improve multichannel algorithms using GARCH modeling and detection and estimation approach. Multivariate GARCH model with cross-band correlation. Optimizing the cost function and the trade-off parameters in case of multi-hypotheses.
96 GARCH modeling Speech Enhancement Voice Activity Detection Speech Dereverberation Blind-Source Separation Thank you!
New Statistical Model for the Enhancement of Noisy Speech
New Statistical Model for the Enhancement of Noisy Speech Electrical Engineering Department Technion - Israel Institute of Technology February 22, 27 Outline Problem Formulation and Motivation 1 Problem
More informationModeling speech signals in the time frequency domain using GARCH
Signal Processing () 53 59 Fast communication Modeling speech signals in the time frequency domain using GARCH Israel Cohen Department of Electrical Engineering, Technion Israel Institute of Technology,
More informationOptimal Speech Enhancement Under Signal Presence Uncertainty Using Log-Spectral Amplitude Estimator
1 Optimal Speech Enhancement Under Signal Presence Uncertainty Using Log-Spectral Amplitude Estimator Israel Cohen Lamar Signal Processing Ltd. P.O.Box 573, Yokneam Ilit 20692, Israel E-mail: icohen@lamar.co.il
More informationModifying Voice Activity Detection in Low SNR by correction factors
Modifying Voice Activity Detection in Low SNR by correction factors H. Farsi, M. A. Mozaffarian, H.Rahmani Department of Electrical Engineering University of Birjand P.O. Box: +98-9775-376 IRAN hfarsi@birjand.ac.ir
More informationSINGLE-CHANNEL SPEECH PRESENCE PROBABILITY ESTIMATION USING INTER-FRAME AND INTER-BAND CORRELATIONS
204 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) SINGLE-CHANNEL SPEECH PRESENCE PROBABILITY ESTIMATION USING INTER-FRAME AND INTER-BAND CORRELATIONS Hajar Momeni,2,,
More informationA POSTERIORI SPEECH PRESENCE PROBABILITY ESTIMATION BASED ON AVERAGED OBSERVATIONS AND A SUPER-GAUSSIAN SPEECH MODEL
A POSTERIORI SPEECH PRESENCE PROBABILITY ESTIMATION BASED ON AVERAGED OBSERVATIONS AND A SUPER-GAUSSIAN SPEECH MODEL Balázs Fodor Institute for Communications Technology Technische Universität Braunschweig
More informationNOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group
NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION M. Schwab, P. Noll, and T. Sikora Technical University Berlin, Germany Communication System Group Einsteinufer 17, 1557 Berlin (Germany) {schwab noll
More informationImproved Speech Presence Probabilities Using HMM-Based Inference, with Applications to Speech Enhancement and ASR
Improved Speech Presence Probabilities Using HMM-Based Inference, with Applications to Speech Enhancement and ASR Bengt J. Borgström, Student Member, IEEE, and Abeer Alwan, IEEE Fellow Abstract This paper
More informationGARCH Models Estimation and Inference
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in
More informationMANY digital speech communication applications, e.g.,
406 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 2, FEBRUARY 2007 An MMSE Estimator for Speech Enhancement Under a Combined Stochastic Deterministic Speech Model Richard C.
More informationIndependent Component Analysis and Unsupervised Learning
Independent Component Analysis and Unsupervised Learning Jen-Tzung Chien National Cheng Kung University TABLE OF CONTENTS 1. Independent Component Analysis 2. Case Study I: Speech Recognition Independent
More informationIntroduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak
Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,
More informationCovariance smoothing and consistent Wiener filtering for artifact reduction in audio source separation
Covariance smoothing and consistent Wiener filtering for artifact reduction in audio source separation Emmanuel Vincent METISS Team Inria Rennes - Bretagne Atlantique E. Vincent (Inria) Artifact reduction
More informationKalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein
Kalman filtering and friends: Inference in time series models Herke van Hoof slides mostly by Michael Rubinstein Problem overview Goal Estimate most probable state at time k using measurement up to time
More informationGARCH Models. Eduardo Rossi University of Pavia. December Rossi GARCH Financial Econometrics / 50
GARCH Models Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 50 Outline 1 Stylized Facts ARCH model: definition 3 GARCH model 4 EGARCH 5 Asymmetric Models 6
More informationVolatility. Gerald P. Dwyer. February Clemson University
Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use
More informationGARCH processes probabilistic properties (Part 1)
GARCH processes probabilistic properties (Part 1) Alexander Lindner Centre of Mathematical Sciences Technical University of Munich D 85747 Garching Germany lindner@ma.tum.de http://www-m1.ma.tum.de/m4/pers/lindner/
More informationDigital Signal Processing
Digital Signal Processing 0 (010) 157 1578 Contents lists available at ScienceDirect Digital Signal Processing www.elsevier.com/locate/dsp Improved minima controlled recursive averaging technique using
More informationMultichannel Deconvolution of Layered Media Using MCMC methods
Multichannel Deconvolution of Layered Media Using MCMC methods Idan Ram Electrical Engineering Department Technion Israel Institute of Technology Supervisors: Prof. Israel Cohen and Prof. Shalom Raz OUTLINE.
More informationIndependent Component Analysis and Unsupervised Learning. Jen-Tzung Chien
Independent Component Analysis and Unsupervised Learning Jen-Tzung Chien TABLE OF CONTENTS 1. Independent Component Analysis 2. Case Study I: Speech Recognition Independent voices Nonparametric likelihood
More informationSystem Identification and Adaptive Filtering in the Short-Time Fourier Transform Domain
System Identification and Adaptive Filtering in the Short-Time Fourier Transform Domain Electrical Engineering Department Technion - Israel Institute of Technology Supervised by: Prof. Israel Cohen Outline
More informationECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications
ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty
More informationGARCH Models Estimation and Inference. Eduardo Rossi University of Pavia
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia Likelihood function The procedure most often used in estimating θ 0 in ARCH models involves the maximization of a likelihood function
More informationA SPEECH PRESENCE PROBABILITY ESTIMATOR BASED ON FIXED PRIORS AND A HEAVY-TAILED SPEECH MODEL
A SPEECH PRESENCE PROBABILITY ESTIMATOR BASED ON FIXED PRIORS AND A HEAVY-TAILED SPEECH MODEL Balázs Fodor Institute for Communications Technology Technische Universität Braunschweig 386 Braunschweig,
More informationNon-Stationary Noise Power Spectral Density Estimation Based on Regional Statistics
Non-Stationary Noise Power Spectral Density Estimation Based on Regional Statistics Xiaofei Li, Laurent Girin, Sharon Gannot, Radu Horaud To cite this version: Xiaofei Li, Laurent Girin, Sharon Gannot,
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationOverview of Single Channel Noise Suppression Algorithms
Overview of Single Channel Noise Suppression Algorithms Matías Zañartu Salas Post Doctoral Research Associate, Purdue University mzanartu@purdue.edu October 4, 2010 General notes Only single-channel speech
More informationHidden Markov models 1
Hidden Markov models 1 Outline Time and uncertainty Markov process Hidden Markov models Inference: filtering, prediction, smoothing Most likely explanation: Viterbi 2 Time and uncertainty The world changes;
More informationRevisiting linear and non-linear methodologies for time series prediction - application to ESTSP 08 competition data
Revisiting linear and non-linear methodologies for time series - application to ESTSP 08 competition data Madalina Olteanu Universite Paris 1 - SAMOS CES 90 Rue de Tolbiac, 75013 Paris - France Abstract.
More informationUniversity of Cambridge. MPhil in Computer Speech Text & Internet Technology. Module: Speech Processing II. Lecture 2: Hidden Markov Models I
University of Cambridge MPhil in Computer Speech Text & Internet Technology Module: Speech Processing II Lecture 2: Hidden Markov Models I o o o o o 1 2 3 4 T 1 b 2 () a 12 2 a 3 a 4 5 34 a 23 b () b ()
More informationLecture 4: State Estimation in Hidden Markov Models (cont.)
EE378A Statistical Signal Processing Lecture 4-04/13/2017 Lecture 4: State Estimation in Hidden Markov Models (cont.) Lecturer: Tsachy Weissman Scribe: David Wugofski In this lecture we build on previous
More informationGARCH Models Estimation and Inference
Università di Pavia GARCH Models Estimation and Inference Eduardo Rossi Likelihood function The procedure most often used in estimating θ 0 in ARCH models involves the maximization of a likelihood function
More information13. Estimation and Extensions in the ARCH model. MA6622, Ernesto Mordecki, CityU, HK, References for this Lecture:
13. Estimation and Extensions in the ARCH model MA6622, Ernesto Mordecki, CityU, HK, 2006. References for this Lecture: Robert F. Engle. GARCH 101: The Use of ARCH/GARCH Models in Applied Econometrics,
More information9 Multi-Model State Estimation
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 9 Multi-Model State
More informationOn Optimal Coding of Hidden Markov Sources
2014 Data Compression Conference On Optimal Coding of Hidden Markov Sources Mehdi Salehifar, Emrah Akyol, Kumar Viswanatha, and Kenneth Rose Department of Electrical and Computer Engineering University
More informationA Priori SNR Estimation Using a Generalized Decision Directed Approach
A Priori SNR Estimation Using a Generalized Decision Directed Approach Aleksej Chinaev, Reinhold Haeb-Umbach Department of Communications Engineering, Paderborn University, 3398 Paderborn, Germany {chinaev,haeb}@nt.uni-paderborn.de
More informationImproved noise power spectral density tracking by a MAP-based postprocessor
Improved noise power spectral density tracking by a MAP-based postprocessor Aleksej Chinaev, Alexander Krueger, Dang Hai Tran Vu, Reinhold Haeb-Umbach University of Paderborn, Germany March 8th, 01 Computer
More informationHidden Markov Models
Hidden Markov Models CI/CI(CS) UE, SS 2015 Christian Knoll Signal Processing and Speech Communication Laboratory Graz University of Technology June 23, 2015 CI/CI(CS) SS 2015 June 23, 2015 Slide 1/26 Content
More informationLecture 6: Univariate Volatility Modelling: ARCH and GARCH Models
Lecture 6: Univariate Volatility Modelling: ARCH and GARCH Models Prof. Massimo Guidolin 019 Financial Econometrics Winter/Spring 018 Overview ARCH models and their limitations Generalized ARCH models
More informationMultivariate GARCH models.
Multivariate GARCH models. Financial market volatility moves together over time across assets and markets. Recognizing this commonality through a multivariate modeling framework leads to obvious gains
More informationA Modified Baum Welch Algorithm for Hidden Markov Models with Multiple Observation Spaces
IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 9, NO. 4, MAY 2001 411 A Modified Baum Welch Algorithm for Hidden Markov Models with Multiple Observation Spaces Paul M. Baggenstoss, Member, IEEE
More informationLecture 11: Hidden Markov Models
Lecture 11: Hidden Markov Models Cognitive Systems - Machine Learning Cognitive Systems, Applied Computer Science, Bamberg University slides by Dr. Philip Jackson Centre for Vision, Speech & Signal Processing
More informationHidden Markov Models and Gaussian Mixture Models
Hidden Markov Models and Gaussian Mixture Models Hiroshi Shimodaira and Steve Renals Automatic Speech Recognition ASR Lectures 4&5 23&27 January 2014 ASR Lectures 4&5 Hidden Markov Models and Gaussian
More informationDetection of Anomalies in Texture Images using Multi-Resolution Features
Detection of Anomalies in Texture Images using Multi-Resolution Features Electrical Engineering Department Supervisor: Prof. Israel Cohen Outline Introduction 1 Introduction Anomaly Detection Texture Segmentation
More informationSoft-Output Trellis Waveform Coding
Soft-Output Trellis Waveform Coding Tariq Haddad and Abbas Yongaçoḡlu School of Information Technology and Engineering, University of Ottawa Ottawa, Ontario, K1N 6N5, Canada Fax: +1 (613) 562 5175 thaddad@site.uottawa.ca
More informationAcoustic MIMO Signal Processing
Yiteng Huang Jacob Benesty Jingdong Chen Acoustic MIMO Signal Processing With 71 Figures Ö Springer Contents 1 Introduction 1 1.1 Acoustic MIMO Signal Processing 1 1.2 Organization of the Book 4 Part I
More informationL11: Pattern recognition principles
L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction
More informationFinancial Econometrics
Financial Econometrics Nonlinear time series analysis Gerald P. Dwyer Trinity College, Dublin January 2016 Outline 1 Nonlinearity Does nonlinearity matter? Nonlinear models Tests for nonlinearity Forecasting
More informationDETECTION theory deals primarily with techniques for
ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for
More informationTIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.
TIME SERIES ANALYSIS Forecasting and Control Fifth Edition GEORGE E. P. BOX GWILYM M. JENKINS GREGORY C. REINSEL GRETA M. LJUNG Wiley CONTENTS PREFACE TO THE FIFTH EDITION PREFACE TO THE FOURTH EDITION
More informationIntroduction to ARMA and GARCH processes
Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More information2D Spectrogram Filter for Single Channel Speech Enhancement
Proceedings of the 7th WSEAS International Conference on Signal, Speech and Image Processing, Beijing, China, September 15-17, 007 89 D Spectrogram Filter for Single Channel Speech Enhancement HUIJUN DING,
More informationON SCALABLE CODING OF HIDDEN MARKOV SOURCES. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose
ON SCALABLE CODING OF HIDDEN MARKOV SOURCES Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose Department of Electrical and Computer Engineering University of California, Santa Barbara, CA, 93106
More information10. Hidden Markov Models (HMM) for Speech Processing. (some slides taken from Glass and Zue course)
10. Hidden Markov Models (HMM) for Speech Processing (some slides taken from Glass and Zue course) Definition of an HMM The HMM are powerful statistical methods to characterize the observed samples of
More informationEnhancement of Noisy Speech. State-of-the-Art and Perspectives
Enhancement of Noisy Speech State-of-the-Art and Perspectives Rainer Martin Institute of Communications Technology (IFN) Technical University of Braunschweig July, 2003 Applications of Noise Reduction
More informationHidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010
Hidden Markov Models Aarti Singh Slides courtesy: Eric Xing Machine Learning 10-701/15-781 Nov 8, 2010 i.i.d to sequential data So far we assumed independent, identically distributed data Sequential data
More informationEcon 423 Lecture Notes: Additional Topics in Time Series 1
Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes
More informationMarkov Chains and Hidden Markov Models
Chapter 1 Markov Chains and Hidden Markov Models In this chapter, we will introduce the concept of Markov chains, and show how Markov chains can be used to model signals using structures such as hidden
More informationA Brief Survey of Speech Enhancement 1
A Brief Survey of Speech Enhancement 1 Yariv Ephraim, Hanoch Lev-Ari and William J.J. Roberts 2 August 2, 2003 Abstract We present a brief overview of the speech enhancement problem for wide-band noise
More informationDynamic Multipath Estimation by Sequential Monte Carlo Methods
Dynamic Multipath Estimation by Sequential Monte Carlo Methods M. Lentmaier, B. Krach, P. Robertson, and T. Thiasiriphet German Aerospace Center (DLR) Slide 1 Outline Multipath problem and signal model
More informationChapter 2 Signal Processing at Receivers: Detection Theory
Chapter Signal Processing at Receivers: Detection Theory As an application of the statistical hypothesis testing, signal detection plays a key role in signal processing at receivers of wireless communication
More informationEEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1
EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle
More informationHidden Markov Models. AIMA Chapter 15, Sections 1 5. AIMA Chapter 15, Sections 1 5 1
Hidden Markov Models AIMA Chapter 15, Sections 1 5 AIMA Chapter 15, Sections 1 5 1 Consider a target tracking problem Time and uncertainty X t = set of unobservable state variables at time t e.g., Position
More informationNonlinear Time Series Modeling
Nonlinear Time Series Modeling Part II: Time Series Models in Finance Richard A. Davis Colorado State University (http://www.stat.colostate.edu/~rdavis/lectures) MaPhySto Workshop Copenhagen September
More informationTRINICON: A Versatile Framework for Multichannel Blind Signal Processing
TRINICON: A Versatile Framework for Multichannel Blind Signal Processing Herbert Buchner, Robert Aichner, Walter Kellermann {buchner,aichner,wk}@lnt.de Telecommunications Laboratory University of Erlangen-Nuremberg
More informationSupervised and Unsupervised Speech Enhancement Using Nonnegative Matrix Factorization
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING Supervised and Unsupervised Speech Enhancement Using Nonnegative Matrix Factorization Nasser Mohammadiha*, Student Member, IEEE, Paris Smaragdis,
More informationBrief Introduction of Machine Learning Techniques for Content Analysis
1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview
More informationA SPECTRAL SUBTRACTION RULE FOR REAL-TIME DSP IMPLEMENTATION OF NOISE REDUCTION IN SPEECH SIGNALS
Proc. of the 1 th Int. Conference on Digital Audio Effects (DAFx-9), Como, Italy, September 1-4, 9 A SPECTRAL SUBTRACTION RULE FOR REAL-TIME DSP IMPLEMENTATION OF NOISE REDUCTION IN SPEECH SIGNALS Matteo
More informationDetecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf
Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf Reading: Ch. 5 in Kay-II. (Part of) Ch. III.B in Poor. EE 527, Detection and Estimation Theory, # 5c Detecting Parametric Signals in Noise
More informationRobust Backtesting Tests for Value-at-Risk Models
Robust Backtesting Tests for Value-at-Risk Models Jose Olmo City University London (joint work with Juan Carlos Escanciano, Indiana University) Far East and South Asia Meeting of the Econometric Society
More informationUCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)
UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable
More informationEstimating Covariance Using Factorial Hidden Markov Models
Estimating Covariance Using Factorial Hidden Markov Models João Sedoc 1,2 with: Jordan Rodu 3, Lyle Ungar 1, Dean Foster 1 and Jean Gallier 1 1 University of Pennsylvania Philadelphia, PA joao@cis.upenn.edu
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7
More informationEconometric Forecasting
Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 1, 2014 Outline Introduction Model-free extrapolation Univariate time-series models Trend
More informationPROBABILISTIC REASONING OVER TIME
PROBABILISTIC REASONING OVER TIME In which we try to interpret the present, understand the past, and perhaps predict the future, even when very little is crystal clear. Outline Time and uncertainty Inference:
More informationMark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.
University of Cambridge Engineering Part IIB & EIST Part II Paper I0: Advanced Pattern Processing Handouts 4 & 5: Multi-Layer Perceptron: Introduction and Training x y (x) Inputs x 2 y (x) 2 Outputs x
More informationMachine Recognition of Sounds in Mixtures
Machine Recognition of Sounds in Mixtures Outline 1 2 3 4 Computational Auditory Scene Analysis Speech Recognition as Source Formation Sound Fragment Decoding Results & Conclusions Dan Ellis
More informationPractical conditions on Markov chains for weak convergence of tail empirical processes
Practical conditions on Markov chains for weak convergence of tail empirical processes Olivier Wintenberger University of Copenhagen and Paris VI Joint work with Rafa l Kulik and Philippe Soulier Toronto,
More information1584 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 6, AUGUST 2011
1584 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 6, AUGUST 2011 Transient Noise Reduction Using Nonlocal Diffusion Filters Ronen Talmon, Student Member, IEEE, Israel Cohen,
More informationRecent Advancements in Speech Enhancement
Recent Advancements in Speech Enhancement Yariv Ephraim and Israel Cohen 1 May 17, 2004 Abstract Speech enhancement is a long standing problem with numerous applications ranging from hearing aids, to coding
More informationA priori SNR estimation and noise estimation for speech enhancement
Yao et al. EURASIP Journal on Advances in Signal Processing (2016) 2016:101 DOI 10.1186/s13634-016-0398-z EURASIP Journal on Advances in Signal Processing RESEARCH A priori SNR estimation and noise estimation
More informationC.M. Liu Perceptual Signal Processing Lab College of Computer Science National Chiao-Tung University
Quantization C.M. Liu Perceptual Signal Processing Lab College of Computer Science National Chiao-Tung University http://www.csie.nctu.edu.tw/~cmliu/courses/compression/ Office: EC538 (03)5731877 cmliu@cs.nctu.edu.tw
More informationLEARNING DYNAMIC SYSTEMS: MARKOV MODELS
LEARNING DYNAMIC SYSTEMS: MARKOV MODELS Markov Process and Markov Chains Hidden Markov Models Kalman Filters Types of dynamic systems Problem of future state prediction Predictability Observability Easily
More informationSpeaker Tracking and Beamforming
Speaker Tracking and Beamforming Dr. John McDonough Spoken Language Systems Saarland University January 13, 2010 Introduction Many problems in science and engineering can be formulated in terms of estimating
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.
More informationExpectation propagation for signal detection in flat-fading channels
Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationSwitching Regime Estimation
Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationMULTISENSORY SPEECH ENHANCEMENT IN NOISY ENVIRONMENTS USING BONE-CONDUCTED AND AIR-CONDUCTED MICROPHONES. Mingzi Li,Israel Cohen and Saman Mousazadeh
MULTISENSORY SPEECH ENHANCEMENT IN NOISY ENVIRONMENTS USING BONE-CONDUCTED AND AIR-CONDUCTED MICROPHONES Mingzi Li,Israel Cohen and Saman Mousazadeh Department of Electrical Engineering, Technion - Israel
More informationHidden Markov Models and Gaussian Mixture Models
Hidden Markov Models and Gaussian Mixture Models Hiroshi Shimodaira and Steve Renals Automatic Speech Recognition ASR Lectures 4&5 25&29 January 2018 ASR Lectures 4&5 Hidden Markov Models and Gaussian
More informationWavelet Based Image Restoration Using Cross-Band Operators
1 Wavelet Based Image Restoration Using Cross-Band Operators Erez Cohen Electrical Engineering Department Technion - Israel Institute of Technology Supervised by Prof. Israel Cohen 2 Layout Introduction
More informationFactor Analysis and Kalman Filtering (11/2/04)
CS281A/Stat241A: Statistical Learning Theory Factor Analysis and Kalman Filtering (11/2/04) Lecturer: Michael I. Jordan Scribes: Byung-Gon Chun and Sunghoon Kim 1 Factor Analysis Factor analysis is used
More informationScalable robust hypothesis tests using graphical models
Scalable robust hypothesis tests using graphical models Umamahesh Srinivas ipal Group Meeting October 22, 2010 Binary hypothesis testing problem Random vector x = (x 1,...,x n ) R n generated from either
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationSTA 414/2104: Machine Learning
STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationSTONY BROOK UNIVERSITY. CEAS Technical Report 829
1 STONY BROOK UNIVERSITY CEAS Technical Report 829 Variable and Multiple Target Tracking by Particle Filtering and Maximum Likelihood Monte Carlo Method Jaechan Lim January 4, 2006 2 Abstract In most applications
More informationStatistical and Adaptive Signal Processing
r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory
More information