LPC methods are the most widely used in. recognition, speaker recognition and verification

Size: px
Start display at page:

Download "LPC methods are the most widely used in. recognition, speaker recognition and verification"

Transcription

1 Digital Seech Processing Lecture 3 Linear Predictive Coding (LPC)- Introduction

2 LPC Methods LPC methods are the most widely used in seech coding, seech synthesis, seech recognition, seaker recognition and verification and for seech storage LPC methods rovide extremely accurate estimates of seech arameters, and does it extremely efficiently basic idea of Linear Prediction: current seech samle can be closely aroximated as a linear combination of ast samles, i.e., αk sn ( ) = sn ( k) for some value of, α 's k 2

3 LPC Methods for eriodic signals with eriod sn ( ) sn ( N ) N, it is obvious that but that is not what LP is doing; it is estimating sn ( ) from the ( << N ) most recent values of s( n) by linearly redicting its value for LP, the redictor coefficients (the αk's) are determined (comuted) by minimizing the sum of squared differences (over a finite interval) between the actual seech samles and the linearly yredicted ones 3

4 LPC Methods LP is based on seech roduction and synthesis models - seech can be modeled as the outut of a linear, time-varying system, excited by either quasi-eriodic ulses or noise; - assume that the model arameters remain constant over seech analysis interval i LP rovides a robust, reliable and accurate method for estimating the arameters of the linear system (the combined vocal tract, glottal ulse, and radiation characteristic for voiced seech) 4

5 LPC Methods LP methods have been used in control and information theory called methods of system estimation and system identification used extensively in seech under grou of names including. covariance method 2. autocorrelation method 3. lattice method 4. inverse filter formulation 5. sectral estimation formulation 6. maximum likelihood method 7. inner roduct method 5

6 Basic Princiles of LP k = sn ( ) = asn ( k) + Gun ( ) k Sz ( ) Hz ( ) = = GU( z) az k k the time-varying digital filter reresents the effects of the glottal ulse shae, the vocal tract IR, and radiation at the lis the system is excited by an imulse train for voiced seech, or a random noise sequence for unvoiced seech this all-ole model is a natural reresentation for non-nasal voiced seech but it also works reasonably well for nasals and unvoiced sounds 6

7 LP Basic Equations th a order linear redictor is a system of theform ( ) ( ) α ( ) ( ) α k Sz sn = ksn k Pz = kz = Sz ( ) k= k= the rediction error, en ( ), is of the form en ( ) = sn ( ) sn ( ) = sn ( ) α sn ( k) the rediction error is the outut of a system with transfer function Ez ( ) Az ( ) = = Pz ( ) = α k z Sz ( ) k = k k if the seech signal obeys the roduction model exactly, and if αk = ak, k en ( ) = Gun ( )and Az ( ) is an inverse filter for Hz ( ), i.e., Hz ( ) = Az ( ) 7

8 LP Estimation Issues need to determine { αk } directly from seech such that they give good estimates of the time-varying sectrum need to estimate { α } from short segments of seech k need to minimize mean-squared rediction error over short segments of seech resulting { αk} assumed to be the actual { ak} in the seech roduction model => intend to show that all of this can be done efficiently, reliably, and accurately for seech 8

9 Solution for {α k } short-time average rediction error is defined as 2 = = m m ( ) 2 E e ( m) s ( m) s ( m) = sˆ( ) α ˆ( ) n m ksn m k m k = select segment of seech s ( m) = s( m+ ) in the vicinity of samle n ˆ the key issue to resolve is the range of m for summation (to be discussed later) 2 9

10 Solution for {α } k can find values of αk that minimize E by setting: E = 0, i = 2,,..., αi giving the set of equations where αˆ k - 2 s ( m i)[ s ( m) ˆ α s ( m k)] = 0, i m k= ( ) ( ), m 2 s m i e m = 0 i are the values of α that minimize E (from now k k on just use αk rather than ˆ αk for the otimum values) rediction error ( e ( m)) is orthogonal to signal ( s ( m i)) for delays ( i) of to 0

11 Solution for {α } k defining i we get φ ( ik, ) = s ( m is ) ( m k) m αkφ φ ( ik, ) = ( i, 0), i= 2,,..., leading to a set of equations in unknowns that can be solved in an efficient manner for the { α } k

12 Solution for {α k } minimum mean-squared rediction error has the form 2 ˆ ˆ α n n k m k= m E = s ( m) s ( m) s ( m k) whichcanbewrittenintheform can written in the E = φ ( 00, ) α φ ( 0, k) k Process:. comute (, ) for φ ik, 0 i k 2. solve matrix equation for αk need to secify range of m to comute φ ( i, k) s need to secify ( m) 2

13 Autocorrelation Method i i assume s ( m ) exists for 0 m L and is exactly zero everywhere else (i.e., window of length L samles) ( Assumtion #) s ( m) = s( m+ ) w( m), 0 m L where wm ( ) is a finite length window of length L samles n^ n+l- ^ m=0 m=l- 3

14 Autocorrelation Method if s ( m) is non-zero only for 0 m L then e ( m ) = s ( m ) α s ( m k ) k L = = m= m= 0 is non-zero only over the interval 0 m L +, giving E e ( m ) e ( m ) i at values of m near 0 (i.e., m = 0,..., ) we are redicting signal from zero-valued samles outside the window range => e ( m) will be (relatively) large i at values near m = L (i.e., m = LL, L+,..., L+ ) we are redicting zero-valu ed samles (outside window range) from non-zero samles => e ( m) will be (relatively) large for these reasons, normally use windows that taer the segment to zero (e.g., Hamming window) m=0 m=l m=- m=l+- 4

15 Autocorrelation Method 5

16 The Autocorrelation Method ^ ^ˆ n+ L s ˆ [ m] = s[ m + n ] w [ m ] L- L k R [ k] = s [ m] s [ m+ k] 2,,, m=0= 0 6

17 The Autocorrelation Method + L s ˆ [ m] = s[ m + n ] wm [ ] ˆn + L Large errors ˆ at ends of window e [ m ] = s [ m ] α s [ m k ] s [ ] [ ˆ ] [ ] n m = s m+ w m k k= L L ( L + ) 7

18 Autocorrelation Method for calculation of φ ( ik, ) since s( m) = 0 outside the range 0 m L, then L + φ ( ik, ) = s( m is ) ( m k), i, 0 k m= 0 which is equivalent to the form L + ( i k) m= 0 φ ( ik, ) = s( ms ) ( m+ i k), i, 0 k there are L i k non-zero terms in the comutation of φ ( i, k) for each value of i and k; can easily show that φ n ˆ ( ik, ) = fi ( k) = Rn ˆ( i k), i, 0 k where R ( i k) is the short-time autocorrelation of s ( m) evaluated at i k where L k R ( k) = s ( m) s ( m+ k) m=0 8

19 Autocorrelation Method since R φ (, ik ) = R ( i k ), i, 0 k ( k) is even, then thus the basic equation becomes αφ ( i k ) = φ ( i, 0 ), i k α k R ( i k ) = R ( i ), i with the minimum mean-squared rediction error of the form E = φ ( 00, ) α φ ( 0, k) k = Rˆ( 0) αkrˆ( k) n n 9

20 Autocorrelation Method as exressed in matrix form Rˆ( 0) ˆ().. ˆ( ) α n Rn Rn R() ˆ() ˆ( 0).. ˆ( 2) α Rn Rn Rn 2 R( 2) = ˆ( ) ˆ( 2).. ˆ( 0) α ˆ( ) Rn Rn Rn Rn Rα=r with solution α = R r R is a Toelitz Matrix => symmetric with all diagonal elements equal => there exist more efficient algorithms to solve for { αk } than simle matrix inversion 20

21 Covariance Method there e is a second basic aroach to defining gthe eseech segment s ( m) and the limits on the sums, namely fix the interval over which the mean-squared error is comuted, giving ( Assumtion #2): 2 L L 2 ˆ = ˆ( ) = ˆ( ) n n n αk ( ) m= 0 m= 0 k= E e m s m s m k L m= 0 φ (, ik) = s( m is ) ( m k), i, 0 k 2

22 Covariance Method changing the summation index gives L i φ ( i, k) = s ( m) s ( m+ i k), i, 0 k m= i L k φ ( i, k ) = s ( m ) s ( m + k i ), i, 0 k m= k key difference from Autocorrelation Method is that limits of summation include terms before m = 0 => window extends samles backwards from sn ( ˆ - ) to sn ( ˆ + L- ) since we are extending window backwards, don't need to taer it using a HW- since there is no transition at window edges m=- m=l- 22

23 Covariance Method 23

24 Covariance Method cannot use autocorrelation formulation => this is a true cross correlation need to solve set of equations of the form αφ k φ E ( ik, ) = ( i, 0), i= 2,,...,, = φ (,) 00 α φ (, 0k) k φ n ˆ (, φ φ α φ ) ( 2, ).. (, ) ( 0, ) ˆ(,) 2 ˆ(,) 22.. ˆ(, 2 ) α φn φn φn 2 φ(,) = φˆ(, ) φˆ(, 2).. φˆ(, ) α φˆ(, 0) n n n n φα = ψ or α = φ ψ 24

25 Covariance Method we have φ ( ik, ) = φ ( ki, ) => symmetric but not Toelitz matrix whose diagonal elements are related as φ ( i +, k + ) = φ ( i, k ) + s ( i ) s ( k ) s ( L i ) s ( L k ) φ (,) 22 = φ (,) + s ( 2) s ( 2) s ( L 2) s ( L 2) all terms φ ( ik, ) have a fixed number of terms contributing to the comuted values ( L terms) ( ik, ) is a covariance matrix => secialized solution for { } φ αk called the Covariance Method 25

26 Summary of LP th use order linear redictor to redict s( ) from revious samles minimize i i mean-squared error, E, over analysis window of duration L-samles solution for otimum redictor coefficients, {αk }, is based on solving a matrix equation => two solutions have evolved - autocorrelation method => signal is windowed by a taering window in order to minimize discontinuities at beginning (redicting seech from zero-valued samles) and end (redicting zero-valued samles from seech samles) of the interval; the matrix φ n ( ik, ) is shown to be an autocorrelation function; the resulting autocorrelation φ n ˆ matrix is Toelitz and can be readily solved using standard matrix solutions - covariance method => the signal is extended by samles outside the normal range of 0 m L- to include samles occurring rior to m = 0; this eliminates large errors in comuting the signal from values rior to m = 0 (they are available) and eliminates the need for a taering window; resulting matrix of correlations is symmetric but not Toelitz => different method of solution with somewhat different set of otimal rediction coefficients, { } α k 26

27 LPC Summary. Seech Production Model: sn ( ) = asn k ( k ) + Gun ( ) Sz ( ) Hz ( ) = = GU ( z ) az k 2. Linear Prediction Model: sn ( ˆ) = α sn ( ˆ k) Sz ( ) Pz ( ) = = S( z) k α z k k k en ( ˆ) = sn ( ˆ) sn ( ˆ) = sn ( ˆ) α sn ( ˆ k) Ez ( ) ( ) = = α k Az kz Sz ( ) k 27

28 LPC Summary 3. LPC Minimization: 2 2 = ( ) = ( ) ( ) m m E e m s m s m = s( m) αks( m k) m k= E = 0, i = 2,,..., α i s ( m i) s ( m) = α s ( m i) s ( m k) k m k= m φ (, ik ) = s ( m is ) ( m k ) m αφ(, k) = φ ( i, 0), i = 2,,..., i k 2 E = k φ (,) 00 α φ (, 0k) 28

29 LPC Summary 4. Autocorrelation Method: s ( m ) = s ( m + ) w ( m ), 0 m L e ( m) = s ( m) α s ( m k), 0 m L + k i s ( m) defined for 0 m L ; e ( m) defined for 0 m L + large errors for 0 m and for L L + E = e ( m ) m= 0 2 L ( i k) m L+ m= 0 ( ) φ (, ik) = R( i k) = s( ms ) ( m+ i k) = R i k α R ( i k ) = R ( i), i k E R ( 0) α R ( k) = k 29

30 LPC Summary 4. Autocorrelation Method: i resulting matrix equation: R α = r or α =R r Rˆ( 0) ˆ().. ˆ( ) α n Rn Rn R() ˆ() ˆ( 0).. ˆ( 2) α Rn Rn Rn 2 R( 2) = ˆ( ) ˆ( 2).. ˆ( 0) α Rn Rn R n Rˆ( ) n i matrix equation solved using Levinson or Durbin method 30

31 5. Covariance Method: i fix interval for error signal i LPC Summary L L 2 E = e( m) = s( m) αks( m k) m= 0 m= 0 k= need signal for from sn ( ˆ ) to sn ( ˆ + L ) L+ samles 2 i αφ k = φ E (, ik) ( i, 0), i = 2,,..., = k φ (,) 00 α φ (, 0 k ) k exressed as a matrix equation: or, symmetric matrix φα = ψ α = φ ψ φ φ(,) φ(, 2).. φ(, ) α φ (, 0) φ(,) 2 φ(,) 22.. φ(, 2 ) α 2 φ ˆ(,) 20 n = φ n ˆ(, ) φˆ(, 2).. φˆ(, ) α φˆ(, 0) n n n 3

32 Comutation of Model Gain it is reasonable to exect the model gain, G, to be determined by matching the signal energy with the energy of the linearly yredicted samles from the basic model equations we have k = Gu( n) = s( n) a s( n k) model whereas for the rediction error we have k en ( ) = sn ( ) α sn ( k) best fit to model k when αk = ak (i.e., erfect match to model), then en ( ) = Gun ( ) since it is virtually imossible to guarantee that α k = ak, cannot use this simle matching roerty for determining the gain; instead use energy matching criterion (energy in error signal=energy in excitation) L + L G u ( m) = e ( m) = E m= 0 m= 0 32

33 Gain Assumtions assumtions about excitation to solve for G voiced seech-- un ( ) = δ ( n) Lorder of a single itch eriod; redictor order,, large enough to model glottal ulse shae, vocal tract IR, and radiation unvoiced seech-- un ( )-zero mean, unity variance, stationary white noise rocess 33

34 Solution for Gain (Voiced) for voiced seech the excitation is Gδ ( n) with outut h ( n) (since it is the IR of the system), ( ) = α ( ) + δ( ); G G hn hn k G n Hz ( ) = = k Az ( ) α k k z with autocorrelation Rm ( ) (of the imulse resonse) satisfying the relation shown below Rm ( ) = hn ( ) hm ( + n) = R [ m], 0 m< n= 0 Rm ( ) = α R ( m k), m< k αk 2 R ( 0 ) = R ( k ) + G, m = 0 34

35 Solution for Gain (Voiced) Since Rm ( ) and Rn ( m) have the identical form, it follows that Rm ( ) = c R ( m ), 0 m where c is a constant to be determined. Since the total energies in the signal ( R ( 0)) and the imulse resonse ( R ( 0)) must be equal, the constant c must be, and we obtain the relation 2 k G = R ( 0 ) α R ( k ) = E since Rm ( ) = R ( m), 0 m, and the energy of the imulse resonse=energy of fthe signal => first + coefficients i of fthe autocorrelation of the imulse resonse of the model are identical to the first + coefficients of the autocorrelation function of the seech signal. This condition called the autocorrelation matching roerty of the autocorrelation method. 35

36 Solution for Gain (Unvoiced) for unvoiced seech the inut is white noise with zero mean and unity variance, i.e., E u ( n ) u ( n m ) = δ ( m ) if we excite the system with inut Gu( n) and call the outut gn ( ) then g ( n) = α g ( n k) + Gu( n) k i Since the autocorrelation function for the outut is the convolution of the autocorrelation function of the imulse resonse with the autocorrelation function of the white noise inut, then Egngn [ [ ] [ m]] = Rm [ ] δ [ m] = Rm [ ] letting Rm ( ) denote the autocorrelation of gn ( ) gives Rm ( ) = E gngn ( ) ( m) = α [ ( ) kegn kgn ( m)] + E Gun ( ) gn ( m) = α Rm ( k), m k since E Gungn ( ) ( m) = 0 for m> 0 because un ( ) is uncorrelated with any signal rior to un ( ) 0 36

37 Solution for Gain (Unvoiced) for m = 0 we get R ( 0) = α ( ) + ( ) kr k GE u n g( n) = α Rk ( ) + G k 2 since E ungn ( ) ( ) = E un ( )( Gu ( n ) + terms rior to n = G 2 since the energy in the signal must equal the energy in the resonse to Gu ( n ) we get Rm ( ) = R( m) 2 ˆ( 0) n αk ( ) G = R R k = E 37

38 Frequency Domain Interretations of Linear Predictive Analysis 38

39 The Resulting LPC Model i i i The final LPC model consists of the LPC arameters, { α } = 2 k, k,,...,, and the gain, G, which together define the system function G Hz ( ) = αkz k with frequency resonse jω G H ( e ) = = ω α j k ke k = G jω Ae ( ) with the gain determined by matching the energy of the model to the short-time energy of the seech signal, i.e., 2 = ( ) 2 ˆ = ˆ = ˆ 0 G En en m Rn αkr k m k= ( ) ( ) ( ) 39

40 LPC Sectrum x = s.* hamming(30); G He ( jω ) = X = fft( x, 000 ) ω α j k ke [ A, G, r ] = autolc( x, 0 ) H = G./ fft(a,000); LP Analysis is seen to be a method of short-time sectrum estimation with removal of excitation fine structure (a form of wideband sectrum analysis) 40

41 LP Short-Time Sectrum Analysis Defined seech segment as: s m = s m+ w m n ˆ [ ] [ ] [ ] The discrete-time Fourier transform of this windowed segment is: ( jω ) Sˆ [ ˆ n e = s m+ n] w[ m] e m= jωm Short-time FT and the LP sectrum are linked via short-time time autocorrelation 4

42 LP Short-Time Sectrum Analysis (a) Voiced seech segment obtained using a Hamming window (b) Corresonding short- time autocorrelation function used in LP analysis (heavy line shows values used in LP analysis) (c) Corresonding shorttime log magnitude Fourier transform and short-time log magnitude LPC sectrum (F S =6 khz) 42

43 LP Short-Time Sectrum Analysis (a) Unvoiced seech segment obtained using a Hamming window (b) Corresonding short- time autocorrelation function used in LP analysis (heavy line shows values used in LP analysis) (c) Corresonding shorttime log magnitude Fourier transform and short-time log magnitude LPC sectrum (F S =6 khz) 43

44 Frequency Domain Interretation of Mean-Squared Prediction Error i The LP sectrum rovides a basis for examining the roerties of the rediction error (or equivalently the excitation of the VT) i The mean-squared rediction error at samle is: E = L + m= 0 e 2 [ m] i which, by Parseval's Theorem, can be exressed as: i π π jω 2 jω 2 jω 2 2 = ( ) ω ( ) ( ) ω 2π E = 2π = π π E e d S e A e d G jω jω where S ( e ) is the FT of s [ m] and A( e ) is the corresonding rediction error frequency resonse jω Ae ( ) = α e k jωk 44

45 Frequency Domain Interretation of Mean-Squared Prediction Error i The LP sectrum is of the form: jω G H ( e ) = jω Ae ( ) i Thus we can exress the mean-squared error as: 2 π jω 2 G S ( e ) jω 2 2 π He ( ) π E = dω = G 2 i We see that minimizing total squared rediction error is equivalent to finding gain and redictor coefficients such that the integral of the ratio of the energy sectrum of the seech segment to the magnitude squared of the frequency resonse of the model linear system is unity. i 2 Thus Sˆ ( j n e ω ) can be interreted as a frequency-domain jω S e jω 2 S ( e ) weighting function LP weights frequencies where is large more heavily than when is small. ( ) 2 45

46 LP Interretation Examle Much better sectral matches to STFT sectral eaks than to STFT sectral valleys as redicted by sectral interretation of error minimization. 46

47 LP Interretation Examle2 Note small differences in sectral shae between STFT, autocorrelation sectrum and covariance sectrum when using short window duration (L=5 samles). 47

48 Effects of Model Order i The AC function, Rn ˆ[ m] of the seech segment, sn ˆ[ m], and the AC function, Rm [ ], of the imulse resonse, hm [ ], corresonding to the system function, H ( z), are equal for the first ( + ) values. Thus, as, the AC functions are equal for all values and thus: j ω 2 j ω 2 = S lim H ( e ) S ( e ) i Thus if is large enough, the FR of the all-ole model, ( j H e ω ), can aroximate the signal sectrum with arbitrarily small error. 48

49 Effects of Model Order 49

50 Effects of Model Order 50

51 Effects of Model Order 5

52 Effects of Model Order lots show Fourier transform of segment and LP sectra for various orders - as increases, more details of the sectrum are reserved - need to choose a value of that reresents the sectral effects of the glottal ulse, vocal tract and radiation--nothing else 52

53 Linear Prediction Sectrogram i Seech sectrogram reviously defined as: L j(2 π / N) km 20log Sr[ k] = 20log s[ rr+ m] w[ m] e m= 0 for set of times, t = rrt, and set of frequencies, F = kf / N,,2,..., N / 2 r k S where R is the time shift (in samles) between adjacent STFTs, T is the samling eriod, F = / T is the samling frequency, S and N is the size of the discrete Fourier transform used to comute each STFT estimate. i Similarly we can define the LP sectrogram as an image lot of: G r r j(2 π / N) k Ar e 20log H [ k ] = 20log ( ) j(2 π / N) k where Gr and Ar( e ) are the gain and rediction error olynomial at analysis time rr. 53

54 Linear Prediction Sectrogram L=8, R=3, N=000, 40 db dynamic range 54

55 Comarison to Other Sectrum Analysis Methods Sectra of synthetic vowel /IY/ (a) Narrowband sectrum using 40 msec window (b) Wideband sectrum using a 0 msec window (c) Cestrally smoothed sectrum (d) LPC sectrum from a 40 msec section using a =2 order LPC analysis 55

56 Comarison to Other Sectrum Analysis Methods Natural seech sectral estimates using cestral smoothing (solid line) and linear rediction analysis (dashed line). Note the fewer (surious) eaks in the LP analysis sectrum since LP used =2 which restricted the sectral match to a maximum of 6 resonance eaks. Note the narrow bandwidths of the LP resonances versus the cestrally smoothed resonances. 56

57 Selective Linear Prediction it is ossible to aly LP methods to selected arts of sectrum - 04kH 0-4 khz for voiced sounds use a redictor of order khz for unvoiced sounds use a redictor of order 2 the key id ea is to ma the frequency region { f, f } linearly to { 0,.} 5 { 2πf 2πf } { 0π} or, equivalently, the region, mas linearly to, via the transformation = ω 2 π f ω A 2πf 2πf 2πf B A ω 2 ω ( ) = j j m R m Sˆ ( ) ω n e e d 2π B we must modify the calculation for the autocorrelation to give: π π A B A B 57

58 Selective Linear Prediction 0-0 khz region modeled using =28 no discontinuity in model sectra at 5 khz 0-5 khz region modeled using =4 5-0 khz region modeled using 2 =5 discontinuity in model sectra at 5 khz 58

59 Solutions of LPC Equations Covariance Method (Cholesky Decomosition Method) 59

60 LPC Solutions-Covariance Method for the covariance method we need to solve the matrix equation αφ k φ ( ik, ) = ( i, 0), i= 2,,..., φα = ψ (in matrix notation) φ is a ositive definite, symmetric matrix with ( i, j) element φ (, i j), and α and ψ are column vectors with elements α and φ ( i, 0) the solution of the matrix equation is called the Cholesky decomosition, or square root method t φ =VDV ; V = lower triangula r matrix ti with ith' 's on the main diagonal D=diagonal matrix i 60

61 LPC Solutions-Covariance Method can readily determine elements of V and D by solving for ( i, j) elements of the matrix equation, as follows giving j ik k jk φ ˆ ( i, j) = V d V, j i n j ij j n ik k jk Vd = φ ˆ ( i, j) V dv, j i and for the diagonal elements giving g φ ˆ ( ii, ) = n i VikdV k ik i d = φ ( i, i) V d, i i with d φ ˆ (, ) = n ik 2 k 2 6

62 Cholesky Decomosition Examle consider examle with = 4, and matrix elements φ ( i, j ) = φ φ φ2 φ3 φ4 φ2 φ22 φ32 φ42 = φ3 φ32 φ33 φ 43 φ 4 φ 42 φ 34 φ d V V V V d V V V3 V d V V4 V42 V d ij 62

63 Cholesky Decomosition Examle solve matrix for d, V, V, V, d, V, V, d, V, d d = φ V d = φ ; V d = φ ; V d = φ V = φ / d ; V = φ / d ; V = φ / d d 2 2 = φ 22 V d 2 ( ) ( φ ) V d = φ V dv V = φ V dv ste ste 2 ste 3 V d = φ V dv V = V dv / d ste 4 / d iterate rocedure to solve for d, V, d

64 LPC Solutions-Covariance Method now need to solve for α using a 2-ste rocedure VDV α = ψ writing this as VY= ψ with t DV α = Y t t or V α = D Y from V (which is now known) solve for column vector Y using a simle recursion of the form i ψ i i ij j j = Y = V Y, i with initial condition Y = ψ 2 64

65 LPC Solutions-Covariance Method now can solve for α using the recursion α i i i ji j j=+ i with initial condition α = = Y / d V α, i Y / d calculation roceeds backwards from i = down to i = 65

66 Cholesky Decomosition Examle continuing the examle we solve for Y Y ψ V2 0 0 Y2 ψ 2 = V3 V32 0 Y 3 ψ 3 V4 V42 V43 Y 4 ψ 4 first solving for Y Y we get Y = ψ Y = ψ V Y Y = ψ V Y V Y Y4 = ψ 4 V4Y V42Y2 V43Y3 66

67 Cholesky Decomosition Examle next solve for α from equation V V V α / d Y Y / d V32 V42 α2 0 / d2 0 0 Y2 Y2 / d2 = = 0 0 V 43 α / d3 0 Y 3 Y3/ d α / d4 Y4 Y4 / d4 giving the results α4 = Y4 / d4 α3 = Y3/ d3 V43α4 α2 = Y2 / d2 V32α3 V42α4 α = Y/ d V2α2 V3α3 V4α4 comleting the solution 67

68 Covariance Method Minimum Error the minimum mean squared error can be written in the form E = φ ( 00, ) α φ ( 0, k) k t = φ (,) 00 α ψ t t since α = YD V can write this as E t 2 k = φ ( 00, ) Y D Y = φ (,) 00 Y / d E n ˆ k this comutation for can be used for all values of LP order from to can understand how LP order reduces mean-squared error 68

69 69

70 Solutions of LPC Equations Autocorrelation Method via Levinson-Durbin Algorithm 70

71 Levinson-Durbin Algorithm i Autocorrelation equations (at each frame ) : αk R[ i k ] = R[ i] i α R=ri is a ositive definite symmetric Toelitz matrix irthe set of otimum redictor coefficients satisfy: Ri [] αk R[ i k ] = 0, i i with minimum mean-squared rediction error of: R[0] α k R [ k ] = E ( ) 7

72 Levinson-Durbin Algorithm 2 i By combining the last two equations we get a larger matrix equation of the form: ( ) R[0] R[] R[2]... R[ ] E ( ) R[] R[0] R[]... R[ ] α 0 ( ) R[2] R[] R[0]... R[ 2] α 2 = ( ) R [ ] R [ ] R [ 2]... R[0] α 0 i exanded matrix is still Toelitz and can be solved iteratively by incororating new correlation value at each iteration and solving for next higher order redictor in terms of new correlation value and revious redictor 72

73 Levinson-Durbin Algorithm 3 th i Show how i order solution can be derived from order solution; i.e., given α, the solution to we derive solution to R α ( i ) ( i ) ( i ) ( i ) ( i ) = e () i () i () i st i The ( i ) solution can be exressed as: R[0] R[] R[2]... R[ i ] E ( i ) R[] R[0] R[]... R[ i 2] α 0 ( i ) R[2] R[] R[0]... R[ i 3] α 2 = ( i ) Ri [ ] Ri [ 2] Ri [ 3]... R[0] α i 0 R α ( i ) st = e 73

74 Levinson-Durbin Algorithm 4 i ( i ) ( i) Aending a 0 to vector α and multilying by the matrix R gives: ( i ) R[0] R[] R[2]... R[ i] E R[] R[0] R[]... R[ i ] ( i ) α 0 ( i ) R[2] R[] R[0]... R[ i 2] α 2 0 = Ri [ ] Ri [ 2] Ri [ 3]... R[] ( i ) α 0 i ( i ) Ri [ ] Ri [ ] Ri [ 2]... R[0] 0 γ i i ( i ) ( i ) α j j= where γ = Ri [] Ri [ j] and Ri [] are introduced 74

75 Levinson-Durbin Algorithm 5 i Key ste is that since Toelitz matrix has secial symmetry we can reverse the order of the equations (first equation last, last equation first), giving: g ( i ) R[0] R[] R[2]... R[ i] 0 γ ( i ) R[] R[0] R[]... R[ i ] α 0 ( i ) R[2] R[] R[0]... R[ i 2] α 2 0 = i ) ( Ri [ ] Ri [ 2] Ri [ 3]... R[] α i 0 ( i ) Ri [] Ri [ ] Ri [ 2]... R[0] E 75

76 Levinson-Durbin Algorithm 6 i To get the equation into the desired form (a single () i e ) comonent in the vector ) we combine the two sets of matrices (with a multilicative factor R i () i ( i ) ( i ) 0 E γ ( ) i α ( i ) α 0 0 ( i ) ( i ) α 2 α ki = ki.... ( i ) α ( i ) i α i 0 0 ( i ) ( i ) 0 γ E ( i ) Choose γ so that non-zero entry, i.e., k i ( i ) Ri [] [ ] ( i ) α j Ri j γ j= i = = ( i ) ( i ) E k i ) giving: vector on right has only a single E 76

77 Levinson-Durbin Algorithm 7 i The first element of the right hand side vector is now: E = E k γ = E ( k ) () i ( i ) ( i ) ( i ) 2 i i i The k arameters are called PARCOR coefficients. i i ( i ) th With this choice of γ, the vector of order coefficients i is: 0 () i ( i ) ( i ) α α α i () i ( i ) ( i ) α 2 α 2 α i 2 = ki... () i ( i ) ( i ) α i αi α () i αi 0 i yielding the udating rocedure α = α kα, j =,2,..., α () i ( i ) ( i ) j j i i j () i i = k i i redictor 77

78 Levinson-Durbin Algorithm 7 i i i i i The final solution for order is: α ( ) j = α j j with rediction error ( ) 2 2 = m = m m= m= E E[0] ( k ) R[0] ( k ) If we use normalized autocorrelation coefficients: rk [ ] = Rk [ ]/ R[0] we get norma lized errors of the form: ν E () i i i () i () i 2 = = αk rk = km R[0] m = () i 0< ν < k i < where or [ ] ( ) 78

79 Levinson-Durbin Algorithm () i ( i ) A ( z) = A ( z) kz A i ( z ) i ( i ) 79

80 Autocorrelation Examle consider a simle = 2 solution of the form R( 0) R() α R() = 0 α R() R( ) R ( 2 ) with solution 2 E ( 0 ) () = R ( 0 ) k = R()/ R( 0) α = R()/ R( 0) 2 2 R 0 R E () = ( ) ( ) R( 0) 80

81 Autocorrelation Examle ( 2) ( 0) ( ) = R R R R ( 0) R ( ) k 2 = 2 2 α α R( 2) R( 0) R ( ) ( 2) 2 = 2 2 R ( 2) = ( 0) R ( ) R() R( 0) R() R( 2) with final coefficients α α E = α ( 2 ) ( 2) 2 = α2 () i R ( 0) R ( ) = rediction error for redictor of order i 8

82 Prediction Error as a Function of V n En Rn[ k] αk R [ 0] R [ 0] = = n n Model order is usually determined by the following rule of thumb: F s /000 oles for vocal tract 2-4 oles for radiation 2 oles for glottal ulse 82

83 Autocorrelation Method Proerties mean-squared rediction error always non-zero decreases monotonically with increasing model order autocorrelation matching roerty model and data match u to order sectrum matching roerty favors eaks of short-time FT minimum-hase roerty zeros of A(z) are inside the unit circle Levinson-Durbin recursion efficient algorithm for finding rediction coefficients PARCOR coefficients and MSE are by-roducts 83

Chapter 9. Linear Predictive Analysis of Speech Signals 语音信号的线性预测分析

Chapter 9. Linear Predictive Analysis of Speech Signals 语音信号的线性预测分析 Chapter 9 Linear Predictive Analysis of Speech Signals 语音信号的线性预测分析 1 LPC Methods LPC methods are the most widely used in speech coding, speech synthesis, speech recognition, speaker recognition and verification

More information

Keywords: Vocal Tract; Lattice model; Reflection coefficients; Linear Prediction; Levinson algorithm.

Keywords: Vocal Tract; Lattice model; Reflection coefficients; Linear Prediction; Levinson algorithm. Volume 3, Issue 6, June 213 ISSN: 2277 128X International Journal of Advanced Research in Comuter Science and Software Engineering Research Paer Available online at: www.ijarcsse.com Lattice Filter Model

More information

Autoregressive (AR) Modelling

Autoregressive (AR) Modelling Autoregressive (AR) Modelling A. Uses of AR Modelling () Alications (a) Seech recognition and coding (storage) (b) System identification (c) Modelling and recognition of sonar, radar, geohysical signals

More information

Lab 9a. Linear Predictive Coding for Speech Processing

Lab 9a. Linear Predictive Coding for Speech Processing EE275Lab October 27, 2007 Lab 9a. Linear Predictive Coding for Speech Processing Pitch Period Impulse Train Generator Voiced/Unvoiced Speech Switch Vocal Tract Parameters Time-Varying Digital Filter H(z)

More information

The Weighted Sum of the Line Spectrum Pair for Noisy Speech

The Weighted Sum of the Line Spectrum Pair for Noisy Speech HELSINKI UNIVERSITY OF TECHNOLOGY Deartment of Electrical and Communications Engineering Laboratory of Acoustics and Audio Signal Processing Zhijian Yuan The Weighted Sum of the Line Sectrum Pair for Noisy

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra Numerous alications in statistics, articularly in the fitting of linear models. Notation and conventions: Elements of a matrix A are denoted by a ij, where i indexes the rows and

More information

L7: Linear prediction of speech

L7: Linear prediction of speech L7: Linear prediction of speech Introduction Linear prediction Finding the linear prediction coefficients Alternative representations This lecture is based on [Dutoit and Marques, 2009, ch1; Taylor, 2009,

More information

Use of Transformations and the Repeated Statement in PROC GLM in SAS Ed Stanek

Use of Transformations and the Repeated Statement in PROC GLM in SAS Ed Stanek Use of Transformations and the Reeated Statement in PROC GLM in SAS Ed Stanek Introduction We describe how the Reeated Statement in PROC GLM in SAS transforms the data to rovide tests of hyotheses of interest.

More information

COURSE OUTLINE. Introduction Signals and Noise: 3) Analysis and Simulation Filtering Sensors and associated electronics. Sensors, Signals and Noise

COURSE OUTLINE. Introduction Signals and Noise: 3) Analysis and Simulation Filtering Sensors and associated electronics. Sensors, Signals and Noise Sensors, Signals and Noise 1 COURSE OUTLINE Introduction Signals and Noise: 3) Analysis and Simulation Filtering Sensors and associated electronics Noise Analysis and Simulation White Noise Band-Limited

More information

representation of speech

representation of speech Digital Speech Processing Lectures 7-8 Time Domain Methods in Speech Processing 1 General Synthesis Model voiced sound amplitude Log Areas, Reflection Coefficients, Formants, Vocal Tract Polynomial, l

More information

A Recursive Block Incomplete Factorization. Preconditioner for Adaptive Filtering Problem

A Recursive Block Incomplete Factorization. Preconditioner for Adaptive Filtering Problem Alied Mathematical Sciences, Vol. 7, 03, no. 63, 3-3 HIKARI Ltd, www.m-hiari.com A Recursive Bloc Incomlete Factorization Preconditioner for Adative Filtering Problem Shazia Javed School of Mathematical

More information

Spectral Analysis by Stationary Time Series Modeling

Spectral Analysis by Stationary Time Series Modeling Chater 6 Sectral Analysis by Stationary Time Series Modeling Choosing a arametric model among all the existing models is by itself a difficult roblem. Generally, this is a riori information about the signal

More information

ESE 524 Detection and Estimation Theory

ESE 524 Detection and Estimation Theory ESE 524 Detection and Estimation heory Joseh A. O Sullivan Samuel C. Sachs Professor Electronic Systems and Signals Research Laboratory Electrical and Systems Engineering Washington University 2 Urbauer

More information

State Estimation with ARMarkov Models

State Estimation with ARMarkov Models Deartment of Mechanical and Aerosace Engineering Technical Reort No. 3046, October 1998. Princeton University, Princeton, NJ. State Estimation with ARMarkov Models Ryoung K. Lim 1 Columbia University,

More information

Voiced Speech. Unvoiced Speech

Voiced Speech. Unvoiced Speech Digital Speech Processing Lecture 2 Homomorphic Speech Processing General Discrete-Time Model of Speech Production p [ n] = p[ n] h [ n] Voiced Speech L h [ n] = A g[ n] v[ n] r[ n] V V V p [ n ] = u [

More information

Feature extraction 2

Feature extraction 2 Centre for Vision Speech & Signal Processing University of Surrey, Guildford GU2 7XH. Feature extraction 2 Dr Philip Jackson Linear prediction Perceptual linear prediction Comparison of feature methods

More information

3.4 Design Methods for Fractional Delay Allpass Filters

3.4 Design Methods for Fractional Delay Allpass Filters Chater 3. Fractional Delay Filters 15 3.4 Design Methods for Fractional Delay Allass Filters Above we have studied the design of FIR filters for fractional delay aroximation. ow we show how recursive or

More information

ON THE USE OF PHASE INFORMATION FOR SPEECH RECOGNITION. Baris Bozkurt and Laurent Couvreur

ON THE USE OF PHASE INFORMATION FOR SPEECH RECOGNITION. Baris Bozkurt and Laurent Couvreur ON THE USE OF PHASE NFOMATON FO SPEECH ECOGNTON Baris Bozkurt and Laurent Couvreur TCTS Lab, Faculté Polytechnique De Mons, nitialis Scientific Park, B-7000 Mons, Belgium, hone: +32 65 374733, fax: +32

More information

General Linear Model Introduction, Classes of Linear models and Estimation

General Linear Model Introduction, Classes of Linear models and Estimation Stat 740 General Linear Model Introduction, Classes of Linear models and Estimation An aim of scientific enquiry: To describe or to discover relationshis among events (variables) in the controlled (laboratory)

More information

Linear Prediction 1 / 41

Linear Prediction 1 / 41 Linear Prediction 1 / 41 A map of speech signal processing Natural signals Models Artificial signals Inference Speech synthesis Hidden Markov Inference Homomorphic processing Dereverberation, Deconvolution

More information

Principles of Computed Tomography (CT)

Principles of Computed Tomography (CT) Page 298 Princiles of Comuted Tomograhy (CT) The theoretical foundation of CT dates back to Johann Radon, a mathematician from Vienna who derived a method in 1907 for rojecting a 2-D object along arallel

More information

Participation Factors. However, it does not give the influence of each state on the mode.

Participation Factors. However, it does not give the influence of each state on the mode. Particiation Factors he mode shae, as indicated by the right eigenvector, gives the relative hase of each state in a articular mode. However, it does not give the influence of each state on the mode. We

More information

The analysis and representation of random signals

The analysis and representation of random signals The analysis and reresentation of random signals Bruno TOÉSNI Bruno.Torresani@cmi.univ-mrs.fr B. Torrésani LTP Université de Provence.1/30 Outline 1. andom signals Introduction The Karhunen-Loève Basis

More information

SPECTRAL ANALYSIS OF GEOPHONE SIGNAL USING COVARIANCE METHOD

SPECTRAL ANALYSIS OF GEOPHONE SIGNAL USING COVARIANCE METHOD International Journal of Pure and Alied Mathematics Volume 114 No. 1 17, 51-59 ISSN: 1311-88 (rinted version); ISSN: 1314-3395 (on-line version) url: htt://www.ijam.eu Secial Issue ijam.eu SPECTRAL ANALYSIS

More information

Radial Basis Function Networks: Algorithms

Radial Basis Function Networks: Algorithms Radial Basis Function Networks: Algorithms Introduction to Neural Networks : Lecture 13 John A. Bullinaria, 2004 1. The RBF Maing 2. The RBF Network Architecture 3. Comutational Power of RBF Networks 4.

More information

Department of Electrical and Computer Engineering Digital Speech Processing Homework No. 7 Solutions

Department of Electrical and Computer Engineering Digital Speech Processing Homework No. 7 Solutions Problem 1 Department of Electrical and Computer Engineering Digital Speech Processing Homework No. 7 Solutions Linear prediction analysis is used to obtain an eleventh-order all-pole model for a segment

More information

Towards understanding the Lorenz curve using the Uniform distribution. Chris J. Stephens. Newcastle City Council, Newcastle upon Tyne, UK

Towards understanding the Lorenz curve using the Uniform distribution. Chris J. Stephens. Newcastle City Council, Newcastle upon Tyne, UK Towards understanding the Lorenz curve using the Uniform distribution Chris J. Stehens Newcastle City Council, Newcastle uon Tyne, UK (For the Gini-Lorenz Conference, University of Siena, Italy, May 2005)

More information

Finite-Sample Bias Propagation in the Yule-Walker Method of Autoregressive Estimation

Finite-Sample Bias Propagation in the Yule-Walker Method of Autoregressive Estimation Proceedings of the 7th World Congress The International Federation of Automatic Control Seoul, Korea, July 6-, 008 Finite-Samle Bias Proagation in the Yule-Walker Method of Autoregressie Estimation Piet

More information

DSP IC, Solutions. The pseudo-power entering into the adaptor is: 2 b 2 2 ) (a 2. Simple, but long and tedious simplification, yields p = 0.

DSP IC, Solutions. The pseudo-power entering into the adaptor is: 2 b 2 2 ) (a 2. Simple, but long and tedious simplification, yields p = 0. 5 FINITE WORD LENGTH EFFECTS 5.4 For a two-ort adator we have: b a + α(a a ) b a + α(a a ) α R R R + R The seudo-ower entering into the adator is: R (a b ) + R (a b ) Simle, but long and tedious simlification,

More information

Minimax Design of Nonnegative Finite Impulse Response Filters

Minimax Design of Nonnegative Finite Impulse Response Filters Minimax Design of Nonnegative Finite Imulse Resonse Filters Xiaoing Lai, Anke Xue Institute of Information and Control Hangzhou Dianzi University Hangzhou, 3118 China e-mail: laix@hdu.edu.cn; akxue@hdu.edu.cn

More information

Hotelling s Two- Sample T 2

Hotelling s Two- Sample T 2 Chater 600 Hotelling s Two- Samle T Introduction This module calculates ower for the Hotelling s two-grou, T-squared (T) test statistic. Hotelling s T is an extension of the univariate two-samle t-test

More information

Convolutional Codes. Lecture 13. Figure 93: Encoder for rate 1/2 constraint length 3 convolutional code.

Convolutional Codes. Lecture 13. Figure 93: Encoder for rate 1/2 constraint length 3 convolutional code. Convolutional Codes Goals Lecture Be able to encode using a convolutional code Be able to decode a convolutional code received over a binary symmetric channel or an additive white Gaussian channel Convolutional

More information

A Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression

A Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Journal of Modern Alied Statistical Methods Volume Issue Article 7 --03 A Comarison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Ghadban Khalaf King Khalid University, Saudi

More information

Chapter 2 Speech Production Model

Chapter 2 Speech Production Model Chapter 2 Speech Production Model Abstract The continuous speech signal (air) that comes out of the mouth and the nose is converted into the electrical signal using the microphone. The electrical speech

More information

FE FORMULATIONS FOR PLASTICITY

FE FORMULATIONS FOR PLASTICITY G These slides are designed based on the book: Finite Elements in Plasticity Theory and Practice, D.R.J. Owen and E. Hinton, 1970, Pineridge Press Ltd., Swansea, UK. 1 Course Content: A INTRODUCTION AND

More information

Filters and Equalizers

Filters and Equalizers Filters and Equalizers By Raymond L. Barrett, Jr., PhD, PE CEO, American Research and Develoment, LLC . Filters and Equalizers Introduction This course will define the notation for roots of olynomial exressions

More information

Frequency-Domain Design of Overcomplete. Rational-Dilation Wavelet Transforms

Frequency-Domain Design of Overcomplete. Rational-Dilation Wavelet Transforms Frequency-Domain Design of Overcomlete 1 Rational-Dilation Wavelet Transforms İlker Bayram and Ivan W. Selesnick Polytechnic Institute of New York University Brooklyn, NY 1121 ibayra1@students.oly.edu,

More information

Speaker Identification and Verification Using Different Model for Text-Dependent

Speaker Identification and Verification Using Different Model for Text-Dependent International Journal of Alied Engineering Research ISSN 0973-4562 Volume 12, Number 8 (2017). 1633-1638 Research India Publications. htt://www.riublication.com Seaker Identification and Verification Using

More information

Approximating min-max k-clustering

Approximating min-max k-clustering Aroximating min-max k-clustering Asaf Levin July 24, 2007 Abstract We consider the roblems of set artitioning into k clusters with minimum total cost and minimum of the maximum cost of a cluster. The cost

More information

A Time-Varying Threshold STAR Model of Unemployment

A Time-Varying Threshold STAR Model of Unemployment A Time-Varying Threshold STAR Model of Unemloyment michael dueker a michael owyang b martin sola c,d a Russell Investments b Federal Reserve Bank of St. Louis c Deartamento de Economia, Universidad Torcuato

More information

COMPARISON OF VARIOUS OPTIMIZATION TECHNIQUES FOR DESIGN FIR DIGITAL FILTERS

COMPARISON OF VARIOUS OPTIMIZATION TECHNIQUES FOR DESIGN FIR DIGITAL FILTERS NCCI 1 -National Conference on Comutational Instrumentation CSIO Chandigarh, INDIA, 19- March 1 COMPARISON OF VARIOUS OPIMIZAION ECHNIQUES FOR DESIGN FIR DIGIAL FILERS Amanjeet Panghal 1, Nitin Mittal,Devender

More information

SPEECH ANALYSIS AND SYNTHESIS

SPEECH ANALYSIS AND SYNTHESIS 16 Chapter 2 SPEECH ANALYSIS AND SYNTHESIS 2.1 INTRODUCTION: Speech signal analysis is used to characterize the spectral information of an input speech signal. Speech signal analysis [52-53] techniques

More information

The Noise Power Ratio - Theory and ADC Testing

The Noise Power Ratio - Theory and ADC Testing The Noise Power Ratio - Theory and ADC Testing FH Irons, KJ Riley, and DM Hummels Abstract This aer develos theory behind the noise ower ratio (NPR) testing of ADCs. A mid-riser formulation is used for

More information

Hidden Predictors: A Factor Analysis Primer

Hidden Predictors: A Factor Analysis Primer Hidden Predictors: A Factor Analysis Primer Ryan C Sanchez Western Washington University Factor Analysis is a owerful statistical method in the modern research sychologist s toolbag When used roerly, factor

More information

4. Score normalization technical details We now discuss the technical details of the score normalization method.

4. Score normalization technical details We now discuss the technical details of the score normalization method. SMT SCORING SYSTEM This document describes the scoring system for the Stanford Math Tournament We begin by giving an overview of the changes to scoring and a non-technical descrition of the scoring rules

More information

Chapter 2 Introductory Concepts of Wave Propagation Analysis in Structures

Chapter 2 Introductory Concepts of Wave Propagation Analysis in Structures Chater 2 Introductory Concets of Wave Proagation Analysis in Structures Wave roagation is a transient dynamic henomenon resulting from short duration loading. Such transient loadings have high frequency

More information

Iterative Methods for Designing Orthogonal and Biorthogonal Two-channel FIR Filter Banks with Regularities

Iterative Methods for Designing Orthogonal and Biorthogonal Two-channel FIR Filter Banks with Regularities R. Bregović and T. Saramäi, Iterative methods for designing orthogonal and biorthogonal two-channel FIR filter bans with regularities, Proc. Of Int. Worsho on Sectral Transforms and Logic Design for Future

More information

Self-Driving Car ND - Sensor Fusion - Extended Kalman Filters

Self-Driving Car ND - Sensor Fusion - Extended Kalman Filters Self-Driving Car ND - Sensor Fusion - Extended Kalman Filters Udacity and Mercedes February 7, 07 Introduction Lesson Ma 3 Estimation Problem Refresh 4 Measurement Udate Quiz 5 Kalman Filter Equations

More information

Pulse Propagation in Optical Fibers using the Moment Method

Pulse Propagation in Optical Fibers using the Moment Method Pulse Proagation in Otical Fibers using the Moment Method Bruno Miguel Viçoso Gonçalves das Mercês, Instituto Suerior Técnico Abstract The scoe of this aer is to use the semianalytic technique of the Moment

More information

Mel-Generalized Cepstral Representation of Speech A Unified Approach to Speech Spectral Estimation. Keiichi Tokuda

Mel-Generalized Cepstral Representation of Speech A Unified Approach to Speech Spectral Estimation. Keiichi Tokuda Mel-Generalized Cepstral Representation of Speech A Unified Approach to Speech Spectral Estimation Keiichi Tokuda Nagoya Institute of Technology Carnegie Mellon University Tamkang University March 13,

More information

On Fractional Predictive PID Controller Design Method Emmanuel Edet*. Reza Katebi.**

On Fractional Predictive PID Controller Design Method Emmanuel Edet*. Reza Katebi.** On Fractional Predictive PID Controller Design Method Emmanuel Edet*. Reza Katebi.** * echnology and Innovation Centre, Level 4, Deartment of Electronic and Electrical Engineering, University of Strathclyde,

More information

The Recursive Fitting of Multivariate. Complex Subset ARX Models

The Recursive Fitting of Multivariate. Complex Subset ARX Models lied Mathematical Sciences, Vol. 1, 2007, no. 23, 1129-1143 The Recursive Fitting of Multivariate Comlex Subset RX Models Jack Penm School of Finance and lied Statistics NU College of Business & conomics

More information

1 University of Edinburgh, 2 British Geological Survey, 3 China University of Petroleum

1 University of Edinburgh, 2 British Geological Survey, 3 China University of Petroleum Estimation of fluid mobility from frequency deendent azimuthal AVO a synthetic model study Yingrui Ren 1*, Xiaoyang Wu 2, Mark Chaman 1 and Xiangyang Li 2,3 1 University of Edinburgh, 2 British Geological

More information

Solution sheet ξi ξ < ξ i+1 0 otherwise ξ ξ i N i,p 1 (ξ) + where 0 0

Solution sheet ξi ξ < ξ i+1 0 otherwise ξ ξ i N i,p 1 (ξ) + where 0 0 Advanced Finite Elements MA5337 - WS7/8 Solution sheet This exercise sheets deals with B-slines and NURBS, which are the basis of isogeometric analysis as they will later relace the olynomial ansatz-functions

More information

Methods for detecting fatigue cracks in gears

Methods for detecting fatigue cracks in gears Journal of Physics: Conference Series Methods for detecting fatigue cracks in gears To cite this article: A Belšak and J Flašker 2009 J. Phys.: Conf. Ser. 181 012090 View the article online for udates

More information

Statics and dynamics: some elementary concepts

Statics and dynamics: some elementary concepts 1 Statics and dynamics: some elementary concets Dynamics is the study of the movement through time of variables such as heartbeat, temerature, secies oulation, voltage, roduction, emloyment, rices and

More information

Damage Identification from Power Spectrum Density Transmissibility

Damage Identification from Power Spectrum Density Transmissibility 6th Euroean Worksho on Structural Health Monitoring - h.3.d.3 More info about this article: htt://www.ndt.net/?id=14083 Damage Identification from Power Sectrum Density ransmissibility Y. ZHOU, R. PERERA

More information

Sinusoidal Modeling. Yannis Stylianou SPCC University of Crete, Computer Science Dept., Greece,

Sinusoidal Modeling. Yannis Stylianou SPCC University of Crete, Computer Science Dept., Greece, Sinusoidal Modeling Yannis Stylianou University of Crete, Computer Science Dept., Greece, yannis@csd.uoc.gr SPCC 2016 1 Speech Production 2 Modulators 3 Sinusoidal Modeling Sinusoidal Models Voiced Speech

More information

MATH 2710: NOTES FOR ANALYSIS

MATH 2710: NOTES FOR ANALYSIS MATH 270: NOTES FOR ANALYSIS The main ideas we will learn from analysis center around the idea of a limit. Limits occurs in several settings. We will start with finite limits of sequences, then cover infinite

More information

AI*IA 2003 Fusion of Multiple Pattern Classifiers PART III

AI*IA 2003 Fusion of Multiple Pattern Classifiers PART III AI*IA 23 Fusion of Multile Pattern Classifiers PART III AI*IA 23 Tutorial on Fusion of Multile Pattern Classifiers by F. Roli 49 Methods for fusing multile classifiers Methods for fusing multile classifiers

More information

MULTI-CHANNEL PARAMETRIC ESTIMATOR FAST BLOCK MATRIX INVERSES

MULTI-CHANNEL PARAMETRIC ESTIMATOR FAST BLOCK MATRIX INVERSES MULTI-CANNEL ARAMETRIC ESTIMATOR FAST BLOCK MATRIX INVERSES S Lawrence Marle Jr School of Electrical Engineering and Comuter Science Oregon State University Corvallis, OR 97331 Marle@eecsoregonstateedu

More information

A PEAK FACTOR FOR PREDICTING NON-GAUSSIAN PEAK RESULTANT RESPONSE OF WIND-EXCITED TALL BUILDINGS

A PEAK FACTOR FOR PREDICTING NON-GAUSSIAN PEAK RESULTANT RESPONSE OF WIND-EXCITED TALL BUILDINGS The Seventh Asia-Pacific Conference on Wind Engineering, November 8-1, 009, Taiei, Taiwan A PEAK FACTOR FOR PREDICTING NON-GAUSSIAN PEAK RESULTANT RESPONSE OF WIND-EXCITED TALL BUILDINGS M.F. Huang 1,

More information

Distributed Rule-Based Inference in the Presence of Redundant Information

Distributed Rule-Based Inference in the Presence of Redundant Information istribution Statement : roved for ublic release; distribution is unlimited. istributed Rule-ased Inference in the Presence of Redundant Information June 8, 004 William J. Farrell III Lockheed Martin dvanced

More information

Vibration Analysis to Determine the Condition of Gear Units

Vibration Analysis to Determine the Condition of Gear Units UDC 61.83.05 Strojniški vestnik - Journal of Mechanical Engineering 54(008)1, 11-4 Paer received: 10.7.006 Paer acceted: 19.1.007 Vibration Analysis to Determine the Condition of Gear Units Aleš Belšak*

More information

PROFIT MAXIMIZATION. π = p y Σ n i=1 w i x i (2)

PROFIT MAXIMIZATION. π = p y Σ n i=1 w i x i (2) PROFIT MAXIMIZATION DEFINITION OF A NEOCLASSICAL FIRM A neoclassical firm is an organization that controls the transformation of inuts (resources it owns or urchases into oututs or roducts (valued roducts

More information

L.J. Faber S.T. Alexander. Department of Electrical and Computer Engineering. North Carolina state University CCSP-TR-85/4

L.J. Faber S.T. Alexander. Department of Electrical and Computer Engineering. North Carolina state University CCSP-TR-85/4 RESDUAL THRESHOLDNG AND CROSS-CORRELATON APPROACHES TO MULTPULSE LPC By L.J. Faber S.T. Alexander CENTER FOR COMMUNCATONS AND SGNAL PROCESSNG Deartment of Electrical and Comuter Engineering North Carolina

More information

convenient means to determine response to a sum of clear evidence of signal properties that are obscured in the original signal

convenient means to determine response to a sum of clear evidence of signal properties that are obscured in the original signal Digital Speech Processing Lecture 9 Short-Time Fourier Analysis Methods- Introduction 1 General Discrete-Time Model of Speech Production Voiced Speech: A V P(z)G(z)V(z)R(z) Unvoiced Speech: A N N(z)V(z)R(z)

More information

Applied Fitting Theory VI. Formulas for Kinematic Fitting

Applied Fitting Theory VI. Formulas for Kinematic Fitting Alied Fitting heory VI Paul Avery CBX 98 37 June 9, 1998 Ar. 17, 1999 (rev.) I Introduction Formulas for Kinematic Fitting I intend for this note and the one following it to serve as mathematical references,

More information

Statistics II Logistic Regression. So far... Two-way repeated measures ANOVA: an example. RM-ANOVA example: the data after log transform

Statistics II Logistic Regression. So far... Two-way repeated measures ANOVA: an example. RM-ANOVA example: the data after log transform Statistics II Logistic Regression Çağrı Çöltekin Exam date & time: June 21, 10:00 13:00 (The same day/time lanned at the beginning of the semester) University of Groningen, Det of Information Science May

More information

Outline. Markov Chains and Markov Models. Outline. Markov Chains. Markov Chains Definitions Huizhen Yu

Outline. Markov Chains and Markov Models. Outline. Markov Chains. Markov Chains Definitions Huizhen Yu and Markov Models Huizhen Yu janey.yu@cs.helsinki.fi Det. Comuter Science, Univ. of Helsinki Some Proerties of Probabilistic Models, Sring, 200 Huizhen Yu (U.H.) and Markov Models Jan. 2 / 32 Huizhen Yu

More information

arxiv: v1 [physics.data-an] 26 Oct 2012

arxiv: v1 [physics.data-an] 26 Oct 2012 Constraints on Yield Parameters in Extended Maximum Likelihood Fits Till Moritz Karbach a, Maximilian Schlu b a TU Dortmund, Germany, moritz.karbach@cern.ch b TU Dortmund, Germany, maximilian.schlu@cern.ch

More information

Optimal Recognition Algorithm for Cameras of Lasers Evanescent

Optimal Recognition Algorithm for Cameras of Lasers Evanescent Otimal Recognition Algorithm for Cameras of Lasers Evanescent T. Gaudo * Abstract An algorithm based on the Bayesian aroach to detect and recognise off-axis ulse laser beams roagating in the atmoshere

More information

%(*)= E A i* eiujt > (!) 3=~N/2

%(*)= E A i* eiujt > (!) 3=~N/2 CHAPTER 58 Estimating Incident and Reflected Wave Fields Using an Arbitrary Number of Wave Gauges J.A. Zelt* A.M. ASCE and James E. Skjelbreia t A.M. ASCE 1 Abstract A method based on linear wave theory

More information

Timbral, Scale, Pitch modifications

Timbral, Scale, Pitch modifications Introduction Timbral, Scale, Pitch modifications M2 Mathématiques / Vision / Apprentissage Audio signal analysis, indexing and transformation Page 1 / 40 Page 2 / 40 Modification of playback speed Modifications

More information

ME scope Application Note 16

ME scope Application Note 16 ME scoe Alication Note 16 Integration & Differentiation of FFs and Mode Shaes NOTE: The stes used in this Alication Note can be dulicated using any Package that includes the VES-36 Advanced Signal Processing

More information

Mobius Functions, Legendre Symbols, and Discriminants

Mobius Functions, Legendre Symbols, and Discriminants Mobius Functions, Legendre Symbols, and Discriminants 1 Introduction Zev Chonoles, Erick Knight, Tim Kunisky Over the integers, there are two key number-theoretic functions that take on values of 1, 1,

More information

Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning

Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning TNN-2009-P-1186.R2 1 Uncorrelated Multilinear Princial Comonent Analysis for Unsuervised Multilinear Subsace Learning Haiing Lu, K. N. Plataniotis and A. N. Venetsanooulos The Edward S. Rogers Sr. Deartment

More information

arxiv:cond-mat/ v2 25 Sep 2002

arxiv:cond-mat/ v2 25 Sep 2002 Energy fluctuations at the multicritical oint in two-dimensional sin glasses arxiv:cond-mat/0207694 v2 25 Se 2002 1. Introduction Hidetoshi Nishimori, Cyril Falvo and Yukiyasu Ozeki Deartment of Physics,

More information

An Inverse Problem for Two Spectra of Complex Finite Jacobi Matrices

An Inverse Problem for Two Spectra of Complex Finite Jacobi Matrices Coyright 202 Tech Science Press CMES, vol.86, no.4,.30-39, 202 An Inverse Problem for Two Sectra of Comlex Finite Jacobi Matrices Gusein Sh. Guseinov Abstract: This aer deals with the inverse sectral roblem

More information

16.2. Infinite Series. Introduction. Prerequisites. Learning Outcomes

16.2. Infinite Series. Introduction. Prerequisites. Learning Outcomes Infinite Series 6.2 Introduction We extend the concet of a finite series, met in Section 6., to the situation in which the number of terms increase without bound. We define what is meant by an infinite

More information

A Dual-Tree Rational-Dilation Complex Wavelet Transform

A Dual-Tree Rational-Dilation Complex Wavelet Transform 1 A Dual-Tree Rational-Dilation Comlex Wavelet Transform İlker Bayram and Ivan W. Selesnick Abstract In this corresondence, we introduce a dual-tree rational-dilation comlex wavelet transform for oscillatory

More information

Digital Speech Processing Lecture 14. Linear Predictive Coding (LPC)-Lattice Methods, Applications

Digital Speech Processing Lecture 14. Linear Predictive Coding (LPC)-Lattice Methods, Applications Dgtal Seech Processng Lecture 14 Lnear Predctve Codng (LPC)-Lattce Methods, Alcatons 1 Predcton Error Sgnal 1. Seech Producton Model sn ( ) = asn ( k) + Gun ( ) 2. LPC Model: k = 1 Sz ( ) Hz ( ) = = Uz

More information

M. Hasegawa-Johnson. DRAFT COPY.

M. Hasegawa-Johnson. DRAFT COPY. Lecture Notes in Speech Production, Speech Coding, and Speech Recognition Mark Hasegawa-Johnson University of Illinois at Urbana-Champaign February 7, 000 M. Hasegawa-Johnson. DRAFT COPY. Chapter Linear

More information

Finite Mixture EFA in Mplus

Finite Mixture EFA in Mplus Finite Mixture EFA in Mlus November 16, 2007 In this document we describe the Mixture EFA model estimated in Mlus. Four tyes of deendent variables are ossible in this model: normally distributed, ordered

More information

Comparative study on different walking load models

Comparative study on different walking load models Comarative study on different walking load models *Jining Wang 1) and Jun Chen ) 1), ) Deartment of Structural Engineering, Tongji University, Shanghai, China 1) 1510157@tongji.edu.cn ABSTRACT Since the

More information

Uniformly best wavenumber approximations by spatial central difference operators: An initial investigation

Uniformly best wavenumber approximations by spatial central difference operators: An initial investigation Uniformly best wavenumber aroximations by satial central difference oerators: An initial investigation Vitor Linders and Jan Nordström Abstract A characterisation theorem for best uniform wavenumber aroximations

More information

Th P10 13 Alternative Misfit Functions for FWI Applied to Surface Waves

Th P10 13 Alternative Misfit Functions for FWI Applied to Surface Waves Th P0 3 Alternative Misfit Functions for FWI Alied to Surface Waves I. Masoni* Total E&P, Joseh Fourier University), R. Brossier Joseh Fourier University), J. Virieux Joseh Fourier University) & J.L. Boelle

More information

Supplementary Information for Quantum nondemolition measurement of mechanical squeezed state beyond the 3 db limit

Supplementary Information for Quantum nondemolition measurement of mechanical squeezed state beyond the 3 db limit Sulementary Information for Quantum nondemolition measurement of mechanical squeezed state beyond the db limit C. U. Lei, A. J. Weinstein, J. Suh, E. E. Wollman, A. Kronwald,, F. Marquardt,, A. A. Clerk,

More information

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi Signal Modeling Techniques in Speech Recognition Hassan A. Kingravi Outline Introduction Spectral Shaping Spectral Analysis Parameter Transforms Statistical Modeling Discussion Conclusions 1: Introduction

More information

Chapter 10. Classical Fourier Series

Chapter 10. Classical Fourier Series Math 344, Male ab Manual Chater : Classical Fourier Series Real and Comle Chater. Classical Fourier Series Fourier Series in PS K, Classical Fourier Series in PS K, are aroimations obtained using orthogonal

More information

An Improved Calibration Method for a Chopped Pyrgeometer

An Improved Calibration Method for a Chopped Pyrgeometer 96 JOURNAL OF ATMOSPHERIC AND OCEANIC TECHNOLOGY VOLUME 17 An Imroved Calibration Method for a Choed Pyrgeometer FRIEDRICH FERGG OtoLab, Ingenieurbüro, Munich, Germany PETER WENDLING Deutsches Forschungszentrum

More information

Convex Optimization methods for Computing Channel Capacity

Convex Optimization methods for Computing Channel Capacity Convex Otimization methods for Comuting Channel Caacity Abhishek Sinha Laboratory for Information and Decision Systems (LIDS), MIT sinhaa@mit.edu May 15, 2014 We consider a classical comutational roblem

More information

Applications of Linear Prediction

Applications of Linear Prediction SGN-4006 Audio and Speech Processing Applications of Linear Prediction Slides for this lecture are based on those created by Katariina Mahkonen for TUT course Puheenkäsittelyn menetelmät in Spring 03.

More information

Analysis of Pressure Transient Response for an Injector under Hydraulic Stimulation at the Salak Geothermal Field, Indonesia

Analysis of Pressure Transient Response for an Injector under Hydraulic Stimulation at the Salak Geothermal Field, Indonesia roceedings World Geothermal Congress 00 Bali, Indonesia, 5-9 Aril 00 Analysis of ressure Transient Resonse for an Injector under Hydraulic Stimulation at the Salak Geothermal Field, Indonesia Jorge A.

More information

Math 104B: Number Theory II (Winter 2012)

Math 104B: Number Theory II (Winter 2012) Math 104B: Number Theory II (Winter 01) Alina Bucur Contents 1 Review 11 Prime numbers 1 Euclidean algorithm 13 Multilicative functions 14 Linear diohantine equations 3 15 Congruences 3 Primes as sums

More information

Multilayer Perceptron Neural Network (MLPs) For Analyzing the Properties of Jordan Oil Shale

Multilayer Perceptron Neural Network (MLPs) For Analyzing the Properties of Jordan Oil Shale World Alied Sciences Journal 5 (5): 546-552, 2008 ISSN 1818-4952 IDOSI Publications, 2008 Multilayer Percetron Neural Network (MLPs) For Analyzing the Proerties of Jordan Oil Shale 1 Jamal M. Nazzal, 2

More information

WAVELETS, PROPERTIES OF THE SCALAR FUNCTIONS

WAVELETS, PROPERTIES OF THE SCALAR FUNCTIONS U.P.B. Sci. Bull. Series A, Vol. 68, No. 4, 006 WAVELETS, PROPERTIES OF THE SCALAR FUNCTIONS C. PANĂ * Pentru a contrui o undină convenabilă Ψ este necesară şi suficientă o analiză multirezoluţie. Analiza

More information

where x i is the ith coordinate of x R N. 1. Show that the following upper bound holds for the growth function of H:

where x i is the ith coordinate of x R N. 1. Show that the following upper bound holds for the growth function of H: Mehryar Mohri Foundations of Machine Learning Courant Institute of Mathematical Sciences Homework assignment 2 October 25, 2017 Due: November 08, 2017 A. Growth function Growth function of stum functions.

More information

PHYS 301 HOMEWORK #9-- SOLUTIONS

PHYS 301 HOMEWORK #9-- SOLUTIONS PHYS 0 HOMEWORK #9-- SOLUTIONS. We are asked to use Dirichlet' s theorem to determine the value of f (x) as defined below at x = 0, ± /, ± f(x) = 0, - < x

More information

On split sample and randomized confidence intervals for binomial proportions

On split sample and randomized confidence intervals for binomial proportions On slit samle and randomized confidence intervals for binomial roortions Måns Thulin Deartment of Mathematics, Usala University arxiv:1402.6536v1 [stat.me] 26 Feb 2014 Abstract Slit samle methods have

More information