Amplitude Modulated Model For Analyzing Non Stationary Speech Signals

Similar documents
ESTIMATION OF PARAMETERS OF PARTIALLY SINUSOIDAL FREQUENCY MODEL

An Efficient and Fast Algorithm for Estimating the Parameters of Sinusoidal Signals

PARAMETER ESTIMATION OF CHIRP SIGNALS IN PRESENCE OF STATIONARY NOISE

PARAMETER ESTIMATION OF CHIRP SIGNALS IN PRESENCE OF STATIONARY NOISE

An Efficient and Fast Algorithm for Estimating the Parameters of Two-Dimensional Sinusoidal Signals

On Two Different Signal Processing Models

ASYMPTOTIC PROPERTIES OF THE LEAST SQUARES ESTIMATORS OF MULTIDIMENSIONAL EXPONENTIAL SIGNALS

On Least Absolute Deviation Estimators For One Dimensional Chirp Model

Asymptotic properties of the least squares estimators of a two dimensional model

On Parameter Estimation of Two Dimensional Chirp Signal

EFFICIENT ALGORITHM FOR ESTIMATING THE PARAMETERS OF CHIRP SIGNAL

Estimating Periodic Signals

Asymptotic of Approximate Least Squares Estimators of Parameters Two-Dimensional Chirp Signal

Analysis of Type-II Progressively Hybrid Censored Data

Asymptotic inference for a nonstationary double ar(1) model

Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring

Analysis of Middle Censored Data with Exponential Lifetime Distributions

Inference for P(Y<X) in Exponentiated Gumbel Distribution

F & B Approaches to a simple model

Design Criteria for the Quadratically Interpolated FFT Method (I): Bias due to Interpolation

A New Two Sample Type-II Progressive Censoring Scheme

A NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A.

Determination of Discrete Spectrum in a Random Field

Wavelet Methods for Time Series Analysis. Part IV: Wavelet-Based Decorrelation of Time Series

IN this paper, we consider the estimation of the frequency

A Subspace Approach to Estimation of. Measurements 1. Carlos E. Davila. Electrical Engineering Department, Southern Methodist University

Noise Space Decomposition Method for two dimensional sinusoidal model

Local Whittle Likelihood Estimators and Tests for non-gaussian Linear Processes

EXTENDED GLRT DETECTORS OF CORRELATION AND SPHERICITY: THE UNDERSAMPLED REGIME. Xavier Mestre 1, Pascal Vallet 2

On Moving Average Parameter Estimation

Covariance function estimation in Gaussian process regression

ROBUST FREQUENCY DOMAIN ARMA MODELLING. Jonas Gillberg Fredrik Gustafsson Rik Pintelon

Non-parametric identification

Quasi Stochastic Approximation American Control Conference San Francisco, June 2011

COMPUTER ALGEBRA DERIVATION OF THE BIAS OF LINEAR ESTIMATORS OF AUTOREGRESSIVE MODELS

Acomplex-valued harmonic with a time-varying phase is a

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data

Estimating the numberof signals of the damped exponential models

On Input Design for System Identification

LIST OF PUBLICATIONS

Hypothesis Testing - Frequentist

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

CLOSED-FORM FREQUENCY ESTIMATION USING SECOND-ORDER NOTCH FILTERS. S.M. Savaresi, S. Bittanti, H.C. So*

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

Ross Bettinger, Analytical Consultant, Seattle, WA

Linear Dependency Between and the Input Noise in -Support Vector Regression

Copyright (c) 2006 Warren Weckesser

UNIVERSITÄT POTSDAM Institut für Mathematik

Math 312 Lecture Notes Linear Two-dimensional Systems of Differential Equations

Residuals in Time Series Models

Time-Varying Parameters

Multi-armed bandit models: a tutorial

Time Series. Anthony Davison. c

MA Advanced Econometrics: Applying Least Squares to Time Series

8.2 Harmonic Regression and the Periodogram

An algorithm for robust fitting of autoregressive models Dimitris N. Politis

AGEC 661 Note Eleven Ximing Wu. Exponential regression model: m (x, θ) = exp (xθ) for y 0

FAST AND ACCURATE DIRECTION-OF-ARRIVAL ESTIMATION FOR A SINGLE SOURCE

Analysis of Gamma and Weibull Lifetime Data under a General Censoring Scheme and in the presence of Covariates

Discriminating Between the Bivariate Generalized Exponential and Bivariate Weibull Distributions

LTI Systems, Additive Noise, and Order Estimation

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015

Comments on New Approaches in Period Analysis of Astronomical Time Series by Pavlos Protopapas (Or: A Pavlosian Response ) Don Percival

Using all observations when forecasting under structural breaks

EECE Adaptive Control

Median Filter Based Realizations of the Robust Time-Frequency Distributions

Change-Point Estimation

Bandit models: a tutorial

Statistics of Stochastic Processes

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno

"ZERO-POINT" IN THE EVALUATION OF MARTENS HARDNESS UNCERTAINTY

Massachusetts Institute of Technology Department of Economics Time Series Lecture 6: Additional Results for VAR s

Generalised AR and MA Models and Applications

ARMA MODELS Herman J. Bierens Pennsylvania State University February 23, 2009

On the Complexity of Best Arm Identification with Fixed Confidence

On the Power of Tests for Regime Switching

Asymptotic Analysis of the Generalized Coherence Estimate

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of

Testing Algebraic Hypotheses

Terence Tai-Leung Chong. Abstract

EE538 Final Exam Fall :20 pm -5:20 pm PHYS 223 Dec. 17, Cover Sheet

Analysis of incomplete data in presence of competing risks

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection

KALMAN-TYPE RECURSIONS FOR TIME-VARYING ARMA MODELS AND THEIR IMPLICATION FOR LEAST SQUARES PROCEDURE ANTONY G AU T I E R (LILLE)

Time Series Examples Sheet

Statistical Methods for Forecasting

The exact bias of S 2 in linear panel regressions with spatial autocorrelation SFB 823. Discussion Paper. Christoph Hanck, Walter Krämer

14 - Gaussian Stochastic Processes

Chapter 7. Hypothesis Testing

Empirical Market Microstructure Analysis (EMMA)

Lecture 3. Inference about multivariate normal distribution

Time Series: Theory and Methods

Nonconcave Penalized Likelihood with A Diverging Number of Parameters

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

and Srikanth K. Iyer Department of Mathematics Indian Institute of Technology, Kanpur, India. ph:

Time series models in the Frequency domain. The power spectrum, Spectral analysis

Worst-Case Bounds for Gaussian Process Models

Regression of Time Series

EIE6207: Estimation Theory

Transcription:

Amplitude Modulated Model For Analyzing on Stationary Speech Signals Swagata andi, Debasis Kundu and Srikanth K. Iyer Institut für Angewandte Mathematik Ruprecht-Karls-Universität Heidelberg Im euenheimer Feld 94 690 Heidelberg, Germany Department of Mathematics Indian Institute of Technology Kanpur Kanpur - 0806 India Abstract Recently Amplitude Modulated model in presence of additive white noise was used to analyze certain non-stationary speech data. It is observed that the assumption of white noise may not be proper in many cases. In this paper we consider the Amplitude Modulated signal model in presence of stationary noise. We consider the least squares estimators and the estimators obtained by maximizing the Periodogram function. The two estimators are asymptotically equivalent. We study the theoretical properties of both estimators and observe their performances through numerical simulations. One speech data is analyzed and it is observed that the performance of the proposed estimators is quite satisfactory. Key Words and Phrases: Strong consistency, frequencies, amplitudes, asymptotic distribution. Short Running Title: AM signal model. Corresponding Author: Debasis Kundu. e:mail:kundu@iitk.ac.in, FAX: 9-5-590007, Phone: 9-5-5974.

. ITRODUCTIO In signal processing, the signal is often assumed to be stationary. In real life, many signals, like speech are non-stationary in nature. Traditionally, the parametric modeling of a non-stationary signal has been carried out using the quasi-stationary models (McAulay and Quatieri; 986 and Isaksson, Wennberg and Zetterberg; 98) where the signal is treated to be stationary only over a short duration of time. The usefulness of these models is restricted due to contradictory requirements for the duration of observations of the signals. On one hand, the duration must be short for the faithfulness of the model; on the other hand, the duration must be long enough to assure accurate estimation of the parameters of the model. It is well known that the time dependent ARMA model provides a general framework for parametric modeling of non-stationary signals (Grenier; 983). There are several nonlinear time series models available in the seminal book of Tong (990). Unfortunately, these approaches are far too general and often lead to difficult problems when estimating a large number of parameters. Fortunately, by exploiting certain known properties for a particular class of signals often it is possible to find a simple model which serves the purpose of representation of signals efficiently. One such model was introduced by Sircar and Syali (996), named as complex Amplitude Modulated (AM) model. It was used to analyze non-stationary speech signals. They proposed certain estimation procedures and the performances were quite satisfactory. They did not study the theoretical properties of the estimators. Moreover, the model validation was also not performed. While re-analyzing the same speech data, we observe that the independent and identically distributed (i.i.d.) error assumptions may not be reasonable. It may be more appropriate to assume that the errors are correlated. Unfortunately in that case the estimation procedure proposed by Sircar and Syali (996) can not be generalized and also obtaining the theoretical properties of these estimators will not be a trivial task. To

make the model more general and also at the same time analytically tractable, we assume that the errors are from a stationary distribution. The main aim of this paper is to define the AM signal model in presence of an additive stationary noise. We propose two estimators. It is observed that both estimators are consistent and we obtain the asymptotic distributions of both the estimators. The asymptotic distribution can be used to construct error bounds, without which the point estimators do not have much value in practice. It is observed that the two estimators are asymptotically equivalent. The small sample performances of the two estimators are compared using numerical simulations. We also analyze a speech data using the proposed method and the performance is quite satisfactory. The rest of the paper is organized as follows. In section, we give the description of the different model assumptions and provide different estimation procedures. The theoretical properties are derived in section 3. A speech data is analyzed in section 4 and finally we conclude the paper in section 5.. MODEL DESCRIPTIO AD ESTIMATIO PROCEDURES The discrete-time complex random process y(t) consisting of M single-tone AM signals is given by M [ y(t) = A k + µk e ] iν kt e iωkt + X(t); t =,...,, (.) k= where A k is the carrier amplitude of constituent signal, µ k is the modulation index, ω k is the carrier angular frequency, ν k is the modulating angular frequency and i =. For physical interpretation of the different parameters see Sircar and Syali (996). The following assumptions are made on the model parameters; Assumption A k 0, µ k 0 and they are bounded and also 0 < ν k < π, 0 < ω k < π for 3

all k. Moreover ω < ω + ν < ω < ω + ν < < ω M < ω M + ν M. (.) The additive error X(t) is a stationary sequence and it satisfies assumption. Assumption : X(t) has the following representation X(t) = a(k)e(t k), k= where e(t) s are i.i.d. complex valued random variables with mean zero and variance σ for both the real and imaginary parts. The real and imaginary parts of e(t) are uncorrelated. a(k) s are arbitrary complex-valued constants such that a(k) <. k= The real and imaginary parts of a(k) will be denoted as a R (k) and a I (k) and of e(t) as e R (t) and e I (t) respectively. We assume M is known. In this paper we mainly consider the estimation of the unknown parameters A k, µ k, ν k and ω k and study their properties. We mainly consider two estimators. The first one is the least squares estimators (LSEs), which can be obtained by minimizing M Q(A, µ, ν, ω) = y(t) A k ( + µ k e iνkt )e iω kt, (.3) k= with respect to A = (A,..., A M ), µ = (µ,..., µ M ), ν = (ν,..., ν M ), ω = (ω,..., ω M ) and subject to the restriction (.). We will denote them as  = (Â,..., ÂM), ˆµ = (ˆµ,..., ˆµ M ), ˆν = (ˆν,..., ˆν M ) and ˆω = (ˆω,..., ˆω M ) respectively. The second estimator is called the approximate least squares estimators (ALSEs) and it can be obtained by maximizing the Periodogram function, defined as follows; M I(ν, ω) = y(t)e iω kt + k= 4 y(t)e i(ω k+ν k )t (.4)

under the restriction (.). Let us denote the estimators as follows; ω < ω + ν < ω < ω + ν < < ω M < ω M + ν M. The ( ω k, ν k ) is the ALSE of (ω k, ν k ), for k =,..., M. The corresponding ALSEs of the linear parameters of A k and µ k can be obtained from the following equations; Ã k = y(t)e i ω kt, Ã k µ k = y(t)e i( ω k+ ν k )t. (.5) In the next section we consider the estimates of the parameters and study their properties. ote that although maximization of (.4) is a M dimensional maximization problem, it can be performed sequentially, i.e. the M dimensional maximization problem can be reduced to M, one dimensional maximization problems. The main idea of using the ALSEs goes back to Walker (97) and Hannan (97). Along the same line as Walker (97) it can be shown by expanding (.3) that the LSEs and the ALSEs are asymptotically equivalent. It indicates that the ALSEs also can be used as an alternative to the LSEs. 3. THEORETICAL RESULTS In this section we mainly consider the asymptotic properties of the LSEs and the ALSEs. We state the main results here, proofs of all the results are provided in the appendix. It may be mentioned that the model (.) does not satisfy the standard sufficient conditions of Jennrich (969), Wu (98) or Kundu (99) for the LSEs to be consistent. Therefore, although the least squares method usually provides satisfactory performance, the complexity of the model makes it unclear, how good the LSEs will be in the present situation. It may be mentioned that when the modulation index µ k = 0 for all k, then the model (.) coincides 5

with the sum of complex exponential models. The theoretical properties of the LSEs of the complex exponential model were discussed by Bai et al. (99), Rao and Zhao (993) and Kundu and Mitra (999) in great details when the errors are i.i.d. random variables. For brevity, first we consider M = in (.), i.e. we have the following model; y(t) = A( + µe iνt )e iωt + X(t). (3.) We use the following notation. A R and A I denote the real and imaginary parts of A, similarly µ R and µ I are defined, θ = (A R, A I, µ R, µ I, ν, ω). The LSE and the ALSE of θ will be denoted by ˆθ = (ÂR, ÂI, ˆµ R, ˆµ I, ˆν, ˆω) and θ = (ÃR, ÃI, µ R, µ I, ν, ω) respectively. For model (3.), the assumption is equivalent to the following assumption. Assumption : A 0 and µ 0 are bounded and ν, ω (0, π). We have the following results for model (3.). Theorem : Under assumptions and, ˆθ is a strongly consistent estimator of θ. Theorem : Under assumptions and, θ is a strongly consistent estimator of θ. Theorem 3: Under assumptions and, { (  R A R ), (  I A I ), (ˆµR µ R ), (ˆµI µ I ), 3 (ˆν ν), 3 (ˆω ω)} converges to a 6-variate normal distribution with mean vector 0 and the dispersion matrix σ Σ (c Σ + c Σ )Σ, where c = k= a(k)e iωk and c = a(k)e ik(ω+ν) k=. 6

Σ = 0 0 0 0 A I A 0 0 0 0 R 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A I A R 0 0 0 3 A, µ 0 Re( µa) Im( µa) A I µ A I µ 0 µ A Im( µa) Re( µa) R µ A R µ Σ = Re( µa) Im( µa) A A 0 µ A I µ I Im( µa) Re( µa) 0 A A µ A R µ. R A I µ A R µ A µ A I µ R 3 µ A 3 µ A A I µ A R µ A µ A I µ R Here µ denotes the complex conjugate of µ and 3 µ A 3 µ A Σ = Σ + Σ. The matrix Σ = σ mn, m, n =,... 6 has the following elements. σ = + 3A I A, σ = σ = 3A IA R A, σ 3 = σ 3 = Re( µa) 3µ IA I A, σ 4 = σ 4 = Im( µa) + 3µ RA I A, σ 5 = σ 5 = 6A I A, σ6 = σ 6 = 6A I A, σ = + 3A R A, σ3 = σ 3 = Im( µa) + 3A Rµ I A, σ 4 = σ 4 = Re( µa) 3A Rµ R A, σ 5 = σ 5 = 6A R A, σ6 = σ 6 = 6A R A, σ33 = ( + µ ) ( + 3µ I A µ ), σ 34 = σ 43 = 3µ Rµ I ( + µ ) A µ, σ 35 = σ 53 = 6µ I( + µ ) A µ, σ 36 = σ 63 = 6µ I A, σ 44 = ( + µ ) ( + 3µ R A µ ), σ45 = σ 54 = 6µ R( + µ ), σ 46 = σ 64 = 6µ R A µ A, σ 55 = ( + µ ) A µ, σ 56 = σ 65 = A, σ66 = A. 7

Theorem 4: Under assumptions and, the ALSEs have the same asymptotic distributions as the LSEs. Theorems and indicate that the LSEs and the ALSEs are reasonable estimators of the unknown parameters. Strong consistency ensures that when the sample size is large, then both the LSEs and the ALSEs should be quite close to the corresponding true parameter values. How good or how close the estimators will be can be found from Theorems 3 and 4. Theorems 3 and 4 indicate that both the LSEs and the ALSEs have the same rate of convergence. It is clear that the rate of convergence of the frequencies is higher compared to the rate of convergence of the amplitudes or modulation indexes. Therefore, for a given sample size the frequency estimators will be much better compared to the amplitude and modulation index estimators. For general M, the results can be easily extended under the assumption that all the frequencies are distinct. Theorems and are still valid replacing θ by the entire set of parameters. Theorems 3 and 4 also can be extended. For general M, the asymptotic dispersion matrix will be a 6M 6M matrix, with block diagonal form of M blocks each of size 6 6. Other blocks have only zero entries. Each diagonal block has the same form as defined in Theorem 3. 4. UMERICAL RESULTS In this section, first we compare the performances of the LSEs and the ALSEs for finite sample by computer simulations and then we analyze one non-stationary real speech data. All the computations are performed at the Indian Institute of Technology Kanpur using FORTRA-77 on the Silicon Graphics machine and they can be obtained from the authors. For computer simulations we use the random deviate generator from Press et al. (993). 8

Example : First we consider the data generated from the model (3.), with A = A R +ia I = 5+i.0, µ = µ R + iµ I =.5 + i.0, ν =.5086, ω =.043. Here X(t) is a stationary sequence which has the following form; X(t) = a 0 e(t) + a e(t ), where a 0 = 0. + i0.4 and a = 0.3 + i0.5. The real and the imaginary parts of e(t) are independent and normally distributed each with mean zero and variance one. e(t) s are i.i.d. The data is generated at 50 points. We compute the LSEs and the ALSEs of the unknown parameters and also compute the 95% confidence bound for each parameter. The process is repeated 5000 times and we compute the average estimates, average biases, variances, the average confidence lengths and the coverage percentages over five thousand replications for all the unknown parameters. The results are reported in Tables and. For comparison purposes we also report the asymptotic variances and the expected confidence lengths, as obtained from Theorems 3 and 4. ote that to compute the confidence intervals of the different parameters, we need to estimate σ, c and c. Although, we can not estimate σ, c and c separately, but it is possible to estimate σ c and σ c, which are needed to compute the confidence bands. By straight forward and lengthy calculations, it can be shown that σ c = E X(t)e iωt, σ c = E X(t)e i(ω+ν)t. Since σ c and σ c are the expected values of the Periodogram function at ω and (ω + ν) respectively, we estimate σ c and σ c by averaging the Periodogram function over a window (-L, L) across the point estimates of ω and (ω + ν). This estimator has been proposed by Hannan (970, page 470) in a different context but it was exploited later on by Quinn and Thomson (99). This estimator works reasonably well. We present the results for LSEs and ALSEs in Tables and respectively. 9

Some of the points are quite clear from Tables and. Both the LSEs and ALSEs work reasonably well even for small samples but the biases and the variances of the ALSEs are slightly larger than the corresponding biases and variances of the LSEs. As the theory suggests, it is observed that for both the LSEs and ALSEs the frequency estimates are much better than the amplitude and modulation index estimates in terms of the biases and variances. The variances of the LSEs are quite close to the asymptotic variances, but the same thing can not be said for the ALSEs. From the results, it is clear that the estimation of σ c and σ c are also quite good at least when the LSEs are used. It reflects in the average confidence length calculations and in the coverage percentages. For the LSEs the average confidence lengths are closer to the expected confidence lengths and also the coverage percentages are quite close to the nominal level. Interestingly most of the times the average confidence lengths based on the ALSEs are larger than the corresponding confidence lengths based on the LSEs but the coverage probabilities for the LSEs are higher than the ALSEs. ote that the expected confidence lengths are based on the true values of σ c and σ c. It may be mentioned that computationally LSEs are more involved than the ALSEs. Comparing all the points we recommend to use the LSEs to estimate the unknown parameters for the AM model if the sample size is not very large even if it is computationally more expensive. If the sample size is large we can use the ALSEs. For better performance, when the sample size is large, LSE can be computed using ALSE as an initial estimate. For illustration purpose, we consider one particular realization of the model presented in example. The real and imaginary parts of the data are plotted in Figure and Figure respectively. The Periodogram function (.4) of the data is provided in Figure 3. The Periodogram function clearly indicates that M =. Assuming M =, from the Periodogram function the initial estimates of ω and ν are.0097 and 0.5063 respectively. Using these initial estimates, the LSEs of A R, A I, µ R, µ I, ω and ν become 5.04679,.068, 0.4975,.00665, 0

.039 and 0.5097 respectively. The real and imaginary parts of the estimated signal are plotted in Figures 4 and 5 respectively. The confidence intervals of A R, A I, µ R, µ I, ω and ν are (4.90734, 5.863), (0.7675,.8647), (0.4358, 0.54968), (0.96795,.04534), (.0096,.036) and (0.5007, 0.50487) respectively. Example : In this example we re-analyze the sustained vowel sound of uuu. It was analyzed by Sircar and Syali (996) also. A total of 5 signal values sampled at 0kHz frequency is available. Sircar and Syali (996) used the model (.) while analyzing the data assuming that X(t) s are i.i.d. random variables. They did not study the residuals to verify the model assumptions. While re-analyzing the data, we observe that the residuals are correlated, therefore the assumptions of i.i.d. errors may not be reasonable. The plot of the original data is provided in Figure 6 and the plot of the Periodogram function is provided in Figure 7. The Periodogram function clearly indicates that M =, therefore, the model is of the form; y(t) = A ( + µ e iνt )e iωt + A ( + µ e iνt )e iωt + X(t). (4.) We obtain the estimates of the different parameters and also the 95% confidence intervals for all the parameters. They are presented in Table 3. ow we obtain the predicted value of y(t) as ŷ(t) = Â( + ˆµ e iˆν t )e iˆω t + Â( + ˆµ e iˆν t )e iˆω t (4.) and the estimated error as ˆX(t) = y(t) ŷ(t). (4.3) The ŷ(t) s are plotted in Figure 8 and the residuals (4.3) are plotted in Figure 9. The predicted values match quite well with the true values. We want to test whether the residuals

are independent or not. We use the run test (Draper and Smith; 98) and z = -.89 confirms that the residuals are dependent. The autocorrelation function and the partial autocorrelation function suggest that the residuals should be an AR(3) process and the parameter can be estimated as X(t) =.0904X(t ) 0.5067X(t ) + 0.065X(t 3) + e(t). (4.4) Performing the run test on ê(t), we obtain z = -.3973. So it does not reject the independent assumptions on e(t) s. Since all the roots of the polynomial equation; z 3.0904z + 0.5067z 0.065 = 0 are less than one in absolute value, therefore, X(t) can be modeled as a stationary AR(3) process, which satisfies assumption and clearly it does not satisfy the error assumption of Sircar and Syali (996). From this data analysis it is clear that the AM model can be used quite effectively for modeling sustained vowel sound uuu with the proper error assumptions. It may be mentioned that without the proper error assumptions the confidence intervals of the unknown parameters will not be correct. 5. COCLUSIOS In this paper we consider the AM signal model originally proposed by Sircar and Syali (996) to analyze certain non-stationary speech data. We assume the errors are from a stationary distribution. It is observed that the usual LSEs and the ALSEs work quite well even when the errors are correlated. The estimated signal matches quite well with the original one. We have the asymptotic distribution of the different estimators and it was used to construct the asymptotic confidence intervals of the different unknown parameters. ote that we have used the Periodogram function to estimate M but no formal result is obtained.

It seems some of the model selection technique like information theoretic criteria or cross validation approach can be used to estimate M. Further work is needed in that direction. ACKOWLEDGMETS The authors would like to thank Professor G.C. Ray of Department of Electrical Engineering, I.I.T. Kanpur for providing the speech data. The authors would also like to thank two referees for some very constructive suggestions and the editor Professor Dr. Olaf Bunke for encouragements. REFERECES Bai, Z.D., Chen X.R., Krishnaiah, P.R., Wu, Y.H. and Zhao, L.C. (99), Strong consistency of the maximum likelihood parameter estimation of the superimposed exponential signals in noise, Theory of Probability and Applications, Vol. 36, o., -7. Brillinger, D. (98), Time Series and data Analysis (Expanded Edn.) San Francisco: Holden-Day. Draper,.R. and Smith, H. (98), Applied Regression Analysis, John Wiley and Sons, ew York. Fuller, W.A. (976), Introduction to Statistical Time Series, John Wiley and Sons, ew York. Grenier, Y. (983), Time-dependent ARMA modeling of non-stationary signals, IEEE Trans. Acoust. Speech and Signal Processing, ASSP-3, o. 4, 899-9. 3

Hannan, E.J. (970), Multiple Time Series, ew York, Wiley. Hannan, E.J. (97), onlinear time series regression, Journal of Applied Probability, Vol. 8, 767-780. Isaksson, A., Wennberg, A. and Zetterberg, L.H. (98), Computer analysis of EEG signals with parametric models, Proc. IEEE, Vol. 69, o. 4, 45-46. Jennrich, R.I. (969), Asymptotic properties of the non-linear least squares estimators, Annals of Mathematical Statistics, Vol. 40, 633-643. Kundu, D. (99), Asymptotic properties of the complex valued non-linear regression model, Communications in Statistics, Ser. A., Vol. 0, o., 3793-3803. Kundu, D. (997), Asymptotic theory of the least squares estimators of sinusoidal signals, Statistics, Vol. 30, -38. Kundu, D. and Mitra, A. (999), On asymptotic behavior of least squares estimators and the confidence intervals of the superimposed exponential signals, Signal Processing, Vol. 7, o. 3, 9-39. McAulay, R.J. and Quatieri, T.F. (986), Speech analysis/ synthesis based on sinusoidal representation, IEEE Trans. Acoust. Speech Processing, ASSP-34, o. 4, 744-754. Press, W.H., Teukolsky, S.A., Vellerling, W.T. and Flannery, B.P. (993), umerical recipes in FORTRA, The Art of Scientific Computing, (nd ed.), Cambridge University Press, Cambridge. 4

Quinn, B.G. and Thomson, P.J. (99), Estimating the frequency of a periodic function, Biometrika, Vol. 78, o., 65-74. Rao, C.R. and Zhao, L.C. (993), Asymptotic behavior of the maximum likelihood estimates of superimposed exponential signals, IEEE Trans. Signal Processing, Vol. 4, 46-463. Sircar, P. and Syali, M.S. (996), Complex AM signal model for non-stationary signals, Signal Processing, Vol. 53, 35-45. Tong, H. (990), on-linear Time Series: A Dynamical System Approach, Clarendon Press, Oxford, 990. Walker, A.M. (97), On the estimation of Harmonic components in a time series with stationary independent residuals, Biometrika, Vol. 58, -6. Wu, C.F.J. (98), Asymptotic theory of non-linear least squares estimators, Annals of Statistics, Vol. 9, 50-53. 5

Appendix In the Appendix we denote θ 0 = (A 0 R, A 0 I, µ 0 R, µ 0 I, ν 0, ω 0 ) as the true parameter value of θ = (A R, A I, µ R, µ I, ν, ω). To prove the different results we need the following lemmas. Lemma : Let U(t) be a real valued stationary sequence such that U(t) = α(k)ɛ(t k), (A.) k= where ɛ(t) s are i.i.d. real valued random variables with mean zero and finite variance σ and k= α(k) <, then lim sup U(t) cos(θt) θ = 0 lim sup U(t) sin(θt) θ = 0 a.s., a.s. Proof of Lemma : See Kundu (997). The lemma also follows from Theorem 4.5. in Brillinger (98; page 98). ote that using Lemma, the following results can be obtained along the same line. Lemma : If X(t) satisfies assumption, then lim sup U(t)t L cos(θt) θ L+ = 0 lim sup U(t)t L sin(θt) θ L+ = 0 for L =,,.... a.s., a.s., Lemma 3: Define S c = {θ : θ θ 0 > c}, then ˆθ, the LSE of θ 0, obtained by minimizing (.3) (when M = ), is a strongly consistent estimator of θ 0 provided { lim inf Q(θ) Q(θ 0 ) } > 0 a.s. θ S c 6

for all c > 0. Proof of Lemma 3: The proof is quite simple and can be obtained along the same line as Wu (98). Proof of Theorem : Let us write S c = A Rc A Ic M Rc M Ic c W c, where ow observe that { Q(θ) Q(θ 0 ) } = A Rc = { θ : A R A 0 R > c }, A Ic = { θ : A I A 0 I > c }, M Rc = { θ : µ R µ 0 R > c }, M Ic = { θ : µ R µ 0 I > c }, c = { θ : ν ν 0 > c }, W c = { θ : ω ω 0 > c }. = { y(t) A( + µe iνt )e iωt X(t) } A 0 ( + µ 0 e iν0t )e iω0t A( + µe iνt )e iωt + { Re X(t) ( Ā 0 ( + µ 0 e iν0t )e iω0t Ā( + µe iνt )e iωt)} = f (θ) + g (θ). Let us write X(t) = X R (t) + ix I (t), where X R (t) = X I (t) = k= k= {a R (k)e R (t k) a I (k)e I (t k)}, {a R (k)e I (t k) + a I (k)e R (t k)}. So both X R (t) and X I (t) are in the form U (t) + U (t) where U k (t), k =, are real-valued stationary sequence satisfying equation (A.) stated in Lemma. ow using Lemma, we have, and for any c > 0, lim inf f (θ) = lim inf θ A Rc lim inf θ g (θ) = 0 a.s., (A.) θ A Rc A 0 ( + µ 0 e iν0t )e iω0t A( + µe iνt )e iωt 7

= lim inf A 0 A ( + µ 0 e iν0t )e iω0t A R A 0 R >c c lim + µ 0 e iν0t c ( + µ 0 ) > 0 a.s. Similarly it can be proved for A Ic, M Rc, M Ic, c and W c which implies that lim inf θ S c f (θ) > 0 a.s. (A.3) So using (A.), (A.3) and Lemma 3, the theorem follows. To prove Theorem, we need the following lemmas. Lemma 4: If η = ( ν, ω) is the ALSE of η 0 = (ν 0, ω 0 ) obtained by maximizing (.4) (for M = ) with respect to ν and ω then ( ν, ω) is a strongly consistent estimator of (ν 0, ω 0 ), provided for all δ > 0. lim sup η 0 η >δ { I(ν, ω) I(ν 0, ω 0 ) } < 0 a.s. Proof of Lemma 4: The proof is quite simple and can be obtained along the same line as Lemma 3. Lemma 5: Under assumptions and, η = ( ν, ω) is a strongly consistent estimator of η 0. Proof of Lemma 5: Define S δ = {η : η η 0 > δ} = S ν δ S ω δ, where S ν δ = { η : ν ν 0 > δ } and S ω δ = { η : ω ω 0 > δ }. ote that because of Lemma, expanding I(η), we have lim sup S ν δ I(η) = lim sup S ν δ lim sup S ν δ y(t)e iωt + y(t)e i(ω+ν)t A 0 e i(ω ω0 )t + A 0 µ 0 e i(ω ν0 ω 0 )t 8

+ A 0 e i(ω+ν ω0 )t + A 0 µ 0 e i(ω+ν ω0 ν 0 )t = lim sup ν ν 0 >δ A 0 + A 0 µ 0 e iν0 t + A 0 e iνt + A 0 µ 0 e i(ν ν0 )t = A 0. Similarly using Lemma and expanding I(η 0 ), we have lim sup y(t)e iω0 t + y(t)e i(ω0 +ν 0 )t = A 0 + A 0 µ 0 > 0. Sδ ν Therefore, lim sup S ν δ Similarly it can be shown that lim sup S ω δ (A.4) and (A.5) imply that { I(η) I(η 0 ) } = A 0 µ 0 < 0 a.s. (A.4) { I(η) I(η 0 ) } = ( + µ 0 ) A 0 < 0 a.s. (A.5) { lim sup I(η) I(η 0 ) } = ( + µ 0 ) A 0 < 0 a.s. S δ and so using Lemma 4, the result follows. Lemma 6: If η = ( ν, ω) is the ALSE of η 0 = (ν 0, ω 0 ) of the model (.) (for M = ), then under assumptions and, ( ν ν 0 ) 0 a.s. ( ω ω 0 ) 0 a.s. Proof of Lemma 6: Expanding I ( ν, ω) = I ( η) around η 0, using multivariate Taylor series expansion up to first order term I ( η) I (η 0 ) = ( η η 0 )I ( η), (A.6) 9

where η is a point between η and η 0. I (η) and I (η) are the vector of first derivatives and the matrix of second derivatives of I(η) w.r.t. η respectively. ote that I ( η) = 0, so from (A.6), we have ( η η 0 ) = I (η 0 ) [I ( η)] ( η η 0 ) = [ ] [ ] I (η 0 ) 3 I ( η). Using Lemma and Lemma, it can be shown that I (η 0 ) Γ, where 3 A 0 µ 0 A 0 µ 0 Γ =, A 0 µ 0 A 0 + A 0 µ 0 which is an invertible matrix because of the assumptions. Elements of I (η) are continuous functions of ν and ω and η is a point between η and η 0. So using the fact that η η 0 a.s., we have lim 3 I ( η) = lim 3 I (η 0 ) = Γ. Also using Lemma, it can be shown that I (η 0 ) 0 a.s. Hence ( η η 0 ) 0 a.s. which implies that ( ν ν 0 ) 0 a.s. and ( ω ω 0 ) 0 a.s. Lemma 7: à and µ, as given in (.5) ( for M = ) are strongly consistent estimators of A 0 and µ 0. Proof of Lemma 7: Let us denote y R (t), y I (t) as the real and imaginary parts of y(t). Therefore, à = [ ] {y R (t) cos( ωt) + y I (t) sin( ωt)} + i [ ] {y I (t) cos( ωt) y R (t) sin( ωt)}. Expanding cos( ωt), sin( ωt) by Taylor series around ω 0 and using Lemmas, and 6, we get à A R 0 + ia I 0 = A 0 a.s. and à µ A 0 µ 0 a.s. 0

which proves the lemma. Proof of Theorem : Combining Lemmas 5 and 7, the result follows immediately. Proof of Theorem 3: Let us denote Q (θ) = [ Q(θ) A R, Q(θ), Q(θ), Q(θ) A I µ R, Q(θ) µ I ν, Q(θ) ] ω and Q (θ) denotes the corresponding 6 6 double derivative matrix of Q(θ). ow expanding Q (ˆθ) around θ 0 by multivariate Taylor series up to the first order term, we get Q (ˆθ) Q (θ 0 ) = (ˆθ θ 0 )Q ( θ), (A.7) where θ is a point between ˆθ and θ 0. Since Q (ˆθ) = 0, (A.7) implies (ˆθ θ 0 ) = Q (θ 0 )[Q ( θ)]. The main idea to prove that (ˆθ θ 0 ) converges to a normal distribution is as follows. Consider the following 6 6 diagonal matrix D; D = diag {,,,, 3, 3 }. Therefore (ˆθ θ 0 )D = Q (θ 0 )D[DQ ( θ)d]. It can be shown by the straight forward but lengthy calculations that lim [DQ ( θ)d] = lim [DQ (θ 0 )D] = Σ, (A.8) where Σ is same as defined in the statement of Theorem 3. Using the central limit theorem of stochastic processes (Fuller; 976, page 5), it can be shown that Q (θ 0 )D 6 [ 0, 4σ (c Σ + c Σ ) ], (A.9)

where c, c, Σ and Σ are same as defined in the statement of Theorem 3. ow combining (A.8) and (A.9), the result follows immediately. Proof of Theorem 4: It can be shown similarly as Hannan (97) or Walker (97) that  R ÃR = O p ( ),  I ÃI = O p ( ), ˆµ R µ R = O p ( ), ˆµ I µ I = O p ( ), ˆω ω = O p ( ), ˆν ν = O p ( ). (A.0) Here the terms O p ( ) and O p ( ) indicate that they converge to zero in probability and also O p ( ) and O p ( ) are both bounded in probability as. Therefore, using Theorem 3 and (A.0) the result follows.

Table : The average LSEs, biases, variances, confidence lengths and coverage probabilities of different parameters. Parameter Average LSE Variance Average Conf. Length Cov. Prob (Bias) (Asymp. Var.) (Expected Conf. Length) (ominal Level) A R 4.99960 5.096e-3 0.9360 0.94 (-0.00040) (7.9057e-3) (0.34854) (0.95) A I.00.4e- 0.54750 0.95 (0.00) (.7534e-) (0.65046) (0.95) µ R 0.50004 6.0334e-4 0.89 0.96 (0.00004) (.53e-3) (0.5339) (0.95) µ I 0.99958 4.937e-4 8.7987e- 0.94 (-0.0004) (7.058e-4) (0.053) (0.95) ω.043 6.7e-07 3.775e-3 0.95 (0.00000) (.3085E-06) (4.484e-3) (0.95) ν 0.5085.074e-06 4.3085e-3 0.95 (-0.0000) (.794e-06) (5.55e-3) (0.95) Table : The average ALSEs, biases, variances, confidence lengths and coverage probabilities of different parameters. Parameter Average LSE Variance Average Conf. Length Cov. Prob (Bias) (Asymp. Var.) (Expected Conf. Length) (ominal Level) A R 4.99958 5.864e-3 0.33484 0.96 (-0.0004) (7.9057e-3) (0.34854) (0.95) A I 0.9774.4587e- 0.6577 0.94 (-0.086) (.7534e-) (0.65046) (0.95) µ R 0.54447.956e-3 0.5338 0.80 (0.04447) (.53e-3) (0.5339) (0.95) µ I 0.97679 6.7943e-4 0.073 0.88 (-0.03) (7.058e-4) (0.053) (0.95) ω.045.478e-6 5.387e-3 0.75 (0.0000) (.3085e-06) (4.484e-3) (0.95) ν 0.50463.403e-6 4.354e-3 0.95 (0.0077) (.794e-6) (5.55e-3) (0.95) 3

Table 3: The least squares estimates and the confidence lengths of the different parameters of the sustained vowel sound uuu. Parameter Estimate Lower Bound Upper Bound A R -5.0508-553.7765-488.35 A I 378.3746 338.8855 47.8636 µ R -.6067 -.7944 -.498 µ I.09645.9369.5660 ω 0.35 0.38 0.376 ν 0.457 0.49 0.485 A R -334.454-369.446-99.463 A I 90.066 5.9665 38.586 µ R 0.75906 0.709 0.8090 µ I -0.876-0.36335-0.008 ω 0.34336 0.3430 0.3437 ν 0.46 0.46 0.498 4