Quadratic Extended Filtering in Nonlinear Systems with Uncertain Observations

Similar documents
Optimal Fusion Estimation with Multi-Step Random Delays and Losses in Transmission

Fixed-interval smoothing algorithm based on covariances with correlation in the uncertainty

RECURSIVE ESTIMATION AND KALMAN FILTERING

4 Derivations of the Discrete-Time Kalman Filter

A Class of Multi-Scales Nonlinear Difference Equations

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53

ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter

A NOVEL OPTIMAL PROBABILITY DENSITY FUNCTION TRACKING FILTER DESIGN 1

City, University of London Institutional Repository

Research Article Weighted Measurement Fusion White Noise Deconvolution Filter with Correlated Noise for Multisensor Stochastic Systems

Statistics 910, #15 1. Kalman Filter

Parameter Estimation for ARCH(1) Models Based on Kalman Filter

New design of estimators using covariance information with uncertain observations in linear discrete-time systems

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49

Linear estimation from uncertain observations with white plus coloured noises using covariance information

Introduction p. 1 Fundamental Problems p. 2 Core of Fundamental Theory and General Mathematical Ideas p. 3 Classical Statistical Decision p.

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance

STOCHASTIC STABILITY OF EXTENDED FILTERING FOR NONLINEAR SYSTEMS WITH MEASUREMENT PACKET LOSSES

Cramér-Rao Bounds for Estimation of Linear System Noise Covariances

ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER. The Kalman Filter. We will be concerned with state space systems of the form

ASIGNIFICANT research effort has been devoted to the. Optimal State Estimation for Stochastic Systems: An Information Theoretic Approach

A Study for the Moment of Wishart Distribution

Research Article Convex Polyhedron Method to Stability of Continuous Systems with Two Additive Time-Varying Delay Components

On Numerical Solutions of Systems of. Ordinary Differential Equations. by Numerical-Analytical Method

X t = a t + r t, (7.1)

A Family of Optimal Multipoint Root-Finding Methods Based on the Interpolating Polynomials

Volume 30, Issue 3. A note on Kalman filter approach to solution of rational expectations models

Short-Term Electricity Price Forecasting

A Generalization of p-rings

Recursive Generalized Eigendecomposition for Independent Component Analysis

A Theoretical Overview on Kalman Filtering

Empirical Power of Four Statistical Tests in One Way Layout

Nonlinear Parameter Estimation for State-Space ARCH Models with Missing Observations

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q

9 Multi-Model State Estimation

Design of FIR Smoother Using Covariance Information for Estimating Signal at Start Time in Linear Continuous Systems

Lessons in Estimation Theory for Signal Processing, Communications, and Control

ARIMA Modelling and Forecasting

Estimation for Nonlinear Dynamical Systems over Packet-Dropping Networks

The Greatest Common Divisor of k Positive Integers

UWB Geolocation Techniques for IEEE a Personal Area Networks

A Note of the Strong Convergence of the Mann Iteration for Demicontractive Mappings

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE

Nonlinear State Estimation! Extended Kalman Filters!

A Matrix Theoretic Derivation of the Kalman Filter

F denotes cumulative density. denotes probability density function; (.)

2 Introduction of Discrete-Time Systems

Ch. 12 Linear Bayesian Estimators

Independent Component Analysis. Contents

Kalman Filtering. Namrata Vaswani. March 29, Kalman Filter as a causal MMSE estimator

A new unscented Kalman filter with higher order moment-matching

Fractal functional filtering ad regularization

Linear Optimal State Estimation in Systems with Independent Mode Transitions

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

Basins of Attraction for Optimal Third Order Methods for Multiple Roots

On Monitoring Shift in the Mean Processes with. Vector Autoregressive Residual Control Charts of. Individual Observation

FUNDAMENTAL FILTERING LIMITATIONS IN LINEAR NON-GAUSSIAN SYSTEMS

Prediction of ESTSP Competition Time Series by Unscented Kalman Filter and RTS Smoother

Recursive Least Squares for an Entropy Regularized MSE Cost Function

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

FIR Filters for Stationary State Space Signal Models

Using the Kalman Filter to Estimate the State of a Maneuvering Aircraft

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares

Simultaneous state and input estimation with partial information on the inputs

Explicit Expressions for Free Components of. Sums of the Same Powers

Expressions for the covariance matrix of covariance data

Time Series Analysis

Notes on Time Series Modeling

Finite-Horizon Optimal State-Feedback Control of Nonlinear Stochastic Systems Based on a Minimum Principle

Parameterized Joint Densities with Gaussian Mixture Marginals and their Potential Use in Nonlinear Robust Estimation

Method of Generation of Chaos Map in the Centre Manifold

Performance assessment of MIMO systems under partial information

Generalized Boolean and Boolean-Like Rings

Stochastic Incremental Approach for Modelling the Claims Reserves

2 Statistical Estimation: Basic Concepts

Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity

1 Kalman Filter Introduction

which are not all zero. The proof in the case where some vector other than combination of the other vectors in S is similar.

ACG M and ACG H Functions

Gaussian, Markov and stationary processes

Filtering for Linear Systems with Error Variance Constraints

Development of a Family of Optimal Quartic-Order Methods for Multiple Roots and Their Dynamics

DESIGNING A KALMAN FILTER WHEN NO NOISE COVARIANCE INFORMATION IS AVAILABLE. Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof

A Study on Linear and Nonlinear Stiff Problems. Using Single-Term Haar Wavelet Series Technique

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

SYSTEMS with multiplicative state noise, also known in

Research Article A Fixed Point Theorem for Mappings Satisfying a Contractive Condition of Rational Type on a Partially Ordered Metric Space

Extension of the Sparse Grid Quadrature Filter

Kalman-Filter-Based Time-Varying Parameter Estimation via Retrospective Optimization of the Process Noise Covariance

Linear Prediction Theory

The Expected Opportunity Cost and Selecting the Optimal Subset

Dynamical Behavior for Optimal Cubic-Order Multiple Solver

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

A NOTE ON SUMS OF INDEPENDENT RANDOM MATRICES AFTER AHLSWEDE-WINTER

Discrete-Time Finite-Horizon Optimal ALM Problem with Regime-Switching for DB Pension Plan

Linearization of Two Dimensional Complex-Linearizable Systems of Second Order Ordinary Differential Equations

in a Rao-Blackwellised Unscented Kalman Filter

Riccati difference equations to non linear extended Kalman filter constraints

k-weyl Fractional Derivative, Integral and Integral Transform

Transcription:

Applied Mathematical Sciences, Vol. 8, 2014, no. 4, 157-172 HIKARI Ltd, www.m-hiari.com http://dx.doi.org/10.12988/ams.2014.311636 Quadratic Extended Filtering in Nonlinear Systems with Uncertain Observations A. Hermoso-Carazo Dpto. Estadística e I. O., Universidad de Granada Campus Fuentenueva, s/n, 18071 Granada, Spain J. Linares-Pérez Dpto. Estadística e I. O., Universidad de Granada Campus Fuentenueva, s/n, 18071 Granada, Spain S.M. Valenzuela-Ruiz Dpto. Estadística e I.O., Universidad de Granada Alfonso XIII, s/n, 52005 Melilla, Spain Copyright c 2014 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original wor is properly cited. Abstract This paper proposes a second order polynomial estimator in nonlinear systems with uncertain observations. It is assumed that the signal and the observations are measured with an additive noise and that the observation equation includes a scalar multiplicative noise, which is described by a sequence of Bernoulli random variables that model the random interruptions which can be present in the observation mechanism. The least mean-squares error quadratic estimation problem is solved, first, by linearizing the system and, then, by defining an augmented system, in which the signal and the observation vectors are obtained by adding the corresponding second-order powers to the signal and observation vectors of the linearized system. The second order polynomial estimator in the nonlinear system is obtained from the linear filtering estimate in the augmented system.

158 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz Mathematics Subject Classification: 62M20, 93E11 Keywords: Uncertain observations, nonlinear systems, quadratic estimation 1 Introduction This paper studies the problem of least mean-squared error (LMSE) quadratic estimation in nonlinear systems with uncertain observations. In this type of systems, the observation equation includes, besides the additive noise, a multiplicative noise, which models the presence or absence of the signal in the observations, described by a sequence of Bernoulli random variables. Systems with uncertain observations are always non-gaussian even if the noises are Gaussian and hence, the optimal estimator of the signal is not easily obtainable. This inconvenience motivates the search for suboptimal solutions under different assumptions about the signal and the noise processes involved in the model. Signal estimation problems in non-gaussian systems have been widely studied. De Santis et al. [11] studied the estimation problem in non-gaussian linear systems, deducing a recursive algorithm that provides the LMSE estimator among all second-degree polynomial transformations of the observations. Since this class of transformations contains the linear ones, the estimators obtained by using this algorithm represented an improvement, in the sense of the LMSE, on those obtained by the Kalman filter in non-gaussian systems. A generalization of this study was proposed in [1], where a recursive algorithm for an arbitrary degree polynomial estimator is proposed. Recent studies based on non-gaussian systems can be seen in [2] and [12]. Nahi [8] was the first to address the LMSE linear estimation problem from the standpoint of uncertain observations, modelling the presence or absence of the signal in the observations using a stochastic process defined by Bernoulli variables. Considering linear systems with independent white additive noises, and under the assumption of independence of the variables that model the uncertainty, Nahi obtained a linear recursive algorithm with a structure similar to that of the Kalman filter but depending on the probability that the observations may be only noises. Later, [3] and [4] generalized the Nahi filter to the case of uncorrelated additive noises. In nonlinear systems with uncertain observations, the volume of contributions is significantly smaller than in the case of linear systems. Some contributions to this study can be seen in the wor of Nanacara and Yaz [10], Hermoso-Carazo and Linares-Pérez [5] and Naamori et al. [9]. Recently, Wu

Quadratic filtering in nonlinear systems with uncertain observations 159 and Song [13] generalized the original extended Kalman filtering, the unscented Kalman filtering and the Gaussian particle filtering for nonlinear systems with uncertain observations. The aim of this paper is to advance the study of the estimation problem in nonlinear systems with uncertain observations. To do so, we use the methodology of the extended Kalman filter in combination with the methodology used in [11], and propose a recursive algorithm that provides the LMSE quadratic filter. Specifically, we linearize the system equations and then obtain the LMSE quadratic estimator of the signal based on the linearized observations and their second-order Kronecer powers. The paper is organized as follows: the hypotheses on the model and the linearized system are formulated in Section 2. The augmented system and the statistical properties of the augmented processes are analyzed in Section 3.1. The filtering algorithm is established in Section 3.2. Finally, in Section 4, we present a numerical simulation example to show the effectiveness of the quadratic estimator in comparison with the linearized mixture filter proposed in [5]. 2 System model and linearized system Let x be the n-dimensional vector which describes the signal to be estimated. Let us suppose that the signal evolves in time through continuous and differentiable nonlinear functions affected by an additive noise {w ; 0}; therefore, the equation that models its dynamics is the following: x = f 1 (x 1 )+w 1, 1. (1) In this paper, we consider m-dimensional nonlinear observations of the signal, y, that contain information of the same, h (x ), with probability p,or only consist of noise, v, with probability 1 p (false alarm probability). Thus, by considering Bernoulli random variables {γ ; 1} with P (γ =1)=p, the observation equation is given by y = γ h (x )+v, 1, (2) where, for all 1, the function h is continuous and differentiable. Our aim is to obtain the least mean-squared error (LMSE) quadratic estimator of the signal x based on the observations {y 1,...,y }. For this purpose, we consider the second-order powers, y [2] j, of the observations y j, j =1,...,, defined by the Kronecer product [7]. Assuming that E[y [2]T j y [2] j ] <, j = 1,..., the quadratic filter of the signal x is its orthogonal projection on the

160 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz space of n-dimensional linear transformations of y 1,...,y and their secondorder powers y [2] 1,...,y[2]. To address the LMSE quadratic estimation problem the following hypotheses are assumed: H.1 The initial signal, x 0, is a random vector with nown moments up to the fourth one: E[x 0 ] = μ 0, Cov[x 0 ] = P 0, Cov[x 0,x [2] 0 ] = P (3) 0 and Cov[x [2] 0,x[2] (4) 0 ]=P 0. H.2 The process {w ; 0} is a zero-mean white noise and its moments up to the fourth one are Cov[w ]=Q, Cov[w,w [2] ]=Q(3) and Cov[w [2],w[2] ]= Q (4). H.3 The noise {v ; 1} is a zero-mean white sequence with Cov[v ]=R, Cov[v,v [2] ]=R(3) and Cov[v [2],v[2] ]=R(4). H.4 The multiplicative noise {γ ; 1} is a sequence of independent Bernoulli variables with probabilities P (γ =1)=p, 1. H.5 The initial signal, x 0, and the noise processes {w ; 0}, {v ; 1} and {γ ; 1} are mutually independent. Our aim is apply the methodology of the extended Kalman filter [6], to derive a recursive algorithm to obtain a quadratic estimator of the signal x based on the observations {y 1,...,y }; that is, we obtain a quadratic filter, ˆx Q /. Accordingly, we linearize the system equations and then obtain the LMSE estimator of the signal based on these linearized observations and their secondorder Kronecer powers. At each sampling time 1, the linearization of the signal equation is performed by applying the Taylor expansion to f 1 in the filter, ˆx Q 1/ 1, and the linearized observation equation is obtained by applying the Taylor expansion to h in the predictor, ˆx Q / 1. Thus, by considering the first order terms of the corresponding expansions, the system equations are approximated by where H = h (x) x F 1 = f 1(x) x x=ˆx Q / 1 x = F 1 x 1 + u 1 + w 1, 1, y = γ (H x + z )+v, 1, x=ˆx Q 1/ 1, u 1 = f 1 (ˆx Q 1/ 1 ) F 1ˆx Q 1/ 1, and z =h (ˆx Q / 1 ) H ˆx Q / 1. (3)

Quadratic filtering in nonlinear systems with uncertain observations 161 For simplicity, we consider the centred signal and observations vectors, x = x E[x ] and ȳ = y E[y ]; thus, we obtain x = F 1 x 1 + w 1, 1, ȳ = γ H x + v, 1, (4) where v =(γ p )(H E[x ]+z )+v. Taing into account the equations of the system (3), the means E[x ] and E[y ], can by obtained by the following expressions: E[x ]=F 1 E[x 1 ]+u 1, 1, E[y ]=p (H E[x ]+z ), 1. The statistical properties of the additive observation noise { v ; 1} are established in the following lemma: Lemma 2.1 The noise process { v ; 1} is a zero-mean white sequence with moments up to the fourth given by R = Cov[ v ]=p (1 p )c c T + R, R (3) = Cov[ v, v [2] ]=p (1 p )(1 2p )c c [2]T + R (3), R (4) = Cov[ v [2], v[2] ]=[ p (1 p )(3p 2 3p +1) p 2 (1 p ) 2] c [2] c[2]t +R (4) +p (1 p )(I m 2 +K m 2) [ (c c T ) R ] (Im 2 + K m 2), where c = H E[x ]+z, K m 2 and I m 2 are the m 2 m 2 commutation and identity matrices, respectively, and is the Kronecer product. Since the linear space generated by {y 1,...,y } is equal to that generated by {ȳ 1,...,ȳ }, the quadratic estimator of x will be obtained from the centred observations, {ȳ 1,...,ȳ }, and the second-order power, {ȳ [2] 1,...,ȳ [2] }. Hence, to address the quadratic estimation problem we consider the system (4). 3 LMSE quadratic estimation problem The quadratic estimation problem can be reformulated as one of linear estimation by increasing the observation vectors with their second-order Kronecer powers. For this purpose, we consider the following augmented signal and observation vectors: ( ) ( ) x ȳ X = x [2] R n+n2, Y = ȳ [2] R m+m2.

162 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz Then, the vector constituted of the first n entries of the LMSE linear estimator of the augmented signal, X, based on the augmented observations, {Y 1,...,Y }, provides the LMSE quadratic estimator of the centred original signal, x, based on {ȳ 1,...,ȳ } and the required estimator will be obtained by adding the mean of the original vector, E[x ]. 3.1 Augmented system In this section, we analyze the dynamics of the augmented signal and observation vectors. Using the Kronecer product properties and the system hypothesis the following equations are obtained x [2] = F [2] 1 +Φ 1, 1, ȳ [2] = γ H [2] x[2] +Ψ, 1, with Φ =(I n 2 + K n 2)[(F x ) w ]+w [2] and Ψ =(I m 2 + K m 2)[(H x ) (γ v )] + v [2]. Then, the vectors X = X E[X ] and Ȳ = Y E[Y ] satisfy the following equations: X = F 1 X 1 + W 1, 1, (5) Ȳ = γ H X + V, 1, ( ) ( ) ( F 0 H 0 where F = 0 F [2], H = 0 H [2], W = ( ) v V = +(γ Ψ E[Ψ ] p ) H E[X ]. w Φ E[Φ ] ), and In the following propositions, we study the properties of the noise processes involved in the system (5), which are required in order to tacle the linear estimation problem. Using the system hypotheses and the Kronecer product properties, the proofs can be deduced without difficulty and, therefore, are omitted here. Proposition 3.1 The noise { W ; 0} is a zero-mean white process with ( ) Q = E[ W T Q W ]= Q (3), where Q 22 =(I n 2 + K n 2) [( F P F T Q (3)T Q 22 ) Q ] (In 2 + K n 2)+Q (4), and where P = E[ x x T ] is obtained recursively by the relation P = F 1 P 1 F 1 T + Q 1, 1; P 0 = Cov[ x 0 ].

Quadratic filtering in nonlinear systems with uncertain observations 163 Proposition 3.2 The process { V ; 1} has zero-mean and R R s = E[ V δ R12 s T V δ s s ]=, R 12T δ R22 s s where R 12 = R (3) + p (1 p )c vec T (P, )H [2]T R 22 s =(I m 2 + K m 2)[(H P,s H T s ) [α sc c T s + p R δ s ]](I m 2 + K m 2)+ + p (1 p )(1 2p )[c [2] vect (P, )H [2]T + H [2] + p (1 p )H [2] vec(p,)vec T (P, )H [2]T δ s, vec(p,)c [2]T ]δ s (4) R δ s and where δ s denotes the Kronecer Delta function and vec( ) denotes the operator that vectorizes a matrix [7]. The matrices P,s = E [ x ] x T s verify P,s = F 1 P 1,s 1 Fs 1 T + Q 1 δ s, and { p p α s = s (1 p )(1 p s ), s, p (1 p ) 2, s =. Proposition 3.3 a) The noises { W ; 0} and { V ; 1} and the initial signal X0 are uncorrelated. b) The noise process {γ ; 1} is independent of the process { W ; 0}, with the initial condition X 0 and ( V 1,..., V 1 ) for any 1. By using the properties established in Propositions 3.1 3.3, in the next Section, we obtain a recursive algorithm for the linear estimators ˆ X/ of the augmented signal X based on {Ȳ1,...,Ȳ}. 3.2 Linear filtering algorithm for the augmented system The properties of the processes involved in the system (5), studied in the previous section, do not permit us to apply the Nahi filter [8], because the additive noise of the observation equation, { V ; 1}, is not white. Therefore, in this section, we derive a linear estimation algorithm for the system (5), which extends Nahi s filter. By the orthogonal projection lemma, the LMSE linear filter ˆ X/ of X satisfies the Wiener-Hopf equation E[ ˆ X/ Ȳi T ]=E[ X Ȳi T ], i =1,...,. (6)

164 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz To derive the algorithm we proceed according to the following stages: Filter in terms of the predictor: we start from the filter expression ˆ X/ as a linear combination of the observations ˆ X / = i=1 K,i Ȳ i. Then, by using the Wiener Hopf equation (6), we have [ E ( ˆ X/ K ], Ȳ )Ȳ s T =E [ ( X K ], Ȳ )Ȳ s T, s =1,..., 1. Denoting the LMSE linear predictor of the signal and the observation by ˆ X / 1 and ˆȲ / 1, respectively, we have E [ ( X K, Ȳ )Ȳ T s ] =E [( ˆ X/ 1 K ], ˆȲ / 1 )Ȳ s T, s =1,..., 1. The following expression for the filter, ˆ X/, in terms of the predictor, ˆ X / 1, is obtained immediately ˆ X / = ˆ X/ 1 + K ν, 1, (7) where ν = Ȳ ˆȲ / 1 is the innovation and K = K, denotes the gain matrix of the filter. Next, denoting the prediction and filtering error covariance matrices by P X /j, j = 1,, from (7) and taing into account that the error X ˆ X/ 1 is uncorrelated with ˆȲ / 1, we obtain P X / = P X / 1 K Π KT, 1, (8) where Π = E [ ] ν ν T is the innovation covariance matrix. Predictor in terms of the filter: considering the Wiener-Hopf equations (6) corresponding to the predictor, ˆ X/ 1, and the filter, ˆ X 1/ 1, E[ X Ȳi T ]=E[ ˆ X/ 1 Ȳi T ], i =1,..., 1, E[ X 1 Ȳi T ]=E[ ˆ X 1/ 1 Ȳi T ], i =1,..., 1, and taing into account that W is centred and uncorrelated with Ȳi, i<,it is proved that ˆ X / 1 = F 1 ˆ X 1/ 1, 1. (9)

Quadratic filtering in nonlinear systems with uncertain observations 165 On the other hand, from Propositions 3.1 and 3.3, the prediction error covariance matrix P X / 1 is recursively calculated from P X / 1 = F 1 P X 1/ 1 F T 1 + Q 1, 1. Innovation process: to determine the innovation, ν = Ȳ ˆȲ / 1,itis sufficient to obtain an expression for the predictor, ˆȲ / 1, which verifies E[ȲȲ i T ]=E[ ˆȲ / 1 Ȳi T ], i =1,..., 1. From Proposition 3.3 it follows that γ and ( X, Ȳ1,...,Ȳ 1) are independent and hence ˆȲ / 1 = p H ˆ X/ 1 + ˆ V / 1, 1, where ˆ V / 1 is the one-stage predictor of V. Therefore, the innovation, ν, is given by ν =(γ p ) H X + p H X/ 1 + V ˆ V / 1, 1. (10) Clearly, the innovation vectors have a zero mean and their covariance matrices, Π = E [ ] ν ν T, are given by Π =p (1 p ) H P X, HT +p 2 H P X / 1 HT +P V,/ 1 +p H P V X +p P H,/ 1 T, 1, X V,/ 1 where P V,/ 1 is the error covariance matrix of ˆ V / 1, P,/ 1 is the error cross-covariance matrix of ˆ X/ 1 and ˆ V / 1, and P X, = E [ ] X XT, which satisfies P X, = F 1 P X 1, 1 F T 1 + Q 1, 1; P X 0,0 = P X 0/0. Gain matrix of the filter: denoting the filtering and predictor errors by X / and X/ 1, respectively, and taing into account the expression (7), we have X / = X/ 1 K ν, 1, X V and, since ν is uncorrelated with X/, clearly the gain matrix is given by [ ] K = p P X X V / 1 HT +P Π 1, 1. (11),/ 1

166 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz One-stage predictor of the observation noise: by a similar reasoning to that used to obtain expression (7), it is deduced that ˆ V /l = ˆ V /l 1 + Ḡ,l Ȳ l/l 1, l <, > 1, (12) where ˆ V /0 = E[ V ]=0, 1. Furthermore, by reasoning similar to that used to obtain (11), the gain matrix Ḡ,l is [ V X Ḡ,l = p l P H ],l/l 1 l T + P V Π 1,l/l 1 l. The error covariance matrices of the successive prediction stages of the observation noise, P V,l/l 1, are obtained recursively from the following relations: P V,l/j =P V,l/j 1 Ḡ,j Π j Ḡ T l,j,,l > 1, j < min(, l), P V,l/0 = R,l,,l 1, which are deduced from the recursive expression (12) using arguments similar to those used to obtain P X / 1. Finally, we obtain the expression of the cross-covariance matrices P taing into account the expressions of ˆ V /l 1 and ˆ Xl/l 1, and so [ V X V X P,l/l 1 = P Π ],l 1/l 2 Ḡ,l 1 l 1 KT l 1 F T l 1, l. 4 Numerical simulation example V X,l/l 1 This section presents a numerical simulation example to illustrate the results obtained in this paper. To do so, we considered the following nonlinear system with uncertain observations: 1 x = x 2 1 +3+ w 1, 1 y = γ (x 2 + exp(x )) + v, 1, where the following hypotheses are assumed: i) The initial signal, x 0, is a zero-mean random variable with variance P 0 = 1, and moments up to the fourth one denoted by P (3) 0 = 0 and P (4) 0 = 3 P0 2. ii) {w ; 0} is a centred white noise with variance Q = 19 3, Q(3) = 128 3 and Q (4) = 1123 Q 2 3.

Quadratic filtering in nonlinear systems with uncertain observations 167 iii) {v ; 1} is a centred white noise with variance R = 19, and third and 3 fourth order moments denoted by R (3) = 128 and R(4) = 1123 R 2 3 3, respectively. iv) {γ ; 1} is a sequence of independent Bernoulli random variables with P (γ =1)=p. v) The initial signal, x 0, and the noise processes {w ; 0}, {v ; 1} and {γ ; 1} are mutually independent. In this example, the performance of the extended quadratic filter algorithm is shown, together with the improvement obtained with respect to the linearized mixture filter proposed by Hermoso and Linares, [5]. Both algorithms were implemented in Matlab, being applied to the estimation of the signal in the previous nonlinear model. For each algorithm, 1000 simulations were performed, with 50 iterations for each simulation. The behaviour of the estimators was compared by the mean square errors (MSE) corresponding to the different simulations of the signal and its estimates: MSE = 1 1000 (x s ˆx s 1000 /) 2, =1,...,50, s=1 where x s denotes the value of the signal generated in simulation s and iteration, and ˆx s / is the filter calculated with the corresponding algorithm in these simulations and iterations.

168 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz 8 7 MSE of quadratic filter, p=0.5 MSE of linearized mixture filter, p=0.5 6 5 4 3 2 1 0 5 10 15 20 25 30 35 40 45 50 Iteration Figure 1: Mean-squared errors of the linearized mixture filter and the extended quadratic filter.

Quadratic filtering in nonlinear systems with uncertain observations 169 The performance of the extended quadratic filter was compared with that of the linearized mixture filter using different false alarm probabilities, and assuming in each case that this probability is constant in all iterations. In all the cases analyzed, the performance of the extended quadratic filter proposed in this paper was clearly superior to that of the linearized mixture of [5]. By way of example, Figure 1 illustrates the results obtained considering P (γ =1)=p = 0.5. This figure shows the evolution of the MSE for both filters, reflecting, as well as the better performance of the extended quadratic filter, manifested by significantly smaller errors than those obtained by the linearized mixture, a lesser dispersion of these errors throughout the different iterations, thus confirming the effectiveness of the proposed algorithm. We also analyzed the behaviour of the estimates obtained by the extended quadratic filter versus the false alarm probability in the observations. These results are illustrated in Figures 2 and 3. 3.2 3 2.8 MSE of quadratic filter, p=0.1 MSE of quadratic filter, p=0.2 MSE of quadratic filter, p=0.3 MSE of quadratic filter, p=0.4 MSE of quadratic filter, p=0.5 2.6 2.4 2.2 2 1.8 0 5 10 15 20 25 30 35 40 45 50 Iteration Figure 2: Mean-squared errors of the extended quadratic filter for p = 0.1, 0.2, 0.3, 0.4, 0.5

170 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz 2.8 2.6 2.4 MSE of quadratic filter, p=0.5 MSE of quadratic filter, p=0.6 MSE of quadratic filter, p=0.7 MSE of quadratic filter, p=0.8 MSE of quadratic filter, p=0.9 2.2 2 1.8 1.6 1.4 1.2 1 0 5 10 15 20 25 30 35 40 45 50 Iteration Figure 3: Mean-squared errors of the extended quadratic filter for p = 0.5, 0.6, 0.7, 0.8, 0.9. As expected, these Figures show that, with an increasing value of p, and therefore a decreasing probability that the observations used in the estimation will lac information about the signal, the mean square error of the extended quadratic filter is significantly reduced. Finally, on comparing these results with those shown in Figure 1 concerning the application to the linearized mixture filter, we also observe that, even for the worst case in the sense of higher mean square errors, corresponding to p = 0.1, the results obtained with the extended quadratic filter are better than those for the mixture filter, even for a higher false alarm probability. This again demonstrated the greater efficiency of the filter proposed in this paper,with respect to the proposal set out in [5]. 5 Conclusion The aim of this wor was to advance in the study of the estimation problem in the case of noisy signals modelled by nonlinear dynamics, using the information provided by the observation of its functions, also nonlinear and also affected by noise.

Quadratic filtering in nonlinear systems with uncertain observations 171 The study focused on the problem of estimation from uncertain observations. In this type of observations there exists a positive probability that they will not contain information about the signal and, therefore, will contribute only noise to the estimation. Our study was aimed at obtaining a recursive algorithm that would provide the estimator based on second-order polynomial functions of the observations, defined by the Kronecer products. The procedure used for obtaining this estimator was to increase the original signal and observation vectors with their second-order powers, thus obtaining a new system in which we could apply a linear filtering algorithm and thus obtain the quadratic filtering algorithm in the original system. The optimality criterion underlying this entire study was to minimize the mean-squared error. Consequently, the efficiency of the proposed algorithm is compared in terms of the mean-squared errors associated with the estimates. ACKNOWLEDGEMENTS. This research was supported by Ministerio de Ciencia e Innovación (Grant No. MTM2011-24718). References [1] F. Carravetta, A. Germani and M. Raimondi, Polynomial filtering for linear discrete time non-gaussian systems. SIAM J. Control and Optimization, 34 (1996), 1666-1690. [2] O. Grothe, A higher order correlation unscented Kalman filter. Applied Mathematics and Computation 219 (2013), 9033-9042. [3] A. Hermoso-Carazo and J. Linares-Pérez, Linear Estimation for discretetime systems in the presence of time-correlated disturbances and uncertain observations. IEEE Transactions on Automatic Control, (1994) AC-39, 8, 1636-1638. [4] A. Hermoso-Carazo and J. Linares-Pérez, Linear smoothing for discretetime systems in the presence of correlated disturbances and uncertain observations. IEEE Transactions on Automatic Control 40 (1995), 1486-1488. [5] A. Hermoso-Carazo and J. Linares-Pérez, Different approaches for state filtering in a nonlinear system with uncertain observations. Applied Mathematics and Computation 187 (2007), 708-724. [6] D. Simon, Optimal State Estimation, Kalman, H and Nonlinear Approaches. John Wiley and Sons, New Jersey, 2006.

172 A. Hermoso-Carazo, J. Linares-Pérez and S.M. Valenzuela-Ruiz [7] J.R. Magnus, H. Neudecer, Matrix Differential Calculus with Applications in Statistics and Econometrics. Revised Ed. Wiley and Sons, New Yor, 1999. [8] N.E. Nahi, Optimal Recursive Estimation with Uncertain Observation. IEEE Transaction on Information Theory, (1969) Vol. IT-15, 3, 462-468. [9] S. Naamori, R. Caballero-Águila, A. Hermoso-Carazo, J.D. Jiménez- Lopez and J. Linares-Pérez, Signal estimation with nonlinear uncertain observations using covariance information. Journal of Statistical Computation and Simulation 79 (1), 55-66 (2009). [10] W. NaNacara and E.E. Yaz, Recursive estimator for linear and nonlinear systems with uncertain observations. Signal Processing, 62 (1997), 215-228. [11] A. De Santis, A. Germani and M. Raimondi, Optimal quadratic filtering of linear discrete-time non-gaussian systems. IEEE Transactions on Automatic Control, 40 (1995), 1274-1278. [12] X. Wang, Y. Liang, Q. Pan and F. Yang, A Gaussian approximation recursive filter for nonlinear system with uncorrelated noises. Automatica 48 (2012) 2290-2297. [13] X. Wu and Z. Song, Multi-step prediction of chaotic time-series with intermittent failures based on the generalized nonlinear filtering methods. Applied Mathematics and Computation 219 (2013) 8584-8594. Received: November 14, 2013