Sequential State Estimation with Interrupted Observation

Similar documents
Square-Root Algorithms of Recursive Least-Squares Wiener Estimators in Linear Discrete-Time Stochastic Systems

6.4 Kalman Filter Equations

Sequential Estimation in Linear Systems with Multiple Time Delays

Design of FIR Smoother Using Covariance Information for Estimating Signal at Start Time in Linear Continuous Systems

State Estimation of Linear and Nonlinear Dynamic Systems

State Estimation using Moving Horizon Estimation and Particle Filtering

Time Series Prediction by Kalman Smoother with Cross-Validated Noise Density

NONUNIFORM SAMPLING FOR DETECTION OF ABRUPT CHANGES*

Data association uncertainty occurs when remote sensing devices, such as radar,

Motion Model Selection in Tracking Humans

UWB Geolocation Techniques for IEEE a Personal Area Networks

Using the Kalman Filter to Estimate the State of a Maneuvering Aircraft

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance

Kalman Filters with Uncompensated Biases

Lessons in Estimation Theory for Signal Processing, Communications, and Control

DESIGNING A KALMAN FILTER WHEN NO NOISE COVARIANCE INFORMATION IS AVAILABLE. Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof

State Estimation for Nonlinear Systems using Restricted Genetic Optimization

State estimation of linear dynamic system with unknown input and uncertain observation using dynamic programming

Linear Optimal State Estimation in Systems with Independent Mode Transitions

Robust Variance Constrained Filter Design for Systems with Non-Gaussian Noises

An Application of the Information Theory to Estimation Problems

Steady State Kalman Filter for Periodic Models: A New Approach. 1 Steady state Kalman filter for periodic models

CS 532: 3D Computer Vision 6 th Set of Notes

ADAPTIVE FILTE APPLICATIONS TO HEAVE COMPENSATION

Evolution Strategies Based Particle Filters for Fault Detection

Optimal Distributed Lainiotis Filter

Performance Analysis of an Adaptive Algorithm for DOA Estimation

FIR Filters for Stationary State Space Signal Models

Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes

Introduction p. 1 Fundamental Problems p. 2 Core of Fundamental Theory and General Mathematical Ideas p. 3 Classical Statistical Decision p.

Discovery Guide. Beautiful, mysterious woman pursued by gunmen. Sounds like a spy story...

Introduction to Unscented Kalman Filter

Quadratic Extended Filtering in Nonlinear Systems with Uncertain Observations

Detection theory 101 ELEC-E5410 Signal Processing for Communications

Fuzzy Adaptive Kalman Filtering for INS/GPS Data Fusion

On Stability of the Kalman Filter for Discrete Time Output Error Systems

State Observers and the Kalman filter

A Comparitive Study Of Kalman Filter, Extended Kalman Filter And Unscented Kalman Filter For Harmonic Analysis Of The Non-Stationary Signals

4 Derivations of the Discrete-Time Kalman Filter

Riccati difference equations to non linear extended Kalman filter constraints

L06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms

Optimal control and estimation

Detection of signal transitions by order statistics filtering

Particle Filters. Outline

Nonlinear State Estimation! Extended Kalman Filters!

STONY BROOK UNIVERSITY. CEAS Technical Report 829

Cramér-Rao Bounds for Estimation of Linear System Noise Covariances

RAO-BLACKWELLISED PARTICLE FILTERS: EXAMPLES OF APPLICATIONS

9 Multi-Model State Estimation

Benjamin L. Pence 1, Hosam K. Fathy 2, and Jeffrey L. Stein 3

Prediction of ESTSP Competition Time Series by Unscented Kalman Filter and RTS Smoother

Data assimilation with and without a model

PRISON POLICY INITIATIVE ANNUAL REPORT

Optimal Polynomial Control for Discrete-Time Systems

Stochastic Models, Estimation and Control Peter S. Maybeck Volumes 1, 2 & 3 Tables of Contents

Lecture 6: Bayesian Inference in SDE Models

Robotics 2 Target Tracking. Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard

Maximum Likelihood Ensemble Filter Applied to Multisensor Systems

Rao-Blackwellized Particle Filter for Multiple Target Tracking

Particle Filter Track Before Detect Algorithms

Outline. 1 Introduction. 2 Problem and solution. 3 Bayesian tracking model of group debris. 4 Simulation results. 5 Conclusions

Lecture 7: Optimal Smoothing

A recursive algorithm based on the extended Kalman filter for the training of feedforward neural models. Isabelle Rivals and Léon Personnaz

TSRT14: Sensor Fusion Lecture 8

Time Series Analysis

ADAPTIVE PERIODIC NOISE CANCELATION FOR COOLING FANS WITH VARYING ROTATIONAL SPEED

A new trial to estimate the noise propagation characteristics of a traffic noise system

Bayesian Methods in Positioning Applications

Kalman Filter and Parameter Identification. Florian Herzog

The Kalman Filter (part 1) Definition. Rudolf Emil Kalman. Why do we need a filter? Definition. HCI/ComS 575X: Computational Perception.

Risk-Sensitive Filtering and Smoothing via Reference Probability Methods

An Introduction to the Kalman Filter

Prediction, filtering and smoothing using LSCR: State estimation algorithms with guaranteed confidence sets

An application of hidden Markov models to asset allocation problems

Abstract. 1 Introduction. 2 Idempotent filters. Idempotent Nonlinear Filters

POLYNOMIAL FILTERING FOR LINEAR DISCRETE TIME NON-GAUSSIAN SYSTEMS

A generalised minimum variance controller for time-varying systems

Adaptive MV ARMA identification under the presence of noise

Lagrangian data assimilation for point vortex systems

Gaussian Filters for Nonlinear Filtering Problems

An Invariance Property of the Generalized Likelihood Ratio Test

Performance of an Adaptive Algorithm for Sinusoidal Disturbance Rejection in High Noise

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. VIII Kalman Filters - Mohinder Singh Grewal

Constrained State Estimation Using the Unscented Kalman Filter

Detection Theory. Chapter 3. Statistical Decision Theory I. Isael Diaz Oct 26th 2010

Sliding Window Test vs. Single Time Test for Track-to-Track Association

Expectation propagation for signal detection in flat-fading channels

Fixed-interval smoothing algorithm based on covariances with correlation in the uncertainty

Tracking and Identification of Multiple targets

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

MULTICHANNEL FAST QR-DECOMPOSITION RLS ALGORITHMS WITH EXPLICIT WEIGHT EXTRACTION

Procedure for Measuring Swing Bearing Movement on 300 A Series Hydraulic Excavators {7063}

APPLICATION OF KALHAN FILTERING TO TRACK FITTING IN THE DELPHI DETECTOR. R. Fruhwirth

Estimation for Nonlinear Dynamical Systems over Packet-Dropping Networks

Open Access Target Tracking Algorithm Based on Improved Unscented Kalman Filter

Data assimilation with and without a model

MULTI-MODEL FILTERING FOR ORBIT DETERMINATION DURING MANOEUVRE

Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf

2. Linear recursions, different cases We discern amongst 5 different cases with respect to the behaviour of the series C(x) (see (. 3)). Let R be the

A6523 Modeling, Inference, and Mining Jim Cordes, Cornell University

Transcription:

INFORMATION AND CONTROL 21, 56--71 (1972) Sequential State Estimation with Interrupted Observation Y. SAWARAGI, T. KATAYAMA AND S. FUJISHIGE Department of Applied Mathematics and Physics, Faculty of Engineering, Kyoto University, Kyoto, Japan This paper deals with the state estimation problem for linear discrete time systems with the interrupted observation mechanisms which can be characterized by the independent binary sequence taking on the values of 0 or 1. On the basis of the Bayesian approach, the approximate minimum variance estimator is derived for the case when at any time the probability of occurrence of interruption is known a priori. An adaptive estimator algorithm is also established when the probability of interruption is unknown but fixed throughout the time interval of interest. Unlike the usual Kalman filter algorithm, all the estimators derived here are nonlinear with respect to observations and the associated covariance equations are related to the actual observations. Digital simulation studies are demonstrated to compare the performance of the approximate nonlinear estimator presented here with that of the best linear estimator due to Nahi (1969). 1. INTRODUCTION The estimation theory developed to date assumes that at any time the signal or state to be estimated is contained in the observations (Kalman, 1963; Bucy and Joseph, 1968). In many practical situations, however, there is a chance that some observations will consist of noise alone; this may be due to a breakdown of the observation mechanisms. As an example, consider the tracking of a target trajectory, and let the observations be made at discrete time instants. Then there may be a possibility that the observed data do not contain the information concerning the target, because of the misalignment of the radar antenna to the object. The problem of determining the optimal control policy in a situation where at any time there is a nonzero probability that the state can not be observed is formulated by Bellman and Kalaba (1961) as the interrupted stochastic control process; and some results have been obtained by Zadeh (1961) and Eaton (1962). Recently, Nahi (1969) considered the class of linear estimation problem with the interrupted observation mechanisms; and Fujita and Copyright 1972 by Academic Press, Inc. All rights of reproduction in any form reserved. 56

SEQUENTIAL STATE ESTIMATION 57 Fukao (1970) applied the result by Nahi (1969) to a linear stochastic control problem. In this paper, on the basis of the Bayesian approach (Ho and Lee, 1964), we shall consider the problem of estimating the state of linear discrete time systems with the interrupted observation mechanisms. The approximate minimum variance estimator is derived for the case when at any time the probability of occurrence of interruption is known a priori. An adaptive estimator algorithm is also presented for the case when the interruption probability is unknown but fixed throughout the time interval of operation. We observe that unlike the usual Kalman filter, the proposed estimator algorithms are nonlinear with respect to observations and the associated covariance equations are directly dependent upon the observations. Digital simulation studies are carried out for illustrating the approximate and adaptive estimator algorithms presented here and comparing the performance of the new estimator with that of the best linear estimator due to Nahi (1969). 2. STATEMENT OF PROBLEM Let the system be represented by the linear difference equations x(k + 1) = F(k) x(k) + w(k), (2.1) y(k) = 7(k) H(k) x(k) + v(k), (2.2) where x(k) is the n 1 state vector at k-th sample time, y(k) is the p 1 observation vector, w(k) is the n 1 Gaussian white noise vector, and v(k) is the p 1 Gaussian white noise vector. The F(k) is the n n state transition matrix, H(k) is the p X n observation matrix, and 7(k), which characterizes the interrupted observation mechanisms, is the scalar random variable that takes on the values of 0 or 1. The noise vector sequences w(k) and v(k) are assumed to have zero means and covariance matrices Q(k) and R(k), respectively. The initial state x(0) is Gaussian with mean ~(0[--1) &= E{x(0)} and covariance matrix P(0i--1)=& coy{x(0)}. It is further assumed that the random variables w(k), v(k), 7(k) and x(0) are mutually independent. We shall consider the following type of interruption mechanisms, which is treated as Case 1 by Nahi (1969); that is, the {7(k), k = 0, 1,..} are mutually independent random sequence with P(y(k) = 1) = p(k), (2.3) P(7(k) = 0) = 1 -- p(k).

58 SAWARAGI, KATAYAMA AND FUJISHIGE Here, P(.) denotes the probability, and p(k) is the probability that the k-th observation conveys the information concerning the state x(k). Consequently, at any time the observation is interrupted with probability 1 -- p(k). The objective of the paper is to find the sequential estimator that produces the minimum variance estimate ~(k l k) of the state x(k) by observing the data {y(0),..., y(k)}. It is well known (Kalman 1963, Bucy and Joseph, 1968) that the best estimate which minimizes the Bayes risk E([x(k) -- ~(k ] k)]' M[x(k) -- k(k [ k)] [ Y~} (2.4) is given by the conditional expectation ~(k [ k) : E{x(k) l yk). (2.5) Here M is the n n symmetric positive definite matrix, yk is the minimal a-field determined by the observations {y(0)... y(k)}, E{. l yk) denotes the conditional expectation given Y~ (Doob, 1953), and the prime denotes the matrix transpose. Hence, if the conditional probability density function p(x(k) I yk) of x(k) given yk is obtained, the optimal estimate ~(k [ k) can be computed from ~(k I k) = f x(k) p(x(k) I Y~) dx(k). (2.6) 3. CONDITIONAL PROBABILITY DENSITY FUNCTION By the smoothing property of conditional expectations (Doob, 1953), we see that p(x(k) l yk) = E{p(x(k) [ 7(k),..., y(0), YT01 yk} 1 1 =... p(x(k) 17(k) =... 7(0) : io, Y9 ik=0 i0=0 P(7(k) = i~,..., 7(0) : io I yk), (3.1) where P('l yk) is the conditional probability given yk. By using the Bayes rule (Ho and Lee, 1964), we can easily show that the conditional probability density function p(x(k) ] 7(k) : ik...,7(0) : i0, yk) is Gaussian for all k. Therefore, the conditional probability density function p(x(k)] yk) which we seek is the weighted sum of 2 ~+1 Gaussian density functions. This implies that for large values of k, the evaluation of p(x(k) [ Y~) through (3.1) is practically impossible due to the requirement of evergrowing amount of

SEQUENTIAL STATE ESTIM.ATION 59 memory. From this point of view, we shall derive the approximate minimum variance estimator under the following ASSUMPTION. The a priori conditional probability density function p(x(k) I yk-1) of x(k) given yk-1 is Gaussian for all k with E(x(h) l yk-~} =?c(k i h -- 1), cov{x(k) ] yk-~} = p(k[ h _ 1). (3.2) This kind of assumption has already been used in deriving the approximate nonlinear filters for the usual nonlinear estimation problems (Jazwinski, 1970). However, the justification of the assumption has not yet been made; the only related works are the comparative studies on several nonlinear approximate filters (Schwartz and Stear, 1968; Wishner et al., 1969). 4. SEQUENTIAL ESTIMATOR 4.1. Derivation of Estimator Algorithm Rewriting (3.1), we have 1 p(x(k)[ yk) = ~ p(x(k) lt(k ) = i, yk)p(7(k) = i] Y~). (4.1) i=0 This expression is convenient for deriving the approximate sequential estimator algorithm as shown below. Let us first evaluate the conditional probability density function p(x(h) [?(k) = i, yk). By use of the Bayes rule (Ho and Lee, 1964), we can see that p(x(k) l T(k ) = i, yk) = P(Y(k) l T(k) = i, x(k)) p(x(k) [ yk-1) p(y(k) l y(k) = i, yk-1) (4.2) From (2.2), p(y(k) ] 7(k) = i, x(k)) is Gaussian. Therefore, we see under Assumption (3.2) that p(x(k) [7(k) = i, yk) is also Gaussian with mean = ~(k [ k -- 1) + i. F(k)[y(k) -- H(k) ~(k ] k -- 1)], coy = P(k ] k -- 1) -- i. F(k) H(k) P(k I k -- 1), (4.3) where i = 0, 1, and F(k) -~ P(k [ k -- 1) H'(k)[H(k) P(k I k -- 1)H'(k) + R(k)] -1. (4.4)

60 SAWARAGI, KATAYAMA AND FUJISHIGE If i = 1, (4.3) is the same as the usual Kalman filter algorithm, and /'(k) defined by (4.4) is the associated optimal gain matrix. Now consider the conditional probability P(~(k) ~ i lye). To simplify the notation, we define p(i l k) zx P(y(k) = i l yk), i = O, 1. (4.5) Application of the Bayes rule (Ho and Lee, 1964) gives p(1 l k) = p(y(k) IT(k) = 1, yk-1) p(v(k) = 1 ] yk-1) 1 X,=o p(y(k) [ v(k) = i, yk-a) p(v(k) = i I yk-1) (4.6) and p(0 l k) = 1 -- p(1 [ k). From Assumption (3.2), the conditional probability density function p(y(k) I ),(k) = i, yk-1) is Gaussian with mean ~ i. H(k) ~(k I k -- 1), coy -~ i. H(k) P(k [ k -- 1) H'(k) + R(k), (4.7) where i = 0, 1. Since ),(k) is independent of yk-1, we see from (2.3) that P(y(k) = 11 y~-l) = p(k). (4.8) Substituting (4.8) into (4.6) yields where p~(k) p(k) p(1 I k) = p~(k)p(k) + po(k)[1 --p(k)] ' (4.9) pi(k) zx p(y(k) J y(k) = i, yk-1), i = 0, 1. (4.10) Now we shall state the main result in this subsection. The approximate minimum variance estimate ~(k l k) of the state x(k) is given by (see Appendix) N(k l k) = N(k l k -- 1) -t- p(1 I k) F(k)[y(k) -- H(k)N(k l k -- 1)], (4.11) where the initial condition is k(0 ] -- 1) = E(x(0)}. (4.12) Moreover, we can show the approximate conditional covariance matrix P(k l k) = E{[x(k) -- k(k l k)][x(k) -- N(k I k)]' 1 yk} (4.13)

SEQUENTIAL STATE ESTIMATION 61 satisfies the following difference equation (see Appendix): where P(k t k) = P(k I h -- 1) -- p(1 ] k) F(k) H(k) P(k I k -- 1) + p(1 [ h)[1 -- p(1 I h)] F(h) T(k) F'(k), (4.14) T(k) = [y(h) -- H(k) k(k [ k -- 1)][y(h) -- H(h) N(k [ k -- 1)]', (4.15) and the initial condition is P(0 [ -- 1) = coy{x(0)}. (4.16) Finally, using (2.1), (4.11) and (4.14), the one step predicted estimate ~(k @ 1 [ h) and its eovarianee matrix P(h 1 I k) are respectively given by and ~(k + 1 i k) = F(k) ~(h I h) (4.17) P(h @ 1 ] h) = F(k) P(k I k)f'(h) + Q(h). (4.18) These quantities are used as the a priori information to the next state x(h + 1). Equations (4.11)-(4.18) completely specify the structure of the estimator which sequentially produces the approximate minimum variance estimates ~(h i h) of the states x(h). It turns out from (4.9), (4.11) and (4.14) that the estimator dynamics are nonlinear with respect to observations and that the corresponding covariance equation is directly related to the observations. It may be noted that since Nahi (1969) considered only the class of linear estimators, his result is entirely different from the above result. It may also be noted that if the conditional probability p(1 l h) defined by (4.9) is reduced to 1, then the resultant estimator (4.11)-(4.18) becomes the usual Kalman filter (Kalman, 1963; Bucy and Joseph, 1968). This shows the plausibility of the proposed estimator algorithm. 4.2. Adaptive Estimator Algorithm The previous result is based upon the fact that the probability of occurrence of interruption is known a priori. In this subsection, we shall consider the case when the value of the probability p(k) is unknown but fixed throughout the time interval of interest: p(k) ~ P(~,(k) = 1) = q, k = 0, 1,... (4.19)

- q)p0(k)] 62 SAWARAGI, KATAYAMA AND FUJISHIGE Here it is assumed that the a priori probability for q is uniformly distributed over [0, 1]. By the Bayes rule (Ho and Lee, 1964), the conditional probability density function p(q [ yl~) becomes p(q [ yk) = with the initial condition p(y(k) l q, yk-~) p(q l yk-~) f~p(y(k) i q, yk-~)p(q I Y~-I) dq ' (4.20) p(e [ y-1) = 1, 0 ~< q ~< 1. (4.21) We can easily show that p(y(k)] q, yk-~) = qp(y(k) [ ~,(k) = 1, yk-z) + (1 -- q)p(y(k) [y(k) = O,Y k-~) = qpx(k) + (1 -- q)po(k). (4.22) Denoting 1 ~(k -- 1) & E{q] yk-1} = f qp(q] yk-1) dq, (4.23) o we can see from (4.20) and (4.22) that [qpl(k) q- (1 - P(ql y~-a) (4.24) P(q 1 yk) = pa(k) q(k -- 1) + P0(k)[1 -- q(k -- 1)]" From (4.21) and (4.24), it may be noted that p(q I yk) is the polynomial in q of order at most k in the interval [0, 1]. Referring to (4.9), it follows that P0 r k) =~ P(~(k) = 1 [ Y~) = E{P(y(k) = 1 [q, yk)[ yk} qp~(k) ykl. (4.25) E,qpl(k) + (1 -- q) po(k) From (4.23), (4.24) and (4.25), we have the desired result: 1 qpl(k) p(1 ] k) = foqpl(k) + (1 -- q)po(k) p(ql Yk)dq p~(k) q(k -- 1) (4.26) pl(k) q(k -- 1) -}- po(k)[1 -- q(k -- 1)]

SEQUENTIAL STATE ESTIMATION 63 It should be noted that although the conditional probability density functions p~(k)zxp(y(k)iv(k)=i, y~-l) appearing in (4.26)are non- Gaussian, we use the Gaussian density functions defined by (4.7). The adaptive estimator algorithm is obtained by replacing p(1 [k) in (4.11) and (4.14) by new expression (4.26). It should be also noted that for the practical computation of p(q ] yl~) by use of (4.24), the continuous range [13, t] of q is approximated by a finite set of grid points, and that the integration in (4.23) is carried out by using some numerical integration technique. 5. EXAMPLE We shall show the example illustrating the application of the present estimator algorithms and compare the performance of the proposed nonlinear estimator with that of the best linear estimator due to Nahi (1969). Consider the scalar system x(k-~ 1) = ax(k)-~w(k), (5.1) y(k) = 7(k) x(k)+v(k), (5.2) where a [ < 1, and letq(k) =Q,R(k) = R, andp(k) = q. From(4.4), F(k) = P(k l k-- By using (4.11) and (4.14), we have 1)/[P(k l k -- 1)-i- R]. (5.3) ~(k I k) = ~(k [ h -- 1) q- p(1 I k) F(k)[y(k) -- ~(k I k -- 1)], (5.4) /~(k [ k) = P(k ] k -- 1) -- p(1 I h) Pz(k I k -- 1)/X(k) where -t- p(1 I k)[1 -- p(1 ] k)] P2(k [ k -- 1)[y(k) -- ~(k I k -- 1)]2/Z2(k), (5.5) z(k) =~ P(klk -- 1) + R. (5.6) Also, from (4.17) and (4.18), ~(k + 1 I k) = a~(k [ k), P(k + 1 I k) = ~P(k [ k) + 9. (5.7) (5.8) 643/2I/I-5

64 SAWARAGI~ KATAYAMA AND FUJISHIGE Moreover, from (4.7) and (4.9), qr1/2 exp{--[y(k) -- ~(k]k -- 1)]~/2Z(k)} (5.9) p(1 r k) ~ qr~/2 exp{--[y(k) -- ~(k r k -- 1)]~/2X(k)} 4- (1 -- q) X1/~(k) exp{--y~(k)/2r) Digital simulation studies are carried out by using the following set of numerical values: a = 0.95, Q = 0.64, R = 1.69, q ~ 0.8, k(0l--1) = 10, P(0f--1) =2. The initial value of the state x(0) is sampled randomly at each experiment from the population with N(30, 2), whereas the initial estimate k(0l --1) is fixed in the whole experiment as above. Figure 1 displays the running values of the estimate ~(k ] k -- 1), together with the estimate x*(k I k -- 1) due to the linear estimator (Nahi, 1969), and the actual state x(k). The covariances are displayed in Fig. 2. We observe that if 7(k) = 0, the covariance P(k I k -- 1) of the next stage takes a large value; on the other hand, since P*(klk -- 1) due to Nahi (1969) is deterministic, it is independent of the actual observations. This is a typical feature of the 30 ~,, X(0)=29.5 q=0.8,~(01-1) =10 a=0.95 ~ #(01-1)=2.0 R=1.69 ~> /o, o, b,~,~-,=\, ~ Q=0.6z. {,o- ~ "~-,, t ~.,~ o-.-.--o X(klk-1), / ^ \,,o ~,.h. "~ - - X(k) 'I~" o... i,,,, I, ~, i,,,,i... t... I 5 t0 15 20 25 30 DISCRETE TIME (t'(k)=l 1010 T 110101001111110111001111 ) FIG. 1. Sample paths of~(klk-- 1),x*(klk-- t) andx(k).

SEQUENTIAL STATE ESTIMATION 65 o.._o ~'(k~k4) 3.0 s--.-.a P(klk-1).,~.z,-~..~.~ /# ~.~. /" ~ "~"~-%..,~. ~ /~ 2.0 < X o I\ ~ I\ I /\/ ~ /\I ~/ ~ i\ I I 1.0 i 0.01... i... f... I... I... i... I 0 5 10 I5 20 25 30 DISCRETE TIME ('i'('k)=l 1 0 1 0 1 1 1 0 1 0 1 0 0 1 1 1 1 1 1 0 1 1 1 0 0 1 1 1 1 ) FIG. 2. Sample covariances/3(k t k -- 1) and P*(k 1 k -- 1). nonlinear estimator which involves the estimation algorithm (5.9). The performance of the two estimators is compared on the basis of the sample root mean squares (_~ N.2)1/2 8(k) = }2 [x~(k) - #(klk - 1)], (5.1o) #=1 C*(k) = ~ [xj(k) -- x*~(k [ k -- 1)], (5.11) 9=1 where C(k) is the performance index of the nonlinear estimator, and C*(k) is that of the best linear estimator (Nahi, 1969), and the superscriptj denotes the number of simulation run. Total of 50 runs are made in each experiment (N = 50); each run has a different noise sample. The results are shown in Figs. 3 and 4. The difference of the two experiments is that in Fig. 4 the sequence {y(k), k ~ 0, 1,...} is fixed throughout the experiment. We can observe that the nonlinear estimator indicates significant improvement over the linear estimator, especially in the earlier stages of estimation. This fact 643/zz/x-5*

66 SAWARAGI, KATAYAMA AND FUJISHIGE is explained as follows. Since, in the earlier stages, the S/N ratio x2(k)/r is extremely high, then whether 7(k) ---- 0 or 1 can be detected by (5.9) with high accuracy. Therefore, the present nonlinear estimator may produce nearly optimal estimates. We now turn to the adaptive case when the actual value of q is unknown. For the computation of p(q] yk) by (4.24), it is approximated by the values on 101 grid points uniformly spaced in [0, 1]. Thus from (4.24), [q~pl(k) -F (1 -- q~)po(k)] P(q81 Y~-O, (5.]2) where qs = silo0, s = 1..., 100, and from (4.23), 100 ~(k -- 1) = ~, qsp(q~ t Yk-1)/lO0. (5.13) 8=1 20 E{X(O)}=30 X(OI-1) :10 q=0.8 a=0.95 15. 15(01-1) = 2.0 R= 1.69 l\ Q:o6 0 5 10 15 20 25 30 DI5CRETE TIME FIG. 3. Performance indices (5.10) and (5.11) for 50 run average.

20J E[X(O)}=30 q=0.8 15 t ~(01-1) = 10 a:0.95 _ ~ #(01-1) =2.0 R=1.69 o = e t :--~ C(k),0 z~-""--'~ d(k) \ A 0 0 5 10 15 20 25 30 DISCRETE TIME (~'(k)=l O I 1 01 1 1 1 0 1 1 1 1 101 1 1 1 1 10101 1 1 11 ) FIG. 4. Performance indices (5.10) and (5.11) for 50 run average, where the realization of y(k)'s is fixed. 0.9 ~0.7 f,,~,,,~,..,a,- ~<.--- ---,, g(k-1) o ~-0.e ~V V o o ~(k 1) o.s q=0.8 O,Z,... I,,,,I,,,,I... I,,,,I,,,,I 5 10 15 2O 25 30 DISCRETE TIME ('((k)=10! 1 0 1 111 1 011 1 1 11 1 1 O1 1 11 100 1 1 1) FIG. 5. Sample path ~(k -- 1) corresponding to the realization of 7@)'s indicated, and 50 run average ~(k -- 1) of ~(k -- 1) for q = 0.8 (unknown).

68 SAWARAGI, KATAYAMA AND FUJISHIGE Then from (4.26), ~(k -- 1) px(k) (5.14) p(1 I k) = q(k - 1)pl(k) + [1 - q(k - 1)] p0(k)" Here it may be noted that p(1 t k) defined by (5.14) is also obtained by replacing q in (5.9) by q(k -- 1) given by (5.13). As stated in subsection 4.2, the adaptive estimator algorithm is formed by using new expression (5.14) for p(1 ] k) appearing in (5.4) and (5.5). Digital simulation studies are also carried out; the sample behavior for k(klk -- 1) and ['(k[k -- 1) is almost the same as those of Figs. 1 and 2. Figure 5 displays the tracking behavior of ~(k -- 1) for the unknown value of q, together with the sample average ~(k -- 1) of q(k -- 1) for 50 simulation runs. We may observe that although the present algorithm is an approximate one, the q(k -- 1) serves as an estimate of the value of unknown probability q. 6. DISCUSSION For the interrupted observation process considered here, we could implement the likelihood ratio test (Van Trees, 1968) in order to detect whether 7(k) is 0 or 1. After the decision has been made, we believe that it is absolutely correct. If the answer is 7(k) ~ 1, then the observed data will be used by the estimator for the purpose of prediction or filtering. However, associated with the decision made, there is a nonzero probability that the observation contains no information concerning the state to be estimated, that is, a nonzero probability of false alarm. It is clear that this information, which is contained in the conditional probabilityp(1 ] k), is useful for improving the performance of the estimator. As is seen from (4.9) and (4.26), thep(1 [ k) is nonlinear with respect to observations. Therefore, if we restrict ourselves to the class of linear estimators, we can not utilize this information. In view of this, the linear estimators developed by Nahi (1969) may not be very effective. However, it should be noted that since the new approximate estimator algorithms are entirely based upon the assumption that the a priori conditional probability density function p(x(k) [ y7~-1) is Gaussian, there is a possibility of divergence of estimation error (Jazwinski, 1970), and the quality of estimate is not known from the covarianee equations defined by (4.14) and (4.18). Therefore, the justification of the assumption must be made in connection with the stability of the new estimator algorithms. These are common problems associated with the nonlinear approximate filtering(jazwinski, 1970). But the evaluation of the accuracy of the approximate nonlinear estimators

SEQUENTIAL STATE ESTIMATION 69 is in general extremely difficult; there are only few papers on this problem so far (Schwartz and Stear, 1968; Wishner et al., 1969). Finally, it may be also noted that the advantage of Nahi's linear estimators is that it is free from the above mentioned difficult problems associated with approximate nonlinear estimators; that is, the quality of estimate is known a priori from the covariance equations. 7. CONCLUSIONS This paper considers the sequential state estimation for discrete time systems with interrrupted observation mechanisms. Since the discussion is not confined to the class of linear estimators, the Bayesian approach can be applied successfully for deriving the sequential estimators. It should be noted that the resultant estimator algorithms are nonlinear with respect to observations and that the covariance equations are dependent upon the actual observations. Furthermore, the present method can easily be extended to the adaptive case when the probability of occurrence of interruption is unknown but fixed during the entire time interval of interest. Although the justification of Assumption (3.2) has not yet been available, digital simulation studies demonstrate the feasibility of the proposed estimator algorithms. APPENDIX Derivation of (4.11) Referring to (4.3), we define zi(k) ~ E{x(k) [ 7(k) = i, yk} = ~(k [ k -- 1) + i. F(k)[y(k) -- H(k) ~(k ]k -- 1)], (A.1) where i = 0, 1. Then use of (2.6), (4.1) and (A.1) yields ~(k[k) = ~ P(7(k) : 1 i=0 i [ Vk) E{x(k) l y(k) -~ i, yk} = p(o ] k) zo(k ) + p(1 ] k) Zl(k ) = ~(k [ k -- 1) + p(1 [ k) F(k)[y(k) -- H(k) ~(k [ k -- 1)], (A.2) where p(i i k) are defined by (4.5).

70 SAWARAGI, KATAYAMA AND FUJISHIGE Derivation of (4.14) We see from (4.1) and (4.13) that 1 P(k I k) = ~ p(i I k) E{[x(k) -- ~(k [ k)][x(k) -- ~(k ] k)]' I 7(k) = i, yk} i=0 where p(o l k) So(k ) -k p(1 I k) Sl(k), (A.3) Si(k) ~ E{[x(k) -- ~(k [ k)][x(k) -- ~(k ] k)]' [ y(k) = i, yk}, i = 0, 1. (A.4) Using (A.1) and (A.4), we can easily show that From (A.1) and (4.3), So(k ) = E{[x(k) -- zo(k)][x(k ) -- zo(k)]' ] y(k ) = O, yk} + [z0(k) -- ~(k[k)][zo(k) -- ~(k i k)]'. (A.5) E{[x(k) -- zo(k)][x(k ) -- zo(k)]' l y(k ) -~ O, Y~} = P(k [ k -- 1), (A.6) and also from (A.1) and (A.2), zo(k ) -- ~(k [ k) = --p(1 ] k) F(k)[y(k) -- H(k) ~(k ] k -- 1)]. (A.7) Application of (A.6) and (A.7) to (A.5) yields So(k ) = P(k [ k -- 1) + p2(1 I k)f(k) T(k) F'(k), (A.8) where T(k) is defined by (4.15). Similarly, we can show that S~(k) = E{[x(k) -- z~(k)][x(k) -- z~(k)]' [ ~(k) = 1, yk} + [zl(k) -- ~(klk)][z~(k) -- ~(k I k)]' = P(k I k -- 1) r(k) H(k) P(k I k -- 1) + p2(o I k) F(k) T(k) F'(k). (A.9) Substituting (A.8) and (A.9) into (A.3) yields the desired result (4.14). ACKNOWLEDGMENT Authors would like to express their thanks to Prof. H. Akashi, who read the manuscript and suggested improvements. I~CEIVED: November 15, 1971

SEQUENTIAL STATE ESTIMATION 71 REFERENCES BELLMAN, R. AND KALABA, R. (1961), A note on interrupted stochastic control processes, Information and Control 4, 346-349. BuoY, R. S. AND JOSEPH, P. D. (1968), "Filtering for Stochastic Processes with Applications to Guidance," Interscience, New York. DooB, J. L. (1953), "Stochastic Processes," John Wiley, New York. EATON, J. H. (1962), Discrete time interrupted stochastic control processes, J. Math. Anal. Appl. 5, 287-305. FUJITA, S. AND FUKAO, T. (1970), Stochastic optimal linear control for discrete-time systems with interrupted observations, Preprints of IFAC Kyoto Symp. on Systems Engrg. Approach to Computer Control, Kyoto, August, 329-334. Ho, Y. C. AND LEE, R. C. K. (1964), A Bayesian approach to problems in stochastic estimation and control, IEEE Trans. Automatic Control AC-9, 333-339. JAZWlNSKI, A. H. (1970), "Stochastic Processes and Filtering Theory," Academic Press, New York. KALMAN, R. E. (1963), New methods in Wiener filtering theory, in "Proc. of 1st Syrup. on Engrg. Applications of Random Function Theory and Probability" (J. L. Bogdanoff and Frank Kozin, Eds.), pp. 270-388, John Wiley, New York. NAHI, N. E. (1969), Optimal recursive estimation with uncertain observation, IEEE Trans. Information Theory IT-lg, 457-462. SCHWARTZ, L. and STEAR, E. B. (1968), A computational comparison of several nonlinear filters, IEEE Trans. Automatic Control AC-13, 504-514. VAN TREES, H. L. (1968), "Detection, Estimation, and Modulation Theory," Part 1, John Wiley, New York. WISHNER, R. P., TABACZYNSKI, J. A., AND ATHANS, M. (1969), A comparison of three nonlinear filters, Automatica 5, 487-496. ZADEH, L. A. (1961), Remarks on the paper by Bellman and Kalaba, Information and Control 4, 350-352.