RECURSIVE OUTLIER-ROBUST FILTERING AND SMOOTHING FOR NONLINEAR SYSTEMS USING THE MULTIVARIATE STUDENT-T DISTRIBUTION

Similar documents
NON-LINEAR NOISE ADAPTIVE KALMAN FILTERING VIA VARIATIONAL BAYES

Recursive Noise Adaptive Kalman Filtering by Variational Bayesian Approximations

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 6: Bayesian Inference in SDE Models

Prediction of ESTSP Competition Time Series by Unscented Kalman Filter and RTS Smoother

Lecture 7: Optimal Smoothing

Extension of the Sparse Grid Quadrature Filter

Lecture 2: From Linear Regression to Kalman Filter and Beyond

On Continuous-Discrete Cubature Kalman Filtering

EM-algorithm for Training of State-space Models with Application to Time Series Prediction

A student's t filter for heavy tailed process and measurement noise

in a Rao-Blackwellised Unscented Kalman Filter

Expectation Propagation in Dynamical Systems

Rao-Blackwellized Particle Filter for Multiple Target Tracking

Robust Non-linear Smoother for State-space Models

Expectation Propagation Algorithm

The Unscented Particle Filter

Optimal filtering with Kalman filters and smoothers a Manual for Matlab toolbox EKF/UKF

A New Nonlinear Filtering Method for Ballistic Target Tracking

SEQUENTIAL MONTE CARLO METHODS FOR THE CALIBRATION OF STOCHASTIC RAINFALL-RUNOFF MODELS

A Systematic Approach for Kalman-type Filtering with non-gaussian Noises

Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations

SQUARE-ROOT CUBATURE-QUADRATURE KALMAN FILTER

Lecture 1a: Basic Concepts and Recaps

State Space Representation of Gaussian Processes

A Comparison of Particle Filters for Personal Positioning

SIGMA POINT GAUSSIAN SUM FILTER DESIGN USING SQUARE ROOT UNSCENTED FILTERS

Estimation and Maintenance of Measurement Rates for Multiple Extended Target Tracking

Lecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Approximations

A Comparison of the EKF, SPKF, and the Bayes Filter for Landmark-Based Localization

Expectation Propagation for Approximate Bayesian Inference

Iterative Filtering and Smoothing In Non-Linear and Non-Gaussian Systems Using Conditional Moments

Online tests of Kalman filter consistency

Nonlinear and/or Non-normal Filtering. Jesús Fernández-Villaverde University of Pennsylvania

Quadrature Filters for Maneuvering Target Tracking

The Kalman Filter ImPr Talk

On the Conditional Distribution of the Multivariate t Distribution

Tyre-road friction potential estimation by data fusion

BAYESIAN ESTIMATION OF TIME-VARYING SYSTEMS:

Gaussian Process Approximations of Stochastic Differential Equations

A Gaussian Mixture PHD Filter for Nonlinear Jump Markov Models

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

Lecture 8: Bayesian Estimation of Parameters in State Space Models

Time Series Prediction by Kalman Smoother with Cross-Validated Noise Density

Efficient Monitoring for Planetary Rovers

Dynamic System Identification using HDMR-Bayesian Technique

Constrained State Estimation Using the Unscented Kalman Filter

Iterative Diffusion-Based Distributed Cubature Gaussian Mixture Filter for Multisensor Estimation

CLOSE-TO-CLEAN REGULARIZATION RELATES

arxiv: v1 [cs.ro] 1 May 2015

Extended Object and Group Tracking with Elliptic Random Hypersurface Models

Smoothers: Types and Benchmarks

STA 4273H: Statistical Machine Learning

Combined Particle and Smooth Variable Structure Filtering for Nonlinear Estimation Problems

Rao-Blackwellised posterior linearisation backward SLAM

Introduction to Unscented Kalman Filter

arxiv: v1 [cs.sy] 12 Dec 2016

2D Image Processing (Extended) Kalman and particle filter

Partitioned Update Kalman Filter

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

Tampere University of Technology Tampere Finland

Sequential Monte Carlo Methods for Bayesian Computation

Bayesian Inference for the Multivariate Normal

TSRT14: Sensor Fusion Lecture 8

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER

Bagging During Markov Chain Monte Carlo for Smoother Predictions

Mobile Robot Localization

Adaptive Monte Carlo methods

Approximate Bayesian inference

Part 1: Expectation Propagation

This is a repository copy of A Novel Robust Rauch-Tung-Striebel Smoother Based on Slash and Generalized Hyperbolic Skew Student s T-Distributions.

EKF, UKF. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics

Outlier-insensitive Kalman Smoothing and Marginal Message Passing

EKF, UKF. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics

Higher-Degree Stochastic Integration Filtering

Randomized Unscented Kalman Filter in Target Tracking

Mini-Course 07 Kalman Particle Filters. Henrique Massard da Fonseca Cesar Cunha Pacheco Wellington Bettencurte Julio Dutra

A NEW NONLINEAR FILTER

Bayesian Methods for Machine Learning

Parameterized Joint Densities with Gaussian Mixture Marginals and their Potential Use in Nonlinear Robust Estimation

1 Kalman Filter Introduction

Recursive Least Squares for an Entropy Regularized MSE Cost Function

Estimating the Shape of Targets with a PHD Filter

Gaussian Mixtures Proposal Density in Particle Filter for Track-Before-Detect

A New Nonlinear State Estimator Using the Fusion of Multiple Extended Kalman Filters

RAO-BLACKWELLIZED PARTICLE FILTER FOR MARKOV MODULATED NONLINEARDYNAMIC SYSTEMS

Expectation propagation for signal detection in flat-fading channels

Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference

Bayesian Filtering for Dynamic Systems with Applications to Tracking

Outlier-insensitive Kalman Smoothing and Marginal Message Passing

Conditional Posterior Cramér-Rao Lower Bounds for Nonlinear Sequential Bayesian Estimation

Using the Kalman Filter to Estimate the State of a Maneuvering Aircraft

Data Assimilation for Dispersion Models

Integrated Non-Factorized Variational Inference

Performance Evaluation of Local State Estimation Methods in Bearings-only Tracking Problems

arxiv: v2 [cs.sy] 13 Aug 2018

Probabilistic Graphical Models

A NOVEL OPTIMAL PROBABILITY DENSITY FUNCTION TRACKING FILTER DESIGN 1

Lecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions

Linear Dynamical Systems

Transcription:

1 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEPT. 3 6, 1, SANTANDER, SPAIN RECURSIVE OUTLIER-ROBUST FILTERING AND SMOOTHING FOR NONLINEAR SYSTEMS USING THE MULTIVARIATE STUDENT-T DISTRIBUTION Robert Piché Tampere University of Technology Department of Mathematics Tampere, Finland Simo Särä and Jouni Hartiainen Aalto University Department of Biomedical Engineering and Computational Science Espoo, Finland ABSTRACT Nonlinear Kalman filter and Rauch Tung Striebel smoother type recursive estimators for nonlinear discrete-time state space models with multivariate Student s t-distributed measurement noise are presented. The methods approximate the posterior state at each time step using the variational Bayes method. The nonlinearities in the dynamic and measurement models are handled using the nonlinear Gaussian filtering and smoothing approach, which encompasses many nown nonlinear Kalman-type filters. The method is compared to alternative methods in a computer simulation. Index Terms Robust filtering, Robust smoothing, Variational Bayes, Gaussian filter, Gaussian smoother 1. INTRODUCTION Because large deviations typically occur in real data more frequently than is modeled by the normal i.e. Gaussian) distribution, there is interest in developing estimation algorithms that can cope with these outliers while retaining the computational efficiency of Gaussian models. In statistics, a popular approach to robustify linear multivariate regression is to use the Student s t-distribution in place of the Gaussian distribution and to compute the estimates using an Expectation-Maximization iteration [1 4]. This outlier-accommodation approach has many advantages: it is easy to implement, computationally light, amenable to Bayesian analysis, and doesn t require parameter tuning. The approach has recently been extended to the sequential linear filtering and smoothing setting by Agamennoni et al. [5]. Their filter approximates the posterior state at each time step using an iterative solution of a variational Bayes formula; each iteration resembles a conventional Kalman filter update. Similar filter equations had been presented earlier in [6, 7], but the former wor focused on on-line learning of the linear model s measurement noise variance parameter, while the latter wor focused on outlier-detection rather than outlier-accommodation. In this wor we extend the outlier-robust filter and smoother to nonlinear state space models. For the sae of generality, we use the Gaussian filtering and smoothing framewor [8 11] also referred to as Gaussian assumed density filtering and smoothing). For Gaussian noise, this framewor includes as special cases many popular nonlinear Kalman-type filters, including the extended Kalman filter EKF) [1], the unscented Kalman filter UKF) [13], the cubature Kalman filter CKF) [14], the Gauss-Hermite Kalman filter GHKF) [9], and the corresponding smoothers [11, 15, 16].. VARIATIONAL BAYES APPROXIMATION FOR FILTERING AND SMOOTHING.1. Problem Formulation Consider the non-linear state space model with t-distributed measurement noise, x x 1 Nfx 1 ), Q ) y x Studenthx ), R, ν), 1a) 1b) where f ) and h ) are non-linear dynamic and measurement model functions 1. The matrix R is symmetric positive definite, and the Student s t-distribution s probability density is py x ) 1 + 1 ν y hx )) T R 1 y hx )) ) ν+d, where d is the dimension of y and ν 1 is a parameter that controls the Student density s urtosis heavy-tailedness); for general practice a value of ν = 4 is recommended [1]. The case ν = 1 is the Cauchy distribution and for ν the distribution converges to Nhx ), R ). In filtering the aim is to estimate the state distribution px y 1: ) at time steps = 1,..., T. In contrast, in smoothing the aim is to estimate the state posterior given 1 For clarity, we assume that f ) and h ) do not depend on, but that dependency can be included easily to the presented algorithms. 978-1-4673-16-/1/$31. c 1 IEEE

future measurements, that is, px y 1:T ) with = 1,..., T. With the model 1) these distributions are analytically intractable, and so in this article we aim to provide approximate algorithms that provide reasonably accurate solutions efficiently... Nonlinear Kalman Filtering and Smoothing Even in the case of Gaussian measurement noise i.e. ν in 1b)), the nonlinear filtering and smoothing problems are analytically intractable [8, 1]. An efficient approach for approximately solving the filtering problem with Gaussian measurement noise is the Gaussian filtering GF) [8 1] framewor. In Gaussian filter, the state posterior density px y 1: ) is approximated with a multivariate Gaussian density Nx m, P ) at each time step = 1,..., T, with the mean m and covariance P computed via the following steps: Prediction: m = fx 1 ) Nx 1 m 1, P 1 ) dx 1, = fx 1 ) m 1 ) fx 1 ) m 1 ) T P Nx 1 m 1, P 1 ) dx 1 + Q. Update: µ = hx ) Nx m, P ) dx S = hx ) µ )hx ) µ ) T Nx m, P ) dx + R C = x m ) hx ) µ ) T Nx m, P ) dx K = C S 1 m = m + K y µ ) P = P K S K T. In [11] it was shown that approximating Gaussian smoothing densities Nx m s, P s ) for the model class 1), but with Gaussian measurement noise can obtained by solving the following Rauch Tung Striebel RTS) smoother type equations in a bacward order = T 1,..., 1: L = x m )fx ) m +1 )T Nx m, P ) dx G = L [P +1 ] 1 m s = m + G m s +1 m +1 ) P s = P + G P s +1 P +1 )GT. These filtering and smoothing methods, however, cannot be directly used in the case of Student s t-distributed measurement noise, but in the following we show how they can be used as part of a Variational Bayes VB) framewor to provides the required approximations efficiently..3. Non-Linear Variational Bayes Filter Introducing an auxiliary random variable λ, the measurement model s Student s t-distribution can be expressed as the Gaussian mixture py x ) = py x, λ )pλ ) dλ, where y x, λ Nhx ), 1 λ R ) and λ Gamma ν, ν ). While assuming that px y 1: 1 ) is approximated with a Gaussian density Nx m, P ) using the prediction step of the non-linear Gaussian filter above, the state posterior can be found by marginalising λ out of the joint posterior px, λ y 1: ) λ d e λ y hx )) T R 1 y hx )) e 1 x m )T P ) 1 x m ) λ ν 1 e νλ. In order to mae this computation tractable we use a standard variational Bayes VB) approach. We see a separable approximation in the form of a product of probability densities, px, λ y 1: ) qx )qλ ). The VB approximation minimizes the Kullbac-Leibler KL) divergence between the product approximation and the true posterior, KLqx )qλ ) px, λ y 1: )) = qx )qλ ) log qx )qλ ) px, λ y 1: ) ) dx dλ. Using the calculus of variations to minimize the KL divergence with respect to qλ ) while eeping qx ) fixed yields [17, p. 466] log qλ ) = E x log px, λ, y 1: ) ) + const. where = 1 λ γ + ν+d 1 ) log λ νλ + const., ) γ = E x y hx )) T R 1 y hx )) ) = tr{e x y hx ))y hx )) T ) R 1 }. 3) This expectation can be approximated with the same methods as used in approximating the Gaussian integrals in non-linear Kalman filtering and smoothing. Thus, qλ ) is the density of a Gamma ν+d, γ +ν ) distribution with mean λ = E λ λ ) = λ qλ ) dλ = ν + d ν + γ. 4)

Similarly, minimizing the KL divergence with respect to qx ), with qλ ) fixed, yields log qx ) = E λ log px, λ, y 1: ) ) + const. = 1 λ y hx )) T R 1 y hx )) 1 x m )T P ) 1 x m ) + const. This can be seen as exactly the same estimation problem as in the update step of non-linear Kalman filter above, and thus we can approximate qx ) with a Gaussian using the update step equations with measurement covariance λ 1 R. The variational parameters of qλ ) and qx ) are coupled, and can be solved with fixed-point iteration such that parameters of qλ ) are solved while eeping qx ) fixed and vice versa, which results in Algorithm 1. Algorithm 1: One time step of the VB filter for the non-linear model with Student t-distributed measurement noise Data: model parameters f), h), Q, R, ν, prior state Gaussian parameters m 1, P 1, and number of VB fixed-point iterations N Result: posterior state Gaussian parameters m, P 1 begin m fx 1 ) Nx 1 m 1, P 1 ) dx 1 P 3 fx 1 ) m ) fx 1) m )T Nx 1 m 1, P 1 ) dx 1 + Q 4 λ 1 5 for n from 1 to N do µ 6 hx ) Nx m, P ) dx S 7 hx ) µ ) hx ) µ ) T Nx m, P ) dx + λ 1 R C 8 x m ) hx ) µ ) T Nx m, P ) dx 9 K C S 1 1 m m + K y µ ) 11 P P K S K T D 1 y hx )) y hx )) T Nx m, P ) dx 13 γ trd R 1 ) 14 λ ν+d ν+ γ 15 endfor 16 end.4. Non-Linear Variational Bayes Smoother The non-linear filtering algorithm can also be extended to provide approximate fixed-interval smoothing solutions, that is, factorizing solutions qx )qλ ) that approximate the true posterior px, λ y 1:T ) at each time step = 1,..., T. Using the standard VB approach, log qλ 1:T ) can be written as log qλ 1:T ) = E x1:t log px 1:T, λ 1:T, y 1:T )) + const. T = E x log py x, λ )) + log pλ ) + const. =1 Because the terms of the sum are jointly independent, we have log qλ ) = 1 λ E x y hx )) T R 1 y hx )) ) + ν+d 1 ) log λ νλ + const., which is of the same form as ) except that the expectation is taen over the smoothed state distribution qx ). Similarly, for log qx 1:T ) we get log qx 1:T ) = E λ1:t log px 1:T, λ 1:T, y 1:T )) + const. T = log px x 1 ) + log py x, λ ) + const. =1 This can be identified as a non-linear smoothing problem with nown measurement covariance, and thus be approximated with non-linear RTS smoothing algorithms [11]. As the result we have an algorithm that alternates between updating the approximations qx ) and qλ ) while eeping the other terms fixed such that qx ) are updated using nonlinear Kalman filtering and smoothing, and qλ ) by solving the parameters from 5) similarly as in the non-linear VB filter above. Several iterations are needed for the parameters to converge towards a fixed value. The resulting fixed-interval smoother is shown Algorithm. Because the smoothing algorithm also updates the estimates of the last time step, it would be possible to use fixedlag smoothing to enhance the filtering results. A Gaussian fixed-lag smoother can be obtained as a simple modification to the fixed-interval smoothers as is shown in [11]..5. Conditionally Independent Measurements In some cases it may be more appropriate to use a marginal t-distribution for each measurement channel. That is, in place of the multivariate t-distribution 1b), the measurement model is taen to be a product of t-marginals py x ) = d Studenty h,i x ), r i,, ν i ). i=1 The VB algorithms developed above can be easily modified for this independent measurement-channel model, by using d- dimensional vectors for λ and γ, and updating each of the components in the algorithms. 3. NUMERICAL EXAMPLE To illustrate the use of the proposed algorithms, we consider a computer simulation of a multi-sensor bearings-only target

Algorithm : Variational Bayes smoother for interval = 1,..., T for the non-linear state-space model with Student t-distributed measurement noise Data: model parameters f), h), Q, R, ν and variational noise parameters λ for each time step = 1,..., T, state prior parameters m and P and number of VB fixed-point iterations N Result: Smoothed state and noise parameters m s, P s and λ for each time step = 1,..., T 1 begin for n from 1 to N do 3 for from 1 to T do 4 m fx 1 ) Nx 1 m 1, P 1 ) dx 1 5 P fx 1) m ) fx 1) m )T Nx 1 m 1, P 1 ) dx 1 + Q µ 6 hx ) Nx m, P ) dx S 7 hx ) µ ) hx ) µ ) T Nx m, P ) dx + λ 1 R C 8 x m ) hx ) µ ) T Nx m, P ) dx 9 K C S 1 1 m m + K y µ ) 11 P P K S K T 1 endfor 13 m s T m T 14 PT s P T 15 for from T 1 to 1 do L 16 x m ) fx ) m +1 )T Nx m, P ) dx 17 G L [P +1 ] 1 18 m s m + G m s +1 m +1 ) 19 P s P + G P+1 s P +1 )GT endfor 1 for from 1 to T do D y hx )) y hx )) T Nx m s, P s) dx 3 γ trd R 1 ) 4 λ ν+d ν+ γ 5 endfor 6 endfor 7 end tracing problem, where the sensors measure the angle between the target and the sensor. Outliers in the sensor measurements are introduced using a clutter model that is not Student-t. The dynamics of the target are modeled with coordinated turning model, where the state x = u, u, v, v, ω) contains the d location u, v) and the corresponding velocities u, v) as well as the turning rate ω of the target. The dynamic model can be written as sinω t) cosω t) 1 1 ω ω cosω t) sinω t) x = 1 cosω t) sinω t) ω 1 ω x 1 + q, sinω t) cosω t) 1 5) where q N, Q ) is the process noise. In our simulations we use the parameters Q = q 1M ) q 1 M t, M = 3 /3 t / t q, t where q 1 =.1, q = 1.75 1 4 and t = 1. We generate measurements for sensors i = 1,..., d according to the mixture model { U π, π), if ci, = y i, x, c i, ) v s Narctan i v u s, σ i i, ), if c i, = 1 u 6) Here c i, = indicates that the measurement from the ith sensor at the th time step is clutter, while c i, = 1 if the measurement is from the actual target. Clutter is uniformly distributed over the angle measurement space [ π, π], whereas the measurements from the actual target are the angle values from sensors located at s i u, s i v) corrupted with additive Gaussian distributed noise with variances σi,. In our simulation scenario we have d = 4 sensors placed at the corners of the square [ 1, 1] [ 1, 1], and the variances are set to σi, =.1. We assume that the probability of clutter is constant over time, that is, c i, is a Bernoulli random variable with probc i, ) = p c. In our simulations we used the clutter probability p c =.1. We simulated the target trajectory 1 times for 4 time steps according to dynamic model 5) and generated measurements on each time step with model 6). With the generated measurement sequences we assessed the performance of following filters and smoothers in estimating the state trajectories: JT: Non-linear Student s t-filter and smoother developed in this article with a joint t-distribution ν = 5) as the measurement model. In VB filter, 5 fixed point iterations were used, which is enough to achieve convergence in all cases. In smoothers we used 5 VB iterations. MT: Same as JT with the joint t-distribution replaced by separate t-marginals for each of the 4 sensors. MC5 and MC: Rao-Blacwellized Monte Carlo data association RBMCDA) algorithm [18] with 5 and particles. This is an ideal method for this simulation

scenario since the clutter model specified above is a special case of model class assumed by RBMCDA. PA: Gaussian ADF with perfect associations, that is, clutter measurements are not given to the filter. 5 4 3 1 5 4 3 1 We used the numerical approximation methods of UKF, CDKF, CKF and GHKF in approximating the necessary Gaussian integrals in every tested filtering algorithm. In UKF we used the transformation parameters α = 1, β =, κ = ), in CDKF the step size h cd = 3 and in GHKF m = 3 n = 43 quadrature points. 1 8 7 6 5 4 3 a) Example filtered estimate True trajectory MT estimate MC estimate 1 8 7 6 5 4 3 b) Example smoothed estimate Fig.. Example trajectory with filtered and smoothed estimates. CPU time s) CPU time s) 15 1 5 15 1 5 JT MT MC5 MC a) CPU time filters) UKF CDKF CKF GHKF JT MT MC5 MC b) CPU time smoothers) Fig. 1. The average CPU times taen by the methods are shown for filters a) and for filter and smoothers combined b). Figure 1 shows the CPU times taen by the methods, showing that the Student s t based filters are many times faster than RBMCDA. We can also see that while GHKF gives the highest estimation accuracy it also much slower than the rest since the number of quadrature points m is so large in general m grows exponentially with the number of state components). Among the other methods CKF provides a good trade-off in accuracy and speed. Both t-filters also provide adequate tracing accuracy, which is further illustrated in Figure, which show typical target trajectories and estimates of it given by MT and MC with CKF/CRTS). The root mean square errors RMSE) of estimating the target positions and turn rates are shown in panels a) d) of Figure 3. We can see that the performance of RBMCDA is nearly optimal with both 5 and particles. The regular Gaussian measurement model was also tested, but it diverged in all simulation cases. Of the tested integration methods GHKF gave the best performance in all cases, followed by CKF, CDKF and UKF in that order. 4. CONCLUDING REMARKS In this article, we have presented a new outlier-robust filter and smoother for nonlinear state space models. The filter and smoother are based on using Student s t-distribution in the measurement model. The methods approximate the posterior state at each time step using the variational Bayes method and handle the non-linearities using the Gaussian filtering and smoothing approach. The method was compared to the particle filtering based method RBMCDA in computer simulation, which showed that the proposed approach provides a good trade-off between accuracy and computational efficiency. 5. REFERENCES [1] K. L. Lange, R. J. A. Little, and J. M. G. Taylor, Robust statistical modeling using the t distribution, Journal of the American Statistical Association, vol. 84, no. 48, pp. 881 896, 1989. [] R. J. A. Little and D. B. Rubin, Statistical Analysis with Missing Data, Wiley Series in Probability and Statistics. Wiley, New Yor, 1987. [3] A. Gelman, J. B. Carlin, H. S. Stern, and D. B. Rubin, Bayesian Data Analysis, Chapman & Hall/CRC, second edition, 4. [4] R. Piché, Robust multivariate linear regression using the Student-t distribution, Tech. Rep., Tampere University of Technology, Tampere, Finland, May 1.

RMSE position) RMSE turn rate) RMSE position) RMSE turn rate) 9 8 7 6 5 4 3 1.9.8.7.6.5.4.3..1 9 8 7 6 5 4 3 1.5.4.3..1 JT MT MC5 MC PA a) Filter RMSE position) JT MT MC5 MC PA b) Filter RMSE turn rate) JT MT MC5 MC PA c) Smoother RMSE position) JT MT MC5 MC PA d) Smoother RMSE turn rate) UKF CDKF CKF GHKF Fig. 3. RMSE values of estimating positions and turn rates with all the tested methods over 1 simulations. [5] G. Agamennoni, J.I. Nieto, and E.M. Nebot, An outlier-robust Kalman filter, in IEEE Int. Conf. on Robotics and Automation ICRA), May 11, pp. 1551 1558. [6] S. Särä and A. Nummenmaa, Recursive noise adaptive Kalman filtering by variational Bayesian approximations, IEEE Transactions on Automatic Control, vol. 54, no. 3, pp. 596 6, 9. [7] J.-A. Ting, E. Theodorou, and S. Schaal, A Kalman filter for robust outlier detection, in Proc. IEEE/RSJ Intl. Conf. Intelligent Robots and Systems, Oct. 7, pp. 1514 1519. [8] P. Maybec, Stochastic Models, Estimation and Control, vol., Academic Press, 198. [9] K. Ito and K. Xiong, Gaussian filters for nonlinear filtering problems, IEEE Transactions on Automatic Control, vol. 45, no. 5, pp. 91 97, May. [1] Y. Wu, D. Hu, M. Wu, and X. Hu, A numericalintegration perspective on Gaussian filters, IEEE Transactions on Signal Processing, vol. 54, no. 8, pp. 91 91, August 6. [11] S. Särä and J. Hartiainen, On Gaussian optimal smoothing of non-linear state space models, IEEE Transactions on Automatic Control, vol. 55, no. 8, pp. 1938 1941, August 1. [1] A. H. Jazwinsi, Stochastic Processes and Filtering Theory, Academic Press, 197. [13] S. J. Julier, J. K. Uhlmann, and H. F. Durrant-Whyte, A new approach for filtering nonlinear systems, in Proceedings of the 1995 American Control, Conference, Seattle, Washington, 1995, pp. 168 163. [14] I. Arasaratnam and S. Hayin, Cubature Kalman filters, IEEE Transactions on Automatic Control, vol. 54, no. 6, pp. 154 169, June 9. [15] S. Särä, Unscented Rauch-Tung-Striebel smoother, IEEE Transactions on Automatic Control, vol. 53, no. 3, pp. 845 849, April 8. [16] I. Arasaratnam and S. Hayin, Cubature Kalman smoothers, Automatica, vol. 47, no. 1, pp. 45 5, 11. [17] C. M. Bishop, Pattern Recognition and Machine Learning, Springer-Verlag New Yor, Inc., Secaucus, NJ, USA, 6. [18] S. Särä, A. Vehtari, and J. Lampinen, Rao- Blacwellized particle filter for multiple target tracing, Information Fusion, vol. 8, no. 1, pp. 15, 7.