Modern Navigation. Thomas Herring

Similar documents
Principles of the Global Positioning System Lecture 11

Modern Navigation. Thomas Herring

Satellite Navigation error sources and position estimation

GPS Geodesy - LAB 7. Neglecting the propagation, multipath, and receiver errors, eq.(1) becomes:

Satellite Navigation PVT estimation and integrity

Modern Navigation

Modern Navigation. Thomas Herring

Principles of the Global Positioning System Lecture 14

STAT2201 Slava Vaisman

Modern Navigation. Thomas Herring

GPS point positioning (using psuedorange from code). and earthquake location (using P or S travel times).

Linear Algebraic Equations

Velocity (Kalman Filter) Velocity (knots) Time (sec) Kalman Filter Accelerations 2.5. Acceleration (m/s2)

EESC Geodesy with the Global Positioning System. Class 7: Relative Positioning using Carrier-Beat Phase

Homework #2 Due Monday, April 18, 2012

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes

Nonparametric Bayesian Methods (Gaussian Processes)

Computational Data Analysis!

Methods for Solving Linear Systems Part 2

Kalman Filter Computer Vision (Kris Kitani) Carnegie Mellon University

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

EESC Geodesy with the Global Positioning System. Class 6: Point Positioning using Pseuduorange

DS-GA 1002 Lecture notes 10 November 23, Linear models

Intelligent Embedded Systems Uncertainty, Information and Learning Mechanisms (Part 1)

IV. Covariance Analysis

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics

Linear Regression (9/11/13)

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

Carrier Phase Integer Ambiguity Resolution Recent results and open issues

Advanced 3 G and 4 G Wireless Communication Prof. Aditya K Jagannathan Department of Electrical Engineering Indian Institute of Technology, Kanpur

GPS Multipath Detection Based on Sequence of Successive-Time Double-Differences

Autonomous Mobile Robot Design

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

DS-GA 1002 Lecture notes 12 Fall Linear regression

TSRT14: Sensor Fusion Lecture 8

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Outline. Math Numerical Analysis. Errors. Lecture Notes Linear Algebra: Part B. Joseph M. Mahaffy,

Autonomous Navigation for Flying Robots

GRACE processing at TU Graz

Lecture 2: Univariate Time Series

ECE521 week 3: 23/26 January 2017

CPE 310: Numerical Analysis for Engineers

Modern Methods of Data Analysis - WS 07/08

Regression. Oscar García

A NEW METHOD FOR ADAPTIVE OPTICS POINT SPREAD FUNCTION RECONSTRUCTION

Expectation-Maximization

1 Cricket chirps: an example

ELEC E7210: Communication Theory. Lecture 10: MIMO systems

Chapter 3 Numerical Methods

Computational Methods. Systems of Linear Equations

Modern Methods of Data Analysis - WS 07/08

Lecture 2: ARMA(p,q) models (part 2)

Principles of the Global Positioning System Lecture 18" Mathematical models in GPS" Mathematical models used in GPS"

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Lecture 23. Lidar Error and Sensitivity Analysis (2)

Linear Equations and Matrix

Stochastic Analogues to Deterministic Optimizers

L03. PROBABILITY REVIEW II COVARIANCE PROJECTION. NA568 Mobile Robotics: Methods & Algorithms

Lessons in Estimation Theory for Signal Processing, Communications, and Control

A. Barbu, J. Laurent-Varin, F. Perosanz, F. Mercier and J. Marty. AVENUE project. June, 20

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

OR MSc Maths Revision Course

Lecture 4: Linear Algebra 1

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

STATISTICAL ORBIT DETERMINATION

Linear Regression (1/1/17)

(Refer Slide Time: 0:36)

EESC Geodesy with the Global Positioning System. Class 4: The pseudorange and phase observables

7.6 The Inverse of a Square Matrix

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

CLASS NOTES Models, Algorithms and Data: Introduction to computing 2018

Estimating Atmospheric Water Vapor with Groundbased. Lecture 12

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression

Dynamic data processing recursive least-squares

Analytical Non-Linear Uncertainty Propagation: Theory And Implementation

Least Squares. Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Winter UCSD

Probability Methods in Civil Engineering Prof. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur

Power System Analysis Prof. A. K. Sinha Department of Electrical Engineering Indian Institute of Technology, Kharagpur. Lecture - 21 Power Flow VI

Likelihood Analysis of Gaussian Graphical Models

The Multivariate Gaussian Distribution [DRAFT]

* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course.

Matrix Assembly in FEA

Use of ground-based GNSS measurements in data assimilation. Reima Eresmaa Finnish Meteorological Institute

MAA507, Power method, QR-method and sparse matrix representation.

FACTOR ANALYSIS AS MATRIX DECOMPOSITION 1. INTRODUCTION

y b where U. matrix inverse A 1 ( L. 1 U 1. L 1 U 13 U 23 U 33 U 13 2 U 12 1

Systems of equation and matrices

Lecture 21: Spectral Learning for Graphical Models

COMP 558 lecture 18 Nov. 15, 2010

Announcements (repeat) Principal Components Analysis

BASIC NOTIONS. x + y = 1 3, 3x 5y + z = A + 3B,C + 2D, DC are not defined. A + C =

Y = ax + b. Numerical Applications Least-squares. Start with Self-test 10-1/459. Linear equation. Error function: E = D 2 = (Y - (ax+b)) 2

Decomposable and Directed Graphical Gaussian Models

Joint GPS and Vision Estimation Using an Adaptive Filter

4. DATA ASSIMILATION FUNDAMENTALS

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Linear Algebra. Solving Linear Systems. Copyright 2005, W.R. Winfrey

Transcription:

12.215 Modern Navigation Thomas Herring

Basic Statistics Summary of last class Statistical description and parameters Probability distributions Descriptions: expectations, variances, moments Covariances Estimates of statistical parameters Propagation of variances Methods for determining the statistical parameters of quantities derived other statistical variables 11/01/2006 12.215 Modern Naviation L13 2

Estimation methods Today s class Restrict to basically linear estimation problems (also non-linear problems that are nearly linear) Restrict to parametric, over determined estimation Concepts in estimation: Mathematical models Statistical models Least squares and Maximum likelihood estimation Covariance matrix of estimates parameters Covariance matrix of post-fit residual 11/01/2006 12.215 Modern Naviation L13 3

Concepts in estimation Given multiple observations of a quantity or related to a set of quantities how do you obtain a best estimate. What do we mean by best How do you quantify of quality of observations and the relationship between errors in observations. The complete answers to the above questions are complex We will limit our discussion to parametric estimation mostly for Gaussian distributed random errors in measurements. In parametric estimation, mathematical relationships between observations and parameters that can be used to model the observations is used (e.g., GPS measures pseudoranges to satellites: These observations can be related to the positions of the ground station and satellites plus other parameters that we discuss later). 11/01/2006 12.215 Modern Naviation L13 4

Basics of parametric estimation All parametric estimation methods can be broken into a few main steps: Observation equations: equations that relate the parameters to be estimated to the observed quantities (observables). Mathematical model. Example: Relationship between pseudorange, receiver position, satellite position (implicit in ρ), clocks, atmospheric and ionosphere delays Stochastic model: Statistical description that describes the random fluctuations in the measurements and maybe the parameters. In some forms the stochastic model is not explicit. Inversion that determines the parameters values from the mathematical model consistent with the statistical model. 11/01/2006 12.215 Modern Naviation L13 5

Observation model Observation model are equations relating observables to parameters of model: Observable = function (parameters) Observables should not appear on right-hand-side of equation The observed values are the observable plus noise of some stochastic nature Often function is non-linear and most common method is linearization of function using Taylor series expansion. Sometimes log linearization for f=a.b.c ie. Products fo parameters 11/01/2006 12.215 Modern Naviation L13 6

Taylor series expansion In most common Taylor series approach: y = f (x 1,x 2,x 3,x 4 ) y 0 +Δy = f (x) x 0 + f (x) x Δx x = (x 1,x 2, x 3,x 4 ) The estimation is made using the difference between the observations and the expected values based on apriori values for the parameters. The estimation returns adjustments to apriori parameter values The observations are y+noise 11/01/2006 12.215 Modern Naviation L13 7

Linearization Since the linearization is only an approximation, the estimation should be iterated until the adjustments to the parameter values are zero. For GPS estimation: Convergence rate is 100-1000:1 typically (ie., a 1 meter error in apriori coordinates could results in 1-10 mm of non-linearity error). To assess, the level on non-linear contribution, the Taylor series expansion is compared to the non-linear evaluation. If the differences are similar in size to the noise in the measurements, then a new Taylor series expansion, about the better estimates of the parameters, is needed. 11/01/2006 12.215 Modern Naviation L13 8

Estimation Most common estimation method is least-squares in which the parameter estimates are the values that minimize the sum of the squares of the differences between the observations and modeled values based on parameter estimates. For linear estimation problems, direct matrix formulation for solution For non-linear problems: Linearization or search technique where parameter space is searched for minimum value Care with search methods that local minimum is not found (will not treat in this course) 11/01/2006 12.215 Modern Naviation L13 9

Least squares estimation Originally formulated by Gauss. Basic equations: Δy is vector of observations; A is linear matrix relating parameters to observables; Δx is vector of parameters; v is residual Δy = AΔx + v minimize ( v T v); superscript T means transpose Δx = (A T A) 1 A T Δy 11/01/2006 12.215 Modern Naviation L13 10

Weighted Least Squares In standard least squares, nothing is assumed about the residuals v except that they are zero expectation. One often sees weight-least-squares in which a weight matrix is assigned to the residuals. Residuals with larger elements in W are given more weight. minimize ( v T Wv); Δx = (A T WA) 1 A T WΔy 11/01/2006 12.215 Modern Naviation L13 11

Statistical approach to least squares If the weight matrix used in weighted least squares is the inverse of the covariance matrix of the residuals, then weighted least squares is a maximum likelihood estimator for Gaussian distributed random errors. This choice maximizes the probability density (called a maximum likelihood estimate, MLE) f (v) = 1 (2π) n V e 1 2 vt V 1 v This latter form of least-squares is most statistically rigorous version. Sometimes weights are chosen empirically 11/01/2006 12.215 Modern Naviation L13 12

Data covariance matrix If we use the inverse of the covariance matrix of the noise in the data, we obtain a MLE if data noise is Gaussian distribution. How do you obtain data covariance matrix? Difficult question to answer completely For sextant measurements: Index error measurements Individual observers Issues to be considered for GPS specifically: Thermal noise in receiver gives on component Multipath could be treated as a noise-like quantity Signal-to-noise ratio of measurements allows an estimate of the noise (discussed later in course). In-complete mathematical model of observables can sometimes be treated as noise-like. Gain of GPS antenna will generate lower SNR at low elevation angles 11/01/2006 12.215 Modern Naviation L13 13

Data covariance matrix In practice in GPS (as well as many other fields), the data covariance matrix is somewhat arbitrarily chosen. Largest problem is temporal correlations in the measurements. Typical GPS data set size for 24-hours of data at 30 second sampling is 8x2880=23000 phase measurements. Since the inverse of the covariance matrix is required, fully accounting for correlations requires the inverse of 23000x23000 matrix. To store the matrix would require, 4Gbytes of memory Even if original covariance matrix is banded (ie., correlations over a time short compared to 24-hours), the inverse of banded matrix is usually a full matrix (However, remember LU decomposition in linear algebra lecture) 11/01/2006 12.215 Modern Naviation L13 14

Covariance matrix of parameter estimates Propagation of covariance can be applied to the weighted least squares problem: x ˆ = (A T V 1 yy A) 1 A T V 1 yy y < x ˆ x ˆ T >= (A T V 1 yy A) 1 A T V 1 yy < yy T > V 1 yy A(A T V 1 yy A) 1 Vˆ x x ˆ = (A T V 1 yy A) 1 Notice that the covariance matrix of parameter estimates is a natural output of the estimator if ATV -1 A is inverted (does not need to be) 11/01/2006 12.215 Modern Naviation L13 15

Covariance matrix of estimated parameters Notice that for the rigorous estimation, the inverse of the data covariance is needed (time consuming if non-diagonal) To compute to parameter estimate covariance, only the covariance matrix of the data is needed (not the inverse) In some cases, a non-rigorous inverse can be done with say a diagonal covariance matrix, but the parameter covariance matrix is rigorously computed using the full covariance matrix. This is a non-mle but the covariance matrix of the parameters should be correct (just not the best estimates that can found). This techniques could be used if storage of the full covariance matrix is possible, but inversion of the matrix is not because it would take too long or inverse can not be performed in place. 11/01/2006 12.215 Modern Naviation L13 16

Covariance matrix of post-fit residuals Post-fit residuals are the differences between the observations and the values computed from the estimated parameters Because some of the noise in the data are absorbed into the parameter estimates, in general, the post-fit residuals are not the same as the errors in the data. In some cases, they can be considerably smaller. The covariance matrix of the post-fit residuals can be computed using propagation of covariances. 11/01/2006 12.215 Modern Naviation L13 17

Covariance matrix of post-fit residuals This can be computed using propagation on covariances: e is the vector of true errors, and v is vector of residuals y = Ax + e x ˆ = (A T V 1 yy A) 1 A T V 1 yy y v = y Aˆ x = I A(A T V 1 yy A) 1 A T 1 V 1444 24443 yy e Eqn 1 Amount error reduced V vv =< vv T >= V yy A(A T V 1 yy A) 1 A T 11/01/2006 12.215 Modern Naviation L13 18

Post-fit residuals Notice that we can compute the compute the covariance matrix of the post-fit residuals (a large matrix in generate) Eqn 1 on previous slide gives an equation of the form v=be; why can we not compute the actual errors with e=b-1v? B is a singular matrix which has no unique inverse (there is in fact one inverse which would generate the true errors) Note: In this case, singularity does not mean that there is no inverse, it means there are an infinite number of inverses. 11/01/2006 12.215 Modern Naviation L13 19

Example Consider the case shown below: When a rate of change is estimated, the slope estimate will absorb error in the last data point particularly as Δt increases. (Try this case yourself) 6 5 Example of fitting slope to non-uniform data distribution Data 4 3 Δt 2 Postfit error bar very small; slope will always pass close 1 Postfit error bar to this data point somewhat reduced 0 0.0 10.0 20.0 30.0 40.0 50.0 Time 11/01/2006 12.215 Modern Naviation L13 20

Estimation methods Summary Restrict to basically linear estimation problems (also non-linear problems that are nearly linear) Restrict to parametric, over determined estimation Concepts in estimation: Mathematical models Statistical models Least squares and Maximum likelihood estimation Covariance matrix of estimated parameters Statistical properties of post-fit residuals 11/01/2006 12.215 Modern Naviation L13 21