ECE295, Data Assimila0on and Inverse Problems, Spring 2015

Similar documents
ECE295, Data Assimila0on and Inverse Problems, Spring 2015

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

ECE295, Data Assimilation and Inverse Problems, Spring 2015

ECE285/SIO209, Machine learning for physical applications, Spring 2017

Miscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier

STA 4273H: Sta-s-cal Machine Learning

Approximate Inference

Mathematical Tools for Neuroscience (NEU 314) Princeton University, Spring 2016 Jonathan Pillow. Homework 8: Logistic Regression & Information Theory

Parallel Particle Filter in Julia

GAUSSIAN PROCESS REGRESSION

Machine Learning for OR & FE

Why do we care? Measurements. Handling uncertainty over time: predicting, estimating, recognizing, learning. Dealing with time

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

Forward Problems and their Inverse Solutions

UVA CS 6316/4501 Fall 2016 Machine Learning. Lecture 6: Linear Regression Model with RegularizaEons. Dr. Yanjun Qi. University of Virginia

STA414/2104 Statistical Methods for Machine Learning II

2D Image Processing. Bayes filter implementation: Kalman filter

Intro to Linear & Nonlinear Optimization

Par$cle Filters Part I: Theory. Peter Jan van Leeuwen Data- Assimila$on Research Centre DARC University of Reading

The Inversion Problem: solving parameters inversion and assimilation problems

Announcements. CS 188: Artificial Intelligence Fall VPI Example. VPI Properties. Reasoning over Time. Markov Models. Lecture 19: HMMs 11/4/2008

CSC2515 Assignment #2

Signal Modeling, Statistical Inference and Data Mining in Astrophysics

CS491/691: Introduction to Aerial Robotics

Physics 2BL: Experiments in Mechanics and Electricity Summer Session I, 2012

Announcements. CS 188: Artificial Intelligence Fall Markov Models. Example: Markov Chain. Mini-Forward Algorithm. Example

MAT 211, Spring 2015, Introduction to Linear Algebra.

Homework 4 will be posted tonight, due Wednesday, May 8.

MITOCW watch?v=dztkqqy2jn4

6. Wiener and Kalman Filters

Nonparametric Bayesian Methods (Gaussian Processes)

SIO223A A Brief Introduction to Geophysical Modeling and Inverse Theory.

Hidden Markov Models. Vibhav Gogate The University of Texas at Dallas

Gaussian Process Approximations of Stochastic Differential Equations

Hidden Markov Models. Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 19 Apr 2012

= v = 2πr. = mv2 r. = v2 r. F g. a c. F c. Text: Chapter 12 Chapter 13. Chapter 13. Think and Explain: Think and Solve:

Tools for Parameter Estimation and Propagation of Uncertainty

Lecture 9. Time series prediction

Markov Models. CS 188: Artificial Intelligence Fall Example. Mini-Forward Algorithm. Stationary Distributions.

CS 188: Artificial Intelligence Spring 2009

Probability in Programming. Prakash Panangaden School of Computer Science McGill University

Fitting a Straight Line to Data

Why do we care? Examples. Bayes Rule. What room am I in? Handling uncertainty over time: predicting, estimating, recognizing, learning

Gaussian Quiz. Preamble to The Humble Gaussian Distribution. David MacKay 1

Gaussian Filtering Strategies for Nonlinear Systems

2D Image Processing. Bayes filter implementation: Kalman filter

Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model. David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC

Introduction to Machine Learning

18.06 Problem Set 8 Due Wednesday, 23 April 2008 at 4 pm in

When one of things that you don t know is the number of things that you don t know

Intro to Linear & Nonlinear Optimization

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi

The Ensemble Kalman Filter:

Fundamental Probability and Statistics

Lecture Note 2: Estimation and Information Theory

At-A-Glance SIAM Events Mobile App

This is Solving Linear Systems, chapter 3 from the book Advanced Algebra (index.html) (v. 1.0).

Newton s Law of Motion

Statistical Methods in Applied Computer Science

Learning Energy-Based Models of High-Dimensional Data

HOW TO WRITE PROOFS. Dr. Min Ru, University of Houston

For final project discussion every afternoon Mark and I will be available

ADAPTIVE FILTER THEORY

Introduction to Stochastic Processes

Linear Models for Regression

Stoichiometry. Cartoon courtesy of NearingZero.net

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Organization. I MCMC discussion. I project talks. I Lecture.

At A Glance. UQ16 Mobile App.

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

The Kalman Filter ImPr Talk

Probabilistic Graphical Models

Expectation Propagation in Dynamical Systems

Detection and Estimation Theory

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

Stochastic Analogues to Deterministic Optimizers

Physics 1140 Fall 2013 Introduction to Experimental Physics

Monday, October 24, Trigonometry, Period 3

OPTIMAL ESTIMATION of DYNAMIC SYSTEMS

Bayes Nets: Sampling

Learning outcomes: You will learn:

CSEP 573: Artificial Intelligence

An introduction to plotting data

Bayesian Concept Learning

4. DATA ASSIMILATION FUNDAMENTALS

Results: MCMC Dancers, q=10, n=500

Math 350: An exploration of HMMs through doodles.

Learning from Data: Regression

Linear Models for Regression CS534

CSE 473: Artificial Intelligence

State Space Models for Wind Forecast Correction

STA 4273H: Statistical Machine Learning

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

Recent Advances in Bayesian Inference Techniques

Data Assimilation Research Testbed Tutorial

(Extended) Kalman Filter

Convex envelopes, cardinality constrained optimization and LASSO. An application in supervised learning: support vector machines (SVMs)

Autonomous Mobile Robot Design

Transcription:

ECE295, Data Assimila0on and Inverse Problems, Spring 2015 Peter Gersto), 534-7768, gersto)@ucsd.edu We meet Wednesday from 5 to 6:20pm in HHS 2305A Text for first 5 classes: Parameter EsGmaGon and Inverse Problems (2nd EdiGon) here under UCSD license Grading S Classes 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1, 2) 8 April, SVD (Aster ch 2 and 3) 15 April, RegularizaGon (ch 4) Numerical Example: Beamforming 22 April, Sparse methods (ch 7.2-7.3) 29 April, Sparse methods 6 May, Bayesian methods and Monte Carlo methods (ch 11) Numerical Example: Ice flow from GPS 13 May, IntroducGon to sequengal Bayesian methods, Kalman Filter 20 May, Data assimilagon, EnKF 27 May, EnKF, Data assimilagon 3 June, Markov Chain Monte Carlo, PF Homework: You can use any programming language, matlab is the most obvious, Just email the code to me (I dont need anything else). Call the files LastName_Ex. Homework is due 8am on Wednesday. That way we can discuss in class. Hw 1: Download the matlab codes for the book (cd_5.3). Run the 3 examples for chapter 2. Come to class with one quesgon about the examples. Due 8 April.

Inverse problems are everywhere When data only indirectly constrain quantities of interest 2

Thinking backwards Most people, if you describe a train of events to them will tell you what the result will be. There are few people, however that if you told them a result, would be able to evolve from their own inner consciousness what the steps were that led to that result. This power is what I mean when I talk of reasoning backward. Sherlock Holmes, A Study in Scarlet, Sir Arthur Conan Doyle (1887) 3

Reversing a forward problem 4

Anatomy of an inverse problem Hunting for gold at the beach with a gravimeter Gravimeter? Courtesy Heiner Igel 5

Forward modelling example: Treasure Hunt We have observed some values: 10, 23, 35, 45, 56 µ gals How can we relate the observed gravity values to the subsurface properties? Gravimeter? We know how to do the forward problem: ( ') ( r ) Gρ r Φ = dv ' r r' This equation relates the (observed) gravitational potential to the subsurface density. -> given a density model we can predict the gravity field at the surface! 6

Treasure Hunt: Trial and error What else do we know? Density sand: 2.2 g/cm 3 Density gold: 19.3 g/cm 3 Do we know these values exactly? Where is the box with gold? Gravimeter? One approach is trial and (t)error forward modelling Use the forward solution to calculate many models for a rectangular box situated somewhere in the ground and compare the theoretical (synthetic) data to the observations. 7

Treasure Hunt: model space But...... we have to define plausible models for the beach. We have to somehow describe the model geometrically. We introduce simplifying approximations Gravimeter? - divide the subsurface into rectangles with variable density - Let us assume a flat surface x x x x x surface gold sand 8

Treasure Hunt: Non- uniqueness Could we compute all possible models and compare the synthetic data with the observations? - at every rectangle two possibilities (sand or gold) - 2 50 ~ 10 15 possible models Gravimeter (Age of universe ~10 17 s) Too many models! - We have 10 15 possible models but only 5 observations! - It is likely that two or more models will fit the data (maybe exactly) Non-uniqueness is likely 9

Treasure hunt: a priori informagon Is there anything we know about the treasure? How large is the box? Is it still intact? Has it possibly disintegrated? What was the shape of the box? Gravimeter This is called a priori (or prior) information. It will allow us to define plausible, possible, and unlikely models: plausible possible unlikely 10

Treasure hunt: data uncertainges Things to consider in formulating the inverse problem Do we have errors in the data? Did the instruments work correctly? Do we have to correct for anything? (e.g. topography, tides,...) Gravimeter Are we using the right theory? Is a 2-D approximation adequate? Are there other materials present other than gold and sand? Are there adjacent masses which could influence observations? Answering these questions often requires introducing more simplifying assumptions and guesses. All inferences are dependent on these assumptions. (GIGO) 11

Treasure Hunt: solugons Models with less than 2% error. 12

Treasure Hunt: solugons Models with less than 1% error. 13

What we have learned from one example Inverse problems = inference about physical systems from data Gravimeter - Data usually contain errors (data uncertainties) - Physical theories require approximations - Infinitely many models will fit the data (non-uniqueness) - Our physical theory may be inaccurate (theoretical uncertainties) - Our forward problem may be highly nonlinear - We always have a finite amount of data Detailed questions are: How accurate are our data? How well can we solve the forward problem? What independent information do we have on the model space (a priori information)? 14

EsGmaGon and Appraisal 15

Ocean circulagon SequenGal inverse problems Observe temperature, current flow, acousgc travelgmes PredicGct ocean circulagon Numerical Weather PredicGon EvoluGon of an earthquake EvoluGon of a oilfield EvoluGon of a stock market Structural monitoring Tracking an airplane

ICE flow 40 RS07 RS08 Lat (meters) 30 20 10 0-10 2 4 6 8 10 12 14 2 4 6 8 10 12 14 2 4 6 8 10 12 14 Lon (meters) 15 10 5 0-5 -10 Elevation (meters) -15 60 40 20 0

Data assimilagon Data assimila0on is the process by which observagons are incorporated into a computer model of a real system. ApplicaGons of data assimilagon arise in many fields of geosciences, perhaps most importantly in weather forecasgng and hydrology. Data assimilagon proceeds by analysis cycles. In each analysis cycle, observagons of the current (and possibly past) state of a system are combined with the results from a numerical model (the forecast) to produce an analysis, which is considered as 'the best' esgmate of the current state of the system. This is called the analysis step. EssenGally, the analysis step tries to balance the uncertainty in the data and in the forecast. The model is then advanced in Gme and its result becomes the forecast in the next analysis cycle.

FormulaGng inverse problems Regression What are d, m and G? Discrete or continuous? Linear or nonlinear? Why? What are the data? What are the model parameters? Unique or non-unique solution? y = a + bx 21

FormulaGng inverse problems Ballistic trajectory What are d, m and G? Discrete or continuous? Linear or nonlinear? Why? What are the data? What are the model parameters? Unique or non-unique solution? 22

Recap: Characterizing inverse problems Inverse problems can be congnuous or discrete ConGnuous problems are o)en discregzed by choosing a set of basis funcgons and projecgng the congnuous funcgon on them. The forward problem is to take a model and predict observables that are compared to actual data. Contains the Physics of the problem. This o)en involves a mathemagcal model which is an approximagon to the real physics. The inverse problem is to take the data and constrain the model in some way. We may want to build a model or we may wish to ask a less precise quesgon of the data!

Over-determined: Linear discrete inverse problem To find the best fit model we can minimize the prediction error of the solution But the data contain errors. Let s assume these are independent and normally distributed, then we weight each residual inversely by the standard deviation of the corresponding (known) error distribution. We can obtain a least squares solution by minimizing the weighted prediction error of the solution. 4

Over-determined: Linear discrete inverse problem We seek the model vector m which minimizes Compare with maximum likelihood Note that this is a quadratic function of the model vector. Solution: Differentiate with respect to m and solve for the model vector which gives a zero gradient in This gives This is the least-squares solution. A solution to the normal equations: 3 5

Over-determined: Linear discrete inverse problem How does the Least-squares solution compare to the standard equations of linear regression? Given N data y i with independent normally distributed errors and standard deviations i what are the expressions for the model parameters m = [a,b] T? 6

Linear discrete inverse problem: Least squares What happens in the under and even-determined cases? Under-determined, N=1: Matrix has a zero determinant What and a is zero m? eigenvalue an infinite number of solutions exist Even-determined, N=2: r = d Gm = 0 What is the prediction error? Prediction error is zero! 7

Example: Over-determined, Linear discrete inverse problem The Ballistics example Given data and noise Calculate G Is the data fit good enough? And how to errors in data propagate into the solution? 8

The two questions in parameter estimation We have our fitted model parameters but we are far from finished! We need to: Assess the quality of the data fit. Goodness of fit: Does the model fit the data to within the statistical uncertainty of the noise? Estimate how errors in the data propagate into the model What are the errors on the model parameters? 9

Goodness of fit Once we have our least squares solution m LS how do we know whether the fit is good enough given the errors in the data? Use the prediction error at the least squares solution! If data errors are Gaussian this as a chi-square statistic 10