MAT Inverse Problems, Part 2: Statistical Inversion

Similar documents
Marginal density. If the unknown is of the form x = (x 1, x 2 ) in which the target of investigation is x 1, a marginal posterior density

Resolving the White Noise Paradox in the Regularisation of Inverse Problems

Gibbs Sampler Componentwise sampling directly from the target density π(x), x R n. Define a transition kernel

Bayesian rules of probability as principles of logic [Cox] Notation: pr(x I) is the probability (or pdf) of x being true given information I

Inference and estimation in probabilistic time series models

Markov Chain Monte Carlo methods

Bayesian Linear Regression [DRAFT - In Progress]

Machine Learning 4771

An example of Bayesian reasoning Consider the one-dimensional deconvolution problem with various degrees of prior information.

ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters

Introduction to Bayesian methods in inverse problems

DRAGON ADVANCED TRAINING COURSE IN ATMOSPHERE REMOTE SENSING. Inversion basics. Erkki Kyrölä Finnish Meteorological Institute

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Markov chain Monte Carlo methods in atmospheric remote sensing

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Density Estimation: ML, MAP, Bayesian estimation

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

Bayesian Learning (II)

Approximate Bayesian Computation and Particle Filters

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation

Treatment and analysis of data Applied statistics Lecture 6: Bayesian estimation

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

Assessing Reliability Using Developmental and Operational Test Data

Markov Chain Monte Carlo (MCMC)

Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors

COS513 LECTURE 8 STATISTICAL CONCEPTS

GAUSSIAN PROCESS REGRESSION

The problem is to infer on the underlying probability distribution that gives rise to the data S.

Inverse problem and optimization

CSC 2541: Bayesian Methods for Machine Learning

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Frequentist-Bayesian Model Comparisons: A Simple Example

Markov chain Monte Carlo methods for visual tracking

Spatial Statistics with Image Analysis. Outline. A Statistical Approach. Johan Lindström 1. Lund October 6, 2016

PATTERN RECOGNITION AND MACHINE LEARNING

UNDERSTANDING DATA ASSIMILATION APPLICATIONS TO HIGH-LATITUDE IONOSPHERIC ELECTRODYNAMICS

An introduction to Bayesian inference and model comparison J. Daunizeau

Inverse Problems in the Bayesian Framework

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Bayesian Learning. Tobias Scheffer, Niels Landwehr

Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model. David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC

2 Two Random Variables

Inverse problems and uncertainty quantification in remote sensing

Bayesian inference J. Daunizeau

A Bayesian Model for Root Computation

CIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions

Bayesian Gaussian / Linear Models. Read Sections and 3.3 in the text by Bishop

Bayesian modelling. Hans-Peter Helfrich. University of Bonn. Theodor-Brinkmann-Graduate School

Variational Bayesian Logistic Regression

A short introduction to INLA and R-INLA

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Bayesian inference J. Daunizeau

Quantile POD for Hit-Miss Data

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

Bayesian reconstruction of free energy profiles from umbrella samples

Strong Lens Modeling (II): Statistical Methods

Lecture 4: Probabilistic Learning

An Introduction to Bayesian Linear Regression

Minimum Message Length Analysis of the Behrens Fisher Problem

Ensemble Data Assimilation and Uncertainty Quantification

Uncertainty quantification for inverse problems with a weak wave-equation constraint

Bayesian Methods in Positioning Applications

The Kalman Filter ImPr Talk

The University of Auckland Applied Mathematics Bayesian Methods for Inverse Problems : why and how Colin Fox Tiangang Cui, Mike O Sullivan (Auckland),

CS340 Winter 2010: HW3 Out Wed. 2nd February, due Friday 11th February

Bayesian PalaeoClimate Reconstruction from proxies:

Bayesian Interpretations of Regularization

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

Chapter 2 Signal Processing at Receivers: Detection Theory

Consider the joint probability, P(x,y), shown as the contours in the figure above. P(x) is given by the integral of P(x,y) over all values of y.

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces.

Recursive Estimation

Machine Learning, Midterm Exam: Spring 2009 SOLUTION

ME 597: AUTONOMOUS MOBILE ROBOTICS SECTION 2 PROBABILITY. Prof. Steven Waslander

Multiple Scenario Inversion of Reflection Seismic Prestack Data

Multivariate Bayesian Linear Regression MLAI Lecture 11

The Expectation-Maximization Algorithm

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

Bayesian approach to image reconstruction in photoacoustic tomography

Bayesian Updating with Continuous Priors Class 13, Jeremy Orloff and Jonathan Bloom

Bayesian Inference. 2 CS295-7 cfl Michael J. Black,

Semantics for Probabilistic Programming

Part 6: Multivariate Normal and Linear Models

Efficient MCMC Sampling for Hierarchical Bayesian Inverse Problems

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

CSC2515 Assignment #2

Robotics. Lecture 4: Probabilistic Robotics. See course website for up to date information.

Stochastic Spectral Approaches to Bayesian Inference

MCMC Sampling for Bayesian Inference using L1-type Priors

Inverse Theory. COST WaVaCS Winterschool Venice, February Stefan Buehler Luleå University of Technology Kiruna

Introduction to Statistical Learning Theory

wissen leben WWU Münster

Retrieval problems in Remote Sensing

Bayesian Inference (A Rough Guide)

The Bayesian Brain. Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester. May 11, 2017

Bayesian Inference for the Multivariate Normal

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008

Uncertainty quantification for Wavefield Reconstruction Inversion

Transcription:

MAT-62006 Inverse Problems, Part 2: Statistical Inversion S. Pursiainen Department of Mathematics, Tampere University of Technology Spring 2015

Overview The statistical inversion approach is based on the following principles: The inverse problem is presented as a statistical quest for information. We have directly observable quantities and others that cannot be observed. In inverse problems, some of the unobservable quantities are of primary interest. These quantities depend on each other through forward and inverse models.

Overview The statistical inversion approach is based on the following principles: 1. All variables included in the model are modelled as random variables. 2. The randomness describes our degree of information concerning their realizations. 3. The degree of information concerning these values is coded in the probability distributions. 4. The solution of the inverse problem is the posterior probability distribution (Bayesian inverse model).

Overview 1. Forming the posterior density often is not enough to enlighten possible realizations of the unknown, since the shape of the posterior distribution is often likely to be somewhat tricky. 2. Factors that can make exploration of the posterior difficult include, among other things, a complicated forward model, a high number dimensions, existence of several local maxima, peaked structure, and/or weak prior information of the unknown. 3. Due to these difficulties numerical optimization and sampling methods are typically needed to find a reconstruction of the unkown based on the posterior.

Assume that we are measuring a quantity y R m in order to get information about another quantity x R n. In order to relate these two quantities, we need a data prediction (forward) model of their dependence, e.g. of the form y = f (x, ε), where f : R n R k R m is the model function and ε R k is a noise vector. In statistical inverse problems, all parameters are viewed as random variables denoted here by upper case letters (e.g. X and Y ). Realizations of the random variables are denoted by lower case letters (e.g. x and y). Thus, the model y = f (x, ε) would lead to a relation Y = f (X, E).

Before performing the measurement of Y there often is some prior (a priori) information, i.e. a foresight, of the unknown variable X. Bayesian approach relies on the assumption that the prior information can be coded into a prior probability density x π pr (x). Joint probability density π(x, y) can be written in the following two ways: π(x, y) = π pr (x)π(y x) and π(x, y) = π(y)π(x y). These together imply the classical Bayes formula π(x y) = π pr(x)π(y x). π(y)

In Bayes formula, π(x y) is known as the posterior probability density of x given the measurement y, which is in this context a solution of an inverse problem. The likelihood function of measuring y with a certain realization of x is given by π(y x), and strongly related to the forward model. Density π(y) can be treated as a constant, when y is given, and therefore, Bayes formula is often expressed in the form π(x y) π pr (x)π(y x). It is possible (e.g. due to nullspace effects) that the likelihood function cannot be scaled so that π(y x) dy = 1, i.e. it is not always a well-defined probability density.

Example 1 (Matlab) Assume that the goal is to detect an electric charge located at the point x = (r, φ) = (0.5, 0.5) (polar coordinates) inside an unit disk D using n equally spaced sensors lying on the boundary of D with one of the sensors placed on positive x-axis. Sensor i experiences the electric potential y i = 1/d i + n i, in which d i is the distance to the charge and n i is a zero mean Gaussian noise random variable with standard deviation σ. From n i = y i 1/d i it follows that π(n i ) = π(y i x) = exp[ (y i 1/d i ) 2 /(2σ 2 )]. The measurements are mutually independed implying that the total likelihood function is a product of individual likelihoods: π(y 1, y 2,..., y n x) π(y 1 x)π(y 2 x) π(y n x) ( exp 1 n 2σ 2 (y i 1/d i ) 2). i=1

Example 1 (Matlab) continued The prior π(x) is, in this case, chosen to be flat (constant) in D and zero outside D. Consequently, the posterior π(x y) is given by the likelihood, i.e. π(x y) π(x)π(y x) π(y x). Plot the posterior with n = 3, 4, 5 and σ = 0.1, 0.2 using Matlab. Use percent scale with respect to the maximum value of the posterior density (MAP). Notice that the posterior is the more focused the more there are sensors and the lower is the standard deviation of the noise.

Example 1 (Matlab) continued Solution 3 sensors, σ = 0.1 4 sensors, σ = 0.1 5 sensors, σ = 0.1 3 sensors, σ = 0.2 4 sensors, σ = 0.2 5 sensors, σ = 0.2 Particle and sensor locations are indicated by the purple and red circles, respectively.