Inverse Theory Methods in Experimental Physics

Similar documents
A study on regularization parameter choice in Near-field Acoustical Holography

REGULARIZATION PARAMETER SELECTION IN DISCRETE ILL POSED PROBLEMS THE USE OF THE U CURVE

IMPROVING THE ACCURACY OF NMR RELAXATION DISTRIBUTION ANALYSIS IN CLAY- RICH RESERVOIRS AND CORE SAMPLES

Spectral Regularization

A novel method for estimating the distribution of convective heat flux in ducts: Gaussian Filtered Singular Value Decomposition

Regularization via Spectral Filtering

Linear Inverse Problems

arxiv: v1 [nucl-th] 1 Dec 2016

Introduction to Bayesian methods in inverse problems

Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology

Reproducing Kernel Hilbert Spaces

Regularization of the Inverse Laplace Transform with Applications in Nuclear Magnetic Resonance Relaxometry Candidacy Exam

Inversion of Phase Data for a Phase Velocity Map 101. Summary for CIDER12 Non-Seismologists

Constructing Approximation Kernels for Non-Harmonic Fourier Data

Analysis of Error Produced by Truncated SVD and Tikhonov Regularization Met hods *

Dynamics of Quantum Many-Body Systems from Monte Carlo simulations

Resolving the White Noise Paradox in the Regularisation of Inverse Problems

Reproducing Kernel Hilbert Spaces

Singular Value Decomposition

SIGNAL AND IMAGE RESTORATION: SOLVING

The Finite Difference Method

Reproducing Kernel Hilbert Spaces

Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution

Inverse problems in statistics

Regularizing inverse problems. Damping and smoothing and choosing...

The Application of Discrete Tikhonov Regularization Inverse Problem in Seismic Tomography

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

Treatment of Error in Experimental Measurements

Multi-Linear Mappings, SVD, HOSVD, and the Numerical Solution of Ill-Conditioned Tensor Least Squares Problems

2 Tikhonov Regularization and ERM

Empirical Risk Minimization as Parameter Choice Rule for General Linear Regularization Methods

Singular value decomposition. If only the first p singular values are nonzero we write. U T o U p =0

Inversion, history matching, clustering and linear algebra

Two-parameter regularization method for determining the heat source

Statistical and Computational Inverse Problems with Applications Part 2: Introduction to inverse problems and example applications

Estimation of Cole-Cole parameters from time-domain electromagnetic data Laurens Beran and Douglas Oldenburg, University of British Columbia.

Initial inverse problem in heat equation with Bessel operator

Choosing the Regularization Parameter

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems

Introduction to Initial Value Problems

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.

NMR Data workup using NUTS

A new method for multi-exponential inversion of NMR relaxation measurements

Chaos and Liapunov exponents

State Space Representation of Gaussian Processes

Chapter 1. Forward and Inverse Problem in Geophysics

Stochastic Spectral Approaches to Bayesian Inference

Quasipatterns in surface wave experiments

Non-negatively constrained least squares and parameter choice by the residual periodogram for the inversion of electrochemical impedance spectroscopy

Inverse problems in statistics

REVIEW: Waves on a String

A Note on the Particle Filter with Posterior Gaussian Resampling

Optimization-based sparse recovery: Compressed sensing vs. super-resolution

Least Squares Approximation

MATH 590: Meshfree Methods

APMA 2811Q. Homework #1. Due: 9/25/13. 1 exp ( f (x) 2) dx, I[f] =

Slow symmetric exchange

Linear Theory of Evolution to an Unstable State

Use of the magnetic field generated by the internal distribution of injected currents for Electrical Impedance Tomography (MR-EIT)

Course on Inverse Problems

Course on Inverse Problems Albert Tarantola

Name: INSERT YOUR NAME HERE. Due to dropbox by 6pm PDT, Wednesday, December 14, 2011

Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics

Estimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition

= (G T G) 1 G T d. m L2

Recovering point sources in unknown environment with differential data

MATH 590: Meshfree Methods

GEOPHYSICAL INVERSE THEORY AND REGULARIZATION PROBLEMS

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR

arxiv: v1 [physics.data-an] 10 Sep 2007

Electrostatic Imaging via Conformal Mapping. R. Kress. joint work with I. Akduman, Istanbul and H. Haddar, Paris

NONLINEAR DIFFUSION PDES

Introduction to numerical schemes

3 Chemical exchange and the McConnell Equations

What is Image Deblurring?

PHYSICS 653 HOMEWORK 1 P.1. Problems 1: Random walks (additional problems)

In the Name of God. Lectures 15&16: Radial Basis Function Networks

Sparsity in system identification and data-driven control

1 Introduction. or equivalently f(z) =

Natural Evolution Strategies for Direct Search

Vollständige Inversion seismischer Wellenfelder - Erderkundung im oberflächennahen Bereich

Discretization-invariant Bayesian inversion and Besov space priors. Samuli Siltanen RICAM Tampere University of Technology

Renormalization of Tensor Network States

Robustness of Principal Components

Bindel, Fall 2009 Matrix Computations (CS 6210) Week 8: Friday, Oct 17

Reproducing Kernel Hilbert Spaces

IDENTIFICATION AND ANALYSIS OF TIME-VARYING MODAL PARAMETERS

CHAPTER 10: Numerical Methods for DAEs

12 The Heat equation in one spatial dimension: Simple explicit method and Stability analysis

Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm

Singular Value Decomposition

An Inverse Mass Expansion for Entanglement Entropy. Free Massive Scalar Field Theory

PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN

Traveltime sensitivity kernels: Banana-doughnuts or just plain bananas? a

Achieving depth resolution with gradient array survey data through transient electromagnetic inversion

The Lorentz Integral Transform (LIT) method and its connections with other approaches

Parameter estimation: A new approach to weighting a priori information

Lecture 1: The Equilibrium Green Function Method

Unfolding Methods in Particle Physics

Structural Aspects of Numerical Loop Calculus

Transcription:

Inverse Theory Methods in Experimental Physics Edward Sternin PHYS 5P10: 2018-02-26 Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 1 / 29

1 Introduction Indirectly observed data A new research paradigm 2 Ill-posed problems Definitions: ill-posed, misfit, pseudo-inverse Uniqueness Stability 3 Regularization Tikhonov regularization A discrete formulation of the FIE problem Example: a distribution of exponential relaxation rates Discretization SVD truncation Tikhonov regularization The L-curve 4 SciLab code example 5 Homework Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 2 / 29

Introduction Indirectly observed data Framework: indirectly observed data Inverting the Fredholm Integral Equation of the 1st kind (FEI): S(ω) = g(x) K(ω, x) dx (1) S(ω) what you measure experimentally (e.g. an NMR spectrum) g(x) what you want to know (e.g. chemical shifts) K(ω, x) a kernel function is, in general, an ill-posed problem. Only a few choice kernels allow a complete inverse calculation: S(ω) g(x), e.g. a Fourier transform. Most other kernels require numerical solutions. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 3 / 29

Introduction A new research paradigm A new research paradigm An unexpected benefit: a new research paradigm 1, where data analysis is an essential part of the experiment itself. The model of the traditional paradigm, where it is kept scrupulously separate from the experiment, becomes a part of it, subject to computer-based tweaking and straining. This yields estimates of the parameters, quantitative analysis of the misfit of the model to the experimental data, and, ultimately, better models. 1 J.V.Beck, Perspective on the Relation of Current Engineering Practice to Inverse Problems, Michigan State University, 1998. http://www.me.ua.edu/research/inverse-problems/perspective.html. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 4 / 29

Introduction A new research paradigm A new research paradigm The language of this new paradigm is the language of the inverse problems, and its range of applicability is vast: medical, industrial, and geophysical imaging, using X-ray or magnetic resonance or electrical impedance measurements; analysis of EEG traces in neuroscience and experimental psychology; inverse heat problems in the casting of steel; scattering problems in physics and geological exploration; extraction of material properties from indirect bulk measurements. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 5 / 29

Ill-posed problems Definitions: ill-posed, misfit, pseudo-inverse Ill-posedness and what to do with it Existence: a solution g(x) may not exist at all Then: ask not for the true g(x), but for a reasonable approximation g(x). Here, reasonable is in the least-squares sense, a minimum of misfit Ψ = S(ω) g(x)k(ω, x)dx This misfit norm, the distance between the measured S(ω) and the approximation calculated as an integral over g(x), is sometimes called the least-squares error norm. A minimum of misfit ensures compatibility of the fit with the measured data. 2 Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 6 / 29

Ill-posed problems Uniqueness Ill-posedness and what to do with it Uniqueness: many g(x) may satisfy Eq.1 equally well Then: bring in additional physical input. Of all compatible solutions, choose those that satisfy those additional constraints. E.g. a reasonable physical function is smooth, so minimize g (x) 2. Each problem is unique and may need a different optimum constraint. Test using numerical modeling (e.g. Monte Carlo). Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 7 / 29

Ill-posed problems Stability Ill-posedness and what to do with it Stability: a small perturbation in S(ω) (e.g. experimental noise) may cause a large jump in g(x) Then: regularize the problem, enforcing local stability of the inverse S(ω) g(x) mapping Formally, lack of stability means that two adjacent forward mappings S 1, S 2 do not have adjacent origins g 1, g 2, i.e. that one cannot find two numbers η and ε(η) so that from g 2 g 1 η follows S 2 S 1 ε(η), independently of the choice of g 1 and g 2. Lack of stability is a distinguishing feature of all ill-posed inverse problems. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 8 / 29

Ill-posed problems Stability Regularization Discretization: just choosing the right set of {x i } for which to ask about g(x i ) may do the trick. Often done implicitly, but there are mathematical implications. cf. Nyquist limit, noise folding, and zero filling in FT NMR Truncation of SVD: expand the transformation matrix in its singular values, and then truncate the expansion, ignoring the lower values which are responsible for most instabilities. cf. Taylor series expansion of functions Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 9 / 29

Tikhonov regularization Regularization Tikhonov regularization: bring in additional constraints, i.e. minimize Ψ{g(x)} = T {g(x)} = S(ω) 2 g(x)k(ω, x)dx + λt {g(x)} g 2 (Tikhonov) g 2 (Phillips) g log g dx (Shannon, maximum entropy) where λ controls the balance between compatibility with the data, and the regularizing effects of T {g} Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 10 / 29

A discrete formulation of the FIE problem Regularization by discretization In the ω-direction, a natural discretization of a FIE occurs when a discrete set of data sampling points ω i is considered. Discretization in the x-direction is realized when, for example, g is approximated with a sum of δ-functions, i g i δ(x x i ), on a pre-selected grid {x i } with the unknown coefficients g i, or when g(x) is assumed to be a linear function within each of the pre-selected intervals [g i, g i+1 ]. This is often done implicitly in the standard analysis of the experimental data, but strictly speaking, such discretizations impose strong restrictions on a possible solution g. This alone may sufficiently stabilize the inverse problem to overcome its ill-posedness. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 11 / 29

A discrete formulation of the FIE problem A discrete formulation of the FIE problem inverse K g = S j i i = j Ranges of i, j define the problem as even-, over-, or under-determined. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 12 / 29

A discrete formulation of the FIE problem Underdetermined?! For an experimentalist, it is somewhat startling to think of a fit to more parameters than there are measured data points, but this simply underscores the fact that regularized solutions of inverse problems are fundamentally different from least-squares fits. The discrete inverse problem is equivalent to calculating the inverse matrix, K 1 such that g = K 1 f. If such an inverse does not exist (a majority of cases), the problem becomes the calculation of the pseudo-inverse. The right kind of numerical strategy is important, and much depends on the experience and judgment of the person performing the calculation: successful inverse problem solving is strongly dependent on the analyst 2 2 J. C. Santamarina and D. Fratta. Discrete signals and inverse problems. An introduction for engineers and scientists. John Wiley & Sons, 2005. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 13 / 29

Example: a distribution of exponential relaxation rates Example: multi-exponential decay f(t) = rmax r min g(r)e rt dr where t represents time and r has the physical meaning of a relaxation rate. The data f(t) is a decay curve in the linearly-sampled time domain, and the desired inverse solution g(r) is a distribution of relaxation rates. Our example: a broad asymmetric distribution of relaxation rates simulated in the logarithmically-sampled range of r min = 10 3 to r max = 10 0, with random normally-distributed noise added at 1% of the maximum f(t) intensity. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 14 / 29

Example: a distribution of exponential relaxation rates Example: multi-exponential decay Two simulated relaxation-rate distribution functions, g a (r) and g b (r) (smooth lines in the insert), are used to generate two time-domain signals, each with 1% random noise is added. The resulting f a (t) and f b (t) are very similar. For now, ignore the points in the insert. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 15 / 29

Discretization Discretization strategy Select enough points in the grid of parameter values to allow for the entire physically relevant range to be covered with sufficient resolution to reproduce all of the essential features in the data; and no more. Acquire enough data points, sufficiently spread out in the observation domain to resolve contributions from different parameter values. Over-determined problems, by at least a factor of 2, are easier to solve. Higher parameter grid densities require better signal-to-noise ratio in the data. Use numerical simulation and testing (trial-and-error) to establish a good discretization scheme for a given problem. Near the optimum, the results of the calculation should be largely independent of the exact choice of the discretization grids. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 16 / 29

Discretization Discretization example Regularizing effects of discretization for a distribution of relaxation rates. Level of discretization is varied by changing the total number of points across the parameter space. At just the right level of regularization (middle plot), the re-calculated g is a faithful representation of the true g. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 17 / 29

SVD truncation SVD truncation strategy Many discrete ill-posed problems exhibit a gradual decrease in the size of their singular values. As the kernel matrix is expanded in its singular values, the lower-valued ones tend to magnify the effects of noise in the measured data. Regularize by truncating the SVD expansion! This has to be done in the region free of rapid changes in the size of the singular values. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 18 / 29

SVD truncation SVD truncation example Overly smooth pseudo-inverse solution (top) fails to reproduce some of the features of the true g, and insufficient regularization produces an unstable pseudo-inverse (bottom). The optimum SVD truncation (middle) faithfully reproduces the true g. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 19 / 29

Tikhonov regularization Tikhonov regularization Applying Tikhonov regularization to a noisy experimental dataset Ψ{g} = f(t) g(r)k(r, t)dr 2 + λ g(r) 2 crucially depends on the right choice for λ: if λ value is too small, a stable solution will not be found; if λ is too large, essential physical features in the solution will be obscured. This form of the regularization term gives preference to small g(x): in the discrete case, those with as few as possible non-zero values in {g i }. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 20 / 29

Tikhonov regularization Tikhonov regularization Overly small pseudo-inverse solution (top) fails to reproduce some of the features of the true g, and insufficient regularization produces an unstable pseudo-inverse (bottom). The optimum choice of λ (middle) faithfully reproduces the true g. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 21 / 29

The L-curve Optimum regularization: the L-curve L-curve is a plot of the norm of the regularized solution g versus the corresponding misfit norm f K g, as the regularization is varied. Upper left: loss of stability in g; lower right: loss of compatibility with the data; the corner corresponds to the optimum regularization. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 22 / 29

The L-curve Optimum regularization: the L-curve L-curve used in selecting optimum grid density in r L-curve used in selecting optimum SVD truncation Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 23 / 29

The L-curve Example: multi-exponential decay Two simulated relaxation-rate distribution functions, g a (r) and g b (r) are successfully recovered from the two, essentially indistinguishable, simulated noisy datasets f a (t) and f b (t). This time, do focus on the points in the insert! The results of the inversion using Tikhonov regularization, of f a (t) and f b (t), reproduce the true exponential rate distributions faithfully. Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 24 / 29

SciLab code example Skeleton code function [g] = regularize(t,f,r,k,svd_n,lambda) m=length(f); n=length(r); if m < n then error("for overdetermined problems only, need more data"); end; [U,S,V]=svd(K); // SVD of the kernel matrix nt=n; if svd_cnt > 0 then // if requested, truncate singular values nt=max(min(n,svd_cnt),2); // less than n, but not too few! end; sl=s(nt,nt); // Tikhonov regularization for k=1:nt sl(k,k)=s(k,k)/(s(k,k)^2+lambda); end; g=v(1:n,1:nt)*sl*u(1:m,1:nt) *f; // return g(r) endfunction; Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 25 / 29

SciLab code example Skeleton code, cont d /////////////////////////////////////////////////////////// // multi-exponential inverse analysis of time decay curves // f(t) = \int g(r) exp(-r*t) dr /////////////////////////////////////////////////////////// // r = vector or r values // n = number of points in r // g = vector of unknowns, g(r) // t = vector of t values // m = number of points in t // f = vector of measured data, f(t) // K = (m x n) kernel matrix // svd_cnt = keep only this many singular values // lambda = Tikhonov regularization parameter // datafile = (noisy) data, in two columns: (t,f) /////////////////////////////////////////////////////////// Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 26 / 29

SciLab code example Skeleton code, cont d datafile= test.dat ; r_steps=30; // r-grid has this many points, r_min=1e-3; // from this minimum, r_max=1e0; // to this maximum svd_cnt=11; // set to 0 for no SVD truncation lambda=2e-4; // set to 0 for no Tikhonov regularization // read in the time-domain data fd=mopen(datafile); [n,t,f]=mfscanf(-1,fd,"%f %f"); mclose(fd); // create a vector of r values, logarithmically spaced r_inc = (log(r_max)-log(r_min)) / (r_steps-1); r = exp([log(r_min):r_inc:log(r_max)]); // set up our kernel matrix, normalize by the step in r K = exp(-t*r) * r_inc; Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 27 / 29

SciLab code example Skeleton code, cont d // call the inversion routine [g] = regularize(t,f,r,k,svd_cnt,lambda); // plot the original data f(t) and our misfit scf(1); clf; plot(t,f, - ); misfit = f - K*g ; Psi = sum(misfit.^2)/(length(f)-1); scale=0.2*max(f)/max(misfit); plot(t,scale*misfit, -r ); legend( input data, f(t), misfit, x +string(scale)); xtitle( LS error norm = +string(psi), t, f(t) ); // separately, plot the result of the inversion scf(2); clf; plot(log10(r),g, :o ); xtitle( inverse solution, log r, g(r) ); Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 28 / 29

Homework Homework Review the concept of SVD decomposition. Analyze the skeleton code of regularize(); make sure you understand every line. Modify the main code by adding appropriate loops, etc. to reproduce the results presented for the exponential example, including the L-curves, on the sample data provided in /work/5p10/test.dat Automate the optimum selection of parameter λ. One possible approach is to seek the value that corresponds to the shortest distance to the origin on the L-curve. Be efficient: vary the step size in λ depending on how strong the dependence on λ is. Optionally, use your program to analyze a real experimental data set (to be provided). Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 29 / 29