Regularizing inverse problems. Damping and smoothing and choosing...
|
|
- Mabel O’Connor’
- 6 years ago
- Views:
Transcription
1 Regularizing inverse problems Damping and smoothing and choosing
2 Regularization The idea behind SVD is to limit the degree of freedom in the model and fit the data to an acceptable level. Retain only those features necessary to fit the data. A general framework for solving non-unique inverse problems is to introduce regularization. Regularization makes a non-unique problem become a unique problem. How does it do this? We want to minimize a combination of data misfit and some property of the model that measures extravagant behaviour, e.g. φ(m) =(d g(m)) T Cd 1 (d g(m))+μ(m m o) T Cm 1 (m m o ) Inverse problem nonlinear optimization problem 142
3 Solutions to inverse problems Optimal data fitting model Extremal model Data acceptable models The general framework is optimization φ(m) =(d g(m) T Cd 1 (d g(m)+μ(m m o) T CM 1 (m m o) 143
4 Damped least squares We seek the model that minimizes φ(d, m) =(d Gm) T Cd 1 (d Gm)+μ(m m o) T CM 1 (m m o) After some algebra we get m DLS =(G T C 1 D G+μC 1 M ) 1 (G T C 1 D d+μc 1 M m o) This is the damped least squares solution. A special case is to minimise a weighted sum of the data misfit and the model norm. min (d Gm) μ m 2 2 Normal equations become (G T G + μi)m = G T d this system of linear equations will have a unique solution which is called the Tikhonov solution 144
5 L-curve solution Choose μ that is nearest the elbow of the curve With this value of μ we can construct a particular Tikhonov solution (G T G + μi)m = G T d Perform repeat solutions of the normal equations for different μ and select the one which lies near the elbow of the trade-off curve Elbow is estimated visually -> potentially subjective if elbow is not clear See example 5.1 of Aster et al. (2005) 145
6 Tikhonov regularization Choose μ that minimizes the length of the model while specifying a maximum value of the prediction min m 2 2 s.t. d Gm 2 2 < δ Choice of μ depends on knowing the errors in the data Discrepancy principle gives us a value of δ Since each solution δ corresponds to a particular μ, we often refer to the Tikhonov solution as a function of μ on the curve min (d Gm) μ m 2 2 (G T G + μi)m = G T d See example 5.1 of Aster et al. (2005) 146
7 Tikhonov: Discrepancy principle Discrepancy principle gives us a value of δ and consequently a value for μ If the N data have Gaussian errors with known variance N(0,σ 2 ) then we would expect that on average that each residual (d i G I,j m j ) would be approximately σ. Hence we have d Gm 2 = Nσ 2 c.f.χ2 N (50%) δ = σ N We get an expected value of the norm of the prediction error from the number of data and the standard deviation of the data (G T G + μi)m = G T d Perform repeated solutions until d Gm = δ 147
8 Tikhonov Example: The Shaw problem Recall the discretized Shaw problem Data d(s) and model m(θ) at N equal angles s i = θ i = (i 0.5)π n π 2, (i =1, 2,...,n) d i = d(s i ) m j = m(θ j ) (i =1,...,n) (j =1,...,n) This gives a system of N N linear equations where d = Gm Ã! G i,j = s(cos(s i )+cos(θ j )) 2 sin(π(sin(si )+sin(θ j ))) 2 π(sin(s i )+sin(θ j )) s = π n We know this problem was ill-posed and previously solved with truncated SVD. 148
9 Example: Shaw problem with Tikhonov regularization Let s revisit the Shaw problem d = Gm N = M =20 s i Eigenvalue spectrum An severely ill-posed inverse problem where G has a condition number = Extreme sensitivity to noise i Resolution is sacrificed to achieve stability Input spike model Data from input spike Amplitude Model units Data units 149
10 Recall SVD applied to this problem (with noise) d = Gm m = V p S 1 p U T p d = px i=1 Ã ui d s i! v i Input spike model Data from spike model Amplitude Model units Data units Recovered model Add Gaussian noise to data σ =10 6 Presence of small eigenvalues means sensitivity of solution to noise 150
11 d = Gm SVD: Shaw problem with p=10 m = V p S 1 p U T p d = px i=1 Ã ui d s i! v i Input spike model Amplitude Model units use first 10 eigenvalues only No noise solution Noise solution Truncating Truncating eigenvalues eigenvalues reduces reduces sensitivity sensitivity to to noise noise but but also also resolving resolving power power of of the the data data 151
12 Tikhonov regularization: Shaw problem (G T G + μi)m = G T d L-curve for the Shaw problem Fortunately a nice clear elbow is visible in this log-log plot, Trade off parameter μ is easily picked. N[0, (10 6 ) 2 ] μl = δ = = μδ = Tikhonov solution for μ δ L-curve solution μ L m =
13 Example: Tikhonov regularization How do these choices for μ relate to the SVD truncation level chosen earlier? μl = δ = σ N δ = μδ = m = px i=1 Ã ui d s i! v i The eigenvalue from the truncation level in SVD is similar to the two choices of in the Tikhonov scheme. μ m = V p S 1 p U T p d = px i=1 Ã! ui d s i v i (G T G + μi)m = G T d 153
14 SVD and Tikhonov regularization Tkihonov solution G = U p S p V T p m =(G T G + μi) 1 G T d It can be shown... that this can be written as Look! MX s 2 Ã! Page of m = i ui d i=1 s 2 i + μ v Aster et al (2005) i s i Now we get very different behaviour m = Ã! px ui d v i i=1 s i SVD Let f i = s2 i s 2 i + μ Filter factors As singular values s i 0 the solution is not highly sensitive to noise Because f i 0 rather than. If s i >> μ then f i 1 If s i << μ then f i 0 SVD s i >s p f i =1 s i <s p f i =0 μ filters out (damps) the unstable influence of the small eigenvalues 154
15 Filter factors Filter factors control influence of singular values u i d s i s i u i d m = f i = c.f. Filter factors MX i=1 s2 i s 2 i + μ Ã! ui d f i v i s i f i =1 As singular values s i 0 the solution is not highly sensitive to noise Because f i 0 rather than. If s i >> μ then f i 1 If s i << μ then f i 0 μ filters out (damps) the unstable influence of the small eigenvalues 155
16 Model resolution for a Tikhonov solution A Tikhonov solution has the form m =(G T G + μi) 1 G T d = G g d Input spike model Model resolution matrix R M = G g G = VFV T F is the M x M diagonal matrix of filter factors m est = R M m true Each row of the resolution matrix can also be found using a spike test! m =(G T G + μi) 1 G T d m = MX i=1 s 2 i s 2 i + μ output model Ã! ui d s i v i This is the row which corresponds to the parameter in the spike model. Resolution tells us how the estimated model is a linear combination of the true model. Relationship to true solution 156
17 Model covariance a Tikhonov solution A Tikhonov solution has the form Input spike model m =(G T G + μi) 1 G T d = G g d Model covariance matrix C M = G g C 1 D (G g ) T Confidence intervals calculated from the diagonal of the covariance matrix Noise propagation into the solution μδ = Error bars too small to see! give unrealistically small errors As regularization strength increases the size of confidence interval decreases Confidence limits do not represent distance to the true solution -> Regularized solutions are ALWAYS BIASED! 157
18 Effect of data noise on regularized solutions Tikhonov m =(G T G + μi) 1 G T d Amplitude Input spike model Model units m = TSVD MX i=1 m = s 2 i s 2 i + μ px i=1 à ui d à ui d s i s i!! v i v i Because regularized solutions are ALWAYS BIASED! No noise solution Noise solution 158
19 Choosing μ with Generalized cross validation As we have seen the choice of trade-off parameter μ determines the level of fit to the data possible with a Tikhonov solution and vice versa δ = σ N What if I really don t have a good idea of how well I should fit my data? How can I choose δ or μ? Can I still get a robust solution? Is all lost? (L-curve is one answer) All is not lost. There is usually information in the variability of data itself about the likely level of fit needed. We can use the data itself to estimate the regularization parameter! This approach is called Cross Validation
20 Choosing μ by cross validation Cross validation is a way of assessing the quality of a solution based on the principle of measuring the fit to missing data The Leave one out principle Lets fit a curve (red) with one data point (green) left out A simple model A complex model Large hyper-parameter μ 1 Small hyper-parameter μ 1 In both cases (on average) the missing data tend to be fit poorly! 160
21 Choosing μ with Cross validation What does cross validation mean? For N data and a chosen value for μ, Drop one data value, d k and use your favourite algorithm to fit the remainder, d i (i=1,.k-1,k+1,..n) m k (μ) =(G T k G k + μi) 1 G T k d = G g k d Compare prediction of recovered model with missing datum r k (μ) =d k Gm k (μ) Repeat for each datum (k=1,...,n) and sum the squares of the residuals to produce a fitness for μ V (μ) = rk 2 k=1 Find the μ that minimizes of V(μ) NX 161
22 Leave some out A general framework: K-fold cross validation V (μ) is the cross validation measure for μ 162
23 Example: GCV in Curve fitting Problem: Find the curve that generated this noisy data. y Data are the y i values for known x i (i=1,,n) and each has Gaussian random noise N[0,0.25] What would you guess is the solution? 163
24 s(x, m) Example: Curve fitting Let represent the continuous solution curve as a function of x, and let its shape be determined by some parameters m. Lets first choose to fit the data with the smoothest possible curve i.e. minimize and also ψ(d, m) = J(s) = NX i=1 Z Ã 2! 2 s x 2 dx (d i s(x i, m)) 2 Duchon (1976) showed that its possible to fit the data d i exactly and obtain a unique curve with minimum J(s) using thin plate splines. s(x, m) =p(x)+ where NX i=1 m i φ(x x i ) p(x) is a polynomial in x φ(x x i ) is a Thin plate spline basis function The parameters m i can be solved for using matrix inversion 164
25 The smoothest thin plate spline The parameters m in thin plate spline solution are given by the solution of a set of linear equations Does this extremal solution look smooth? What has gone wrong? Answer: Data is over fit 165
26 Choosing the trade-off parameter with GCV We should not try to fit noisy data exactly but only to an acceptable level. To relax data fit we want to minimize φ(d, m) = NX i=1 (d i s(x i, m)) 2 + μj(s) How to choose the trade off parameter μ? Answer: there are many ways. If we do not know what an acceptable fit to the data is we can use Generalized Cross Validation. In this approach we use the data to find a value for μ NX Leave each datum out in turn and compute the solution V (μ) = i=1 (d i s i (x i, m)) 2 s i (x i, m) = thin plate spline interpolant with i-th datum missing Look how V(μ) behaves as a function of μ μ V (μ) μ 0 V (μ) The solution is very smooth The solution is very rough 166
27 Example: Curve fitting with GCV. We find μ that minimizes V(μ) by repeated solutions of the TPS problem We get a unique minimum in V(μ)! 167
28 Example: Curve fitting with CV. The TPS solution that corresponds to the minimum of μ This is achieved without knowing in advance what a suitable acceptable data fit was. The GCV is in effect a way of estimating the error in the data and using it in model fitting. 168
29 Regularization by GCV in general For a general linear inverse problem with the k-th datum left out a solution is obtained by minimizing This is second order Tikhonov regularization NX φ k (d, m) = ((Gm) i d i ) 2 + μm T L T Lm i6=k If is the solution model for given μwith k-th datum missing m k (μ) m k (μ) =(G T k G k + μl T L) 1 G T k d k = G g k d k then the cross validation function is the prediction error for all N solutions. V (μ) = 1 N NX k=1 (G k m k d k ) 2 We find μ that minimized V(μ). Evaluation of V for any μ would appear to require N solutions to the inverse problem but its possible to show that V (μ) N (d Gm(μ))T (d Gm(μ)) Tr(I GG g See Aster et al. (p108, 2005) μ ) 2 Which requires only a single optimization solution to evaluate for any μ 169
30 Generalized cross validation Note that calculation of a single value of the cross validation function V(¹ 1 ) requires N-repeat inversion solutions. Where we choose N. This might become expensive. Craven & Wahba (1979) devised Generalized cross validation, which removes the need to perform many repeat inversions. Hence its much more efficient and gives the same answer. Minimize φ k (d, m) = NX i6=k ((Gm) i d i ) 2 + μm T L T Lm CV function V (μ) = 1 N NX k=1 (G k m k d k ) 2 N solutions for each μ GCV function V (μ) N (d Gm(μ))T (d Gm(μ)) Tr(I GG g μ ) 2 1 solution for each μ 170
31 Example: Cross validation in Finite Fault Slip inversions Fault slip model 2004 Sumatra-Andaman earthquake Vertical Sea floor movement Chen Ji (Caltech) Australian Tsunami Warning System Model of 1833 event Courtesy Phil Cummins (Geoscience Australia) 171
32 How does Finite Fault Slip inversion work? Try to fit the data (body wave, surface wave seismograms, geodetic observations etc) Body wave seismograms Also try to seek the least complex model Least complex may mean Minimum length Flattest model Smoothest model Most compact slip distribution (Lohman & Simons, 2004) Fault slip model d = Gm Posed as a regularized linear inverse problem: Minimize a combination of fit to data and control on the model φ = {Data misfit} + r {Model Roughness} 172
33 Solving for a finite fault slip model φ = d Gm 2 + μ 2 1 L 1m 2 + μ 2 2 L om 2 For given values of the hyper-parameters we find the m model which minimizes φ. (μ 1, μ 2 ) Use non-negative least squares method (Lawson and Hansen 1972) Solves the linear system d 0 0 = G μ 1 L 1 μ 2 L 0 m A good stable, reliable approach d 0 = G 0 m in a least squares sense subject to positivity constraints on all variables m 0 Existence of inequality constraints means GCV not applicable 173
34 2006 Kuril earthquake Fault plane solution obtained with interactively chosen hyper-parameter 40 waveforms to fit 4677 data equations 1676 model parameters Toshikata Baba s solution Deep ocean Tsunami records Regularization: Minimum length + positivity constraints 174
35 Interactive selection of hyper-parameters is not efficient enough! Poor data fit simple solutions Baba s solution Good data fit complex solutions Time consuming, subjective and difficult 175
36 2006 Kuril earthquake with CV 176
37 2006 Kuril earthquake: comparison between CV and interactive Interactive solution (r 1 = 150) Cross validation solution (r 1 = 210) 40 waveforms to fit 4677 data equations 1676 model parameters Regularization: Minimum length + positivity constraints 177
38 2007 Solomons earthquake with CV A clear minimum again! Baba s Cross validation solution o fold waveforms to fit 2562 data equations 896 model parameters r 1 = 70, r 2 = 0 Regularization: Minimum length + positivity constraints 178
39 What about the computational cost of Cross validation? Ideally suited to parallelization Parallelization possible over data space and model space T. Baba parallelized this calculation over data space only and was able to reduce computation from 3 days to 2 hours. 179
40 Grid search is not the only option φ = d Gm 2 + μ 2 1 L 1m 2 + μ 2 2 L om 2 More than one hyper-parameter requires multi-dimensional direct search to minimize V(¹ 1, ¹ 2 ) V (μ 1, μ 2 ) Uniform grid search μ 2 μ 1 Adaptive search (NA, SA, GA) 180
41 Cross validation properties Cross validation is a way of assessing the quality of a solution based on the principle of measuring the fit to missing data On the plus side Cross validation can be used in cases where a suitable level of data misfit is not known in advance (it is in effect a bootstrap measure of data variance) Provides an objective criterion for picking a preferred solution. Can be automated. (GCV for linear problems with no inequality constraints) On the minus side Has been criticized for over-smoothing models. Computational burden has often seen it ignored (e.g. by geophysicists!) 181
ECE295, Data Assimila0on and Inverse Problems, Spring 2015
ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch
More informationSingular value decomposition. If only the first p singular values are nonzero we write. U T o U p =0
Singular value decomposition If only the first p singular values are nonzero we write G =[U p U o ] " Sp 0 0 0 # [V p V o ] T U p represents the first p columns of U U o represents the last N-p columns
More informationChoosing the Regularization Parameter
Choosing the Regularization Parameter At our disposal: several regularization methods, based on filtering of the SVD components. Often fairly straightforward to eyeball a good TSVD truncation parameter
More informationRegularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution
Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Rosemary Renaut, Jodi Mead Arizona State and Boise State September 2007 Renaut and Mead (ASU/Boise) Scalar
More informationLinearized inverse Problems
Linearized inverse Problems (Weakly nonlinear problems) Using a Taylor expansion then away we go... Linearized inverse problems Nonlinear inverse problem d obs,i = g i (m) Choose a reference model m o
More informationOption 1: Landmark Registration We can try to align specific points. The Registration Problem. Landmark registration. Time-Warping Functions
The Registration Problem Most analyzes only account for variation in amplitude. Frequently, observed data exhibit features that vary in time. Option 1: Landmark Registration We can try to align specific
More informationInversion of Phase Data for a Phase Velocity Map 101. Summary for CIDER12 Non-Seismologists
Inversion of Phase Data for a Phase Velocity Map 101 Summary for CIDER12 Non-Seismologists 1. Setting up a Linear System of Equations This is a quick-and-dirty, not-peer reviewed summary of how a dataset
More informationforms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms
Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.
More informationLinear Regression and Its Applications
Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start
More informationSparsity in system identification and data-driven control
1 / 40 Sparsity in system identification and data-driven control Ivan Markovsky This signal is not sparse in the "time domain" 2 / 40 But it is sparse in the "frequency domain" (it is weighted sum of six
More informationLinear Least-Squares Data Fitting
CHAPTER 6 Linear Least-Squares Data Fitting 61 Introduction Recall that in chapter 3 we were discussing linear systems of equations, written in shorthand in the form Ax = b In chapter 3, we just considered
More informationOne Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017
One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 1 One Picture and a Thousand Words Using Matrix Approximations Dianne P. O Leary
More informationModeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop
Modeling Data with Linear Combinations of Basis Functions Read Chapter 3 in the text by Bishop A Type of Supervised Learning Problem We want to model data (x 1, t 1 ),..., (x N, t N ), where x i is a vector
More information= (G T G) 1 G T d. m L2
The importance of the Vp/Vs ratio in determining the error propagation and the resolution in linear AVA inversion M. Aleardi, A. Mazzotti Earth Sciences Department, University of Pisa, Italy Introduction.
More informationDimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining
Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Combinations of features Given a data matrix X n p with p fairly large, it can
More informationInversion Methods for Determining. Tsunami Source Amplitudes from DART Buoy Data. Don Percival
Inversion Methods for Determining Tsunami Source Amplitudes from DART Buoy Data Don Percival Applied Physics Laboratory University of Washington, Seattle NOAA-sponsored collaborative effort overheads for
More informationUncertainty Quantification for Inverse Problems. November 7, 2011
Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance
More informationNewton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve
Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Rosemary Renaut, Jodi Mead Arizona State and Boise State Copper Mountain Conference on Iterative Methods
More informationDiscrete ill posed problems
Discrete ill posed problems Gérard MEURANT October, 2008 1 Introduction to ill posed problems 2 Tikhonov regularization 3 The L curve criterion 4 Generalized cross validation 5 Comparisons of methods Introduction
More informationParameter estimation: A new approach to weighting a priori information
c de Gruyter 27 J. Inv. Ill-Posed Problems 5 (27), 2 DOI.55 / JIP.27.xxx Parameter estimation: A new approach to weighting a priori information J.L. Mead Communicated by Abstract. We propose a new approach
More information5 Linear Algebra and Inverse Problem
5 Linear Algebra and Inverse Problem 5.1 Introduction Direct problem ( Forward problem) is to find field quantities satisfying Governing equations, Boundary conditions, Initial conditions. The direct problem
More informationMATH 590: Meshfree Methods
MATH 590: Meshfree Methods Chapter 34: Improving the Condition Number of the Interpolation Matrix Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu
More informationA Modern Look at Classical Multivariate Techniques
A Modern Look at Classical Multivariate Techniques Yoonkyung Lee Department of Statistics The Ohio State University March 16-20, 2015 The 13th School of Probability and Statistics CIMAT, Guanajuato, Mexico
More informationNumerical Methods of Approximation
Contents 31 Numerical Methods of Approximation 31.1 Polynomial Approximations 2 31.2 Numerical Integration 28 31.3 Numerical Differentiation 58 31.4 Nonlinear Equations 67 Learning outcomes In this Workbook
More informationTime Series. Anthony Davison. c
Series Anthony Davison c 2008 http://stat.epfl.ch Periodogram 76 Motivation............................................................ 77 Lutenizing hormone data..................................................
More informationApplication of Principal Component Analysis to TES data
Application of Principal Component Analysis to TES data Clive D Rodgers Clarendon Laboratory University of Oxford Madison, Wisconsin, 27th April 2006 1 My take on the PCA business 2/41 What is the best
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More informationMultiple realizations using standard inversion techniques a
Multiple realizations using standard inversion techniques a a Published in SEP report, 105, 67-78, (2000) Robert G Clapp 1 INTRODUCTION When solving a missing data problem, geophysicists and geostatisticians
More informationStatistically-Based Regularization Parameter Estimation for Large Scale Problems
Statistically-Based Regularization Parameter Estimation for Large Scale Problems Rosemary Renaut Joint work with Jodi Mead and Iveta Hnetynkova March 1, 2010 National Science Foundation: Division of Computational
More informationMathematical Methods in Engineering and Science Prof. Bhaskar Dasgupta Department of Mechanical Engineering Indian Institute of Technology, Kanpur
Mathematical Methods in Engineering and Science Prof. Bhaskar Dasgupta Department of Mechanical Engineering Indian Institute of Technology, Kanpur Module - I Solution of Linear Systems Lecture - 05 Ill-Conditioned
More informationLinear inversion methods and generalized cross-validation
Online supplement to Using generalized cross-validation to select parameters in inversions for regional caron fluxes Linear inversion methods and generalized cross-validation NIR Y. KRAKAUER AND TAPIO
More informationStanford Exploration Project, Report 105, September 5, 2000, pages 41 53
Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53 40 Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53 Short Note Multiple realizations using standard inversion
More informationCS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares
CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search
More informationUncertainty Quantification for Machine Learning and Statistical Models
Uncertainty Quantification for Machine Learning and Statistical Models David J. Stracuzzi Joint work with: Max Chen, Michael Darling, Stephen Dauphin, Matt Peterson, and Chris Young Sandia National Laboratories
More information15 Singular Value Decomposition
15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationDiscrete Ill Posed and Rank Deficient Problems. Alistair Boyle, Feb 2009, SYS5906: Directed Studies Inverse Problems 1
Discrete Ill Posed and Rank Deficient Problems Alistair Boyle, Feb 2009, SYS5906: Directed Studies Inverse Problems 1 Definitions Overview Inversion, SVD, Picard Condition, Rank Deficient, Ill-Posed Classical
More informationOutline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren
1 / 34 Metamodeling ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 1, 2015 2 / 34 1. preliminaries 1.1 motivation 1.2 ordinary least square 1.3 information
More informationISyE 691 Data mining and analytics
ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)
More informationStructural Damage Detection Using Time Windowing Technique from Measured Acceleration during Earthquake
Structural Damage Detection Using Time Windowing Technique from Measured Acceleration during Earthquake Seung Keun Park and Hae Sung Lee ABSTRACT This paper presents a system identification (SI) scheme
More informationA study on regularization parameter choice in Near-field Acoustical Holography
Acoustics 8 Paris A study on regularization parameter choice in Near-field Acoustical Holography J. Gomes a and P.C. Hansen b a Brüel & Kjær Sound and Vibration Measurement A/S, Skodsborgvej 37, DK-285
More informationNumerical Methods I Solving Nonlinear Equations
Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)
More informationSIO223A A Brief Introduction to Geophysical Modeling and Inverse Theory.
SIO223A A Brief Introduction to Geophysical Modeling and Inverse Theory. Forward modeling: data model space space ˆd = f(x, m) m =(m 1,m 2,..., m N ) x =(x 1,x 2,x 3,..., x km ) ˆd =(ˆd 1, ˆd 2, ˆd 3,...,
More informationSIGNAL AND IMAGE RESTORATION: SOLVING
1 / 55 SIGNAL AND IMAGE RESTORATION: SOLVING ILL-POSED INVERSE PROBLEMS - ESTIMATING PARAMETERS Rosemary Renaut http://math.asu.edu/ rosie CORNELL MAY 10, 2013 2 / 55 Outline Background Parameter Estimation
More informationIn the Name of God. Lectures 15&16: Radial Basis Function Networks
1 In the Name of God Lectures 15&16: Radial Basis Function Networks Some Historical Notes Learning is equivalent to finding a surface in a multidimensional space that provides a best fit to the training
More informationThe Application of Discrete Tikhonov Regularization Inverse Problem in Seismic Tomography
The Application of Discrete Tikhonov Regularization Inverse Problem in Seismic Tomography KAMBIZ TEIMOORNEGAD, NEDA POROOHAN 2, Geology Department Islamic Azad University, Lahijan Branch 2 Islamic Azad
More informationStatistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1
Week 2 Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Part I Other datatypes, preprocessing 2 / 1 Other datatypes Document data You might start with a collection of
More informationTHE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR
THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR 1. Definition Existence Theorem 1. Assume that A R m n. Then there exist orthogonal matrices U R m m V R n n, values σ 1 σ 2... σ p 0 with p = min{m, n},
More informationPart I. Other datatypes, preprocessing. Other datatypes. Other datatypes. Week 2 Based in part on slides from textbook, slides of Susan Holmes
Week 2 Based in part on slides from textbook, slides of Susan Holmes Part I Other datatypes, preprocessing October 3, 2012 1 / 1 2 / 1 Other datatypes Other datatypes Document data You might start with
More informationDetection theory. H 0 : x[n] = w[n]
Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal
More information[POLS 8500] Review of Linear Algebra, Probability and Information Theory
[POLS 8500] Review of Linear Algebra, Probability and Information Theory Professor Jason Anastasopoulos ljanastas@uga.edu January 12, 2017 For today... Basic linear algebra. Basic probability. Programming
More informationMetric-based classifiers. Nuno Vasconcelos UCSD
Metric-based classifiers Nuno Vasconcelos UCSD Statistical learning goal: given a function f. y f and a collection of eample data-points, learn what the function f. is. this is called training. two major
More informationMachine Learning (Spring 2012) Principal Component Analysis
1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in
More informationMSA220 Statistical Learning for Big Data
MSA220 Statistical Learning for Big Data Lecture 4 Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology More on Discriminant analysis More on Discriminant
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationSYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I
SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability
More informationA MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS
A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS SILVIA NOSCHESE AND LOTHAR REICHEL Abstract. Truncated singular value decomposition (TSVD) is a popular method for solving linear discrete ill-posed
More informationLeast Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e.
Least Squares with Eamples in Signal Processing Ivan Selesnick March 7, 3 NYU-Poly These notes address (approimate) solutions to linear equations by least squares We deal with the easy case wherein the
More informationINVERSE PROBEMS: AN INTRODUCTION
SEISMOLOGY Master Degree Programme in Physics - UNITS Physics of the Earth and of the Environment INVERSE PROBEMS: AN INTRODUCTION FABIO ROMANELLI Department of Mathematics & Geosciences University of
More information(a)
Chapter 8 Subspace Methods 8. Introduction Principal Component Analysis (PCA) is applied to the analysis of time series data. In this context we discuss measures of complexity and subspace methods for
More informationIrregular Solutions of an Ill-Posed Problem
Irregular Solutions of an Ill-Posed Problem Peter Linz University of California at Davis Richard L.C. Wang Clearsight Systems Inc. Abstract: Tikhonov regularization is a popular and effective method for
More informationIntroduction. Chapter One
Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and
More informationGetting Started with Communications Engineering
1 Linear algebra is the algebra of linear equations: the term linear being used in the same sense as in linear functions, such as: which is the equation of a straight line. y ax c (0.1) Of course, if we
More informationSlip distributions of the 1944 Tonankai and 1946 Nankai earthquakes including the horizontal movement effect on tsunami generation
Slip distributions of the 1944 Tonankai and 1946 Nankai earthquakes including the horizontal movement effect on tsunami generation Toshitaka Baba Research Program for Plate Dynamics, Institute for Frontier
More informationRegularization methods for large-scale, ill-posed, linear, discrete, inverse problems
Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Silvia Gazzola Dipartimento di Matematica - Università di Padova January 10, 2012 Seminario ex-studenti 2 Silvia Gazzola
More informationQueens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.
Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 2018 3 Lecture 3 3.1 General remarks March 4, 2018 This
More information4. DATA ASSIMILATION FUNDAMENTALS
4. DATA ASSIMILATION FUNDAMENTALS... [the atmosphere] "is a chaotic system in which errors introduced into the system can grow with time... As a consequence, data assimilation is a struggle between chaotic
More informationSpatial Process Estimates as Smoothers: A Review
Spatial Process Estimates as Smoothers: A Review Soutir Bandyopadhyay 1 Basic Model The observational model considered here has the form Y i = f(x i ) + ɛ i, for 1 i n. (1.1) where Y i is the observed
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V
More informationnonlinear simultaneous equations of type (1)
Module 5 : Solving Nonlinear Algebraic Equations Section 1 : Introduction 1 Introduction Consider set of nonlinear simultaneous equations of type -------(1) -------(2) where and represents a function vector.
More informationIntroduction to Supervised Learning. Performance Evaluation
Introduction to Supervised Learning Performance Evaluation Marcelo S. Lauretto Escola de Artes, Ciências e Humanidades, Universidade de São Paulo marcelolauretto@usp.br Lima - Peru Performance Evaluation
More informationREGULARIZATION PARAMETER SELECTION IN DISCRETE ILL POSED PROBLEMS THE USE OF THE U CURVE
Int. J. Appl. Math. Comput. Sci., 007, Vol. 17, No., 157 164 DOI: 10.478/v10006-007-0014-3 REGULARIZATION PARAMETER SELECTION IN DISCRETE ILL POSED PROBLEMS THE USE OF THE U CURVE DOROTA KRAWCZYK-STAŃDO,
More informationChapter 2 Interpolation
Chapter 2 Interpolation Experiments usually produce a discrete set of data points (x i, f i ) which represent the value of a function f (x) for a finite set of arguments {x 0...x n }. If additional data
More informationChapter 9. Non-Parametric Density Function Estimation
9-1 Density Estimation Version 1.2 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least
More informationLinear Inverse Problems
Linear Inverse Problems Ajinkya Kadu Utrecht University, The Netherlands February 26, 2018 Outline Introduction Least-squares Reconstruction Methods Examples Summary Introduction 2 What are inverse problems?
More informationLecture No. 1 Introduction to Method of Weighted Residuals. Solve the differential equation L (u) = p(x) in V where L is a differential operator
Lecture No. 1 Introduction to Method of Weighted Residuals Solve the differential equation L (u) = p(x) in V where L is a differential operator with boundary conditions S(u) = g(x) on Γ where S is a differential
More informationMachine Learning. Lecture 9: Learning Theory. Feng Li.
Machine Learning Lecture 9: Learning Theory Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2018 Why Learning Theory How can we tell
More information5.1 2D example 59 Figure 5.1: Parabolic velocity field in a straight two-dimensional pipe. Figure 5.2: Concentration on the input boundary of the pipe. The vertical axis corresponds to r 2 -coordinate,
More informationLinear Models 1. Isfahan University of Technology Fall Semester, 2014
Linear Models 1 Isfahan University of Technology Fall Semester, 2014 References: [1] G. A. F., Seber and A. J. Lee (2003). Linear Regression Analysis (2nd ed.). Hoboken, NJ: Wiley. [2] A. C. Rencher and
More informationThe Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation
The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut Collaborators: Jodi Mead and Iveta Hnetynkova DEPARTMENT OF MATHEMATICS
More informationCS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation
Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80
More informationNoise & Data Reduction
Noise & Data Reduction Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum Dimension Reduction 1 Remember: Central Limit
More informationAlgebra of Random Variables: Optimal Average and Optimal Scaling Minimising
Review: Optimal Average/Scaling is equivalent to Minimise χ Two 1-parameter models: Estimating < > : Scaling a pattern: Two equivalent methods: Algebra of Random Variables: Optimal Average and Optimal
More informationDeconvolution. Parameter Estimation in Linear Inverse Problems
Image Parameter Estimation in Linear Inverse Problems Chair for Computer Aided Medical Procedures & Augmented Reality Department of Computer Science, TUM November 10, 2006 Contents A naive approach......with
More informationM. Holschneider 1 A. Eicker 2 R. Schachtschneider 1 T. Mayer-Guerr 2 K. Ilk 2
SPP project TREGMAT: Tailored Gravity Field Models for Mass Distributions and Mass Transport Phenomena in the Earth System by by M. 1 A. Eicker 2 R. Schachtschneider 1 T. Mayer-Guerr 2 K. Ilk 2 1 University
More informationThe Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation
The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut DEPARTMENT OF MATHEMATICS AND STATISTICS Prague 2008 MATHEMATICS AND STATISTICS
More informationGeneralized Cross Validation
26 november 2009 Plan 1 Generals 2 What Regularization Parameter λ is Optimal? Examples Dening the Optimal λ 3 Generalized Cross-Validation (GCV) Convergence Result 4 Discussion Plan 1 Generals 2 What
More informationInverse Theory Methods in Experimental Physics
Inverse Theory Methods in Experimental Physics Edward Sternin PHYS 5P10: 2018-02-26 Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 1 / 29 1 Introduction Indirectly observed data A new research
More informationScalable kernel methods and their use in black-box optimization
with derivatives Scalable kernel methods and their use in black-box optimization David Eriksson Center for Applied Mathematics Cornell University dme65@cornell.edu November 9, 2018 1 2 3 4 1/37 with derivatives
More informationPartititioned Methods for Multifield Problems
Partititioned Methods for Multifield Problems Joachim Rang, 22.7.215 22.7.215 Joachim Rang Partititioned Methods for Multifield Problems Seite 1 Non-matching matches usually the fluid and the structure
More information11 More Regression; Newton s Method; ROC Curves
More Regression; Newton s Method; ROC Curves 59 11 More Regression; Newton s Method; ROC Curves LEAST-SQUARES POLYNOMIAL REGRESSION Replace each X i with feature vector e.g. (X i ) = [X 2 i1 X i1 X i2
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More information8 The SVD Applied to Signal and Image Deblurring
8 The SVD Applied to Signal and Image Deblurring We will discuss the restoration of one-dimensional signals and two-dimensional gray-scale images that have been contaminated by blur and noise. After an
More informationSome general observations.
Modeling and analyzing data from computer experiments. Some general observations. 1. For simplicity, I assume that all factors (inputs) x1, x2,, xd are quantitative. 2. Because the code always produces
More informationChapter 9. Non-Parametric Density Function Estimation
9-1 Density Estimation Version 1.1 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least
More informationKalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q
Kalman Filter Kalman Filter Predict: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Update: K = P k k 1 Hk T (H k P k k 1 Hk T + R) 1 x k k = x k k 1 + K(z k H k x k k 1 ) P k k =(I
More informationDegrees of Freedom in Regression Ensembles
Degrees of Freedom in Regression Ensembles Henry WJ Reeve Gavin Brown University of Manchester - School of Computer Science Kilburn Building, University of Manchester, Oxford Rd, Manchester M13 9PL Abstract.
More informationBasic Linear Inverse Method Theory - DRAFT NOTES
Basic Linear Inverse Method Theory - DRAFT NOTES Peter P. Jones 1 1 Centre for Compleity Science, University of Warwick, Coventry CV4 7AL, UK (Dated: 21st June 2012) BASIC LINEAR INVERSE METHOD THEORY
More informationWe consider the problem of finding a polynomial that interpolates a given set of values:
Chapter 5 Interpolation 5. Polynomial Interpolation We consider the problem of finding a polynomial that interpolates a given set of values: x x 0 x... x n y y 0 y... y n where the x i are all distinct.
More informationSTATISTICS 174: APPLIED STATISTICS TAKE-HOME FINAL EXAM POSTED ON WEBPAGE: 6:00 pm, DECEMBER 6, 2004 HAND IN BY: 6:00 pm, DECEMBER 7, 2004 This is a
STATISTICS 174: APPLIED STATISTICS TAKE-HOME FINAL EXAM POSTED ON WEBPAGE: 6:00 pm, DECEMBER 6, 2004 HAND IN BY: 6:00 pm, DECEMBER 7, 2004 This is a take-home exam. You are expected to work on it by yourself
More informationStephen Scott.
1 / 35 (Adapted from Ethem Alpaydin and Tom Mitchell) sscott@cse.unl.edu In Homework 1, you are (supposedly) 1 Choosing a data set 2 Extracting a test set of size > 30 3 Building a tree on the training
More information