Regularizing inverse problems. Damping and smoothing and choosing...

Size: px
Start display at page:

Download "Regularizing inverse problems. Damping and smoothing and choosing..."

Transcription

1 Regularizing inverse problems Damping and smoothing and choosing

2 Regularization The idea behind SVD is to limit the degree of freedom in the model and fit the data to an acceptable level. Retain only those features necessary to fit the data. A general framework for solving non-unique inverse problems is to introduce regularization. Regularization makes a non-unique problem become a unique problem. How does it do this? We want to minimize a combination of data misfit and some property of the model that measures extravagant behaviour, e.g. φ(m) =(d g(m)) T Cd 1 (d g(m))+μ(m m o) T Cm 1 (m m o ) Inverse problem nonlinear optimization problem 142

3 Solutions to inverse problems Optimal data fitting model Extremal model Data acceptable models The general framework is optimization φ(m) =(d g(m) T Cd 1 (d g(m)+μ(m m o) T CM 1 (m m o) 143

4 Damped least squares We seek the model that minimizes φ(d, m) =(d Gm) T Cd 1 (d Gm)+μ(m m o) T CM 1 (m m o) After some algebra we get m DLS =(G T C 1 D G+μC 1 M ) 1 (G T C 1 D d+μc 1 M m o) This is the damped least squares solution. A special case is to minimise a weighted sum of the data misfit and the model norm. min (d Gm) μ m 2 2 Normal equations become (G T G + μi)m = G T d this system of linear equations will have a unique solution which is called the Tikhonov solution 144

5 L-curve solution Choose μ that is nearest the elbow of the curve With this value of μ we can construct a particular Tikhonov solution (G T G + μi)m = G T d Perform repeat solutions of the normal equations for different μ and select the one which lies near the elbow of the trade-off curve Elbow is estimated visually -> potentially subjective if elbow is not clear See example 5.1 of Aster et al. (2005) 145

6 Tikhonov regularization Choose μ that minimizes the length of the model while specifying a maximum value of the prediction min m 2 2 s.t. d Gm 2 2 < δ Choice of μ depends on knowing the errors in the data Discrepancy principle gives us a value of δ Since each solution δ corresponds to a particular μ, we often refer to the Tikhonov solution as a function of μ on the curve min (d Gm) μ m 2 2 (G T G + μi)m = G T d See example 5.1 of Aster et al. (2005) 146

7 Tikhonov: Discrepancy principle Discrepancy principle gives us a value of δ and consequently a value for μ If the N data have Gaussian errors with known variance N(0,σ 2 ) then we would expect that on average that each residual (d i G I,j m j ) would be approximately σ. Hence we have d Gm 2 = Nσ 2 c.f.χ2 N (50%) δ = σ N We get an expected value of the norm of the prediction error from the number of data and the standard deviation of the data (G T G + μi)m = G T d Perform repeated solutions until d Gm = δ 147

8 Tikhonov Example: The Shaw problem Recall the discretized Shaw problem Data d(s) and model m(θ) at N equal angles s i = θ i = (i 0.5)π n π 2, (i =1, 2,...,n) d i = d(s i ) m j = m(θ j ) (i =1,...,n) (j =1,...,n) This gives a system of N N linear equations where d = Gm Ã! G i,j = s(cos(s i )+cos(θ j )) 2 sin(π(sin(si )+sin(θ j ))) 2 π(sin(s i )+sin(θ j )) s = π n We know this problem was ill-posed and previously solved with truncated SVD. 148

9 Example: Shaw problem with Tikhonov regularization Let s revisit the Shaw problem d = Gm N = M =20 s i Eigenvalue spectrum An severely ill-posed inverse problem where G has a condition number = Extreme sensitivity to noise i Resolution is sacrificed to achieve stability Input spike model Data from input spike Amplitude Model units Data units 149

10 Recall SVD applied to this problem (with noise) d = Gm m = V p S 1 p U T p d = px i=1 Ã ui d s i! v i Input spike model Data from spike model Amplitude Model units Data units Recovered model Add Gaussian noise to data σ =10 6 Presence of small eigenvalues means sensitivity of solution to noise 150

11 d = Gm SVD: Shaw problem with p=10 m = V p S 1 p U T p d = px i=1 Ã ui d s i! v i Input spike model Amplitude Model units use first 10 eigenvalues only No noise solution Noise solution Truncating Truncating eigenvalues eigenvalues reduces reduces sensitivity sensitivity to to noise noise but but also also resolving resolving power power of of the the data data 151

12 Tikhonov regularization: Shaw problem (G T G + μi)m = G T d L-curve for the Shaw problem Fortunately a nice clear elbow is visible in this log-log plot, Trade off parameter μ is easily picked. N[0, (10 6 ) 2 ] μl = δ = = μδ = Tikhonov solution for μ δ L-curve solution μ L m =

13 Example: Tikhonov regularization How do these choices for μ relate to the SVD truncation level chosen earlier? μl = δ = σ N δ = μδ = m = px i=1 Ã ui d s i! v i The eigenvalue from the truncation level in SVD is similar to the two choices of in the Tikhonov scheme. μ m = V p S 1 p U T p d = px i=1 Ã! ui d s i v i (G T G + μi)m = G T d 153

14 SVD and Tikhonov regularization Tkihonov solution G = U p S p V T p m =(G T G + μi) 1 G T d It can be shown... that this can be written as Look! MX s 2 Ã! Page of m = i ui d i=1 s 2 i + μ v Aster et al (2005) i s i Now we get very different behaviour m = Ã! px ui d v i i=1 s i SVD Let f i = s2 i s 2 i + μ Filter factors As singular values s i 0 the solution is not highly sensitive to noise Because f i 0 rather than. If s i >> μ then f i 1 If s i << μ then f i 0 SVD s i >s p f i =1 s i <s p f i =0 μ filters out (damps) the unstable influence of the small eigenvalues 154

15 Filter factors Filter factors control influence of singular values u i d s i s i u i d m = f i = c.f. Filter factors MX i=1 s2 i s 2 i + μ Ã! ui d f i v i s i f i =1 As singular values s i 0 the solution is not highly sensitive to noise Because f i 0 rather than. If s i >> μ then f i 1 If s i << μ then f i 0 μ filters out (damps) the unstable influence of the small eigenvalues 155

16 Model resolution for a Tikhonov solution A Tikhonov solution has the form m =(G T G + μi) 1 G T d = G g d Input spike model Model resolution matrix R M = G g G = VFV T F is the M x M diagonal matrix of filter factors m est = R M m true Each row of the resolution matrix can also be found using a spike test! m =(G T G + μi) 1 G T d m = MX i=1 s 2 i s 2 i + μ output model Ã! ui d s i v i This is the row which corresponds to the parameter in the spike model. Resolution tells us how the estimated model is a linear combination of the true model. Relationship to true solution 156

17 Model covariance a Tikhonov solution A Tikhonov solution has the form Input spike model m =(G T G + μi) 1 G T d = G g d Model covariance matrix C M = G g C 1 D (G g ) T Confidence intervals calculated from the diagonal of the covariance matrix Noise propagation into the solution μδ = Error bars too small to see! give unrealistically small errors As regularization strength increases the size of confidence interval decreases Confidence limits do not represent distance to the true solution -> Regularized solutions are ALWAYS BIASED! 157

18 Effect of data noise on regularized solutions Tikhonov m =(G T G + μi) 1 G T d Amplitude Input spike model Model units m = TSVD MX i=1 m = s 2 i s 2 i + μ px i=1 à ui d à ui d s i s i!! v i v i Because regularized solutions are ALWAYS BIASED! No noise solution Noise solution 158

19 Choosing μ with Generalized cross validation As we have seen the choice of trade-off parameter μ determines the level of fit to the data possible with a Tikhonov solution and vice versa δ = σ N What if I really don t have a good idea of how well I should fit my data? How can I choose δ or μ? Can I still get a robust solution? Is all lost? (L-curve is one answer) All is not lost. There is usually information in the variability of data itself about the likely level of fit needed. We can use the data itself to estimate the regularization parameter! This approach is called Cross Validation

20 Choosing μ by cross validation Cross validation is a way of assessing the quality of a solution based on the principle of measuring the fit to missing data The Leave one out principle Lets fit a curve (red) with one data point (green) left out A simple model A complex model Large hyper-parameter μ 1 Small hyper-parameter μ 1 In both cases (on average) the missing data tend to be fit poorly! 160

21 Choosing μ with Cross validation What does cross validation mean? For N data and a chosen value for μ, Drop one data value, d k and use your favourite algorithm to fit the remainder, d i (i=1,.k-1,k+1,..n) m k (μ) =(G T k G k + μi) 1 G T k d = G g k d Compare prediction of recovered model with missing datum r k (μ) =d k Gm k (μ) Repeat for each datum (k=1,...,n) and sum the squares of the residuals to produce a fitness for μ V (μ) = rk 2 k=1 Find the μ that minimizes of V(μ) NX 161

22 Leave some out A general framework: K-fold cross validation V (μ) is the cross validation measure for μ 162

23 Example: GCV in Curve fitting Problem: Find the curve that generated this noisy data. y Data are the y i values for known x i (i=1,,n) and each has Gaussian random noise N[0,0.25] What would you guess is the solution? 163

24 s(x, m) Example: Curve fitting Let represent the continuous solution curve as a function of x, and let its shape be determined by some parameters m. Lets first choose to fit the data with the smoothest possible curve i.e. minimize and also ψ(d, m) = J(s) = NX i=1 Z Ã 2! 2 s x 2 dx (d i s(x i, m)) 2 Duchon (1976) showed that its possible to fit the data d i exactly and obtain a unique curve with minimum J(s) using thin plate splines. s(x, m) =p(x)+ where NX i=1 m i φ(x x i ) p(x) is a polynomial in x φ(x x i ) is a Thin plate spline basis function The parameters m i can be solved for using matrix inversion 164

25 The smoothest thin plate spline The parameters m in thin plate spline solution are given by the solution of a set of linear equations Does this extremal solution look smooth? What has gone wrong? Answer: Data is over fit 165

26 Choosing the trade-off parameter with GCV We should not try to fit noisy data exactly but only to an acceptable level. To relax data fit we want to minimize φ(d, m) = NX i=1 (d i s(x i, m)) 2 + μj(s) How to choose the trade off parameter μ? Answer: there are many ways. If we do not know what an acceptable fit to the data is we can use Generalized Cross Validation. In this approach we use the data to find a value for μ NX Leave each datum out in turn and compute the solution V (μ) = i=1 (d i s i (x i, m)) 2 s i (x i, m) = thin plate spline interpolant with i-th datum missing Look how V(μ) behaves as a function of μ μ V (μ) μ 0 V (μ) The solution is very smooth The solution is very rough 166

27 Example: Curve fitting with GCV. We find μ that minimizes V(μ) by repeated solutions of the TPS problem We get a unique minimum in V(μ)! 167

28 Example: Curve fitting with CV. The TPS solution that corresponds to the minimum of μ This is achieved without knowing in advance what a suitable acceptable data fit was. The GCV is in effect a way of estimating the error in the data and using it in model fitting. 168

29 Regularization by GCV in general For a general linear inverse problem with the k-th datum left out a solution is obtained by minimizing This is second order Tikhonov regularization NX φ k (d, m) = ((Gm) i d i ) 2 + μm T L T Lm i6=k If is the solution model for given μwith k-th datum missing m k (μ) m k (μ) =(G T k G k + μl T L) 1 G T k d k = G g k d k then the cross validation function is the prediction error for all N solutions. V (μ) = 1 N NX k=1 (G k m k d k ) 2 We find μ that minimized V(μ). Evaluation of V for any μ would appear to require N solutions to the inverse problem but its possible to show that V (μ) N (d Gm(μ))T (d Gm(μ)) Tr(I GG g See Aster et al. (p108, 2005) μ ) 2 Which requires only a single optimization solution to evaluate for any μ 169

30 Generalized cross validation Note that calculation of a single value of the cross validation function V(¹ 1 ) requires N-repeat inversion solutions. Where we choose N. This might become expensive. Craven & Wahba (1979) devised Generalized cross validation, which removes the need to perform many repeat inversions. Hence its much more efficient and gives the same answer. Minimize φ k (d, m) = NX i6=k ((Gm) i d i ) 2 + μm T L T Lm CV function V (μ) = 1 N NX k=1 (G k m k d k ) 2 N solutions for each μ GCV function V (μ) N (d Gm(μ))T (d Gm(μ)) Tr(I GG g μ ) 2 1 solution for each μ 170

31 Example: Cross validation in Finite Fault Slip inversions Fault slip model 2004 Sumatra-Andaman earthquake Vertical Sea floor movement Chen Ji (Caltech) Australian Tsunami Warning System Model of 1833 event Courtesy Phil Cummins (Geoscience Australia) 171

32 How does Finite Fault Slip inversion work? Try to fit the data (body wave, surface wave seismograms, geodetic observations etc) Body wave seismograms Also try to seek the least complex model Least complex may mean Minimum length Flattest model Smoothest model Most compact slip distribution (Lohman & Simons, 2004) Fault slip model d = Gm Posed as a regularized linear inverse problem: Minimize a combination of fit to data and control on the model φ = {Data misfit} + r {Model Roughness} 172

33 Solving for a finite fault slip model φ = d Gm 2 + μ 2 1 L 1m 2 + μ 2 2 L om 2 For given values of the hyper-parameters we find the m model which minimizes φ. (μ 1, μ 2 ) Use non-negative least squares method (Lawson and Hansen 1972) Solves the linear system d 0 0 = G μ 1 L 1 μ 2 L 0 m A good stable, reliable approach d 0 = G 0 m in a least squares sense subject to positivity constraints on all variables m 0 Existence of inequality constraints means GCV not applicable 173

34 2006 Kuril earthquake Fault plane solution obtained with interactively chosen hyper-parameter 40 waveforms to fit 4677 data equations 1676 model parameters Toshikata Baba s solution Deep ocean Tsunami records Regularization: Minimum length + positivity constraints 174

35 Interactive selection of hyper-parameters is not efficient enough! Poor data fit simple solutions Baba s solution Good data fit complex solutions Time consuming, subjective and difficult 175

36 2006 Kuril earthquake with CV 176

37 2006 Kuril earthquake: comparison between CV and interactive Interactive solution (r 1 = 150) Cross validation solution (r 1 = 210) 40 waveforms to fit 4677 data equations 1676 model parameters Regularization: Minimum length + positivity constraints 177

38 2007 Solomons earthquake with CV A clear minimum again! Baba s Cross validation solution o fold waveforms to fit 2562 data equations 896 model parameters r 1 = 70, r 2 = 0 Regularization: Minimum length + positivity constraints 178

39 What about the computational cost of Cross validation? Ideally suited to parallelization Parallelization possible over data space and model space T. Baba parallelized this calculation over data space only and was able to reduce computation from 3 days to 2 hours. 179

40 Grid search is not the only option φ = d Gm 2 + μ 2 1 L 1m 2 + μ 2 2 L om 2 More than one hyper-parameter requires multi-dimensional direct search to minimize V(¹ 1, ¹ 2 ) V (μ 1, μ 2 ) Uniform grid search μ 2 μ 1 Adaptive search (NA, SA, GA) 180

41 Cross validation properties Cross validation is a way of assessing the quality of a solution based on the principle of measuring the fit to missing data On the plus side Cross validation can be used in cases where a suitable level of data misfit is not known in advance (it is in effect a bootstrap measure of data variance) Provides an objective criterion for picking a preferred solution. Can be automated. (GCV for linear problems with no inequality constraints) On the minus side Has been criticized for over-smoothing models. Computational burden has often seen it ignored (e.g. by geophysicists!) 181

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

ECE295, Data Assimila0on and Inverse Problems, Spring 2015 ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch

More information

Singular value decomposition. If only the first p singular values are nonzero we write. U T o U p =0

Singular value decomposition. If only the first p singular values are nonzero we write. U T o U p =0 Singular value decomposition If only the first p singular values are nonzero we write G =[U p U o ] " Sp 0 0 0 # [V p V o ] T U p represents the first p columns of U U o represents the last N-p columns

More information

Choosing the Regularization Parameter

Choosing the Regularization Parameter Choosing the Regularization Parameter At our disposal: several regularization methods, based on filtering of the SVD components. Often fairly straightforward to eyeball a good TSVD truncation parameter

More information

Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution

Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Rosemary Renaut, Jodi Mead Arizona State and Boise State September 2007 Renaut and Mead (ASU/Boise) Scalar

More information

Linearized inverse Problems

Linearized inverse Problems Linearized inverse Problems (Weakly nonlinear problems) Using a Taylor expansion then away we go... Linearized inverse problems Nonlinear inverse problem d obs,i = g i (m) Choose a reference model m o

More information

Option 1: Landmark Registration We can try to align specific points. The Registration Problem. Landmark registration. Time-Warping Functions

Option 1: Landmark Registration We can try to align specific points. The Registration Problem. Landmark registration. Time-Warping Functions The Registration Problem Most analyzes only account for variation in amplitude. Frequently, observed data exhibit features that vary in time. Option 1: Landmark Registration We can try to align specific

More information

Inversion of Phase Data for a Phase Velocity Map 101. Summary for CIDER12 Non-Seismologists

Inversion of Phase Data for a Phase Velocity Map 101. Summary for CIDER12 Non-Seismologists Inversion of Phase Data for a Phase Velocity Map 101 Summary for CIDER12 Non-Seismologists 1. Setting up a Linear System of Equations This is a quick-and-dirty, not-peer reviewed summary of how a dataset

More information

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Sparsity in system identification and data-driven control

Sparsity in system identification and data-driven control 1 / 40 Sparsity in system identification and data-driven control Ivan Markovsky This signal is not sparse in the "time domain" 2 / 40 But it is sparse in the "frequency domain" (it is weighted sum of six

More information

Linear Least-Squares Data Fitting

Linear Least-Squares Data Fitting CHAPTER 6 Linear Least-Squares Data Fitting 61 Introduction Recall that in chapter 3 we were discussing linear systems of equations, written in shorthand in the form Ax = b In chapter 3, we just considered

More information

One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017

One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 1 One Picture and a Thousand Words Using Matrix Approximations Dianne P. O Leary

More information

Modeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop

Modeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop Modeling Data with Linear Combinations of Basis Functions Read Chapter 3 in the text by Bishop A Type of Supervised Learning Problem We want to model data (x 1, t 1 ),..., (x N, t N ), where x i is a vector

More information

= (G T G) 1 G T d. m L2

= (G T G) 1 G T d. m L2 The importance of the Vp/Vs ratio in determining the error propagation and the resolution in linear AVA inversion M. Aleardi, A. Mazzotti Earth Sciences Department, University of Pisa, Italy Introduction.

More information

Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining

Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Combinations of features Given a data matrix X n p with p fairly large, it can

More information

Inversion Methods for Determining. Tsunami Source Amplitudes from DART Buoy Data. Don Percival

Inversion Methods for Determining. Tsunami Source Amplitudes from DART Buoy Data. Don Percival Inversion Methods for Determining Tsunami Source Amplitudes from DART Buoy Data Don Percival Applied Physics Laboratory University of Washington, Seattle NOAA-sponsored collaborative effort overheads for

More information

Uncertainty Quantification for Inverse Problems. November 7, 2011

Uncertainty Quantification for Inverse Problems. November 7, 2011 Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance

More information

Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve

Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Rosemary Renaut, Jodi Mead Arizona State and Boise State Copper Mountain Conference on Iterative Methods

More information

Discrete ill posed problems

Discrete ill posed problems Discrete ill posed problems Gérard MEURANT October, 2008 1 Introduction to ill posed problems 2 Tikhonov regularization 3 The L curve criterion 4 Generalized cross validation 5 Comparisons of methods Introduction

More information

Parameter estimation: A new approach to weighting a priori information

Parameter estimation: A new approach to weighting a priori information c de Gruyter 27 J. Inv. Ill-Posed Problems 5 (27), 2 DOI.55 / JIP.27.xxx Parameter estimation: A new approach to weighting a priori information J.L. Mead Communicated by Abstract. We propose a new approach

More information

5 Linear Algebra and Inverse Problem

5 Linear Algebra and Inverse Problem 5 Linear Algebra and Inverse Problem 5.1 Introduction Direct problem ( Forward problem) is to find field quantities satisfying Governing equations, Boundary conditions, Initial conditions. The direct problem

More information

MATH 590: Meshfree Methods

MATH 590: Meshfree Methods MATH 590: Meshfree Methods Chapter 34: Improving the Condition Number of the Interpolation Matrix Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu

More information

A Modern Look at Classical Multivariate Techniques

A Modern Look at Classical Multivariate Techniques A Modern Look at Classical Multivariate Techniques Yoonkyung Lee Department of Statistics The Ohio State University March 16-20, 2015 The 13th School of Probability and Statistics CIMAT, Guanajuato, Mexico

More information

Numerical Methods of Approximation

Numerical Methods of Approximation Contents 31 Numerical Methods of Approximation 31.1 Polynomial Approximations 2 31.2 Numerical Integration 28 31.3 Numerical Differentiation 58 31.4 Nonlinear Equations 67 Learning outcomes In this Workbook

More information

Time Series. Anthony Davison. c

Time Series. Anthony Davison. c Series Anthony Davison c 2008 http://stat.epfl.ch Periodogram 76 Motivation............................................................ 77 Lutenizing hormone data..................................................

More information

Application of Principal Component Analysis to TES data

Application of Principal Component Analysis to TES data Application of Principal Component Analysis to TES data Clive D Rodgers Clarendon Laboratory University of Oxford Madison, Wisconsin, 27th April 2006 1 My take on the PCA business 2/41 What is the best

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

Multiple realizations using standard inversion techniques a

Multiple realizations using standard inversion techniques a Multiple realizations using standard inversion techniques a a Published in SEP report, 105, 67-78, (2000) Robert G Clapp 1 INTRODUCTION When solving a missing data problem, geophysicists and geostatisticians

More information

Statistically-Based Regularization Parameter Estimation for Large Scale Problems

Statistically-Based Regularization Parameter Estimation for Large Scale Problems Statistically-Based Regularization Parameter Estimation for Large Scale Problems Rosemary Renaut Joint work with Jodi Mead and Iveta Hnetynkova March 1, 2010 National Science Foundation: Division of Computational

More information

Mathematical Methods in Engineering and Science Prof. Bhaskar Dasgupta Department of Mechanical Engineering Indian Institute of Technology, Kanpur

Mathematical Methods in Engineering and Science Prof. Bhaskar Dasgupta Department of Mechanical Engineering Indian Institute of Technology, Kanpur Mathematical Methods in Engineering and Science Prof. Bhaskar Dasgupta Department of Mechanical Engineering Indian Institute of Technology, Kanpur Module - I Solution of Linear Systems Lecture - 05 Ill-Conditioned

More information

Linear inversion methods and generalized cross-validation

Linear inversion methods and generalized cross-validation Online supplement to Using generalized cross-validation to select parameters in inversions for regional caron fluxes Linear inversion methods and generalized cross-validation NIR Y. KRAKAUER AND TAPIO

More information

Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53

Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53 Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53 40 Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53 Short Note Multiple realizations using standard inversion

More information

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search

More information

Uncertainty Quantification for Machine Learning and Statistical Models

Uncertainty Quantification for Machine Learning and Statistical Models Uncertainty Quantification for Machine Learning and Statistical Models David J. Stracuzzi Joint work with: Max Chen, Michael Darling, Stephen Dauphin, Matt Peterson, and Chris Young Sandia National Laboratories

More information

15 Singular Value Decomposition

15 Singular Value Decomposition 15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Discrete Ill Posed and Rank Deficient Problems. Alistair Boyle, Feb 2009, SYS5906: Directed Studies Inverse Problems 1

Discrete Ill Posed and Rank Deficient Problems. Alistair Boyle, Feb 2009, SYS5906: Directed Studies Inverse Problems 1 Discrete Ill Posed and Rank Deficient Problems Alistair Boyle, Feb 2009, SYS5906: Directed Studies Inverse Problems 1 Definitions Overview Inversion, SVD, Picard Condition, Rank Deficient, Ill-Posed Classical

More information

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren 1 / 34 Metamodeling ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 1, 2015 2 / 34 1. preliminaries 1.1 motivation 1.2 ordinary least square 1.3 information

More information

ISyE 691 Data mining and analytics

ISyE 691 Data mining and analytics ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)

More information

Structural Damage Detection Using Time Windowing Technique from Measured Acceleration during Earthquake

Structural Damage Detection Using Time Windowing Technique from Measured Acceleration during Earthquake Structural Damage Detection Using Time Windowing Technique from Measured Acceleration during Earthquake Seung Keun Park and Hae Sung Lee ABSTRACT This paper presents a system identification (SI) scheme

More information

A study on regularization parameter choice in Near-field Acoustical Holography

A study on regularization parameter choice in Near-field Acoustical Holography Acoustics 8 Paris A study on regularization parameter choice in Near-field Acoustical Holography J. Gomes a and P.C. Hansen b a Brüel & Kjær Sound and Vibration Measurement A/S, Skodsborgvej 37, DK-285

More information

Numerical Methods I Solving Nonlinear Equations

Numerical Methods I Solving Nonlinear Equations Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)

More information

SIO223A A Brief Introduction to Geophysical Modeling and Inverse Theory.

SIO223A A Brief Introduction to Geophysical Modeling and Inverse Theory. SIO223A A Brief Introduction to Geophysical Modeling and Inverse Theory. Forward modeling: data model space space ˆd = f(x, m) m =(m 1,m 2,..., m N ) x =(x 1,x 2,x 3,..., x km ) ˆd =(ˆd 1, ˆd 2, ˆd 3,...,

More information

SIGNAL AND IMAGE RESTORATION: SOLVING

SIGNAL AND IMAGE RESTORATION: SOLVING 1 / 55 SIGNAL AND IMAGE RESTORATION: SOLVING ILL-POSED INVERSE PROBLEMS - ESTIMATING PARAMETERS Rosemary Renaut http://math.asu.edu/ rosie CORNELL MAY 10, 2013 2 / 55 Outline Background Parameter Estimation

More information

In the Name of God. Lectures 15&16: Radial Basis Function Networks

In the Name of God. Lectures 15&16: Radial Basis Function Networks 1 In the Name of God Lectures 15&16: Radial Basis Function Networks Some Historical Notes Learning is equivalent to finding a surface in a multidimensional space that provides a best fit to the training

More information

The Application of Discrete Tikhonov Regularization Inverse Problem in Seismic Tomography

The Application of Discrete Tikhonov Regularization Inverse Problem in Seismic Tomography The Application of Discrete Tikhonov Regularization Inverse Problem in Seismic Tomography KAMBIZ TEIMOORNEGAD, NEDA POROOHAN 2, Geology Department Islamic Azad University, Lahijan Branch 2 Islamic Azad

More information

Statistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1

Statistics 202: Data Mining. c Jonathan Taylor. Week 2 Based in part on slides from textbook, slides of Susan Holmes. October 3, / 1 Week 2 Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Part I Other datatypes, preprocessing 2 / 1 Other datatypes Document data You might start with a collection of

More information

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR 1. Definition Existence Theorem 1. Assume that A R m n. Then there exist orthogonal matrices U R m m V R n n, values σ 1 σ 2... σ p 0 with p = min{m, n},

More information

Part I. Other datatypes, preprocessing. Other datatypes. Other datatypes. Week 2 Based in part on slides from textbook, slides of Susan Holmes

Part I. Other datatypes, preprocessing. Other datatypes. Other datatypes. Week 2 Based in part on slides from textbook, slides of Susan Holmes Week 2 Based in part on slides from textbook, slides of Susan Holmes Part I Other datatypes, preprocessing October 3, 2012 1 / 1 2 / 1 Other datatypes Other datatypes Document data You might start with

More information

Detection theory. H 0 : x[n] = w[n]

Detection theory. H 0 : x[n] = w[n] Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal

More information

[POLS 8500] Review of Linear Algebra, Probability and Information Theory

[POLS 8500] Review of Linear Algebra, Probability and Information Theory [POLS 8500] Review of Linear Algebra, Probability and Information Theory Professor Jason Anastasopoulos ljanastas@uga.edu January 12, 2017 For today... Basic linear algebra. Basic probability. Programming

More information

Metric-based classifiers. Nuno Vasconcelos UCSD

Metric-based classifiers. Nuno Vasconcelos UCSD Metric-based classifiers Nuno Vasconcelos UCSD Statistical learning goal: given a function f. y f and a collection of eample data-points, learn what the function f. is. this is called training. two major

More information

Machine Learning (Spring 2012) Principal Component Analysis

Machine Learning (Spring 2012) Principal Component Analysis 1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in

More information

MSA220 Statistical Learning for Big Data

MSA220 Statistical Learning for Big Data MSA220 Statistical Learning for Big Data Lecture 4 Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology More on Discriminant analysis More on Discriminant

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability

More information

A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS

A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS SILVIA NOSCHESE AND LOTHAR REICHEL Abstract. Truncated singular value decomposition (TSVD) is a popular method for solving linear discrete ill-posed

More information

Least Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e.

Least Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e. Least Squares with Eamples in Signal Processing Ivan Selesnick March 7, 3 NYU-Poly These notes address (approimate) solutions to linear equations by least squares We deal with the easy case wherein the

More information

INVERSE PROBEMS: AN INTRODUCTION

INVERSE PROBEMS: AN INTRODUCTION SEISMOLOGY Master Degree Programme in Physics - UNITS Physics of the Earth and of the Environment INVERSE PROBEMS: AN INTRODUCTION FABIO ROMANELLI Department of Mathematics & Geosciences University of

More information

(a)

(a) Chapter 8 Subspace Methods 8. Introduction Principal Component Analysis (PCA) is applied to the analysis of time series data. In this context we discuss measures of complexity and subspace methods for

More information

Irregular Solutions of an Ill-Posed Problem

Irregular Solutions of an Ill-Posed Problem Irregular Solutions of an Ill-Posed Problem Peter Linz University of California at Davis Richard L.C. Wang Clearsight Systems Inc. Abstract: Tikhonov regularization is a popular and effective method for

More information

Introduction. Chapter One

Introduction. Chapter One Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and

More information

Getting Started with Communications Engineering

Getting Started with Communications Engineering 1 Linear algebra is the algebra of linear equations: the term linear being used in the same sense as in linear functions, such as: which is the equation of a straight line. y ax c (0.1) Of course, if we

More information

Slip distributions of the 1944 Tonankai and 1946 Nankai earthquakes including the horizontal movement effect on tsunami generation

Slip distributions of the 1944 Tonankai and 1946 Nankai earthquakes including the horizontal movement effect on tsunami generation Slip distributions of the 1944 Tonankai and 1946 Nankai earthquakes including the horizontal movement effect on tsunami generation Toshitaka Baba Research Program for Plate Dynamics, Institute for Frontier

More information

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Silvia Gazzola Dipartimento di Matematica - Università di Padova January 10, 2012 Seminario ex-studenti 2 Silvia Gazzola

More information

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane. Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 2018 3 Lecture 3 3.1 General remarks March 4, 2018 This

More information

4. DATA ASSIMILATION FUNDAMENTALS

4. DATA ASSIMILATION FUNDAMENTALS 4. DATA ASSIMILATION FUNDAMENTALS... [the atmosphere] "is a chaotic system in which errors introduced into the system can grow with time... As a consequence, data assimilation is a struggle between chaotic

More information

Spatial Process Estimates as Smoothers: A Review

Spatial Process Estimates as Smoothers: A Review Spatial Process Estimates as Smoothers: A Review Soutir Bandyopadhyay 1 Basic Model The observational model considered here has the form Y i = f(x i ) + ɛ i, for 1 i n. (1.1) where Y i is the observed

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

nonlinear simultaneous equations of type (1)

nonlinear simultaneous equations of type (1) Module 5 : Solving Nonlinear Algebraic Equations Section 1 : Introduction 1 Introduction Consider set of nonlinear simultaneous equations of type -------(1) -------(2) where and represents a function vector.

More information

Introduction to Supervised Learning. Performance Evaluation

Introduction to Supervised Learning. Performance Evaluation Introduction to Supervised Learning Performance Evaluation Marcelo S. Lauretto Escola de Artes, Ciências e Humanidades, Universidade de São Paulo marcelolauretto@usp.br Lima - Peru Performance Evaluation

More information

REGULARIZATION PARAMETER SELECTION IN DISCRETE ILL POSED PROBLEMS THE USE OF THE U CURVE

REGULARIZATION PARAMETER SELECTION IN DISCRETE ILL POSED PROBLEMS THE USE OF THE U CURVE Int. J. Appl. Math. Comput. Sci., 007, Vol. 17, No., 157 164 DOI: 10.478/v10006-007-0014-3 REGULARIZATION PARAMETER SELECTION IN DISCRETE ILL POSED PROBLEMS THE USE OF THE U CURVE DOROTA KRAWCZYK-STAŃDO,

More information

Chapter 2 Interpolation

Chapter 2 Interpolation Chapter 2 Interpolation Experiments usually produce a discrete set of data points (x i, f i ) which represent the value of a function f (x) for a finite set of arguments {x 0...x n }. If additional data

More information

Chapter 9. Non-Parametric Density Function Estimation

Chapter 9. Non-Parametric Density Function Estimation 9-1 Density Estimation Version 1.2 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least

More information

Linear Inverse Problems

Linear Inverse Problems Linear Inverse Problems Ajinkya Kadu Utrecht University, The Netherlands February 26, 2018 Outline Introduction Least-squares Reconstruction Methods Examples Summary Introduction 2 What are inverse problems?

More information

Lecture No. 1 Introduction to Method of Weighted Residuals. Solve the differential equation L (u) = p(x) in V where L is a differential operator

Lecture No. 1 Introduction to Method of Weighted Residuals. Solve the differential equation L (u) = p(x) in V where L is a differential operator Lecture No. 1 Introduction to Method of Weighted Residuals Solve the differential equation L (u) = p(x) in V where L is a differential operator with boundary conditions S(u) = g(x) on Γ where S is a differential

More information

Machine Learning. Lecture 9: Learning Theory. Feng Li.

Machine Learning. Lecture 9: Learning Theory. Feng Li. Machine Learning Lecture 9: Learning Theory Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2018 Why Learning Theory How can we tell

More information

5.1 2D example 59 Figure 5.1: Parabolic velocity field in a straight two-dimensional pipe. Figure 5.2: Concentration on the input boundary of the pipe. The vertical axis corresponds to r 2 -coordinate,

More information

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

Linear Models 1. Isfahan University of Technology Fall Semester, 2014 Linear Models 1 Isfahan University of Technology Fall Semester, 2014 References: [1] G. A. F., Seber and A. J. Lee (2003). Linear Regression Analysis (2nd ed.). Hoboken, NJ: Wiley. [2] A. C. Rencher and

More information

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut Collaborators: Jodi Mead and Iveta Hnetynkova DEPARTMENT OF MATHEMATICS

More information

CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation

CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80

More information

Noise & Data Reduction

Noise & Data Reduction Noise & Data Reduction Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum Dimension Reduction 1 Remember: Central Limit

More information

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising Review: Optimal Average/Scaling is equivalent to Minimise χ Two 1-parameter models: Estimating < > : Scaling a pattern: Two equivalent methods: Algebra of Random Variables: Optimal Average and Optimal

More information

Deconvolution. Parameter Estimation in Linear Inverse Problems

Deconvolution. Parameter Estimation in Linear Inverse Problems Image Parameter Estimation in Linear Inverse Problems Chair for Computer Aided Medical Procedures & Augmented Reality Department of Computer Science, TUM November 10, 2006 Contents A naive approach......with

More information

M. Holschneider 1 A. Eicker 2 R. Schachtschneider 1 T. Mayer-Guerr 2 K. Ilk 2

M. Holschneider 1 A. Eicker 2 R. Schachtschneider 1 T. Mayer-Guerr 2 K. Ilk 2 SPP project TREGMAT: Tailored Gravity Field Models for Mass Distributions and Mass Transport Phenomena in the Earth System by by M. 1 A. Eicker 2 R. Schachtschneider 1 T. Mayer-Guerr 2 K. Ilk 2 1 University

More information

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut DEPARTMENT OF MATHEMATICS AND STATISTICS Prague 2008 MATHEMATICS AND STATISTICS

More information

Generalized Cross Validation

Generalized Cross Validation 26 november 2009 Plan 1 Generals 2 What Regularization Parameter λ is Optimal? Examples Dening the Optimal λ 3 Generalized Cross-Validation (GCV) Convergence Result 4 Discussion Plan 1 Generals 2 What

More information

Inverse Theory Methods in Experimental Physics

Inverse Theory Methods in Experimental Physics Inverse Theory Methods in Experimental Physics Edward Sternin PHYS 5P10: 2018-02-26 Edward Sternin Inverse Theory Methods PHYS 5P10: 2018-02-26 1 / 29 1 Introduction Indirectly observed data A new research

More information

Scalable kernel methods and their use in black-box optimization

Scalable kernel methods and their use in black-box optimization with derivatives Scalable kernel methods and their use in black-box optimization David Eriksson Center for Applied Mathematics Cornell University dme65@cornell.edu November 9, 2018 1 2 3 4 1/37 with derivatives

More information

Partititioned Methods for Multifield Problems

Partititioned Methods for Multifield Problems Partititioned Methods for Multifield Problems Joachim Rang, 22.7.215 22.7.215 Joachim Rang Partititioned Methods for Multifield Problems Seite 1 Non-matching matches usually the fluid and the structure

More information

11 More Regression; Newton s Method; ROC Curves

11 More Regression; Newton s Method; ROC Curves More Regression; Newton s Method; ROC Curves 59 11 More Regression; Newton s Method; ROC Curves LEAST-SQUARES POLYNOMIAL REGRESSION Replace each X i with feature vector e.g. (X i ) = [X 2 i1 X i1 X i2

More information

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Spectral Approaches to Bayesian Inference Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to

More information

8 The SVD Applied to Signal and Image Deblurring

8 The SVD Applied to Signal and Image Deblurring 8 The SVD Applied to Signal and Image Deblurring We will discuss the restoration of one-dimensional signals and two-dimensional gray-scale images that have been contaminated by blur and noise. After an

More information

Some general observations.

Some general observations. Modeling and analyzing data from computer experiments. Some general observations. 1. For simplicity, I assume that all factors (inputs) x1, x2,, xd are quantitative. 2. Because the code always produces

More information

Chapter 9. Non-Parametric Density Function Estimation

Chapter 9. Non-Parametric Density Function Estimation 9-1 Density Estimation Version 1.1 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least

More information

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q

Kalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Kalman Filter Kalman Filter Predict: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Update: K = P k k 1 Hk T (H k P k k 1 Hk T + R) 1 x k k = x k k 1 + K(z k H k x k k 1 ) P k k =(I

More information

Degrees of Freedom in Regression Ensembles

Degrees of Freedom in Regression Ensembles Degrees of Freedom in Regression Ensembles Henry WJ Reeve Gavin Brown University of Manchester - School of Computer Science Kilburn Building, University of Manchester, Oxford Rd, Manchester M13 9PL Abstract.

More information

Basic Linear Inverse Method Theory - DRAFT NOTES

Basic Linear Inverse Method Theory - DRAFT NOTES Basic Linear Inverse Method Theory - DRAFT NOTES Peter P. Jones 1 1 Centre for Compleity Science, University of Warwick, Coventry CV4 7AL, UK (Dated: 21st June 2012) BASIC LINEAR INVERSE METHOD THEORY

More information

We consider the problem of finding a polynomial that interpolates a given set of values:

We consider the problem of finding a polynomial that interpolates a given set of values: Chapter 5 Interpolation 5. Polynomial Interpolation We consider the problem of finding a polynomial that interpolates a given set of values: x x 0 x... x n y y 0 y... y n where the x i are all distinct.

More information

STATISTICS 174: APPLIED STATISTICS TAKE-HOME FINAL EXAM POSTED ON WEBPAGE: 6:00 pm, DECEMBER 6, 2004 HAND IN BY: 6:00 pm, DECEMBER 7, 2004 This is a

STATISTICS 174: APPLIED STATISTICS TAKE-HOME FINAL EXAM POSTED ON WEBPAGE: 6:00 pm, DECEMBER 6, 2004 HAND IN BY: 6:00 pm, DECEMBER 7, 2004 This is a STATISTICS 174: APPLIED STATISTICS TAKE-HOME FINAL EXAM POSTED ON WEBPAGE: 6:00 pm, DECEMBER 6, 2004 HAND IN BY: 6:00 pm, DECEMBER 7, 2004 This is a take-home exam. You are expected to work on it by yourself

More information

Stephen Scott.

Stephen Scott. 1 / 35 (Adapted from Ethem Alpaydin and Tom Mitchell) sscott@cse.unl.edu In Homework 1, you are (supposedly) 1 Choosing a data set 2 Extracting a test set of size > 30 3 Building a tree on the training

More information