Application of Principal Component Analysis to TES data

Size: px
Start display at page:

Download "Application of Principal Component Analysis to TES data"

Transcription

1 Application of Principal Component Analysis to TES data Clive D Rodgers Clarendon Laboratory University of Oxford Madison, Wisconsin, 27th April

2 My take on the PCA business 2/41

3 What is the best estimate of a spectrum given many measured spectra? Simple approach using optimal estimation: Forward model is y=x+ ε x is true spectrum, y is measurement, ε is noise Maximum a posteriori estimate of x is x r = x a + S a (S a + S ε ) -1 (y - x a ) 3/41

4 Continued If we have a large sample of spectra: Expect that x a = <x + ε> = <y> Can estimate S a +S ε from statistics of y Should have a good idea of S ε a priori But: S a (S a +S ε ) -1 will be a large matrix S a found from S a +S ε and S ε is likely to be ill-conditioned S a is likely to have a small number of eigenvalues greater than noise 4/41

5 Singular Vectors (or Principal Components) Let the ensemble of (y -<y>) be N columns of a matrix Y Represent Y as its singular vector decomposition: Y = UΛV T where Λ is diagonal, U T U=I and V T V=I The j th individual spectrum y j is then y j = <y> + Σ i u i λ i v ij T The spectrum is represented as a sum of columns u i of U, with coefficients λ i v ijt. Because U T U=I, we can compute λ i v ijt for any spectrum as U T (y j -<y>). 5/41

6 What do we expect? 1 / N YY T is the covariance matrix S y of the spectra Left singular vectors U are the same as eigenvectors of YY T, singular values are the square roots of its eigenvalues In the linear case with independent constant noise, S y would be S y = S a + σ ε 2 I S a has rank n, I is of dimension m >> n, where n is the degrees of freedom of the atmosphere The eigenvalues of S y are λ i2 /N The eigenvalues of S a should be λ i2 /N - σ ε 2 6/41

7 Reconstructing Spectra We can drop terms with λ i2 /N ~ σ ε2, i.e. i > n, without significant loss they correspond to noise only Better, multiply retained terms by something like (λ i2 - N σ ε 2)/ λ i 2 So spectra can be reconstructed from the first few coefficients. The noise can be reconstructed from the rest Reconstructed spectra have much reduced noise 7/41

8 Information Content The Shannon information content of a single spectrum relative to the ensemble is H = 1 / 2 i ln(λ i 2 /N σ ε2 ) The degrees of freedom for signal is d s = i (1 - N σ ε2 /λ i2 ) But reality isn t quite like that... 8/41

9 Tropospheric Emission Spectrometer Connes-type four-port Fourier transform spectrometer Nadir view: 0.06 cm -1 Limb view: cm element detector arrays in 4 focal planes: 1A ( ), 1B ( ), 2A ( ), 2B ( ) 0.5 X 5 km nadir; 2.3 X 23 km limb Global Survey mode: Along track nadir, 216 views/orbit. 9/41

10 SVD Applications Information content analysis Validation Instrument performance Forward model De-noised spectra Retrieval? 10/41

11 A typical example Run Sept 2004? Channel 1B cm views * 16 detectors 11/41

12 An example of Information Content from Channel 1B2 Using one orbit, 1152 spectra each with 3951 elements cm -1 n Lambda Information d.f.s cum. cum Total: 14.5 bits 7.05 d.f. Sing. Vals. Residual Fit to noise asymptote SV number /41

13 How do you fit noise to the asymptote? I havn t used enough spectra Expected distribution of singular values of independent gaussian noise for a finite number of samples Simulation Theory: there is an explicit formula Fit expected distribution to the long tail 13/41

14 Validation an early example Run Sept 2004? Channel 1B cm views * 16 detectors 14/41

15 Singular Vectors 15/41

16 Singular vectors * Lambda 16/41

17 V-vectors 17/41

18 Features Most of variation is in the first singular vector. First six are: Data spikes - identified Data spikes - unidentified Pixel-dependent variation in the spectra Scan-direction dependence 18/41

19 Singular Vector 6 Systematic variation across the detector array Must be an artefact Suggests systematic error in ILS How is it related to mean spectrum? Least squares fit to find function that when convolved with mean spectrum gives SV6 19/41

20 SV6 20/41

21 SV6 21/41

22 SV6 Suggests the derivative of the ILS 22/41

23 Curious behaviour of the singular values of an ensemble of B1 spectra What is this? This is the atmospheric variability cm -1 23/41

24 Features It occurs at around SV 200. I have used 100 sequences, 1600 spectra. Implies two of these SV s per sequence. Confirmed by trying other numbers of sequences The vectors look like noise Only in 2B1 does this 24/41

25 With 64 sequences, plotted in 9/04 The shoulder here is at 64 there was only 1 per sequence then 25/41

26 Examine pixel noise Look at the spectrum for each pixel minus the mean over all pixels removes atmospheric structure, but differences will be slightly correlated Singular vectors of the difference: look like noise Coefficients of the singular vectors for one set of 16: don t look like noise for two of the SV s Covariance & correlation matrices 26/41

27 SV coefficients 27/41

28 2B1 noise is significantly correlated between pixels 28/41

29 Deductions The change from 1 to 2 SVs per sequence implies something in the processing rather than the instrument Each pixel has its own independent noise Plus noise correlated between all the pixels of each sequence The correlations depend on distance between the pixels, quasi-sinusoidally Not an odd pixel/even pixel effect 29/41

30 SV s of Residual Spectra Difference between measured spectrum and calculated from retrieval No just in the microwindows used Courtesy Reinhard Beer & Susan S. Kulawik 30/41

31 1.0e-7 Mean Residual Radiance for Run 2147 (2004 September 21) from 220 Spectra between 30 o S & 30 o N 5.0e-8 Residual Radiance, watts/cm 2 /sr/cm e-8-1.0e-7-1.5e-7-2.0e-7-2.5e-7 Residual O 3 Residual Silicate Surface Feature "Over-fitted" HDO Unmodeled F12 Residual CH 4-3.0e-7 Retrieval Microwindows -3.5e Frequency, cm -1 31/41

32 Singular Values of Run 2147 Full-Filter Residuals ("W" Matrix Diagonal; First 51 Elements of 568 only) 3.0e-5 2.5e-5 2.0e-5 Singular Value 1.5e-5 1.0e-5 5.0e Singular Value Number 32/41

33 0.08 U Singular Vector Amplitude The quasi-periodicity is due to the orbital variation of total radiance Index (Time order) V Singular Vector Amplitude Note the strong resemblance to the grand average Frequency, cm -1 33

34 0.15 U Singular Vector 1 Amplitude This pattern of roughly orbits has been observed in other diagnostics. The cause is under investigation Index (Time order) V Singular Vector 1 Amplitude The discontinuity at 1120 cm -1 occurs at the overlap of two different filters Frequency, cm -1 34/41

35 Amplitude U Singular Vector 2 Positive peaks align with land scenes (green bars) Index (Time order) Amplitude V Singular Vector Broad feature is a component of the unmodeled surface silicate emissivity Frequency, cm -1 35/41

36 Retrieval some untested thoughts Possible methods include: Retrieve from denoised spectrum in the same way as usual, but with noise covariance reduced in some ad-hoc way not optimal, inefficient Ditto, but with the correct error covariance covariance matrix is singular Select a subset of channels according to information content straightforward if linear, not otherwise Retrieve from SV representation coefficients needs a special forward model for efficiency PCRTM or a regression or neural net method 36/41

37 Characterising a denoised spectrum Reconstruct y with the first r singular vectors, U r, giving y r y r = U r U rt y Measurement function is: y r = U r U rt F(x) + U r U rt ε We assume that U r U rt F(x) ~ F(x), so y r = F(x) + U r U rt ε = F(x) + ε r Covariance of ε r is S εr = U r U rt S ε U r U r T which is of rank r 37/41

38 Retrieve from a denoised spectrum We can use all or some of the spectrum. Microwindows Channels selected by information content Channels selected by Xu Liu s approach Define a selection operator M, and retrieve from the m element subset y m = My r The subset depends on the whole of the original spectrum, y m = MU r U rt y, but we assume that we can model it as MF(x). The error covariance of ε m is S m = M U r U rt S ε U r U rt M T which may be singular if m > r. 38/41

39 Do another SVD Use the SVD of MU r U rt = LΛR T to give L T y m = L T MF(x) + ΛR T ε Drop elements of L T y m with zero (or small) singular values, leaving p=min(r,m) elements at most. Gives L pt y m Retrieve from the rest of L pt y m with L pt MF(x) as the forward model. L pt can be precomputed. Error covariance is Λ p R pt S ε R p Λ p. If S ε = σ 2 I, this reduces to σ 2 Λ p2. 39/41

40 Comments This allows us to retrieve with the minimum number of evaluations of the forward model And the minimum length of measurement vector With very little extra matrix manipulation I havn t tried it yet, so I don t know what the catch is 40/41

41 Summary/Comments PCA does not improve the information content of a measurement But it does allow us to use it more efficiently Very useful for validation - separates out independent sources of variation, e.g. artefacts, omitted physics,... Denoised spectra guide the eye to real features that may otherwise not be seen Data compression 41/41

Inverse Theory. COST WaVaCS Winterschool Venice, February Stefan Buehler Luleå University of Technology Kiruna

Inverse Theory. COST WaVaCS Winterschool Venice, February Stefan Buehler Luleå University of Technology Kiruna Inverse Theory COST WaVaCS Winterschool Venice, February 2011 Stefan Buehler Luleå University of Technology Kiruna Overview Inversion 1 The Inverse Problem 2 Simple Minded Approach (Matrix Inversion) 3

More information

Retrieval Algorithm Using Super channels

Retrieval Algorithm Using Super channels Retrieval Algorithm Using Super channels Xu Liu NASA Langley Research Center, Hampton VA 23662 D. K. Zhou, A. M. Larar (NASA LaRC) W. L. Smith (HU and UW) P. Schluessel (EUMETSAT) Hank Revercomb (UW) Jonathan

More information

IASI PC compression Searching for signal in the residuals. Thomas August, Nigel Atkinson, Fiona Smith

IASI PC compression Searching for signal in the residuals. Thomas August, Nigel Atkinson, Fiona Smith IASI PC compression Searching for signal in the residuals Tim.Hultberg@EUMETSAT.INT Thomas August, Nigel Atkinson, Fiona Smith Raw radiance (minus background) Reconstructed radiance (minus background)

More information

Compression of IASI Data and Representation of FRTM in EOF Space

Compression of IASI Data and Representation of FRTM in EOF Space Compression of IASI Data and Representation of FRTM in EOF Space Peter Schlüssel EUMETSAT Am Kavalleriesand 31 64295 Darmstadt Germany Anthony C. L. Lee Met Office FitzRoy Road Exeter, Devon EX1 3PB United

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

COMBINED OZONE RETRIEVAL USING THE MICHELSON INTERFEROMETER FOR PASSIVE ATMOSPHERIC SOUNDING (MIPAS) AND THE TROPOSPHERIC EMISSION SPECTROMETER (TES)

COMBINED OZONE RETRIEVAL USING THE MICHELSON INTERFEROMETER FOR PASSIVE ATMOSPHERIC SOUNDING (MIPAS) AND THE TROPOSPHERIC EMISSION SPECTROMETER (TES) COMBINED OZONE RETRIEVAL USING THE MICHELSON INTERFEROMETER FOR PASSIVE ATMOSPHERIC SOUNDING (MIPAS) AND THE TROPOSPHERIC EMISSION SPECTROMETER (TES) Claire Waymark, Anu Dudhia, John Barnett, Fred Taylor,

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

Machine Learning (Spring 2012) Principal Component Analysis

Machine Learning (Spring 2012) Principal Component Analysis 1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in

More information

Principal Component Analysis (PCA) of AIRS Data

Principal Component Analysis (PCA) of AIRS Data Principal Component Analysis (PCA) of AIRS Data Mitchell D. Goldberg 1, Lihang Zhou 2, Walter Wolf 2 and Chris Barnet 1 NOAA/NESDIS/Office of Research and Applications, Camp Springs, MD 1 QSS Group Inc.

More information

Satellite data information content, channel selection and density

Satellite data information content, channel selection and density Satellite data information content, channel selection and density Élisabeth Gérard Météo-France/CNRM/GMAP Toulouse, France [EG2], 2 nd ENVISAT summer school Frascati, Italy, 16-26 Aug 2004 1 Overview A.

More information

Retrievals of methane from IASI radiance spectra and comparisons with ground-based FTIR measurements

Retrievals of methane from IASI radiance spectra and comparisons with ground-based FTIR measurements ACCENT-AT2 follow-up meeting, 22-23 June 2009, Mainz, Germany Retrievals of methane from IASI radiance spectra and comparisons with ground-based FTIR measurements Tobias Kerzenmacher 1, Nicolas Kumps 1,

More information

Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining

Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Combinations of features Given a data matrix X n p with p fairly large, it can

More information

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

ECE295, Data Assimila0on and Inverse Problems, Spring 2015 ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch

More information

Principal Component Analysis CS498

Principal Component Analysis CS498 Principal Component Analysis CS498 Today s lecture Adaptive Feature Extraction Principal Component Analysis How, why, when, which A dual goal Find a good representation The features part Reduce redundancy

More information

Intercomparison of remote sounding instruments

Intercomparison of remote sounding instruments JOURNAL OF GEOPHYSICAL RESEARCH, VOL. 108, NO. D3, 4116, doi:10.1029/2002jd002299, 2003 Intercomparison of remote sounding instruments Clive D. Rodgers 1 and Brian J. Connor National Institute for Water

More information

ECE 501b Homework #6 Due: 11/26

ECE 501b Homework #6 Due: 11/26 ECE 51b Homework #6 Due: 11/26 1. Principal Component Analysis: In this assignment, you will explore PCA as a technique for discerning whether low-dimensional structure exists in a set of data and for

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

CIMSS Hyperspectral IR Sounding Retrieval (CHISR) Processing & Applications

CIMSS Hyperspectral IR Sounding Retrieval (CHISR) Processing & Applications CIMSS Hyperspectral IR Sounding Retrieval (CHISR) Processing & Applications Jun Li @, Elisabeth Weisz @, Jinlong Li @, Hui Liu #, Timothy J. Schmit &, Jason Otkin @ and many other CIMSS collaborators @Cooperative

More information

Singular value decomposition. If only the first p singular values are nonzero we write. U T o U p =0

Singular value decomposition. If only the first p singular values are nonzero we write. U T o U p =0 Singular value decomposition If only the first p singular values are nonzero we write G =[U p U o ] " Sp 0 0 0 # [V p V o ] T U p represents the first p columns of U U o represents the last N-p columns

More information

Exercise Sheet 1. 1 Probability revision 1: Student-t as an infinite mixture of Gaussians

Exercise Sheet 1. 1 Probability revision 1: Student-t as an infinite mixture of Gaussians Exercise Sheet 1 1 Probability revision 1: Student-t as an infinite mixture of Gaussians Show that an infinite mixture of Gaussian distributions, with Gamma distributions as mixing weights in the following

More information

Linear Subspace Models

Linear Subspace Models Linear Subspace Models Goal: Explore linear models of a data set. Motivation: A central question in vision concerns how we represent a collection of data vectors. The data vectors may be rasterized images,

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Regression Retrieval Overview. Larry McMillin

Regression Retrieval Overview. Larry McMillin Regression Retrieval Overview Larry McMillin Climate Research and Applications Division National Environmental Satellite, Data, and Information Service Washington, D.C. Larry.McMillin@noaa.gov Pick one

More information

Gopalkrishna Veni. Project 4 (Active Shape Models)

Gopalkrishna Veni. Project 4 (Active Shape Models) Gopalkrishna Veni Project 4 (Active Shape Models) Introduction Active shape Model (ASM) is a technique of building a model by learning the variability patterns from training datasets. ASMs try to deform

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

Principal Components Analysis (PCA)

Principal Components Analysis (PCA) Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

CS540 Machine learning Lecture 5

CS540 Machine learning Lecture 5 CS540 Machine learning Lecture 5 1 Last time Basis functions for linear regression Normal equations QR SVD - briefly 2 This time Geometry of least squares (again) SVD more slowly LMS Ridge regression 3

More information

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to: System 2 : Modelling & Recognising Modelling and Recognising Classes of Classes of Shapes Shape : PDM & PCA All the same shape? System 1 (last lecture) : limited to rigidly structured shapes System 2 :

More information

Application of PCA to IASI: An NWP Perspective. Andrew Collard, ECMWF. Acknowledgements to Tony McNally, Jean-Noel Thépaut

Application of PCA to IASI: An NWP Perspective. Andrew Collard, ECMWF. Acknowledgements to Tony McNally, Jean-Noel Thépaut Application of PCA to IASI: An NWP Perspective Andrew Collard, ECMWF Acknowledgements to Tony McNally, Jean-Noel Thépaut Overview Introduction Why use PCA/RR? Expected performance of PCA/RR for IASI Assimilation

More information

Retrieval problems in Remote Sensing

Retrieval problems in Remote Sensing Maarten Krol August 2006 m.c.krol@phys.uu.nl 1. Goal To show that remote sensing is an indirect measurement, i.e. that we do not measure directly, but by means of (for instance) electromagnetic radiation

More information

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:

More information

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS LAST TIME Intro to cudnn Deep neural nets using cublas and cudnn TODAY Building a better model for image classification Overfitting

More information

Singular Value Decompsition

Singular Value Decompsition Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost

More information

The Retrieval of Infrared Cooling Rate Profiles from Thermal Infrared Remote Sounder Radiance Measurements

The Retrieval of Infrared Cooling Rate Profiles from Thermal Infrared Remote Sounder Radiance Measurements The Retrieval of Infrared Cooling Rate Profiles from Thermal Infrared Remote Sounder Radiance Measurements Daniel Feldman Candidacy Presentation May 13, 2005 Outline 1. Background 2. Motivation 3. Cooling

More information

Announcements (repeat) Principal Components Analysis

Announcements (repeat) Principal Components Analysis 4/7/7 Announcements repeat Principal Components Analysis CS 5 Lecture #9 April 4 th, 7 PA4 is due Monday, April 7 th Test # will be Wednesday, April 9 th Test #3 is Monday, May 8 th at 8AM Just hour long

More information

Covariance to PCA. CS 510 Lecture #8 February 17, 2014

Covariance to PCA. CS 510 Lecture #8 February 17, 2014 Covariance to PCA CS 510 Lecture 8 February 17, 2014 Status Update Programming Assignment 2 is due March 7 th Expect questions about your progress at the start of class I still owe you Assignment 1 back

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 12 Jan-Willem van de Meent (credit: Yijun Zhao, Percy Liang) DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Linear Dimensionality

More information

Lecture 17: variance in a band = log(s xx (f)) df (2) If we want to plot something that is more directly representative of variance, we can try this:

Lecture 17: variance in a band = log(s xx (f)) df (2) If we want to plot something that is more directly representative of variance, we can try this: UCSD SIOC 221A: (Gille) 1 Lecture 17: Recap We ve now spent some time looking closely at coherence and how to assign uncertainties to coherence. Can we think about coherence in a different way? There are

More information

Linear Regression Linear Regression with Shrinkage

Linear Regression Linear Regression with Shrinkage Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle

More information

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary

More information

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17 Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 17 Outline Filters and Rotations Generating co-varying random fields Translating co-varying fields into

More information

Designing Information Devices and Systems II Fall 2018 Elad Alon and Miki Lustig Homework 9

Designing Information Devices and Systems II Fall 2018 Elad Alon and Miki Lustig Homework 9 EECS 16B Designing Information Devices and Systems II Fall 18 Elad Alon and Miki Lustig Homework 9 This homework is due Wednesday, October 31, 18, at 11:59pm. Self grades are due Monday, November 5, 18,

More information

Frequentist-Bayesian Model Comparisons: A Simple Example

Frequentist-Bayesian Model Comparisons: A Simple Example Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal

More information

Covariance to PCA. CS 510 Lecture #14 February 23, 2018

Covariance to PCA. CS 510 Lecture #14 February 23, 2018 Covariance to PCA CS 510 Lecture 14 February 23, 2018 Overview: Goal Assume you have a gallery (database) of images, and a probe (test) image. The goal is to find the database image that is most similar

More information

MLCC 2015 Dimensionality Reduction and PCA

MLCC 2015 Dimensionality Reduction and PCA MLCC 2015 Dimensionality Reduction and PCA Lorenzo Rosasco UNIGE-MIT-IIT June 25, 2015 Outline PCA & Reconstruction PCA and Maximum Variance PCA and Associated Eigenproblem Beyond the First Principal Component

More information

COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017

COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University PRINCIPAL COMPONENT ANALYSIS DIMENSIONALITY

More information

Systematic strategies for real time filtering of turbulent signals in complex systems

Systematic strategies for real time filtering of turbulent signals in complex systems Systematic strategies for real time filtering of turbulent signals in complex systems Statistical inversion theory for Gaussian random variables The Kalman Filter for Vector Systems: Reduced Filters and

More information

Theorems. Least squares regression

Theorems. Least squares regression Theorems In this assignment we are trying to classify AML and ALL samples by use of penalized logistic regression. Before we indulge on the adventure of classification we should first explain the most

More information

Signal Analysis. Principal Component Analysis

Signal Analysis. Principal Component Analysis Multi dimensional Signal Analysis Lecture 2E Principal Component Analysis Subspace representation Note! Given avector space V of dimension N a scalar product defined by G 0 a subspace U of dimension M

More information

Principal Components Analysis

Principal Components Analysis Principal Components Analysis Santiago Paternain, Aryan Mokhtari and Alejandro Ribeiro March 29, 2018 At this point we have already seen how the Discrete Fourier Transform and the Discrete Cosine Transform

More information

Application of PCA to Hyperspectral Data for Noise Characterization, Noise Filtering and Instrument Monitoring

Application of PCA to Hyperspectral Data for Noise Characterization, Noise Filtering and Instrument Monitoring Application of PCA to Hyperspectral Data for Noise Characterization, Noise Filtering and Instrument Monitoring Paolo Antonelli, Dave Tobin, Hank Revercomb, Dave Turner, Steve Dutcher, Ben Howell, Bob Knuteson,

More information

Regularizing inverse problems. Damping and smoothing and choosing...

Regularizing inverse problems. Damping and smoothing and choosing... Regularizing inverse problems Damping and smoothing and choosing... 141 Regularization The idea behind SVD is to limit the degree of freedom in the model and fit the data to an acceptable level. Retain

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

(a)

(a) Chapter 8 Subspace Methods 8. Introduction Principal Component Analysis (PCA) is applied to the analysis of time series data. In this context we discuss measures of complexity and subspace methods for

More information

Modeling Classes of Shapes Suppose you have a class of shapes with a range of variations: System 2 Overview

Modeling Classes of Shapes Suppose you have a class of shapes with a range of variations: System 2 Overview 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 Modeling Classes of Shapes Suppose you have a class of shapes with a range of variations: System processes System Overview Previous Systems:

More information

Hale Collage. Spectropolarimetric Diagnostic Techniques!!!!!!!! Rebecca Centeno

Hale Collage. Spectropolarimetric Diagnostic Techniques!!!!!!!! Rebecca Centeno Hale Collage Spectropolarimetric Diagnostic Techniques Rebecca Centeno March 1-8, 2016 Today Simple solutions to the RTE Milne-Eddington approximation LTE solution The general inversion problem Spectral

More information

Computer Vision Group Prof. Daniel Cremers. 3. Regression

Computer Vision Group Prof. Daniel Cremers. 3. Regression Prof. Daniel Cremers 3. Regression Categories of Learning (Rep.) Learnin g Unsupervise d Learning Clustering, density estimation Supervised Learning learning from a training data set, inference on the

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

LECTURE 16: PCA AND SVD

LECTURE 16: PCA AND SVD Instructor: Sael Lee CS549 Computational Biology LECTURE 16: PCA AND SVD Resource: PCA Slide by Iyad Batal Chapter 12 of PRML Shlens, J. (2003). A tutorial on principal component analysis. CONTENT Principal

More information

Fitting functions to data

Fitting functions to data 1 Fitting functions to data 1.1 Exact fitting 1.1.1 Introduction Suppose we have a set of real-number data pairs x i, y i, i = 1, 2,, N. These can be considered to be a set of points in the xy-plane. They

More information

Dimensionality Reduction and Principle Components Analysis

Dimensionality Reduction and Principle Components Analysis Dimensionality Reduction and Principle Components Analysis 1 Outline What is dimensionality reduction? Principle Components Analysis (PCA) Example (Bishop, ch 12) PCA vs linear regression PCA as a mixture

More information

Emission Limb sounders (MIPAS)

Emission Limb sounders (MIPAS) Emission Limb sounders (MIPAS) Bruno Carli ENVISAT ATMOSPHERIC PACKAGE MIPAS Michelson Interferometric Passive Atmospheric Sounder GOMOS Global Ozone Monitoring by Occultation of Stars SCIAMACHY Scanning

More information

What is Image Deblurring?

What is Image Deblurring? What is Image Deblurring? When we use a camera, we want the recorded image to be a faithful representation of the scene that we see but every image is more or less blurry, depending on the circumstances.

More information

Lecture VIII Dim. Reduction (I)

Lecture VIII Dim. Reduction (I) Lecture VIII Dim. Reduction (I) Contents: Subset Selection & Shrinkage Ridge regression, Lasso PCA, PCR, PLS Lecture VIII: MLSC - Dr. Sethu Viayakumar Data From Human Movement Measure arm movement and

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

Optimum Sampling Vectors for Wiener Filter Noise Reduction

Optimum Sampling Vectors for Wiener Filter Noise Reduction 58 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 1, JANUARY 2002 Optimum Sampling Vectors for Wiener Filter Noise Reduction Yukihiko Yamashita, Member, IEEE Absact Sampling is a very important and

More information

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection

More information

5 Linear Algebra and Inverse Problem

5 Linear Algebra and Inverse Problem 5 Linear Algebra and Inverse Problem 5.1 Introduction Direct problem ( Forward problem) is to find field quantities satisfying Governing equations, Boundary conditions, Initial conditions. The direct problem

More information

A principal component noise filter for high spectral resolution infrared measurements

A principal component noise filter for high spectral resolution infrared measurements JOURNAL OF GEOPHYSICAL RESEARCH, VOL. 109,, doi:10.1029/2004jd004862, 2004 A principal component noise filter for high spectral resolution infrared measurements P. Antonelli, H. E. Revercomb, and L. A.

More information

Principal components analysis COMS 4771

Principal components analysis COMS 4771 Principal components analysis COMS 4771 1. Representation learning Useful representations of data Representation learning: Given: raw feature vectors x 1, x 2,..., x n R d. Goal: learn a useful feature

More information

Principal Component Analysis (PCA) CSC411/2515 Tutorial

Principal Component Analysis (PCA) CSC411/2515 Tutorial Principal Component Analysis (PCA) CSC411/2515 Tutorial Harris Chan Based on previous tutorial slides by Wenjie Luo, Ladislav Rampasek University of Toronto hchan@cs.toronto.edu October 19th, 2017 (UofT)

More information

Validation of Complex Data Assimilation Methods

Validation of Complex Data Assimilation Methods Fairmode Technical Meeting 24.-25.6.2015, DAO, Univ. Aveiro Validation of Complex Data Assimilation Methods Hendrik Elbern, Elmar Friese, Nadine Goris, Lars Nieradzik and many others Rhenish Institute

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Consider the joint probability, P(x,y), shown as the contours in the figure above. P(x) is given by the integral of P(x,y) over all values of y.

Consider the joint probability, P(x,y), shown as the contours in the figure above. P(x) is given by the integral of P(x,y) over all values of y. ATMO/OPTI 656b Spring 009 Bayesian Retrievals Note: This follows the discussion in Chapter of Rogers (000) As we have seen, the problem with the nadir viewing emission measurements is they do not contain

More information

Joint Temperature, Humidity, and Sea Surface Temperature Retrieval from IASI Sensor Data

Joint Temperature, Humidity, and Sea Surface Temperature Retrieval from IASI Sensor Data Joint Temperature, Humidity, and Sea Surface Temperature Retrieval from IASI Sensor Data Marc Schwaerz and Gottfried Kirchengast Institute for Geophysics, Astrophysics, and Meteorology (IGAM), University

More information

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection Instructor: Herke van Hoof (herke.vanhoof@cs.mcgill.ca) Based on slides by:, Jackie Chi Kit Cheung Class web page:

More information

c Springer, Reprinted with permission.

c Springer, Reprinted with permission. Zhijian Yuan and Erkki Oja. A FastICA Algorithm for Non-negative Independent Component Analysis. In Puntonet, Carlos G.; Prieto, Alberto (Eds.), Proceedings of the Fifth International Symposium on Independent

More information

Linear Regression Linear Regression with Shrinkage

Linear Regression Linear Regression with Shrinkage Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle

More information

Image Registration Lecture 2: Vectors and Matrices

Image Registration Lecture 2: Vectors and Matrices Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this

More information

Chapter 4: Factor Analysis

Chapter 4: Factor Analysis Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.

More information

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26 Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1

More information

Noise & Data Reduction

Noise & Data Reduction Noise & Data Reduction Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum Dimension Reduction 1 Remember: Central Limit

More information

Inverse problem and optimization

Inverse problem and optimization Inverse problem and optimization Laurent Condat, Nelly Pustelnik CNRS, Gipsa-lab CNRS, Laboratoire de Physique de l ENS de Lyon Decembre, 15th 2016 Inverse problem and optimization 2/36 Plan 1. Examples

More information

Homework 1. Yuan Yao. September 18, 2011

Homework 1. Yuan Yao. September 18, 2011 Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to

More information

Two-dimensional characterization of atmospheric prole retrievals from limb sounding observations

Two-dimensional characterization of atmospheric prole retrievals from limb sounding observations Journal of Quantitative Spectroscopy & Radiative Transfer 86 (2004) 45 71 www.elsevier.com/locate/jqsrt Two-dimensional characterization of atmospheric prole retrievals from limb sounding observations

More information

A few applications of the SVD

A few applications of the SVD A few applications of the SVD Many methods require to approximate the original data (matrix) by a low rank matrix before attempting to solve the original problem Regularization methods require the solution

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Introduction Consider a zero mean random vector R n with autocorrelation matri R = E( T ). R has eigenvectors q(1),,q(n) and associated eigenvalues λ(1) λ(n). Let Q = [ q(1)

More information

Chemometrics. Matti Hotokka Physical chemistry Åbo Akademi University

Chemometrics. Matti Hotokka Physical chemistry Åbo Akademi University Chemometrics Matti Hotokka Physical chemistry Åbo Akademi University Linear regression Experiment Consider spectrophotometry as an example Beer-Lamberts law: A = cå Experiment Make three known references

More information

Estimation of Mars surface physical properties from hyperspectral images using the SIR method

Estimation of Mars surface physical properties from hyperspectral images using the SIR method Estimation of Mars surface physical properties from hyperspectral images using the SIR method Caroline Bernard-Michel, Sylvain Douté, Laurent Gardes and Stéphane Girard Source: ESA Outline I. Context Hyperspectral

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

UWPHYSRET an SSEC inversion package for high resolution infrared data based on LBLRTM.

UWPHYSRET an SSEC inversion package for high resolution infrared data based on LBLRTM. UWPHYSRET an SSEC inversion package for high resolution infrared data based on LBLRTM. Paolo Antonelli, R. Knuteson, R. Garcia, S. Bedka, D. Tobin, J.Taylor, W. Smith, and H. Revercomb QuickTime and a

More information

1 Linearity and Linear Systems

1 Linearity and Linear Systems Mathematical Tools for Neuroscience (NEU 34) Princeton University, Spring 26 Jonathan Pillow Lecture 7-8 notes: Linear systems & SVD Linearity and Linear Systems Linear system is a kind of mapping f( x)

More information

Principal Component Analysis

Principal Component Analysis B: Chapter 1 HTF: Chapter 1.5 Principal Component Analysis Barnabás Póczos University of Alberta Nov, 009 Contents Motivation PCA algorithms Applications Face recognition Facial expression recognition

More information

Modelling temporal structure (in noise and signal)

Modelling temporal structure (in noise and signal) Modelling temporal structure (in noise and signal) Mark Woolrich, Christian Beckmann*, Salima Makni & Steve Smith FMRIB, Oxford *Imperial/FMRIB temporal noise: modelling temporal autocorrelation temporal

More information

HST.582J/6.555J/16.456J

HST.582J/6.555J/16.456J Blind Source Separation: PCA & ICA HST.582J/6.555J/16.456J Gari D. Clifford gari [at] mit. edu http://www.mit.edu/~gari G. D. Clifford 2005-2009 What is BSS? Assume an observation (signal) is a linear

More information

Ground and On-Orbit Characterization and Calibration of the Geosynchronous Imaging Fourier Transform Spectrometer (GIFTS)

Ground and On-Orbit Characterization and Calibration of the Geosynchronous Imaging Fourier Transform Spectrometer (GIFTS) Ground and On-Orbit Characterization and Calibration of the Geosynchronous Imaging Fourier Transform Spectrometer (GIFTS) John D. Elwell 1, Deron K. Scott 1 Henry E. Revercomb 2, Fred A. Best 2, Robert

More information

Synergistic Cloud Clearing Using Aqua Sounding and Imaging Infrared Measurements

Synergistic Cloud Clearing Using Aqua Sounding and Imaging Infrared Measurements Synergistic Cloud Clearing Using Aqua Sounding and Imaging Infrared Measurements H-L Allen Huang, et al. CIMSS/SSEC Univ. of Wisconsin-Madison and William L. Smith, LaRC, NASA Cloud Clearing/Cloud Detection

More information

BINGO simulations and updates on the performance of. the instrument

BINGO simulations and updates on the performance of. the instrument BINGO simulations and updates on the performance of BINGO telescope the instrument M.-A. Bigot-Sazy BINGO collaboration Paris 21cm Intensity Mapping Workshop June 2014 21cm signal Observed sky Credit:

More information

Computational Methods. Eigenvalues and Singular Values

Computational Methods. Eigenvalues and Singular Values Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations

More information