Application of Principal Component Analysis to TES data
|
|
- Isaac Woods
- 5 years ago
- Views:
Transcription
1 Application of Principal Component Analysis to TES data Clive D Rodgers Clarendon Laboratory University of Oxford Madison, Wisconsin, 27th April
2 My take on the PCA business 2/41
3 What is the best estimate of a spectrum given many measured spectra? Simple approach using optimal estimation: Forward model is y=x+ ε x is true spectrum, y is measurement, ε is noise Maximum a posteriori estimate of x is x r = x a + S a (S a + S ε ) -1 (y - x a ) 3/41
4 Continued If we have a large sample of spectra: Expect that x a = <x + ε> = <y> Can estimate S a +S ε from statistics of y Should have a good idea of S ε a priori But: S a (S a +S ε ) -1 will be a large matrix S a found from S a +S ε and S ε is likely to be ill-conditioned S a is likely to have a small number of eigenvalues greater than noise 4/41
5 Singular Vectors (or Principal Components) Let the ensemble of (y -<y>) be N columns of a matrix Y Represent Y as its singular vector decomposition: Y = UΛV T where Λ is diagonal, U T U=I and V T V=I The j th individual spectrum y j is then y j = <y> + Σ i u i λ i v ij T The spectrum is represented as a sum of columns u i of U, with coefficients λ i v ijt. Because U T U=I, we can compute λ i v ijt for any spectrum as U T (y j -<y>). 5/41
6 What do we expect? 1 / N YY T is the covariance matrix S y of the spectra Left singular vectors U are the same as eigenvectors of YY T, singular values are the square roots of its eigenvalues In the linear case with independent constant noise, S y would be S y = S a + σ ε 2 I S a has rank n, I is of dimension m >> n, where n is the degrees of freedom of the atmosphere The eigenvalues of S y are λ i2 /N The eigenvalues of S a should be λ i2 /N - σ ε 2 6/41
7 Reconstructing Spectra We can drop terms with λ i2 /N ~ σ ε2, i.e. i > n, without significant loss they correspond to noise only Better, multiply retained terms by something like (λ i2 - N σ ε 2)/ λ i 2 So spectra can be reconstructed from the first few coefficients. The noise can be reconstructed from the rest Reconstructed spectra have much reduced noise 7/41
8 Information Content The Shannon information content of a single spectrum relative to the ensemble is H = 1 / 2 i ln(λ i 2 /N σ ε2 ) The degrees of freedom for signal is d s = i (1 - N σ ε2 /λ i2 ) But reality isn t quite like that... 8/41
9 Tropospheric Emission Spectrometer Connes-type four-port Fourier transform spectrometer Nadir view: 0.06 cm -1 Limb view: cm element detector arrays in 4 focal planes: 1A ( ), 1B ( ), 2A ( ), 2B ( ) 0.5 X 5 km nadir; 2.3 X 23 km limb Global Survey mode: Along track nadir, 216 views/orbit. 9/41
10 SVD Applications Information content analysis Validation Instrument performance Forward model De-noised spectra Retrieval? 10/41
11 A typical example Run Sept 2004? Channel 1B cm views * 16 detectors 11/41
12 An example of Information Content from Channel 1B2 Using one orbit, 1152 spectra each with 3951 elements cm -1 n Lambda Information d.f.s cum. cum Total: 14.5 bits 7.05 d.f. Sing. Vals. Residual Fit to noise asymptote SV number /41
13 How do you fit noise to the asymptote? I havn t used enough spectra Expected distribution of singular values of independent gaussian noise for a finite number of samples Simulation Theory: there is an explicit formula Fit expected distribution to the long tail 13/41
14 Validation an early example Run Sept 2004? Channel 1B cm views * 16 detectors 14/41
15 Singular Vectors 15/41
16 Singular vectors * Lambda 16/41
17 V-vectors 17/41
18 Features Most of variation is in the first singular vector. First six are: Data spikes - identified Data spikes - unidentified Pixel-dependent variation in the spectra Scan-direction dependence 18/41
19 Singular Vector 6 Systematic variation across the detector array Must be an artefact Suggests systematic error in ILS How is it related to mean spectrum? Least squares fit to find function that when convolved with mean spectrum gives SV6 19/41
20 SV6 20/41
21 SV6 21/41
22 SV6 Suggests the derivative of the ILS 22/41
23 Curious behaviour of the singular values of an ensemble of B1 spectra What is this? This is the atmospheric variability cm -1 23/41
24 Features It occurs at around SV 200. I have used 100 sequences, 1600 spectra. Implies two of these SV s per sequence. Confirmed by trying other numbers of sequences The vectors look like noise Only in 2B1 does this 24/41
25 With 64 sequences, plotted in 9/04 The shoulder here is at 64 there was only 1 per sequence then 25/41
26 Examine pixel noise Look at the spectrum for each pixel minus the mean over all pixels removes atmospheric structure, but differences will be slightly correlated Singular vectors of the difference: look like noise Coefficients of the singular vectors for one set of 16: don t look like noise for two of the SV s Covariance & correlation matrices 26/41
27 SV coefficients 27/41
28 2B1 noise is significantly correlated between pixels 28/41
29 Deductions The change from 1 to 2 SVs per sequence implies something in the processing rather than the instrument Each pixel has its own independent noise Plus noise correlated between all the pixels of each sequence The correlations depend on distance between the pixels, quasi-sinusoidally Not an odd pixel/even pixel effect 29/41
30 SV s of Residual Spectra Difference between measured spectrum and calculated from retrieval No just in the microwindows used Courtesy Reinhard Beer & Susan S. Kulawik 30/41
31 1.0e-7 Mean Residual Radiance for Run 2147 (2004 September 21) from 220 Spectra between 30 o S & 30 o N 5.0e-8 Residual Radiance, watts/cm 2 /sr/cm e-8-1.0e-7-1.5e-7-2.0e-7-2.5e-7 Residual O 3 Residual Silicate Surface Feature "Over-fitted" HDO Unmodeled F12 Residual CH 4-3.0e-7 Retrieval Microwindows -3.5e Frequency, cm -1 31/41
32 Singular Values of Run 2147 Full-Filter Residuals ("W" Matrix Diagonal; First 51 Elements of 568 only) 3.0e-5 2.5e-5 2.0e-5 Singular Value 1.5e-5 1.0e-5 5.0e Singular Value Number 32/41
33 0.08 U Singular Vector Amplitude The quasi-periodicity is due to the orbital variation of total radiance Index (Time order) V Singular Vector Amplitude Note the strong resemblance to the grand average Frequency, cm -1 33
34 0.15 U Singular Vector 1 Amplitude This pattern of roughly orbits has been observed in other diagnostics. The cause is under investigation Index (Time order) V Singular Vector 1 Amplitude The discontinuity at 1120 cm -1 occurs at the overlap of two different filters Frequency, cm -1 34/41
35 Amplitude U Singular Vector 2 Positive peaks align with land scenes (green bars) Index (Time order) Amplitude V Singular Vector Broad feature is a component of the unmodeled surface silicate emissivity Frequency, cm -1 35/41
36 Retrieval some untested thoughts Possible methods include: Retrieve from denoised spectrum in the same way as usual, but with noise covariance reduced in some ad-hoc way not optimal, inefficient Ditto, but with the correct error covariance covariance matrix is singular Select a subset of channels according to information content straightforward if linear, not otherwise Retrieve from SV representation coefficients needs a special forward model for efficiency PCRTM or a regression or neural net method 36/41
37 Characterising a denoised spectrum Reconstruct y with the first r singular vectors, U r, giving y r y r = U r U rt y Measurement function is: y r = U r U rt F(x) + U r U rt ε We assume that U r U rt F(x) ~ F(x), so y r = F(x) + U r U rt ε = F(x) + ε r Covariance of ε r is S εr = U r U rt S ε U r U r T which is of rank r 37/41
38 Retrieve from a denoised spectrum We can use all or some of the spectrum. Microwindows Channels selected by information content Channels selected by Xu Liu s approach Define a selection operator M, and retrieve from the m element subset y m = My r The subset depends on the whole of the original spectrum, y m = MU r U rt y, but we assume that we can model it as MF(x). The error covariance of ε m is S m = M U r U rt S ε U r U rt M T which may be singular if m > r. 38/41
39 Do another SVD Use the SVD of MU r U rt = LΛR T to give L T y m = L T MF(x) + ΛR T ε Drop elements of L T y m with zero (or small) singular values, leaving p=min(r,m) elements at most. Gives L pt y m Retrieve from the rest of L pt y m with L pt MF(x) as the forward model. L pt can be precomputed. Error covariance is Λ p R pt S ε R p Λ p. If S ε = σ 2 I, this reduces to σ 2 Λ p2. 39/41
40 Comments This allows us to retrieve with the minimum number of evaluations of the forward model And the minimum length of measurement vector With very little extra matrix manipulation I havn t tried it yet, so I don t know what the catch is 40/41
41 Summary/Comments PCA does not improve the information content of a measurement But it does allow us to use it more efficiently Very useful for validation - separates out independent sources of variation, e.g. artefacts, omitted physics,... Denoised spectra guide the eye to real features that may otherwise not be seen Data compression 41/41
Inverse Theory. COST WaVaCS Winterschool Venice, February Stefan Buehler Luleå University of Technology Kiruna
Inverse Theory COST WaVaCS Winterschool Venice, February 2011 Stefan Buehler Luleå University of Technology Kiruna Overview Inversion 1 The Inverse Problem 2 Simple Minded Approach (Matrix Inversion) 3
More informationRetrieval Algorithm Using Super channels
Retrieval Algorithm Using Super channels Xu Liu NASA Langley Research Center, Hampton VA 23662 D. K. Zhou, A. M. Larar (NASA LaRC) W. L. Smith (HU and UW) P. Schluessel (EUMETSAT) Hank Revercomb (UW) Jonathan
More informationIASI PC compression Searching for signal in the residuals. Thomas August, Nigel Atkinson, Fiona Smith
IASI PC compression Searching for signal in the residuals Tim.Hultberg@EUMETSAT.INT Thomas August, Nigel Atkinson, Fiona Smith Raw radiance (minus background) Reconstructed radiance (minus background)
More informationCompression of IASI Data and Representation of FRTM in EOF Space
Compression of IASI Data and Representation of FRTM in EOF Space Peter Schlüssel EUMETSAT Am Kavalleriesand 31 64295 Darmstadt Germany Anthony C. L. Lee Met Office FitzRoy Road Exeter, Devon EX1 3PB United
More informationLinear Algebra & Geometry why is linear algebra useful in computer vision?
Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia
More informationCOMBINED OZONE RETRIEVAL USING THE MICHELSON INTERFEROMETER FOR PASSIVE ATMOSPHERIC SOUNDING (MIPAS) AND THE TROPOSPHERIC EMISSION SPECTROMETER (TES)
COMBINED OZONE RETRIEVAL USING THE MICHELSON INTERFEROMETER FOR PASSIVE ATMOSPHERIC SOUNDING (MIPAS) AND THE TROPOSPHERIC EMISSION SPECTROMETER (TES) Claire Waymark, Anu Dudhia, John Barnett, Fred Taylor,
More informationLinear Algebra & Geometry why is linear algebra useful in computer vision?
Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia
More informationMachine Learning (Spring 2012) Principal Component Analysis
1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in
More informationPrincipal Component Analysis (PCA) of AIRS Data
Principal Component Analysis (PCA) of AIRS Data Mitchell D. Goldberg 1, Lihang Zhou 2, Walter Wolf 2 and Chris Barnet 1 NOAA/NESDIS/Office of Research and Applications, Camp Springs, MD 1 QSS Group Inc.
More informationSatellite data information content, channel selection and density
Satellite data information content, channel selection and density Élisabeth Gérard Météo-France/CNRM/GMAP Toulouse, France [EG2], 2 nd ENVISAT summer school Frascati, Italy, 16-26 Aug 2004 1 Overview A.
More informationRetrievals of methane from IASI radiance spectra and comparisons with ground-based FTIR measurements
ACCENT-AT2 follow-up meeting, 22-23 June 2009, Mainz, Germany Retrievals of methane from IASI radiance spectra and comparisons with ground-based FTIR measurements Tobias Kerzenmacher 1, Nicolas Kumps 1,
More informationDimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes. October 3, Statistics 202: Data Mining
Dimension reduction, PCA & eigenanalysis Based in part on slides from textbook, slides of Susan Holmes October 3, 2012 1 / 1 Combinations of features Given a data matrix X n p with p fairly large, it can
More informationECE295, Data Assimila0on and Inverse Problems, Spring 2015
ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch
More informationPrincipal Component Analysis CS498
Principal Component Analysis CS498 Today s lecture Adaptive Feature Extraction Principal Component Analysis How, why, when, which A dual goal Find a good representation The features part Reduce redundancy
More informationIntercomparison of remote sounding instruments
JOURNAL OF GEOPHYSICAL RESEARCH, VOL. 108, NO. D3, 4116, doi:10.1029/2002jd002299, 2003 Intercomparison of remote sounding instruments Clive D. Rodgers 1 and Brian J. Connor National Institute for Water
More informationECE 501b Homework #6 Due: 11/26
ECE 51b Homework #6 Due: 11/26 1. Principal Component Analysis: In this assignment, you will explore PCA as a technique for discerning whether low-dimensional structure exists in a set of data and for
More informationDimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas
Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx
More informationCIMSS Hyperspectral IR Sounding Retrieval (CHISR) Processing & Applications
CIMSS Hyperspectral IR Sounding Retrieval (CHISR) Processing & Applications Jun Li @, Elisabeth Weisz @, Jinlong Li @, Hui Liu #, Timothy J. Schmit &, Jason Otkin @ and many other CIMSS collaborators @Cooperative
More informationSingular value decomposition. If only the first p singular values are nonzero we write. U T o U p =0
Singular value decomposition If only the first p singular values are nonzero we write G =[U p U o ] " Sp 0 0 0 # [V p V o ] T U p represents the first p columns of U U o represents the last N-p columns
More informationExercise Sheet 1. 1 Probability revision 1: Student-t as an infinite mixture of Gaussians
Exercise Sheet 1 1 Probability revision 1: Student-t as an infinite mixture of Gaussians Show that an infinite mixture of Gaussian distributions, with Gamma distributions as mixing weights in the following
More informationLinear Subspace Models
Linear Subspace Models Goal: Explore linear models of a data set. Motivation: A central question in vision concerns how we represent a collection of data vectors. The data vectors may be rasterized images,
More informationIntroduction to Machine Learning
10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what
More informationRegression Retrieval Overview. Larry McMillin
Regression Retrieval Overview Larry McMillin Climate Research and Applications Division National Environmental Satellite, Data, and Information Service Washington, D.C. Larry.McMillin@noaa.gov Pick one
More informationGopalkrishna Veni. Project 4 (Active Shape Models)
Gopalkrishna Veni Project 4 (Active Shape Models) Introduction Active shape Model (ASM) is a technique of building a model by learning the variability patterns from training datasets. ASMs try to deform
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More informationPrincipal Components Analysis (PCA)
Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More informationCS540 Machine learning Lecture 5
CS540 Machine learning Lecture 5 1 Last time Basis functions for linear regression Normal equations QR SVD - briefly 2 This time Geometry of least squares (again) SVD more slowly LMS Ridge regression 3
More informationSystem 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:
System 2 : Modelling & Recognising Modelling and Recognising Classes of Classes of Shapes Shape : PDM & PCA All the same shape? System 1 (last lecture) : limited to rigidly structured shapes System 2 :
More informationApplication of PCA to IASI: An NWP Perspective. Andrew Collard, ECMWF. Acknowledgements to Tony McNally, Jean-Noel Thépaut
Application of PCA to IASI: An NWP Perspective Andrew Collard, ECMWF Acknowledgements to Tony McNally, Jean-Noel Thépaut Overview Introduction Why use PCA/RR? Expected performance of PCA/RR for IASI Assimilation
More informationRetrieval problems in Remote Sensing
Maarten Krol August 2006 m.c.krol@phys.uu.nl 1. Goal To show that remote sensing is an indirect measurement, i.e. that we do not measure directly, but by means of (for instance) electromagnetic radiation
More informationUnsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent
Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:
More informationCS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS
CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS LAST TIME Intro to cudnn Deep neural nets using cublas and cudnn TODAY Building a better model for image classification Overfitting
More informationSingular Value Decompsition
Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost
More informationThe Retrieval of Infrared Cooling Rate Profiles from Thermal Infrared Remote Sounder Radiance Measurements
The Retrieval of Infrared Cooling Rate Profiles from Thermal Infrared Remote Sounder Radiance Measurements Daniel Feldman Candidacy Presentation May 13, 2005 Outline 1. Background 2. Motivation 3. Cooling
More informationAnnouncements (repeat) Principal Components Analysis
4/7/7 Announcements repeat Principal Components Analysis CS 5 Lecture #9 April 4 th, 7 PA4 is due Monday, April 7 th Test # will be Wednesday, April 9 th Test #3 is Monday, May 8 th at 8AM Just hour long
More informationCovariance to PCA. CS 510 Lecture #8 February 17, 2014
Covariance to PCA CS 510 Lecture 8 February 17, 2014 Status Update Programming Assignment 2 is due March 7 th Expect questions about your progress at the start of class I still owe you Assignment 1 back
More informationData Mining Techniques
Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 12 Jan-Willem van de Meent (credit: Yijun Zhao, Percy Liang) DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Linear Dimensionality
More informationLecture 17: variance in a band = log(s xx (f)) df (2) If we want to plot something that is more directly representative of variance, we can try this:
UCSD SIOC 221A: (Gille) 1 Lecture 17: Recap We ve now spent some time looking closely at coherence and how to assign uncertainties to coherence. Can we think about coherence in a different way? There are
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationDATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD
DATA MINING LECTURE 8 Dimensionality Reduction PCA -- SVD The curse of dimensionality Real data usually have thousands, or millions of dimensions E.g., web documents, where the dimensionality is the vocabulary
More informationPrincipal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17
Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 17 Outline Filters and Rotations Generating co-varying random fields Translating co-varying fields into
More informationDesigning Information Devices and Systems II Fall 2018 Elad Alon and Miki Lustig Homework 9
EECS 16B Designing Information Devices and Systems II Fall 18 Elad Alon and Miki Lustig Homework 9 This homework is due Wednesday, October 31, 18, at 11:59pm. Self grades are due Monday, November 5, 18,
More informationFrequentist-Bayesian Model Comparisons: A Simple Example
Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal
More informationCovariance to PCA. CS 510 Lecture #14 February 23, 2018
Covariance to PCA CS 510 Lecture 14 February 23, 2018 Overview: Goal Assume you have a gallery (database) of images, and a probe (test) image. The goal is to find the database image that is most similar
More informationMLCC 2015 Dimensionality Reduction and PCA
MLCC 2015 Dimensionality Reduction and PCA Lorenzo Rosasco UNIGE-MIT-IIT June 25, 2015 Outline PCA & Reconstruction PCA and Maximum Variance PCA and Associated Eigenproblem Beyond the First Principal Component
More informationCOMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017
COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University PRINCIPAL COMPONENT ANALYSIS DIMENSIONALITY
More informationSystematic strategies for real time filtering of turbulent signals in complex systems
Systematic strategies for real time filtering of turbulent signals in complex systems Statistical inversion theory for Gaussian random variables The Kalman Filter for Vector Systems: Reduced Filters and
More informationTheorems. Least squares regression
Theorems In this assignment we are trying to classify AML and ALL samples by use of penalized logistic regression. Before we indulge on the adventure of classification we should first explain the most
More informationSignal Analysis. Principal Component Analysis
Multi dimensional Signal Analysis Lecture 2E Principal Component Analysis Subspace representation Note! Given avector space V of dimension N a scalar product defined by G 0 a subspace U of dimension M
More informationPrincipal Components Analysis
Principal Components Analysis Santiago Paternain, Aryan Mokhtari and Alejandro Ribeiro March 29, 2018 At this point we have already seen how the Discrete Fourier Transform and the Discrete Cosine Transform
More informationApplication of PCA to Hyperspectral Data for Noise Characterization, Noise Filtering and Instrument Monitoring
Application of PCA to Hyperspectral Data for Noise Characterization, Noise Filtering and Instrument Monitoring Paolo Antonelli, Dave Tobin, Hank Revercomb, Dave Turner, Steve Dutcher, Ben Howell, Bob Knuteson,
More informationRegularizing inverse problems. Damping and smoothing and choosing...
Regularizing inverse problems Damping and smoothing and choosing... 141 Regularization The idea behind SVD is to limit the degree of freedom in the model and fit the data to an acceptable level. Retain
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More information(a)
Chapter 8 Subspace Methods 8. Introduction Principal Component Analysis (PCA) is applied to the analysis of time series data. In this context we discuss measures of complexity and subspace methods for
More informationModeling Classes of Shapes Suppose you have a class of shapes with a range of variations: System 2 Overview
4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 4 4 4 6 Modeling Classes of Shapes Suppose you have a class of shapes with a range of variations: System processes System Overview Previous Systems:
More informationHale Collage. Spectropolarimetric Diagnostic Techniques!!!!!!!! Rebecca Centeno
Hale Collage Spectropolarimetric Diagnostic Techniques Rebecca Centeno March 1-8, 2016 Today Simple solutions to the RTE Milne-Eddington approximation LTE solution The general inversion problem Spectral
More informationComputer Vision Group Prof. Daniel Cremers. 3. Regression
Prof. Daniel Cremers 3. Regression Categories of Learning (Rep.) Learnin g Unsupervise d Learning Clustering, density estimation Supervised Learning learning from a training data set, inference on the
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationLECTURE 16: PCA AND SVD
Instructor: Sael Lee CS549 Computational Biology LECTURE 16: PCA AND SVD Resource: PCA Slide by Iyad Batal Chapter 12 of PRML Shlens, J. (2003). A tutorial on principal component analysis. CONTENT Principal
More informationFitting functions to data
1 Fitting functions to data 1.1 Exact fitting 1.1.1 Introduction Suppose we have a set of real-number data pairs x i, y i, i = 1, 2,, N. These can be considered to be a set of points in the xy-plane. They
More informationDimensionality Reduction and Principle Components Analysis
Dimensionality Reduction and Principle Components Analysis 1 Outline What is dimensionality reduction? Principle Components Analysis (PCA) Example (Bishop, ch 12) PCA vs linear regression PCA as a mixture
More informationEmission Limb sounders (MIPAS)
Emission Limb sounders (MIPAS) Bruno Carli ENVISAT ATMOSPHERIC PACKAGE MIPAS Michelson Interferometric Passive Atmospheric Sounder GOMOS Global Ozone Monitoring by Occultation of Stars SCIAMACHY Scanning
More informationWhat is Image Deblurring?
What is Image Deblurring? When we use a camera, we want the recorded image to be a faithful representation of the scene that we see but every image is more or less blurry, depending on the circumstances.
More informationLecture VIII Dim. Reduction (I)
Lecture VIII Dim. Reduction (I) Contents: Subset Selection & Shrinkage Ridge regression, Lasso PCA, PCR, PLS Lecture VIII: MLSC - Dr. Sethu Viayakumar Data From Human Movement Measure arm movement and
More informationMachine learning for pervasive systems Classification in high-dimensional spaces
Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version
More informationOptimum Sampling Vectors for Wiener Filter Noise Reduction
58 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 1, JANUARY 2002 Optimum Sampling Vectors for Wiener Filter Noise Reduction Yukihiko Yamashita, Member, IEEE Absact Sampling is a very important and
More informationGI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil
GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection
More information5 Linear Algebra and Inverse Problem
5 Linear Algebra and Inverse Problem 5.1 Introduction Direct problem ( Forward problem) is to find field quantities satisfying Governing equations, Boundary conditions, Initial conditions. The direct problem
More informationA principal component noise filter for high spectral resolution infrared measurements
JOURNAL OF GEOPHYSICAL RESEARCH, VOL. 109,, doi:10.1029/2004jd004862, 2004 A principal component noise filter for high spectral resolution infrared measurements P. Antonelli, H. E. Revercomb, and L. A.
More informationPrincipal components analysis COMS 4771
Principal components analysis COMS 4771 1. Representation learning Useful representations of data Representation learning: Given: raw feature vectors x 1, x 2,..., x n R d. Goal: learn a useful feature
More informationPrincipal Component Analysis (PCA) CSC411/2515 Tutorial
Principal Component Analysis (PCA) CSC411/2515 Tutorial Harris Chan Based on previous tutorial slides by Wenjie Luo, Ladislav Rampasek University of Toronto hchan@cs.toronto.edu October 19th, 2017 (UofT)
More informationValidation of Complex Data Assimilation Methods
Fairmode Technical Meeting 24.-25.6.2015, DAO, Univ. Aveiro Validation of Complex Data Assimilation Methods Hendrik Elbern, Elmar Friese, Nadine Goris, Lars Nieradzik and many others Rhenish Institute
More informationHST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007
MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationConsider the joint probability, P(x,y), shown as the contours in the figure above. P(x) is given by the integral of P(x,y) over all values of y.
ATMO/OPTI 656b Spring 009 Bayesian Retrievals Note: This follows the discussion in Chapter of Rogers (000) As we have seen, the problem with the nadir viewing emission measurements is they do not contain
More informationJoint Temperature, Humidity, and Sea Surface Temperature Retrieval from IASI Sensor Data
Joint Temperature, Humidity, and Sea Surface Temperature Retrieval from IASI Sensor Data Marc Schwaerz and Gottfried Kirchengast Institute for Geophysics, Astrophysics, and Meteorology (IGAM), University
More informationCOMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection
COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection Instructor: Herke van Hoof (herke.vanhoof@cs.mcgill.ca) Based on slides by:, Jackie Chi Kit Cheung Class web page:
More informationc Springer, Reprinted with permission.
Zhijian Yuan and Erkki Oja. A FastICA Algorithm for Non-negative Independent Component Analysis. In Puntonet, Carlos G.; Prieto, Alberto (Eds.), Proceedings of the Fifth International Symposium on Independent
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationImage Registration Lecture 2: Vectors and Matrices
Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this
More informationChapter 4: Factor Analysis
Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.
More informationLecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26
Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1
More informationNoise & Data Reduction
Noise & Data Reduction Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum Dimension Reduction 1 Remember: Central Limit
More informationInverse problem and optimization
Inverse problem and optimization Laurent Condat, Nelly Pustelnik CNRS, Gipsa-lab CNRS, Laboratoire de Physique de l ENS de Lyon Decembre, 15th 2016 Inverse problem and optimization 2/36 Plan 1. Examples
More informationHomework 1. Yuan Yao. September 18, 2011
Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to
More informationTwo-dimensional characterization of atmospheric prole retrievals from limb sounding observations
Journal of Quantitative Spectroscopy & Radiative Transfer 86 (2004) 45 71 www.elsevier.com/locate/jqsrt Two-dimensional characterization of atmospheric prole retrievals from limb sounding observations
More informationA few applications of the SVD
A few applications of the SVD Many methods require to approximate the original data (matrix) by a low rank matrix before attempting to solve the original problem Regularization methods require the solution
More informationPrincipal Component Analysis
Principal Component Analysis Introduction Consider a zero mean random vector R n with autocorrelation matri R = E( T ). R has eigenvectors q(1),,q(n) and associated eigenvalues λ(1) λ(n). Let Q = [ q(1)
More informationChemometrics. Matti Hotokka Physical chemistry Åbo Akademi University
Chemometrics Matti Hotokka Physical chemistry Åbo Akademi University Linear regression Experiment Consider spectrophotometry as an example Beer-Lamberts law: A = cå Experiment Make three known references
More informationEstimation of Mars surface physical properties from hyperspectral images using the SIR method
Estimation of Mars surface physical properties from hyperspectral images using the SIR method Caroline Bernard-Michel, Sylvain Douté, Laurent Gardes and Stéphane Girard Source: ESA Outline I. Context Hyperspectral
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationUWPHYSRET an SSEC inversion package for high resolution infrared data based on LBLRTM.
UWPHYSRET an SSEC inversion package for high resolution infrared data based on LBLRTM. Paolo Antonelli, R. Knuteson, R. Garcia, S. Bedka, D. Tobin, J.Taylor, W. Smith, and H. Revercomb QuickTime and a
More information1 Linearity and Linear Systems
Mathematical Tools for Neuroscience (NEU 34) Princeton University, Spring 26 Jonathan Pillow Lecture 7-8 notes: Linear systems & SVD Linearity and Linear Systems Linear system is a kind of mapping f( x)
More informationPrincipal Component Analysis
B: Chapter 1 HTF: Chapter 1.5 Principal Component Analysis Barnabás Póczos University of Alberta Nov, 009 Contents Motivation PCA algorithms Applications Face recognition Facial expression recognition
More informationModelling temporal structure (in noise and signal)
Modelling temporal structure (in noise and signal) Mark Woolrich, Christian Beckmann*, Salima Makni & Steve Smith FMRIB, Oxford *Imperial/FMRIB temporal noise: modelling temporal autocorrelation temporal
More informationHST.582J/6.555J/16.456J
Blind Source Separation: PCA & ICA HST.582J/6.555J/16.456J Gari D. Clifford gari [at] mit. edu http://www.mit.edu/~gari G. D. Clifford 2005-2009 What is BSS? Assume an observation (signal) is a linear
More informationGround and On-Orbit Characterization and Calibration of the Geosynchronous Imaging Fourier Transform Spectrometer (GIFTS)
Ground and On-Orbit Characterization and Calibration of the Geosynchronous Imaging Fourier Transform Spectrometer (GIFTS) John D. Elwell 1, Deron K. Scott 1 Henry E. Revercomb 2, Fred A. Best 2, Robert
More informationSynergistic Cloud Clearing Using Aqua Sounding and Imaging Infrared Measurements
Synergistic Cloud Clearing Using Aqua Sounding and Imaging Infrared Measurements H-L Allen Huang, et al. CIMSS/SSEC Univ. of Wisconsin-Madison and William L. Smith, LaRC, NASA Cloud Clearing/Cloud Detection
More informationBINGO simulations and updates on the performance of. the instrument
BINGO simulations and updates on the performance of BINGO telescope the instrument M.-A. Bigot-Sazy BINGO collaboration Paris 21cm Intensity Mapping Workshop June 2014 21cm signal Observed sky Credit:
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More information