Diffusion Wavelets on Graphs and Manifolds

Size: px
Start display at page:

Download "Diffusion Wavelets on Graphs and Manifolds"

Transcription

1 Diffusion Wavelets on Graphs and Manifolds R.R. Coifman, MM, J.C. Bremer Jr., A.D. Szlam Other collaborators on related projects: P.W. Jones, S. Lafon, R. Schul

2 Outline of the talk 1) Relate learning and approximation under smoothness constraints, motivate need for good 'building blocks for approximation 2) Some very classical building blocks in Euclidean spaces: Fourier and wavelets (a comparison) 3) New multiscale building blocks for manifolds: for this we need to talk about about diffusion geometries: 3a) Global diffusion geometries, eigenfunctions of the Laplacian 3b) Multiscale diffusion wavelets 4) Stylized applications 5) Future work

3 Learning & Function Approximation on graphs and manifolds In many cases learning algorithms (or learning itself?) can be viewed as an approximation problem, under smoothness constraints, on a set of points/state space represented as a manifold or graph. [Think of SMVs as an example] Approximation Smoothness constraint Fitting the function of interest Want generalization power, complexity control, avoid overfitting Manifold/graph Belief: the structure of the state space has to do with the learning task.

4 Goal Develop tools for multiscale analysis of functions on manifolds, varifolds, graphs, datasets. Motivations Analyze large amount of data, and functions on this data, intrinsically rather low-dimensional, embedded in high dimensions. Paradigm: we have a large number of documents (e.g.: web pages, gene array data, (hyper)spectral data, molecular dynamics data etc...) and a way of measuring similarity between pairs. Model: a graph (G,E,W) In important practical cases: vertices are points in high-dimensional Euclidean space, weights may be a function of Euclidean distance. Difficulties Data sets in high-dimensions are complicated, as are classes of interesting functions on them. We want to do approximation ( learning ) of such functions. Parametrize low dimensional subspaces of smooth functions on sets embedded in highdimension. Fast algorithms

5 High dimensional data: examples Documents, web searching Customer databases Satellite imagery Transaction logs Social networks Gene arrays, proteomics data Art transactions data Traffic (automobilistic, network) statistics

6

7

8 Function with a discontinuity Rate of decay of coefficients onto Fourier (red) and wavelets (blue)

9 Laplacian, diffusion geometries [RR Coifman, S. Lafon] Related previous and current work by M. Belkin, P. Niyogi on eigenmaps and applications to semi-supervised learning on manifolds and graphs. Part of the material in the next few slides is courtesy of Stephane Lafon. From local to global: diffusion distances Motto: Diffusion distance measures and averages connections of all lengths. It is more stable and finer than geodesic distance, uses a preponderance of evidence

10 A local similarity operator on the set X = { x1, x2,..., xn } data set. k ( x, y ) kernel defined on the data and being symmetric k ( x, y) = k ( x, y) positivity-preserving: k ( x, y ) positive semi-definite: α ( x)α ( y)k ( x, y) x X y X The kernel k describes the geometry of X by defining the relationship between the data points.

11 Examples of similarity kernels Examples: If X lies in n-dimensional Euclidean space, k(x,y) could be: - exponentially weighted distance: exp(-( x-y /a)^2 - angle: <x,y>/( x y ) - harmonic potential: 1/(a+ x-y ), or powers of it - feature distances : any of the above applied in the range f(x), f nonlinear, possibly mapping X to higher dimension - if a model for the data is available (probabilistic, or from a (stochastic) dynamical system), k may be chosen to be consistent with that model If X is an abstract graph: we need to be given some weights on edges, measuring similarity. This could wildly vary: from the graph obtained by discretizing a PDE, to ways of measuring similarity between two web pages, or two protein chains...

12 Diffusion Distances x y Diffusion embedding mapping X with diffusion distance into Euclidean space with Euclidean distance:

13 Phi1 Phi2 Phi3

14 Diffusion embedding mapping X with diffusion distance into Euclidean space with Euclidean distance:

15 Link with other kernel methods Recent kernel methods: LLE (Roweis, Saul 2), Laplacian Eigenmaps (Belkin, Niyogi 22), Hessian Eigenmaps (Donoho, Grimes 23), LTSA (Zhang, Zha 22)... all based on the following paradigm: minimize Q( f ) where Q( f ) = Qx ( f ) x X Qx ( f ) : quadratic form measuring local variation of f in a neighborhood of x Solution: compute eigenfunctions {ϕl }of Q and map data points via x (ϕ ( x), ϕ1 ( x),..., ϕ p ( x)) T In our case we minimize k ( x, y)( f ( x) f ( y)) 2 x X

16 Applications - Classifiers in the semi-supervised learning context (M. Belkin, P. Nyogi) - fmri data (F. Meyer, X. Shen) - Art data (W Goetzmann, PW Jones, MM, J Walden) - Hyperspectral Imaging in Pathology (MM, GL Davis, F Warner, F. Geshwind, A Coppi, R. DeVerse, RR Coifman) - Molecular dynamics simulations (RR. Coifman, G.Hummer, I. Kevrekidis, MM) - Text documents classification (RR. Coifman, MM)

17 Art data [W Goetzmann, PW Jones, MM, J Walden] Given the sales of art works (paintings and drawings) of 58 artists in for 9 years. View them as 58 points in 9 dimensions, try to find clusters that indicate how the works of groups of artists are 'traded' together. Does this correlate to style? Projection of the data on the top 3 Principal Components

18 Image under the top 3 eigenfunction of Laplacian + K-means

19 Hyper-spectral Pathology Data [MM, GL Davis, F Warner, F. Geshwind, A Coppi, R. DeVerse, RR Coifman]

20

21

22

23 So far.....we have seen that: - it seems useful to consider a framework in which, for a given data set, local similarities are given only between very similar points - it is possible to organize these local information by diffusion into global parametrizations, - these parametrizations can be found by looking at the eigenvectors of a diffusion operator, - these eigenvectors in turn yield a nonlinear embedding into low-dimensional Euclidean space, - the eigenvectors can be used for global Fourier analysis on the set/manifold PROBLEM: Either very local information or very global information: in many problems the intermediate scales are very interesting (think of web pages)! Would like MULTISCALE information! Solution 1: proceed bottom up: repeatedly cluster together in a multi-scale fashion, in a way that is faithful to the operator: diffusion wavelets. Solution 2: proceed top bottom: cut greedily according to global information, and repeat procedure on the pieces: recursive partitioning, local cosines... Solution 3: do both!

24 From global Diffusion Geometries... We are given a graph X with weights W. There is a natural random walk P on X induced by these weights. P maps probability distributions on X to probability distributions on X. The reversibility condition on P implies that P is conjugate to a selfadjoint operator T. We assume T is renormalized to have 2-norm 1. The spectra of T and its powers look like: σ (T ).5 4 σ (T ).4 8 σ (T ) ) σ (T ε V 3 V 2 1 V 1 15 V

25 ... to Multiresolution Diffusion [Coifman,MM] Multiscale compression scheme: - Random walk for 1=2^ step - Collect together random walkers into representatives - Write a random walk on the representatives - Let the representatives random walk 2=2^1 steps... The decay in the spectrum of T says powers of T are low-rank, hence compressible σ (T ).5 4 σ (T ).4 8 σ (T ) σ (T ) ε V 3 V 2 1 V 1 15 V

26 Multiscale Random Walkers =

27 Dilations, translations, downsampling We have frequencies: the eigenvalues of the diffusion T. What about dilations, translations, downsampling? We may have minimal information about the geometry, and only locally. Let's think in terms of functions on the set X. Dilations: Use the diffusion operator T and its dyadic powers as dilations. Translations and downsampling: Idea: diffusing a basis of scaling functions at a certain scale by a power of T should yield a redundant set of coarser scaling functions at the next coarser scale: reduce this set to a Riesz (i.e. well-conditioned) or even orthonormal basis. This is downsampling in the function space, and corresponds to finding a well-conditioned subset of translates.

28 Potential Theory, Green's function

29

30 Diffusion Wavelets on the sphere

31 Diffusion Wavelets on a dumbbell

32

33 Connections... Wavelets: - Lifting (W. Sweldens, I. Daubechies,...) - Continuous wavelets from group actions (Grossman, Morlet, Ali, Antoine, Gazeau, many physicists...) Classical Harmonic Analysis: - Littlewood-Paley on semigroups (Stein), Markov diffusion semigroups (large literature) - Martingales associated to the above, Brownian motion - Heat kernel estimates, generalized Heisenberg principles (A. Nahmod) - Harmonic Analysis of eigenfunctions of the Laplacian on domains, manifolds and graphs - Atomic decompositions (Coifman, Weiss, Rochberg...) Numerics: - Algebraic Multigrid (Brandt, large literature since the 8's...) - Kind of inverse FMM (Rohklin, Beylkin-Coifman-Rohklin,...) - Multiscale matrix compression techniques (Gu-Eisenstat, Chaen-GimbutasMartinsson-Rohklin,...) - Randomizable - FFTs?

34 Diffusion Wavelet Packets [JC Bremer, RR Coifman, MM, AD Szlam] We can split the wavelet subspaces further, in a hierarchical dyadic fashion, very much like in the classical case. The splittings are generated by numerical kernel and numerical range operations.

35

36 Compression example I

37 Denoising [a la Coifman-Wickerhauser & Donoho-Johnstone]

38 Analysis of a document corpora Given 1, documents, each of which is associated with 1, words, with a value indicating the relevance of each word in that document. View this as a set of 1, in 1, dimensions, construct a graph with 1, vertices, each of which with few selected edges to very close-by documents

39

40 Physics Paleontology Climate Astronomy, planets Comets, asteroids

41

42 Examples of multiscale scaling functions on documents Documents - By Jessica Gorman--"Pan scrapings!" announced G. Kenneth Sams. "When you come down to it, that is -what has brought us all here tonight-pan scrapings."-- It was probably not the most appetizing introduction for a $15per-plate -museum fundraiser. Some 15 guests, many in black tie, had just sat down to -enjoy their first course at lavish banquet tables in the Upper Egyptian Gallery -of the University of Pennsylvania Museum of Archaeology and Anthropology in -Philadelphia. Did they r Can chemical analysis confirm a wine's authenticity?-damaris Christensen-- There's a romance to wine unmatched by most agricultural products. Making -wine, with all its complex flavors, remains as much an art as a science.-- Hints of citrus, peach, raspberry, pear, oak, grass, or flowers may show up in -the taste or smell of wines. The catalog of factors that determine the flavor -and bouquet is almost as long as the list of adjectives that connoisseurs use -to describe them.-- Words 'wine' 'region' 'chemical' 'food' 'ancient'

43 Examples of multiscale scaling functions on documents Documents Words 'nitrogen' 'plant' 'ecologist' 'carbon' 'global' by J. Raloff-- Acid rain and agricultural pollution 'specy' 'native' 'change' 'pollution' both spew nitrogen into the air. Though -plants need 'university' 'growth' 'soil' 'human' 'activity' nitrogen to grow, a new study finds that even small 'nutrient' 'dioxide' 'fuel' 'acid' 'cycle' additions of -this fertilizing pollutant can perturb the 'natural' landscape.-- In plots of Minnesota prairie to which ecologists applied nitrogen for 12 -years, native grasses showed a dramatically impaired ability to compete against -weeds that had immigrated from Europe centuries ago.-- The nitrogen treatment triggered the terrestrial equ ans = by C. Mlot-- For most of agricultural history, nitrogen has been a precious commodity. Only -specialized bacteria and lightning could convert atmospheric nitrogen into -biologically usable forms. Today, however, fertilizers and fossil fuels have -made nitrogen so freely available that it has become too much of a good thing.-- In a review of nitrogen's effects across the environmental spectrum, a team of -ecologists headed by Peter M. Vitousek of Stanford University has concluded in -no

44 Examples of multiscale scaling functions on documents Documents Words 'earthquake' 'wave' 'fault' 'quake' 'tsunami' New technologies convert the motion of waves into 'california' 'southern' 'rock' 'seismologist' watts-- Peter Weiss-'ocean' 'lake' 'coast' 'tsunamis' 'north' R. Monastersky-- The underwater volcano growing 'region' 'island' 'seismic' 'plate' 'crust' off Hawaii's south coastline collapsed 'power' - The Mush Zone A slurpy layer lurks deep inside the planet-by I. Peterson-- In late 1942, carrying 15, U.S. soldiers bound for England, the Queen Mary -hit a storm about 7 miles off the coast of Scotland. Without warning amid the -tumult, a single, mountainous wave struck the ocean liner, rolling it over and -washing water across its upper decks. Luckily, the ship managed to right itself -and continue on its voyage.-- Gargantuan waves, which appear unexpectedly even under calm conditions in the -open ocean, have damaged and sunk numerous s - Racing the Waves Seismologists try to catch quake tremors quickly enough to -save lives-- geologists have been snapping a very different picture of -the lake lately. Far beneath Lake Tahoe's gentle surface, they say, several -hidden earthquake faults snake across the lake's flat bottom. How a middling quake made a giant tsunami Waves of Death-- Why the New Guinea tsunami

45 Comments, Applications, etc... This is a wavelet analysis on manifolds, graphs, markov chains, certain fractals, while Laplacian eigenfunctions do the corresponding Fourier Analysis, adapted to a given diffusion operator. We are compressing : powers of the operator, functions of the operators, subspaces of the function subspaces on which its powers act. This yields sampling formulas, quadrature formulas. Does not require the diffusion to be self-adjoint, nor needs the eigenvectors. We are constructing a biorthogonal version of the transform (better adapted to studying Markov chains): this will allow efficient denoising, compression, discrimination on all the spaces mentioned above. Preliminary: n log(n) with good constants for symmetric problems. The multiscale spaces are a natural scale of complexity spaces for learning empirical functions on the data set. Diffusion scaling functions extend outside the set, in a natural multiscale fashion. Exploring ties with measure-geometric considerations used for embedding metric spaces in Euclidean spaces with small distortion. Study and compression of dynamical systems.

46 Current & Future Work - Use for regularization and learning on graphs and manifolds: semisupervised and plain supervised. - Robustness to noise, perturbations, extension of the functions to test points. - Biorthogonal construction (better localization, better constants) - Multiscale embeddings for graphs, measure-geometric implications - Martingale aspects and fast Brownian motion simulation in nonhomogeneous media - Compression of data sets - Going nonlinear This talk, papers, Matlab code available at: (or Google for Mauro Maggioni ) Thank you!

Diffusion Wavelets and Applications

Diffusion Wavelets and Applications Diffusion Wavelets and Applications J.C. Bremer, R.R. Coifman, P.W. Jones, S. Lafon, M. Mohlenkamp, MM, R. Schul, A.D. Szlam Demos, web pages and preprints available at: S.Lafon: www.math.yale.edu/~sl349

More information

Diffusion Geometries, Global and Multiscale

Diffusion Geometries, Global and Multiscale Diffusion Geometries, Global and Multiscale R.R. Coifman, S. Lafon, MM, J.C. Bremer Jr., A.D. Szlam, P.W. Jones, R.Schul Papers, talks, other materials available at: www.math.yale.edu/~mmm82 Data and functions

More information

Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets.

Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets. Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets. R.R. Coifman, S. Lafon, MM Mathematics Department Program of Applied Mathematics. Yale University Motivations The main

More information

Diffusion Wavelets for multiscale analysis on manifolds and graphs: constructions and applications

Diffusion Wavelets for multiscale analysis on manifolds and graphs: constructions and applications Diffusion Wavelets for multiscale analysis on manifolds and graphs: constructions and applications Mauro Maggioni EPFL, Lausanne Dec. 19th, 5 EPFL Multiscale Analysis and Diffusion Wavelets - Mauro Maggioni,

More information

Towards Multiscale Harmonic Analysis of Graphs and Document Corpora

Towards Multiscale Harmonic Analysis of Graphs and Document Corpora Towards Multiscale Harmonic Analysis of Graphs and Document Corpora Mauro Maggioni Diffusion geometry group at Yale: R.R. Coifman, P. Jones, A.D. Szlam, J. Bremer, A. Singer, F. Warner, S.W. Zucker Jan.

More information

Multiscale analysis on graphs

Multiscale analysis on graphs Mathematics and Computer Science Duke University I.P.A.M. 11/17/08 In collaboration with R.R. Coifman, P.W. Jones, Y-M. Jung, R. Schul, A.D. Szlam & J.C. Bremer Jr. Funding: NSF/DHS-FODAVA, DMS, IIS, CCF;

More information

Nonlinear Dimensionality Reduction. Jose A. Costa

Nonlinear Dimensionality Reduction. Jose A. Costa Nonlinear Dimensionality Reduction Jose A. Costa Mathematics of Information Seminar, Dec. Motivation Many useful of signals such as: Image databases; Gene expression microarrays; Internet traffic time

More information

Contribution from: Springer Verlag Berlin Heidelberg 2005 ISBN

Contribution from: Springer Verlag Berlin Heidelberg 2005 ISBN Contribution from: Mathematical Physics Studies Vol. 7 Perspectives in Analysis Essays in Honor of Lennart Carleson s 75th Birthday Michael Benedicks, Peter W. Jones, Stanislav Smirnov (Eds.) Springer

More information

Clustering in kernel embedding spaces and organization of documents

Clustering in kernel embedding spaces and organization of documents Clustering in kernel embedding spaces and organization of documents Stéphane Lafon Collaborators: Raphy Coifman (Yale), Yosi Keller (Yale), Ioannis G. Kevrekidis (Princeton), Ann B. Lee (CMU), Boaz Nadler

More information

Graphs, Geometry and Semi-supervised Learning

Graphs, Geometry and Semi-supervised Learning Graphs, Geometry and Semi-supervised Learning Mikhail Belkin The Ohio State University, Dept of Computer Science and Engineering and Dept of Statistics Collaborators: Partha Niyogi, Vikas Sindhwani In

More information

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels March 13, 2008 Paper: R.R. Coifman, S. Lafon, maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels Figure: Example Application from [LafonWWW] meaningful geometric

More information

A Multiscale Framework for Markov Decision Processes using Diffusion Wavelets

A Multiscale Framework for Markov Decision Processes using Diffusion Wavelets A Multiscale Framework for Markov Decision Processes using Diffusion Wavelets Mauro Maggioni Program in Applied Mathematics Department of Mathematics Yale University New Haven, CT 6 mauro.maggioni@yale.edu

More information

Multiscale Wavelets on Trees, Graphs and High Dimensional Data

Multiscale Wavelets on Trees, Graphs and High Dimensional Data Multiscale Wavelets on Trees, Graphs and High Dimensional Data ICML 2010, Haifa Matan Gavish (Weizmann/Stanford) Boaz Nadler (Weizmann) Ronald Coifman (Yale) Boaz Nadler Ronald Coifman Motto... the relationships

More information

Data-dependent representations: Laplacian Eigenmaps

Data-dependent representations: Laplacian Eigenmaps Data-dependent representations: Laplacian Eigenmaps November 4, 2015 Data Organization and Manifold Learning There are many techniques for Data Organization and Manifold Learning, e.g., Principal Component

More information

Learning gradients: prescriptive models

Learning gradients: prescriptive models Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan

More information

Multiscale Manifold Learning

Multiscale Manifold Learning Multiscale Manifold Learning Chang Wang IBM T J Watson Research Lab Kitchawan Rd Yorktown Heights, New York 598 wangchan@usibmcom Sridhar Mahadevan Computer Science Department University of Massachusetts

More information

Non-linear Dimensionality Reduction

Non-linear Dimensionality Reduction Non-linear Dimensionality Reduction CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Introduction Laplacian Eigenmaps Locally Linear Embedding (LLE)

More information

Learning on Graphs and Manifolds. CMPSCI 689 Sridhar Mahadevan U.Mass Amherst

Learning on Graphs and Manifolds. CMPSCI 689 Sridhar Mahadevan U.Mass Amherst Learning on Graphs and Manifolds CMPSCI 689 Sridhar Mahadevan U.Mass Amherst Outline Manifold learning is a relatively new area of machine learning (2000-now). Main idea Model the underlying geometry of

More information

Data dependent operators for the spatial-spectral fusion problem

Data dependent operators for the spatial-spectral fusion problem Data dependent operators for the spatial-spectral fusion problem Wien, December 3, 2012 Joint work with: University of Maryland: J. J. Benedetto, J. A. Dobrosotskaya, T. Doster, K. W. Duke, M. Ehler, A.

More information

Diffusion/Inference geometries of data features, situational awareness and visualization. Ronald R Coifman Mathematics Yale University

Diffusion/Inference geometries of data features, situational awareness and visualization. Ronald R Coifman Mathematics Yale University Diffusion/Inference geometries of data features, situational awareness and visualization Ronald R Coifman Mathematics Yale University Digital data is generally converted to point clouds in high dimensional

More information

Fast Direct Policy Evaluation using Multiscale Analysis of Markov Diffusion Processes

Fast Direct Policy Evaluation using Multiscale Analysis of Markov Diffusion Processes Fast Direct Policy Evaluation using Multiscale Analysis of Markov Diffusion Processes Mauro Maggioni mauro.maggioni@yale.edu Department of Mathematics, Yale University, P.O. Box 88, New Haven, CT,, U.S.A.

More information

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics

More information

Multiscale Analysis and Diffusion Semigroups With Applications

Multiscale Analysis and Diffusion Semigroups With Applications Multiscale Analysis and Diffusion Semigroups With Applications Karamatou Yacoubou Djima Advisor: Wojciech Czaja Norbert Wiener Center Department of Mathematics University of Maryland, College Park http://www.norbertwiener.umd.edu

More information

Multiscale bi-harmonic Analysis of Digital Data Bases and Earth moving distances.

Multiscale bi-harmonic Analysis of Digital Data Bases and Earth moving distances. Multiscale bi-harmonic Analysis of Digital Data Bases and Earth moving distances. R. Coifman, Department of Mathematics, program of Applied Mathematics Yale University Joint work with M. Gavish and W.

More information

Bi-stochastic kernels via asymmetric affinity functions

Bi-stochastic kernels via asymmetric affinity functions Bi-stochastic kernels via asymmetric affinity functions Ronald R. Coifman, Matthew J. Hirn Yale University Department of Mathematics P.O. Box 208283 New Haven, Connecticut 06520-8283 USA ariv:1209.0237v4

More information

Global vs. Multiscale Approaches

Global vs. Multiscale Approaches Harmonic Analysis on Graphs Global vs. Multiscale Approaches Weizmann Institute of Science, Rehovot, Israel July 2011 Joint work with Matan Gavish (WIS/Stanford), Ronald Coifman (Yale), ICML 10' Challenge:

More information

CSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13

CSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13 CSE 291. Assignment 3 Out: Wed May 23 Due: Wed Jun 13 3.1 Spectral clustering versus k-means Download the rings data set for this problem from the course web site. The data is stored in MATLAB format as

More information

Invariant Scattering Convolution Networks

Invariant Scattering Convolution Networks Invariant Scattering Convolution Networks Joan Bruna and Stephane Mallat Submitted to PAMI, Feb. 2012 Presented by Bo Chen Other important related papers: [1] S. Mallat, A Theory for Multiresolution Signal

More information

Value Function Approximation with Diffusion Wavelets and Laplacian Eigenfunctions

Value Function Approximation with Diffusion Wavelets and Laplacian Eigenfunctions Value Function Approximation with Diffusion Wavelets and Laplacian Eigenfunctions Sridhar Mahadevan Department of Computer Science University of Massachusetts Amherst, MA 13 mahadeva@cs.umass.edu Mauro

More information

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Overview Introduction Linear Methods for Dimensionality Reduction Nonlinear Methods and Manifold

More information

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian

More information

Sparse Approximation: from Image Restoration to High Dimensional Classification

Sparse Approximation: from Image Restoration to High Dimensional Classification Sparse Approximation: from Image Restoration to High Dimensional Classification Bin Dong Beijing International Center for Mathematical Research Beijing Institute of Big Data Research Peking University

More information

Spectral Processing. Misha Kazhdan

Spectral Processing. Misha Kazhdan Spectral Processing Misha Kazhdan [Taubin, 1995] A Signal Processing Approach to Fair Surface Design [Desbrun, et al., 1999] Implicit Fairing of Arbitrary Meshes [Vallet and Levy, 2008] Spectral Geometry

More information

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Introduction and Data Representation Mikhail Belkin & Partha Niyogi Department of Electrical Engieering University of Minnesota Mar 21, 2017 1/22 Outline Introduction 1 Introduction 2 3 4 Connections to

More information

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold. Nonlinear Methods Data often lies on or near a nonlinear low-dimensional curve aka manifold. 27 Laplacian Eigenmaps Linear methods Lower-dimensional linear projection that preserves distances between all

More information

Nonlinear Dimensionality Reduction

Nonlinear Dimensionality Reduction Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Kernel PCA 2 Isomap 3 Locally Linear Embedding 4 Laplacian Eigenmap

More information

Mathematical Methods in Machine Learning

Mathematical Methods in Machine Learning UMD, Spring 2016 Outline Lecture 2: Role of Directionality 1 Lecture 2: Role of Directionality Anisotropic Harmonic Analysis Harmonic analysis decomposes signals into simpler elements called analyzing

More information

Apprentissage non supervisée

Apprentissage non supervisée Apprentissage non supervisée Cours 3 Higher dimensions Jairo Cugliari Master ECD 2015-2016 From low to high dimension Density estimation Histograms and KDE Calibration can be done automacally But! Let

More information

Stable MMPI-2 Scoring: Introduction to Kernel. Extension Techniques

Stable MMPI-2 Scoring: Introduction to Kernel. Extension Techniques 1 Stable MMPI-2 Scoring: Introduction to Kernel Extension Techniques Liberty,E., Almagor,M., Zucker,S., Keller,Y., and Coifman,R.R. Abstract The current study introduces a new technique called Geometric

More information

The Scope of Earth Science

The Scope of Earth Science Earth science Earth Science The Scope of Earth Science The field of Earth Science can be broken into four major areas of specialization: astronomy, meteorology, geology, and oceanography. 1. Astronomy

More information

IFT LAPLACIAN APPLICATIONS. Mikhail Bessmeltsev

IFT LAPLACIAN APPLICATIONS.   Mikhail Bessmeltsev IFT 6112 09 LAPLACIAN APPLICATIONS http://www-labs.iro.umontreal.ca/~bmpix/teaching/6112/2018/ Mikhail Bessmeltsev Rough Intuition http://pngimg.com/upload/hammer_png3886.png You can learn a lot about

More information

Unsupervised dimensionality reduction

Unsupervised dimensionality reduction Unsupervised dimensionality reduction Guillaume Obozinski Ecole des Ponts - ParisTech SOCN course 2014 Guillaume Obozinski Unsupervised dimensionality reduction 1/30 Outline 1 PCA 2 Kernel PCA 3 Multidimensional

More information

Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center

Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center 1 Outline Part I - Applications Motivation and Introduction Patient similarity application Part II

More information

Math 307 Learning Goals. March 23, 2010

Math 307 Learning Goals. March 23, 2010 Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent

More information

The Generalized Haar-Walsh Transform (GHWT) for Data Analysis on Graphs and Networks

The Generalized Haar-Walsh Transform (GHWT) for Data Analysis on Graphs and Networks The Generalized Haar-Walsh Transform (GHWT) for Data Analysis on Graphs and Networks Jeff Irion & Naoki Saito Department of Mathematics University of California, Davis SIAM Annual Meeting 2014 Chicago,

More information

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Yoshua Bengio Pascal Vincent Jean-François Paiement University of Montreal April 2, Snowbird Learning 2003 Learning Modal Structures

More information

Justin Solomon MIT, Spring 2017

Justin Solomon MIT, Spring 2017 Justin Solomon MIT, Spring 2017 http://pngimg.com/upload/hammer_png3886.png You can learn a lot about a shape by hitting it (lightly) with a hammer! What can you learn about its shape from vibration frequencies

More information

Is Manifold Learning for Toy Data only?

Is Manifold Learning for Toy Data only? s Manifold Learning for Toy Data only? Marina Meilă University of Washington mmp@stat.washington.edu MMDS Workshop 2016 Outline What is non-linear dimension reduction? Metric Manifold Learning Estimating

More information

Robust Laplacian Eigenmaps Using Global Information

Robust Laplacian Eigenmaps Using Global Information Manifold Learning and its Applications: Papers from the AAAI Fall Symposium (FS-9-) Robust Laplacian Eigenmaps Using Global Information Shounak Roychowdhury ECE University of Texas at Austin, Austin, TX

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 23 1 / 27 Overview

More information

Proto-value Functions: A Laplacian Framework for Learning Representation and Control in Markov Decision Processes

Proto-value Functions: A Laplacian Framework for Learning Representation and Control in Markov Decision Processes Proto-value Functions: A Laplacian Framework for Learning Representation and Control in Markov Decision Processes Sridhar Mahadevan Department of Computer Science University of Massachusetts Amherst, MA

More information

PARAMETERIZATION OF NON-LINEAR MANIFOLDS

PARAMETERIZATION OF NON-LINEAR MANIFOLDS PARAMETERIZATION OF NON-LINEAR MANIFOLDS C. W. GEAR DEPARTMENT OF CHEMICAL AND BIOLOGICAL ENGINEERING PRINCETON UNIVERSITY, PRINCETON, NJ E-MAIL:WGEAR@PRINCETON.EDU Abstract. In this report we consider

More information

6/9/2010. Feature-based methods and shape retrieval problems. Structure. Combining local and global structures. Photometric stress

6/9/2010. Feature-based methods and shape retrieval problems. Structure. Combining local and global structures. Photometric stress 1 2 Structure Feature-based methods and shape retrieval problems Global Metric Local Feature descriptors Alexander & Michael Bronstein, 2006-2009 Michael Bronstein, 2010 tosca.cs.technion.ac.il/book 048921

More information

Nonlinear Dimensionality Reduction

Nonlinear Dimensionality Reduction Nonlinear Dimensionality Reduction Piyush Rai CS5350/6350: Machine Learning October 25, 2011 Recap: Linear Dimensionality Reduction Linear Dimensionality Reduction: Based on a linear projection of the

More information

Learning Representation & Behavior:

Learning Representation & Behavior: Learning Representation & Behavior: Manifold and Spectral Methods for Markov Decision Processes and Reinforcement Learning Sridhar Mahadevan, U. Mass, Amherst Mauro Maggioni, Yale Univ. June 25, 26 ICML

More information

Dimensionality Reduction:

Dimensionality Reduction: Dimensionality Reduction: From Data Representation to General Framework Dong XU School of Computer Engineering Nanyang Technological University, Singapore What is Dimensionality Reduction? PCA LDA Examples:

More information

The Nature of Science

The Nature of Science Chapter 1 Earth Science Lesson 1 The Nature of Science Main idea: Earth science encompasses five areas of study: astronomy, meteorology, geology oceanography, and environmental science. Earth has four

More information

Lecture: Some Practical Considerations (3 of 4)

Lecture: Some Practical Considerations (3 of 4) Stat260/CS294: Spectral Graph Methods Lecture 14-03/10/2015 Lecture: Some Practical Considerations (3 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough.

More information

Lecture: Local Spectral Methods (1 of 4)

Lecture: Local Spectral Methods (1 of 4) Stat260/CS294: Spectral Graph Methods Lecture 18-03/31/2015 Lecture: Local Spectral Methods (1 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough. They provide

More information

Analysis of Spectral Kernel Design based Semi-supervised Learning

Analysis of Spectral Kernel Design based Semi-supervised Learning Analysis of Spectral Kernel Design based Semi-supervised Learning Tong Zhang IBM T. J. Watson Research Center Yorktown Heights, NY 10598 Rie Kubota Ando IBM T. J. Watson Research Center Yorktown Heights,

More information

Regression on Manifolds Using Kernel Dimension Reduction

Regression on Manifolds Using Kernel Dimension Reduction Jens Nilsson JENSN@MATHS.LTH.SE Centre for Mathematical Sciences, Lund University, Box 118, SE-221 00 Lund, Sweden Fei Sha FEISHA@CS.BERKELEY.EDU Computer Science Division, University of California, Berkeley,

More information

Exploiting Sparse Non-Linear Structure in Astronomical Data

Exploiting Sparse Non-Linear Structure in Astronomical Data Exploiting Sparse Non-Linear Structure in Astronomical Data Ann B. Lee Department of Statistics and Department of Machine Learning, Carnegie Mellon University Joint work with P. Freeman, C. Schafer, and

More information

Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking p. 1/31

Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking p. 1/31 Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking Dengyong Zhou zhou@tuebingen.mpg.de Dept. Schölkopf, Max Planck Institute for Biological Cybernetics, Germany Learning from

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

From graph to manifold Laplacian: The convergence rate

From graph to manifold Laplacian: The convergence rate Appl. Comput. Harmon. Anal. 2 (2006) 28 34 www.elsevier.com/locate/acha Letter to the Editor From graph to manifold Laplacian: The convergence rate A. Singer Department of athematics, Yale University,

More information

Manifold Coarse Graining for Online Semi-supervised Learning

Manifold Coarse Graining for Online Semi-supervised Learning for Online Semi-supervised Learning Mehrdad Farajtabar, Amirreza Shaban, Hamid R. Rabiee, Mohammad H. Rohban Digital Media Lab, Department of Computer Engineering, Sharif University of Technology, Tehran,

More information

Kernel-Based Contrast Functions for Sufficient Dimension Reduction

Kernel-Based Contrast Functions for Sufficient Dimension Reduction Kernel-Based Contrast Functions for Sufficient Dimension Reduction Michael I. Jordan Departments of Statistics and EECS University of California, Berkeley Joint work with Kenji Fukumizu and Francis Bach

More information

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Shuyang Ling Courant Institute of Mathematical Sciences, NYU Aug 13, 2018 Joint

More information

Graph Partitioning Using Random Walks

Graph Partitioning Using Random Walks Graph Partitioning Using Random Walks A Convex Optimization Perspective Lorenzo Orecchia Computer Science Why Spectral Algorithms for Graph Problems in practice? Simple to implement Can exploit very efficient

More information

Filtering via a Reference Set. A.Haddad, D. Kushnir, R.R. Coifman Technical Report YALEU/DCS/TR-1441 February 21, 2011

Filtering via a Reference Set. A.Haddad, D. Kushnir, R.R. Coifman Technical Report YALEU/DCS/TR-1441 February 21, 2011 Patch-based de-noising algorithms and patch manifold smoothing have emerged as efficient de-noising methods. This paper provides a new insight on these methods, such as the Non Local Means or the image

More information

Manifold Learning and it s application

Manifold Learning and it s application Manifold Learning and it s application Nandan Dubey SE367 Outline 1 Introduction Manifold Examples image as vector Importance Dimension Reduction Techniques 2 Linear Methods PCA Example MDS Perception

More information

The Illustrated Wavelet Transform Handbook. Introductory Theory and Applications in Science, Engineering, Medicine and Finance.

The Illustrated Wavelet Transform Handbook. Introductory Theory and Applications in Science, Engineering, Medicine and Finance. The Illustrated Wavelet Transform Handbook Introductory Theory and Applications in Science, Engineering, Medicine and Finance Paul S Addison Napier University, Edinburgh, UK IoP Institute of Physics Publishing

More information

Intrinsic Structure Study on Whale Vocalizations

Intrinsic Structure Study on Whale Vocalizations 1 2015 DCLDE Conference Intrinsic Structure Study on Whale Vocalizations Yin Xian 1, Xiaobai Sun 2, Yuan Zhang 3, Wenjing Liao 3 Doug Nowacek 1,4, Loren Nolte 1, Robert Calderbank 1,2,3 1 Department of

More information

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH Hoang Trang 1, Tran Hoang Loc 1 1 Ho Chi Minh City University of Technology-VNU HCM, Ho Chi

More information

Discrete Wavelet Transform

Discrete Wavelet Transform Discrete Wavelet Transform [11] Kartik Mehra July 2017 Math 190s Duke University "1 Introduction Wavelets break signals up and then analyse them separately with a resolution that is matched with scale.

More information

Conference in Honor of Aline Bonami Orleans, June 2014

Conference in Honor of Aline Bonami Orleans, June 2014 Conference in Honor of Aline Bonami Orleans, June 2014 Harmonic Analysis and functional duality, as a tool for organization of information, and learning. R. Coifman Department of Mathematics, program of

More information

Using the computer to select the right variables

Using the computer to select the right variables Using the computer to select the right variables Rationale: Straight Line Distance Curved Transition Distance Actual transition difficulty represented by curved path Lake Carnegie, Princeton, NJ Using

More information

High-Dimensional Pattern Recognition using Low-Dimensional Embedding and Earth Mover s Distance

High-Dimensional Pattern Recognition using Low-Dimensional Embedding and Earth Mover s Distance High-Dimensional Pattern Recognition using Low-Dimensional Embedding and Earth Mover s Distance Linh Lieu,a, Naoki Saito a a Department of Mathematics, University of California, Davis, CA 95616, USA Abstract

More information

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis Alvina Goh Vision Reading Group 13 October 2005 Connection of Local Linear Embedding, ISOMAP, and Kernel Principal

More information

Earthquakes. Written by: Lillie Hamstick

Earthquakes. Written by: Lillie Hamstick Earthquakes Written by: Lillie Hamstick Table of Contents Meeting an Earthquake. 1 Dangerous Destruction.. 2 Exciting Earthquakes 3 Yo, What Causes an Earthquake... 4 Where Are You Earthquake. 5 How to

More information

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction ECE 521 Lecture 11 (not on midterm material) 13 February 2017 K-means clustering, Dimensionality reduction With thanks to Ruslan Salakhutdinov for an earlier version of the slides Overview K-means clustering

More information

THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING

THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING Luis Rademacher, Ohio State University, Computer Science and Engineering. Joint work with Mikhail Belkin and James Voss This talk A new approach to multi-way

More information

L26: Advanced dimensionality reduction

L26: Advanced dimensionality reduction L26: Advanced dimensionality reduction The snapshot CA approach Oriented rincipal Components Analysis Non-linear dimensionality reduction (manifold learning) ISOMA Locally Linear Embedding CSCE 666 attern

More information

Manifold Regularization

Manifold Regularization 9.520: Statistical Learning Theory and Applications arch 3rd, 200 anifold Regularization Lecturer: Lorenzo Rosasco Scribe: Hooyoung Chung Introduction In this lecture we introduce a class of learning algorithms,

More information

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations. Previously Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations y = Ax Or A simply represents data Notion of eigenvectors,

More information

Advanced Machine Learning & Perception

Advanced Machine Learning & Perception Advanced Machine Learning & Perception Instructor: Tony Jebara Topic 1 Introduction, researchy course, latest papers Going beyond simple machine learning Perception, strange spaces, images, time, behavior

More information

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR

More information

Improved Local Coordinate Coding using Local Tangents

Improved Local Coordinate Coding using Local Tangents Improved Local Coordinate Coding using Local Tangents Kai Yu NEC Laboratories America, 10081 N. Wolfe Road, Cupertino, CA 95129 Tong Zhang Rutgers University, 110 Frelinghuysen Road, Piscataway, NJ 08854

More information

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline

More information

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008 Advances in Manifold Learning Presented by: Nakul Verma June 10, 008 Outline Motivation Manifolds Manifold Learning Random projection of manifolds for dimension reduction Introduction to random projections

More information

Semi-Supervised Learning

Semi-Supervised Learning Semi-Supervised Learning getting more for less in natural language processing and beyond Xiaojin (Jerry) Zhu School of Computer Science Carnegie Mellon University 1 Semi-supervised Learning many human

More information

Diffusion Wavelets. Ronald R. Coifman, Mauro Maggioni

Diffusion Wavelets. Ronald R. Coifman, Mauro Maggioni Diffusion Wavelets Ronald R. Coifman, Mauro Maggioni Program in Applied Mathematics Department of Mathematics Yale University New Haven,CT,65 U.S.A. Abstract We present a multiresolution construction for

More information

Localized Sliced Inverse Regression

Localized Sliced Inverse Regression Localized Sliced Inverse Regression Qiang Wu, Sayan Mukherjee Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University, Durham NC 2778-251,

More information

Unit 2: The World in Spatial Terms (Lessons 6-7)

Unit 2: The World in Spatial Terms (Lessons 6-7) Unit 2: The World in Spatial Terms (Lessons 6-7) Physical Features of Earth Create Challenges and Opportunities The physical features of Earth create both challenges and opportunities for the people living

More information

Unit 2 Ecology Study Guide. Niche Autotrophs Heterotrophs Decomposers Demography Dispersion

Unit 2 Ecology Study Guide. Niche Autotrophs Heterotrophs Decomposers Demography Dispersion Vocabulary to know: Ecology Ecosystem Abiotic Factors Biotic Factors Communities Population Unit 2 Ecology Study Guide Niche Autotrophs Heterotrophs Decomposers Demography Dispersion Growth Rate Carrying

More information

Spectral Algorithms I. Slides based on Spectral Mesh Processing Siggraph 2010 course

Spectral Algorithms I. Slides based on Spectral Mesh Processing Siggraph 2010 course Spectral Algorithms I Slides based on Spectral Mesh Processing Siggraph 2010 course Why Spectral? A different way to look at functions on a domain Why Spectral? Better representations lead to simpler solutions

More information

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture 7 What is spectral

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction

More information

STA 414/2104: Lecture 8

STA 414/2104: Lecture 8 STA 414/2104: Lecture 8 6-7 March 2017: Continuous Latent Variable Models, Neural networks With thanks to Russ Salakhutdinov, Jimmy Ba and others Outline Continuous latent variable models Background PCA

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information