Subspace Arrangements in Theory and Practice
|
|
- Matilda Skinner
- 6 years ago
- Views:
Transcription
1 Subspace Arrangements in Theory and Practice Robert M. Fossum Department of Mathematics Beckman Institute University of Illinios at Urbana-Champaign Applications in Biology, Dynamics, and Statistics 09 March 2007
2 Outline 1 Introduction Abstract Motivating Examples of Data Types and Uses 2 Subspace Arrangements Definitions Quick Course on Algebraic Sets Hilbert Functions and Hilbert Series Hilbert Series and Subspace Arrangements 3 Estimating Hybrid Subspace Models from Noisy Samples Stability Estimation of Vanishing Polynomials GPCA-Voting 4 Robust Component Analysis 5 Other Recent Advances Occluded Face Recognition 6 Future Directions Further Investigations Summary and Conclusions Thanks
3 Outline 1 Introduction Abstract Motivating Examples of Data Types and Uses 2 Subspace Arrangements Definitions Quick Course on Algebraic Sets Hilbert Functions and Hilbert Series Hilbert Series and Subspace Arrangements 3 Estimating Hybrid Subspace Models from Noisy Samples Stability Estimation of Vanishing Polynomials GPCA-Voting 4 Robust Component Analysis 5 Other Recent Advances Occluded Face Recognition 6 Future Directions Further Investigations Summary and Conclusions Thanks
4 Abstract Imagery A subspace arrangement is a union of a finite number of subspaces of a vector space. We will discuss the importance of subspace arrangements first as mathematical objects and now as a popular class of models for engineering.
5 Abstract Imagery A subspace arrangement is a union of a finite number of subspaces of a vector space. We will discuss the importance of subspace arrangements first as mathematical objects and now as a popular class of models for engineering. We will then introduce some of new theoretical results that were motivated from practice. Using these results we will address the computational issue about how to extract subspace arrangements from noisy or corrupted data.
6 Abstract Imagery A subspace arrangement is a union of a finite number of subspaces of a vector space. We will discuss the importance of subspace arrangements first as mathematical objects and now as a popular class of models for engineering. We will then introduce some of new theoretical results that were motivated from practice. Using these results we will address the computational issue about how to extract subspace arrangements from noisy or corrupted data. Finally we will turn to the importance of subspace arrangements by briefly discussing the connections to sparse representations, manifold learning, etc...
7 Collaborators and Support Primary Collaborators Yi Ma, ECE, UIUC Allen Yang, ECE, UIUC and EECS, UC Berkeley Harm Derksen, Math, U Michigan Collaborators René Vidal, Johns Hopkins Kun Huang, OSU John Wright, ECE, UIUC Shankar Rao, ECE, UIUC Drew Wagner, ECE, UIUC Support UIUC ECE, CSL, and CSE. NSF CAREER, CRS, CCF ONR
8 Problems and Motivation Goal Process, represent, store, retrieve, and interpret multidimensional and multivariate data. Example Imagery Multivariate data (extracted from images and videos) tend to be inhomogeneous and multi-modal. Often desirable to segment such data into uni-modal or homogeneous subsets, and model each subset with a different model. Linear Models are easiest to use but of course our data is usually not linear.
9 Problems and Motivation Goal Process, represent, store, retrieve, and interpret multidimensional and multivariate data.
10 Example (Reconstruction of Dynamic Scenes) Segment feature points that belong to objects that have different 2D or 3D motions. (a) 2D translations are hyperplanes in C 2 (b) 3D homographies are (c) 3D rigid-bodies are bilinear forms in R bilinear forms in C Figure: Subspace constraints in pairwise images [Vidal & Ma, ECCV 2004]
11 Example (Reconstruction of Dynamic Scenes) Segment feature points that belong to objects that have different 2D or 3D motions. parking-lot movie Figure: Stacked image features in an affine camera sequence lie in 4-D subspaces [Kanatani 2003]
12 DNA Sequences Clustering of DNA sequences Figure: Helicobacter pylori Zinovyev, Andreï: Visualizing the spatial structure of triplet distributions in genetic texts, IHES Technical Report 2002.
13 DNA Sequences Clustering of DNA sequences
14 Example Finger Print Identification Example of finger print data (a) Final Result (b) Original Figure: pk/research/matlabfns/fingerprints/docs/
15 Introduction Subspace Arrangements Noise Robust GPCA Other Recent Advances Future Image Segmentation Segment an image into regions of homogeneous texture. (a) GPCA (without post-processing) (b) Human Reference: Huang, Ma, & Vidal. Minimum Effective Dimension for Mixtures of Subspaces. CVPR, IMA 09 March 2007 Robert M. Fossum Subspace Arrangements in Theory and Practice
16 Introduction Subspace Arrangements Noise Robust GPCA Other Recent Advances Future Video Segmentation Event Detection Segment a video sequence into clips of homogeneous scenes. PingPong movie (c) Ping-Pong sequence (d) First three PCs (e) Segments Reference: 1. Huang, Wagner, & Ma. Hybrid linear system identification. CDC, Vidal & Ravichandran. Segmentation and Optical Flow for Multiple Moving Dynamic Textures. CVPR, IMA 09 March 2007 Robert M. Fossum Subspace Arrangements in Theory and Practice
17 Introduction Subspace Arrangements Noise Robust GPCA Other Recent Advances Future Other Applications Scientific studies on a large class of multivariate mixed data require effective segmentation methods to represent the data into multiple subsets with simpler quantitative models. Face Recognition. Hyperspectral Images. Linear Switching Systems. Human Kinematics. Handwritten Digits. trackers movie IMA 09 March 2007 Robert M. Fossum Subspace Arrangements in Theory and Practice
18 Segmentation? Example (What does segmentation mean?) Is there a formal mathematical definition of segmentation? What should the proper criterion for segmentation be? How should we measure the gain or loss of the segmentation? First choose a simple class of models with which each of the subsets is supposed to fit. Some of the popular models are either probabilistic distributions (e.g., Gaussian distributions) or geometric/algebraic sets (e.g., linear subspaces). Then the complete set of mixed data is assumed to consist of samples drawn from a mixture of such probabilistic distributions [5, 15] or geometric/algebraic sets [17]. Then estimate the mixture of all the models and simultaneously or subsequently decompose them into individual ones. Thus data segmentation is essentially identified with a (mixture) model estimation problem. Segmenting the data and estimating the model are therefore coupled together. Ma, Derksen, et al. Lossy Coding and segmentation of multivariate mixed data. Tech, Report
19 Method of Attack Model Find a model that fits the data and that fits the application. The goal of dimension reduction is to find a low-dimensional model for the high-dimensional data points. This problem has been studied for at least a century. One of the most commonly used technique is principal component analysis (PCA) (aka the Karhunen-Loéve transform (KLT). Given a finite set of data points, the goal is to find a low-dimensional linear model such as a subspace to fit the data points. Mathematically this is achieved by performing singular value decomposition on the matrix of data points to obtain the basis for the subspace. PCA (or SVD) has been widely used. In bioinformatics it has been frequently used to cluster sequences, microarray data, or even spectroscopy signals.
20 Method of Attack Model Find a model that fits the data and that fits the application. The goal of dimension reduction is to find a low-dimensional model for the high-dimensional data points. This problem has been studied for at least a century. One of the most commonly used technique is principal component analysis (PCA) (aka the Karhunen-Loéve transform (KLT). Given a finite set of data points, the goal is to find a low-dimensional linear model such as a subspace to fit the data points. Mathematically this is achieved by performing singular value decomposition on the matrix of data points to obtain the basis for the subspace. PCA (or SVD) has been widely used. In bioinformatics it has been frequently used to cluster sequences, microarray data, or even spectroscopy signals.
21 Method of Attack Model Find a model that fits the data and that fits the application. The goal of dimension reduction is to find a low-dimensional model for the high-dimensional data points. This problem has been studied for at least a century. One of the most commonly used technique is principal component analysis (PCA) (aka the Karhunen-Loéve transform (KLT). Given a finite set of data points, the goal is to find a low-dimensional linear model such as a subspace to fit the data points. Mathematically this is achieved by performing singular value decomposition on the matrix of data points to obtain the basis for the subspace. PCA (or SVD) has been widely used. In bioinformatics it has been frequently used to cluster sequences, microarray data, or even spectroscopy signals.
22 Method of Attack Model Find a model that fits the data and that fits the application. The goal of dimension reduction is to find a low-dimensional model for the high-dimensional data points. This problem has been studied for at least a century. One of the most commonly used technique is principal component analysis (PCA) (aka the Karhunen-Loéve transform (KLT). Given a finite set of data points, the goal is to find a low-dimensional linear model such as a subspace to fit the data points. Mathematically this is achieved by performing singular value decomposition on the matrix of data points to obtain the basis for the subspace. PCA (or SVD) has been widely used. In bioinformatics it has been frequently used to cluster sequences, microarray data, or even spectroscopy signals.
23 Method of Attack Model Find a model that fits the data and that fits the application. The goal of dimension reduction is to find a low-dimensional model for the high-dimensional data points. This problem has been studied for at least a century. One of the most commonly used technique is principal component analysis (PCA) (aka the Karhunen-Loéve transform (KLT). Given a finite set of data points, the goal is to find a low-dimensional linear model such as a subspace to fit the data points. Mathematically this is achieved by performing singular value decomposition on the matrix of data points to obtain the basis for the subspace. PCA (or SVD) has been widely used. In bioinformatics it has been frequently used to cluster sequences, microarray data, or even spectroscopy signals.
24 Principal Component Analysis (PCA) and Extensions PCA Problem: Dimension reduction for high-dimensional data to a low-dimensional subspace [Pearson 1901, Eckart-Yong 1930, Jolliffe 2002] Solution: Singular Value Decomposition (SVD) UΣV T = svd([x 1, x 2,..., x N ]). Extensions to PCA: Nonlinear Kernel PCA [Scholkopf-Smola-Muller 1998] Probabilistic PCA [Tipping-Bishop 1999, Collins et al. 2001] Higher-order SVD [Tucker 1966, Davis 2002] Independent Component Analysis (ICA) [Hyvarinen-Karhunen-Oja 2001] Punch line All of them do not have the capacity to estimate multiple low-dimensional models.
25 Iterative Schemes to the Subspace-Segmentation Problem: EM-Type Solutions: EM [Dempster et al. 1977] & K-Subspaces [Ho et al. 2003] Complete variable: (x, z) R D N, where z is the missing membership. Model parameters: Θ = {θ 1,..., θ n}, where θ i = (B i, σ i, π i ) for each V i, B i is the basis, and π i = p(z = i). 1 E step: Calculate w (m) ik. = p(z k = i x k, ˆΘ (m) ). 2 M step: Use the expected values w (m) ik, compute ˆΘ (m+1). Difficulties: Sensitive to the initialization (greedy). Random Sample Consensus (RANSAC): [Fischler & Bolles 1981, Torr 1998, Forsyth et al. 2001, Bartoli 2001, Schindler & Suter 2005] 1 Sample a (minimal) subset to estimate an estimate of the parameters. 2 Compute the consensus of the estimated parameters with the total sample set. Difficulties: 1 Higher-dimensional models overfit samples on lower-dimensional subspaces. 2 Lower-dimensional models may fit samples at intersections.
26 Outline 1 Introduction Abstract Motivating Examples of Data Types and Uses 2 Subspace Arrangements Definitions Quick Course on Algebraic Sets Hilbert Functions and Hilbert Series Hilbert Series and Subspace Arrangements 3 Estimating Hybrid Subspace Models from Noisy Samples Stability Estimation of Vanishing Polynomials GPCA-Voting 4 Robust Component Analysis 5 Other Recent Advances Occluded Face Recognition 6 Future Directions Further Investigations Summary and Conclusions Thanks
27 Subspace Arrangements Definition Let W be a vector space over a scalar field F of dimension D. A subspace arrangement A in W is a union of subspaces V 1, V 2,..., V m of W. A = V 1 V 2 V m (1) Definition For each subset S {1, 2,..., m} let V S = s S V s (2) n S = dim V S (3) c S = dim W dim V S. (4) The last integer is called the codimenson of V S.
28 Subspace Arrangements Definition Let W be a vector space over a scalar field F of dimension D. A subspace arrangement A in W is a union of subspaces V 1, V 2,..., V m of W. A = V 1 V 2 V m (1) Transversal Arrangements Definition The arrangmement is said to be transversal if c S = min(dim W, X c i ) (5) i S for all non-empty subsets S of {1, 2,..., m}. This means that the intersections of the subspaces are as small as possible.
29 Subspace Arrangements Definition Let W be a vector space over a scalar field F of dimension D. A subspace arrangement A in W is a union of subspaces V 1, V 2,..., V m of W. A = V 1 V 2 V m (1) Simple Case of two subspaces Suppose V 1, V 2 are subspaces of W. Then or, in terms of codimensions, dim V 1 + dim V 2 = dim(v 1 + V 2 ) + dim(v 1 V 2 ) codimv 1 + codimv 2 = codim(v 1 + V 2 ) + codim(v 1 V 2 ). We conclude that when D codim(v 1 ) + codim(v 2 ), the arrangement is transversal if V 1 V 2 = 0. When D > codim(v 1 ) + codim(v 2 ), then it is transveral if V 1 + V 2 = W.
30 Subspace Arrangements Definition Let W be a vector space over a scalar field F of dimension D. A subspace arrangement A in W is a union of subspaces V 1, V 2,..., V m of W. A = V 1 V 2 V m (1) Problem Given data {z 1, z 2,..., z N } R D, find a subspace arrangement A, the constituent subspaces V 1,..., V m, their bases, and then the segmentation of the data into these constituent subspaces.
31 In the first frame we see a collection of sample points that are arranged on a subspace arrangement pictured in the second frame. The third frame illustrates the separated subspaces.
32 Algebraic sets and Vanishing Ideals Definitions The vector space W = R D. The algebra of multivariate polynomial functions on W is R [D]. = R[X1, X 2,..., X D ]. A function q(x 1,..., X D ) R [D] is homogeneous of degree h if it a linear combination of monomials of the form where h = e 1 + e 2 + e D. X e 1 1 X e 2 2 X e D D The subspace of homogeneous functions of degree h is denoted by R [D] h. There is a decomposition R [D] = R [D] 0 R [D] 1 R [D] h. The dimension dim(r [D] h D 1 + h ) =. h
33 Algebraic sets and Vanishing Ideals Definitions The vector space W = R D. The algebra of multivariate polynomial functions on W is R [D]. = R[X1, X 2,..., X D ]. A function q(x 1,..., X D ) R [D] is homogeneous of degree h if it a linear combination of monomials of the form where h = e 1 + e 2 + e D. X e 1 1 X e 2 2 X e D D The subspace of homogeneous functions of degree h is denoted by R [D] h. There is a decomposition R [D] = R [D] 0 R [D] 1 R [D] h. The dimension dim(r [D] h D 1 + h ) =. h
34 Algebraic sets and Vanishing Ideals Definitions The vector space W = R D. The algebra of multivariate polynomial functions on W is R [D]. = R[X1, X 2,..., X D ]. A function q(x 1,..., X D ) R [D] is homogeneous of degree h if it a linear combination of monomials of the form where h = e 1 + e 2 + e D. X e 1 1 X e 2 2 X e D D The subspace of homogeneous functions of degree h is denoted by R [D] h. There is a decomposition R [D] = R [D] 0 R [D] 1 R [D] h. The dimension dim(r [D] h D 1 + h ) =. h
35 Algebraic sets and Vanishing Ideals Definitions The vector space W = R D. The algebra of multivariate polynomial functions on W is R [D]. = R[X1, X 2,..., X D ]. A function q(x 1,..., X D ) R [D] is homogeneous of degree h if it a linear combination of monomials of the form where h = e 1 + e 2 + e D. X e 1 1 X e 2 2 X e D D The subspace of homogeneous functions of degree h is denoted by R [D] h. There is a decomposition R [D] = R [D] 0 R [D] 1 R [D] h. The dimension dim(r [D] h D 1 + h ) =. h
36 Algebraic sets and Vanishing Ideals Definitions The vector space W = R D. The algebra of multivariate polynomial functions on W is R [D]. = R[X1, X 2,..., X D ]. A function q(x 1,..., X D ) R [D] is homogeneous of degree h if it a linear combination of monomials of the form where h = e 1 + e 2 + e D. X e 1 1 X e 2 2 X e D D The subspace of homogeneous functions of degree h is denoted by R [D] h. There is a decomposition R [D] = R [D] 0 R [D] 1 R [D] h. The dimension dim(r [D] h D 1 + h ) =. h
37 Vanishing Polynomials More Definitions If X W, then a vanishing polynomial on X is a polynomial f R [D]. = R[X1,..., X D ] such that f (x) = 0 for all x X. The vanishing ideal consists of all vanishing polynomials. They form an ideal I(X) = {f R [D] f (x) = 0, for all x X} The ideal generated by the polynomials {f 1,..., f n} is denoted by (f 1,..., f n) and is the set {f 1 g f ng n : g i R [D] for all i}. The zero set of a subset F of R [D] is Z(F ) = {x R D : f (x) = 0 for all f F }. An ideal I in R [D] is said to be homogeneous if the homogeneous component of any function in I is also in I. If I is homogeneous, then there is a decomposition where I h = I R [D]. I = I 0 I 1 I h
38 Vanishing Polynomials More Definitions If X W, then a vanishing polynomial on X is a polynomial f R [D]. = R[X1,..., X D ] such that f (x) = 0 for all x X. The vanishing ideal consists of all vanishing polynomials. They form an ideal I(X) = {f R [D] f (x) = 0, for all x X} The ideal generated by the polynomials {f 1,..., f n} is denoted by (f 1,..., f n) and is the set {f 1 g f ng n : g i R [D] for all i}. The zero set of a subset F of R [D] is Z(F ) = {x R D : f (x) = 0 for all f F }. An ideal I in R [D] is said to be homogeneous if the homogeneous component of any function in I is also in I. If I is homogeneous, then there is a decomposition where I h = I R [D]. I = I 0 I 1 I h
39 Vanishing Polynomials More Definitions If X W, then a vanishing polynomial on X is a polynomial f R [D]. = R[X1,..., X D ] such that f (x) = 0 for all x X. The vanishing ideal consists of all vanishing polynomials. They form an ideal I(X) = {f R [D] f (x) = 0, for all x X} The ideal generated by the polynomials {f 1,..., f n} is denoted by (f 1,..., f n) and is the set {f 1 g f ng n : g i R [D] for all i}. The zero set of a subset F of R [D] is Z(F ) = {x R D : f (x) = 0 for all f F }. An ideal I in R [D] is said to be homogeneous if the homogeneous component of any function in I is also in I. If I is homogeneous, then there is a decomposition where I h = I R [D]. I = I 0 I 1 I h
40 Vanishing Polynomials More Definitions If X W, then a vanishing polynomial on X is a polynomial f R [D]. = R[X1,..., X D ] such that f (x) = 0 for all x X. The vanishing ideal consists of all vanishing polynomials. They form an ideal I(X) = {f R [D] f (x) = 0, for all x X} The ideal generated by the polynomials {f 1,..., f n} is denoted by (f 1,..., f n) and is the set {f 1 g f ng n : g i R [D] for all i}. The zero set of a subset F of R [D] is Z(F ) = {x R D : f (x) = 0 for all f F }. An ideal I in R [D] is said to be homogeneous if the homogeneous component of any function in I is also in I. If I is homogeneous, then there is a decomposition where I h = I R [D]. I = I 0 I 1 I h
41 Vanishing Polynomials More Definitions If X W, then a vanishing polynomial on X is a polynomial f R [D]. = R[X1,..., X D ] such that f (x) = 0 for all x X. The vanishing ideal consists of all vanishing polynomials. They form an ideal I(X) = {f R [D] f (x) = 0, for all x X} The ideal generated by the polynomials {f 1,..., f n} is denoted by (f 1,..., f n) and is the set {f 1 g f ng n : g i R [D] for all i}. The zero set of a subset F of R [D] is Z(F ) = {x R D : f (x) = 0 for all f F }. An ideal I in R [D] is said to be homogeneous if the homogeneous component of any function in I is also in I. If I is homogeneous, then there is a decomposition where I h = I R [D]. I = I 0 I 1 I h
42 Vanishing Polynomials More Definitions If X W, then a vanishing polynomial on X is a polynomial f R [D]. = R[X1,..., X D ] such that f (x) = 0 for all x X. The vanishing ideal consists of all vanishing polynomials. They form an ideal I(X) = {f R [D] f (x) = 0, for all x X} The ideal generated by the polynomials {f 1,..., f n} is denoted by (f 1,..., f n) and is the set {f 1 g f ng n : g i R [D] for all i}. The zero set of a subset F of R [D] is Z(F ) = {x R D : f (x) = 0 for all f F }. An ideal I in R [D] is said to be homogeneous if the homogeneous component of any function in I is also in I. If I is homogeneous, then there is a decomposition where I h = I R [D]. I = I 0 I 1 I h
43 Introduction Subspace Arrangements Noise Robust GPCA Other Recent Advances Future Example Models {0} Ambient Space Geometry Zero Sets {0} V = {x nti x = 0, i = 1,..., c} Algebra Vanishing Ideals I(0) = {f (0) = 0 : f R[D] } ` I(V ) = nt1 x,..., ntc x A = V1 Vn RD I(A) = I(V1 ) I(Vn ) I(RD ) = {0} IMA 09 March 2007 Robert M. Fossum Subspace Arrangements in Theory and Practice
44 Algebraic Sets and their Vanishing Polynomials Basics Theorem (Correspondence between Algebraic Sets and Ideals) Suppose that X R D and that J R [D] is an ideal. Then Z(I(X)) X I(Z(J)) J with equality when X = Z(J) for some ideal J or when J = I(X) for some subset X. Definition A set of the form Z(J) is called an algebraic set.
45 Examples Example (1) Suppose V is a subspace of W (so a simple subspace arrangement). Then V is the set of solutions of a homogeneous system of linear equations. If dim V = n, then the system of equations has codimv linearly independent equations. The corresponding ideal of vanishing functions is the ideal generated by these linear forms. Denote this ideal by I(V ). This is a homogeneous ideal, generated by its homogeneous forms of degree 1. So I(V ) = I 1 (V ) I 2 (V ) where 0 1 x 1 x 2 I 1 (V ) = {c 1 X 1 + c 2 X c D X D c 1 x 1 + c D x D = 0 for all B A V }. x D Example (2) If A is the subspace arrangement A = V 1 V m, then I(A) = I(V 1 ) I(V m).
46 Hilbert Functions Purpose We are given a data set {z 1, z 2,..., z N } whose vanishing ideal we wish to find. We assume or know that the vanishing ideal is homogeneous. We would like to know the dimensions of homogeneous components of this ideal. This dimension has been studied intensely for at least a century. Definition (Hilbert Function) Suppose J is a homogeneous ideal. The function H J : N N with value H J (h) = dim J h is called the Hilbert Function of J.
47 Hilbert Series Putting things together Definition Suppose J is a homogeneous ideal. The Hilbert Series for J, denoted by H J (t) is defined to the the formal power series H J (t) = X h 0 H J (h)t h. Example Let the ideal be the algebra R [D] = R[X 1,..., X D ]. Then D 1 + h H R [D](h) = h H R [D](t) = X D 1 + h t h h h 0 = 1 (1 t) D.
48 Examples Example (of a Subspace) Suppose V is a subspace of R D of dimension d and codimension c = D d. Then I(V ) is generated as an ideal by V. It is relatively easy to see that the Hilbert series of this ideal is. H I(V ) (t) = = 1 (1 t) D 1 (1 t) d 1 (1 t)c (1 t) D
49 Examples Example (of a Subspace Arrangement) Now suppose A = V 1 V m is a subspace arrangement. It follows that I(A) = I(V 1 ) I(V m). This ideal is homogeneous. Another ideal that is also important for us is the product ideal P(A) = I(V 1 )I(V 2 ) I(V m). The reason for this is that the zeros of this ideal comprise the subspace arrangement. Theorem (Ma et.al) The zero set of the product ideal above is A. Z(P(A)) = A.
50 Examples The Hilbert Series for the product ideal is determined by the codimensions of the intersections. Theorem (Derksen) Let A = V 1 V 2 V m be a subspace arrangement. Then the Hilbert Function of the product ideal is a function of the codimensions of the individual subspaces and is given by H P(A) (h) = X ( 1) T D 1 c T + h, D 1 c T T where the sum is over all subsets T {1, 2,..., m} (including the empty subset) and c T = P j T c j. When the arrangement is transversal even more can be said. Theorem (Derksen) If the subspace arrangement is transversal, then the Hilbert functions for the vanishing ideal and the product ideal agree in all dimensions where they are defined. In particular for all h m. I h (A) = P h (A)
51 Product Ideal and Vanishing Ideal This equality tell us that the homogeneous components of the vanishing ideal I h (A) and the product ideal P h (A) agree for all h m in the case of a transversal arrangement. Knowing the values of the Hilbert Functions facilitates the task of finding the correct subspace arrangement model for a given data set. This is especially important if the set consists of noisy or corrupted data. This formula was found by Harm Derksen (U Michigan).
52 Product Ideal and Vanishing Ideal This equality tell us that the homogeneous components of the vanishing ideal I h (A) and the product ideal P h (A) agree for all h m in the case of a transversal arrangement. Knowing the values of the Hilbert Functions facilitates the task of finding the correct subspace arrangement model for a given data set. This is especially important if the set consists of noisy or corrupted data. This formula was found by Harm Derksen (U Michigan).
53 Product Ideal and Vanishing Ideal This equality tell us that the homogeneous components of the vanishing ideal I h (A) and the product ideal P h (A) agree for all h m in the case of a transversal arrangement. Knowing the values of the Hilbert Functions facilitates the task of finding the correct subspace arrangement model for a given data set. This is especially important if the set consists of noisy or corrupted data. This formula was found by Harm Derksen (U Michigan).
54 Table of Hilbert Functions Suppose that A = V 1 V 2 V 3 is a transversal arrangement in R 4. Let d 1, d 2, d 3 (respectively c 1, c 2, c 3 ) be the dimensions (resp. codimensions) of V 1, V 2, V 3. We make a table of h I (h) for h = 3, 4, 5. c 1, c 2, c 3 d 1, d 2, d 3 h I (3) h I (4) h I (5) 1, 1, 1 3, 3, , 1, 2 3, 3, , 1, 3 3, 3, , 2, 2 3, 3, , 2, 3 3, 2, , 3, 3 3, 1, , 2, 2 2, 2, , 2, 3 2, 2, , 3, 3 2, 1, , 3, 3 1, 1, Note that the codimensions c 1, c 2, c 3 are almost determined by h I (3). They are uniquely determined h I (3) and h I (4).
55 Veronese Maps Definition (Veronese Map) For each integer h we define the Veronese of order h by ν h : R D R D 1+h h 0 a h a a h a 2 a 2 ν h B A =. a h 1 1 a. D a D C. A ad h The entries on the right are the monomials in the a i of degree h arranged in some order.
56 Veronese Maps D = 3, h = 3; D = 4, h = 2 0 a a1 2a 2 a1 2a 3 0 ν a 1 a 1 a a 2 A = a 1 a 2 a 3 a 3 a 1 a3 2 a2 3 B a2 a 3 C a 2 a3 2 A a3 3 0 a a 1 a a 1 a 3 a 1 a 1 a 4 ν 2 Ba 2 a 3 A = a 2 2 a 2 a 3 a 4 a 2 a 4 B a3 2 a 3 a 4 A a4 2
57 Homogeneous Polynomials A homogeneous polynomial of degree h is an element where c R c T ν h (X) = X c e1 e D X e 1 1 X e 2 2 X e D D e 1 + e e D = h D 1+h h and X = (X 1,..., X D ) T.
58 Finding Vanishing Polynomials from a Data Set Z = {z 1,..., z N } is a given data set. A basis for the vector space of homogeneous polynomial functions of degree h that vanish on Z is found by finding a basis for the left null space of the `D 1+h h N matrix ` ν h (z 1 ) ν h (z 2 ) ν h (z N )
59 Why knowing the value of h I (n) is crucial in estimating subspace arrangements? 1 H I (m) determines the number of linearly dependent vanishing polynomials of degree m. M = H I (h) {f 1, f 2,, f M } I m = (f 1, f 2,, f M ) A = Z (I m). 2 H I (m) facilitates model selection within a class of candidate arrangement models. Figure: Data space R 3.M [3] 3 = 10. c 1 c 2 c 3 H I (3) I 3 is a subspace of R [3] 3 of dim 10 dim(i 3 ) can only be 1, 2, 4, or 7. Given the subspace dimensions uniquely determine dim(i m). If H I (m) correctly estimated the possible dimensions of the individual subspaces. Reference: Yang, Rao, Wagner, Fossum, & Ma. Hilbert functions and applications to the estimation of subspace arrangements. ICCV, 2005.
60 GPCA: (a walk-through example) x V 1 V 2 (x 3 = 0)or(x 1 = x 2 = 0) I 2 = (x 1 x 3, x 2 x 3 ). Veronese Map: Given N samples x 1,..., x N R 3,. L 2 = [ν2 (x 1 ),..., ν 2 (x N )] R M[3] 2 N 2 3 = 6 4 (x 1 ) 2 (x 1 x 2 ) (x 1 x 3 ) (x 2 ) 2 (x 2 x 3 ) (x 3 ) V1 x1 x3 V2 R 3 x2 The null space of L 2 is c 1 = [0, 0, 1, 0, 0, 0] c 2 = [0, 0, 0, 0, 1, 0] p 1 = c T 1 ν 2(x) = x 1 x 3 p 2 = c T 2 ν 2(x) = x 2 x 3 Punch line 1 The null space of L n give coefficients of the vanishing polynomials. 2 The dimension of Null(L n) is given by h I (n). Vidal, Ma, & Sastry. Generalized Principal Component Analysis, PAMI, 2005.
61 GPCA: a walk-through example (continued) P(x) =. [p 1 (x)p 2 (x)] = [x 1 x 3, x 2 x 3 ], then x p J (P)(x) = 1 x3 0 x x p = x 3 x. 2 Evaluating J (P)(x) at sample points gives normal vectors that span V1 and V 2, respectively: V1 V2 b22 b21 b11 z = (a, b, 0) T V 1 J (P)(x) z = ` 0 0 a 0 0 b. y = (0, 0, c) T V 2 J (P)(x) y = ` c 0 0 c 0 0. Using V1 and V 2 V 1 and V 2. to segment samples and recover Polynomial Differentiation Algorithm (PDA) V1 V2 R D νn(x) = R M[D] n R M[D] n p(x) = c T x Null(Ln) x = = V1 V2 R D Rank(Ln) = M [D] n hi(n) MATLAB code available at:
62 Generalized Principal Component Analysis Algorithm Given a set of (noise free) samples {z 1, z 2,..., z N } from m linear subspaces of dimensions d 1, d 2,..., d m in R D : 1: Construct the matrix L m = `ν m(z 1 ), ν m(z 2 ),..., ν m(z N ). 2: Compute the singular value decomposition (SVD) of L m and let C be the singular vectors associated with the M = H I (m) smallest singular values. 3: Construct the polynomials P(X) = C T ν m(x). 4: for all 1 i m do 5: Pick one point z i per subspace, and compute the Jacobian J (P)(z i ). 6: Compute a basis B i = `b 1, b 2,..., b di of Vi from the right null space of J (P)(z i ) via the singular value decomposition of J (P)(z i ). 7: Assign samples z j that satisfy B T i z j = 0 to the subspace V i. 8: end for
63 Outline 1 Introduction Abstract Motivating Examples of Data Types and Uses 2 Subspace Arrangements Definitions Quick Course on Algebraic Sets Hilbert Functions and Hilbert Series Hilbert Series and Subspace Arrangements 3 Estimating Hybrid Subspace Models from Noisy Samples Stability Estimation of Vanishing Polynomials GPCA-Voting 4 Robust Component Analysis 5 Other Recent Advances Occluded Face Recognition 6 Future Directions Further Investigations Summary and Conclusions Thanks
64 From Ideal to Noisy Noisy Data We have given some detail showing how to use the theory of subspace arrangements to segment a pure set of data into its compoenent subspaces. The problem now is how to handle noisy data. Estimating the vanishing polynomials and subsequently retrieving the subspaces become a statistical problem (which algebraists are not equipped to handle). The data matrix is now usually of full rank. When the vanishing polynomials are found, their Jacobian at noisy sample points no longer span the orthogonal complement to the subspaces. Yang, Ma, et al. have found techniques that apply in this case a subject of Yang s thesis. 1 Noiseless samples from the subspaces. 2 Corrupted by (Gaussian) noise. 3 Contaminated by both noise and outliers.
65 From Ideal to Noisy Noisy Data We have given some detail showing how to use the theory of subspace arrangements to segment a pure set of data into its compoenent subspaces. The problem now is how to handle noisy data. Estimating the vanishing polynomials and subsequently retrieving the subspaces become a statistical problem (which algebraists are not equipped to handle). The data matrix is now usually of full rank. When the vanishing polynomials are found, their Jacobian at noisy sample points no longer span the orthogonal complement to the subspaces. Yang, Ma, et al. have found techniques that apply in this case a subject of Yang s thesis. 1 Noiseless samples from the subspaces. 2 Corrupted by (Gaussian) noise. 3 Contaminated by both noise and outliers.
66 From Ideal to Noisy Noisy Data We have given some detail showing how to use the theory of subspace arrangements to segment a pure set of data into its compoenent subspaces. The problem now is how to handle noisy data. Estimating the vanishing polynomials and subsequently retrieving the subspaces become a statistical problem (which algebraists are not equipped to handle). The data matrix is now usually of full rank. When the vanishing polynomials are found, their Jacobian at noisy sample points no longer span the orthogonal complement to the subspaces. Yang, Ma, et al. have found techniques that apply in this case a subject of Yang s thesis. 1 Noiseless samples from the subspaces. 2 Corrupted by (Gaussian) noise. 3 Contaminated by both noise and outliers.
67 Estimation of Arrangements from Noisy Samples An algorithm is stable if it tolerates noise, robust if it tolerates outliers (together with noise). Polynomial Differentiation Algorithm (PDA) is stable on hyperplanes:[yang] (a) 0% noise (b) 4% noise (c) 8% noise (d) 16% noise However, PDA cannot handle subspaces of different dimensions in noisy data. 1 P(x) estimated by PCA is noisy P(x) 0. 2 x is perturbed away from subspace x P x is not normal to the subspace. Derivatives x P x is always full rank. Best we can do is to treat each V i as hyperplane. (e) a priori, 8% noise (f) estimation result
68 We propose a new algorithm, GPCA-Voting, for subspaces of different dimensions. The noise effects in the process of algebraic GPCA. V1 V2 R D νn(x) = R M[D] n R M[D] n p(x) = c T x Null(Ln) x = = R D V1 V2 Improve the vanishing polynomials, i.e., estimation of Null(L n): 1 L n is always full rank Eigenspace. 2 Uses Fisher discriminant to min within-subspace fitting error and max between-subspace fitting error. (g) eigenvalue (noise-free) (h) eigenvalue (8% noise) Improve the normal vectors: A voting scheme. 1 More accurate estimates of the bases. 2 Determine different dimensions of the subspaces.
69 Improving Vanishing Polynomials Since L n is full rank, replace null space by eigenspace to the smallest eigenvalues. Assuming ν n(x) is zero-mean Gaussian distributed, the last eigenvector c of L n = [ν n(x i )] is optimal for least-square fitting c = arg min ν c n(x) T c. However, it is more practical to assume x to be corrupted by Gaussian noise: x i = ˆx i (c) + n i, i = 1, 2,..., N, where ˆx(c) is the projection of x onto the zero set of the polynomial c. Hence, we want to estimate the mean square distance c 1 = arg min c N NX x i ˆx i (c) 2. i=1
70 Introduction Subspace Arrangements Noise Robust GPCA Other Recent Advances Future A Voting Scheme Goal: Averaging derivatives at all samples of a subspace will produce more stable basis. Difficulty: We do not know which samples belong to the same subspace, yet. GPCA-Voting (a walk-through example) 1 Assume samples are drawn from 3 subspaces of dimensions 2, 1, 1 in R3. 2 Estimate hi (3) = 6 linearly independent 3rd degree polynomials via eig(a, B). 3 Subspace voting on two codimension classes (rank-1 & rank-2) with a tolerance threshold τ 4 Segment the samples based on the estimated bases. 5 (optional) Iteratively refine the segmentation via EM or K-Subspaces. IMA 09 March 2007 Robert M. Fossum Subspace Arrangements in Theory and Practice
71 Simulation Results 1 Improving vanishing polynomials. (i) eigenvalue (noise-free) (j) eigenvalue (8% noise) (k) eig(a,b) (8% noise) 2 Segmentation simulations Table: Segmentation errors. 4% Gaussian noise is added. Subspace Dimensions EM K-Subspaces PDA Voting Voting+K-Subspaces (2, 2, 1) in R 3 29% 27% 13.2% 6.4% 5.4% (4, 2, 2, 1) in R 5 53% 57% 39.8% 5.7% 5.7% (4, 4, 4, 4) in R 5 20% 25% 25.3% 17% 11% (l) 8% (m) 12% (n) 16% Figure: (2, 1, 1) R 3. (a) 8% (b) 12% (c) 16% Figure: (2, 2, 1) R 3.
72 Outline 1 Introduction Abstract Motivating Examples of Data Types and Uses 2 Subspace Arrangements Definitions Quick Course on Algebraic Sets Hilbert Functions and Hilbert Series Hilbert Series and Subspace Arrangements 3 Estimating Hybrid Subspace Models from Noisy Samples Stability Estimation of Vanishing Polynomials GPCA-Voting 4 Robust Component Analysis 5 Other Recent Advances Occluded Face Recognition 6 Future Directions Further Investigations Summary and Conclusions Thanks
73 Noisy Data Corrupted by Outliers There isn t time to discuss the problems involved in estimating the vanishing polynomials in the case of outliers. Again, this is a topic that Yang has handled in his thesis. See Yi Ma, Allen Y. Yang, Harm Derksen, and Robert Fossum. Estimation of Subspace Arrangements with Applications in Modeling and Segmenting Mixed Data. To Appear SIAM Review, 2007.
74 Outline 1 Introduction Abstract Motivating Examples of Data Types and Uses 2 Subspace Arrangements Definitions Quick Course on Algebraic Sets Hilbert Functions and Hilbert Series Hilbert Series and Subspace Arrangements 3 Estimating Hybrid Subspace Models from Noisy Samples Stability Estimation of Vanishing Polynomials GPCA-Voting 4 Robust Component Analysis 5 Other Recent Advances Occluded Face Recognition 6 Future Directions Further Investigations Summary and Conclusions Thanks
75 Problem Given labeled training images X i R m n i i = 1... K of the faces of K individuals, and a new test image, x T, infer the identity of the person in the test image. Under idealized conditions, images of a face (in a fixed pose) under all possible lighting conditions lie on a linear subspace (see fig 8) [?].
76 Figure: Subspaces spanned by face images of three individuals under varying illumination.
77 Recovering Sparse Data Thus, if x T belongs to the i-th individual, then x T X i w for some w R n i. We can then view face recognition as identifying a point on an arrangement of K subspaces, range(x 1 )... range(x K ). Concatenating the training data into one large matrix A = [X 1... X K ] R m n, we have that x T = Aw for some sparse w in R n. How can we efficiently recover such a sparse w? Theorem (Donoho, Candes and Tao) Let A be m n, with m < n, and suppose y = Aw 0 for some sufficiently sparse w 0 ( supp(w) < ρ(a)). Then the sparsest solution w 0 is also the minimum l 1 -norm solution: w 0 = arg min w 1 subject to Aw = y. (6) w and can therefore be computed by linear programming. Thus, minimizing the l 1 norm gives an efficient way of locating a point in the (combinatorially large) subspace arrangement given by all ` n ρ(a) subsets of ρ(a) columns of A.
78 Recognition under occlusion. Now suppose the some fraction of the pixels of the test image are occluded: x T = Aw + e (7) where e is some vector of errors which are arbitrary magnitude, uncorrelated with the image, but concentrated on only a subset of the image pixels. That is, e is sparse as well. We can estimate such simultaneously sparse w, e in the same manner. Rewriting, «w x T = [A I]. (8) e We then wish to locate x T in the subspace arrangement given by sparse linear combinations of the columns of [A I]. The next figure shows some results for images with 30% of the pixels occluded.
79 Figure Image Error Reconstruction Image Error Reconstruction Image Error Reconstruction (a) (b) (c) Figure: Some visual results. (a) Occluded face image. (b) Estimated error, e. (c) Reconstruction, Aw.
80 Coefficients, red > same class Coefficients, red > same class Coefficients, red > same class Figure: Recovered coefficients w for the above three examples. The red entries correspond to training images of the same person. Notice that in all cases the largest coefficient corresponds to the correct identity.
81 Outline 1 Introduction Abstract Motivating Examples of Data Types and Uses 2 Subspace Arrangements Definitions Quick Course on Algebraic Sets Hilbert Functions and Hilbert Series Hilbert Series and Subspace Arrangements 3 Estimating Hybrid Subspace Models from Noisy Samples Stability Estimation of Vanishing Polynomials GPCA-Voting 4 Robust Component Analysis 5 Other Recent Advances Occluded Face Recognition 6 Future Directions Further Investigations Summary and Conclusions Thanks
82 Further Investigation 1 Investigate arrangements of manifolds higher order. 2 Investiage multilinear arrangements. That is arrangements within spaces of the form R D 1 R D 2 R Dr. Video sequences are examples where r 3. 3 Scale the method up to handle larger data sets. 4 Apply to various data sets. As an example, consider a scene in which two or more vehicles with occupants conversing are moving. Try to segment the two vehicles AND the sound from the two vehicles. (Huang)
83 Further Investigation 1 Investigate arrangements of manifolds higher order. 2 Investiage multilinear arrangements. That is arrangements within spaces of the form R D 1 R D 2 R Dr. Video sequences are examples where r 3. 3 Scale the method up to handle larger data sets. 4 Apply to various data sets. As an example, consider a scene in which two or more vehicles with occupants conversing are moving. Try to segment the two vehicles AND the sound from the two vehicles. (Huang)
84 Further Investigation 1 Investigate arrangements of manifolds higher order. 2 Investiage multilinear arrangements. That is arrangements within spaces of the form R D 1 R D 2 R Dr. Video sequences are examples where r 3. 3 Scale the method up to handle larger data sets. 4 Apply to various data sets. As an example, consider a scene in which two or more vehicles with occupants conversing are moving. Try to segment the two vehicles AND the sound from the two vehicles. (Huang)
85 Further Investigation 1 Investigate arrangements of manifolds higher order. 2 Investiage multilinear arrangements. That is arrangements within spaces of the form R D 1 R D 2 R Dr. Video sequences are examples where r 3. 3 Scale the method up to handle larger data sets. 4 Apply to various data sets. As an example, consider a scene in which two or more vehicles with occupants conversing are moving. Try to segment the two vehicles AND the sound from the two vehicles. (Huang)
86 Summary Summary GPCA and its stable and robust versions offer a top down method for analysing data. One studies the algebraic structure first and then studies the individual subspaces using both algebraic and statistical methods. A new greedy algorithm due to Derksen and Ma shows promise in segmenting data. It may be used in conjunction with GPCA and other techniques. The confluence of algebra, statistics, and computation is crucial for a complete and thorough understanding of the modeling of mixed data. Problems Dimensionality Curse. The dimension of the ambient space of the Veronese embedding is huge. Outlier Detection Number of motions (does not work well for large numbers of moving objects).
87 Summary Summary GPCA and its stable and robust versions offer a top down method for analysing data. One studies the algebraic structure first and then studies the individual subspaces using both algebraic and statistical methods. A new greedy algorithm due to Derksen and Ma shows promise in segmenting data. It may be used in conjunction with GPCA and other techniques. The confluence of algebra, statistics, and computation is crucial for a complete and thorough understanding of the modeling of mixed data. Problems Dimensionality Curse. The dimension of the ambient space of the Veronese embedding is huge. Outlier Detection Number of motions (does not work well for large numbers of moving objects).
88 Summary Summary GPCA and its stable and robust versions offer a top down method for analysing data. One studies the algebraic structure first and then studies the individual subspaces using both algebraic and statistical methods. A new greedy algorithm due to Derksen and Ma shows promise in segmenting data. It may be used in conjunction with GPCA and other techniques. The confluence of algebra, statistics, and computation is crucial for a complete and thorough understanding of the modeling of mixed data. Problems Dimensionality Curse. The dimension of the ambient space of the Veronese embedding is huge. Outlier Detection Number of motions (does not work well for large numbers of moving objects).
89 Summary Summary GPCA and its stable and robust versions offer a top down method for analysing data. One studies the algebraic structure first and then studies the individual subspaces using both algebraic and statistical methods. A new greedy algorithm due to Derksen and Ma shows promise in segmenting data. It may be used in conjunction with GPCA and other techniques. The confluence of algebra, statistics, and computation is crucial for a complete and thorough understanding of the modeling of mixed data. Problems Dimensionality Curse. The dimension of the ambient space of the Veronese embedding is huge. Outlier Detection Number of motions (does not work well for large numbers of moving objects).
90 Summary Summary GPCA and its stable and robust versions offer a top down method for analysing data. One studies the algebraic structure first and then studies the individual subspaces using both algebraic and statistical methods. A new greedy algorithm due to Derksen and Ma shows promise in segmenting data. It may be used in conjunction with GPCA and other techniques. The confluence of algebra, statistics, and computation is crucial for a complete and thorough understanding of the modeling of mixed data. Problems Dimensionality Curse. The dimension of the ambient space of the Veronese embedding is huge. Outlier Detection Number of motions (does not work well for large numbers of moving objects).
91 Summary Summary GPCA and its stable and robust versions offer a top down method for analysing data. One studies the algebraic structure first and then studies the individual subspaces using both algebraic and statistical methods. A new greedy algorithm due to Derksen and Ma shows promise in segmenting data. It may be used in conjunction with GPCA and other techniques. The confluence of algebra, statistics, and computation is crucial for a complete and thorough understanding of the modeling of mixed data. Problems Dimensionality Curse. The dimension of the ambient space of the Veronese embedding is huge. Outlier Detection Number of motions (does not work well for large numbers of moving objects).
92 Summary Summary GPCA and its stable and robust versions offer a top down method for analysing data. One studies the algebraic structure first and then studies the individual subspaces using both algebraic and statistical methods. A new greedy algorithm due to Derksen and Ma shows promise in segmenting data. It may be used in conjunction with GPCA and other techniques. The confluence of algebra, statistics, and computation is crucial for a complete and thorough understanding of the modeling of mixed data. Problems Dimensionality Curse. The dimension of the ambient space of the Veronese embedding is huge. Outlier Detection Number of motions (does not work well for large numbers of moving objects).
93 Acknowledgements/Collaborators Acknowledgements Director Pierre Wiltzius and Beckman Institute for its hospitality. Yi Ma, Allen Y. Yang, and others in our group for graciously allowing me to use slides. Frank Sottile for pointing us to subspace arrangements. Bernt Sturmfels for pointing us to the Veronese map. Collaborators Professor Tom Huang (ECE) Professor Yi Ma (ECE) Professor Harm Derksen University of Michigan Professor Rene Vidal at Johns Hopkins Professor Kun Huang at Ohio State Allen Yang, Wei Hong, Shankar Rao, Andrew Wagner, John Wright, Nemanja Petrovic, David Murphy, Joseph Brennan, and many others Supported by NSF CCF-TF
94 THANK YOU!
Segmentation of Subspace Arrangements III Robust GPCA
Segmentation of Subspace Arrangements III Robust GPCA Berkeley CS 294-6, Lecture 25 Dec. 3, 2006 Generalized Principal Component Analysis (GPCA): (an overview) x V 1 V 2 (x 3 = 0)or(x 1 = x 2 = 0) {x 1x
More informationPart I Generalized Principal Component Analysis
Part I Generalized Principal Component Analysis René Vidal Center for Imaging Science Institute for Computational Medicine Johns Hopkins University Principal Component Analysis (PCA) Given a set of points
More informationSparse Subspace Clustering
Sparse Subspace Clustering Based on Sparse Subspace Clustering: Algorithm, Theory, and Applications by Elhamifar and Vidal (2013) Alex Gutierrez CSCI 8314 March 2, 2017 Outline 1 Motivation and Background
More informationESTIMATION OF SUBSPACE ARRANGEMENTS: ITS ALGEBRA AND STATISTICS ALLEN YANG YANG
ESTIMATION OF SUBSPACE ARRANGEMENTS: ITS ALGEBRA AND STATISTICS BY ALLEN YANG YANG B.Eng., University of Science and Technology of China, 2001 M.S., University of Illinois at Urbana-Champaign, 2003 M.S.,
More informationGeneralized Principal Component Analysis (GPCA)
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 27, NO. 12, PAGE 1945-1959, DECEMBER 2005 1 Generalized Principal Component Analysis (GPCA) René Vidal, Member, IEEE, Yi Ma, Member,
More informationSegmentation of Dynamic Scenes from the Multibody Fundamental Matrix
ECCV Workshop on Vision and Modeling of Dynamic Scenes, Copenhagen, Denmark, May 2002 Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix René Vidal Dept of EECS, UC Berkeley Berkeley,
More informationTwo-View Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix
Two-View Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix René Vidal Stefano Soatto Shankar Sastry Department of EECS, UC Berkeley Department of Computer Sciences, UCLA 30 Cory Hall,
More informationThe Multibody Trifocal Tensor: Motion Segmentation from 3 Perspective Views
The Multibody Trifocal Tensor: Motion Segmentation from 3 Perspective Views Richard Hartley 1,2 and RenéVidal 2,3 1 Dept. of Systems Engineering 3 Center for Imaging Science Australian National University
More informationFILTRATED ALGEBRAIC SUBSPACE CLUSTERING
FILTRATED ALGEBRAIC SUBSPACE CLUSTERING MANOLIS C. TSAKIRIS AND RENÉ VIDAL Abstract. Subspace clustering is the problem of clustering data that lie close to a union of linear subspaces. Existing algebraic
More informationA New GPCA Algorithm for Clustering Subspaces by Fitting, Differentiating and Dividing Polynomials
A New GPCA Algorithm for Clustering Subspaces by Fitting, Differentiating and Dividing Polynomials René Vidal Yi Ma Jacopo Piazzi Center for Imaging Science Electrical & Computer Engineering Mechanical
More informationCS 4495 Computer Vision Principle Component Analysis
CS 4495 Computer Vision Principle Component Analysis (and it s use in Computer Vision) Aaron Bobick School of Interactive Computing Administrivia PS6 is out. Due *** Sunday, Nov 24th at 11:55pm *** PS7
More informationMACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA
1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR
More informationGeneralized Principal Component Analysis
This is page i Printer: Opaque this Generalized Principal Component Analysis Estimation & Segmentation of Hybrid Models René Vidal (JOHNS HOPKINS UNIVERSITY) Yi Ma (UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN)
More informationStructured matrix factorizations. Example: Eigenfaces
Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix
More informationProbabilistic Latent Semantic Analysis
Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationScalable Subspace Clustering
Scalable Subspace Clustering René Vidal Center for Imaging Science, Laboratory for Computational Sensing and Robotics, Institute for Computational Medicine, Department of Biomedical Engineering, Johns
More informationCS 231A Section 1: Linear Algebra & Probability Review
CS 231A Section 1: Linear Algebra & Probability Review 1 Topics Support Vector Machines Boosting Viola-Jones face detector Linear Algebra Review Notation Operations & Properties Matrix Calculus Probability
More informationCS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang
CS 231A Section 1: Linear Algebra & Probability Review Kevin Tang Kevin Tang Section 1-1 9/30/2011 Topics Support Vector Machines Boosting Viola Jones face detector Linear Algebra Review Notation Operations
More informationTwo-View Multibody Structure from Motion
International Journal of Computer Vision 68(), 7 5, 006 c 006 Springer Science + Business Media, LLC Manufactured in he Netherlands DOI: 0007/s63-005-4839-7 wo-view Multibody Structure from Motion RENÉ
More informationIntroduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin
1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)
More informationMachine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling
Machine Learning B. Unsupervised Learning B.2 Dimensionality Reduction Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University
More informationIntroduction to Machine Learning
10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what
More informationPCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani
PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given
More informationFace Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition
ace Recognition Identify person based on the appearance of face CSED441:Introduction to Computer Vision (2017) Lecture10: Subspace Methods and ace Recognition Bohyung Han CSE, POSTECH bhhan@postech.ac.kr
More informationExample: Face Detection
Announcements HW1 returned New attendance policy Face Recognition: Dimensionality Reduction On time: 1 point Five minutes or more late: 0.5 points Absent: 0 points Biometrics CSE 190 Lecture 14 CSE190,
More informationPrincipal Component Analysis
Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used
More informationVector Space Concepts
Vector Space Concepts ECE 174 Introduction to Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 25 Vector Space Theory
More informationRobust Motion Segmentation by Spectral Clustering
Robust Motion Segmentation by Spectral Clustering Hongbin Wang and Phil F. Culverhouse Centre for Robotics Intelligent Systems University of Plymouth Plymouth, PL4 8AA, UK {hongbin.wang, P.Culverhouse}@plymouth.ac.uk
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationMachine learning for pervasive systems Classification in high-dimensional spaces
Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version
More informationPrincipal Component Analysis (PCA)
Principal Component Analysis (PCA) Additional reading can be found from non-assessed exercises (week 8) in this course unit teaching page. Textbooks: Sect. 6.3 in [1] and Ch. 12 in [2] Outline Introduction
More informationLinear Subspace Models
Linear Subspace Models Goal: Explore linear models of a data set. Motivation: A central question in vision concerns how we represent a collection of data vectors. The data vectors may be rasterized images,
More informationA Factorization Method for 3D Multi-body Motion Estimation and Segmentation
1 A Factorization Method for 3D Multi-body Motion Estimation and Segmentation René Vidal Department of EECS University of California Berkeley CA 94710 rvidal@eecs.berkeley.edu Stefano Soatto Dept. of Computer
More informationData Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395
Data Mining Dimensionality reduction Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1395 1 / 42 Outline 1 Introduction 2 Feature selection
More informationCollaborative Filtering: A Machine Learning Perspective
Collaborative Filtering: A Machine Learning Perspective Chapter 6: Dimensionality Reduction Benjamin Marlin Presenter: Chaitanya Desai Collaborative Filtering: A Machine Learning Perspective p.1/18 Topics
More informationFace Detection and Recognition
Face Detection and Recognition Face Recognition Problem Reading: Chapter 18.10 and, optionally, Face Recognition using Eigenfaces by M. Turk and A. Pentland Queryimage face query database Face Verification
More informationMachine Learning - MT & 14. PCA and MDS
Machine Learning - MT 2016 13 & 14. PCA and MDS Varun Kanade University of Oxford November 21 & 23, 2016 Announcements Sheet 4 due this Friday by noon Practical 3 this week (continue next week if necessary)
More informationStatistical Geometry Processing Winter Semester 2011/2012
Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian
More informationUncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization
Uncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization Haiping Lu 1 K. N. Plataniotis 1 A. N. Venetsanopoulos 1,2 1 Department of Electrical & Computer Engineering,
More informationReview of Some Concepts from Linear Algebra: Part 2
Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set
More informationSubspace Methods for Visual Learning and Recognition
This is a shortened version of the tutorial given at the ECCV 2002, Copenhagen, and ICPR 2002, Quebec City. Copyright 2002 by Aleš Leonardis, University of Ljubljana, and Horst Bischof, Graz University
More informationGeometric Modeling Summer Semester 2010 Mathematical Tools (1)
Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric
More informationPrincipal Component Analysis -- PCA (also called Karhunen-Loeve transformation)
Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) PCA transforms the original input space into a lower dimensional space, by constructing dimensions that are linear combinations
More informationLecture 3: Review of Linear Algebra
ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,
More informationIV. Matrix Approximation using Least-Squares
IV. Matrix Approximation using Least-Squares The SVD and Matrix Approximation We begin with the following fundamental question. Let A be an M N matrix with rank R. What is the closest matrix to A that
More informationMULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES
MULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES S. Visuri 1 H. Oja V. Koivunen 1 1 Signal Processing Lab. Dept. of Statistics Tampere Univ. of Technology University of Jyväskylä P.O.
More informationNotes on Latent Semantic Analysis
Notes on Latent Semantic Analysis Costas Boulis 1 Introduction One of the most fundamental problems of information retrieval (IR) is to find all documents (and nothing but those) that are semantically
More informationHybrid System Identification via Sparse Polynomial Optimization
2010 American Control Conference Marriott Waterfront, Baltimore, MD, USA June 30-July 02, 2010 WeA046 Hybrid System Identification via Sparse Polynomial Optimization Chao Feng, Constantino M Lagoa and
More informationSingular Value Decompsition
Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost
More informationSTA 414/2104: Lecture 8
STA 414/2104: Lecture 8 6-7 March 2017: Continuous Latent Variable Models, Neural networks With thanks to Russ Salakhutdinov, Jimmy Ba and others Outline Continuous latent variable models Background PCA
More informationCOMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017
COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University PRINCIPAL COMPONENT ANALYSIS DIMENSIONALITY
More informationLinear Regression and Its Applications
Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start
More informationPrincipal Component Analysis and Linear Discriminant Analysis
Principal Component Analysis and Linear Discriminant Analysis Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1/29
More informationNormed & Inner Product Vector Spaces
Normed & Inner Product Vector Spaces ECE 174 Introduction to Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 27 Normed
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationLinear Algebra and Robot Modeling
Linear Algebra and Robot Modeling Nathan Ratliff Abstract Linear algebra is fundamental to robot modeling, control, and optimization. This document reviews some of the basic kinematic equations and uses
More informationAlgebraic Clustering of Affine Subspaces
1 Algebraic Clustering of Affine Subspaces Manolis C. Tsakiris and René Vidal, Fellow, IEEE Abstract Subspace clustering is an important problem in machine learning with many applications in computer vision
More informationMaximum variance formulation
12.1. Principal Component Analysis 561 Figure 12.2 Principal component analysis seeks a space of lower dimensionality, known as the principal subspace and denoted by the magenta line, such that the orthogonal
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction
ECE 521 Lecture 11 (not on midterm material) 13 February 2017 K-means clustering, Dimensionality reduction With thanks to Ruslan Salakhutdinov for an earlier version of the slides Overview K-means clustering
More informationGeneralized Principal Component Analysis (GPCA): an Algebraic Geometric Approach to Subspace Clustering and Motion Segmentation. René Esteban Vidal
Generalized Principal Component Analysis (GPCA): an Algebraic Geometric Approach to Subspace Clustering and Motion Segmentation by René Esteban Vidal B.S. (P. Universidad Católica de Chile) 1995 M.S. (University
More informationFocus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.
Previously Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations y = Ax Or A simply represents data Notion of eigenvectors,
More informationLecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University
Lecture 4: Principal Component Analysis Aykut Erdem May 016 Hacettepe University This week Motivation PCA algorithms Applications PCA shortcomings Autoencoders Kernel PCA PCA Applications Data Visualization
More informationImage Analysis. PCA and Eigenfaces
Image Analysis PCA and Eigenfaces Christophoros Nikou cnikou@cs.uoi.gr Images taken from: D. Forsyth and J. Ponce. Computer Vision: A Modern Approach, Prentice Hall, 2003. Computer Vision course by Svetlana
More informationStatistical Pattern Recognition
Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationLow Rank Matrix Completion Formulation and Algorithm
1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9
More informationDimension Reduction Techniques. Presented by Jie (Jerry) Yu
Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage
More informationCoding the Matrix Index - Version 0
0 vector, [definition]; (2.4.1): 68 2D geometry, transformations in, [lab]; (4.15.0): 196-200 A T (matrix A transpose); (4.5.4): 157 absolute value, complex number; (1.4.1): 43 abstract/abstracting, over
More informationIntroduction PCA classic Generative models Beyond and summary. PCA, ICA and beyond
PCA, ICA and beyond Summer School on Manifold Learning in Image and Signal Analysis, August 17-21, 2009, Hven Technical University of Denmark (DTU) & University of Copenhagen (KU) August 18, 2009 Motivation
More informationMachine Learning 2nd Edition
INTRODUCTION TO Lecture Slides for Machine Learning 2nd Edition ETHEM ALPAYDIN, modified by Leonardo Bobadilla and some parts from http://www.cs.tau.ac.il/~apartzin/machinelearning/ The MIT Press, 2010
More informationLinear Algebra Practice Problems
Linear Algebra Practice Problems Page of 7 Linear Algebra Practice Problems These problems cover Chapters 4, 5, 6, and 7 of Elementary Linear Algebra, 6th ed, by Ron Larson and David Falvo (ISBN-3 = 978--68-78376-2,
More informationCS 540: Machine Learning Lecture 1: Introduction
CS 540: Machine Learning Lecture 1: Introduction AD January 2008 AD () January 2008 1 / 41 Acknowledgments Thanks to Nando de Freitas Kevin Murphy AD () January 2008 2 / 41 Administrivia & Announcement
More informationPrincipal Component Analysis
B: Chapter 1 HTF: Chapter 1.5 Principal Component Analysis Barnabás Póczos University of Alberta Nov, 009 Contents Motivation PCA algorithms Applications Face recognition Facial expression recognition
More informationLecture: Face Recognition and Feature Reduction
Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the
More informationLecture 7: Con3nuous Latent Variable Models
CSC2515 Fall 2015 Introduc3on to Machine Learning Lecture 7: Con3nuous Latent Variable Models All lecture slides will be available as.pdf on the course website: http://www.cs.toronto.edu/~urtasun/courses/csc2515/
More informationDimensionality Reduction Using the Sparse Linear Model: Supplementary Material
Dimensionality Reduction Using the Sparse Linear Model: Supplementary Material Ioannis Gkioulekas arvard SEAS Cambridge, MA 038 igkiou@seas.harvard.edu Todd Zickler arvard SEAS Cambridge, MA 038 zickler@seas.harvard.edu
More informationDimension Reduction. David M. Blei. April 23, 2012
Dimension Reduction David M. Blei April 23, 2012 1 Basic idea Goal: Compute a reduced representation of data from p -dimensional to q-dimensional, where q < p. x 1,...,x p z 1,...,z q (1) We want to do
More informationseries. Utilize the methods of calculus to solve applied problems that require computational or algebraic techniques..
1 Use computational techniques and algebraic skills essential for success in an academic, personal, or workplace setting. (Computational and Algebraic Skills) MAT 203 MAT 204 MAT 205 MAT 206 Calculus I
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationLecture: Face Recognition and Feature Reduction
Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed
More informationThe University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.
The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment Two Caramanis/Sanghavi Due: Tuesday, Feb. 19, 2013. Computational
More informationMachine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang.
Machine Learning CUNY Graduate Center, Spring 2013 Lectures 11-12: Unsupervised Learning 1 (Clustering: k-means, EM, mixture models) Professor Liang Huang huang@cs.qc.cuny.edu http://acl.cs.qc.edu/~lhuang/teaching/machine-learning
More informationVectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =
Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.
More informationLecture 3: Review of Linear Algebra
ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak, scribe: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters,
More informationSPECTRAL CLUSTERING AND KERNEL PRINCIPAL COMPONENT ANALYSIS ARE PURSUING GOOD PROJECTIONS
SPECTRAL CLUSTERING AND KERNEL PRINCIPAL COMPONENT ANALYSIS ARE PURSUING GOOD PROJECTIONS VIKAS CHANDRAKANT RAYKAR DECEMBER 5, 24 Abstract. We interpret spectral clustering algorithms in the light of unsupervised
More informationCovariance-Based PCA for Multi-Size Data
Covariance-Based PCA for Multi-Size Data Menghua Zhai, Feiyu Shi, Drew Duncan, and Nathan Jacobs Department of Computer Science, University of Kentucky, USA {mzh234, fsh224, drew, jacobs}@cs.uky.edu Abstract
More informationSEQUENTIAL SUBSPACE FINDING: A NEW ALGORITHM FOR LEARNING LOW-DIMENSIONAL LINEAR SUBSPACES.
SEQUENTIAL SUBSPACE FINDING: A NEW ALGORITHM FOR LEARNING LOW-DIMENSIONAL LINEAR SUBSPACES Mostafa Sadeghi a, Mohsen Joneidi a, Massoud Babaie-Zadeh a, and Christian Jutten b a Electrical Engineering Department,
More informationCompressed Sensing Meets Machine Learning - Classification of Mixture Subspace Models via Sparse Representation
Compressed Sensing Meets Machine Learning - Classification of Mixture Subspace Models via Sparse Representation Allen Y Yang Feb 25, 2008 UC Berkeley What is Sparsity Sparsity A
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationIdentification of PWARX Hybrid Models with Unknown and Possibly Different Orders
Identification of PWARX Hybrid Models with Unknown and Possibly Different Orders René Vidal Center for Imaging Science, Department of Biomedical Engineering, Johns Hopkins University 308B Clark Hall, 3400
More informationMath 4A Notes. Written by Victoria Kala Last updated June 11, 2017
Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...
More informationStatistical Machine Learning
Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 218 Outlines Overview Introduction Linear Algebra Probability Linear Regression 1
More informationA Statistical Analysis of Fukunaga Koontz Transform
1 A Statistical Analysis of Fukunaga Koontz Transform Xiaoming Huo Dr. Xiaoming Huo is an assistant professor at the School of Industrial and System Engineering of the Georgia Institute of Technology,
More informationChemometrics: Classification of spectra
Chemometrics: Classification of spectra Vladimir Bochko Jarmo Alander University of Vaasa November 1, 2010 Vladimir Bochko Chemometrics: Classification 1/36 Contents Terminology Introduction Big picture
More informationA Multi-Affine Model for Tensor Decomposition
Yiqing Yang UW Madison breakds@cs.wisc.edu A Multi-Affine Model for Tensor Decomposition Hongrui Jiang UW Madison hongrui@engr.wisc.edu Li Zhang UW Madison lizhang@cs.wisc.edu Chris J. Murphy UC Davis
More informationLEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach
LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach Dr. Guangliang Chen February 9, 2016 Outline Introduction Review of linear algebra Matrix SVD PCA Motivation The digits
More informationRobust Principal Component Analysis
ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M
More informationUnsupervised learning: beyond simple clustering and PCA
Unsupervised learning: beyond simple clustering and PCA Liza Rebrova Self organizing maps (SOM) Goal: approximate data points in R p by a low-dimensional manifold Unlike PCA, the manifold does not have
More information