Machine Learning. Spectral Clustering. Lecture 23, April 14, Reading: Eric Xing 1

Similar documents
Correspondence Analysis & Related Methods

Set of square-integrable function 2 L : function space F

Khintchine-Type Inequalities and Their Applications in Optimization

Machine Learning 4771

Learning the structure of Bayesian belief networks

Multistage Median Ranked Set Sampling for Estimating the Population Median

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

Thermodynamics of solids 4. Statistical thermodynamics and the 3 rd law. Kwangheon Park Kyung Hee University Department of Nuclear Engineering

8 Baire Category Theorem and Uniform Boundedness

Clustering. Outline. Supervised vs. Unsupervised Learning. Clustering. Clustering Example. Applications of Clustering

4 Recursive Linear Predictor

4 SingularValue Decomposition (SVD)

Chapter Fifiteen. Surfaces Revisited

Physics 11b Lecture #2. Electric Field Electric Flux Gauss s Law

V. Principles of Irreversible Thermodynamics. s = S - S 0 (7.3) s = = - g i, k. "Flux": = da i. "Force": = -Â g a ik k = X i. Â J i X i (7.

Classification. Representing data: Hypothesis (classifier) Lecture 2, September 14, Reading: Eric CMU,

P 365. r r r )...(1 365

Background. 3D Object recognition. Modeling polyhedral objects. Modeling polyhedral objects. Object and world coordinate systems

General Variance Covariance Structures in Two-Way Random Effects Models

PHYS 705: Classical Mechanics. Derivation of Lagrange Equations from D Alembert s Principle

A. Thicknesses and Densities

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13

A Novel Approach to Expression Recognition from Non-frontal Face Images

Linear Approximation with Regularization and Moving Least Squares

DISC-GLASSO: DISCRIMINATIVE GRAPH LEARNING WITH SPARSITY REGULARIZATION. 201 Broadway, Cambridge, MA 02139, USA

Density Functional Theory I

Using DP for hierarchical discretization of continuous attributes. Amit Goyal (31 st March 2008)

Spectral Clustering. Shannon Quinn

PHY126 Summer Session I, 2008

Pattern Analyses (EOF Analysis) Introduction Definition of EOFs Estimation of EOFs Inference Rotated EOFs

Detection and Estimation Theory

Energy in Closed Systems

Vibration Input Identification using Dynamic Strain Measurement

Generating Functions, Weighted and Non-Weighted Sums for Powers of Second-Order Recurrence Sequences

If there are k binding constraints at x then re-label these constraints so that they are the first k constraints.

Dirichlet Mixture Priors: Inference and Adjustment

Research Article Incremental Tensor Principal Component Analysis for Handwritten Digit Recognition

6.6 The Marquardt Algorithm

ρ some λ THE INVERSE POWER METHOD (or INVERSE ITERATION) , for , or (more usually) to

19 The Born-Oppenheimer Approximation

CS 468 Lecture 16: Isometry Invariance and Spectral Techniques

Chapter I Matrices, Vectors, & Vector Calculus 1-1, 1-9, 1-10, 1-11, 1-17, 1-18, 1-25, 1-27, 1-36, 1-37, 1-41.

Integral Vector Operations and Related Theorems Applications in Mechanics and E&M

3. A Review of Some Existing AW (BT, CT) Algorithms

Kernel Methods and SVMs Extension

Errors for Linear Systems

Impact of Polarimetric Dimensionality of Forest Parameter Estimation by Means of Polarimetric SAR interferometry

CS649 Sensor Networks IP Track Lecture 3: Target/Source Localization in Sensor Networks

Lossy Compression. Compromise accuracy of reconstruction for increased compression.

Part V: Velocity and Acceleration Analysis of Mechanisms

Announcements. Stereo (Part 3) Summary of Stereo Constraints. Features on same epipolar line. Stereo matching. Truco Fig. 7.5

Scalars and Vectors Scalar

Physics Exam II Chapters 25-29

Error Bars in both X and Y

Physics 2A Chapter 11 - Universal Gravitation Fall 2017

Information Retrieval

Stanford University CS259Q: Quantum Computing Handout 8 Luca Trevisan October 18, 2012

Efficiency of the principal component Liu-type estimator in logistic

Asymptotic Waves for a Non Linear System

Lecture 10 Support Vector Machines. Oct

Information Retrieval Advanced IR models. Luca Bondi

Contact, information, consultations

COORDINATE SYSTEMS, COORDINATE TRANSFORMS, AND APPLICATIONS

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

Rigid Bodies: Equivalent Systems of Forces

Experimental study on parameter choices in norm-r support vector regression machines with noisy input

Robust Feature Induction for Support Vector Machines

EE 5337 Computational Electromagnetics (CEM)

Constraint Score: A New Filter Method for Feature Selection with Pairwise Constraints

Tian Zheng Department of Statistics Columbia University

1 Matrix representations of canonical matrices

LECTURE 9 CANONICAL CORRELATION ANALYSIS

Backward Haplotype Transmission Association (BHTA) Algorithm. Tian Zheng Department of Statistics Columbia University. February 5 th, 2002

Potential Fields in Cooperative Motion Control and Formations

Norms, Condition Numbers, Eigenvalues and Eigenvectors

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

Event Shape Update. T. Doyle S. Hanlon I. Skillicorn. A. Everett A. Savin. Event Shapes, A. Everett, U. Wisconsin ZEUS Meeting, October 15,

Optimization Methods: Linear Programming- Revised Simplex Method. Module 3 Lecture Notes 5. Revised Simplex Method, Duality and Sensitivity analysis

Report on Image warping

VParC: A Compression Scheme for Numeric Data in Column-Oriented Databases

Multipole Radiation. March 17, 2014

Singular Value Decomposition: Theory and Applications

Tutorial Chemical Reaction Engineering:

Anouncements. Multigrid Solvers. Multigrid Solvers. Multigrid Solvers. Multigrid Solvers. Multigrid Solvers

Solutions to exam in SF1811 Optimization, Jan 14, 2015

CS4495/6495 Introduction to Computer Vision. 3C-L3 Calibrating cameras

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

Effective Discriminative Feature Selection with Non-trivial Solutions

( m is the length of columns of A ) spanned by the columns of A : . Select those columns of B that contain a pivot; say those are Bi

Exact Simplification of Support Vector Solutions

An Approach to Inverse Fuzzy Arithmetic

Ranks of quotients, remainders and p-adic digits of matrices

Information Retrieval (Relevance Feedback & Query Expansion)

COMPLEMENTARY ENERGY METHOD FOR CURVED COMPOSITE BEAMS

Test 1 phy What mass of a material with density ρ is required to make a hollow spherical shell having inner radius r i and outer radius r o?

Generalized Linear Methods

SOME NEW SELF-DUAL [96, 48, 16] CODES WITH AN AUTOMORPHISM OF ORDER 15. KEYWORDS: automorphisms, construction, self-dual codes

E For K > 0. s s s s Fall Physical Chemistry (II) by M. Lim. singlet. triplet

Pushdown Automata (PDAs)

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 16

Transcription:

Machne Leanng -7/5 7/5-78, 78, Spng 8 Spectal Clusteng Ec Xng Lectue 3, pl 4, 8 Readng: Ec Xng Data Clusteng wo dffeent ctea Compactness, e.g., k-means, mxtue models Connectvty, e.g., spectal clusteng Compactness Connectvty Ec Xng

Ec Xng 3 Spectal Clusteng Data Smlates Ec Xng 4

Weghted Gaph Pattonng Some gaph temnology Objects (e.g., pxels, data ponts) W j j I = vetces of gaph G Edges (j) = pxel pas wth W j > Smlaty matx W = [ W j ] Degee d = Σ j G W j d = Σ d degee of G B ssoc(, = Σ Σ j B W j Ec Xng 5 Cuts n a Gaph (edge) cut = set of edges whose emoval makes a gaph dsconnected weght of a cut: cut(, B ) = Σ Σ j B W j =ssoc(, Nomalzed Cut ctea: mnmum cut(,ā) cut(, ) cut(, ) Ncut(, ) = + d d Moe geneally: k Ncut(, K k ) = =, j V \, j V W j = W j = cut(, ) d Ec Xng 6 k 3

Gaph-based Clusteng Data Goupng W j j W j W = f d( x, x )) j ( j G = {V,E} Image sgmentaton ffnty matx: Degee matx: Laplacan matx: W = [ w, j ] D = dag( d ) L = D W (bpatte) patton vecto: x = [ x,...,x = [,,K,,, K ] Ec Xng 7 N ] ffnty Functon W = e, j X X j σ ffntes gow as σ gows How the choce of σ value affects the esults? What would be the optmal choce fo σ? Ec Xng 8 4

Clusteng va Optmzng Nomalzed Cut he nomalzed cut: cut(, cut(, Ncut(, = + d d B Computng an optmal nomalzed cut ove all possble y (.e., patton) s NP had ansfom Ncut equaton to a matx fom (Sh & Malk ): mn x y ( D W) y Ncut( x) = mny y Dy n y {, b} y D = Subject to: Raylegh quotent cut(, cut(, Stll an NP had poblem Ncut(, = + deg() deg( ( + x) ( D S)( + x) ( x) ( D S)( x) > D(, ) x = + ; k = k D ( k) D D(, ) Ec Xng =... 9 Relaxaton mn Ncut( x) = mn x Subject to: y y {, b} y D = y ( D W) y y Dy n Raylegh quotent Instead, elax nto the contnuous doman by solvng genealzed egenvalue system: mn y ( D W) y, s.t. y Dy = y Whch gves: ( D W ) y = λdy Raylegh quotent theoem ( D W ) = Note that so, the fst egenvecto s y = wth egenvalue. he second smallest egenvecto s the eal valued soluton to ths poblem!! Ec Xng 5

lgothm. Defne a smlaty functon between nodes..e.: w = e, j X ( ) X ( j ) σ X. Compute affnty matx (W) and degee matx (D). 3. Solve ( D W ) y = λdy Do sngula value decomposton (SVD) of the gaph Laplacan L V * = ΛV y L = D W 4. Use the egenvecto wth the second smallest egenvalue,, to bpatton the gaph. Fo each theshold k, k ={ y among k lagest element of y*} B k ={ y among n-k smallest element of y*} Compute Ncut( k,b k ) * Output k = ag max Ncut( k, B k ) and k *, B * Ec Xng k y * Ideally y ( D S) y Ncut(, =, wth y {, b}, y D =. y Dy y ( D S) y = λdy y Ec Xng 6

Example (Xng et al, ) Ec Xng 3 Poo featues can lead to poo outcome (Xng et al ) Ec Xng 4 7

Cluste vs. Block matx cut(, cut(, Ncut(, = + d Degee( ) = d B W, j V, j B cut(, cut(, Ncut(, = + d d B B Ec Xng 5 Compae to Mnmum cut Cteon fo patton: mn cut(, = mn Poblem! Weght Weght of of cut cut s s dectly dectly popotonal popotonal to to the the numbe numbe of of edges edges n n the the cut. cut., B W, j B, j B Ideal Cut Cuts wth lesse weght than the deal cut Fst poposed by Wu and Leahy Ec Xng 6 8

Supeo Pefomance? K-means and Gaussan mxtue methods ae based towad convex clustes Ec Xng 7 Ncut s supeo n cetan cases Ec Xng 8 9

Why? Ec Xng 9 Geneal Spectal Clusteng Data Smlates Ec Xng

Repesentaton Patton matx X: [ X ] X =,..., X K segments pxels Pa-wse smlaty matx W: W (, j) = aff (, j) Degee matx D: Laplacan matx L: D w j, j (, ) = L = D W Ec Xng Egenvectos and blocks Block matces have block egenvectos: λ = λ = λ 3 =.7 λ 4 = egensolve.7.7.7 Nea-block matces have nea-block egenvectos: λ =. λ =. λ 3 = -...7 λ 4 = -.. -. egensolve.69.4 -.4.69 -..7 Ec Xng

Spectal Space Can put tems nto blocks by egenvectos: e..7 -..69 -.4..4.69 e -..7 e e Clustes clea egadless of ow odeng: e..7..4.69 -..69 -.4 e -..7 e e Ec Xng 3 Spectal Clusteng lgothms that cluste ponts usng egenvectos of matces deved fom the data Obtan data epesentaton n the low-dmensonal space that can be easly clusteed Vaety of methods that use the egenvectos dffeently (we have seen an example) Empcally vey successful uthos dsagee: Whch egenvectos to use How to deve clustes fom these egenvectos wo geneal methods Ec Xng 4

Method # Patton usng only one egenvecto at a tme Use pocedue ecusvely Example: Image Segmentaton Uses nd (smallest) egenvecto to defne optmal cut Recusvely geneates two clustes wth each cut Ec Xng 5 Method # Use k egenvectos (k chosen by use) Dectly compute k-way pattonng Expementally has been seen to be bette Ec Xng 6 3

Spectal Clusteng lgothm Ng, Jodan, and Wess 3 Gven a set of ponts S={s, s n } Fom the affnty matx w, j = e S S j σ, j, w, = Defne dagonal matx D = Σ κ a k Fom the matx L = D WD / / Stack the k lagest egenvectos of L to fo the columns of the new matx X: X = x x Renomalze each of X s ows to have unt length and get new matx Y. Cluste ows of Y as ponts n R k Ec Xng 7 L x k SC vs Kmeans Ec Xng 8 4

5 Ec Xng 9 Why t woks? K-means n the spectum space! Ec Xng 3 Moe fomally Recall genealzed Ncut Mnmzng ths s equvalent to spectal clusteng = = = = k k V j j V j j k d W W ), cut( ), Ncut(, \, K = = k k d ), cut( ), Ncut( mn K Y WD D Y / / mn I Y Y = s.t. segments pxels Y

oy examples Images fom Matthew Band (R--4) Ec Xng 3 Use s Peogatve Choce of k, the numbe of clustes Choce of scalng facto σ Realstcally, seach ove and pck value that gves the tghtest clustes Choce of clusteng method: k-way o ecusve bpatte Kenel affnty matx w = K( S, S, j j ) Ec Xng 3 6

Conclusons Good news: Smple and poweful methods to segment mages. Flexble and easy to apply to othe clusteng poblems. Bad news: Hgh memoy equements (use spase matces). Vey dependant on the scale facto fo a specfc poblem. W X ( ) X ( j ) σ X (, j) = e Ec Xng 33 7