Outline. EM Algorithm and its Applications. K-Means Classifier. K-Means Classifier (Cont.) Introduction of EM K-Means EM EM Applications.

Similar documents
Mixture of Gaussians Expectation Maximization (EM) Part 2

Mixture o f of Gaussian Gaussian clustering Nov

Lecture Nov

Classification Bayesian Classifiers

Expectation Maximization Mixture Models HMMs

Machine Learning. Classification. Theory of Classification and Nonparametric Classifier. Representing data: Hypothesis (classifier) Eric Xing

Hidden Markov Model Cheat Sheet

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015

Gaussian Mixture Models

Chapter Newton s Method

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin

Multilayer Perceptron (MLP)

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition

Natural Images, Gaussian Mixtures and Dead Leaves Supplementary Material

10-701/ Machine Learning, Fall 2005 Homework 3

Regularized Discriminant Analysis for Face Recognition

Gaussian Conditional Random Field Network for Semantic Segmentation - Supplementary Material

Clustering Techniques for Information Retrieval

I529: Machine Learning in Bioinformatics (Spring 2017) Markov Models

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

6 Supplementary Materials

Web-Mining Agents Probabilistic Information Retrieval

Unified Subspace Analysis for Face Recognition

CHAPTER 4d. ROOTS OF EQUATIONS

EM and Structure Learning

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Chapter 7 Clustering Analysis (1)

Course 395: Machine Learning - Lectures

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics

Hopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen

Clustering & Unsupervised Learning

Machine Learning for Signal Processing Linear Gaussian Models

Bayesian Decision Theory

Lecture 2 Solution of Nonlinear Equations ( Root Finding Problems )

Pattern Classification (II) 杜俊

Image classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them?

: Numerical Analysis Topic 2: Solution of Nonlinear Equations Lectures 5-11:

Pattern Recognition. Approximating class densities, Bayesian classifier, Errors in Biometric Systems

Clustering & (Ken Kreutz-Delgado) UCSD

Semi-supervised Classification with Active Query Selection

An Experiment/Some Intuition (Fall 2006): Lecture 18 The EM Algorithm heads coin 1 tails coin 2 Overview Maximum Likelihood Estimation

Retrieval Models: Language models

Line Drawing and Clipping Week 1, Lecture 2

Generalized Linear Methods

MATH 567: Mathematical Techniques in Data Science Lab 8

VQ widely used in coding speech, image, and video

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

CSC401/2511 Spring CSC401/2511 Natural Language Computing Spring 2019 Lecture 5 Frank Rudzicz and Chloé Pou-Prom University of Toronto

Grover s Algorithm + Quantum Zeno Effect + Vaidman

Lecture Notes on Linear Regression

Instance-Based Learning and Clustering

Automatic Object Trajectory- Based Motion Recognition Using Gaussian Mixture Models

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

Video Data Analysis. Video Data Analysis, B-IT

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester

Confidence intervals for weighted polynomial calibrations

Non-Ideality Through Fugacity and Activity

Ensemble Methods: Boosting

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980

Large-Margin HMM Estimation for Speech Recognition

Object Localization by Subspace Clustering of Local Descriptors

Newton s Method for One - Dimensional Optimization - Theory

CHAPTER-5 INFORMATION MEASURE OF FUZZY MATRIX AND FUZZY BINARY RELATION

Generative classification models

Bayesian classification CISC 5800 Professor Daniel Leeds

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities

Tracking with Kalman Filter

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018

Clustering with Gaussian Mixtures

Outline. Clustering: Similarity-Based Clustering. Supervised Learning vs. Unsupervised Learning. Clustering. Applications of Clustering

Segmentation Method of MRI Using Fuzzy Gaussian Basis Neural Network

Lecture 12: Classification

Graphical Models and Conditional Random Fields

Spatial Statistics and Analysis Methods (for GEOG 104 class).

A quantum-statistical-mechanical extension of Gaussian mixture model

Bayesian networks for scenario analysis of nuclear waste repositories

Generative and Discriminative Models. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

CISE301: Numerical Methods Topic 2: Solution of Nonlinear Equations

A New Scrambling Evaluation Scheme based on Spatial Distribution Entropy and Centroid Difference of Bit-plane

En Route Traffic Optimization to Reduce Environmental Impact

Relevance Vector Machines Explained

RBF Neural Network Model Training by Unscented Kalman Filter and Its Application in Mechanical Fault Diagnosis

Machine Learning: and 15781, 2003 Assignment 4

Classification learning II

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

An Accurate Heave Signal Prediction Using Artificial Neural Network

Probability Theory (revisited)

Boostrapaggregating (Bagging)

Probability Density Function Estimation by different Methods

An Improved multiple fractal algorithm

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

The Basic Idea of EM

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

Efficient, General Point Cloud Registration with Kernel Feature Maps

Expectation Maximization Mixture Models

Transcription:

EM Algorthm and ts Alcatons Y L Deartment of omuter Scence and Engneerng Unversty of Washngton utlne Introducton of EM K-Means EM EM Alcatons Image Segmentaton usng EM bect lass Recognton n BIR olor lusterng by K-means Algorthm K-Means lassfer Form K-means clusters from a set of n-dmensonal vectors. Set c teraton count to 2. hoose randomly a set of K means m,, m K. 3. For each vector, comute D,m k c, k,k and assgn to the cluster wth nearest mean. 4. Increment c by, udate the means to get m c,,m K c. 5. Reeat stes 3 and 4 untl k c k c for all k. {r, g, b } 2 {r 2, g 2, b 2 } {r, g, b } lassfer K-Means lassfcaton Results 2 2 luster Parameters m for m 2 for 2 m k for k K-Means lassfer ont. {r, g, b } Inut Known {r, g, b } 2 {r 2, g 2, b 2 } {r, g, b } luster Parameters m for m 2 for 2 m k for k utut Unknown lassfcaton Results 2 2 Inut Known 2 {r 2, g 2, b 2 } {r, g, b } Intal Guess of luster Parameters m, m 2,, m k utut Unknown lassfcaton Results, 2,, luster Parameters m, m 2,, m k lassfcaton Results 2 luster Parameters 2 m, m 2,, m k, 2,, luster Parameters c m, m 2,, m k lassfcaton Results c, 2,,

K-Means ont. K-Means Eamle Boot Ste: Intalze K clusters:,, K Each luster s reresented by ts mean m Iteraton Ste: Estmate the cluster of each data Re-estmate the cluster arameters m mean } { K-Means Eamle K-Means EM Boot Ste: Intalze K clusters:,, K, Σ and P for each cluster. Iteraton Ste: Estmate the cluster of each data Re-estmate the cluster arameters, Σ, For each cluster Eectaton Mamzaton EM lassfer EM lassfer ont. {r, g, b } 2 {r 2, g 2, b 2 } {r, g, b } lassfer EM lassfcaton Results 2 luster Parameters,Σ, for 2,Σ 2, 2 for 2 Inut Known {r, g, b } 2 {r 2, g 2, b 2 } {r, g, b } luster Parameters,Σ, for 2,Σ 2, 2 for 2 k,σ k, k for k utut Unknown lassfcaton Results 2 k,σ k, k for k 2

3 {r, g, b } 2 {r 2, g 2, b 2 } {r, g, b } luster Parameters,Σ, for 2,Σ 2, 2 for 2 k,σ k, k for k Eectaton Ste Inut Known Inut Estmaton utut lassfcaton Results 2 {r, g, b } 2 {r 2, g 2, b 2 } {r, g, b } luster Parameters,Σ, for 2,Σ 2, 2 for 2 k,σ k, k for k Mamzaton Ste Σ T Inut Known Inut Estmaton utut lassfcaton Results 2 EM Algorthm Boot Ste: Intalze K clusters:,, K Iteraton Ste: Eectaton Ste Mamzaton Ste, Σ and P for each cluster. Σ T EM Demo Demo htt://www.neurosc.ast.go./~akaho/mtureem.html Eamle htt://www-2.cs.cmu.edu/~awm/tutorals/gmm3.df EM Alcatons Blobworld: Image Segmentaton Usng Eectaton-Mamzaton and ts Alcaton to Image Queryng Image Segmentaton usng EM Ste : Feature Etracton Ste 2: Image Segmentaton usng EM

Symbols The feature vector for el s called. There are gong to be K segments; K s gven. The -th segment has a Gaussan dstrbuton wth arameters θ,σ. α 's are the weghts whch sum to of Gaussans. Θ s the collecton of arameters: Θ α,, α k, θ,, θ k Intalzaton Each of the K Gaussans wll have arameters θ,σ, where s the mean of the -th Gaussan. Σ s the covarance matr of the -th Gaussan. The covarance matrces are ntaled to be the dentty matr. The means can be ntalzed by fndng the average feature vectors n each of K wndows n the mage; ths s data-drven ntalzaton. E-Ste, Θ α f θ K k k f θ d / 2 2π Σ k α f θ T Σ 2 e / 2 k M-Ste new Σ new, Θ, Θ, Θ, Θ new new T α new, Θ Samle Results bect lass Recognton n BIR The Goal: Automatc mage labelng annotaton to enable obect-based mage retreval 4

Problem Statement Known: Some mages and ther corresondng descrtons Abstract Regons rgnal Images olor Regons Teture Regons Lne lusters {trees, grass, cherry trees} {cheetah, trunk} {mountans, sky} {beach, sky, trees, water} To solve: What obect classes are resent n new mages bect Model Learnng Ideal Boat, Water, Sky,! Sky Tree Sky boat buldng sky tree Water Tree Water Boat water Boat Learned Models boat regon attrbutes obect bect Model Learnng Real Model Intal Estmaton Sky Estmate the ntal model of an obect usng all the regon features from all mages that contan the obect {sky, tree, water, boat} Tree Water Boat Tree Learned Models regon attrbutes obect Sky 5

EM Varant bect Model Learnng Intal Model for trees Intal Model for sky EM Fnal Model for trees Fnal Model for sky Assumtons The feature dstrbuton of each obect wthn a regon s a Gaussan; Each mage s a set of regons, each of whch can be modeled as a mture of multvarate Gaussan dstrbutons.. Intalzaton Ste Eamle Image & descrton I 2 I 2 3 I 3 2 3 2. Iteraton Ste Eamle I I 2 I 2 2 3 3 3 E-Ste q q 2 q 3 0 q 0 q 2 0 q 3 W0.8 W0.2 W0.2 W0.8 W0.2 W0.8 M-Ste W0.8 W0.2 W0.8 W0.2 W0.2 W0.8 q q q 3 2 Image Labelng Test Image To calculate tree mage olor Regons tree a a tree mage ma tree o FI ma o r a r a F I tree tree comare bect Model Database Tree Sky Eerments 860 mages 8 keywords: mountans 30, orangutan 37, track 40, tree trunk 43, football feld 43, beach 45, rare grass 53, cherry tree 53, snow 54, zebra 56, olar bear 56, lon 7, water 76, chmanzee 79, cheetah 2, sky 259, grass 272, tree 36. A set of cross-valdaton eerments 80% as the tranng set and the other 20% as the test set 6

R harts Samle Results True Postve Rate 0.8 0.6 0.4 0.2 True Postve Rate 0.8 0.6 0.4 0.2 cheetah 0 0 0.2 0.4 0.6 0.8 False Postve Rate Indeendent Treatment of olor and Teture 0 0 0.2 0.4 0.6 0.8 False Postve Rate Usng Intersecton of olor and Teture Samle Results ont. Samle Results ont. grass lon 7