Mean-Shift Tracker Computer Vision (Kris Kitani) Carnegie Mellon University

Size: px
Start display at page:

Download "Mean-Shift Tracker Computer Vision (Kris Kitani) Carnegie Mellon University"

Transcription

1 Mean-Shift Tracker Computer Vision (Kris Kitani) Carnegie Mellon University

2

3 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975)

4 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Find the region of highest density

5 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Pick a point

6 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Draw a window

7 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Compute the mean

8 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Shift the window

9 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Compute the mean

10 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Shift the window

11 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Compute the mean

12 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Shift the window

13 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Compute the mean

14 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Shift the window

15 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Compute the mean

16 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Shift the window

17 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Compute the mean

18 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Shift the window

19 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Compute the mean

20 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Shift the window

21 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Compute the mean

22 Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Shift the window

23 Kernel Density Estimation Computer Vision (Kris Kitani) Carnegie Mellon University

24 To understand the mean shift algorithm Kernel Density Estimation Approximate the underlying PDF from samples Put bump on every sample to approximate the PDF

25 p(x) probability density function cumulative density function

26 1 randomly sample 0

27 1 randomly sample 0

28 1 randomly sample 0 samples

29 place Gaussian bumps on the samples samples

30 samples

31 samples

32 samples

33 Kernel Density Estimate approximates the original PDF samples

34 Kernel Density Estimation Approximate the underlying PDF from samples from it p(x) = X i (x x i ) 2 c i e 2 2 Gaussian bump aka kernel Put bump on every sample to approximate the PDF

35 Kernel Function K(x, x 0 ) a distance between two points

36 Epanechnikov kernel K(x, x 0 )= c(1 kx x 0 k 2 ) kx x 0 k 2 apple 1 0 otherwise Uniform kernel c kx x 0 k 2 apple 1 K(x, x 0 )= 0 otherwise Normal kernel K(x, x 0 )=c exp 1 2 kx x0 k 2 Radially symmetric kernels

37 Radially symmetric kernels can be written in terms of its profile K(x, x 0 )=c k(kx x 0 k 2 ) profile

38 Connecting KDE and the Mean Shift Algorithm

39 Mean-Shift Tracker Computer Vision (Kris Kitani) Carnegie Mellon University

40 Mean-Shift Tracking Given a set of points: {x s } S s=1 x s 2R d and a kernel: K(x, x 0 ) Find the mean sample point: x

41 Mean-Shift Algorithm Initialize x While v(x) > 1. Compute mean-shift P m(x) = s K(x, x s)x P s s K(x, x s) v(x) =m(x) x 2. Update x x + v(x) Where does this algorithm come from?

42 Mean-Shift Algorithm Initialize x While v(x) > 1. Compute mean-shift P m(x) = s K(x, x s)x P s s K(x, x s) Where does this come from? v(x) =m(x) x 2. Update x x + v(x) Where does this algorithm come from?

43 How is the KDE related to the mean shift algorithm? Recall: Kernel density estimate (radially symmetric kernels) P (x) = 1 N c X n k(kx x n k 2 ) We can show that: Gradient of the PDF is related to the mean shift vector rp (x) / m(x) The mean shift is a step in the direction of the gradient of the KDE

44 In mean-shift tracking, we are trying to find this which means we are trying to

45 We are trying to optimize this: x = arg max x = arg max x P (x) 1 N c X n k( x x n 2 ) usually non-linear non-parametric How do we optimize this non-linear function?

46 We are trying to optimize this: x = arg max x = arg max x P (x) 1 N c X n k( x x n 2 ) usually non-linear non-parametric How do we optimize this non-linear function? compute partial derivatives, gradient descent

47 P (x) = 1 N c X n k(kx x n k 2 ) Compute the gradient

48 P (x) = 1 N c X n k(kx x n k 2 ) Gradient rp (x) = 1 N c X n rk(kx x n k 2 ) Expand the gradient (algebra)

49 P (x) = 1 N c X n k(kx x n k 2 ) Gradient rp (x) = 1 N c X n rk(kx x n k 2 ) Expand gradient rp (x) = 1 N 2c X n (x x n )k 0 (kx x n k 2 )

50 P (x) = 1 N c X n k(kx x n k 2 ) Gradient rp (x) = 1 N c X n rk(kx x n k 2 ) Expand gradient rp (x) = 1 N 2c X n (x x n )k 0 (kx x n k 2 ) Call the gradient of the kernel function g k 0 ( ) = g( )

51 P (x) = 1 N c X n k(kx x n k 2 ) Gradient rp (x) = 1 N c X n rk(kx x n k 2 ) Expand gradient rp (x) = 1 N 2c X n (x x n )k 0 (kx x n k 2 ) change of notation (kernel-shadow pairs) rp (x) = 1 N 2c X n (x n x)g(kx x n k 2 ) keep this in memory: k 0 ( ) = g( )

52 rp (x) = 1 N 2c X n (x n x)g(kx x n k 2 ) multiply it out rp (x) = 1 N 2c X n x n g(kx x n k 2 ) 1 xg(kx x n k 2 ) N 2c X n too long! (use short hand notation) rp (x) = 1 N 2c X n x n g n 1 N 2c X n xg n

53 rp (x) = 1 N 2c X n x n g n 1 N 2c X n xg n rp (x) = 1 N 2c X n multiply by one! x n g n P P n g n n g n 1 N 2c X n xg n rp (x) = 1 N 2c X n collecting like terms P g n x ng n n P n g n x Does this look familiar?

54 mean shift rp (x) = 1 N 2c X n P g n x ng n n P n g n x { constant { mean shift! The mean shift is a step in the direction of the gradient of the KDE m(x) v(x) = P n x ng n P n g n x = rp (x) 1 N 2c P n g n Gradient ascent with adaptive step size

55 Mean-Shift Algorithm Initialize x While v(x) > 1. Compute mean-shift P m(x) = s K(x, x s)x P s s K(x, x s) v(x) =m(x) x 2. Update x x + v(x) gradient with adaptive step size rp (x) 1 N 2c P n g n

56 Everything up to now has been about distributions over samples

57 Dealing with images Pixels for a lattice, spatial density is the same everywhere! What can we do?

58 Consider a set of points: {x s } S s=1 x s 2R d Associated weights: w(x s ) Sample mean: m(x) = P s K(x, x s)w(x s )x s P s K(x, x s)w(x s ) Mean shift: m(x) x

59 Mean-Shift Algorithm Initialize x While v(x) > 1. Compute mean-shift P m(x) = s K(x, x s)w(x s )x P s s K(x, x s)w(x s ) v(x) =m(x) x 2. Update x x + v(x)

60 For images, each pixel is point with a weight

61 For images, each pixel is point with a weight

62 For images, each pixel is point with a weight

63 For images, each pixel is point with a weight

64 For images, each pixel is point with a weight

65 For images, each pixel is point with a weight

66 For images, each pixel is point with a weight

67 For images, each pixel is point with a weight

68 For images, each pixel is point with a weight

69 For images, each pixel is point with a weight

70 For images, each pixel is point with a weight

71 For images, each pixel is point with a weight

72 For images, each pixel is point with a weight

73 For images, each pixel is point with a weight

74 Finally mean shift tracking in video

75 Goal: find the best candidate location in frame 2 x center coordinate of target center coordinate y of candidate target candidate there are many candidates but only one target Frame 1 Frame 2 Use the mean shift algorithm to find the best candidate location

76 Non-rigid object tracking

77 Compute a descriptor for the target Target

78 Search for similar descriptor in neighborhood in next frame Target Candidate

79 Compute a descriptor for the new target Target

80 Search for similar descriptor in neighborhood in next frame Target Candidate

81 How do we model the target and candidate regions?

82 Modeling the target M-dimensional target descriptor q = {q 1,...,q M } (centered at target center) a fancy (confusing) way to write a weighted histogram q m = C X n Normalization factor sum over all pixels k(kx n k 2 ) [b(x n ) m] function of inverse distance (weight) Kronecker delta function quantization function bin ID A normalized color histogram (weighted by distance)

83 Modeling the candidate M-dimensional candidate descriptor p(y) ={p 1 (y),...,p M (y} (centered at location y) y 0 a weighted histogram at y p m = C h X n k y x n h 2! [b(x n ) m] bandwidth

84 Similarity between the target and candidate Distance function d(y) = p 1 [p(y), q] Bhattacharyya Coefficient (y) [p(y), q] = X m p pm (y)q u Just the Cosine distance between two unit vectors p(y) (y) = cos y = p(y)> q kpkkqk = X m p pm (y)q m q

85 Now we can compute the similarity between a target and multiple candidate regions

86 q target p(y) [p(y), q] image similarity over image

87 q target we want to find this peak p(y) [p(y), q] image similarity over image

88 Objective function min y d(y) same as max y [p(y), q] Assuming a good initial guess [p(y 0 + y), q] Linearize around the initial guess (Taylor series expansion) [p(y), q] 1 2 X p pm (y 0 )q m + 1 X r qm p m (y) 2 p m (y 0 ) m m function at specified value derivative

89 Linearized objective [p(y), q] 1 2 X m p pm (y 0 )q m X m p m (y) r qm p m (y 0 ) p m = C h X n k y x n h 2! [b(x n ) m] Remember definition of this? [p(y), q] 1 2 X m Fully expanded ( p pm (y 0 )q m + 1 X X C h k 2 m n y x n h 2! [b(x n ) m] ) r qm p m (y 0 )

90 [p(y), q] 1 2 X m Fully expanded linearized objective ( p pm (y 0 )q m + 1 X X C h k 2 m n y x n h 2! [b(x n ) m] ) r qm p m (y 0 ) [p(y), q] 1 2 X m Moving terms around p pm (y 0 )q m + C h 2 X n w n k y x n h 2! Does not depend on unknown y X r where w n = m Weighted kernel density estimate qm p m (y 0 ) [b(x n) m] Weight is bigger when q m >p m (y 0 )

91 OK, why are we doing all this math?

92 We want to maximize this max y [p(y), q]

93 We want to maximize this max y [p(y), q] [p(y), q] 1 2 X m where Fully expanded linearized objective p pm (y 0 )q m + C h 2 w n = X m r qm X n w n k y x n h p m (y 0 ) [b(x n) m] 2!

94 We want to maximize this max y [p(y), q] [p(y), q] 1 2 X m where Fully expanded linearized objective p pm (y 0 )q m + C h 2 doesn t depend on unknown y w n = X m r qm X n w n k y x n h p m (y 0 ) [b(x n) m] 2!

95 We want to maximize this max y [p(y), q] [p(y), q] 1 2 X m where Fully expanded linearized objective p pm (y 0 )q m + C h 2 doesn t depend on unknown y w n = X m r qm X n w n k y x n h p m (y 0 ) [b(x n) m] only need to maximize this! 2!

96 We want to maximize this max y [p(y), q] [p(y), q] 1 2 X m where Fully expanded linearized objective p pm (y 0 )q m + C h 2 doesn t depend on unknown y w n = X m r qm X n w n k y x n h p m (y 0 ) [b(x n) m] 2! Mean Shift Algorithm! what can we use to solve this weighted KDE?

97 C h 2 X n w n k y h x n 2! the new sample of mean of this KDE is y 1 = (new candidate location) P n x nw n g P n w ng y0 y0 x n h x n h 2 2 (this was derived earlier)

98 Mean-Shift Object Tracking For each frame: 1. Initialize location Compute q Compute p(y 0 ) y 0 2. Derive weights w n 3. Shift to new candidate location (mean shift) y 1 4. Compute p(y 1 ) 5. If ky 0 y 1 k < return Otherwise and go back to 2 y 0 y 1

99 Compute a descriptor for the target Target q

100 Search for similar descriptor in neighborhood in next frame Target Candidate max y [p(y), q]

101 Compute a descriptor for the new target Target q

102 Search for similar descriptor in neighborhood in next frame Target Candidate max y [p(y), q]

103

CS 556: Computer Vision. Lecture 21

CS 556: Computer Vision. Lecture 21 CS 556: Computer Vision Lecture 21 Prof. Sinisa Todorovic sinisa@eecs.oregonstate.edu 1 Meanshift 2 Meanshift Clustering Assumption: There is an underlying pdf governing data properties in R Clustering

More information

Kalman Filter Computer Vision (Kris Kitani) Carnegie Mellon University

Kalman Filter Computer Vision (Kris Kitani) Carnegie Mellon University Kalman Filter 16-385 Computer Vision (Kris Kitani) Carnegie Mellon University Examples up to now have been discrete (binary) random variables Kalman filtering can be seen as a special case of a temporal

More information

CS 534: Computer Vision Segmentation III Statistical Nonparametric Methods for Segmentation

CS 534: Computer Vision Segmentation III Statistical Nonparametric Methods for Segmentation CS 534: Computer Vision Segmentation III Statistical Nonparametric Methods for Segmentation Ahmed Elgammal Dept of Computer Science CS 534 Segmentation III- Nonparametric Methods - - 1 Outlines Density

More information

Video and Motion Analysis Computer Vision Carnegie Mellon University (Kris Kitani)

Video and Motion Analysis Computer Vision Carnegie Mellon University (Kris Kitani) Video and Motion Analysis 16-385 Computer Vision Carnegie Mellon University (Kris Kitani) Optical flow used for feature tracking on a drone Interpolated optical flow used for super slow-mo optical flow

More information

Lucas-Kanade Optical Flow. Computer Vision Carnegie Mellon University (Kris Kitani)

Lucas-Kanade Optical Flow. Computer Vision Carnegie Mellon University (Kris Kitani) Lucas-Kanade Optical Flow Computer Vision 16-385 Carnegie Mellon University (Kris Kitani) I x u + I y v + I t =0 I x = @I @x I y = @I u = dx v = dy I @y t = @I dt dt @t spatial derivative optical flow

More information

Improved Fast Gauss Transform. Fast Gauss Transform (FGT)

Improved Fast Gauss Transform. Fast Gauss Transform (FGT) 10/11/011 Improved Fast Gauss Transform Based on work by Changjiang Yang (00), Vikas Raykar (005), and Vlad Morariu (007) Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Kernel Density Estimation, Factor Analysis Mark Schmidt University of British Columbia Winter 2017 Admin Assignment 2: 2 late days to hand it in tonight. Assignment 3: Due Feburary

More information

Chapter 9. Non-Parametric Density Function Estimation

Chapter 9. Non-Parametric Density Function Estimation 9-1 Density Estimation Version 1.1 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least

More information

Non-parametric Methods

Non-parametric Methods Non-parametric Methods Machine Learning Torsten Möller Möller/Mori 1 Reading Chapter 2 of Pattern Recognition and Machine Learning by Bishop (with an emphasis on section 2.5) Möller/Mori 2 Outline Last

More information

Regression with Numerical Optimization. Logistic

Regression with Numerical Optimization. Logistic CSG220 Machine Learning Fall 2008 Regression with Numerical Optimization. Logistic regression Regression with Numerical Optimization. Logistic regression based on a document by Andrew Ng October 3, 204

More information

More on Estimation. Maximum Likelihood Estimation.

More on Estimation. Maximum Likelihood Estimation. More on Estimation. In the previous chapter we looked at the properties of estimators and the criteria we could use to choose between types of estimators. Here we examine more closely some very popular

More information

12 - Nonparametric Density Estimation

12 - Nonparametric Density Estimation ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning 3. Instance Based Learning Alex Smola Carnegie Mellon University http://alex.smola.org/teaching/cmu2013-10-701 10-701 Outline Parzen Windows Kernels, algorithm Model selection

More information

Chapter 9. Non-Parametric Density Function Estimation

Chapter 9. Non-Parametric Density Function Estimation 9-1 Density Estimation Version 1.2 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least

More information

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall Machine Learning Gaussian Mixture Models Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall 2012 1 The Generative Model POV We think of the data as being generated from some process. We assume

More information

Improved Fast Gauss Transform

Improved Fast Gauss Transform Improved Fast Gauss Transform Changjiang Yang Department of Computer Science University of Maryland Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to efficiently evaluate

More information

Review for Exam 1. Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA

Review for Exam 1. Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA Review for Exam Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA 0003 March 26, 204 Abstract Here are some things you need to know for the in-class

More information

Improved Fast Gauss Transform. (based on work by Changjiang Yang and Vikas Raykar)

Improved Fast Gauss Transform. (based on work by Changjiang Yang and Vikas Raykar) Improved Fast Gauss Transform (based on work by Changjiang Yang and Vikas Raykar) Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to efficiently evaluate the weighted sum

More information

Image Alignment Computer Vision (Kris Kitani) Carnegie Mellon University

Image Alignment Computer Vision (Kris Kitani) Carnegie Mellon University Lucas Kanade Image Alignment 16-385 Comuter Vision (Kris Kitani) Carnegie Mellon University htt://www.humansensing.cs.cmu.edu/intraface/ How can I find in the image? Idea #1: Temlate Matching Slow, combinatory,

More information

CS8803: Statistical Techniques in Robotics Byron Boots. Hilbert Space Embeddings

CS8803: Statistical Techniques in Robotics Byron Boots. Hilbert Space Embeddings CS8803: Statistical Techniques in Robotics Byron Boots Hilbert Space Embeddings 1 Motivation CS8803: STR Hilbert Space Embeddings 2 Overview Multinomial Distributions Marginal, Joint, Conditional Sum,

More information

Generative v. Discriminative classifiers Intuition

Generative v. Discriminative classifiers Intuition Logistic Regression Machine Learning 070/578 Carlos Guestrin Carnegie Mellon University September 24 th, 2007 Generative v. Discriminative classifiers Intuition Want to Learn: h:x a Y X features Y target

More information

Nonparametric Methods

Nonparametric Methods Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis

More information

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning 12. Gaussian Processes Alex Smola Carnegie Mellon University http://alex.smola.org/teaching/cmu2013-10-701 10-701 The Normal Distribution http://www.gaussianprocess.org/gpml/chapters/

More information

Gaussian mean-shift algorithms

Gaussian mean-shift algorithms Gaussian mean-shift algorithms Miguel Á. Carreira-Perpiñán Dept. of Computer Science & Electrical Engineering, OGI/OHSU http://www.csee.ogi.edu/~miguel Gaussian mean-shift (GMS) as an EM algorithm Gaussian

More information

Robert Collins CSE598G Mean-Shift Blob Tracking through Scale Space

Robert Collins CSE598G Mean-Shift Blob Tracking through Scale Space Mean-Shift Blob Tracking through Scale Space Robert Collins, CVPR 03 Abstract Mean-shift tracking Choosing scale of kernel is an issue Scale-space feature selection provides inspiration Perform mean-shift

More information

Modelling Non-linear and Non-stationary Time Series

Modelling Non-linear and Non-stationary Time Series Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September

More information

CMU-Q Lecture 24:

CMU-Q Lecture 24: CMU-Q 15-381 Lecture 24: Supervised Learning 2 Teacher: Gianni A. Di Caro SUPERVISED LEARNING Hypotheses space Hypothesis function Labeled Given Errors Performance criteria Given a collection of input

More information

Midterm exam CS 189/289, Fall 2015

Midterm exam CS 189/289, Fall 2015 Midterm exam CS 189/289, Fall 2015 You have 80 minutes for the exam. Total 100 points: 1. True/False: 36 points (18 questions, 2 points each). 2. Multiple-choice questions: 24 points (8 questions, 3 points

More information

Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University. September 20, 2012

Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University. September 20, 2012 Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University September 20, 2012 Today: Logistic regression Generative/Discriminative classifiers Readings: (see class website)

More information

Mixture Models and EM

Mixture Models and EM Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering

More information

Gaussian Mixture Models

Gaussian Mixture Models Gaussian Mixture Models Pradeep Ravikumar Co-instructor: Manuela Veloso Machine Learning 10-701 Some slides courtesy of Eric Xing, Carlos Guestrin (One) bad case for K- means Clusters may overlap Some

More information

Expectation maximization

Expectation maximization Expectation maximization Subhransu Maji CMSCI 689: Machine Learning 14 April 2015 Motivation Suppose you are building a naive Bayes spam classifier. After your are done your boss tells you that there is

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

EKF and SLAM. McGill COMP 765 Sept 18 th, 2017

EKF and SLAM. McGill COMP 765 Sept 18 th, 2017 EKF and SLAM McGill COMP 765 Sept 18 th, 2017 Outline News and information Instructions for paper presentations Continue on Kalman filter: EKF and extension to mapping Example of a real mapping system:

More information

An Adaptive Test of Independence with Analytic Kernel Embeddings

An Adaptive Test of Independence with Analytic Kernel Embeddings An Adaptive Test of Independence with Analytic Kernel Embeddings Wittawat Jitkrittum 1 Zoltán Szabó 2 Arthur Gretton 1 1 Gatsby Unit, University College London 2 CMAP, École Polytechnique ICML 2017, Sydney

More information

An Adaptive Test of Independence with Analytic Kernel Embeddings

An Adaptive Test of Independence with Analytic Kernel Embeddings An Adaptive Test of Independence with Analytic Kernel Embeddings Wittawat Jitkrittum Gatsby Unit, University College London wittawat@gatsby.ucl.ac.uk Probabilistic Graphical Model Workshop 2017 Institute

More information

A Discriminatively Trained, Multiscale, Deformable Part Model

A Discriminatively Trained, Multiscale, Deformable Part Model A Discriminatively Trained, Multiscale, Deformable Part Model P. Felzenszwalb, D. McAllester, and D. Ramanan Edward Hsiao 16-721 Learning Based Methods in Vision February 16, 2009 Images taken from P.

More information

Bias-Variance Tradeoff

Bias-Variance Tradeoff What s learning, revisited Overfitting Generative versus Discriminative Logistic Regression Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University September 19 th, 2007 Bias-Variance Tradeoff

More information

Shape of Gaussians as Feature Descriptors

Shape of Gaussians as Feature Descriptors Shape of Gaussians as Feature Descriptors Liyu Gong, Tianjiang Wang and Fang Liu Intelligent and Distributed Computing Lab, School of Computer Science and Technology Huazhong University of Science and

More information

2D Image Processing (Extended) Kalman and particle filter

2D Image Processing (Extended) Kalman and particle filter 2D Image Processing (Extended) Kalman and particle filter Prof. Didier Stricker Dr. Gabriele Bleser Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz

More information

Generative v. Discriminative classifiers Intuition

Generative v. Discriminative classifiers Intuition Logistic Regression Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University September 24 th, 2007 1 Generative v. Discriminative classifiers Intuition Want to Learn: h:x a Y X features

More information

Support Vector Machines

Support Vector Machines Two SVM tutorials linked in class website (please, read both): High-level presentation with applications (Hearst 1998) Detailed tutorial (Burges 1998) Support Vector Machines Machine Learning 10701/15781

More information

3.1. QUADRATIC FUNCTIONS AND MODELS

3.1. QUADRATIC FUNCTIONS AND MODELS 3.1. QUADRATIC FUNCTIONS AND MODELS 1 What You Should Learn Analyze graphs of quadratic functions. Write quadratic functions in standard form and use the results to sketch graphs of functions. Find minimum

More information

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means

More information

Nearest Neighbor. Machine Learning CSE546 Kevin Jamieson University of Washington. October 26, Kevin Jamieson 2

Nearest Neighbor. Machine Learning CSE546 Kevin Jamieson University of Washington. October 26, Kevin Jamieson 2 Nearest Neighbor Machine Learning CSE546 Kevin Jamieson University of Washington October 26, 2017 2017 Kevin Jamieson 2 Some data, Bayes Classifier Training data: True label: +1 True label: -1 Optimal

More information

Machine Learning (CSE 446): Multi-Class Classification; Kernel Methods

Machine Learning (CSE 446): Multi-Class Classification; Kernel Methods Machine Learning (CSE 446): Multi-Class Classification; Kernel Methods Sham M Kakade c 2018 University of Washington cse446-staff@cs.washington.edu 1 / 12 Announcements HW3 due date as posted. make sure

More information

Announcements. Proposals graded

Announcements. Proposals graded Announcements Proposals graded Kevin Jamieson 2018 1 Bayesian Methods Machine Learning CSE546 Kevin Jamieson University of Washington November 1, 2018 2018 Kevin Jamieson 2 MLE Recap - coin flips Data:

More information

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means

More information

Expectation Propagation Algorithm

Expectation Propagation Algorithm Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,

More information

Mean Shift is a Bound Optimization

Mean Shift is a Bound Optimization IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL.??, NO.??, MONTH YEAR 1 Mean Shift is a Bound Optimization Mark Fashing and Carlo Tomasi, Member, IEEE M. Fashing and C. Tomasi are with

More information

The Kernel Trick, Gram Matrices, and Feature Extraction. CS6787 Lecture 4 Fall 2017

The Kernel Trick, Gram Matrices, and Feature Extraction. CS6787 Lecture 4 Fall 2017 The Kernel Trick, Gram Matrices, and Feature Extraction CS6787 Lecture 4 Fall 2017 Momentum for Principle Component Analysis CS6787 Lecture 3.1 Fall 2017 Principle Component Analysis Setting: find the

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

Kaggle.

Kaggle. Administrivia Mini-project 2 due April 7, in class implement multi-class reductions, naive bayes, kernel perceptron, multi-class logistic regression and two layer neural networks training set: Project

More information

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

LOCAL SEARCH. Today. Reading AIMA Chapter , Goals Local search algorithms. Introduce adversarial search 1/31/14

LOCAL SEARCH. Today. Reading AIMA Chapter , Goals Local search algorithms. Introduce adversarial search 1/31/14 LOCAL SEARCH Today Reading AIMA Chapter 4.1-4.2, 5.1-5.2 Goals Local search algorithms n hill-climbing search n simulated annealing n local beam search n genetic algorithms n gradient descent and Newton-Rhapson

More information

L5 Support Vector Classification

L5 Support Vector Classification L5 Support Vector Classification Support Vector Machine Problem definition Geometrical picture Optimization problem Optimization Problem Hard margin Convexity Dual problem Soft margin problem Alexander

More information

Visual Tracking via Geometric Particle Filtering on the Affine Group with Optimal Importance Functions

Visual Tracking via Geometric Particle Filtering on the Affine Group with Optimal Importance Functions Monday, June 22 Visual Tracking via Geometric Particle Filtering on the Affine Group with Optimal Importance Functions Junghyun Kwon 1, Kyoung Mu Lee 1, and Frank C. Park 2 1 Department of EECS, 2 School

More information

Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis

Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 4 Spatial Point Patterns Definition Set of point locations with recorded events" within study

More information

10-701/15-781, Machine Learning: Homework 4

10-701/15-781, Machine Learning: Homework 4 10-701/15-781, Machine Learning: Homewor 4 Aarti Singh Carnegie Mellon University ˆ The assignment is due at 10:30 am beginning of class on Mon, Nov 15, 2010. ˆ Separate you answers into five parts, one

More information

Comments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms

Comments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms Neural networks Comments Assignment 3 code released implement classification algorithms use kernels for census dataset Thought questions 3 due this week Mini-project: hopefully you have started 2 Example:

More information

Linear Models for Regression CS534

Linear Models for Regression CS534 Linear Models for Regression CS534 Prediction Problems Predict housing price based on House size, lot size, Location, # of rooms Predict stock price based on Price history of the past month Predict the

More information

Outline. Supervised Learning. Hong Chang. Institute of Computing Technology, Chinese Academy of Sciences. Machine Learning Methods (Fall 2012)

Outline. Supervised Learning. Hong Chang. Institute of Computing Technology, Chinese Academy of Sciences. Machine Learning Methods (Fall 2012) Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Linear Models for Regression Linear Regression Probabilistic Interpretation

More information

MATH 2250 Final Exam Solutions

MATH 2250 Final Exam Solutions MATH 225 Final Exam Solutions Tuesday, April 29, 28, 6: 8:PM Write your name and ID number at the top of this page. Show all your work. You may refer to one double-sided sheet of notes during the exam

More information

CS325 Artificial Intelligence Chs. 18 & 4 Supervised Machine Learning (cont)

CS325 Artificial Intelligence Chs. 18 & 4 Supervised Machine Learning (cont) CS325 Artificial Intelligence Cengiz Spring 2013 Model Complexity in Learning f(x) x Model Complexity in Learning f(x) x Let s start with the linear case... Linear Regression Linear Regression price =

More information

Non-parametric Methods

Non-parametric Methods Non-parametric Methods Machine Learning Alireza Ghane Non-Parametric Methods Alireza Ghane / Torsten Möller 1 Outline Machine Learning: What, Why, and How? Curve Fitting: (e.g.) Regression and Model Selection

More information

Diffeomorphic Warping. Ben Recht August 17, 2006 Joint work with Ali Rahimi (Intel)

Diffeomorphic Warping. Ben Recht August 17, 2006 Joint work with Ali Rahimi (Intel) Diffeomorphic Warping Ben Recht August 17, 2006 Joint work with Ali Rahimi (Intel) What Manifold Learning Isn t Common features of Manifold Learning Algorithms: 1-1 charting Dense sampling Geometric Assumptions

More information

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem

More information

CS 4495 Computer Vision Principle Component Analysis

CS 4495 Computer Vision Principle Component Analysis CS 4495 Computer Vision Principle Component Analysis (and it s use in Computer Vision) Aaron Bobick School of Interactive Computing Administrivia PS6 is out. Due *** Sunday, Nov 24th at 11:55pm *** PS7

More information

Machine Learning Lecture 5

Machine Learning Lecture 5 Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory

More information

Machine Learning

Machine Learning Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 1, 2011 Today: Generative discriminative classifiers Linear regression Decomposition of error into

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 2, 2015 Today: Logistic regression Generative/Discriminative classifiers Readings: (see class website)

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should

More information

Density Estimation (II)

Density Estimation (II) Density Estimation (II) Yesterday Overview & Issues Histogram Kernel estimators Ideogram Today Further development of optimization Estimating variance and bias Adaptive kernels Multivariate kernel estimation

More information

CS 147: Computer Systems Performance Analysis

CS 147: Computer Systems Performance Analysis CS 147: Computer Systems Performance Analysis Summarizing Variability and Determining Distributions CS 147: Computer Systems Performance Analysis Summarizing Variability and Determining Distributions 1

More information

Review Smoothing Spatial Filters Sharpening Spatial Filters. Spatial Filtering. Dr. Praveen Sankaran. Department of ECE NIT Calicut.

Review Smoothing Spatial Filters Sharpening Spatial Filters. Spatial Filtering. Dr. Praveen Sankaran. Department of ECE NIT Calicut. Spatial Filtering Dr. Praveen Sankaran Department of ECE NIT Calicut January 7, 203 Outline 2 Linear Nonlinear 3 Spatial Domain Refers to the image plane itself. Direct manipulation of image pixels. Figure:

More information

(Extended) Kalman Filter

(Extended) Kalman Filter (Extended) Kalman Filter Brian Hunt 7 June 2013 Goals of Data Assimilation (DA) Estimate the state of a system based on both current and all past observations of the system, using a model for the system

More information

Computational Physics (6810): Session 13

Computational Physics (6810): Session 13 Computational Physics (6810): Session 13 Dick Furnstahl Nuclear Theory Group OSU Physics Department April 14, 2017 6810 Endgame Various recaps and followups Random stuff (like RNGs :) Session 13 stuff

More information

Over-enhancement Reduction in Local Histogram Equalization using its Degrees of Freedom. Alireza Avanaki

Over-enhancement Reduction in Local Histogram Equalization using its Degrees of Freedom. Alireza Avanaki Over-enhancement Reduction in Local Histogram Equalization using its Degrees of Freedom Alireza Avanaki ABSTRACT A well-known issue of local (adaptive) histogram equalization (LHE) is over-enhancement

More information

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model

More information

Distances in R 3. Last time we figured out the (parametric) equation of a line and the (scalar) equation of a plane:

Distances in R 3. Last time we figured out the (parametric) equation of a line and the (scalar) equation of a plane: Distances in R 3 Last time we figured out the (parametric) equation of a line and the (scalar) equation of a plane: Definition: The equation of a line through point P(x 0, y 0, z 0 ) with directional vector

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

CSC2515 Assignment #2

CSC2515 Assignment #2 CSC2515 Assignment #2 Due: Nov.4, 2pm at the START of class Worth: 18% Late assignments not accepted. 1 Pseudo-Bayesian Linear Regression (3%) In this question you will dabble in Bayesian statistics and

More information

Analysis methods of heavy-tailed data

Analysis methods of heavy-tailed data Institute of Control Sciences Russian Academy of Sciences, Moscow, Russia February, 13-18, 2006, Bamberg, Germany June, 19-23, 2006, Brest, France May, 14-19, 2007, Trondheim, Norway PhD course Chapter

More information

UNDETERMINED COEFFICIENTS SUPERPOSITION APPROACH *

UNDETERMINED COEFFICIENTS SUPERPOSITION APPROACH * 4.4 UNDETERMINED COEFFICIENTS SUPERPOSITION APPROACH 19 Discussion Problems 59. Two roots of a cubic auxiliary equation with real coeffi cients are m 1 1 and m i. What is the corresponding homogeneous

More information

CSE 598C Vision-based Tracking Seminar. Times: MW 10:10-11:00AM Willard 370 Instructor: Robert Collins Office Hours: Tues 2-4PM, Wed 9-9:50AM

CSE 598C Vision-based Tracking Seminar. Times: MW 10:10-11:00AM Willard 370 Instructor: Robert Collins Office Hours: Tues 2-4PM, Wed 9-9:50AM CSE 598C Vision-based Tracking Seminar Times: MW 10:10-11:00AM Willard 370 Instructor: Robert Collins Office Hours: Tues 2-4PM, Wed 9-9:50AM What is Tracking? typical idea: tracking a single target in

More information

Functional Gradient Descent

Functional Gradient Descent Statistical Techniques in Robotics (16-831, F12) Lecture #21 (Nov 14, 2012) Functional Gradient Descent Lecturer: Drew Bagnell Scribe: Daniel Carlton Smith 1 1 Goal of Functional Gradient Descent We have

More information

Quarter 2 400, , , , , , ,000 50,000

Quarter 2 400, , , , , , ,000 50,000 Algebra 2 Quarter 2 Quadratic Functions Introduction to Polynomial Functions Hybrid Electric Vehicles Since 1999, there has been a growing trend in the sales of hybrid electric vehicles. These data show

More information

Geometry Summer Assignment 2018

Geometry Summer Assignment 2018 Geometry Summer Assignment 2018 The following packet contains topics and definitions that you will be required to know in order to succeed in Geometry this year. You are advised to be familiar with each

More information

Joint Optimization of Segmentation and Appearance Models

Joint Optimization of Segmentation and Appearance Models Joint Optimization of Segmentation and Appearance Models David Mandle, Sameep Tandon April 29, 2013 David Mandle, Sameep Tandon (Stanford) April 29, 2013 1 / 19 Overview 1 Recap: Image Segmentation 2 Optimization

More information

Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart

Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart 1 Motivation and Problem In Lecture 1 we briefly saw how histograms

More information

A-Level Maths Induction Summer Work

A-Level Maths Induction Summer Work A-Level Maths Induction Summer Work Name:. (Show workings for every question in this booklet) This booklet contains GCSE Algebra skills that you will need in order to successfully complete the A-Level

More information

SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE

SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE SIFT: SCALE INVARIANT FEATURE TRANSFORM BY DAVID LOWE Overview Motivation of Work Overview of Algorithm Scale Space and Difference of Gaussian Keypoint Localization Orientation Assignment Descriptor Building

More information

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C,

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C, Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: PMF case: p(x, y, z) is the joint Probability Mass Function of X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) PDF case: f(x, y, z) is

More information

MAT Mathematics in Today's World

MAT Mathematics in Today's World MAT 1000 Mathematics in Today's World Last Time We discussed the four rules that govern probabilities: 1. Probabilities are numbers between 0 and 1 2. The probability an event does not occur is 1 minus

More information

Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems

Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems 1 Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems V. Estivill-Castro 2 Perception Concepts Vision Chapter 4 (textbook) Sections 4.3 to 4.5 What is the course

More information

Chapter 7. Functions and onto. 7.1 Functions

Chapter 7. Functions and onto. 7.1 Functions Chapter 7 Functions and onto This chapter covers functions, including function composition and what it means for a function to be onto. In the process, we ll see what happens when two dissimilar quantifiers

More information

Nonparametric Estimation of Luminosity Functions

Nonparametric Estimation of Luminosity Functions x x Nonparametric Estimation of Luminosity Functions Chad Schafer Department of Statistics, Carnegie Mellon University cschafer@stat.cmu.edu 1 Luminosity Functions The luminosity function gives the number

More information