Machine Learning for Signal Processing Bayes Classification and Regression


 Merryl Lynch
 9 months ago
 Views:
Transcription
1 Machine Learning for Signal Processing Bayes Classification and Regression Instructor: Bhiksha Raj 11755/
2 Recap: KNN A very effective and simple way of performing classification Simple model: For any instance, select the class from the instances close to it in feature space 11755/
3 Multiclass Image Classification 3
4 knearest Neighbors Given a query item: Find k closest matches in a labeled dataset 4
5 knearest Neighbors Given a query item: Find k closest matches Return the most Frequent label 5
6 knearest Neighbors 2 votes for cat, 1 each for Buffalo, Cat wins Deer, Lion 7
7 Nearest neighbor method Weighted majority vote within the k nearest neighbors Y x = argmax c x i Nk x,y i =c w(x, x i )y i new K= 1: blue K= 3: green 8
8 But what happens if.. You have many training instances at exactly that value of x? Majority vote on nearest neighbors: new Test Training y = argmax y count(y x) 9
9 But what happens if.. You have many training instances at exactly that value of x? Majority vote on nearest neighbors: new Test Training y = argmax y N y (x) N(x) 10
10 But what happens if.. You have many training instances at exactly that value of x? Majority vote on nearest neighbors: new Test Training y = argmax y P(y x) 11
11 But what happens if.. You have many training instances at exactly that value of x? Majority vote on nearest neighbors: new Test Training y = argmax y P(y x) Bayes Classification Rule 12
12 Bayes Classification Rule For any observed feature X, select the class value Y that is most frequent Also applies to continuous valued predicted variables I.e. regression Select Y to maximize the a posteriori probability P(Y X) Bayes classification is an instance of maximum a posteriori estimation 11755/
13 Bayes classification What happens if there are no exact neighbors No training instances with exactly the same X value? new 14
14 Given Bayes Classification Rule a set of classes C = {C 1, C 2,, C N } Conditional probability distributions P(C X) Classification performed as መC = argmax C C P(C X) Problem: How do you characterize P C X Require a function that, given an X, computes P(C X) for every class C 11755/
15 Modelling P(C X) Each pixel is a combination of red green and blue weighted by the a posteriori probability of the classes Blue: Class 1 Red: Class 2 Green: Class 3 Assumption: there s a continuous function that, at every X, produces a vector of outputs P(C X) for every class C The decision boundary for any class is the boundary within which its own posterior has the highest value This function accurately represents the actual a posteriori probabilities for the classes Objective: Estimate this function 11755/
16 Modelling the posterior To model the posterior, we need a functional form for P(C X) which can be learned Typically this functional form is expressed in terms of distance from a decision boundary The simplest decision boundary is the linear boundary 11755/
17 Bayesian Linear Classification: Two class case First: Twoclass classification Assumption: the decision boundary between the classes is a simple hyperplane As you go away from the hyperplane, the fraction of data from one class increases, while that from the other decreases Will also hold for any sample of data 11755/
18 1D binary class example y x Onedimensional example for visualization Only two classes (represented by y=0 and y=1) All (red) dots at Y=1 represent instances of class Y=1 All (blue) dots at Y=0 are from class Y=0 The data are not linearly separable In this 1D example, a linear separator is a threshold No threshold will cleanly separate red and blue dots 19
19 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of Y=1 at that point 20
20 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 21
21 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 22
22 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 23
23 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 24
24 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 25
25 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 26
26 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 27
27 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 28
28 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 29
29 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 30
30 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 31
31 The probability of y=1 y x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 32
32 The probability of y=1 Need a function to model the shape of this curve. y A good choice that works under many conditions: Logistic function x Consider this differently: at each point look at a small window around that point Plot the average value within the window This is an approximation of the probability of 1 at that point 33
33 The logistic regression model y=1 P ( y x) = w ( 0 + x w 1+ e 1 ) y=0 x Class 1 becomes increasingly probable going left to right Very typical in many problems The logistic is a function of the distance from the P(y x) = 0.5 boundary 34
34 For twodimensional input Decision: y > 0.5? x 2 1 When X is a 2D variable P( y x) = w ( 0 + x ) 1+ e w The decision boundary for P(Y X)=0.5 is a hyperplane It is a linear model x 1 35
35 The logistic regression model y=1 P e w0 ( y x) = w 0 + w x = w ( 0 + w x 1+ e + w x 1+ e 1 ) y=0 x Note how it varies with w 0 + w 1 x 36
36 Estimating the model y P ( y x) = f ( x) = ( w w x) 1+ e 1 + x Given the training data (many (x, y) pairs represented by the dots), estimate w 0 and w 1 for the curve 37
37 Estimating the model 38 x y ) ( 1 1 ) 1 ( x w w e x y P + + = = ) ( 1 1 ) 1 ( x w w e x y P + + = = ) ( 1 1 ) ( x w w y e x y P + + = Easier to represent using a y = +1/1 notation
38 Estimating the model Given: Training data X 1, y 1, X 2, y 2,, X N, y N Xs are vectors, ys are binary (0/1) class values Total probability of data P X 1, y 1, X 2, y 2,, X N, y N = P X i, y i 1 = P y i X i P X i = 1 + e P X y i(w 0 +w T X i ) i i i i 39
39 Estimating the model Likelihood P Training data = Log likelihood log P Training data = i e y i(w 0 +w T X i ) P X i log P X i log 1 + e y i(w 0 +w T X i ) i i 40
40 Maximum Likelihood Estimate w 0, w 1 = argmax w 0,w 1 log P Training data Equals (note argmin rather than argmax) w 0, w 1 = argmin w 0,w i log 1 + e y i(w 0 +w T X i ) Minimizing the KL divergence between the desired output y and actual output 1 1+e (w 0+w T X i ) Cannot be solved directly, needs gradient descent 41
41 Model learned by logistic regression Pure Red: 0 Pure Green: 1 The figure shows the class probability over a twodimensional feature space Any decision threshold P(C X) = Const is a hyperplane Diagonal line in this case 11755/
42 Multiclass logistic regression The simple logistic regression model can be extended to multiple classes: P C X = exp(w C T X) σ C exp(w T C X) W T C X is, in fact, the discriminant function of the classes We ve encountered discriminant functions earlier Also called a softmax Each class C i has a probability that is exponentially related to the closeness of the vector X to a representative vector w i for the class This too can be learned from training data via maximum likelihood estimation Just like the twoclass case 11755/
43 Multiclass logistic regression The boundary between adjacent classes is a hyperplane (line) The decision boundary for any class is convex polytope with hyperplane segments I.e. still a linear classifier 11755/
44 Multiclass Bayes In many classification problems, linear boundaries are not sufficient We need to be able to model more complex boundaries This too can be supported by the logistic regression classifier 11755/
45 Logistic regression with nonlinear boundaries The logistic regression can be modified to have nonlinear discriminants: P C X = exp(f(x; θ C)) σ C exp(f(x; θ C )) f(x; θ C ) is the discriminant for class C, and has parameter θ C The discriminants determine the shape of the decision boundary Nonlinear discriminants result in nonlinear decision boundaries The parameters θ C for all classes can be learned by maximum likelihood (or MAP) estimation as before 11755/
46 Quadratic discriminant P C X = exp(f(x; θ C)) σ C exp(f(x; θ C )) With quadratic discriminants: f X; θ C = (X α C )β C (X α C ) T Note that decision boundaries are quadratic The probability of a class increases (or decreases) as we go away from a boundary 11755/
47 Logistic regression with nonlinear boundaries P C X = exp(f(x; θ C)) σ C exp(f(x; θ C )) For complex decision boundaries, the function f(x; θ C ) must be correspondingly complex Currently the most successful approach in these cases is to model f(x; θ C ) by a neural network In fact neural networks with softmax decision layers may be seen as an instance of a logistic regression with a nonlinear discriminant Topic for a later class 11755/
48 Logistic regression with nonlinear boundaries The logistic regression can be modified to have nonlinear discriminants: P C X = exp(f(x; θ C)) σ C exp(f(x; θ C )) Note: This can also be viewed as nonlinearly transforming the data X into a space where a simple linear logistic regression models posteriors well Z X = [f X; θ 1 f X; θ 2 f(x; θ K )] T I.e. into a space where the data are most linearly separable We will discuss this in a later lecture on neural networks 11755/
49 Problem with modelling P(C X) We have considered modelling the a posteriori probability of the classes directly This implicitly assumes that The characteristics of the data for any class remain the same between train and test The relative proportions of the classes too remain the same Often the second assumption will not hold The data characteristics remain, but the relative proportions change E.g. the shapes of the differently colored coins don t change, but the relative proportions of the colors changes between train and test We must then modify our approach to Bayes classification to a generative framework 11755/
50 The Bayesian Classifier.. መC = argmax P(C X) C C Choose the class that is most frequent for the given X P(C X) = P(C)P(X C) P(X) argmax C C P(C X) = argmax C C P(C)P(X C) Choose the class that is most likely to have produced X While accounting for the relative frequency of C 11755/
51 P(X) Bayes Classification Rule Given a set of classes C = {C 1, C 2,, C N } መC = argmax P(C)P(X C) C C P(X C i ) measures the probability that a random instance of class C i will take the value X P(X C 1 ) X 11755/
52 P(X) Bayes Classification Rule Given a set of classes C = {C 1, C 2,, C N } መC = argmax P(C)P(X C) C C P(X C i ) measures the probability that a random instance of class C i will take the value X P(X C 2 ) P(X C 1 ) X 11755/
53 P(X) Bayes Classification Rule Given a set of classes C = {C 1, C 2,, C N } መC = argmax C C P(C)P(X C) P(C i ) scales them up to match the expected relative proportions of the classes P(C 2 )P(X C 2 ) P(C 1 )P(X C 1 ) X 11755/
54 P(X) Bayes Classification Rule Given a set of classes C = {C 1, C 2,, C N } መC = argmax C C P(C)P(X C) P(C i ) scales them up to match the expected relative proportions of the classes Decision boundary P(C 2 )P(X C 2 ) P(C 1 )P(X C 1 ) X 11755/
55 P(X) Bayes Classification Rule Given a set of classes C = {C 1, C 2,, C N } መC = argmax C C P(C)P(X C) P(C i ) scales them up to match the expected relative proportions of the classes Fraction of all instances that belong to C 1 and fall on the wrong Decision boundary side of the boundary and are misclassified P(C 1 )P(X C 1 ) P(C 2 )P(X C 2 ) X 11755/
56 P(X) Bayes Classification Rule Given a set of classes C = {C 1, C 2,, C N } መC = argmax C C P(C)P(X C) P(C i ) scales them up to match the expected relative proportions of the classes Fraction of all instances that belong to C 2 and fall on the wrong Decision boundary side of the boundary and are misclassified P(C 1 )P(X C 1 ) P(C 2 )P(X C 2 ) X 11755/
57 P(X) Bayes Classification Rule Given a set of classes C = {C 1, C 2,, C N } መC = argmax C C P(C)P(X C) P(C i ) scales them up to match the expected relative proportions of the classes Decision boundary Total misclassification probability P(C 2 )P(X C 2 ) P(C 1 )P(X C 1 ) X 11755/
58 P(X) Bayes Classification Rule The Bayes classification rule is the statistically optimal classification rule Moving the boundary in either direction will always increase the classification error Excess classification error with shifted boundary P(C 2 )P(X C 2 ) P(C 1 )P(X C 1 ) X 11755/
59 The Bayesian Classifier.. መC = argmax C C P(C X) = argmax C C P(C)P(X C) We can now directly learn the classspecific statistical characteristics P(X C) from the training data The relative frequency of C, P(C), can be independently adjusted to our expectations of these frequencies in the test data These need not match the training data 11755/
60 Modeling P(X C) Challenge: How to learn P(X C) This will not be known beforehand and must be learned from examples of X that belong to class C Will generally have unknown and unknowable shape We only observe samples of X Must make some assumptions about the form of P(X C) 11755/
61 The problem of dependent variables P(X C) = P(X 1, X 2,, X D C) must be defined for every combination of X 1, X 2,, X D Too many parameters to describe explicitly Most combinations unseen in training data P(X C) may have an arbitrary scatter/shape Hard to characterize mathematically Typically do so by assigning a functional form to it 11755/
62 The problem of dependent variables P(X C) = P(X 1, X 2,, X D C) must be defined for every combination of X 1, X 2,, X D Too many parameters to describe explicitly Most combinations unseen in training data P(X C) may have an arbitrary scatter/shape Hard to characterize mathematically Typically do so by assigning a functional form to it 11755/
63 The Naïve Bayes assumption C C X 1 X 5 X2 X 4 X 1 X 2 X 3 X 4 X 5 X 3 Assume all the components are independent of one another The joint probability is the product of the marginal P X C = P X 1, X 2,, X D C = P X i C Sufficient to learn marginal distributions P X i C The problem of having to observe all combinations of X 1, X 2,, X D never arises i 11755/
64 Naïve Bayes estimating P X i C P X i C may be estimated using conventional maximum likelihood estimation Given a number of training instances belonging to class C Select the ith component of all instances Estimate P X i C For discretevalued X i this will be a multinomial distribution For continuous valued X i a form must be assumed E.g Gaussian, Laplacian etc 11755/
65 Naïve Bayes Binary Case 11755/
66 The problem of dependent variables P(X C) = P(X 1, X 2,, X D C) must be defined for every combination of X 1, X 2,, X D Too many parameters Most combinations unseen in training data P(X C) may have an arbitrary scatter/shape Hard to characterize mathematically Typically do so by assigning a functional form to it 11755/
67 Assigning a functional form to P(X C) Assign a functional form to P(X C) Common assumptions: Unimodal forms: Gaussian, Laplacian Multimodal forms: Gaussian mixtures Time series: Hidden Markov models Multidimensional structures: Markov random fields 11755/
68 Assigning a functional form to P(X C) Assign a functional form to P(X C) Common assumptions: Unimodal forms: Gaussian, Laplacian Most common of all Multimodal forms: Gaussian mixtures Time series: Hidden Markov models Multidimensional structures: Markov random fields 11755/
69 Gaussian Distribution Mean Vector Covariance Matrix  Symmetric  Positive Definite 11755/
70 Gaussian Distribution 11755/
71 Parameter Estimation 11755/
72 Gaussian classifier Different Classes, different Gaussians 11755/
73 Gaussian Classifier For each class we need: Mean Vector Covariance Matrix Training Fit a Gaussian to each class Classification: Problem: Find the best Gaussian to explain the distribution for the class arg max i Many parameters to train! P C i p x, μ i, Σ i Dominated by covariance: for Ddimensional data the covariance matrices requires D 2 parameters each For N c classes, a total of N c D 2 parameters 11755/
74 Homoskedasticity assumption Assume all distributions have the same covariance Σ i = Σ i Assumption, may not be true But still works in many cases Fewer parameters to train One common covariance matrix for all classes Only D 2 total parameters As opposed to N c D 2 if each class has its own covariance matrix 11755/
75 Homoskedastic Gaussians Σ 1 = Σ 2 = = Σ K = Σ For the binary classification case (K = 2) Decision boundary: P X C 2 P C 2 = P X C 2 P C /
76 Homoskedastic Gaussians Linear Boundary!!! 11755/
77 Homoskedastic Gaussians 11755/
78 Homoskedastic Gaussians, K > 2 Case K > 2 (more than two classes) Classification performed as: argmax P C i p x, μ i, Σ i i Taking logs and ignoring the common constant argmax i 1 2 x μ i T Σ 1 x μ i + log P C i Expanding out and ignoring common terms argmax i This is just a linear classifier 1 2 xt Σ 1 μ i 1 2 μ i T Σ 1 μ i + log P C i 11755/
79 Homoskedastic Gaussians, K > 2 Decision boundaries for argmax P C i p x, μ i, Σ i i Linear classifier: Decision boundaries are hyperplanes 11755/
80 Homoskedastic Gaussians, K > 2 Case K > 2 (more than two classes) Classification performed as: argmax P C i p x, μ i, Σ i i Taking logs and ignoring the common constant argmax i 1 2 x μ i T Σ 1 x μ i + log P C i Changing the sign and rewriting it as argmin argmin i x μ i T Σ 1 x μ i 2log P C i 11755/
81 Homoskedastic Gaussians Mahalanobis Distance A Gaussian Classifier with common Covariance Matrix is similar to a Nearest Neighbor Classifier Classification corresponds to the nearest mean vector 11755/
82 How to estimate the Covariance Matrix? Maximum likelihood estimate of covariances of individual classes: Σ C = 1 N C N C x i (c) μc x i (c) μc T i=1 Estimate of common covariance for all classes Σ = 1 σk C =1 K N C Σ C N C C= /
83 Hetero skedastic Gaussians.. Homoskedastic Gaussians do not capture nonlinear decision boundaries Also, the assumption that all Gaussians have the same covariance is questionable Permitting each Gaussian to have its own covariance results in nonlinear decision boundaries Hetero skedastic Gaussians 11755/
84 Heteroskedastic Gaussians Different Covariance Matrices 1D case. K = 2 Decision Boundary Quadratic Boundary!!! 11755/
85 Heteroskedastic Gaussians 11755/
86 Digit recognition Handwriting recognition Money detection in check deposit $ /
87 now is to take a probabilistic view of the problem: We treat the variations in the d Gaussian distribution Classifier over the images. So for each for digit j 2 Digit { 0,1,...,9}, we have a probabil ion of the distribution P 1 (x) that captures our uncertainty of the variations among digit 1. over the images x, which represents our uncertainty or imperfect knowledge about the P j are imperfect models of the digits. Nevertheless, images containing since they are digit concentrated j. around ch digit, they are still a useful representation of our recognition knowledge. At this point is still not t the distributions P j from, but for the moment Under let usdistribution assume we have P j (x), them. we are a lot more likely to generate images x that look like digit j; however, there is also a small probability that P j (x) generates images that are e can phrase the digit classification problem in terms of balls and bins as follows: We have digit j, but by random chance may look like other digits. Figure 3 shows an illustrati corresponds to a digit. Within each bin we have all possible balls (images) corresponding Weassume the images within bin j are distributed P 1 (x). according The bulk to the of the probability mass (images distribution that are assigned high probability) are good image he images in bin j are more likely to have a digit the tails j written (images in them. withthen small weprobability) are given an are some ambiguous images that could be int taken from one of the bins (the image has apossibly digit written other in it), digits. but we don t know which is to guess which bin the image was likely to have come from (which digit the image s and bins analogy of digit classification. Balls Figure are the3: images, Illustration and bins of are the the distribution digits. P 1 (x) that captures our uncertainty of the vari The distributions P j are imperfect models of the digits. Nevertheless, since they are good images of each digit, they are still a useful representation of our knowledge. At t clear where we get the distributions P j from, but for the moment let us assume we hav ecture 21 3 With this setup, we can phrase the digit classification problem in terms of balls and bin 10 bins, each one corresponds to a digit. Within each bin we have all possible balls (im to that bin (digit). Weassume the images within bin j aredistributed according to the pr P j, which means / the images in bin j are more likely to have a digit j written 88 in them.
88 Showing the average of each digit Average digit Figure 7: Mean vectors for each digit (the estimated p ji s) /
89 Bayes classifcation Back to balls and bins Suppose we are just given a ball (image) x Which bin (digit) y did it come from? Normalize the Posterior prior posterior Bayes rule /
90 Inadequacy of Gaussian classifiers Gaussian classifiers can only capture simple linear or quadratic decision boundaries Often, the decision boundaries required are more complex In this case we must employ a Gaussian Mixture classifier argmax P C i p x C i i p x C i is modelled by a Gaussian mixture 11755/
91 GMM classifier  For each class, train a GMM (with EM)  Classify according to 11755/
92 Bayesian Classification with Gaussian Mixtures Plotting P C i p x C i for all classes Left: Twoclass classification, Right: Threeclass classification Each class modelled by a mixture of three Gaussians Note the complex decision boundary 11755/
93 Estimating P(C) argmax P C i p x, μ i, Σ i i Have not explained where the class prior P C i from comes This can be dependent on the test data Typical solutions: Estimate from training data Optimize on development or heldout test data Heuristic guess Conservative estimates Set the prior of classes that have high cost if incorrectly detected to be low Set prior of classes that have high cost if incorrectly missed to be low Etc /
94 Topics not covered Maximum a posteriori estimation When we make assumptions about the parameters (means, covariances) themselves MAP regression with Gaussians Predicting continuousvalued RVs assuming Gaussian distributions MAP regression with Gaussian Mixtures Predicting continuousvalued RVs with Gaussian mixture distributions Timeseries and other structured data Partially covered in a later lecture 11755/