INF Introduction to classifiction Anne Solberg

Size: px
Start display at page:

Download "INF Introduction to classifiction Anne Solberg"

Transcription

1 INF Introduction to classifiction Anne Solberg Introduction to classification Based on handout from Pattern Recognition b Theodoridis, available after the lecture INF

2 Plan for this lecture: Eplain the relation between thresholding and classification with classes Background in probabilit theor Baes rule Classification with a Gaussian densit and a single feature Linear boundaries in feature space Briefl: training and testing a classifier INF 4300

3 Scatter plots A D scatter plot is a plot of feature values for two different features. Each obect s feature values are plotted in the position given b the features values, and with a class label telling its obect class. A scatter plott visualize the space spanned b or more features: called the feature space Matlab: gscatterfeature1, feature, labelvector Classification is done based on more than two features, but this is difficult to visualize. Features with good class separation show clusters for each class, but different clusters should ideall be separated. INF 4300 Feature : maor ais length Feature 1: minor ais length INF

4 Discriminating between classes in a scatter plot INF

5 Introduction to classification One of the most challenging topics in image analsis is recognizing a specific obect in an image. To do that, several steps are normall used. Some kind of segmentation can delimit the spatial etent of the foreground obects, then a set of features to describe the obect characteristics are computed, before the recognition step that decides which obect tpe this is. The focus of the net lectures is the recognition step. The starting point is that we have a set of K features computed for an obect. These features can either describe the shape or the gra level/color/teture properties of the obect. Based on these features we need to decide what kind of an obect this is. Statistical classification is the process of estimating the probabilit that an obect belongs to one of S obect classes based on the observed value of K features. Classification can be done both unsupervised or supervised. In unsupervised classification the categories or classes are not known, and the classification process with be based on grouping similar obects together. In supervised classification the categories or classes are known, and the classification will of consist of estimating the probabilit that the obect belongs to each of the S obect classes. The obect is assigned to the class with highest probabilit. For supervised classifcation, training data is needed. Training data consists of a set of obects with know class tpe, and the are used to estimate the parameters of the classifier. Classification can be piel-based or region-based. The performance of a classifier is normall computed as the accurac it gets when classifing a different set of obects with known class labels called the test set. INF

6 Concepts in classification In the following three lectures we will cover these topics related to classification: Training set Validation set Test set Classifier accurac/confusion matrices. Computing the probabilit that an obect belongs to a class. Let each class be represented b a probabilit densit function. In general man probabilit densities can be used but we use the multivariate normal distribution which is commonl used. Baes rule Discriminant functions/decision boundaries Normal distribution, mean vector and covariance matrices knn classification Unsupervised classification/clustering INF

7 Training data set A set of known obects from all classes for the application. Tpicall a high number of samples from each class. Used to determine the parameters of the given classification algorithm. Eample 1: fit a Gaussian densit to the data Eample : find a Support Vector Model that fits the data Eample 3: find the weights in a neural net INF

8 Validation data set A set of known obects from all classes for the application. Tpicall a sufficientl high number of samples from each class. Used to optimize hperparameters, e.g. Compare feature selection algorithms Compare classifiers Compare network architectures INF

9 Test data set A set of known obects from all classes for the application. Tpicall a high number of samples from each class. Used ONCE to estimate the accurac of a trained model. Should never optimize an parameters on the test data set. INF

10 From INF310: Thresholding Gonzalez and Woods 10.3 Basic thresholding assigns all piels in the image to one of classes: foreground or background g, 0 if 1 if f, f, T T This can be seen as a -class classification problem based on a single feature, the gra level. The classes are background and foreground, and the threshold T defines the border between them. INF

11 Classification error for thresholding Threshold t Foreground tet probabilit distribution Background probabilit distribution In this region, foreground piels are misclassified as background In this region, background piels are misclassified as foreground INF

12 Classification error for thresholding We assume that bz is the normalized histogram for background bz and fz is the normalized histogram for foreground. The histograms are estimates of the probabilit distribution of the gra levels in the image. Let F and B be the prior probabilities for background and foregroundb+f=1 The normalized histogram for the image is then given b p z B b z F f z The probabilit for misclassification given a treshold t is: E E B F t t t t f z dz b z dz INF

13 Find T that minimizes the error t E t F f z dz B b z dz t de t dt 0 F f T B b T Minimum error is achieved b setting T equal to the point where the probabilities for foreground and background are equal. The locations in feature space where the probabilities are equal is called decision boundar in classification. The goal of classification is to find the boundaries. INF

14 Partitioning the feature space using thresholding 1 feature and 1 threshold Tresholding one feature gives a linear boundar separating the feature space INF

15 Partitioning the feature space using thresholding features and thresholds Tresholding two features independentl gives a rectangular boes in feature space INF

16 Can we find a line with better separation? We cannot isolate all classes using straight lines. This problem is not linearl separable INF

17 The goal of classification: partitioning the feature space with smooth boundaries INF

18 Distributions, standard deviation and variance A univariate one feature Gaussian distribution normal distribution is specified given the mean value and the variance : p z Variance, standard deviation 1 e INF

19 Two Gaussian distributions for thresholding a single feature Assume that bz and fz are Gaussian distributions, then p B F B F B F z B e F e B and F are the mean values for background and foreground. B and F are the variance for background and foreground. INF

20 The -class classification problem summarized Given two Gaussian distributions bz and fz. The classes have prior probabilities F and B. Ever piel should be assigned to the class that minimizes the classification error. The classification error is minimized at the point where F fz = B bz. What we will do now is to generalize to K classes and D features. INF

21 The goal of classification We estimate the decision boundaries equivalent to the threshold for multivariate data based on training data. Classification performance is alwas estimated on a separate test data set. We tr to measure the generalization performance. The classifier should perform well when classifing new samples Have lowest possible classification error. We often face a tradeoff between classification error on the training set and generalization abilit when determining the compleit of the decision boundar. INF

22 Probabilit theor Let be a discrete random variable that can assume an of a finite number of M different values. The probabilit that belongs to class i is p i = Pr=i, i=1,...m A probabilit distribution must sum to 1 and probabilities must be positive so p i 0 and M i1 p i 1 INF 4300

23 Epected values - definition The epected value or mean of a random variable is: The variance or second order moment is: These will be estimated from training data where we know the true class labels: Maimum likelihood estimates. INF M i ip i P E 1 P u u E Var P E 1 1 k N i i i k k N i i i k k k k N N

24 Pairs of random variables - definitions Let and be two random variables. The oint probabilit of observing a pair of values =i,= is p i. Alternativel we can define a oint probabilit distribution function P, for which, 0, P, P 1 The marginal distributions for and if we want to eliminate one of them is: P P P P,, INF

25 Statistical independence - definitions Variables and are statistical independent if and onl if P, P P In words: two variables are indepentent if the occurrence of one does not affect the other. If two variables are not independent, the are dependent. If two variables are independent, the are also uncorrelated. For more than two variables: all pairs must be independent. Two variables are uncorrelated if 0 If Cov[X,Y] = E[X Y] E[X ]E[Y] =0, we must have E[X Y] = E[X ]E[Y] If two variables are uncorrelated, the can still be dependent. INF

26 Epected values of two variables Epected values of two variables: INF P E P E P E P E P E P f f E,,,,,,,, Where in this course have ou seen similar formulas?

27 Feature Can ou sketch approimatel,,? Which is largest, or? Feature 1 INF

28 Conditional probabilit If two variables are statisticall dependent, knowing the value of one of them lets us get a better estimate of the value of the other one. We need to consider their covariance. The conditional probabilit of given is defined: Pr i, Pr i Pr and for distributions : P P, P Eample: Draw two cards from a deck. Drawing a king in the first draw has probabilit 4/5. Drawing a king in the second draw given that the first draw gave a king is 3/51. INF

29 Baesian decision theor A fundamental statistical approach to pattern classification. Named after Thomas Baes , an english priest and matematician. It combines prior knowledge about the problem with a probabilit distribution function. The most central concept for us is Baes decision rule. INF

30 Baes rule for a classification problem Suppose we have J, =1,...J classes. is the class label for a piel, and is the observed gra level or feature vector. We can use Baes rule to find an epression for the class with the highest probabilit: p P P p prior probabilit posterior probabilit likelihood normalizing factor For thresholding, P is the prior probabilit for background or foreground. If we don't have special knowledge that one of the classes occur more frequent than other classes, we set them equal for all classes. P =1/J, =1.,,,J. Small p means a probabilit distribution Capital P means a probabilit scalar value between 0 and 1 INF

31 Baes rule eplained P p P p p is the probabilit densit function that models the likelihood for observing gra level if the piel belongs to class. Tpicall we assume a tpe of distribution, e.g. Gaussian, and the mean and covariance of that distribution is fitted to some data that we know belong to that class. This fitting is called classifier training. P is the posterior probabilit that the piel actuall belongs to class. We will soon se that the the classifier that achieves the minimum error is a classifier that assigns each piel to the class that has the highest posterior probabilit. p is ust a scaling factor that assures that the probabilities sum to 1. INF

32 Probabilit of error If we have classes, we make an error either if we decide 1 if the true class is, and if we decide if the true class is 1. If P 1 > P we have more belief that belongs to 1, and we decide 1. The probabilit of error is then: P error P 1 if P if we decide we decide 1 INF

33 Back to classification error for thresholding - Background - Foreground P error P error, d P error p d In this region, foreground piels are misclassified as background In this region, background piels are misclassified as foreground INF

34 Minimizing the error P error P error, d P error p d When we derived the optimal threshold INF 310, we showed that the minimum error was achieved for placing the threshold or decision boundar as we will call it now at the point where P 1 = P This is still valid. INF

35 Baes decision rule In the class case, our goal of minimizing the error implies a decision rule: Decide ω 1 if Pω 1 >Pω ; otherwise ω For J classes, the rule analogusl etends to choose the class with maimum a posteriori probabilit The decision boundar is the border between classes i and, simpl where Pω i =Pω Eactl where the threshold was set in minimum error thresholding! INF

36 Baes classification with J classes How do we generalize: and D features To more the one feature at a time To J classes To consider loss functions that some errors are more costl than others INF

37 Baes rule with J classes and d features If we measure d features, will be a d-dimensional feature vector. Let { 1,..., J } be a set of J classes. The posterior probabilit for class is now computed as Still, we assign a piel with feature vector to the class that has the highest posterior probabilit: INF c P p p p P p P 1 i P for all, P if Decide 1 1

38 Feature space and decision regions The minimum error criterior or a different classifier criterion partitions feature space into disoint region R i. The decision surface separates the regions in multidimensional space. For minimum error, the equation for the decision surface is P P k INF

39 Discriminant functions The decision rule Decide1 if P 1 P, for all i can be written as assign to 1 if g i g The classifier computes J discriminant functions g i and selects the class corresponding to the largest value of the discriminant function. Since classification consists of choosing the class that has the largest value, a scaling of the discriminant function g i b fg i will not effect the decision if f is a monotonicall increasing function. This can lead to simplifications as we will soon see. INF

40 Equivalent discriminant functions The following choices of discriminant functions give equivalent decisions: The effect of the decision rules is to divide the feature space into c decision regions R 1,...R c. If g i >g for all i, then is in region R i. The regions are separated b decision boundaries, surfaces in features space where the discriminant functions for two classes are equal INF ln ln i i i i i i i i i i P p g P p g p P p P g

41 The Gaussian densit - univariate case a single feature To use a classifier we need to select a probabilit densit function p i. The most commonl used probabilit densit is the normal Gaussian distribution: p 1 ep E with epected value or mean E and variance 1 p d p d INF

42 Eample: image and training masks The masks contain labels for the training data. If label=1, then the piel belongs to class 1 red, and so on. If a piel is not part of the training data, it will have label 0. A piel belonging to class k will have value k in the mask image. The mask is often visualized in pseudo-colors on top of the input image, where each class is assign a color. We should have a similar mask for the test data. INF

43 Training a univariate Gaussian classifier To be able to compute the value of the discriminant function, we need to have an estimate of and for each class. Assume that we know the true class labels for some piels and that this is given in a mask image. The mask has N k piels for each class. Training the classifier then consists of computing and for all piels with class label in the mask file. The are computed from training data as: For all piels i with label k in the training mask, compute k k 1 N k 1 N k N k ii N k ii i i k INF

44 Training for i=1:n for =i:m if maski,>==k increment nof. Samples in class K store the feature vector fi, in a vector of training samples from class K end end end For class k=1:k compute meank and sigmak k k 1 N k 1 N k N k ii N k ii i i k INF

45 How do to classification with a univariate Gaussian 1 feature Decide on values for the prior probabilities, P. If we have no prior information, assume that all classes are equall probable and P =1/J. Estimate and based on training data based on the formulae on the previous slide. Training For each piel in a new image: For class =1,.J, compute the discriminant function 1 1 P p P ep P Assign piel to the class C with the highest value of P b setting label_image,= C The result after classification is an image with class labels corresponding to the most probable class for each piel. We compute the classification error rate from an independent test mask. INF

46 Estimating classification error A simple measure of classification accurac can be to count the percentage of correctl classified piels overall averaged for all classes, or per. class. If a piel has true class label k, it is correctl classified if =k. Normall we use different piels to train and test a classifier, so we have a disoint training mask and test mask. Estimate the classification error b classifing all piels in the test set and count the percentage of wrongl classified piels. INF

47 Validating classifier performance Classification performance is evaluated on a different set of samples with known class - the test set. The training set and the test set must be independent! Normall, the set of ground truth piels with known class is partionioned into a set of training piels and a set of test piels of approimatel the same size. If the classifier has hperparameters, we use a separate validation set to find the best value of them. More on this in a later lecture. INF

48 Confusion matrices A matri with the true class label versus the estimated class labels for each class Estimated class labels True class labels Class 1 Class Class 3 Total #sampl es Class Class Class Total INF

49 Confusion matri - cont. Alternatives: Report nof. correctl classified piels for each class. Report the percentage of correctl classified piels for each class. Report the percentage of correctl classified piels in total. Wh is this not a good measure if the number of test piels from each class varies between classes? Class 1 Class Class 3 Total #sam ples Class Class Class Total INF

50 Using more than one feature The power of the computer lies in deciding based on more than 1 feature at a time A simple trick to do this is to assume that features i and and independent, then pi, c=pi cp c The oint decision based on D independent features is then: INF ,...,,..., D D D p P p p p P

51 Upcoming lectures Multivariate Gaussian Classifier evaluation Support vector machine classification Feature selection/feature transforms Knn-classification Unsupervised classification INF

52 If time.. We start with multivariate Gaussian classification INF

INF Introduction to classifiction Anne Solberg Based on Chapter 2 ( ) in Duda and Hart: Pattern Classification

INF Introduction to classifiction Anne Solberg Based on Chapter 2 ( ) in Duda and Hart: Pattern Classification INF 4300 151014 Introduction to classifiction Anne Solberg anne@ifiuiono Based on Chapter 1-6 in Duda and Hart: Pattern Classification 151014 INF 4300 1 Introduction to classification One of the most challenging

More information

INF Anne Solberg One of the most challenging topics in image analysis is recognizing a specific object in an image.

INF Anne Solberg One of the most challenging topics in image analysis is recognizing a specific object in an image. INF 4300 700 Introduction to classifiction Anne Solberg anne@ifiuiono Based on Chapter -6 6inDuda and Hart: attern Classification 303 INF 4300 Introduction to classification One of the most challenging

More information

INF Introduction to classifiction Anne Solberg Based on Chapter 2 ( ) in Duda and Hart: Pattern Classification

INF Introduction to classifiction Anne Solberg Based on Chapter 2 ( ) in Duda and Hart: Pattern Classification INF 4300 90 Itroductio to classifictio Ae Solberg ae@ifiuioo Based o Chapter -6 i Duda ad Hart: atter Classificatio 90 INF 4300 Madator proect Mai task: classificatio You must implemet a classificatio

More information

Bayesian Decision Theory

Bayesian Decision Theory Bayesian Decision Theory Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Bayesian Decision Theory Bayesian classification for normal distributions Error Probabilities

More information

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout :. The Multivariate Gaussian & Decision Boundaries..15.1.5 1 8 6 6 8 1 Mark Gales mjfg@eng.cam.ac.uk Lent

More information

Intro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation

Intro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation Lecture 15. Pattern Classification (I): Statistical Formulation Outline Statistical Pattern Recognition Maximum Posterior Probability (MAP) Classifier Maximum Likelihood (ML) Classifier K-Nearest Neighbor

More information

Machine Learning Lecture 2

Machine Learning Lecture 2 Announcements Machine Learning Lecture 2 Eceptional number of lecture participants this year Current count: 449 participants This is very nice, but it stretches our resources to their limits Probability

More information

Random Vectors. 1 Joint distribution of a random vector. 1 Joint distribution of a random vector

Random Vectors. 1 Joint distribution of a random vector. 1 Joint distribution of a random vector Random Vectors Joint distribution of a random vector Joint distributionof of a random vector Marginal and conditional distributions Previousl, we studied probabilit distributions of a random variable.

More information

Lecture 3: Pattern Classification. Pattern classification

Lecture 3: Pattern Classification. Pattern classification EE E68: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mitures and

More information

SGN (4 cr) Chapter 5

SGN (4 cr) Chapter 5 SGN-41006 (4 cr) Chapter 5 Linear Discriminant Analysis Jussi Tohka & Jari Niemi Department of Signal Processing Tampere University of Technology January 21, 2014 J. Tohka & J. Niemi (TUT-SGN) SGN-41006

More information

Cross-validation for detecting and preventing overfitting

Cross-validation for detecting and preventing overfitting Cross-validation for detecting and preventing overfitting A Regression Problem = f() + noise Can we learn f from this data? Note to other teachers and users of these slides. Andrew would be delighted if

More information

HTF: Ch4 B: Ch4. Linear Classifiers. R Greiner Cmput 466/551

HTF: Ch4 B: Ch4. Linear Classifiers. R Greiner Cmput 466/551 HTF: Ch4 B: Ch4 Linear Classifiers R Greiner Cmput 466/55 Outline Framework Eact Minimize Mistakes Perceptron Training Matri inversion LMS Logistic Regression Ma Likelihood Estimation MLE of P Gradient

More information

Machine Learning Linear Classification. Prof. Matteo Matteucci

Machine Learning Linear Classification. Prof. Matteo Matteucci Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)

More information

Biostatistics in Research Practice - Regression I

Biostatistics in Research Practice - Regression I Biostatistics in Research Practice - Regression I Simon Crouch 30th Januar 2007 In scientific studies, we often wish to model the relationships between observed variables over a sample of different subjects.

More information

6. Vector Random Variables

6. Vector Random Variables 6. Vector Random Variables In the previous chapter we presented methods for dealing with two random variables. In this chapter we etend these methods to the case of n random variables in the following

More information

CSE 546 Midterm Exam, Fall 2014

CSE 546 Midterm Exam, Fall 2014 CSE 546 Midterm Eam, Fall 2014 1. Personal info: Name: UW NetID: Student ID: 2. There should be 14 numbered pages in this eam (including this cover sheet). 3. You can use an material ou brought: an book,

More information

Machine Learning Lecture 2

Machine Learning Lecture 2 Machine Perceptual Learning and Sensory Summer Augmented 6 Computing Announcements Machine Learning Lecture 2 Course webpage http://www.vision.rwth-aachen.de/teaching/ Slides will be made available on

More information

14.1 Systems of Linear Equations in Two Variables

14.1 Systems of Linear Equations in Two Variables 86 Chapter 1 Sstems of Equations and Matrices 1.1 Sstems of Linear Equations in Two Variables Use the method of substitution to solve sstems of equations in two variables. Use the method of elimination

More information

Systems of Linear Equations: Solving by Graphing

Systems of Linear Equations: Solving by Graphing 8.1 Sstems of Linear Equations: Solving b Graphing 8.1 OBJECTIVE 1. Find the solution(s) for a set of linear equations b graphing NOTE There is no other ordered pair that satisfies both equations. From

More information

CONTINUOUS SPATIAL DATA ANALYSIS

CONTINUOUS SPATIAL DATA ANALYSIS CONTINUOUS SPATIAL DATA ANALSIS 1. Overview of Spatial Stochastic Processes The ke difference between continuous spatial data and point patterns is that there is now assumed to be a meaningful value, s

More information

p(x ω i 0.4 ω 2 ω

p(x ω i 0.4 ω 2 ω p( ω i ). ω.3.. 9 3 FIGURE.. Hypothetical class-conditional probability density functions show the probability density of measuring a particular feature value given the pattern is in category ω i.if represents

More information

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential CHAPTER : Partial Derivatives.1 Definition of a Partial Derivative. Increments and Differential.3 Chain Rules.4 Local Etrema.5 Absolute Etrema 1 Chapter : Partial Derivatives.1 Definition of a Partial

More information

Language and Statistics II

Language and Statistics II Language and Statistics II Lecture 19: EM for Models of Structure Noah Smith Epectation-Maimization E step: i,, q i # p r $ t = p r i % ' $ t i, p r $ t i,' soft assignment or voting M step: r t +1 # argma

More information

Unit 12 Study Notes 1 Systems of Equations

Unit 12 Study Notes 1 Systems of Equations You should learn to: Unit Stud Notes Sstems of Equations. Solve sstems of equations b substitution.. Solve sstems of equations b graphing (calculator). 3. Solve sstems of equations b elimination. 4. Solve

More information

Pattern Recognition and Machine Learning. Learning and Evaluation of Pattern Recognition Processes

Pattern Recognition and Machine Learning. Learning and Evaluation of Pattern Recognition Processes Pattern Recognition and Machine Learning James L. Crowley ENSIMAG 3 - MMIS Fall Semester 2016 Lesson 1 5 October 2016 Learning and Evaluation of Pattern Recognition Processes Outline Notation...2 1. The

More information

Machine Learning for Signal Processing Bayes Classification and Regression

Machine Learning for Signal Processing Bayes Classification and Regression Machine Learning for Signal Processing Bayes Classification and Regression Instructor: Bhiksha Raj 11755/18797 1 Recap: KNN A very effective and simple way of performing classification Simple model: For

More information

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1 EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle

More information

Machine Learning Lecture 3

Machine Learning Lecture 3 Announcements Machine Learning Lecture 3 Eam dates We re in the process of fiing the first eam date Probability Density Estimation II 9.0.207 Eercises The first eercise sheet is available on L2P now First

More information

MACHINE LEARNING 2012 MACHINE LEARNING

MACHINE LEARNING 2012 MACHINE LEARNING 1 MACHINE LEARNING kernel CCA, kernel Kmeans Spectral Clustering 2 Change in timetable: We have practical session net week! 3 Structure of toda s and net week s class 1) Briefl go through some etension

More information

Lecture 4 Discriminant Analysis, k-nearest Neighbors

Lecture 4 Discriminant Analysis, k-nearest Neighbors Lecture 4 Discriminant Analysis, k-nearest Neighbors Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University. Email: fredrik.lindsten@it.uu.se fredrik.lindsten@it.uu.se

More information

1 History of statistical/machine learning. 2 Supervised learning. 3 Two approaches to supervised learning. 4 The general learning procedure

1 History of statistical/machine learning. 2 Supervised learning. 3 Two approaches to supervised learning. 4 The general learning procedure Overview Breiman L (2001). Statistical modeling: The two cultures Statistical learning reading group Aleander L 1 Histor of statistical/machine learning 2 Supervised learning 3 Two approaches to supervised

More information

Copyright, 2008, R.E. Kass, E.N. Brown, and U. Eden REPRODUCTION OR CIRCULATION REQUIRES PERMISSION OF THE AUTHORS

Copyright, 2008, R.E. Kass, E.N. Brown, and U. Eden REPRODUCTION OR CIRCULATION REQUIRES PERMISSION OF THE AUTHORS Copright, 8, RE Kass, EN Brown, and U Eden REPRODUCTION OR CIRCULATION REQUIRES PERMISSION OF THE AUTHORS Chapter 6 Random Vectors and Multivariate Distributions 6 Random Vectors In Section?? we etended

More information

0.24 adults 2. (c) Prove that, regardless of the possible values of and, the covariance between X and Y is equal to zero. Show all work.

0.24 adults 2. (c) Prove that, regardless of the possible values of and, the covariance between X and Y is equal to zero. Show all work. 1 A socioeconomic stud analzes two discrete random variables in a certain population of households = number of adult residents and = number of child residents It is found that their joint probabilit mass

More information

Bayesian Decision Theory

Bayesian Decision Theory Bayesian Decision Theory Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University) 1 / 46 Bayesian

More information

Limits 4: Continuity

Limits 4: Continuity Limits 4: Continuit 55 Limits 4: Continuit Model : Continuit I. II. III. IV. z V. VI. z a VII. VIII. IX. Construct Your Understanding Questions (to do in class). Which is the correct value of f (a) in

More information

Derivatives 2: The Derivative at a Point

Derivatives 2: The Derivative at a Point Derivatives 2: The Derivative at a Point 69 Derivatives 2: The Derivative at a Point Model 1: Review of Velocit In the previous activit we eplored position functions (distance versus time) and learned

More information

Uncertainty and Parameter Space Analysis in Visualization -

Uncertainty and Parameter Space Analysis in Visualization - Uncertaint and Parameter Space Analsis in Visualiation - Session 4: Structural Uncertaint Analing the effect of uncertaint on the appearance of structures in scalar fields Rüdiger Westermann and Tobias

More information

BAYESIAN DECISION THEORY

BAYESIAN DECISION THEORY Last updated: September 17, 2012 BAYESIAN DECISION THEORY Problems 2 The following problems from the textbook are relevant: 2.1 2.9, 2.11, 2.17 For this week, please at least solve Problem 2.3. We will

More information

CLASSIFICATION (DISCRIMINATION)

CLASSIFICATION (DISCRIMINATION) CLASSIFICATION (DISCRIMINATION) Caren Marzban http://www.nhn.ou.edu/ marzban Generalities This lecture relies on the previous lecture to some etent. So, read that one first. Also, as in the previous lecture,

More information

Chemometrics: Classification of spectra

Chemometrics: Classification of spectra Chemometrics: Classification of spectra Vladimir Bochko Jarmo Alander University of Vaasa November 1, 2010 Vladimir Bochko Chemometrics: Classification 1/36 Contents Terminology Introduction Big picture

More information

Linear Discriminant Functions

Linear Discriminant Functions Linear Discriminant Functions Linear discriminant functions and decision surfaces Definition It is a function that is a linear combination of the components of g() = t + 0 () here is the eight vector and

More information

Pattern recognition. "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher

Pattern recognition. To understand is to perceive patterns Sir Isaiah Berlin, Russian philosopher Pattern recognition "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher The more relevant patterns at your disposal, the better your decisions will be. This is hopeful news to

More information

Kernel PCA, clustering and canonical correlation analysis

Kernel PCA, clustering and canonical correlation analysis ernel PCA, clustering and canonical correlation analsis Le Song Machine Learning II: Advanced opics CSE 8803ML, Spring 2012 Support Vector Machines (SVM) 1 min w 2 w w + C j ξ j s. t. w j + b j 1 ξ j,

More information

MRC: The Maximum Rejection Classifier for Pattern Detection. With Michael Elad, Renato Keshet

MRC: The Maximum Rejection Classifier for Pattern Detection. With Michael Elad, Renato Keshet MRC: The Maimum Rejection Classifier for Pattern Detection With Michael Elad, Renato Keshet 1 The Problem Pattern Detection: Given a pattern that is subjected to a particular type of variation, detect

More information

Eigenvectors and Eigenvalues 1

Eigenvectors and Eigenvalues 1 Ma 2015 page 1 Eigenvectors and Eigenvalues 1 In this handout, we will eplore eigenvectors and eigenvalues. We will begin with an eploration, then provide some direct eplanation and worked eamples, and

More information

Speech and Language Processing

Speech and Language Processing Speech and Language Processing Lecture 5 Neural network based acoustic and language models Information and Communications Engineering Course Takahiro Shinoaki 08//6 Lecture Plan (Shinoaki s part) I gives

More information

Exact Differential Equations. The general solution of the equation is f x, y C. If f has continuous second partials, then M y 2 f

Exact Differential Equations. The general solution of the equation is f x, y C. If f has continuous second partials, then M y 2 f APPENDIX C Additional Topics in Differential Equations APPENDIX C. Eact First-Order Equations Eact Differential Equations Integrating Factors Eact Differential Equations In Chapter 6, ou studied applications

More information

Dependence and scatter-plots. MVE-495: Lecture 4 Correlation and Regression

Dependence and scatter-plots. MVE-495: Lecture 4 Correlation and Regression Dependence and scatter-plots MVE-495: Lecture 4 Correlation and Regression It is common for two or more quantitative variables to be measured on the same individuals. Then it is useful to consider what

More information

EUSIPCO

EUSIPCO EUSIPCO 3 56974469 APPLICATION OF THE SAMPLE-CONDITIONED MSE TO NON-LINEAR CLASSIFICATION AND CENSORED SAMPLING Lori A. Dalton Department of Electrical and Computer Engineering and Department of Biomedical

More information

Multilayer Neural Networks

Multilayer Neural Networks Pattern Recognition Multilaer Neural Networs Lecture 4 Prof. Daniel Yeung School of Computer Science and Engineering South China Universit of Technolog Outline Introduction (6.) Artificial Neural Networ

More information

Chapter 3: Maximum-Likelihood & Bayesian Parameter Estimation (part 1)

Chapter 3: Maximum-Likelihood & Bayesian Parameter Estimation (part 1) HW 1 due today Parameter Estimation Biometrics CSE 190 Lecture 7 Today s lecture was on the blackboard. These slides are an alternative presentation of the material. CSE190, Winter10 CSE190, Winter10 Chapter

More information

Recall Discrete Distribution. 5.2 Continuous Random Variable. A probability histogram. Density Function 3/27/2012

Recall Discrete Distribution. 5.2 Continuous Random Variable. A probability histogram. Density Function 3/27/2012 3/7/ Recall Discrete Distribution 5. Continuous Random Variable For a discrete distribution, for eample Binomial distribution with n=5, and p=.4, the probabilit distribution is f().7776.59.3456.34.768.4.3

More information

Probability Theory Refresher

Probability Theory Refresher Machine Learning WS24 Module IN264 Sheet 2 Page Machine Learning Worksheet 2 Probabilit Theor Refresher Basic Probabilit Problem : A secret government agenc has developed a scanner which determines whether

More information

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II)

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II) Contents Lecture Lecture Linear Discriminant Analysis Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University Email: fredriklindsten@ituuse Summary of lecture

More information

Generative classifiers: The Gaussian classifier. Ata Kaban School of Computer Science University of Birmingham

Generative classifiers: The Gaussian classifier. Ata Kaban School of Computer Science University of Birmingham Generative classifiers: The Gaussian classifier Ata Kaban School of Computer Science University of Birmingham Outline We have already seen how Bayes rule can be turned into a classifier In all our examples

More information

Chapter 18 Quadratic Function 2

Chapter 18 Quadratic Function 2 Chapter 18 Quadratic Function Completed Square Form 1 Consider this special set of numbers - the square numbers or the set of perfect squares. 4 = = 9 = 3 = 16 = 4 = 5 = 5 = Numbers like 5, 11, 15 are

More information

MACHINE LEARNING ADVANCED MACHINE LEARNING

MACHINE LEARNING ADVANCED MACHINE LEARNING MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 2 2 MACHINE LEARNING Overview Definition pdf Definition joint, condition, marginal,

More information

f x, y x 2 y 2 2x 6y 14. Then

f x, y x 2 y 2 2x 6y 14. Then SECTION 11.7 MAXIMUM AND MINIMUM VALUES 645 absolute minimum FIGURE 1 local maimum local minimum absolute maimum Look at the hills and valles in the graph of f shown in Figure 1. There are two points a,

More information

L11: Pattern recognition principles

L11: Pattern recognition principles L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction

More information

STA 414/2104, Spring 2014, Practice Problem Set #1

STA 414/2104, Spring 2014, Practice Problem Set #1 STA 44/4, Spring 4, Practice Problem Set # Note: these problems are not for credit, and not to be handed in Question : Consider a classification problem in which there are two real-valued inputs, and,

More information

x y plane is the plane in which the stresses act, yy xy xy Figure 3.5.1: non-zero stress components acting in the x y plane

x y plane is the plane in which the stresses act, yy xy xy Figure 3.5.1: non-zero stress components acting in the x y plane 3.5 Plane Stress This section is concerned with a special two-dimensional state of stress called plane stress. It is important for two reasons: () it arises in real components (particularl in thin components

More information

Perturbation Theory for Variational Inference

Perturbation Theory for Variational Inference Perturbation heor for Variational Inference Manfred Opper U Berlin Marco Fraccaro echnical Universit of Denmark Ulrich Paquet Apple Ale Susemihl U Berlin Ole Winther echnical Universit of Denmark Abstract

More information

Semi-Supervised Laplacian Regularization of Kernel Canonical Correlation Analysis

Semi-Supervised Laplacian Regularization of Kernel Canonical Correlation Analysis Semi-Supervised Laplacian Regularization of Kernel Canonical Correlation Analsis Matthew B. Blaschko, Christoph H. Lampert, & Arthur Gretton Ma Planck Institute for Biological Cbernetics Tübingen, German

More information

Bayesian decision theory Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory

Bayesian decision theory Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory Bayesian decision theory 8001652 Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory Jussi Tohka jussi.tohka@tut.fi Institute of Signal Processing Tampere University of Technology

More information

6.867 Machine learning

6.867 Machine learning 6.867 Machine learning Mid-term eam October 8, 6 ( points) Your name and MIT ID: .5.5 y.5 y.5 a).5.5 b).5.5.5.5 y.5 y.5 c).5.5 d).5.5 Figure : Plots of linear regression results with different types of

More information

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers Engineering Part IIB: Module 4F0 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers Phil Woodland: pcw@eng.cam.ac.uk Michaelmas 202 Engineering Part IIB:

More information

5. Zeros. We deduce that the graph crosses the x-axis at the points x = 0, 1, 2 and 4, and nowhere else. And that s exactly what we see in the graph.

5. Zeros. We deduce that the graph crosses the x-axis at the points x = 0, 1, 2 and 4, and nowhere else. And that s exactly what we see in the graph. . Zeros Eample 1. At the right we have drawn the graph of the polnomial = ( 1) ( 2) ( 4). Argue that the form of the algebraic formula allows ou to see right awa where the graph is above the -ais, where

More information

Bayes Decision Theory

Bayes Decision Theory Bayes Decision Theory Minimum-Error-Rate Classification Classifiers, Discriminant Functions and Decision Surfaces The Normal Density 0 Minimum-Error-Rate Classification Actions are decisions on classes

More information

Sample Problems For Grade 9 Mathematics. Grade. 1. If x 3

Sample Problems For Grade 9 Mathematics. Grade. 1. If x 3 Sample roblems For 9 Mathematics DIRECTIONS: This section provides sample mathematics problems for the 9 test forms. These problems are based on material included in the New York Cit curriculum for 8.

More information

Lecture 1: Bayesian Framework Basics

Lecture 1: Bayesian Framework Basics Lecture 1: Bayesian Framework Basics Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de April 21, 2014 What is this course about? Building Bayesian machine learning models Performing the inference of

More information

Handout for Adequacy of Solutions Chapter SET ONE The solution to Make a small change in the right hand side vector of the equations

Handout for Adequacy of Solutions Chapter SET ONE The solution to Make a small change in the right hand side vector of the equations Handout for dequac of Solutions Chapter 04.07 SET ONE The solution to 7.999 4 3.999 Make a small change in the right hand side vector of the equations 7.998 4.00 3.999 4.000 3.999 Make a small change in

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 1

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 1 CS434a/541a: Pattern Recognition Prof. Olga Veksler Lecture 1 1 Outline of the lecture Syllabus Introduction to Pattern Recognition Review of Probability/Statistics 2 Syllabus Prerequisite Analysis of

More information

Notes on Discriminant Functions and Optimal Classification

Notes on Discriminant Functions and Optimal Classification Notes on Discriminant Functions and Optimal Classification Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Discriminant Functions Consider a classification problem

More information

Generative Learning. INFO-4604, Applied Machine Learning University of Colorado Boulder. November 29, 2018 Prof. Michael Paul

Generative Learning. INFO-4604, Applied Machine Learning University of Colorado Boulder. November 29, 2018 Prof. Michael Paul Generative Learning INFO-4604, Applied Machine Learning University of Colorado Boulder November 29, 2018 Prof. Michael Paul Generative vs Discriminative The classification algorithms we have seen so far

More information

Bayes Decision Theory - I

Bayes Decision Theory - I Bayes Decision Theory - I Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Statistical Learning from Data Goal: Given a relationship between a feature vector and a vector y, and iid data samples ( i,y i ), find

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 7 Unsupervised Learning Statistical Perspective Probability Models Discrete & Continuous: Gaussian, Bernoulli, Multinomial Maimum Likelihood Logistic

More information

Minimum Error Rate Classification

Minimum Error Rate Classification Minimum Error Rate Classification Dr. K.Vijayarekha Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur-613 401 Table of Contents 1.Minimum Error Rate Classification...

More information

D u f f x h f y k. Applying this theorem a second time, we have. f xx h f yx k h f xy h f yy k k. f xx h 2 2 f xy hk f yy k 2

D u f f x h f y k. Applying this theorem a second time, we have. f xx h f yx k h f xy h f yy k k. f xx h 2 2 f xy hk f yy k 2 93 CHAPTER 4 PARTIAL DERIVATIVES We close this section b giving a proof of the first part of the Second Derivatives Test. Part (b) has a similar proof. PROOF OF THEOREM 3, PART (A) We compute the second-order

More information

V-0. Review of Probability

V-0. Review of Probability V-0. Review of Probabilit Random Variable! Definition Numerical characterization of outcome of a random event!eamles Number on a rolled die or dice Temerature at secified time of da 3 Stock Market at close

More information

Computer Vision & Digital Image Processing

Computer Vision & Digital Image Processing Computer Vision & Digital Image Processing Image Segmentation Dr. D. J. Jackson Lecture 6- Image segmentation Segmentation divides an image into its constituent parts or objects Level of subdivision depends

More information

For questions 5-8, solve each inequality and graph the solution set. You must show work for full credit. (2 pts each)

For questions 5-8, solve each inequality and graph the solution set. You must show work for full credit. (2 pts each) Alg Midterm Review Practice Level 1 C 1. Find the opposite and the reciprocal of 0. a. 0, 1 b. 0, 1 0 0 c. 0, 1 0 d. 0, 1 0 For questions -, insert , or = to make the sentence true. (1pt each) A. 5

More information

Section 5.1: Functions

Section 5.1: Functions Objective: Identif functions and use correct notation to evaluate functions at numerical and variable values. A relationship is a matching of elements between two sets with the first set called the domain

More information

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 2: Multivariate Gaussians

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 2: Multivariate Gaussians Engineering Part IIB: Module F Statistical Pattern Processing University of Cambridge Engineering Part IIB Module F: Statistical Pattern Processing Handout : Multivariate Gaussians. Generative Model Decision

More information

L20: MLPs, RBFs and SPR Bayes discriminants and MLPs The role of MLP hidden units Bayes discriminants and RBFs Comparison between MLPs and RBFs

L20: MLPs, RBFs and SPR Bayes discriminants and MLPs The role of MLP hidden units Bayes discriminants and RBFs Comparison between MLPs and RBFs L0: MLPs, RBFs and SPR Bayes discriminants and MLPs The role of MLP hidden units Bayes discriminants and RBFs Comparison between MLPs and RBFs CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna CSE@TAMU

More information

IN Pratical guidelines for classification Evaluation Feature selection Principal component transform Anne Solberg

IN Pratical guidelines for classification Evaluation Feature selection Principal component transform Anne Solberg IN 5520 30.10.18 Pratical guidelines for classification Evaluation Feature selection Principal component transform Anne Solberg (anne@ifi.uio.no) 30.10.18 IN 5520 1 Literature Practical guidelines of classification

More information

3.2 Understanding Relations and Functions-NOTES

3.2 Understanding Relations and Functions-NOTES Name Class Date. Understanding Relations and Functions-NOTES Essential Question: How do ou represent relations and functions? Eplore A1.1.A decide whether relations represented verball, tabularl, graphicall,

More information

Curve Fitting Re-visited, Bishop1.2.5

Curve Fitting Re-visited, Bishop1.2.5 Curve Fitting Re-visited, Bishop1.2.5 Maximum Likelihood Bishop 1.2.5 Model Likelihood differentiation p(t x, w, β) = Maximum Likelihood N N ( t n y(x n, w), β 1). (1.61) n=1 As we did in the case of the

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

The Bayes classifier

The Bayes classifier The Bayes classifier Consider where is a random vector in is a random variable (depending on ) Let be a classifier with probability of error/risk given by The Bayes classifier (denoted ) is the optimal

More information

3.7 InveRSe FUnCTIOnS

3.7 InveRSe FUnCTIOnS CHAPTER functions learning ObjeCTIveS In this section, ou will: Verif inverse functions. Determine the domain and range of an inverse function, and restrict the domain of a function to make it one-to-one.

More information

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE GRAVIS 2016 BASEL. Logistic Regression. Pattern Recognition 2016 Sandro Schönborn University of Basel

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE GRAVIS 2016 BASEL. Logistic Regression. Pattern Recognition 2016 Sandro Schönborn University of Basel Logistic Regression Pattern Recognition 2016 Sandro Schönborn University of Basel Two Worlds: Probabilistic & Algorithmic We have seen two conceptual approaches to classification: data class density estimation

More information

1.6 ELECTRONIC STRUCTURE OF THE HYDROGEN ATOM

1.6 ELECTRONIC STRUCTURE OF THE HYDROGEN ATOM 1.6 ELECTRONIC STRUCTURE OF THE HYDROGEN ATOM 23 How does this wave-particle dualit require us to alter our thinking about the electron? In our everda lives, we re accustomed to a deterministic world.

More information

Algebra II Notes Unit Six: Polynomials Syllabus Objectives: 6.2 The student will simplify polynomial expressions.

Algebra II Notes Unit Six: Polynomials Syllabus Objectives: 6.2 The student will simplify polynomial expressions. Algebra II Notes Unit Si: Polnomials Sllabus Objectives: 6. The student will simplif polnomial epressions. Review: Properties of Eponents (Allow students to come up with these on their own.) Let a and

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012 Classification CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Topics Discriminant functions Logistic regression Perceptron Generative models Generative vs. discriminative

More information

Machine Learning for Signal Processing Bayes Classification

Machine Learning for Signal Processing Bayes Classification Machine Learning for Signal Processing Bayes Classification Class 16. 24 Oct 2017 Instructor: Bhiksha Raj - Abelino Jimenez 11755/18797 1 Recap: KNN A very effective and simple way of performing classification

More information

2. This problem is very easy if you remember the exponential cdf; i.e., 0, y 0 F Y (y) = = ln(0.5) = ln = ln 2 = φ 0.5 = β ln 2.

2. This problem is very easy if you remember the exponential cdf; i.e., 0, y 0 F Y (y) = = ln(0.5) = ln = ln 2 = φ 0.5 = β ln 2. FALL 8 FINAL EXAM SOLUTIONS. Use the basic counting rule. There are n = number of was to choose 5 white balls from 7 = n = number of was to choose ellow ball from 5 = ( ) 7 5 ( ) 5. there are n n = ( )(

More information

SF2935: MODERN METHODS OF STATISTICAL LECTURE 3 SUPERVISED CLASSIFICATION, LINEAR DISCRIMINANT ANALYSIS LEARNING. Tatjana Pavlenko.

SF2935: MODERN METHODS OF STATISTICAL LECTURE 3 SUPERVISED CLASSIFICATION, LINEAR DISCRIMINANT ANALYSIS LEARNING. Tatjana Pavlenko. SF2935: MODERN METHODS OF STATISTICAL LEARNING LECTURE 3 SUPERVISED CLASSIFICATION, LINEAR DISCRIMINANT ANALYSIS Tatjana Pavlenko 5 November 2015 SUPERVISED LEARNING (REP.) Starting point: we have an outcome

More information

Summary of Random Variable Concepts March 17, 2000

Summary of Random Variable Concepts March 17, 2000 Summar of Random Variable Concepts March 17, 2000 This is a list of important concepts we have covered, rather than a review that devives or eplains them. Tpes of random variables discrete A random variable

More information

An Information Theory For Preferences

An Information Theory For Preferences An Information Theor For Preferences Ali E. Abbas Department of Management Science and Engineering, Stanford Universit, Stanford, Ca, 94305 Abstract. Recent literature in the last Maimum Entrop workshop

More information