STAD68: Machine Learning

Similar documents
STA 4273H: Sta-s-cal Machine Learning

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model

CS 6140: Machine Learning Spring 2016

Learning Deep Genera,ve Models

Computer Vision. Pa0ern Recogni4on Concepts Part I. Luis F. Teixeira MAP- i 2012/13

Bayesian networks Lecture 18. David Sontag New York University

CS 6140: Machine Learning Spring What We Learned Last Week 2/26/16

An Introduc+on to Sta+s+cs and Machine Learning for Quan+ta+ve Biology. Anirvan Sengupta Dept. of Physics and Astronomy Rutgers University

Founda'ons of Large- Scale Mul'media Informa'on Management and Retrieval. Lecture #3 Machine Learning. Edward Chang

CSE 473: Ar+ficial Intelligence

Bias/variance tradeoff, Model assessment and selec+on

Latent Dirichlet Alloca/on

ECE521 lecture 4: 19 January Optimization, MLE, regularization

CSE446: Linear Regression Regulariza5on Bias / Variance Tradeoff Winter 2015

STA 4273H: Statistical Machine Learning

COMP 562: Introduction to Machine Learning

Priors in Dependency network learning

Regression.

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

UVA CS 4501: Machine Learning. Lecture 6: Linear Regression Model with Dr. Yanjun Qi. University of Virginia

Mixture Models. Michael Kuhn

ECE521 week 3: 23/26 January 2017

Recent Advances in Bayesian Inference Techniques

CS 6140: Machine Learning Spring 2017

CS 6375 Machine Learning

CSCI 360 Introduc/on to Ar/ficial Intelligence Week 2: Problem Solving and Op/miza/on

Graphical Models. Lecture 1: Mo4va4on and Founda4ons. Andrew McCallum

STA414/2104 Statistical Methods for Machine Learning II

STA 4273H: Sta-s-cal Machine Learning

CSE 473: Ar+ficial Intelligence. Probability Recap. Markov Models - II. Condi+onal probability. Product rule. Chain rule.

Regulatory Inferece from Gene Expression. CMSC858P Spring 2012 Hector Corrada Bravo

Lecture 7: Con3nuous Latent Variable Models

Sta$s$cal sequence recogni$on

STA 4273H: Statistical Machine Learning

Machine Learning and Data Mining. Bayes Classifiers. Prof. Alexander Ihler

Natural Language Processing with Deep Learning. CS224N/Ling284

Part 2. Representation Learning Algorithms

Classifica(on and predic(on omics style. Dr Nicola Armstrong Mathema(cs and Sta(s(cs Murdoch University

Outline. What is Machine Learning? Why Machine Learning? 9/29/08. Machine Learning Approaches to Biological Research: Bioimage Informa>cs and Beyond

STA 4273H: Statistical Machine Learning

INTRODUCTION TO PATTERN RECOGNITION

Quan&fying Uncertainty. Sai Ravela Massachuse7s Ins&tute of Technology

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Machine Learning & Data Mining CS/CNS/EE 155. Lecture 8: Hidden Markov Models

STA 4273H: Statistical Machine Learning

Introduction to Machine Learning Midterm Exam

Last Lecture Recap UVA CS / Introduc8on to Machine Learning and Data Mining. Lecture 3: Linear Regression

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Generative Clustering, Topic Modeling, & Bayesian Inference

Generative Model (Naïve Bayes, LDA)

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

STA 4273H: Sta-s-cal Machine Learning

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

Pattern Recognition and Machine Learning

Machine Learning & Data Mining CS/CNS/EE 155. Lecture 11: Hidden Markov Models

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

Introduction to Machine Learning

CS 540: Machine Learning Lecture 1: Introduction

Brief Introduction of Machine Learning Techniques for Content Analysis

Slides modified from: PATTERN RECOGNITION AND MACHINE LEARNING CHRISTOPHER M. BISHOP

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Mathematical Formulation of Our Example

Gene Regulatory Networks II Computa.onal Genomics Seyoung Kim

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari

Natural Language Processing with Deep Learning. CS224N/Ling284

Latent Variable Models

Detec%ng and Analyzing Urban Regions with High Impact of Weather Change on Transport

UVA CS / Introduc8on to Machine Learning and Data Mining

Probabilistic Graphical Models for Image Analysis - Lecture 1

Computer Vision. Pa0ern Recogni4on Concepts. Luis F. Teixeira MAP- i 2014/15

Neural Networks. William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]

Least Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression

Unsupervised Learning

Lecture 3. Linear Regression II Bastian Leibe RWTH Aachen

STA 414/2104: Machine Learning

Introduction to Machine Learning

STA 414/2104: Machine Learning

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

Least Mean Squares Regression. Machine Learning Fall 2017

Discrimina)ve Latent Variable Models. SPFLODD November 15, 2011

Introduction. Le Song. Machine Learning I CSE 6740, Fall 2013

Lecture : Probabilistic Machine Learning

CPSC 540: Machine Learning

STA 4273H: Sta-s-cal Machine Learning

Machine Learning and Data Mining. Linear regression. Prof. Alexander Ihler

CSCI 360 Introduc/on to Ar/ficial Intelligence Week 2: Problem Solving and Op/miza/on. Professor Wei-Min Shen Week 8.1 and 8.2

BAYESIAN DECISION THEORY

Unsupervised Learning

Latent Dirichlet Allocation Introduction/Overview

Introduction to Particle Filters for Data Assimilation

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a

Introduction. Chapter 1

Statistical learning. Chapter 20, Sections 1 4 1

Modelling Time- varying effec3ve Brain Connec3vity using Mul3regression Dynamic Models. Thomas Nichols University of Warwick

STA 414/2104: Lecture 8

Stat 406: Algorithms for classification and prediction. Lecture 1: Introduction. Kevin Murphy. Mon 7 January,

Transcription:

STAD68: Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! h0p://www.cs.toronto.edu/~rsalakhu/ Lecture 1

Evalua;on 3 Assignments worth 40%. Midterm worth 20%. Final worth 40% Tenta4ve Dates Check the website for updates! Assignment 1: Handed out: Jan 13, Due: Jan 27. Assignment 2: Handed out: Jan 27, Due: Feb 10. Assignment 3: Handed out: March 3, Due: March 17.

Text Books Christopher M. Bishop (2006) PaAern Recogni4on and Machine Learning, Springer. Addi;onal Books Trevor Has;e, Robert Tibshirani, Jerome Friedman (2009) The Elements of Sta;s;cal Learning David MacKay (2003) Informa;on Theory, Inference, and Learning Algorithms Most of the figures and material will come from these books.

Sta;s;cal Machine Learning Sta;s;cal machine learning is a very dynamic field that lies at the intersec;on of Sta;s;cs and computa;onal sciences. The goal of sta;s;cal machine learning is to develop algorithms that can learn from data by construc;ng stochas;c models that can be used for making predic;ons and decisions.

Machine Learning s Successes Biosta;s;cs / Computa;onal Biology. Neuroscience. Medical Imaging: - computer- aided diagnosis, image- guided therapy. - image registra;on, image fusion. Informa;on Retrieval / Natural Language Processing: - Text, audio, and image retrieval. - Parsing, machine transla;on, text analysis. Speech processing: - Speech recogni;on, voice iden;fica;on. Robo;cs: - Autonomous car driving, planning, control.

Mining for Structure Massive increase in both computa;onal power and the amount of data available from web, video cameras, laboratory measurements. Images & Video Text & Language Speech & Audio Gene Expression Develop sta;s;cal models that can discover Rela;onal underlying structure, cause, Data/ Climate Change Product Social Network or sta;s;cal correla;ons from data. Geological Data Recommenda;on

Example: Boltzmann Machine Model parameters Latent (hidden) variables Input data (e.g. pixel intensi;es of an image, words from webpages, speech signal). Target variables (response) (e.g. class labels, categories, phonemes). Markov Random Fields, Undirected Graphical Models.

Boltzmann Machine Observed Data 25,000 characters from 50 alphabets around the world. Simulate a Markov chain whose sta;onary distribu;on is

Boltzmann Machine Pa0ern Comple;on: P(image par;al image)

Boltzmann Machine Pa0ern Recogni;on Op;cal Character Recogni;on Learning Algorithm Error Logis;c regression 22.14% K- NN 18.92% Neural Network 14.62% SVM (Larochelle et.al. 2009) 9.70% Deep Autoencoder (Bengio et. al. 2007) Deep Belief Net (Larochelle et. al. 2009) 10.05% 9.68% This model 8.40%

Finding Structure in Data Vector of word counts on a webpage Latent variables: hidden topics Interbank Markets European Community Monetary/Economic Energy Markets Disasters and Accidents 804,414 newswire stories Leading Economic Indicators Accounts/ Earnings Government Borrowings Legal/Judicial

Matrix Factoriza;on Collabora;ve Filtering/ Matrix Factoriza;on/ Ra;ng value of user i for item j Hierarchical Bayesian Model Latent user feature (preference) vector Latent item feature vector Predic4on: predict a ra;ng r * ij for user i and query movie j. Latent variables that we infer from observed ra;ngs. Posterior over Latent Variables Infer latent variables and make predic;ons using Markov chain Monte Carlo.

Finding Structure in Data Collabora;ve Filtering/ Matrix Factoriza;on/ Product Recommenda;on Learned ``genre Nemlix dataset: 480,189 users 17,770 movies Over 100 million ra;ngs. Fahrenheit 9/11 Bowling for Columbine The People vs. Larry Flynt Canadian Bacon La Dolce Vita Friday the 13th The Texas Chainsaw Massacre Children of the Corn Child's Play The Return of Michael Myers Independence Day The Day Aoer Tomorrow Con Air Men in Black II Men in Black Part of the wining solu;on in the Nemlix contest (1 million dollar prize).

Tenta;ve List of Topics Linear methods for regression, Bayesian linear regression Linear models for classifica;on Probabilis;c Genera;ve and Discrimina;ve models Regulariza;on methods Model Comparison and BIC Neural Networks Radial basis func;on networks Kernel Methods, Gaussian processes, Support Vector Machines Mixture models and EM algorithm Graphical Models and Bayesian Networks

Types of Learning Consider observing a series of input vectors: Supervised Learning: We are also given target outputs (labels, responses): y 1,y 2,, and the goal is to predict correct output given a new input. Unsupervised Learning: The goal is to build a sta;s;cal model of x, which can be used for making predic;ons, decisions. Reinforcement Learning: the model (agent) produces a set of ac;ons: a 1, a 2, that affect the state of the world, and received rewards r 1, r 2 The goal is to learn ac;ons that maximize the reward (we will not cover this topic in this course). Semi- supervised Learning: We are given only a limited amount of labels, but lots of unlabeled data.

Supervised Learning Classifica4on: target outputs y i are discrete class labels. The goal is to correctly classify new inputs. Regression: target outputs y i are con;nuous. The goal is to predict the output given new inputs.

Handwri0en Digit Classifica;on

Unsupervised Learning The goal is to construct sta;s;cal model that finds useful representa;on of data: Clustering Dimensionality reduc;on Modeling the data density Finding hidden causes (useful explana;on) of the data Unsupervised Learning can be used for: Structure discovery Anomaly detec;on / Outlier detec;on Data compression, Data visualiza;on Used to aid classifica;on/regression tasks

DNA Microarray Data Expression matrix of 6830 genes (rows) and 64 samples (columns) for the human tumor data. The display is a heat map ranging from bright green (under expressed) to bright red (over expressed). Ques;ons we may ask: Which samples are similar to other samples in terms of their expression levels across genes. Which genes are similar to each other in terms of their expression levels across samples.

Linear Least Squares Given a vector of d- dimensional inputs to predict the target (response) using the linear model: we want The term w 0 is the intercept, or ooen called bias term. It will be convenient to include the constant variable 1 in x and write: Observe a training set consis;ng of N observa;ons together with corresponding target values Note that X is an matrix.

Linear Least Squares One op;on is to minimize the sum of the squares of the errors between the predic;ons for each data point x n and the corresponding real- valued targets t n. Loss func;on: sum- of- squared error func;on: Source: Wikipedia

Linear Least Squares If is nonsingular, then the unique solu;on is given by: op;mal weights vector of target values the design matrix has one input vector per row Source: Wikipedia At an arbitrary input, the predic;on is The en;re model is characterized by d+1 parameters w *.

Example: Polynomial Curve Fivng Consider observing a training set consis;ng of N 1- dimensional observa;ons: together with corresponding real- valued targets: Goal: Fit the data using a polynomial func;on of the form: The green plot is the true func;on The training data was generated by taking x n spaced uniformly between [0 1]. The target set (blue circles) was obtained by first compu;ng the corresponding values of the sin func;on, and then adding a small Gaussian noise. Note: the polynomial func;on is a nonlinear func;on of x, but it is a linear func;on of the coefficients w! Linear Models.

Example: Polynomial Curve Fivng As for the least squares example: we can minimize the sum of the squares of the errors between the predic;ons for each data point x n and the corresponding target values t n. Loss func;on: sum- of- squared error func;on: Similar to the linear least squares: Minimizing sum- of- squared error func;on has a unique solu;on w *. The model is characterized by M+1 parameters w *. How do we choose M?! Model Selec4on.

Some Fits to the Data For M=9, we have fi0ed the training data perfectly.

Overfivng Consider a separate test set containing 100 new data points generated using the same procedure that was used to generate the training data. For M=9, the training error is zero! The polynomial contains 10 degrees of freedom corresponding to 10 parameters w, and so can be fi0ed exactly to the 10 data points. However, the test error has become very large. Why?

Overfivng As M increases, the magnitude of coefficients gets larger. For M=9, the coefficients have become finely tuned to the data. Between data points, the func;on exhibits large oscilla;ons. More flexible polynomials with larger M tune to the random noise on the target values.

Varying the Size of the Data 9th order polynomial For a given model complexity, the overfivng problem becomes less severe as the size of the dataset increases. However, the number of parameters is not necessarily the most appropriate measure of model complexity.

Generaliza;on The goal is achieve good generaliza4on by making accurate predic;ons for new test data that is not known during learning. Choosing the values of parameters that minimize the loss func;on on the training data may not be the best op;on. We would like to model the true regulari;es in the data and ignore the noise in the data: - It is hard to know which regulari;es are real and which are accidental due to the par;cular training examples we happen to pick. Intui;on: We expect the model to generalize if it explains the data well given the complexity of the model. If the model has as many degrees of freedom as the data, it can fit the data perfectly. But this is not very informa;ve. Some theory on how to control model complexity to op;mize generaliza;on.

A Simple Way to Penalize Complexity One technique for controlling over- fivng phenomenon is regulariza4on, which amounts to adding a penalty term to the error func;on. penalized error func;on target value regulariza;on parameter where and is called the regulariza;on term. Note that we do not penalize the bias term w 0. The idea is to shrink es;mated parameters toward zero (or towards the mean of some other weights). Shrinking to zero: penalize coefficients based on their size. For a penalty func;on which is the sum of the squares of the parameters, this is known as weight decay, or ridge regression.

Regulariza;on Graph of the root- mean- squared training and test errors vs. ln for the M=9 polynomial. How to choose?

Cross Valida;on If the data is plen;ful, we can divide the dataset into three subsets: Training Data: used to fivng/learning the parameters of the model. Valida;on Data: not used for learning but for selec;ng the model, or choose the amount of regulariza;on that works best. Test Data: used to get performance of the final model. For many applica;ons, the supply of data for training and tes;ng is limited. To build good models, we may want to use as much training data as possible. If the valida;on set is small, we get noisy es;mate of the predic;ve performance. S fold cross- valida;on The data is par;;oned into S groups. Then S- 1 of the groups are used for training the model, which is evaluated on the remaining group. Repeat procedure for all S possible choices of the held- out group. Performance from the S runs are averaged.

Basics of Probability Theory Consider two random variables X and Y: - X takes any values x i, where i=1,..,m. - Y takes any values y j, j=1,,l. Consider a total of N trials and let the number of trials in which X = x i and Y = y j is n ij. Joint Probability: Marginal Probability: where

Basics of Probability Theory Consider two random variables X and Y: - X takes any values x i, where i=1,..,m. - Y takes any values y j, j=1,,l. Consider a total of N trials and let the number of trials in which X = x i and Y = y j is n ij. Marginal probability can be wri0en as: Called marginal probability because it is obtained by marginalizing, or summing out, the other variables.

Basics of Probability Theory Condi;onal Probability: We can derive the following rela;onship: which is the product rule of probability.

The Rules of Probability Sum Rule Product Rule

Bayes Rule From the product rule, together with symmetric property: Remember the sum rule: We will revisit Bayes Rule later in class.

Illustra;ve Example Distribu;on over two variables: X takes on 9 possible values, and Y takes on 2 possible values.

Probability Density If the probability of a real- valued variable x falling in the interval is given by then p(x) is called the probability density over x. The probability density must sa;sfy the following two condi;ons

Probability Density Cumula;ve distribu;on func;on is defined as: which also sa;sfies: The sum and product rules take similar forms:

Expecta;ons The average value of some func;on f(x) under a probability distribu;on (density) p(x) is called the expecta;on of f(x): If we are given a finite number N of points drawn from the probability distribu;on (density), then the expecta;on can be approximated as: Condi;onal Expecta;on with respect to the condi;onal distribu;on.

Variances and Covariances The variance of f(x) is defined as: which measures how much variability there is in f(x) around its mean value E[f(x)]. Note that if f(x) = x, then

Variances and Covariances For two random variables x and y, the covariance is defined as: which measures the extent to which x and y vary together. If x and y are independent, then their covariance vanishes. For two vectors of random variables x and y, the covariance is a matrix:

The Gaussian Distribu;on For the case of single real- valued variable x, the Gaussian distribu;on is defined as: which is governed by two parameters: - µ (mean) - ¾ 2 (variance) is called the precision. Next class, we will look at various distribu;ons as well as at mul;variate extension of the Gaussian distribu;on.

The Gaussian Distribu;on For the case of single real- valued variable x, the Gaussian distribu;on is defined as: The Gaussian distribu;on sa;sfies: which sa;sfies the two requirements for a valid probability density

Mean and Variance Expected value of x takes the following form: Because the parameter µ represents the average value of x under the distribu;on, it is referred to as the mean. Similarly, the second order moment takes form: It then follows that the variance of x is given by:

Sampling Assump;ons Suppose we have a dataset of observa;ons x = (x 1,,x N ) T, represen;ng N 1- dimensional observa;ons. Assume that the training examples are drawn independently from the set of all possible examples, or from the same underlying distribu;on We also assume that the training examples are iden;cally distributed! i.i.d assump;on. Assume that the test samples are drawn in exactly the same way - - i.i.d from the same distribu;on as the test data. These assump;ons make it unlikely that some strong regularity in the training data will be absent in the test data.

Gaussian Parameter Es;ma;on Suppose we have a dataset of i.i.d. observa;ons x = (x 1,,x N ) T, represen;ng N 1- dimensional observa;ons. Because out dataset x is i.i.d., we can write down the joint probability of all the data points as given µ and ¾ 2 : Likelihood func;on When viewed as a func;on of µ and ¾ 2, this is called the likelihood func;on for the Gaussian.

Maximum (log) likelihood The log- likelihood can be wri0en as: Maximizing w.r.t µ gives: Sample mean Likelihood func;on Maximizing w.r.t ¾ 2 gives: Sample variance

Proper;es of the ML es;ma;on ML approach systema;cally underes;mates the variance of the distribu;on. This is an example of a phenomenon called bias. It is straighmorward to show that: It follows that the following es;mate is unbiased:

Proper;es of the ML es;ma;on Example of how bias arises in using ML to determine the variance of a Gaussian distribu;on: The green curve shows the true Gaussian distribu;on. Fit 3 datasets, each consis;ng of two blue points. Averaged across 3 datasets, the mean is correct. But the variance is under- es;mated because it is measured rela;ve to the sample (and not the true) mean.

Probabilis;c Perspec;ve So far we saw that polynomial curve fivng can be expressed in terms of error minimiza;on. We now view it from probabilis;c perspec;ve. Suppose that our model arose from a sta;s;cal model: where ² is a random error having Gaussian distribu;on with zero mean, and is independent of x. Thus we have: where is a precision parameter, corresponding to the inverse variance. I will use probability distribution and probability density interchangeably. It should be obvious from the context.!

Maximum Likelihood If the data are assumed to be independently and iden;cally distributed (i.i.d assump*on), the likelihood func;on takes form: It is ooen convenient to maximize the log of the likelihood func;on: Maximizing log- likelihood with respect to w (under the assump;on of a Gaussian noise) is equivalent to minimizing the sum- of- squared error func;on. Determine w.r.t. : by maximizing log- likelihood. Then maximizing

Predic;ve Distribu;on Once we determined the parameters w and, we can make predic;on for new values of x: Later we will consider Bayesian linear regression.

Sta;s;cal Decision Theory We now develop a small amount of theory that provides a framework for developing many of the models we consider. Suppose we have a real- valued input vector x and a corresponding target (output) value t with joint probability distribu;on: Our goal is predict target t given a new value for x: - for regression: t is a real- valued con;nuous target. - for classifica;on: t a categorical variable represen;ng class labels. The joint probability distribu;on provides a complete summary of uncertain;es associated with these random variables. Determining from training data is known as the inference problem.