Lecture Slides for. ETHEM ALPAYDIN The MIT Press,

Similar documents
E. Alpaydın AERFAISS

Machine Learning. Lecture Slides for. ETHEM ALPAYDIN The MIT Press, h1p://

Lecture Slides for INTRODUCTION TO. Machine Learning. ETHEM ALPAYDIN The MIT Press,

Kernel Methods and SVMs Extension

Excess Error, Approximation Error, and Estimation Error

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

1 Definition of Rademacher Complexity

PROBABILITY AND STATISTICS Vol. III - Analysis of Variance and Analysis of Covariance - V. Nollau ANALYSIS OF VARIANCE AND ANALYSIS OF COVARIANCE

Lecture 6 More on Complete Randomized Block Design (RBD)

Chapter 12 Analysis of Covariance

Statistics for Economics & Business

COS 511: Theoretical Machine Learning

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

PGM Learning Tasks and Metrics

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

Ensemble Methods: Boosting

Lecture 6: Introduction to Linear Regression

Reliability estimation in Pareto-I distribution based on progressively type II censored sample with binomial removals

Regression Analysis. Regression Analysis

# c i. INFERENCE FOR CONTRASTS (Chapter 4) It's unbiased: Recall: A contrast is a linear combination of effects with coefficients summing to zero:

Statistics II Final Exam 26/6/18

Chapter 11: I = 2 samples independent samples paired samples Chapter 12: I 3 samples of equal size J one-way layout two-way layout

STAT 3008 Applied Regression Analysis

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

UCLA STAT 13 Introduction to Statistical Methods for the Life and Health Sciences. Chapter 11 Analysis of Variance - ANOVA. Instructor: Ivo Dinov,

Modeling and Simulation NETW 707

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION MTH352/MH3510 Regression Analysis

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

Lecture 4 Hypothesis Testing

Economics 130. Lecture 4 Simple Linear Regression Continued

Chapter 5: Hypothesis Tests, Confidence Intervals & Gauss-Markov Result

THE CHINESE REMAINDER THEOREM. We should thank the Chinese for their wonderful remainder theorem. Glenn Stevens

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

Hydrological statistics. Hydrological statistics and extremes

Methods in Epidemiology. Medical statistics 02/11/2014. Estimation How large is the effect? At the end of the lecture students should be able

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. Introductory Econometrics 1 hour 30 minutes

Basic Business Statistics, 10/e

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

Topic- 11 The Analysis of Variance

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

x i1 =1 for all i (the constant ).

Comparative Analysis of Bradley-Terry and Thurstone-Mosteller Paired Comparison Models for Image Quality Assessment

First Year Examination Department of Statistics, University of Florida

10-701/ Machine Learning, Fall 2005 Homework 3

Homework 10 Stat 547. Problem ) Z D!

Answers Problem Set 2 Chem 314A Williamsen Spring 2000

Statistical tables are provided Two Hours UNIVERSITY OF MANCHESTER. Date: Wednesday 4 th June 2008 Time: 1400 to 1600

Comparison of Regression Lines

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

Chapter 5 Multilevel Models

Support Vector Machines

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup

Generative classification models

17 - LINEAR REGRESSION II

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

F statistic = s2 1 s 2 ( F for Fisher )

a. (All your answers should be in the letter!

Which Separator? Spring 1

January Examinations 2015

Machine Learning. What is a good Decision Boundary? Support Vector Machines

Statistics for Business and Economics

The Impact of the Earth s Movement through the Space on Measuring the Velocity of Light

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

experimenteel en correlationeel onderzoek

Topic 7: Analysis of Variance

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Chapter 13 Analysis of Variance and Experimental Design

Development of a Semi-Automated Approach for Regional Corrector Surface Modeling in GPS-Levelling

Reduced slides. Introduction to Analysis of Variance (ANOVA) Part 1. Single factor

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

CHAPT II : Prob-stats, estimation

Learning Objectives for Chapter 11

Question 1 carries a weight of 25%; question 2 carries 20%; question 3 carries 25%; and question 4 carries 30%.

Collaborative Filtering Recommendation Algorithm

Face Recognition Using Ada-Boosted Gabor Features

Perceptual Organization (IV)

STATISTICS QUESTIONS. Step by Step Solutions.

Computational and Statistical Learning theory Assignment 4

Chapter 6 Support vector machine. Séparateurs à vaste marge

Negative Binomial Regression

Two-factor model. Statistical Models. Least Squares estimation in LM two-factor model. Rats

ANOVA. The Observations y ij

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

Outline. EEC 686/785 Modeling & Performance Evaluation of Computer Systems. Lecture k r Factorial Designs with Replication

+, where 0 x N - n. k k

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Bayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County

Regularized Discriminant Analysis for Face Recognition

Reminder: Nested models. Lecture 9: Interactions, Quadratic terms and Splines. Effect Modification. Model 1

CHAPTER IV RESEARCH FINDING AND DISCUSSIONS

ANALYSIS OF SIMULATION EXPERIMENTS BY BOOTSTRAP RESAMPLING. Russell C.H. Cheng

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

Chapter 14 Simple Linear Regression

Study of Classification Methods Based on Three Learning Criteria and Two Basis Functions

Transcription:

ecture Sldes for ETHEM APAYDI The MIT Press, 00 alpaydn@boun.edu.tr http://www.cpe.boun.edu.tr/~ethe/le

Introducton Questons: Assessent of the expected error of a learnng algorth: Is the error rate of - less than %? Coparng the expected errors of two algorths: Is k- ore accurate than MP? Tranng/valdaton/test sets Resaplng ethods: -fold cross-valdaton ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 3

Algorth Preference Crtera (Applcaton-dependent): Msclassfcaton error, or rsk (loss functons) Tranng te/space coplexty Testng te/space coplexty Interpretablty Easy prograablty Cost-senstve learnng ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 4

Factors and Response ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 5

Strateges of Experentaton Response surface desgn for approxatng and axzng the response functon n ters of the controllable factors ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 6

Gudelnes for M experents A. A of the study B. Selecton of the response varable C. Choce of factors and levels D. Choce of experental desgn E. Perforng the experent F. Statstcal Analyss of the Data G. Conclusons and Recoendatons ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 7

The need for ultple tranng/valdaton sets {,V } : Tranng/valdaton sets of fold -fold cross-valdaton: Dvde nto k,,=,..., T share - parts Resaplng and -Fold Cross-Valdaton 3 3 T V T V T V 8 ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

5 Cross-Valdaton 5 0 5 0 5 9 5 9 4 4 3 3 V T V T V T V T V T V T 9 5 tes fold cross-valdaton (Detterch, 998) ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

Bootstrappng Draw nstances fro a dataset wth replaceent Prob that we do not pck an nstance after draws e 0. 368 that s, only 36.8% s new! ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 0

Measurng Error Error rate = # of errors / # of nstances = (F+FP) / Recall = # of found postves / # of postves = TP / (TP+F) = senstvty = ht rate Precson = # of found postves / # of found = TP / (TP+FP) Specfcty = T / (T+FP) False alar rate = FP / (FP+T) = - Specfcty ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

ROC Curve ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 3

Precson and Recall ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 4

Interval Estaton 0 95 96 96 0 95 96 96 z z P P P / /...... ~ Z 5 = { x t } t where x t ~ ( μ, σ ) ~ ( μ, σ /) 00(- α) percent confdence nterval ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

6 When σ s not known: 0 95 64 0 95 64 z P P P.... S t S t P t S x S t t, /, / ~ / ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

Hypothess Testng Reect a null hypothess f not supported by the saple wth enough confdence = { x t } t where x t ~ ( μ, σ ) H 0 : μ = μ 0 vs. H : μ μ 0 Accept H 0 wth level of sgnfcance α f μ 0 s n the 00(- α) confdence nterval Two-sded test 0 z z /, / ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 7

One-sded test: H 0 : μ μ 0 vs. H : μ > μ 0 Accept f Varance unknown: Use t, nstead of z Accept H 0 : μ = μ 0 f 0, z 0 S t t /,, /, ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 8

Assessng Error: H 0 : p p 0 vs. H : p > p 0 Sngle tranng/valdaton set: Bnoal Test If error prob s p 0, prob that there are e errors or less n valdaton trals s e P e p0 p0 Accept f ths prob s less than - α - α =00, e=0 ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 9

oral Approxaton to the Bnoal uber of errors s approx wth ean p 0 and var p 0 (-p 0 ) p p 0 0 p 0 ~ Z Accept f ths prob for = e s less than z -α - α ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 0

t Test Multple tranng/valdaton sets x t = f nstance t sclassfed on fold t Error rate of fold : x p t Wth and s average and var of p, we accept p 0 or less error f p0 ~ t S s less than t α,- ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

Coparng Classfers: H 0 : μ 0 = μ vs. H : μ 0 μ Sngle tranng/valdaton set: Mcear s Test Under H 0, we expect e 0 = e 0 =(e 0 + e 0 )/ Accept f < α, e0 e0 e 0 e 0 ~ ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

-Fold CV Pared t Test 0 0 0 0 0 t t t s s p s p H H, /, /, ~ : : n f Accept vs. 3 Use -fold cv to get tranng/valdaton folds p, p : Errors of classfers and on fold p = p p : Pared dfference on fold The null hypothess s whether p has ean 0 ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

5 cv Pared t Test Use 5 cv to get folds of 5 tra/val replcatons (Detterch, 998) p () : dfference btw errors of and on fold =, of replcaton =,...,5 p p p / s p p 5 s p p / 5 ~ t Two-sded test: Accept H 0 : μ 0 = μ f n (-t α/,5,t α/,5 ) One-sded test: Accept H 0 : μ 0 μ f < t α,5 5 p ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 4

5 cv Pared F Test 5 5 s p ~ F 0, 5 Two-sded test: Accept H 0 : μ 0 = μ f < F α,0,5 ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 5

Coparng > Algorths: Analyss of Varance (Anova) H : 0 Errors of algorths on folds,,,...,, ~,..., We construct two estators to σ. One s vald f H 0 s true, the other s always vald. We reect H 0 f the two estators dsagree. ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 6

7 0 0 SSb H SSb S S H ~ ~ / ˆ /, ~ : we have s true, So when, naely, s Thus anestatorof s true If ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

8 0 0 F H F SSw SSb SSw SSb SSw S SSw S S S H,,, : ~ / / / / / ~ ~ ˆ f : group varances average of s the secondestator to our Regardlessof ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

AOVA table 9 If AOVA reects, we do parwse posthoc tests ) ( ~ : vs : 0 w t t H H ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0)

Coparson over Multple Datasets Coparng two algorths: Sgn test: Count how any tes A beats B over datasets, and check f ths could have been by chance f A and B dd have the sae error rate Coparng ultple algorths ruskal-walls test: Calculate the average rank of all algorths on datasets, and check f these could have been by chance f they all had equal error If W reects, we do parwse posthoc tests to fnd whch ones have sgnfcant rank dfference ecture otes for E Alpaydın 00 Introducton to Machne earnng e The MIT Press (V.0) 30