Data Mining with Linear Discriminants. Exercise: Business Intelligence (Part 6) Summer Term 2014 Stefan Feuerriegel

Size: px
Start display at page:

Download "Data Mining with Linear Discriminants. Exercise: Business Intelligence (Part 6) Summer Term 2014 Stefan Feuerriegel"

Transcription

1 Data Mining with Linear Discriminants Exercise: Business Intelligence (Part 6) Summer Term 2014 Stefan Feuerriegel

2 Tday s Lecture Objectives 1 Recgnizing the ideas f artificial neural netwrks and their use in R 2 Understanding the cncept and the usage f supprt vectr machines 3 Being able t evaluate the predictive perfrmance in terms f bth metrics and the receiver perating characteristic curve 4 Distinguishing predictive and explanatry pwer Data Mining, with Linear Discriminants 2

3 Outline 1 Recap 2 Linear Discriminants 3 Artificial Neural Netwrks 4 Supprt Vectr Machines 5 Predictin Perfrmance 6 Wrap-Up Data Mining, with Linear Discriminants 3

4 Outline 1 Recap 2 Linear Discriminants 3 Artificial Neural Netwrks 4 Supprt Vectr Machines 5 Predictin Perfrmance 6 Wrap-Up Data Mining, with Linear Discriminants: Recap 4

5 Supervised vs. Unsupervised Learning Supervised learning Machine learning task f inferring a functin frm labeled training data Training data includes bth the input and the desired results crrect results (target values) are given Unsupervised learning Methds try t find hidden structure in unlabeled data The mdel is nt prvided with the crrect results during the training N errr r reward signal t evaluate a ptential slutin Examples: Hidden Markv mdels Dimensin reductin (e. g. by principal cmpnent analysis) Clustering (e. g. by k-means algrithm) grup int classes n the basis f their statistical prperties nly Data Mining, with Linear Discriminants: Recap 5

6 Feature B Feature B Taxnmy f Machine Learning Machine learning estimates functin and parameter in y = f(x,w) Type f methd varies depending n the nature f what is predicted Regressin Predicted value refers t a real number Cntinuus y Classificatin Clustering Predicted value refers t a class label Discrete y (e. g. class membership) Grup pints int clusters based n hw "near" they are t ne anther Identify structure in data Feature A Data Mining, with Linear Discriminants: Recap Feature A 6

7 K -Nearest Neighbr Classificatin Input: training examples as vectrs in a multidimensinal feature space, each with a class label N training phase t calculate internal parameters Testing: Assign t class accrding t k-nearest neighbrs Classificatin as majrity vte Prblematic Skewed data Unequal frequency f classes What label t assign t the circle?? Data Mining, with Linear Discriminants: Recap 7

8 Decisin Trees Flwchart-like structure in which ndes represent tests n attributes End ndes (leaves) f each branch represent class labels Example: Decisin tree fr playing tennis Outlk Sunny Overcast Rain Humidity Yes Wind High Nrmal Strng Weak N Yes N Yes Data Mining, with Linear Discriminants: Recap 8

9 Decisin Trees Issues Hw deep t grw? Hw t handle cntinuus attributes? Hw t chse an apprpriate attributes selectin measure? Hw t handle data with missing attributes values? Advantages Simple t understand and interpret Requires nly few bservatins Wrds, best and expected values can be determined fr different scenaris Disadvantages Infrmatin Gain criterin is biased in favr f attributes with mre levels Calculatins becme cmplex if values are uncertain and/r utcmes are linked Data Mining, with Linear Discriminants: Recap 9

10 Feature B k-means Clustering Partitin n bservatins int k clusters in which each bservatin belngs t the cluster with the nearest mean, serving as a prttype f the cluster Feature A Cmputatinally expensive; instead, we use efficient heuristics Default: Euclidean distance as metric and variance as a measure f cluster scatter Data Mining, with Linear Discriminants: Recap 10

11 Outline 1 Recap 2 Linear Discriminants 3 Artificial Neural Netwrks 4 Supprt Vectr Machines 5 Predictin Perfrmance 6 Wrap-Up Data Mining, with Linear Discriminants: Linear Discriminants 11

12 Classificatin Prblems General classificatin prblem Gal: Take a new input x and assign it t ne f K classes C k Given training set X = [x 1... x n ] T with target values T = [t 1,...,t n ] T Number f dimensins D, i. e. x i R D Learn a discriminant functin y(x) t perfrm the classificatin 2-class prblem with binary target values t i {0,1} Decide fr class C 1 if y(x) > 0, else fr class C 2 K -class prblem with 1-f-K cding scheme, i. e. t i [0,1,0,0,0] T Data Mining, with Linear Discriminants: Linear Discriminants 12

13 Feature B Feature B Linear Separability If a data set can be perfectly classified by a linear discriminant, then we call it linearly separable Feature A Linearly separable Feature A Nt linearly separable Data Mining, with Linear Discriminants: Linear Discriminants 13

14 Linear Discriminant Functins Decisin bundary given by y(x) = w T x + b = 0 defines a hyperplane Classes labeled accrding t sign ( w T x + b ) Nrmal vectr w and ffset b w y(x) > 0 y(x) = 0 x 2 y(x) < 0 w x 1 Data Mining, with Linear Discriminants: Linear Discriminants 14

15 Learning Discriminant Functins Linear discriminant functins given by y(x) = w T x + b = = D i=0 D i=1 w i x i + b w i x i with x 0 = 1 Weight vectr w "Bias" b, i. e. threshld Gal: Chse w and b, r w respectively, such that b t 1... w T X +. w T X T is minimal b t n Data Mining, with Linear Discriminants: Linear Discriminants 15

16 Chsing Discriminant Functin Slving w T X T by least-squares has drawbacks Least-squares is very sensitive t utliers Errr functin penalizes predictins that are "t crrect" Wrks nly fr linearly separable prblems Least-squares assumes Gaussian distributin c c 2 0 c c Alternative slutins (e. g. in blue): Generalized linear mdels ( neural netwrks), supprt vectr machines, etc. frm Leibe (2010). Data Mining, with Linear Discriminants: Linear Discriminants 16

17 Outline 1 Recap 2 Linear Discriminants 3 Artificial Neural Netwrks 4 Supprt Vectr Machines 5 Predictin Perfrmance 6 Wrap-Up Data Mining, with Linear Discriminants: Neural Netwrks 17

18 Generalized Linear Mdel Linear mdel y(x) = w T x + b Generalized linear mdel with activatin functin A y(x) = A ( w T x + b ) Other than least-squares, chice f activatin functin shuld limit influence f utliers, e. g. using a threshld as A A(z) Data Mining, with Linear Discriminants: Neural Netwrks 18

19 Relatinship t Neural Netwrks In 2-class case y(x) = with x 0 = 1 D i=0 A(w i x i ) Single-layer perceptrn threshld w0 w1 y(x) wd utput weights In multi-class case y k (x) = with x 0 = 1 D i=0 A(w k,i x i ) Multi-class perceptrn threshlds wk0 y1(x) yk(x) utputs weights wki x0=1 x1 xd inputs X0 = 1 x1 xd inputs Data Mining, with Linear Discriminants: Neural Netwrks 19

20 Artificial Neural Netwrks Artificial neural netwrks (ANN) are cmputatinal mdels t cmpute f : X Y, inspired by the central nervus system Cmpute f by feeding infrmatin thrugh the netwrk Represented as a system f cnnected neurns ANNs are universal apprximatrs amng cntinuus functins (under certain mild assumptins) Input Hidden Output x 1 z 1 x N z M y 1 y 2 Data Mining, with Linear Discriminants: Neural Netwrks 20

21 Layers in Neural Netwrks Neurns are arranged in three (r mre) layers First layer: Input neurns receive the input vectr x X Hidden layer(s): Cnnect input and utput neurns Final layer: Output neurns cmpute a respnse ỹ Y Input Hidden Output x 1 z 1 x N z M y 1 y 2 When neurns are cnnected as a directed graph withut cycles, this is called a feed-frward ANN Data Mining, with Linear Discriminants: Neural Netwrks 21

22 Feeding Infrmatin thrugh Neural Netwrks Input z j f each neurn j = 1,...,M is a weighted sum f all previus neurns calculated as ( where z j = A w 0,j + N i=1 w i,j x i ) = A ( w 0,j + w T j x ) x i are the values frm the input layer suitable cefficients w i,j fr i = 1,...,N and j = 1,...,M Predefined nn-linear functin A is referred t as the activatin functin Frequent chice: Lgistic functin 1 A(z) = 1 + e z Cefficients w i,j learned frm e. g. a back-prpagatin algrithm Data Mining, with Linear Discriminants: Neural Netwrks 22

23 Lgistic Functin Lgistic A(z) z Resembles a threshld functin { 0, z < 0, A(z) 1, z > 0 Data Mining, with Linear Discriminants: Neural Netwrks 23

24 Neural Netwrks in R Lading required library nnet library(nnet) Accessing credit scres library(caret) data(germancredit) Split data int index subset fr training (20%) and testing (80%) instances intrain <- runif(nrw(germancredit)) < 0.2 Data Mining, with Linear Discriminants: Neural Netwrks 24

25 Neural Netwrks in R Train neural netwrk with n ndes in the hidden layer via nnet(frmula, data=d, size=n...) nn <- nnet(class ~., data=germancredit[intrain,], size=15, maxit=100, rang=0.1, decay=5e-4) ## # weights: 946 ## initial value ## iter 10 value ## iter 20 value ## iter 30 value ## iter 40 value ## iter 50 value ## iter 60 value ## iter 70 value ## iter 80 value ## iter 90 value ## iter 100 value ## final value ## stpped after 100 iteratins Data Mining, with Linear Discriminants: Neural Netwrks 25

26 Neural Netwrks in R Predict credit scres via predict(nn, test, type="class") pred <- predict(nn, GermanCredit[-inTrain,], type="class") Cnfusin matrix via table(pred=pred_classes, true=true_classes) cm <- table(pred=pred, true=germancredit$class[-intrain]) Calculate accuracy sum(diag(cm))/sum(sum(cm)) ## [1] Data Mining, with Linear Discriminants: Neural Netwrks 26

27 Outline 1 Recap 2 Linear Discriminants 3 Artificial Neural Netwrks 4 Supprt Vectr Machines 5 Predictin Perfrmance 6 Wrap-Up Data Mining, with Linear Discriminants: Supprt Vectr Machines 27

28 Supprt Vectr Machine (SVM) Which f these linear separatrs is ptimal? Idea: Maximize separating margin (here: A) Data pints n the margin are called supprt vectrs When calculating decisin bundary, nly supprt vectrs matter; ther training data is ignred Frmulatin as cnvex ptimizatin prblem with glbal slutin y B A x Data Mining, with Linear Discriminants: Supprt Vectr Machines 28

29 SVM in R Lading required library e1071 library(e1071) Accessing credit scres library(caret) data(germancredit) Split data int index subset fr training (20%) and testing (80%) instances intrain <- runif(nrw(germancredit)) < 0.2 Data Mining, with Linear Discriminants: Supprt Vectr Machines 29

30 SVM in R Train supprt vectr machine fr classificatin via svm(frmula, data=d, type="c-classificatin") mdel <- svm(class ~., data=germancredit[intrain,], type="c-classificatin") Predict credit scres fr testing instances test via predict(svm, test) pred <- predict(mdel, GermanCredit[-inTrain, ]) head(cbind(pred, GermanCredit$Class[-inTrain])) ## pred ## ## ## ## ## ## First rw gives predicted utcmes, secnd are actual (true) values Data Mining, with Linear Discriminants: Supprt Vectr Machines 30

31 SVM in R Cnfusin matrix via table(pred=pred_classes, true=true_classes) cm <- table(pred=pred, true=germancredit$class[-intrain]) cm ## true ## pred Bad Gd ## Bad 57 1 ## Gd Calculate accuracy sum(diag(cm))/sum(sum(cm)) ## [1] Data Mining, with Linear Discriminants: Supprt Vectr Machines 31

32 Outline 1 Recap 2 Linear Discriminants 3 Artificial Neural Netwrks 4 Supprt Vectr Machines 5 Predictin Perfrmance 6 Wrap-Up Data Mining, with Linear Discriminants: Predictin Perfrmance 32

33 Assessment f Mdels 1 Predictive perfrmance (measured by accuracy, recall, F1, ROC,... ) 2 Cmputatin time fr bth mdel building and predicting 3 Rbustness t nise in predictr values 4 Scalability 5 Interpretability transparency, ease f understanding Mdel Allws n < k Interpret. # Tuning Parameters Rbust t Predictr Nise Cmp. Time Linear Regressin 0 ANN 1 2 SVM/SVR 1 3 k-nn 1 Single Tree 1 Randm Frest 0 1 Bsted Trees 3 frm Kuhn & Jhnsn (2013). Applied Predictive Mdeling. Data Mining, with Linear Discriminants: Predictin Perfrmance 33

34 Accuracy vs. Precisin Accuracy Clseness t the actual (true) value Precisin Similarity f repeated measurements under unchanged cnditins High accuracy, but lw precisin High precisin, but lw accuracy Data Mining, with Linear Discriminants: Predictin Perfrmance 34

35 Cnfusin Matrix Cnfusin matrix (als named cntingency table r errr matrix) displays predictive perfrmance Cnditin (as determined by Gld standard) True False Psitive Outcme True Psitive (TP) False Psitive (FP) Type I Errr False Alarm Precisin r Psitive Predictive Value = TP TP+FP Negative Outcme False Negative (FN) Type II Errr / Miss True Negative (TN) Sensitivity = TP Rate = TP TP+FN Specificity = TN Rate = TN FP+TN Accuracy = TP+TN Ttal Equivalent with hit rate and recall Data Mining, with Linear Discriminants: Predictin Perfrmance 35

36 Cnfusin Matrix Example: Bld prbe t test fr cancer Patient with Cancer True False Psitive Bld Test Outcme TP: Cancer crrectly diagnsed FP: Healthy persn diagnsed cancer Precisin = TP TP+FP Negative Bld Test Outcme FN: Cancer nt diagnsed TN: Healthy persn diagnsed as healthy Sensitivity = TP Rate = TP TP+FN Specificity = TN Rate = TN FP+TN Accuracy = TP+TN Ttal Different lss functins: Redundant, e 1000 check in FP case, cmpared t lethal utcme in FN case Data Mining, with Linear Discriminants: Predictin Perfrmance 36

37 Assessing Predictin Perfrmance Imagine the fllwing cnfusin matrix with an accuracy f 65% Patient with Cancer True False Psitive Bld Test Outcme Negative Bld Test Outcme TP = 60 FP = 5 Precisin = TP TP+FP 0.92 FN = 30 TN = 5 Sensitivity = TP TP+FN 0.67 Specificity = TN FP+TN = 0.50 Accuracy = TP+TN Ttal = 0.65 Is this a "gd" result? N, because f unevenly distributed data, a mdel which always guesses psitive will scre an accuracy f 60% Data Mining, with Linear Discriminants: Predictin Perfrmance 37

38 Predictin Perfrmance in R Cnfusin matrix in variable cm ## True False ## Ps ## Neg Calculating accuracy (cm[1,1]+cm[2,2])/ (cm[1,1]+cm[1,2]+cm[2,1]+cm[2,2]) ## [1] 0.7 # Alternative that wrks als fr multi-class data sum(diag(cm))/sum(sum(cm)) ## [1] 0.7 Calculating precisin cm[1, 1]/(cm[1, 1] + cm[1, 2]) ## [1] 0.75 Data Mining, with Linear Discriminants: Predictin Perfrmance 38

39 Trade-Off: Sensitivity vs. Specificity/Precisin Perfrmance gals frequently place mre emphasis n either sensitivity r specificity/precisin Example: Airprt scanners triggered n lw-risk items like belts (lw precisin), but reduce risk f missing risky bjects (high sensitivity) Trade-Off: F1 scre is the harmnic mean f precisin and sensitivity 2TP F1 = 2TP + FP + FN Visualized by receiver perating characteristic (ROC) curve Data Mining, with Linear Discriminants: Predictin Perfrmance 39

40 Sensitivity Receiver Operating Characteristic (ROC) ROC illustrates perfrmance f binary classifier as its discriminatin threshld y(x) is varied Interpretatin: Curve A is randm guessing (50% crrect guesses) Curve frm mdel B perfrms better than A, but wrse than C Curve C frm perfect predictin Area suth-east f curve is named area under the curve and shuld be maximized C B A Specifity Data Mining, with Linear Discriminants: Predictin Perfrmance 40

41 ROC in R Lad required library proc library(proc) Perfrm predictin and retrieve decisin values dv pred <- predict(mdel, GermanCredit[-inTrain,], decisin.values=true) dv <- attributes(pred)$decisin.values Data Mining, with Linear Discriminants: Predictin Perfrmance 41

42 ROC in R Plt ROC curve via plt.rc(classes, dv) plt.rc(as.numeric(germancredit$class[-intrain]), dv, xlab = "1 - Specificity") Sensitivity Specificity ## ## Call: ## plt.rc.default(x = as.numeric(germancredit$class[-intrain]), predictr = dv, xlab = ## ## Data: dv in 300 cntrls (as.numeric(germancredit$class[-intrain]) 1) > 699 cases (as.num ## Area under the curve: Data Mining, with Linear Discriminants: Predictin Perfrmance 42

43 Predictive vs. Explanatry Pwer Significant difference between predicting and explaining: 1 Empirical Mdels fr Predictin Empirical predictive mdels (e. g. statistical mdels, methds frm data mining) designed t predict new/future bservatins Predictive Analytics describes the evaluatin f the predictive pwer, such as accuracy r precisin 2 Empirical Mdels fr Explanatin Any type f statistical mdel used fr testing causal hypthesis Use methds fr evaluating the explanatry pwer, such as statistical tests r measures like R 2 Data Mining, with Linear Discriminants: Predictin Perfrmance 43

44 Predictive vs. Explanatry Pwer N = 15 N = 100 Explanatry pwer des nt imply predictive pwer Red is the best explanatry mdel; gray the best predictive In particular, dummies d nt translate well t predictive mdes D nt write smething like "the regressin prves the predictive pwer f regressr x i " 44 Data Mining, with Linear Discriminants: Predictin Perfrmance

45 Overfitting When learning algrithm is perfrmed fr t lng, the learner may adjust t very specific randm features nt related t the target functin Overfitting: Perfrmance n training data (in gray) still increases, while the perfrmance n unseen data (in red) becmes wrse y Mean Squared Errr x Flexibility Data Mining, with Linear Discriminants: Predictin Perfrmance 45

46 Overfitting in R Given data x t predict y plt(x, y, pch = 20) Estimate plynmials f rder P = 1,...,6 and cmpare errrs n training and testing data intrain <- runif(length(x)) < 0.2 err.fit <- c() err.pred <- c() y fr (i in 1:6) { # Explanatry mdel xx <- x[intrain] m <- lm(y[intrain] ~ ply(xx, i, raw=true)) err.fit[i] <- sqrt(mean(m$residuals^2)) # Predictive mdel xx <- x[-intrain] err <- predict(m, ply(xx, i, raw=true))-y[-intrain] err.pred[i] <- sqrt(mean(err^2)) } x Data Mining, with Linear Discriminants: Predictin Perfrmance 46

47 Overfitting in R Perfrmance n training data (in gray) still increases, while the perfrmance n unseen data (in red) becmes wrse plt(1:6, err.fit, type='l', ylim=c(3.5, 31), cl="gray", xlab="order f plynmial", ylab="errr n training/testing data") lines(1:6, err.pred, cl="firebrick3") Errr n training/testing data Order f plynmial Data Mining, with Linear Discriminants: Predictin Perfrmance 47

48 Supprt Vectr Regressin in R Supprt vectr regressin (SVR) predicts cntinuus values Accessing weekly stck market returns fr 21 years library(islr) data(weekly) intrain <- runif(nrw(weekly)) < 0.2 When training, use type="eps-regressin" instead mdel <- svm(tday ~ Lag1 + Lag2 + Lag3 + Lag4 + Lag5, data=weekly[intrain,], type="eps-regressin") Cmpute rt mean square errr (RMSE) pred <- predict(mdel, Weekly[-inTrain, ]) sqrt(mean(pred - Weekly$Tday[-inTrain])^2) ## [1] given by RMSE = 1 N N i=1 (pred i true i ) 2 and cmpare t standard deviatin f σ 2 = (N in denminatr, nt N 1) Data Mining, with Linear Discriminants: Predictin Perfrmance 48

49 Outline 1 Recap 2 Linear Discriminants 3 Artificial Neural Netwrks 4 Supprt Vectr Machines 5 Predictin Perfrmance 6 Wrap-Up Data Mining, with Linear Discriminants: Wrap-Up 49

50 Summary: Data Mining with Linear Discriminants Linearly separable Artificial neural netwrk Supprt vectr machine Cnfusin matrix Data can be perfectly classified by linear discriminant Feeding infrmatin thrugh the netwrk nnet(frmula, data=d, size=n...) Maximize separating margin svm(frmula, data=d, type=...) Tabular utline f crrect/incrrect guesses Predictive perfrmance Measured by accuracy, precisin,... ROC curve Overfitting Visualize trade-ff between sensitivity and specificity Perfrmance n training data increases, different frm unseen bservatins Data Mining, with Linear Discriminants: Wrap-Up 50

Chapter 3: Cluster Analysis

Chapter 3: Cluster Analysis Chapter 3: Cluster Analysis } 3.1 Basic Cncepts f Clustering 3.1.1 Cluster Analysis 3.1. Clustering Categries } 3. Partitining Methds 3..1 The principle 3.. K-Means Methd 3..3 K-Medids Methd 3..4 CLARA

More information

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017 Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with

More information

Pattern Recognition 2014 Support Vector Machines

Pattern Recognition 2014 Support Vector Machines Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft

More information

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels Mtivating Example Memry-Based Learning Instance-Based Learning K-earest eighbr Inductive Assumptin Similar inputs map t similar utputs If nt true => learning is impssible If true => learning reduces t

More information

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares

More information

Simple Linear Regression (single variable)

Simple Linear Regression (single variable) Simple Linear Regressin (single variable) Intrductin t Machine Learning Marek Petrik January 31, 2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse

More information

Text Mining. Exercise: Business Intelligence (Part 7) Summer Term 2014 Stefan Feuerriegel

Text Mining. Exercise: Business Intelligence (Part 7) Summer Term 2014 Stefan Feuerriegel Text Mining Exercise: Business Intelligence (Part 7) Summer Term 2014 Stefan Feuerriegel Tday s Lecture Objectives 1 Being able t perfrm preprcessing steps fr text mining 2 Learning the representatin as

More information

COMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d)

COMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d) COMP 551 Applied Machine Learning Lecture 9: Supprt Vectr Machines (cnt d) Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Class web page: www.cs.mcgill.ca/~hvanh2/cmp551 Unless therwise

More information

IAML: Support Vector Machines

IAML: Support Vector Machines 1 / 22 IAML: Supprt Vectr Machines Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester 1 2 / 22 Outline Separating hyperplane with maimum margin Nn-separable training data Epanding the input int

More information

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must M.E. Aggune, M.J. Dambrg, M.A. El-Sharkawi, R.J. Marks II and L.E. Atlas, "Dynamic and static security assessment f pwer systems using artificial neural netwrks", Prceedings f the NSF Wrkshp n Applicatins

More information

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter Midwest Big Data Summer Schl: Machine Learning I: Intrductin Kris De Brabanter kbrabant@iastate.edu Iwa State University Department f Statistics Department f Cmputer Science June 24, 2016 1/24 Outline

More information

Agenda. What is Machine Learning? Learning Type of Learning: Supervised, Unsupervised and semi supervised Classification

Agenda. What is Machine Learning? Learning Type of Learning: Supervised, Unsupervised and semi supervised Classification Agenda Artificial Intelligence and its applicatins Lecture 6 Supervised Learning Prfessr Daniel Yeung danyeung@ieee.rg Dr. Patrick Chan patrickchan@ieee.rg Suth China University f Technlgy, China Learning

More information

The blessing of dimensionality for kernel methods

The blessing of dimensionality for kernel methods fr kernel methds Building classifiers in high dimensinal space Pierre Dupnt Pierre.Dupnt@ucluvain.be Classifiers define decisin surfaces in sme feature space where the data is either initially represented

More information

In SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw:

In SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw: In SMV I IAML: Supprt Vectr Machines II Nigel Gddard Schl f Infrmatics Semester 1 We sa: Ma margin trick Gemetry f the margin and h t cmpute it Finding the ma margin hyperplane using a cnstrained ptimizatin

More information

COMP 551 Applied Machine Learning Lecture 4: Linear classification

COMP 551 Applied Machine Learning Lecture 4: Linear classification COMP 551 Applied Machine Learning Lecture 4: Linear classificatin Instructr: Jelle Pineau (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted

More information

What is Statistical Learning?

What is Statistical Learning? What is Statistical Learning? Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radi 0 20 40 60 80 100 Newspaper Shwn are Sales vs TV, Radi and Newspaper,

More information

Support-Vector Machines

Support-Vector Machines Supprt-Vectr Machines Intrductin Supprt vectr machine is a linear machine with sme very nice prperties. Haykin chapter 6. See Alpaydin chapter 13 fr similar cntent. Nte: Part f this lecture drew material

More information

Resampling Methods. Chapter 5. Chapter 5 1 / 52

Resampling Methods. Chapter 5. Chapter 5 1 / 52 Resampling Methds Chapter 5 Chapter 5 1 / 52 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and

More information

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came.

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came. MATH 1342 Ch. 24 April 25 and 27, 2013 Page 1 f 5 CHAPTER 24: INFERENCE IN REGRESSION Chapters 4 and 5: Relatinships between tw quantitative variables. Be able t Make a graph (scatterplt) Summarize the

More information

Data Mining: Concepts and Techniques. Classification and Prediction. Chapter February 8, 2007 CSE-4412: Data Mining 1

Data Mining: Concepts and Techniques. Classification and Prediction. Chapter February 8, 2007 CSE-4412: Data Mining 1 Data Mining: Cncepts and Techniques Classificatin and Predictin Chapter 6.4-6 February 8, 2007 CSE-4412: Data Mining 1 Chapter 6 Classificatin and Predictin 1. What is classificatin? What is predictin?

More information

Part 3 Introduction to statistical classification techniques

Part 3 Introduction to statistical classification techniques Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms

More information

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551

More information

Linear Classification

Linear Classification Linear Classificatin CS 54: Machine Learning Slides adapted frm Lee Cper, Jydeep Ghsh, and Sham Kakade Review: Linear Regressin CS 54 [Spring 07] - H Regressin Given an input vectr x T = (x, x,, xp), we

More information

Multiple Source Multiple. using Network Coding

Multiple Source Multiple. using Network Coding Multiple Surce Multiple Destinatin Tplgy Inference using Netwrk Cding Pegah Sattari EECS, UC Irvine Jint wrk with Athina Markpulu, at UCI, Christina Fraguli, at EPFL, Lausanne Outline Netwrk Tmgraphy Gal,

More information

A Scalable Recurrent Neural Network Framework for Model-free

A Scalable Recurrent Neural Network Framework for Model-free A Scalable Recurrent Neural Netwrk Framewrk fr Mdel-free POMDPs April 3, 2007 Zhenzhen Liu, Itamar Elhanany Machine Intelligence Lab Department f Electrical and Cmputer Engineering The University f Tennessee

More information

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression 4th Indian Institute f Astrphysics - PennState Astrstatistics Schl July, 2013 Vainu Bappu Observatry, Kavalur Crrelatin and Regressin Rahul Ry Indian Statistical Institute, Delhi. Crrelatin Cnsider a tw

More information

Fall 2013 Physics 172 Recitation 3 Momentum and Springs

Fall 2013 Physics 172 Recitation 3 Momentum and Springs Fall 03 Physics 7 Recitatin 3 Mmentum and Springs Purpse: The purpse f this recitatin is t give yu experience wrking with mmentum and the mmentum update frmula. Readings: Chapter.3-.5 Learning Objectives:.3.

More information

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9. Sectin 7 Mdel Assessment This sectin is based n Stck and Watsn s Chapter 9. Internal vs. external validity Internal validity refers t whether the analysis is valid fr the ppulatin and sample being studied.

More information

Tree Structured Classifier

Tree Structured Classifier Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients

More information

STATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours

STATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours STATS216v Intrductin t Statistical Learning Stanfrd University, Summer 2016 Practice Final (Slutins) Duratin: 3 hurs Instructins: (This is a practice final and will nt be graded.) Remember the university

More information

Enhancing Performance of MLP/RBF Neural Classifiers via an Multivariate Data Distribution Scheme

Enhancing Performance of MLP/RBF Neural Classifiers via an Multivariate Data Distribution Scheme Enhancing Perfrmance f / Neural Classifiers via an Multivariate Data Distributin Scheme Halis Altun, Gökhan Gelen Nigde University, Electrical and Electrnics Engineering Department Nigde, Turkey haltun@nigde.edu.tr

More information

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa There are tw parts t this lab. The first is intended t demnstrate hw t request and interpret the spatial diagnstics f a standard OLS regressin mdel using GeDa. The diagnstics prvide infrmatin abut the

More information

Checking the resolved resonance region in EXFOR database

Checking the resolved resonance region in EXFOR database Checking the reslved resnance regin in EXFOR database Gttfried Bertn Sciété de Calcul Mathématique (SCM) Oscar Cabells OECD/NEA Data Bank JEFF Meetings - Sessin JEFF Experiments Nvember 0-4, 017 Bulgne-Billancurt,

More information

The Kullback-Leibler Kernel as a Framework for Discriminant and Localized Representations for Visual Recognition

The Kullback-Leibler Kernel as a Framework for Discriminant and Localized Representations for Visual Recognition The Kullback-Leibler Kernel as a Framewrk fr Discriminant and Lcalized Representatins fr Visual Recgnitin Nun Vascncels Purdy H Pedr Mren ECE Department University f Califrnia, San Dieg HP Labs Cambridge

More information

7 TH GRADE MATH STANDARDS

7 TH GRADE MATH STANDARDS ALGEBRA STANDARDS Gal 1: Students will use the language f algebra t explre, describe, represent, and analyze number expressins and relatins 7 TH GRADE MATH STANDARDS 7.M.1.1: (Cmprehensin) Select, use,

More information

COMP9444 Neural Networks and Deep Learning 3. Backpropagation

COMP9444 Neural Networks and Deep Learning 3. Backpropagation COMP9444 Neural Netwrks and Deep Learning 3. Backprpagatin Tetbk, Sectins 4.3, 5.2, 6.5.2 COMP9444 17s2 Backprpagatin 1 Outline Supervised Learning Ockham s Razr (5.2) Multi-Layer Netwrks Gradient Descent

More information

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical mdel fr micrarray data analysis David Rssell Department f Bistatistics M.D. Andersn Cancer Center, Hustn, TX 77030, USA rsselldavid@gmail.cm

More information

Artificial Neural Networks MLP, Backpropagation

Artificial Neural Networks MLP, Backpropagation Artificial Neural Netwrks MLP, Backprpagatin 01001110 01100101 01110101 01110010 01101111 01101110 01101111 01110110 01100001 00100000 01110011 01101011 01110101 01110000 01101001 01101110 01100001 00100000

More information

Early detection of mining truck failure by modelling its operation with neural networks classification algorithms

Early detection of mining truck failure by modelling its operation with neural networks classification algorithms RU, Rand GOLOSINSKI, T.S. Early detectin f mining truck failure by mdelling its peratin with neural netwrks classificatin algrithms. Applicatin f Cmputers and Operatins Research ill the Minerals Industries,

More information

Computational modeling techniques

Computational modeling techniques Cmputatinal mdeling techniques Lecture 4: Mdel checing fr ODE mdels In Petre Department f IT, Åb Aademi http://www.users.ab.fi/ipetre/cmpmd/ Cntent Stichimetric matrix Calculating the mass cnservatin relatins

More information

Turing Machines. Human-aware Robotics. 2017/10/17 & 19 Chapter 3.2 & 3.3 in Sipser Ø Announcement:

Turing Machines. Human-aware Robotics. 2017/10/17 & 19 Chapter 3.2 & 3.3 in Sipser Ø Announcement: Turing Machines Human-aware Rbtics 2017/10/17 & 19 Chapter 3.2 & 3.3 in Sipser Ø Annuncement: q q q q Slides fr this lecture are here: http://www.public.asu.edu/~yzhan442/teaching/cse355/lectures/tm-ii.pdf

More information

5 th grade Common Core Standards

5 th grade Common Core Standards 5 th grade Cmmn Cre Standards In Grade 5, instructinal time shuld fcus n three critical areas: (1) develping fluency with additin and subtractin f fractins, and develping understanding f the multiplicatin

More information

Stats Classification Ji Zhu, Michigan Statistics 1. Classification. Ji Zhu 445C West Hall

Stats Classification Ji Zhu, Michigan Statistics 1. Classification. Ji Zhu 445C West Hall Stats 415 - Classificatin Ji Zhu, Michigan Statistics 1 Classificatin Ji Zhu 445C West Hall 734-936-2577 jizhu@umich.edu Stats 415 - Classificatin Ji Zhu, Michigan Statistics 2 Examples f Classificatin

More information

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007 CS 477/677 Analysis f Algrithms Fall 2007 Dr. Gerge Bebis Curse Prject Due Date: 11/29/2007 Part1: Cmparisn f Srting Algrithms (70% f the prject grade) The bjective f the first part f the assignment is

More information

Math Foundations 20 Work Plan

Math Foundations 20 Work Plan Math Fundatins 20 Wrk Plan Units / Tpics 20.8 Demnstrate understanding f systems f linear inequalities in tw variables. Time Frame December 1-3 weeks 6-10 Majr Learning Indicatrs Identify situatins relevant

More information

Linear programming III

Linear programming III Linear prgramming III Review 1/33 What have cvered in previus tw classes LP prblem setup: linear bjective functin, linear cnstraints. exist extreme pint ptimal slutin. Simplex methd: g thrugh extreme pint

More information

Image Processing 1 (IP1) Bildverarbeitung 1

Image Processing 1 (IP1) Bildverarbeitung 1 MIN-Fakultät Fachbereich Infrmatik Arbeitsbereich SAV/BV (KOGS) Image Prcessing 1 (IP1) Bildverarbeitung 1 Lecture 15 Pa;ern Recgni=n Winter Semester 2014/15 Dr. Benjamin Seppke Prf. Siegfried S=ehl What

More information

Differentiation Applications 1: Related Rates

Differentiation Applications 1: Related Rates Differentiatin Applicatins 1: Related Rates 151 Differentiatin Applicatins 1: Related Rates Mdel 1: Sliding Ladder 10 ladder y 10 ladder 10 ladder A 10 ft ladder is leaning against a wall when the bttm

More information

The standards are taught in the following sequence.

The standards are taught in the following sequence. B L U E V A L L E Y D I S T R I C T C U R R I C U L U M MATHEMATICS Third Grade In grade 3, instructinal time shuld fcus n fur critical areas: (1) develping understanding f multiplicatin and divisin and

More information

Administrativia. Assignment 1 due thursday 9/23/2004 BEFORE midnight. Midterm exam 10/07/2003 in class. CS 460, Sessions 8-9 1

Administrativia. Assignment 1 due thursday 9/23/2004 BEFORE midnight. Midterm exam 10/07/2003 in class. CS 460, Sessions 8-9 1 Administrativia Assignment 1 due thursday 9/23/2004 BEFORE midnight Midterm eam 10/07/2003 in class CS 460, Sessins 8-9 1 Last time: search strategies Uninfrmed: Use nly infrmatin available in the prblem

More information

T Algorithmic methods for data mining. Slide set 6: dimensionality reduction

T Algorithmic methods for data mining. Slide set 6: dimensionality reduction T-61.5060 Algrithmic methds fr data mining Slide set 6: dimensinality reductin reading assignment LRU bk: 11.1 11.3 PCA tutrial in mycurses (ptinal) ptinal: An Elementary Prf f a Therem f Jhnsn and Lindenstrauss,

More information

Reinforcement Learning" CMPSCI 383 Nov 29, 2011!

Reinforcement Learning CMPSCI 383 Nov 29, 2011! Reinfrcement Learning" CMPSCI 383 Nv 29, 2011! 1 Tdayʼs lecture" Review f Chapter 17: Making Cmple Decisins! Sequential decisin prblems! The mtivatin and advantages f reinfrcement learning.! Passive learning!

More information

CS 109 Lecture 23 May 18th, 2016

CS 109 Lecture 23 May 18th, 2016 CS 109 Lecture 23 May 18th, 2016 New Datasets Heart Ancestry Netflix Our Path Parameter Estimatin Machine Learning: Frmally Many different frms f Machine Learning We fcus n the prblem f predictin Want

More information

Biplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 13 Offprint

Biplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 13 Offprint Biplts in Practice MICHAEL GREENACRE Prfessr f Statistics at the Pmpeu Fabra University Chapter 13 Offprint CASE STUDY BIOMEDICINE Cmparing Cancer Types Accrding t Gene Epressin Arrays First published:

More information

Lab 1 The Scientific Method

Lab 1 The Scientific Method INTRODUCTION The fllwing labratry exercise is designed t give yu, the student, an pprtunity t explre unknwn systems, r universes, and hypthesize pssible rules which may gvern the behavir within them. Scientific

More information

A Matrix Representation of Panel Data

A Matrix Representation of Panel Data web Extensin 6 Appendix 6.A A Matrix Representatin f Panel Data Panel data mdels cme in tw brad varieties, distinct intercept DGPs and errr cmpnent DGPs. his appendix presents matrix algebra representatins

More information

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur Cdewrd Distributin fr Frequency Sensitive Cmpetitive Learning with One Dimensinal Input Data Aristides S. Galanpuls and Stanley C. Ahalt Department f Electrical Engineering The Ohi State University Abstract

More information

Name: Block: Date: Science 10: The Great Geyser Experiment A controlled experiment

Name: Block: Date: Science 10: The Great Geyser Experiment A controlled experiment Science 10: The Great Geyser Experiment A cntrlled experiment Yu will prduce a GEYSER by drpping Ments int a bttle f diet pp Sme questins t think abut are: What are yu ging t test? What are yu ging t measure?

More information

, which yields. where z1. and z2

, which yields. where z1. and z2 The Gaussian r Nrmal PDF, Page 1 The Gaussian r Nrmal Prbability Density Functin Authr: Jhn M Cimbala, Penn State University Latest revisin: 11 September 13 The Gaussian r Nrmal Prbability Density Functin

More information

Time, Synchronization, and Wireless Sensor Networks

Time, Synchronization, and Wireless Sensor Networks Time, Synchrnizatin, and Wireless Sensr Netwrks Part II Ted Herman University f Iwa Ted Herman/March 2005 1 Presentatin: Part II metrics and techniques single-hp beacns reginal time znes ruting-structure

More information

Churn Prediction using Dynamic RFM-Augmented node2vec

Churn Prediction using Dynamic RFM-Augmented node2vec Churn Predictin using Dynamic RFM-Augmented nde2vec Sandra Mitrvić, Jchen de Weerdt, Bart Baesens & Wilfried Lemahieu Department f Decisin Sciences and Infrmatin Management, KU Leuven 18 September 2017,

More information

Lecture 10, Principal Component Analysis

Lecture 10, Principal Component Analysis Principal Cmpnent Analysis Lecture 10, Principal Cmpnent Analysis Ha Helen Zhang Fall 2017 Ha Helen Zhang Lecture 10, Principal Cmpnent Analysis 1 / 16 Principal Cmpnent Analysis Lecture 10, Principal

More information

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Emanuela Galass The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Emanuela Galass fr the purpse f this wrkshp When can we use

More information

Elements of Machine Intelligence - I

Elements of Machine Intelligence - I ECE-175A Elements f Machine Intelligence - I Ken Kreutz-Delgad Nun Vascncels ECE Department, UCSD Winter 2011 The curse The curse will cver basic, but imprtant, aspects f machine learning and pattern recgnitin

More information

AP Statistics Practice Test Unit Three Exploring Relationships Between Variables. Name Period Date

AP Statistics Practice Test Unit Three Exploring Relationships Between Variables. Name Period Date AP Statistics Practice Test Unit Three Explring Relatinships Between Variables Name Perid Date True r False: 1. Crrelatin and regressin require explanatry and respnse variables. 1. 2. Every least squares

More information

Video Encoder Control

Video Encoder Control Vide Encder Cntrl Thmas Wiegand Digital Image Cmmunicatin 1 / 41 Outline Intrductin Encder Cntrl using Lagrange multipliers Lagrangian ptimizatin Lagrangian bit allcatin Lagrangian Optimizatin in Hybrid

More information

Department of Electrical Engineering, University of Waterloo. Introduction

Department of Electrical Engineering, University of Waterloo. Introduction Sectin 4: Sequential Circuits Majr Tpics Types f sequential circuits Flip-flps Analysis f clcked sequential circuits Mre and Mealy machines Design f clcked sequential circuits State transitin design methd

More information

Building to Transformations on Coordinate Axis Grade 5: Geometry Graph points on the coordinate plane to solve real-world and mathematical problems.

Building to Transformations on Coordinate Axis Grade 5: Geometry Graph points on the coordinate plane to solve real-world and mathematical problems. Building t Transfrmatins n Crdinate Axis Grade 5: Gemetry Graph pints n the crdinate plane t slve real-wrld and mathematical prblems. 5.G.1. Use a pair f perpendicular number lines, called axes, t define

More information

Physical Layer: Outline

Physical Layer: Outline 18-: Intrductin t Telecmmunicatin Netwrks Lectures : Physical Layer Peter Steenkiste Spring 01 www.cs.cmu.edu/~prs/nets-ece Physical Layer: Outline Digital Representatin f Infrmatin Characterizatin f Cmmunicatin

More information

Admin. MDP Search Trees. Optimal Quantities. Reinforcement Learning

Admin. MDP Search Trees. Optimal Quantities. Reinforcement Learning Admin Reinfrcement Learning Cntent adapted frm Berkeley CS188 MDP Search Trees Each MDP state prjects an expectimax-like search tree Optimal Quantities The value (utility) f a state s: V*(s) = expected

More information

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards:

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards: MODULE FOUR This mdule addresses functins SC Academic Standards: EA-3.1 Classify a relatinship as being either a functin r nt a functin when given data as a table, set f rdered pairs, r graph. EA-3.2 Use

More information

ENSC Discrete Time Systems. Project Outline. Semester

ENSC Discrete Time Systems. Project Outline. Semester ENSC 49 - iscrete Time Systems Prject Outline Semester 006-1. Objectives The gal f the prject is t design a channel fading simulatr. Upn successful cmpletin f the prject, yu will reinfrce yur understanding

More information

SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES

SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES 1 SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES Wlfgang HÄRDLE Ruslan MORO Center fr Applied Statistics and Ecnmics (CASE), Humbldt-Universität zu Berlin Mtivatin 2 Applicatins in Medicine estimatin f

More information

CC283 Intelligent Problem Solving 28/10/2013

CC283 Intelligent Problem Solving 28/10/2013 Machine Learning What is the research agenda? How to measure success? How to learn? Machine Learning Overview Unsupervised Learning Supervised Learning Training Testing Unseen data Data Observed x 1 x

More information

Lecture 8: Multiclass Classification (I)

Lecture 8: Multiclass Classification (I) Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Lecture 8: Multiclass Classificatin (I) Ha Helen Zhang Fall 07 Ha Helen Zhang Lecture 8: Multiclass Classificatin

More information

Module 3: Gaussian Process Parameter Estimation, Prediction Uncertainty, and Diagnostics

Module 3: Gaussian Process Parameter Estimation, Prediction Uncertainty, and Diagnostics Mdule 3: Gaussian Prcess Parameter Estimatin, Predictin Uncertainty, and Diagnstics Jerme Sacks and William J Welch Natinal Institute f Statistical Sciences and University f British Clumbia Adapted frm

More information

Five Whys How To Do It Better

Five Whys How To Do It Better Five Whys Definitin. As explained in the previus article, we define rt cause as simply the uncvering f hw the current prblem came int being. Fr a simple causal chain, it is the entire chain. Fr a cmplex

More information

MACHINE LEARNING FOR CLUSTER- GALAXY CLASSIFICATION

MACHINE LEARNING FOR CLUSTER- GALAXY CLASSIFICATION MACHINE LEARNING FOR CLUSTER- GALAXY CLASSIFICATION Silvia de Castr García Directres: Dr. Ricard Pérez Martínez, Dra. Ana María Pérez García 16/03/2018 Machine Learning fr cluster-galaxy classificatin

More information

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS 1 Influential bservatins are bservatins whse presence in the data can have a distrting effect n the parameter estimates and pssibly the entire analysis,

More information

You need to be able to define the following terms and answer basic questions about them:

You need to be able to define the following terms and answer basic questions about them: CS440/ECE448 Sectin Q Fall 2017 Midterm Review Yu need t be able t define the fllwing terms and answer basic questins abut them: Intr t AI, agents and envirnments Pssible definitins f AI, prs and cns f

More information

NAME: Prof. Ruiz. 1. [5 points] What is the difference between simple random sampling and stratified random sampling?

NAME: Prof. Ruiz. 1. [5 points] What is the difference between simple random sampling and stratified random sampling? CS4445 ata Mining and Kwledge iscery in atabases. B Term 2014 Exam 1 Nember 24, 2014 Prf. Carlina Ruiz epartment f Cmputer Science Wrcester Plytechnic Institute NAME: Prf. Ruiz Prblem I: Prblem II: Prblem

More information

Lecture 02 CSE 40547/60547 Computing at the Nanoscale

Lecture 02 CSE 40547/60547 Computing at the Nanoscale PN Junctin Ntes: Lecture 02 CSE 40547/60547 Cmputing at the Nanscale Letʼs start with a (very) shrt review f semi-cnducting materials: - N-type material: Obtained by adding impurity with 5 valence elements

More information

Chapter 11: Neural Networks

Chapter 11: Neural Networks Chapter 11: Neural Netwrks DD3364 December 16, 2012 Prjectin Pursuit Regressin Prjectin Pursuit Regressin mdel: Prjectin Pursuit Regressin f(x) = M g m (wmx) t i=1 where X R p and have targets Y R. Additive

More information

Support Vector Machines and Flexible Discriminants

Support Vector Machines and Flexible Discriminants 12 Supprt Vectr Machines and Flexible Discriminants This is page 417 Printer: Opaque this 12.1 Intrductin In this chapter we describe generalizatins f linear decisin bundaries fr classificatin. Optimal

More information

MATHEMATICS SYLLABUS SECONDARY 5th YEAR

MATHEMATICS SYLLABUS SECONDARY 5th YEAR Eurpean Schls Office f the Secretary-General Pedaggical Develpment Unit Ref. : 011-01-D-8-en- Orig. : EN MATHEMATICS SYLLABUS SECONDARY 5th YEAR 6 perid/week curse APPROVED BY THE JOINT TEACHING COMMITTEE

More information

CHM112 Lab Graphing with Excel Grading Rubric

CHM112 Lab Graphing with Excel Grading Rubric Name CHM112 Lab Graphing with Excel Grading Rubric Criteria Pints pssible Pints earned Graphs crrectly pltted and adhere t all guidelines (including descriptive title, prperly frmatted axes, trendline

More information

Experiment #3. Graphing with Excel

Experiment #3. Graphing with Excel Experiment #3. Graphing with Excel Study the "Graphing with Excel" instructins that have been prvided. Additinal help with learning t use Excel can be fund n several web sites, including http://www.ncsu.edu/labwrite/res/gt/gt-

More information

Hypothesis Tests for One Population Mean

Hypothesis Tests for One Population Mean Hypthesis Tests fr One Ppulatin Mean Chapter 9 Ala Abdelbaki Objective Objective: T estimate the value f ne ppulatin mean Inferential statistics using statistics in rder t estimate parameters We will be

More information

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Céline Ferré The Wrld Bank When can we use matching? What if the assignment t the treatment is nt dne randmly r based n an eligibility index, but n the basis

More information

Chapter 15 & 16: Random Forests & Ensemble Learning

Chapter 15 & 16: Random Forests & Ensemble Learning Chapter 15 & 16: Randm Frests & Ensemble Learning DD3364 Nvember 27, 2012 Ty Prblem fr Bsted Tree Bsted Tree Example Estimate this functin with a sum f trees with 9-terminal ndes by minimizing the sum

More information

Inference in the Multiple-Regression

Inference in the Multiple-Regression Sectin 5 Mdel Inference in the Multiple-Regressin Kinds f hypthesis tests in a multiple regressin There are several distinct kinds f hypthesis tests we can run in a multiple regressin. Suppse that amng

More information

The general linear model and Statistical Parametric Mapping I: Introduction to the GLM

The general linear model and Statistical Parametric Mapping I: Introduction to the GLM The general linear mdel and Statistical Parametric Mapping I: Intrductin t the GLM Alexa Mrcm and Stefan Kiebel, Rik Hensn, Andrew Hlmes & J-B J Pline Overview Intrductin Essential cncepts Mdelling Design

More information

Dataflow Analysis and Abstract Interpretation

Dataflow Analysis and Abstract Interpretation Dataflw Analysis and Abstract Interpretatin Cmputer Science and Artificial Intelligence Labratry MIT Nvember 9, 2015 Recap Last time we develped frm first principles an algrithm t derive invariants. Key

More information

Determining the Accuracy of Modal Parameter Estimation Methods

Determining the Accuracy of Modal Parameter Estimation Methods Determining the Accuracy f Mdal Parameter Estimatin Methds by Michael Lee Ph.D., P.E. & Mar Richardsn Ph.D. Structural Measurement Systems Milpitas, CA Abstract The mst cmmn type f mdal testing system

More information

AP Statistics Notes Unit Two: The Normal Distributions

AP Statistics Notes Unit Two: The Normal Distributions AP Statistics Ntes Unit Tw: The Nrmal Distributins Syllabus Objectives: 1.5 The student will summarize distributins f data measuring the psitin using quartiles, percentiles, and standardized scres (z-scres).

More information

INTRODUCTION TO MACHINE LEARNING FOR MEDICINE

INTRODUCTION TO MACHINE LEARNING FOR MEDICINE Fall 2017 INTRODUCTION TO MACHINE LEARNING FOR MEDICINE Carla E. Brdley Prfessr & Dean Cllege f Cmputer and Infrmatin Science Nrtheastern University WHAT IS MACHINE LEARNING/DATA MINING? Figure is frm

More information

Emphases in Common Core Standards for Mathematical Content Kindergarten High School

Emphases in Common Core Standards for Mathematical Content Kindergarten High School Emphases in Cmmn Cre Standards fr Mathematical Cntent Kindergarten High Schl Cntent Emphases by Cluster March 12, 2012 Describes cntent emphases in the standards at the cluster level fr each grade. These

More information