Data Mining: Concepts and Techniques. Classification and Prediction. Chapter February 8, 2007 CSE-4412: Data Mining 1

Size: px
Start display at page:

Download "Data Mining: Concepts and Techniques. Classification and Prediction. Chapter February 8, 2007 CSE-4412: Data Mining 1"

Transcription

1 Data Mining: Cncepts and Techniques Classificatin and Predictin Chapter February 8, 2007 CSE-4412: Data Mining 1 Chapter 6 Classificatin and Predictin 1. What is classificatin? What is predictin? 2. Issues regarding classificatin and predictin 3. Classificatin by decisin tree inductin 4. Bayesian classificatin 5. Rule-based classificatin 6. Classificatin by back prpagatin 7. Supprt Vectr Machines (SVM) 8. Summary February 8, 2007 CSE-4412: Data Mining 2 1

2 Basic Idea (Again) Use ld tuples with knwn classes t classify new tuples with unknwn classes. E.g., tuples: custmers ld tuples: previus and current custmers new tuples: prspective custmers questin: Is custmer a gd credit risk? classes (answers): gd, fair, pr Why nt ust use the class prir prbabilities ver the ld tuples? Yes, why nt? February 8, 2007 CSE-4412: Data Mining 3 Use the Attributes Okay, the tuples have attributes. Use the attribute values t d better classificatin. Idea: Given a new tuple (e.g., <25, $72k, student>), use ust the ld tuples that match exactly t decide. Wuld this wrk? What are the prblems with this apprach? Still a gd idea. Hw can we fix this apprach? February 8, 2007 CSE-4412: Data Mining 4 2

3 Bayesian Classificatin A statistical classifier Perfrms prbabilistic predictin; i.e., predicts class membership prbabilities. Based n Bayes Therem. Perfrmance A simple Bayesian classifier, naïve Bayesian classifier, has cmparable perfrmance with decisin tree and selected neural netwrk classifiers. Incremental Each training example can incrementally increase / decrease the prbability that a hypthesis is crrect. Prir knwledge can be cmbined with bserved data. Standard Can be cmputatinally intractable, but Prvide a standard f ptimal decisin making, against which ther methds can be measured. February 8, 2007 CSE-4412: Data Mining 5 Bayes Therem Basics Let X be a data sample ( evidence ): class label is unknwn. Let H be a hypthesis that X belngs t class C. Classificatin is t determine P(H X), the prbability that the hypthesis hlds given the bserved data sample X. P(H) (prir prbability), the initial prbability. E.g., X will buy cmputer, regardless f age, incme, P(X): prbability that sample data is bserved. P(X H) (psteriri prbability), the prbability f bserving the sample X, given that the hypthesis hlds. E.g., Given that X will buy cmputer, the prbability that X is age , incme = medium,... February 8, 2007 CSE-4412: Data Mining 6 3

4 Bayes Therem Given training data X, psteriri prbability f a hypthesis H, P(H X), fllws the Bayes therem: P ( H X) = P( X H) P( H) P( X) Infrmally, this can be written as psteriri likelihd prir / evidence Predicts X belngs t C i iff the prbability P(C i X) is the highest amng all the P(C k X) fr all the k classes. Practical difficulty: Require initial knwledge f many prbabilities, significant cmputatinal cst. February 8, 2007 CSE-4412: Data Mining 7 Twards a Naïve Bayesian Classifier Let D be a training set f tuples and their assciated class labels, and each tuple is represented by an n-d attribute vectr X = (x 1, x 2,, x n ). Suppse there are k classes C 1, C 2,, C k. Classificatin is t derive the maximum psteriri; i.e., the maximal P(C i X). This can be derived frm Bayes Therem: P( X C ) P( C ) P ( C X) = i i i P( X) Since P(X) is cnstant fr all classes, nly P( C X ) = P( X C ) P( C ) i i i needs t be maximized. February 8, 2007 CSE-4412: Data Mining 8 4

5 Derivatin f Naïve Bayes Classifier A simplified assumptin: attributes are cnditinally independent (i.e., n dependence relatin between attributes): n P( X Ci) = " P( x Ci) = P( x ) ( )... ( ) 1 Ci! P x 2 Ci!! P x k n Ci k = 1 This greatly reduces the cmputatinal cst: Only cunts the class distributin. If A k is categrical, P(x k C i ) is the # f tuples in C i having value x k fr A k divided by C i, D (# f tuples f C i in D). If A k is cntinuus-valued, P(x k C i ) is usually cmputed based n Gaussian distributin with a mean μ and standard deviatin σ and P(x C i ) is 2 ( x# µ ) 1 # 2 2! g( x, µ,! ) = e 2"! P(x k C i ) = g(x k, µ ci, σ ci ) February 8, 2007 CSE-4412: Data Mining 9 Naïve Bayesian Classifier: Training Dataset Class: C1:buys_cmputer = yes C2:buys_cmputer = n Data sample X = (age <=30, Incme = medium, Student = yes Credit_rating = Fair) age incme studentcredit_rating_cmp <=30 high n fair n <=30 high n excellent n high n fair yes >40 medium n fair yes >40 lw yes fair yes >40 lw yes excellent n lw yes excellent yes <=30 medium n fair n <=30 lw yes fair yes >40 medium yes fair yes <=30 medium yes excellent yes medium n excellent yes high yes fair yes >40 medium n excellent n February 8, 2007 CSE-4412: Data Mining 10 5

6 Naïve Bayesian Classifier An Example P(C i ): P(buys_cmputer = yes ) = 9/14 = P(buys_cmputer = n ) = 5/14= Cmpute P(X C i ) fr each class P(age = <=30 buys_cmputer = yes ) = 2/9 = P(age = <= 30 buys_cmputer = n ) = 3/5 = 0.6 P(incme = medium buys_cmputer = yes ) = 4/9 = P(incme = medium buys_cmputer = n ) = 2/5 = 0.4 P(student = yes buys_cmputer = yes) = 6/9 = P(student = yes buys_cmputer = n ) = 1/5 = 0.2 P(credit_rating = fair buys_cmputer = yes ) = 6/9 = P(credit_rating = fair buys_cmputer = n ) = 2/5 = 0.4 X = (age <= 30, incme = medium, student = yes, credit_rating = fair) P(X C i ) : P(X buys_cmputer = yes ) = x x x = P(X buys_cmputer = n ) = 0.6 x 0.4 x 0.2 x 0.4 = P(X C i )*P(C i ) : P(X buys_cmputer = yes ) * P(buys_cmputer = yes ) = P(X buys_cmputer = n ) * P(buys_cmputer = n ) = Therefre, X belngs t class ( buys_cmputer = yes ) February 8, 2007 CSE-4412: Data Mining 11 The Zer-Prbability Prblem Naïve Bayesian predictin requires each cnditinal prbability be nn-zer. Otherwise, the predicted prbability will be zer! P( X Ci) = n! P( xk Ci) k = 1 E.g., Suppse a dataset with 1000 tuples, incme=lw (0), incme= medium (990), and incme = high (10). Use Laplacian crrectin (r Laplacian estimatr): Add 1 t each case. E.g., Prb(incme = lw) = 1 / 1003 Prb(incme = medium) = 991 / 1003 Prb(incme = high) = 11 / 1003 The crrected prbability estimates are clse t their uncrrected cunterparts. February 8, 2007 CSE-4412: Data Mining 12 6

7 Naïve Bayesian Classifier Evaluatin Advantages Easy t implement and maintain. Easy t update incrementally with new training tuples. Gd results btained in many dmains. N issues with verfitting the mdel. Reasnably immune t nise. Can wrk with missing values in data, bth training and fr classifying. Nise in data (incrrect values) get balanced ut (t sme extent). February 8, 2007 CSE-4412: Data Mining 13 Naïve Bayesian Classifier Evaluatin Disadvantages Assumptin: Class cnditinal independence, therefre lss f accuracy. In practice, hwever, dependencies d exist amng variables. E.g., hspital patients prfiles: age, family histry, etc. Symptms: fever, cugh etc. Disease: lung cancer, diabetes, etc. Dependencies amng these cannt be mdeled by a Naïve Bayesian Classifier. Blackbx: Cannt interpret the mdel. Hw t deal with these dependencies? Bayesian Belief Netwrks February 8, 2007 CSE-4412: Data Mining 14 7

8 Bayesian Belief Netwrks Bayesian belief netwrk allws a subset f the variables cnditinally independent. A graphical mdel f causal relatinships Represents dependencies amng the variables. Gives a specificatin f int prbability distributin. Ndes: randm variables X Z Y P Links: dependency X and Y are the parents f Z, and Y is the parent f P N dependency between Z and P Has n lps r cycles February 8, 2007 CSE-4412: Data Mining 15 Bayesian Belief Netwrk An Example Family Histry Smker The cnditinal prbability table (CPT) fr variable LungCancer: LC (FH, S) 0.8 (FH, ~S) (~FH, S) (~FH, ~S) LungCancer Emphysema ~LC CPT shws the cnditinal prbability fr each pssible cmbinatin f its parents PsitiveXRay Dyspnea Bayesian Belief Netwrk Derivatin f the prbability f a particular cmbinatin f values f X, frm CPT: P( x x = n 1,..., n)! P( xi Parents( Y i)) i = 1 February 8, 2007 CSE-4412: Data Mining 16 8

9 Training Bayesian Netwrks Several scenaris: Given bth the netwrk structure (knwn) and all variables bservable: learn nly the CPTs. Netwrk structure knwn, sme hidden variables: gradient descent (greedy hill-climbing) methd, analgus t neural netwrk learning. Netwrk structure unknwn, all variables bservable: search thrugh the mdel space t recnstruct netwrk tplgy. Unknwn structure, hidden variables: N gd algrithms knwn fr this purpse! Ref.: D. Heckerman: Bayesian netwrks fr data mining February 8, 2007 CSE-4412: Data Mining 17 Chapter 6 Classificatin and Predictin 1. What is classificatin? What is predictin? 2. Issues regarding classificatin and predictin 3. Classificatin by decisin tree inductin 4. Bayesian classificatin 5. Rule-based classificatin 6. Classificatin by back prpagatin 7. Supprt Vectr Machines (SVM) 8. Summary February 8, 2007 CSE-4412: Data Mining 18 9

10 Using IF-THEN Rules fr Classificatin Represent the knwledge in the frm f IF-THEN rules R: IF age = yuth AND student = yes THEN buys_cmputer = yes Rule antecedent/precnditin vs. rule cnsequent Assessment f a rule: cverage and accuracy n cvers = # f tuples cvered by R n crrect = # f tuples crrectly classified by R cverage(r) = n cvers / D /* D: training data set */ accuracy(r) = n crrect / n cvers If mre than ne rule is triggered, need cnflict reslutin. size rdering: Assign the highest pririty t the triggering rules that has the tughest requirement (i.e., with the mst attribute test). class-based rdering: Decreasing rder f prevalence r misclassificatin cst per class. rule-based rdering (decisin list): Rules are rganized int ne lng pririty list, accrding t sme measure f rule quality r by experts. February 8, 2007 CSE-4412: Data Mining 19 Rule Extractin frm a Decisin Tree Rules are easier t understand than large trees. One rule is created fr each path frm the rt t a leaf. Each attribute-value pair alng a path frms a cnunctin: the leaf hlds the class predictin. Rules are mutually exclusive and exhaustive. n student? Example: Rule extractin frm ur buys_cmputer decisin-tree: IF age = yung AND student = n IF age = yung AND student = yes IF age = mid-age age? <= >40 yes yes n credit rating? excellent fair n yes yes THEN buys_cmputer = n THEN buys_cmputer = yes THEN buys_cmputer = yes IF age = ld AND credit_rating = excellent THEN buys_cmputer = yes IF age = yung AND credit_rating = fair THEN buys_cmputer = n February 8, 2007 CSE-4412: Data Mining 20 10

11 Rule Extractin frm the Training Data Sequential cvering algrithm: Extracts rules directly frm training data. Typical sequential cvering algrithms: FOIL, AQ, CN2, RIPPER. Rules are learned sequentially, each fr a given class C i will cver many tuples f C i but nne (r few) f the tuples f ther classes. Steps: Rules are learned ne at a time. Each time a rule is learned, the tuples cvered by the rules are remved. The prcess repeats n the remaining tuples unless terminatin cnditin, e.g., when n mre training examples r when the quality f a rule returned is belw a user-specified threshld. Cmparisn w/ decisin-tree inductin: Learning a set f rules simultaneusly. February 8, 2007 CSE-4412: Data Mining 21 Learn-One-Rule Start with the mst general rule pssible: cnditin = empty. Add new attributes by adpting a greedy depth-first strategy. Pick the ne that mst imprves the rule quality. Rule-Quality measures: cnsider bth cverage and accuracy. Fil-gain (in FOIL & RIPPER): assesses inf_gain by extending cnditin. ps' ps FOIL _ Gain = ps' "(lg2! lg2 ) ps' + neg' ps + neg It favrs rules that have high accuracy and cver many psitive tuples Rule pruning based n an independent set f test tuples. ps! neg FOIL _ Prune( R) = ps + neg Ps/neg are # f psitive/negative tuples cvered by R. If FOIL_Prune is higher fr the pruned versin f R, prune R. February 8, 2007 CSE-4412: Data Mining 22 11

12 Chapter 6 Classificatin and Predictin 1. What is classificatin? What is predictin? 2. Issues regarding classificatin and predictin 3. Classificatin by decisin tree inductin 4. Bayesian classificatin 5. Rule-based classificatin 6. Classificatin by back prpagatin 7. Supprt Vectr Machines (SVM) 8. Summary February 8, 2007 CSE-4412: Data Mining 23 Classificatin as a Mathematical Mapping Classificatin: Predicts categrical class labels. E.g., Persnal hmepage classificatin. = (x 1, x 2, x 3, ), y i = +1 r 1 x 1 : # f the wrd hmepage x 2 : # f the wrd welcme Mathematically X = R n, y Y = {+1, 1} We want a functin f: X Y February 8, 2007 CSE-4412: Data Mining 24 12

13 Linear Classificatin x x x x x x x x x x Binary Classificatin prblem. The data abve the red line belngs t class x. The data belw red line belngs t class Examples: SVM, Perceptrn, Prbabilistic Classifiers February 8, 2007 CSE-4412: Data Mining 25 Discriminative Classifiers Advantages: predictin accuracy is generally high As cmpared t Bayesian methds in general rbust, wrks when training examples cntain errrs fast evaluatin f the learned target functin Bayesian netwrks are nrmally slw. Disadvantages: lng training time difficult t understand the learned functin (weights) Bayesian netwrks can be used easily fr pattern discvery. nt easy t incrprate dmain knwledge Easy in the frm f prirs n the data r distributins February 8, 2007 CSE-4412: Data Mining 26 13

14 Perceptrn & Winnw x 2 Vectr: x, w Scalar: x, y, w Input: {( 1, y 1 ), } Output: classificatin functin f() f( i ) > 0 fr y i = +1 f( i ) < 0 fr y i = -1 f() => + b = 0 r w 1 x 1 +w 2 x 2 +b = 0 x 1 Perceptrn: Update W additively. Winnw: Update W multiplicatively. February 8, 2007 CSE-4412: Data Mining 27 Classificatin by Backprpagatin Nnlinear Neural netwrk: A set f cnnected input/utput units where each cnnectin has a weight assciated with it. During the learning phase, the netwrk learns by adusting the weights s as t be able t predict the crrect class label f the input tuples. Als referred t as cnnectinist learning due t the cnnectins between units. Backprpagatin: A neural netwrk learning algrithm. Started by psychlgists and neurbilgists t develp and test cmputatinal analgues f neurns. February 8, 2007 CSE-4412: Data Mining 28 14

15 Neural Netwrk as a Classifier Advantages: High tlerance t nisy data. Ability t classify untrained patterns. Well-suited fr cntinuus-valued inputs and utputs. Successful n a wide array f real-wrld data. Algrithms are inherently parallel. Disadvantages: Lng training time. Require a number f parameters typically best determined empirically; e.g., the netwrk tplgy r structure. Pr interpretability: Difficult t interpret the symblic meaning. behind the learned weights and f hidden units in the netwrk. February 8, 2007 CSE-4412: Data Mining 29 A Neurn (= a perceptrn) - θ k x 0 w 0 x 1 x n w 1 w n f utput y Input vectr x weight vectr w weighted sum Activatin functin Fr Example n y = sign(! w ix i + µ k ) i= 0 The n-dimensinal input vectr x is mapped int variable y by means f the scalar prduct and a nnlinear functin mapping. February 8, 2007 CSE-4412: Data Mining 30 15

16 Output vectr A Multi-Layer Feed-Frward Neural Netwrk Output layer Hidden layer Input layer Input vectr: X Err = O (1 " O )! Err w Err = O ( 1! O )( T! O ) w i! i 1 O =! I 1 + e I =! w O + " =! w = w + ( l) Err O i i i k + (l) Err i i k k February 8, 2007 CSE-4412: Data Mining 31 Hw des a Multi-Layer Neural Netwrk Wrk? The inputs t the netwrk crrespnd t the attributes measured fr each training tuple. Inputs are fed simultaneusly int the units making up the input layer. They are then weighted and fed simultaneusly t a hidden layer. The number f hidden layers is arbitrary, althugh usually nly ne. The weighted utputs f the last hidden layer are input t units making up the utput layer, which emits the netwrk's predictin. The netwrk is feed-frward in that nne f the weights cycles back t an input unit r t an utput unit f a previus layer. Frm a statistical pint f view, netwrks perfrm nnlinear regressin: Given enugh hidden units and enugh training samples, they can clsely apprximate any functin. February 8, 2007 CSE-4412: Data Mining 32 16

17 Defining a Netwrk Tplgy First decide the netwrk tplgy: # f units in the input layer, # f hidden layers (if > 1), # f units in each hidden layer, and # f units in the utput layer. Nrmalizing the input values fr each attribute measured in the training tuples t [ ]. One input unit per dmain value, each initialized t 0. Output, if fr classificatin and mre than tw classes, ne utput unit per class is used. Once a netwrk has been trained and its accuracy is unacceptable, repeat the training prcess with a different netwrk tplgy r a different set f initial weights. February 8, 2007 CSE-4412: Data Mining 33 Backprpagatin Iteratively prcess a set f training tuples & cmpare the netwrk's predictin with the actual knwn target value. Fr each training tuple, the weights are mdified t minimize the mean squared errr between the netwrk's predictin and the actual target value. Mdificatins are made in the backwards directin: frm the utput layer, thrugh each hidden layer dwn t the first hidden layer, hence backprpagatin. Steps 1. Initialize weights (t small randm #s) and biases in the netwrk. 2. Prpagate the inputs frward (by applying activatin functin). 3. Backprpagate the errr (by updating weights and biases). 4. Terminating cnditin (when errr is very small, etc.). February 8, 2007 CSE-4412: Data Mining 34 17

18 Backprpagatin and Interpretability Efficiency f backprpagatin: Each epch (ne interatin thrugh the training set) takes O( D * w), with D tuples and w weights. Hwever, # epchs can be expnential in n, the number f inputs, in the wrst case. Rule extractin frm netwrks (netwrk pruning): Simplify the netwrk structure by remving weighted links that have the least effect n the trained netwrk. Then perfrm link, unit, r activatin value clustering. The set f input and activatin values are studied t derive rules describing the relatinship between the input and hidden unit layers. Sensitivity analysis: Assess the impact that a given input variable has n a netwrk utput. The knwledge gained frm this analysis can be represented as rules. February 8, 2007 CSE-4412: Data Mining 35 18

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551

More information

Tree Structured Classifier

Tree Structured Classifier Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients

More information

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) >

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) > Btstrap Methd > # Purpse: understand hw btstrap methd wrks > bs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(bs) > mean(bs) [1] 21.64625 > # estimate f lambda > lambda = 1/mean(bs);

More information

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares

More information

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017 Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with

More information

Chapter 3: Cluster Analysis

Chapter 3: Cluster Analysis Chapter 3: Cluster Analysis } 3.1 Basic Cncepts f Clustering 3.1.1 Cluster Analysis 3.1. Clustering Categries } 3. Partitining Methds 3..1 The principle 3.. K-Means Methd 3..3 K-Medids Methd 3..4 CLARA

More information

What is Statistical Learning?

What is Statistical Learning? What is Statistical Learning? Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radi 0 20 40 60 80 100 Newspaper Shwn are Sales vs TV, Radi and Newspaper,

More information

Enhancing Performance of MLP/RBF Neural Classifiers via an Multivariate Data Distribution Scheme

Enhancing Performance of MLP/RBF Neural Classifiers via an Multivariate Data Distribution Scheme Enhancing Perfrmance f / Neural Classifiers via an Multivariate Data Distributin Scheme Halis Altun, Gökhan Gelen Nigde University, Electrical and Electrnics Engineering Department Nigde, Turkey haltun@nigde.edu.tr

More information

The blessing of dimensionality for kernel methods

The blessing of dimensionality for kernel methods fr kernel methds Building classifiers in high dimensinal space Pierre Dupnt Pierre.Dupnt@ucluvain.be Classifiers define decisin surfaces in sme feature space where the data is either initially represented

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Pattern Recognition 2014 Support Vector Machines

Pattern Recognition 2014 Support Vector Machines Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft

More information

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must M.E. Aggune, M.J. Dambrg, M.A. El-Sharkawi, R.J. Marks II and L.E. Atlas, "Dynamic and static security assessment f pwer systems using artificial neural netwrks", Prceedings f the NSF Wrkshp n Applicatins

More information

A Scalable Recurrent Neural Network Framework for Model-free

A Scalable Recurrent Neural Network Framework for Model-free A Scalable Recurrent Neural Netwrk Framewrk fr Mdel-free POMDPs April 3, 2007 Zhenzhen Liu, Itamar Elhanany Machine Intelligence Lab Department f Electrical and Cmputer Engineering The University f Tennessee

More information

COMP 551 Applied Machine Learning Lecture 4: Linear classification

COMP 551 Applied Machine Learning Lecture 4: Linear classification COMP 551 Applied Machine Learning Lecture 4: Linear classificatin Instructr: Jelle Pineau (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted

More information

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse

More information

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical mdel fr micrarray data analysis David Rssell Department f Bistatistics M.D. Andersn Cancer Center, Hustn, TX 77030, USA rsselldavid@gmail.cm

More information

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9. Sectin 7 Mdel Assessment This sectin is based n Stck and Watsn s Chapter 9. Internal vs. external validity Internal validity refers t whether the analysis is valid fr the ppulatin and sample being studied.

More information

Lecture 13: Markov Chain Monte Carlo. Gibbs sampling

Lecture 13: Markov Chain Monte Carlo. Gibbs sampling Lecture 13: Markv hain Mnte arl Gibbs sampling Gibbs sampling Markv chains 1 Recall: Apprximate inference using samples Main idea: we generate samples frm ur Bayes net, then cmpute prbabilities using (weighted)

More information

Support-Vector Machines

Support-Vector Machines Supprt-Vectr Machines Intrductin Supprt vectr machine is a linear machine with sme very nice prperties. Haykin chapter 6. See Alpaydin chapter 13 fr similar cntent. Nte: Part f this lecture drew material

More information

Resampling Methods. Chapter 5. Chapter 5 1 / 52

Resampling Methods. Chapter 5. Chapter 5 1 / 52 Resampling Methds Chapter 5 Chapter 5 1 / 52 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and

More information

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur Cdewrd Distributin fr Frequency Sensitive Cmpetitive Learning with One Dimensinal Input Data Aristides S. Galanpuls and Stanley C. Ahalt Department f Electrical Engineering The Ohi State University Abstract

More information

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels Mtivating Example Memry-Based Learning Instance-Based Learning K-earest eighbr Inductive Assumptin Similar inputs map t similar utputs If nt true => learning is impssible If true => learning reduces t

More information

Part 3 Introduction to statistical classification techniques

Part 3 Introduction to statistical classification techniques Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms

More information

NAME: Prof. Ruiz. 1. [5 points] What is the difference between simple random sampling and stratified random sampling?

NAME: Prof. Ruiz. 1. [5 points] What is the difference between simple random sampling and stratified random sampling? CS4445 ata Mining and Kwledge iscery in atabases. B Term 2014 Exam 1 Nember 24, 2014 Prf. Carlina Ruiz epartment f Cmputer Science Wrcester Plytechnic Institute NAME: Prf. Ruiz Prblem I: Prblem II: Prblem

More information

Admin. MDP Search Trees. Optimal Quantities. Reinforcement Learning

Admin. MDP Search Trees. Optimal Quantities. Reinforcement Learning Admin Reinfrcement Learning Cntent adapted frm Berkeley CS188 MDP Search Trees Each MDP state prjects an expectimax-like search tree Optimal Quantities The value (utility) f a state s: V*(s) = expected

More information

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts

Data Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification

More information

Five Whys How To Do It Better

Five Whys How To Do It Better Five Whys Definitin. As explained in the previus article, we define rt cause as simply the uncvering f hw the current prblem came int being. Fr a simple causal chain, it is the entire chain. Fr a cmplex

More information

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank CAUSAL INFERENCE Technical Track Sessin I Phillippe Leite The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Phillippe Leite fr the purpse f this wrkshp Plicy questins are causal

More information

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007 CS 477/677 Analysis f Algrithms Fall 2007 Dr. Gerge Bebis Curse Prject Due Date: 11/29/2007 Part1: Cmparisn f Srting Algrithms (70% f the prject grade) The bjective f the first part f the assignment is

More information

COMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d)

COMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d) COMP 551 Applied Machine Learning Lecture 9: Supprt Vectr Machines (cnt d) Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Class web page: www.cs.mcgill.ca/~hvanh2/cmp551 Unless therwise

More information

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards:

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards: MODULE FOUR This mdule addresses functins SC Academic Standards: EA-3.1 Classify a relatinship as being either a functin r nt a functin when given data as a table, set f rdered pairs, r graph. EA-3.2 Use

More information

IAML: Support Vector Machines

IAML: Support Vector Machines 1 / 22 IAML: Supprt Vectr Machines Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester 1 2 / 22 Outline Separating hyperplane with maimum margin Nn-separable training data Epanding the input int

More information

Distributions, spatial statistics and a Bayesian perspective

Distributions, spatial statistics and a Bayesian perspective Distributins, spatial statistics and a Bayesian perspective Dug Nychka Natinal Center fr Atmspheric Research Distributins and densities Cnditinal distributins and Bayes Thm Bivariate nrmal Spatial statistics

More information

You need to be able to define the following terms and answer basic questions about them:

You need to be able to define the following terms and answer basic questions about them: CS440/ECE448 Sectin Q Fall 2017 Midterm Review Yu need t be able t define the fllwing terms and answer basic questins abut them: Intr t AI, agents and envirnments Pssible definitins f AI, prs and cns f

More information

Department of Electrical Engineering, University of Waterloo. Introduction

Department of Electrical Engineering, University of Waterloo. Introduction Sectin 4: Sequential Circuits Majr Tpics Types f sequential circuits Flip-flps Analysis f clcked sequential circuits Mre and Mealy machines Design f clcked sequential circuits State transitin design methd

More information

Simple Linear Regression (single variable)

Simple Linear Regression (single variable) Simple Linear Regressin (single variable) Intrductin t Machine Learning Marek Petrik January 31, 2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins

More information

CS 109 Lecture 23 May 18th, 2016

CS 109 Lecture 23 May 18th, 2016 CS 109 Lecture 23 May 18th, 2016 New Datasets Heart Ancestry Netflix Our Path Parameter Estimatin Machine Learning: Frmally Many different frms f Machine Learning We fcus n the prblem f predictin Want

More information

Elements of Machine Intelligence - I

Elements of Machine Intelligence - I ECE-175A Elements f Machine Intelligence - I Ken Kreutz-Delgad Nun Vascncels ECE Department, UCSD Winter 2011 The curse The curse will cver basic, but imprtant, aspects f machine learning and pattern recgnitin

More information

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Emanuela Galass The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Emanuela Galass fr the purpse f this wrkshp When can we use

More information

Agenda. What is Machine Learning? Learning Type of Learning: Supervised, Unsupervised and semi supervised Classification

Agenda. What is Machine Learning? Learning Type of Learning: Supervised, Unsupervised and semi supervised Classification Agenda Artificial Intelligence and its applicatins Lecture 6 Supervised Learning Prfessr Daniel Yeung danyeung@ieee.rg Dr. Patrick Chan patrickchan@ieee.rg Suth China University f Technlgy, China Learning

More information

COMP9444 Neural Networks and Deep Learning 3. Backpropagation

COMP9444 Neural Networks and Deep Learning 3. Backpropagation COMP9444 Neural Netwrks and Deep Learning 3. Backprpagatin Tetbk, Sectins 4.3, 5.2, 6.5.2 COMP9444 17s2 Backprpagatin 1 Outline Supervised Learning Ockham s Razr (5.2) Multi-Layer Netwrks Gradient Descent

More information

Eric Klein and Ning Sa

Eric Klein and Ning Sa Week 12. Statistical Appraches t Netwrks: p1 and p* Wasserman and Faust Chapter 15: Statistical Analysis f Single Relatinal Netwrks There are fur tasks in psitinal analysis: 1) Define Equivalence 2) Measure

More information

Artificial Neural Networks MLP, Backpropagation

Artificial Neural Networks MLP, Backpropagation Artificial Neural Netwrks MLP, Backprpagatin 01001110 01100101 01110101 01110010 01101111 01101110 01101111 01110110 01100001 00100000 01110011 01101011 01110101 01110000 01101001 01101110 01100001 00100000

More information

Kinetic Model Completeness

Kinetic Model Completeness 5.68J/10.652J Spring 2003 Lecture Ntes Tuesday April 15, 2003 Kinetic Mdel Cmpleteness We say a chemical kinetic mdel is cmplete fr a particular reactin cnditin when it cntains all the species and reactins

More information

Early detection of mining truck failure by modelling its operation with neural networks classification algorithms

Early detection of mining truck failure by modelling its operation with neural networks classification algorithms RU, Rand GOLOSINSKI, T.S. Early detectin f mining truck failure by mdelling its peratin with neural netwrks classificatin algrithms. Applicatin f Cmputers and Operatins Research ill the Minerals Industries,

More information

Linear Classification

Linear Classification Linear Classificatin CS 54: Machine Learning Slides adapted frm Lee Cper, Jydeep Ghsh, and Sham Kakade Review: Linear Regressin CS 54 [Spring 07] - H Regressin Given an input vectr x T = (x, x,, xp), we

More information

The Kullback-Leibler Kernel as a Framework for Discriminant and Localized Representations for Visual Recognition

The Kullback-Leibler Kernel as a Framework for Discriminant and Localized Representations for Visual Recognition The Kullback-Leibler Kernel as a Framewrk fr Discriminant and Lcalized Representatins fr Visual Recgnitin Nun Vascncels Purdy H Pedr Mren ECE Department University f Califrnia, San Dieg HP Labs Cambridge

More information

Checking the resolved resonance region in EXFOR database

Checking the resolved resonance region in EXFOR database Checking the reslved resnance regin in EXFOR database Gttfried Bertn Sciété de Calcul Mathématique (SCM) Oscar Cabells OECD/NEA Data Bank JEFF Meetings - Sessin JEFF Experiments Nvember 0-4, 017 Bulgne-Billancurt,

More information

Statistical Learning. 2.1 What Is Statistical Learning?

Statistical Learning. 2.1 What Is Statistical Learning? 2 Statistical Learning 2.1 What Is Statistical Learning? In rder t mtivate ur study f statistical learning, we begin with a simple example. Suppse that we are statistical cnsultants hired by a client t

More information

Chapter 6 Classification and Prediction (2)

Chapter 6 Classification and Prediction (2) Chapter 6 Classification and Prediction (2) Outline Classification and Prediction Decision Tree Naïve Bayes Classifier Support Vector Machines (SVM) K-nearest Neighbors Accuracy and Error Measures Feature

More information

Collocation Map for Overcoming Data Sparseness

Collocation Map for Overcoming Data Sparseness Cllcatin Map fr Overcming Data Sparseness Mnj Kim, Yung S. Han, and Key-Sun Chi Department f Cmputer Science Krea Advanced Institute f Science and Technlgy Taejn, 305-701, Krea mj0712~eve.kaist.ac.kr,

More information

Analysis on the Stability of Reservoir Soil Slope Based on Fuzzy Artificial Neural Network

Analysis on the Stability of Reservoir Soil Slope Based on Fuzzy Artificial Neural Network Research Jurnal f Applied Sciences, Engineering and Technlgy 5(2): 465-469, 2013 ISSN: 2040-7459; E-ISSN: 2040-7467 Maxwell Scientific Organizatin, 2013 Submitted: May 08, 2012 Accepted: May 29, 2012 Published:

More information

Naïve Bayesian. From Han Kamber Pei

Naïve Bayesian. From Han Kamber Pei Naïve Bayesian From Han Kamber Pei Bayesian Theorem: Basics Let X be a data sample ( evidence ): class label is unknown Let H be a hypothesis that X belongs to class C Classification is to determine H

More information

Modelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA

Modelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA Mdelling f Clck Behaviur Dn Percival Applied Physics Labratry University f Washingtn Seattle, Washingtn, USA verheads and paper fr talk available at http://faculty.washingtn.edu/dbp/talks.html 1 Overview

More information

Department of Economics, University of California, Davis Ecn 200C Micro Theory Professor Giacomo Bonanno. Insurance Markets

Department of Economics, University of California, Davis Ecn 200C Micro Theory Professor Giacomo Bonanno. Insurance Markets Department f Ecnmics, University f alifrnia, Davis Ecn 200 Micr Thery Prfessr Giacm Bnann Insurance Markets nsider an individual wh has an initial wealth f. ith sme prbability p he faces a lss f x (0

More information

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression 4th Indian Institute f Astrphysics - PennState Astrstatistics Schl July, 2013 Vainu Bappu Observatry, Kavalur Crrelatin and Regressin Rahul Ry Indian Statistical Institute, Delhi. Crrelatin Cnsider a tw

More information

Lab 1 The Scientific Method

Lab 1 The Scientific Method INTRODUCTION The fllwing labratry exercise is designed t give yu, the student, an pprtunity t explre unknwn systems, r universes, and hypthesize pssible rules which may gvern the behavir within them. Scientific

More information

A Matrix Representation of Panel Data

A Matrix Representation of Panel Data web Extensin 6 Appendix 6.A A Matrix Representatin f Panel Data Panel data mdels cme in tw brad varieties, distinct intercept DGPs and errr cmpnent DGPs. his appendix presents matrix algebra representatins

More information

NUROP CONGRESS PAPER CHINESE PINYIN TO CHINESE CHARACTER CONVERSION

NUROP CONGRESS PAPER CHINESE PINYIN TO CHINESE CHARACTER CONVERSION NUROP Chinese Pinyin T Chinese Character Cnversin NUROP CONGRESS PAPER CHINESE PINYIN TO CHINESE CHARACTER CONVERSION CHIA LI SHI 1 AND LUA KIM TENG 2 Schl f Cmputing, Natinal University f Singapre 3 Science

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Matrix Data: Classification: Part 2 Instructor: Yizhou Sun yzsun@ccs.neu.edu September 21, 2014 Methods to Learn Matrix Data Set Data Sequence Data Time Series Graph & Network

More information

CS:4420 Artificial Intelligence

CS:4420 Artificial Intelligence CS:4420 Artificial Intelligence Spring 2017 Learning frm Examples Cesare Tinelli The University f Iwa Cpyright 2004 17, Cesare Tinelli and Stuart Russell a a These ntes were riginally develped by Stuart

More information

BLAST / HIDDEN MARKOV MODELS

BLAST / HIDDEN MARKOV MODELS CS262 (Winter 2015) Lecture 5 (January 20) Scribe: Kat Gregry BLAST / HIDDEN MARKOV MODELS BLAST CONTINUED HEURISTIC LOCAL ALIGNMENT Use Cmmnly used t search vast bilgical databases (n the rder f terabases/tetrabases)

More information

Churn Prediction using Dynamic RFM-Augmented node2vec

Churn Prediction using Dynamic RFM-Augmented node2vec Churn Predictin using Dynamic RFM-Augmented nde2vec Sandra Mitrvić, Jchen de Weerdt, Bart Baesens & Wilfried Lemahieu Department f Decisin Sciences and Infrmatin Management, KU Leuven 18 September 2017,

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Chapter 8&9: Classification: Part 3 Instructor: Yizhou Sun yzsun@ccs.neu.edu March 12, 2013 Midterm Report Grade Distribution 90-100 10 80-89 16 70-79 8 60-69 4

More information

15-381/781 Bayesian Nets & Probabilistic Inference

15-381/781 Bayesian Nets & Probabilistic Inference 15-381/781 Bayesian Nets & Prbabilistic Inference Emma Brunskill (this time) Ariel Prcaccia With thanks t Dan Klein (Berkeley), Percy Liang (Stanfrd) and Past 15-381 Instructrs fr sme slide cntent, and

More information

Hypothesis Tests for One Population Mean

Hypothesis Tests for One Population Mean Hypthesis Tests fr One Ppulatin Mean Chapter 9 Ala Abdelbaki Objective Objective: T estimate the value f ne ppulatin mean Inferential statistics using statistics in rder t estimate parameters We will be

More information

Least Squares Optimal Filtering with Multirate Observations

Least Squares Optimal Filtering with Multirate Observations Prc. 36th Asilmar Cnf. n Signals, Systems, and Cmputers, Pacific Grve, CA, Nvember 2002 Least Squares Optimal Filtering with Multirate Observatins Charles W. herrien and Anthny H. Hawes Department f Electrical

More information

7 TH GRADE MATH STANDARDS

7 TH GRADE MATH STANDARDS ALGEBRA STANDARDS Gal 1: Students will use the language f algebra t explre, describe, represent, and analyze number expressins and relatins 7 TH GRADE MATH STANDARDS 7.M.1.1: (Cmprehensin) Select, use,

More information

COMP9414/ 9814/ 3411: Artificial Intelligence. 14. Course Review. COMP3411 c UNSW, 2014

COMP9414/ 9814/ 3411: Artificial Intelligence. 14. Course Review. COMP3411 c UNSW, 2014 COMP9414/ 9814/ 3411: Artificial Intelligence 14. Curse Review COMP9414/9814/3411 14s1 Review 1 Assessment Assessable cmpnents f the curse: Assignment 1 10% Assignment 2 8% Assignment 3 12% Written Eam

More information

ECEN 4872/5827 Lecture Notes

ECEN 4872/5827 Lecture Notes ECEN 4872/5827 Lecture Ntes Lecture #5 Objectives fr lecture #5: 1. Analysis f precisin current reference 2. Appraches fr evaluating tlerances 3. Temperature Cefficients evaluatin technique 4. Fundamentals

More information

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS 1 Influential bservatins are bservatins whse presence in the data can have a distrting effect n the parameter estimates and pssibly the entire analysis,

More information

Logistic Regression. and Maximum Likelihood. Marek Petrik. Feb

Logistic Regression. and Maximum Likelihood. Marek Petrik. Feb Lgistic Regressin and Maximum Likelihd Marek Petrik Feb 09 2017 S Far in ML Regressin vs Classificatin Linear regressin Bias-variance decmpsitin Practical methds fr linear regressin Simple Linear Regressin

More information

Chapter Summary. Mathematical Induction Strong Induction Recursive Definitions Structural Induction Recursive Algorithms

Chapter Summary. Mathematical Induction Strong Induction Recursive Definitions Structural Induction Recursive Algorithms Chapter 5 1 Chapter Summary Mathematical Inductin Strng Inductin Recursive Definitins Structural Inductin Recursive Algrithms Sectin 5.1 3 Sectin Summary Mathematical Inductin Examples f Prf by Mathematical

More information

Dataflow Analysis and Abstract Interpretation

Dataflow Analysis and Abstract Interpretation Dataflw Analysis and Abstract Interpretatin Cmputer Science and Artificial Intelligence Labratry MIT Nvember 9, 2015 Recap Last time we develped frm first principles an algrithm t derive invariants. Key

More information

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Céline Ferré The Wrld Bank When can we use matching? What if the assignment t the treatment is nt dne randmly r based n an eligibility index, but n the basis

More information

making triangle (ie same reference angle) ). This is a standard form that will allow us all to have the X= y=

making triangle (ie same reference angle) ). This is a standard form that will allow us all to have the X= y= Intrductin t Vectrs I 21 Intrductin t Vectrs I 22 I. Determine the hrizntal and vertical cmpnents f the resultant vectr by cunting n the grid. X= y= J. Draw a mangle with hrizntal and vertical cmpnents

More information

Performance Bounds for Detect and Avoid Signal Sensing

Performance Bounds for Detect and Avoid Signal Sensing Perfrmance unds fr Detect and Avid Signal Sensing Sam Reisenfeld Real-ime Infrmatin etwrks, University f echnlgy, Sydney, radway, SW 007, Australia samr@uts.edu.au Abstract Detect and Avid (DAA) is a Cgnitive

More information

In SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw:

In SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw: In SMV I IAML: Supprt Vectr Machines II Nigel Gddard Schl f Infrmatics Semester 1 We sa: Ma margin trick Gemetry f the margin and h t cmpute it Finding the ma margin hyperplane using a cnstrained ptimizatin

More information

Maximum A Posteriori (MAP) CS 109 Lecture 22 May 16th, 2016

Maximum A Posteriori (MAP) CS 109 Lecture 22 May 16th, 2016 Maximum A Psteriri (MAP) CS 109 Lecture 22 May 16th, 2016 Previusly in CS109 Game f Estimatrs Maximum Likelihd Nn spiler: this didn t happen Side Plt argmax argmax f lg Mther f ptimizatins? Reviving an

More information

Sequential Allocation with Minimal Switching

Sequential Allocation with Minimal Switching In Cmputing Science and Statistics 28 (1996), pp. 567 572 Sequential Allcatin with Minimal Switching Quentin F. Stut 1 Janis Hardwick 1 EECS Dept., University f Michigan Statistics Dept., Purdue University

More information

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001)

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001) CN700 Additive Mdels and Trees Chapter 9: Hastie et al. (2001) Madhusudana Shashanka Department f Cgnitive and Neural Systems Bstn University CN700 - Additive Mdels and Trees March 02, 2004 p.1/34 Overview

More information

, which yields. where z1. and z2

, which yields. where z1. and z2 The Gaussian r Nrmal PDF, Page 1 The Gaussian r Nrmal Prbability Density Functin Authr: Jhn M Cimbala, Penn State University Latest revisin: 11 September 13 The Gaussian r Nrmal Prbability Density Functin

More information

Technical Bulletin. Generation Interconnection Procedures. Revisions to Cluster 4, Phase 1 Study Methodology

Technical Bulletin. Generation Interconnection Procedures. Revisions to Cluster 4, Phase 1 Study Methodology Technical Bulletin Generatin Intercnnectin Prcedures Revisins t Cluster 4, Phase 1 Study Methdlgy Release Date: Octber 20, 2011 (Finalizatin f the Draft Technical Bulletin released n September 19, 2011)

More information

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came.

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came. MATH 1342 Ch. 24 April 25 and 27, 2013 Page 1 f 5 CHAPTER 24: INFERENCE IN REGRESSION Chapters 4 and 5: Relatinships between tw quantitative variables. Be able t Make a graph (scatterplt) Summarize the

More information

Reinforcement Learning" CMPSCI 383 Nov 29, 2011!

Reinforcement Learning CMPSCI 383 Nov 29, 2011! Reinfrcement Learning" CMPSCI 383 Nv 29, 2011! 1 Tdayʼs lecture" Review f Chapter 17: Making Cmple Decisins! Sequential decisin prblems! The mtivatin and advantages f reinfrcement learning.! Passive learning!

More information

Computational modeling techniques

Computational modeling techniques Cmputatinal mdeling techniques Lecture 2: Mdeling change. In Petre Department f IT, Åb Akademi http://users.ab.fi/ipetre/cmpmd/ Cntent f the lecture Basic paradigm f mdeling change Examples Linear dynamical

More information

I.S. 239 Mark Twain. Grade 7 Mathematics Spring Performance Task: Proportional Relationships

I.S. 239 Mark Twain. Grade 7 Mathematics Spring Performance Task: Proportional Relationships I.S. 239 Mark Twain 7 ID Name: Date: Grade 7 Mathematics Spring Perfrmance Task: Prprtinal Relatinships Directins: Cmplete all parts f each sheet fr each given task. Be sure t read thrugh the rubrics s

More information

MATHEMATICS SYLLABUS SECONDARY 5th YEAR

MATHEMATICS SYLLABUS SECONDARY 5th YEAR Eurpean Schls Office f the Secretary-General Pedaggical Develpment Unit Ref. : 011-01-D-8-en- Orig. : EN MATHEMATICS SYLLABUS SECONDARY 5th YEAR 6 perid/week curse APPROVED BY THE JOINT TEACHING COMMITTEE

More information

ENSC Discrete Time Systems. Project Outline. Semester

ENSC Discrete Time Systems. Project Outline. Semester ENSC 49 - iscrete Time Systems Prject Outline Semester 006-1. Objectives The gal f the prject is t design a channel fading simulatr. Upn successful cmpletin f the prject, yu will reinfrce yur understanding

More information

Time, Synchronization, and Wireless Sensor Networks

Time, Synchronization, and Wireless Sensor Networks Time, Synchrnizatin, and Wireless Sensr Netwrks Part II Ted Herman University f Iwa Ted Herman/March 2005 1 Presentatin: Part II metrics and techniques single-hp beacns reginal time znes ruting-structure

More information

Math Foundations 20 Work Plan

Math Foundations 20 Work Plan Math Fundatins 20 Wrk Plan Units / Tpics 20.8 Demnstrate understanding f systems f linear inequalities in tw variables. Time Frame December 1-3 weeks 6-10 Majr Learning Indicatrs Identify situatins relevant

More information

Fall 2013 Physics 172 Recitation 3 Momentum and Springs

Fall 2013 Physics 172 Recitation 3 Momentum and Springs Fall 03 Physics 7 Recitatin 3 Mmentum and Springs Purpse: The purpse f this recitatin is t give yu experience wrking with mmentum and the mmentum update frmula. Readings: Chapter.3-.5 Learning Objectives:.3.

More information

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter Midwest Big Data Summer Schl: Machine Learning I: Intrductin Kris De Brabanter kbrabant@iastate.edu Iwa State University Department f Statistics Department f Cmputer Science June 24, 2016 1/24 Outline

More information

On classifier behavior in the presence of mislabeling noise

On classifier behavior in the presence of mislabeling noise Data Min Knwl Disc DOI 10.1007/s10618-016-0484-8 On classifier behavir in the presence f mislabeling nise Katsiaryna Mirylenka 1 Gerge Giannakpuls 2 Le Minh D 3 Themis Palpanas 4 Received: 12 Nvember 2015

More information

Pure adaptive search for finite global optimization*

Pure adaptive search for finite global optimization* Mathematical Prgramming 69 (1995) 443-448 Pure adaptive search fr finite glbal ptimizatin* Z.B. Zabinskya.*, G.R. Wd b, M.A. Steel c, W.P. Baritmpa c a Industrial Engineering Prgram, FU-20. University

More information

Particle Size Distributions from SANS Data Using the Maximum Entropy Method. By J. A. POTTON, G. J. DANIELL AND B. D. RAINFORD

Particle Size Distributions from SANS Data Using the Maximum Entropy Method. By J. A. POTTON, G. J. DANIELL AND B. D. RAINFORD 3 J. Appl. Cryst. (1988). 21,3-8 Particle Size Distributins frm SANS Data Using the Maximum Entrpy Methd By J. A. PTTN, G. J. DANIELL AND B. D. RAINFRD Physics Department, The University, Suthamptn S9

More information

Graduate AI Lecture 16: Planning 2. Teachers: Martial Hebert Ariel Procaccia (this time)

Graduate AI Lecture 16: Planning 2. Teachers: Martial Hebert Ariel Procaccia (this time) Graduate AI Lecture 16: Planning 2 Teachers: Martial Hebert Ariel Prcaccia (this time) Reminder State is a cnjunctin f cnditins, e.g., at(truck 1,Shadyside) at(truck 2,Oakland) States are transfrmed via

More information

11. DUAL NATURE OF RADIATION AND MATTER

11. DUAL NATURE OF RADIATION AND MATTER 11. DUAL NATURE OF RADIATION AND MATTER Very shrt answer and shrt answer questins 1. Define wrk functin f a metal? The minimum energy required fr an electrn t escape frm the metal surface is called the

More information

NUMBERS, MATHEMATICS AND EQUATIONS

NUMBERS, MATHEMATICS AND EQUATIONS AUSTRALIAN CURRICULUM PHYSICS GETTING STARTED WITH PHYSICS NUMBERS, MATHEMATICS AND EQUATIONS An integral part t the understanding f ur physical wrld is the use f mathematical mdels which can be used t

More information