An Introduction. Statistical Learning. The Elements of Statistical Learning. Data Mining, Inference, and Prediction.
|
|
- Marilyn McDonald
- 5 years ago
- Views:
Transcription
1 1 Intrductin CS 189 / 289A [Spring 217] Machine Learning Jnathan Shewchuk jrs/189 TAs: Daylen Yang, Ting-Chun Wang, Muls Vretts, Mstafa Rhaninejad, Michael Zhang, Anurag Ajay, Alvin Wan, Srush Nasiriany, Garrett Thmas, Nah Glmant, Adam Villaflr, Raul Puri, Alex Francis Questins: Please use Piazza, nt [Piazza has an ptin fr private questins, but please use public fr mst questins s ther peple can benefit] Fr persnal matters nly, jrs@cryeecsberkeleyedu Discussin sectins: 12 nw; mre will be added Attend any sectin If the rm is t full, please g t anther ne [Hwever, t get int the curse, yu have t pick sme sectin with space Desn t matter which ne!] N sectins this week [Enrllment: We re trying t raise it t 54 After enugh students drp, it s pssible that everyne might get in Cncurrent enrllment students have the lwest pririty; nn-cs grad students the secnd-lwest] [Textbks: Available free nline Linked frm class web page] f the field nd cmplex arketing t t imprtant pics include, tree-based d real-wrld is textbk ners in scienting the e statistical STS Springer Texts in Statistics Springer Series in Statistics James Witten Hastie Tibshirani 1 An Intrductin t Statistical Learning Gareth James Daniela Witten Trevr Hastie Rbert Tibshirani Trevr Hastie Rbert Tibshirani Jerme Friedman The Elements f Statictical Learning An Intrductin t Statistical Learning During the past decade there has been an explsin in cmputatin and infrmatin technlgy With it have cme vast amunts f data in a variety f fields such as medicine, bilgy, finance, and marketing The challenge f understanding these data has led t the devel-, Tibshirani nd machine pment f new tls in the field f statistics, and spawned new areas such as data mining, f the same machine learning, and biinfrmatics Many f these tls have cmmn underpinnings but targeted are at ften expressed with different terminlgy This bk describes the imprtant ideas in stical learnrse in linear these areas in a cmmn cnceptual framewrk While the apprach is statistical, the emphasis is n cncepts rather than mathematics Many examples are given, with a liberal use f clr graphics It shuld be a valuable resurce fr statisticians and anyne interested in data mining in science r industry The bk s cverage is brad, frm supervised learning nia He has (predictin) t unsupervised learning The many tpics include neural netwrks, supprt stical learncnceptual tpic in any bk vectr machines, classificatin trees and bsting the first cmprehensive treatment f this This majr new editin features many tpics nt cvered in the riginal, including graphical ingtn Her mdels, randm frests, ensemble methds, least angle regressin & path algrithms fr the ng She has lass, nn-negative matrix factrizatin, and spectral clustering There is als a chapter n f genmics, methds fr wide data (p bigger than n), including multiple testing and false discvery rates mittee that Trevr Hastie, Rbert Tibshirani, and Jerme Friedman are prfessrs f statistics at Stanfrd University They are prminent researchers in this area: Hastie and Tibshirani iversity, and develped generalized additive mdels and wrte a ppular bk f that title Hastie cdevelped much f the statistical mdeling sftware and envirnment in R/S-PLUS and Hastie and k f that invented principal curves and surfaces Tibshirani prpsed the lass and is c-authr f the nvirnment very successful An Intrductin t the Btstrap Friedman is the c-inventr f many datamining tls including CART, MARS, prjectin pursuit and gradient ed the lass bsting STATISTICS ISBN springercm with Applicatins in R Hastie Tibshirani Friedman The Elements f Statistical Learning Trevr Hastie Rbert Tibshirani Jerme Friedman Springer Series in Statistics The Elements f Statistical Learning Data Mining, Inference, and Predictin Secnd Editin 1
2 Prerequisites Math 5 (vectr calculus) Math 54 r 11 (linear algebra) CS 7 (prbability) NOT CS 188 [Might still be listed as a prerequisite, but we re having it remved] [BUT be aware that 189 midterm starts 1 minutes after 188 midterm ends] Grading: 189 4% 7 Hmewrks Late plicy: 5 slip days ttal 2% Midterm: Wednesday, March 15, in class (6: 8 pm) 4% Final Exam: MOVED t Mnday, May 8, 6 PM (Exam grup ) [Gd news fr sme f yu wh had final exam cnflicts] Grading: 289A 4% HW 2% Midterm 2% Final 2% Prject Cheating Discussin f HW prblems is encuraged All hmewrks, including prgramming, must be written individually We will actively check fr plagiarism Typical penalty is a large NEGATIVE scre, but I reserve right t give an instant F fr even ne vilatin, and will always give an F fr tw [Last time I taught CS 61B, we had t punish rughly 1 peple fr cheating It was very painful Please dn t put me thrugh that again] CORE MATERIAL Finding patterns in data; using them t make predictins Mdels and statistics help us understand patterns Optimizatin algrithms learn the patterns [The mst imprtant part f this is the data Data drives everything else Yu cannt learn much if yu dn t have enugh data Yu cannt learn much if yur data sucks But it s amazing what yu can d if yu have lts f gd data Machine learning has changed a lt in the last decade because the internet has made truly vast quantities f data available Fr instance, with a little patience yu can dwnlad tens f millins f phtgraphs Then yu can build a D mdel f Paris Sme techniques that had fallen ut f favr, like neural nets, have cme back big in the last few years because researchers fund that they wrk s much better when yu have vast quantities f data] 2
3 CLASSIFICATION 42 Why Nt Linear Regressin? 129 Incme Balance Incme Balance N Yes Default N Yes Default FIGURE 41 The Default data set Left: The annual incmes and mnthly credit card balances f a number f individuals The individuals wh defaulted n their credit card payments are shwn in range, and thse wh did nt are shwn in blue Center: Bxplts f balance as a functin f default status Right: Bxplts f incme as a functin f default status creditcardspdf (ISL, Figure 41) [The prblem f classificatin We are given data pints, each 42belnging Whyt ne Nt f tw Linear classes Regressin? Then we are given additinal pints whse class is unknwn, and we are asked t predict what class each new pint is in Given the credit card balance We have andstated annual incme that linear f a cardhlder, regressin predict is nt whether apprpriate they will default in the ncase their debt] f a Cllect qualitative training respnse data: reliable Why debtrs nt? & defaulted debtrs Evaluate Suppse new applicants that we(predictin) are trying t predict the medical cnditin f a patient in the emergency rm n the basis f her symptms In this simplified example, there are three pssible diagnses: decisin strke, bundary drug verdse, and epileptic seizure Weculdcnsiderencdingthesevaluesasaquantitative respnse variable, Y, asfllws: 1 if strke; Y = 2 if drug verdse; if epileptic seizure Using this cding, least squares culd be used t fit a linear regressin mdel t predict Y n the basis f a set f predictrs X 1,,X p Unfrtunately, this cding [Drawimplies this figure anbyrdering hand classifypdf n the utcmes, ] putting drug verdse in between [Draw strke 2 clrs andfepileptic dts, almst but seizure, nt quite andinsistingthatthedifference linearly separable] between [ Hw strke d we and classify druga new verdse pint? isdraw the asame pint in asa the thirddifference clr] between drug verdse [One pssibility: and epileptic lk at its nearest seizure neighbr] Inpracticethereisnparticular [Anther pssibility: draw a linear decisin bundary; label it] reasn that this needs t be the case Fr instance, ne culd chse an [Thse are tw different mdels fr the data] equally reasnable cding, [We ll learn sme ways t draw these linear decisin bundaries in the next several lectures But fr nw, let s cmpare this methd with anther methd] 1 if epileptic seizure; Y = 2 if strke; if drug verdse
4 16 2 Overview f Supervised Learning 2 Least Squares and Nearest Neighbrs 1 1 Nearest Neighbr Classifier Linear Regressin f /1 Respnse classnearpdf, classlinearpdf (ESL, Figures 2 & 21) [Here are tw examples f classifiers fr the same data At left we have a nearest neighbr classifier, which classifies a pint by finding the nearest pint in the input data, and assigning it the same class At right we have a linear classifier, which guesses that everything abve the line is brwn, and everything belw the line is blue The decisin bundaries are in black] FIGURE 21 A classificatin example in tw dimensins The classes are cded FIGURE 2 The same classificatin example in tw dimensins as in Figure 21 The classes are cded as a binary variable (BLUE =, ORANGE =1),and as a binary variable (BLUE =, ORANGE =1), and then fit by linear regressin The line is the decisin bundary defined by x T then predicted by 1-nearest-neighbr classificatin ˆβ =5 Therangeshadedregin dentes that part f input space classified as ORANGE, while the blue regin is classified as BLUE Frm Overview Least f Supervised Squares t Learning Nearest Neighbrs 2 Least Squares and Nearest Neighbrs The linear decisin bundary frm least squares is very The smth, set fand pints ap-iparently stable t fit It 1 Nearest des appear Neighbr t Classifier rely heavily nindicated the assumptin in Figure 21, and 15-Nearest the tw Neighbr predicted Classifier classes are separated by the IR 2 classified as ORANGE crrespnds t {x : x T ˆβ 15 > 5}, that a linear decisin bundary is apprpriate In language decisin we willbundary develp {x : x T ˆβ =5}, whichislinearinthiscasewesee shrtly, it has lw variance and ptentially high bias that fr these data there are several misclassificatins n bth sides f the On the ther hand, the k-nearest-neighbr prcedures decisin d nt appear bundary t Perhaps ur linear mdel is t rigid rare such errrs rely n any stringent assumptins abut the underlying data,andcanadapt unavidable? Remember that these are errrs n the training data itself, t any situatin Hwever, any particular subregin f the and decisin we have bundary depends n a handful f input pints and their particular nt said where the cnstructed data came frm Cnsider the tw pssible psitins, scenaris: and is thus wiggly and unstable high variance and lw bias Scenari 1: The training data in each class were generated frm bivariate Each methd has its wn situatins fr which it wrks best; ingaussian particular distributins with uncrrelated cmpnents and different linear regressin is mre apprpriate fr Scenari 1 abve, while means nearest neighbrs are mre suitable fr Scenari 2 The time has cme t expse the racle! The data in fact were simulated frm a mdel Scenari smewhere 2: The between the tw, but clser t Scenari 2 training data in each class came frm a mixture f 1 lw- means m Gaussian k distributins, with individual means themselves First we generated 1variance frm a bivariate Gaussian distributin N((1, ) T, I) andlabeledthisclass distributed as Gaussian BLUE Similarly,1mreweredrawnfrmN((, 1) T, I) andlabeledclass ORANGE Thenfreachclasswegenerated1bservatinsasfllws: A mixture f frgaussians is best described in terms f the generative each bservatin, we picked an m k at randm with prbability mdel 1/1, One first and generates a discrete variable that determines which f FIGURE classnearpdf, 2 The same classnear15pdf classificatin example(esl, in tw dimensins Figures FIGURE as2 22 in Figure 21 The classes are cded a binary variable (BLUE =, & The22) same classificatin [At example right in twwe dimensins have as in Fig-ure 21 The classes are cded as a binary variable (BLUE =, ORANGE =1)and then ORANGE fit by =1),and 15-nearest neighbr classifier Instead f lking 15-nearest-neighbr at the averaging nearest as in (28) neighbr The predictedf classais hence new then predicted by 1-nearest-neighbr classificatin chsen by majrity vte amngst the 15-nearest neighbrs pint, it lks at the 15 nearest neighbrs and lets them vte fr the crrect class The 1-nearest neighbr classifier at left has a big advantage: it classifies all the training data crrectly, whereas the 15-nearest neighbr classifier at right figure des nt But the right 2 Frm Least Squares t Nearest Neighbrs In Figure 22 we see that far fewer training bservatins are misclassified The linear decisin bundary frm least squares is very than smth, in Figure and apparently stable t fit It des appear t rely heavily ninthe Figure assumptin 2 nne f the training data are misclassified A little thught 21 This shuld nt give us t much cmfrt, thugh, since that a linear decisin bundary is apprpriate In language suggests we will that develp fr k-nearest-neighbr fits, the errr n the training data figure has an advantage t Smebdy shrtly, it has lw variance and ptentially high bias shuld please be apprximately tell me what] an increasing functin f k, andwillalwaysbe On the ther hand, the k-nearest-neighbr prcedures fr d nt k =1Anindependenttestsetwuldgiveusamresatisfactrymeans appear t rely any stringent assumptins abut the underlying data,andcanadapt fr cmparing the different methds t any situatin Hwever, any particular subregin f the decisin It appears bundary depends a handful f input pints and their particular that k-nearest-neighbr fits have a single parameter, the number f neighbrs psitins, k, cmpared t the p parameters in least-squares fits Althugh this is the case, we will see that the effective number f parameters Classifica9n(Pipeline( and is thus wiggly and unstable high variance and lw bias Each methd has its wn situatins fr which it wrks best; f k-nearest in particular neighbrs is N/k and is generally bigger than p, anddecreases with increasing k Tgetanideafwhy,ntethatiftheneighbrhds linear regressin is mre apprpriate fr Scenari 1 abve, while nearest were nnverlapping, there wuld be N/k neighbrhds and we wuld fit neighbrs are mre suitable fr Scenari 2 The time has cme t expse ne parameter (a mean) in each neighbrhd the racle! The data in fact were simulated frm a mdel smewhere between the tw, but clser t Scenari 2 First we generated 1 means m k It is als clear that we cannt use sum-f-squared errrs n the training frm a bivariate Gaussian distributin N((1, ) T set as a criterin fr picking k, since we wuld always pick k =1!Itwuld, I) andlabeledthisclass BLUE Similarly,1mreweredrawnfrmN((, 1) T seem that k-nearest-neighbr methds wuld be mre apprpriate fr the, I) andlabeledclass mixture Scenari 2 described abve, while fr Gaussian data the decisin ORANGE Thenfreachclasswegenerated1bservatinsasfllws: fr bundaries f k-nearest neighbrs wuld be unnecessarily nisy each bservatin, we picked an m k at randm with prbability 1/1, and [The left figure is an example f what s called verfitting In the left figure, bserve hw intricate the decisin bundary is that separates the psitive examples frm the negative examples It s a bit t intricate t reflect reality In the right figure, the decisin bundary is smther Intuitively, that smthness is prbably mre likely t crrespnd t reality] Cllect(Training(Images( Classifying Digits Psi9ve:(( Nega9ve:(( Training(Time( sevensnespdf [In this simplified digit recgnitin prblem, we are given handwritten 7 s and 1 s, and we are asked t learn t distinguish the 7 s frm the 1 s] Cmpute(feature(vectrs(fr(psi9ve(and(nega9ve( 4 example(images( Train(a(classifier(
5 Express these images as vectrs Images are pints in 16-dimensinal space Linear decisin bundary is a hyperplane Validatin Train a classifier: it learns t distinguish 7 frm nt 7 Test the classifier n NEW images 2 kinds f errr: Training set errr: fractin f training images nt classified crrectly [This is zer with the 1-nearest neighbr classifier, but nnzer with the 15-nearest neighbr and linear classifiers we ve just seen] Test set errr: fractin f misclassified NEW images, nt seen during training [When I underline a wrd r phrase, that usually means it s a definitin If yu want t d well in this curse, my advice t yu is t memrize the definitins I cver in class] utliers: pints whse labels are atypical (eg slvent brrwer wh defaulted anyway) verfitting: when the test errr deterirates because the classifier becmes t sensitive t utliers r ther spurius patterns [In machine learning, the gal is t create a classifier that generalizes t new examples we haven t seen yet Overfitting is cunterprductive t that gal S we re always seeking a cmprmise: we want decisin bundaries that make fine distinctins withut being dwnright superstitius] 5
6 Mst ML algrithms have a few hyperparameters that cntrl ver/underfitting, eg k in k-nearest neighbrs underfit k: # f nearest neighbrs errr rate Train Test Bayes Linear test errr training errr verfit! best (7) verfitlabeledpdf (mdified frm ESL, Figure 24) We select them by validatin: Hld back a subset f training data, called the validatin set Train the classifier multiple times with different hyperparameter settings Chse the settings that wrk best n validatin set Nw we have sets: training set used t learn mdel weights validatin set used t tune hyperparameters, chse amng different mdels test set used as FINAL evaluatin f mdel Keep in a vault Run ONCE, at the very end [It s very bad when researchers in medicine r pharmaceuticals peek int the test set prematurely!] Kagglecm: Runs ML cmpetitins, including ur HWs We use 2 data sets: public set results available during cmpetitin private set revealed nly after due date [If yur public results are a lt better than yur private results, we will knw that yu verfitted] Techniques [taught in this class, NOT a cmplete list] Supervised learning: Classificatin: is this spam? Regressin: hw likely des this patient have cancer? Unsupervised learning: Clustering: which DNA sequences are similar t each ther? Dimensinality reductin: what are cmmn features f faces? cmmn differences? 6
What is Statistical Learning?
What is Statistical Learning? Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radi 0 20 40 60 80 100 Newspaper Shwn are Sales vs TV, Radi and Newspaper,
More informationResampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017
Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with
More informationIAML: Support Vector Machines
1 / 22 IAML: Supprt Vectr Machines Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester 1 2 / 22 Outline Separating hyperplane with maimum margin Nn-separable training data Epanding the input int
More informationCAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank
CAUSAL INFERENCE Technical Track Sessin I Phillippe Leite The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Phillippe Leite fr the purpse f this wrkshp Plicy questins are causal
More informationCOMP 551 Applied Machine Learning Lecture 4: Linear classification
COMP 551 Applied Machine Learning Lecture 4: Linear classificatin Instructr: Jelle Pineau (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted
More informationPattern Recognition 2014 Support Vector Machines
Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft
More informationResampling Methods. Chapter 5. Chapter 5 1 / 52
Resampling Methds Chapter 5 Chapter 5 1 / 52 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and
More informationLecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff
Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised
More informationLecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff
Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised
More informationk-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels
Mtivating Example Memry-Based Learning Instance-Based Learning K-earest eighbr Inductive Assumptin Similar inputs map t similar utputs If nt true => learning is impssible If true => learning reduces t
More informationSimple Linear Regression (single variable)
Simple Linear Regressin (single variable) Intrductin t Machine Learning Marek Petrik January 31, 2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins
More informationCOMP 551 Applied Machine Learning Lecture 11: Support Vector Machines
COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse
More informationChapter 3: Cluster Analysis
Chapter 3: Cluster Analysis } 3.1 Basic Cncepts f Clustering 3.1.1 Cluster Analysis 3.1. Clustering Categries } 3. Partitining Methds 3..1 The principle 3.. K-Means Methd 3..3 K-Medids Methd 3..4 CLARA
More informationx 1 Outline IAML: Logistic Regression Decision Boundaries Example Data
Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares
More informationCOMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification
COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551
More informationThe Law of Total Probability, Bayes Rule, and Random Variables (Oh My!)
The Law f Ttal Prbability, Bayes Rule, and Randm Variables (Oh My!) Administrivia Hmewrk 2 is psted and is due tw Friday s frm nw If yu didn t start early last time, please d s this time. Gd Milestnes:
More informationMath Foundations 20 Work Plan
Math Fundatins 20 Wrk Plan Units / Tpics 20.8 Demnstrate understanding f systems f linear inequalities in tw variables. Time Frame December 1-3 weeks 6-10 Majr Learning Indicatrs Identify situatins relevant
More informationInternal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.
Sectin 7 Mdel Assessment This sectin is based n Stck and Watsn s Chapter 9. Internal vs. external validity Internal validity refers t whether the analysis is valid fr the ppulatin and sample being studied.
More informationBiplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 13 Offprint
Biplts in Practice MICHAEL GREENACRE Prfessr f Statistics at the Pmpeu Fabra University Chapter 13 Offprint CASE STUDY BIOMEDICINE Cmparing Cancer Types Accrding t Gene Epressin Arrays First published:
More informationCS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007
CS 477/677 Analysis f Algrithms Fall 2007 Dr. Gerge Bebis Curse Prject Due Date: 11/29/2007 Part1: Cmparisn f Srting Algrithms (70% f the prject grade) The bjective f the first part f the assignment is
More informationFive Whys How To Do It Better
Five Whys Definitin. As explained in the previus article, we define rt cause as simply the uncvering f hw the current prblem came int being. Fr a simple causal chain, it is the entire chain. Fr a cmplex
More informationINSTRUMENTAL VARIABLES
INSTRUMENTAL VARIABLES Technical Track Sessin IV Sergi Urzua University f Maryland Instrumental Variables and IE Tw main uses f IV in impact evaluatin: 1. Crrect fr difference between assignment f treatment
More informationBootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) >
Btstrap Methd > # Purpse: understand hw btstrap methd wrks > bs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(bs) > mean(bs) [1] 21.64625 > # estimate f lambda > lambda = 1/mean(bs);
More informationPhys. 344 Ch 7 Lecture 8 Fri., April. 10 th,
Phys. 344 Ch 7 Lecture 8 Fri., April. 0 th, 009 Fri. 4/0 8. Ising Mdel f Ferrmagnets HW30 66, 74 Mn. 4/3 Review Sat. 4/8 3pm Exam 3 HW Mnday: Review fr est 3. See n-line practice test lecture-prep is t
More informationENSC Discrete Time Systems. Project Outline. Semester
ENSC 49 - iscrete Time Systems Prject Outline Semester 006-1. Objectives The gal f the prject is t design a channel fading simulatr. Upn successful cmpletin f the prject, yu will reinfrce yur understanding
More informationCHM112 Lab Graphing with Excel Grading Rubric
Name CHM112 Lab Graphing with Excel Grading Rubric Criteria Pints pssible Pints earned Graphs crrectly pltted and adhere t all guidelines (including descriptive title, prperly frmatted axes, trendline
More informationExperiment #3. Graphing with Excel
Experiment #3. Graphing with Excel Study the "Graphing with Excel" instructins that have been prvided. Additinal help with learning t use Excel can be fund n several web sites, including http://www.ncsu.edu/labwrite/res/gt/gt-
More informationLesson Plan. Recode: They will do a graphic organizer to sequence the steps of scientific method.
Lessn Plan Reach: Ask the students if they ever ppped a bag f micrwave ppcrn and nticed hw many kernels were unppped at the bttm f the bag which made yu wnder if ther brands pp better than the ne yu are
More informationMedium Scale Integrated (MSI) devices [Sections 2.9 and 2.10]
EECS 270, Winter 2017, Lecture 3 Page 1 f 6 Medium Scale Integrated (MSI) devices [Sectins 2.9 and 2.10] As we ve seen, it s smetimes nt reasnable t d all the design wrk at the gate-level smetimes we just
More informationThis section is primarily focused on tools to aid us in finding roots/zeros/ -intercepts of polynomials. Essentially, our focus turns to solving.
Sectin 3.2: Many f yu WILL need t watch the crrespnding vides fr this sectin n MyOpenMath! This sectin is primarily fcused n tls t aid us in finding rts/zers/ -intercepts f plynmials. Essentially, ur fcus
More informationSUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis
SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical mdel fr micrarray data analysis David Rssell Department f Bistatistics M.D. Andersn Cancer Center, Hustn, TX 77030, USA rsselldavid@gmail.cm
More informationThe blessing of dimensionality for kernel methods
fr kernel methds Building classifiers in high dimensinal space Pierre Dupnt Pierre.Dupnt@ucluvain.be Classifiers define decisin surfaces in sme feature space where the data is either initially represented
More informationGetting Involved O. Responsibilities of a Member. People Are Depending On You. Participation Is Important. Think It Through
f Getting Invlved O Literature Circles can be fun. It is exciting t be part f a grup that shares smething. S get invlved, read, think, and talk abut bks! Respnsibilities f a Member Remember a Literature
More informationActivity Guide Loops and Random Numbers
Unit 3 Lessn 7 Name(s) Perid Date Activity Guide Lps and Randm Numbers CS Cntent Lps are a relatively straightfrward idea in prgramming - yu want a certain chunk f cde t run repeatedly - but it takes a
More informationLab 1 The Scientific Method
INTRODUCTION The fllwing labratry exercise is designed t give yu, the student, an pprtunity t explre unknwn systems, r universes, and hypthesize pssible rules which may gvern the behavir within them. Scientific
More informationDifferentiation Applications 1: Related Rates
Differentiatin Applicatins 1: Related Rates 151 Differentiatin Applicatins 1: Related Rates Mdel 1: Sliding Ladder 10 ladder y 10 ladder 10 ladder A 10 ft ladder is leaning against a wall when the bttm
More informationMidwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter
Midwest Big Data Summer Schl: Machine Learning I: Intrductin Kris De Brabanter kbrabant@iastate.edu Iwa State University Department f Statistics Department f Cmputer Science June 24, 2016 1/24 Outline
More informationEvaluating enterprise support: state of the art and future challenges. Dirk Czarnitzki KU Leuven, Belgium, and ZEW Mannheim, Germany
Evaluating enterprise supprt: state f the art and future challenges Dirk Czarnitzki KU Leuven, Belgium, and ZEW Mannheim, Germany Intrductin During the last decade, mircecnmetric ecnmetric cunterfactual
More informationTree Structured Classifier
Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients
More information, which yields. where z1. and z2
The Gaussian r Nrmal PDF, Page 1 The Gaussian r Nrmal Prbability Density Functin Authr: Jhn M Cimbala, Penn State University Latest revisin: 11 September 13 The Gaussian r Nrmal Prbability Density Functin
More informationThe standards are taught in the following sequence.
B L U E V A L L E Y D I S T R I C T C U R R I C U L U M MATHEMATICS Third Grade In grade 3, instructinal time shuld fcus n fur critical areas: (1) develping understanding f multiplicatin and divisin and
More information4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression
4th Indian Institute f Astrphysics - PennState Astrstatistics Schl July, 2013 Vainu Bappu Observatry, Kavalur Crrelatin and Regressin Rahul Ry Indian Statistical Institute, Delhi. Crrelatin Cnsider a tw
More informationAdmin. MDP Search Trees. Optimal Quantities. Reinforcement Learning
Admin Reinfrcement Learning Cntent adapted frm Berkeley CS188 MDP Search Trees Each MDP state prjects an expectimax-like search tree Optimal Quantities The value (utility) f a state s: V*(s) = expected
More informationData Mining: Concepts and Techniques. Classification and Prediction. Chapter February 8, 2007 CSE-4412: Data Mining 1
Data Mining: Cncepts and Techniques Classificatin and Predictin Chapter 6.4-6 February 8, 2007 CSE-4412: Data Mining 1 Chapter 6 Classificatin and Predictin 1. What is classificatin? What is predictin?
More informationData mining/machine learning large data sets. STA 302 or 442 (Applied Statistics) :, 1
Data mining/machine learning large data sets STA 302 r 442 (Applied Statistics) :, 1 Data mining/machine learning large data sets high dimensinal spaces STA 302 r 442 (Applied Statistics) :, 2 Data mining/machine
More informationWeathering. Title: Chemical and Mechanical Weathering. Grade Level: Subject/Content: Earth and Space Science
Weathering Title: Chemical and Mechanical Weathering Grade Level: 9-12 Subject/Cntent: Earth and Space Science Summary f Lessn: Students will test hw chemical and mechanical weathering can affect a rck
More informationCOMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d)
COMP 551 Applied Machine Learning Lecture 9: Supprt Vectr Machines (cnt d) Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Class web page: www.cs.mcgill.ca/~hvanh2/cmp551 Unless therwise
More informationIn SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw:
In SMV I IAML: Supprt Vectr Machines II Nigel Gddard Schl f Infrmatics Semester 1 We sa: Ma margin trick Gemetry f the margin and h t cmpute it Finding the ma margin hyperplane using a cnstrained ptimizatin
More informationCHAPTER 3 INEQUALITIES. Copyright -The Institute of Chartered Accountants of India
CHAPTER 3 INEQUALITIES Cpyright -The Institute f Chartered Accuntants f India INEQUALITIES LEARNING OBJECTIVES One f the widely used decisin making prblems, nwadays, is t decide n the ptimal mix f scarce
More information2004 AP CHEMISTRY FREE-RESPONSE QUESTIONS
2004 AP CHEMISTRY FREE-RESPONSE QUESTIONS 6. An electrchemical cell is cnstructed with an pen switch, as shwn in the diagram abve. A strip f Sn and a strip f an unknwn metal, X, are used as electrdes.
More informationPlease Stop Laughing at Me and Pay it Forward Final Writing Assignment
Kirk Please Stp Laughing at Me and Pay it Frward Final Writing Assignment Our fcus fr the past few mnths has been n bullying and hw we treat ther peple. We ve played sme games, read sme articles, read
More informationKinetic Model Completeness
5.68J/10.652J Spring 2003 Lecture Ntes Tuesday April 15, 2003 Kinetic Mdel Cmpleteness We say a chemical kinetic mdel is cmplete fr a particular reactin cnditin when it cntains all the species and reactins
More informationComputational modeling techniques
Cmputatinal mdeling techniques Lecture 2: Mdeling change. In Petre Department f IT, Åb Akademi http://users.ab.fi/ipetre/cmpmd/ Cntent f the lecture Basic paradigm f mdeling change Examples Linear dynamical
More informationAP Statistics Practice Test Unit Three Exploring Relationships Between Variables. Name Period Date
AP Statistics Practice Test Unit Three Explring Relatinships Between Variables Name Perid Date True r False: 1. Crrelatin and regressin require explanatry and respnse variables. 1. 2. Every least squares
More informationCHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came.
MATH 1342 Ch. 24 April 25 and 27, 2013 Page 1 f 5 CHAPTER 24: INFERENCE IN REGRESSION Chapters 4 and 5: Relatinships between tw quantitative variables. Be able t Make a graph (scatterplt) Summarize the
More informationMODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards:
MODULE FOUR This mdule addresses functins SC Academic Standards: EA-3.1 Classify a relatinship as being either a functin r nt a functin when given data as a table, set f rdered pairs, r graph. EA-3.2 Use
More informationName: Block: Date: Science 10: The Great Geyser Experiment A controlled experiment
Science 10: The Great Geyser Experiment A cntrlled experiment Yu will prduce a GEYSER by drpping Ments int a bttle f diet pp Sme questins t think abut are: What are yu ging t test? What are yu ging t measure?
More informationAP Statistics Notes Unit Two: The Normal Distributions
AP Statistics Ntes Unit Tw: The Nrmal Distributins Syllabus Objectives: 1.5 The student will summarize distributins f data measuring the psitin using quartiles, percentiles, and standardized scres (z-scres).
More informationIf (IV) is (increased, decreased, changed), then (DV) will (increase, decrease, change) because (reason based on prior research).
Science Fair Prject Set Up Instructins 1) Hypthesis Statement 2) Materials List 3) Prcedures 4) Safety Instructins 5) Data Table 1) Hw t write a HYPOTHESIS STATEMENT Use the fllwing frmat: If (IV) is (increased,
More informationTrigonometric Ratios Unit 5 Tentative TEST date
1 U n i t 5 11U Date: Name: Trignmetric Ratis Unit 5 Tentative TEST date Big idea/learning Gals In this unit yu will extend yur knwledge f SOH CAH TOA t wrk with btuse and reflex angles. This extensin
More informationYou need to be able to define the following terms and answer basic questions about them:
CS440/ECE448 Sectin Q Fall 2017 Midterm Review Yu need t be able t define the fllwing terms and answer basic questins abut them: Intr t AI, agents and envirnments Pssible definitins f AI, prs and cns f
More informationAP Physics Kinematic Wrap Up
AP Physics Kinematic Wrap Up S what d yu need t knw abut this mtin in tw-dimensin stuff t get a gd scre n the ld AP Physics Test? First ff, here are the equatins that yu ll have t wrk with: v v at x x
More informationLifting a Lion: Using Proportions
Overview Students will wrk in cperative grups t slve a real-wrd prblem by using the bk Hw D yu Lift a Lin? Using a ty lin and a lever, students will discver hw much wrk is needed t raise the ty lin. They
More information5 th grade Common Core Standards
5 th grade Cmmn Cre Standards In Grade 5, instructinal time shuld fcus n three critical areas: (1) develping fluency with additin and subtractin f fractins, and develping understanding f the multiplicatin
More information[COLLEGE ALGEBRA EXAM I REVIEW TOPICS] ( u s e t h i s t o m a k e s u r e y o u a r e r e a d y )
(Abut the final) [COLLEGE ALGEBRA EXAM I REVIEW TOPICS] ( u s e t h i s t m a k e s u r e y u a r e r e a d y ) The department writes the final exam s I dn't really knw what's n it and I can't very well
More informationNUMBERS, MATHEMATICS AND EQUATIONS
AUSTRALIAN CURRICULUM PHYSICS GETTING STARTED WITH PHYSICS NUMBERS, MATHEMATICS AND EQUATIONS An integral part t the understanding f ur physical wrld is the use f mathematical mdels which can be used t
More informationPart 3 Introduction to statistical classification techniques
Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms
More informationMathematics and Computer Sciences Department. o Work Experience, General. o Open Entry/Exit. Distance (Hybrid Online) for online supported courses
SECTION A - Curse Infrmatin 1. Curse ID: 2. Curse Title: 3. Divisin: 4. Department: 5. Subject: 6. Shrt Curse Title: 7. Effective Term:: MATH 70S Integrated Intermediate Algebra Natural Sciences Divisin
More informationStandard Title: Frequency Response and Frequency Bias Setting. Andrew Dressel Holly Hawkins Maureen Long Scott Miller
Template fr Quality Review f NERC Reliability Standard BAL-003-1 Frequency Respnse and Frequency Bias Setting Basic Infrmatin: Prject number: 2007-12 Standard number: BAL-003-1 Prject title: Frequency
More informationStatistical Learning. 2.1 What Is Statistical Learning?
2 Statistical Learning 2.1 What Is Statistical Learning? In rder t mtivate ur study f statistical learning, we begin with a simple example. Suppse that we are statistical cnsultants hired by a client t
More informationSupport-Vector Machines
Supprt-Vectr Machines Intrductin Supprt vectr machine is a linear machine with sme very nice prperties. Haykin chapter 6. See Alpaydin chapter 13 fr similar cntent. Nte: Part f this lecture drew material
More informationComprehensive Exam Guidelines Department of Chemical and Biomolecular Engineering, Ohio University
Cmprehensive Exam Guidelines Department f Chemical and Bimlecular Engineering, Ohi University Purpse In the Cmprehensive Exam, the student prepares an ral and a written research prpsal. The Cmprehensive
More informationHypothesis Tests for One Population Mean
Hypthesis Tests fr One Ppulatin Mean Chapter 9 Ala Abdelbaki Objective Objective: T estimate the value f ne ppulatin mean Inferential statistics using statistics in rder t estimate parameters We will be
More informationPhysics 2B Chapter 23 Notes - Faraday s Law & Inductors Spring 2018
Michael Faraday lived in the Lndn area frm 1791 t 1867. He was 29 years ld when Hand Oersted, in 1820, accidentally discvered that electric current creates magnetic field. Thrugh empirical bservatin and
More informationEASTERN ARIZONA COLLEGE Introduction to Statistics
EASTERN ARIZONA COLLEGE Intrductin t Statistics Curse Design 2014-2015 Curse Infrmatin Divisin Scial Sciences Curse Number PSY 220 Title Intrductin t Statistics Credits 3 Develped by Adam Stinchcmbe Lecture/Lab
More informationEric Klein and Ning Sa
Week 12. Statistical Appraches t Netwrks: p1 and p* Wasserman and Faust Chapter 15: Statistical Analysis f Single Relatinal Netwrks There are fur tasks in psitinal analysis: 1) Define Equivalence 2) Measure
More informationFall 2013 Physics 172 Recitation 3 Momentum and Springs
Fall 03 Physics 7 Recitatin 3 Mmentum and Springs Purpse: The purpse f this recitatin is t give yu experience wrking with mmentum and the mmentum update frmula. Readings: Chapter.3-.5 Learning Objectives:.3.
More informationOverview of Supervised Learning
2 Overview f Supervised Learning 2.1 Intrductin The first three examples described in Chapter 1 have several cmpnents in cmmn. Fr each there is a set f variables that might be dented as inputs, which are
More informationPipetting 101 Developed by BSU CityLab
Discver the Micrbes Within: The Wlbachia Prject Pipetting 101 Develped by BSU CityLab Clr Cmparisns Pipetting Exercise #1 STUDENT OBJECTIVES Students will be able t: Chse the crrect size micrpipette fr
More informationSubject description processes
Subject representatin 6.1.2. Subject descriptin prcesses Overview Fur majr prcesses r areas f practice fr representing subjects are classificatin, subject catalging, indexing, and abstracting. The prcesses
More informationProfessional Development. Implementing the NGSS: High School Physics
Prfessinal Develpment Implementing the NGSS: High Schl Physics This is a dem. The 30-min vide webinar is available in the full PD. Get it here. Tday s Learning Objectives NGSS key cncepts why this is different
More informationWe can see from the graph above that the intersection is, i.e., [ ).
MTH 111 Cllege Algebra Lecture Ntes July 2, 2014 Functin Arithmetic: With nt t much difficulty, we ntice that inputs f functins are numbers, and utputs f functins are numbers. S whatever we can d with
More informationT Algorithmic methods for data mining. Slide set 6: dimensionality reduction
T-61.5060 Algrithmic methds fr data mining Slide set 6: dimensinality reductin reading assignment LRU bk: 11.1 11.3 PCA tutrial in mycurses (ptinal) ptinal: An Elementary Prf f a Therem f Jhnsn and Lindenstrauss,
More informationNAME: Prof. Ruiz. 1. [5 points] What is the difference between simple random sampling and stratified random sampling?
CS4445 ata Mining and Kwledge iscery in atabases. B Term 2014 Exam 1 Nember 24, 2014 Prf. Carlina Ruiz epartment f Cmputer Science Wrcester Plytechnic Institute NAME: Prf. Ruiz Prblem I: Prblem II: Prblem
More information24 Multiple Eigenvectors; Latent Factor Analysis; Nearest Neighbors
Multiple Eigenvectrs; Latent Factr Analysis; Nearest Neighbrs 47 24 Multiple Eigenvectrs; Latent Factr Analysis; Nearest Neighbrs Clustering w/multiple Eigenvectrs [When we use the Fiedler vectr fr spectral
More informationDepartment of Economics, University of California, Davis Ecn 200C Micro Theory Professor Giacomo Bonanno. Insurance Markets
Department f Ecnmics, University f alifrnia, Davis Ecn 200 Micr Thery Prfessr Giacm Bnann Insurance Markets nsider an individual wh has an initial wealth f. ith sme prbability p he faces a lss f x (0
More informationLHS Mathematics Department Honors Pre-Calculus Final Exam 2002 Answers
LHS Mathematics Department Hnrs Pre-alculus Final Eam nswers Part Shrt Prblems The table at the right gives the ppulatin f Massachusetts ver the past several decades Using an epnential mdel, predict the
More informationIntroduction to Spacetime Geometry
Intrductin t Spacetime Gemetry Let s start with a review f a basic feature f Euclidean gemetry, the Pythagrean therem. In a twdimensinal crdinate system we can relate the length f a line segment t the
More informationThermodynamics Partial Outline of Topics
Thermdynamics Partial Outline f Tpics I. The secnd law f thermdynamics addresses the issue f spntaneity and invlves a functin called entrpy (S): If a prcess is spntaneus, then Suniverse > 0 (2 nd Law!)
More informationSTATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours
STATS216v Intrductin t Statistical Learning Stanfrd University, Summer 2016 Practice Final (Slutins) Duratin: 3 hurs Instructins: (This is a practice final and will nt be graded.) Remember the university
More informationDiscussion on Regularized Regression for Categorical Data (Tutz and Gertheiss)
Discussin n Regularized Regressin fr Categrical Data (Tutz and Gertheiss) Peter Bühlmann, Ruben Dezeure Seminar fr Statistics, Department f Mathematics, ETH Zürich, Switzerland Address fr crrespndence:
More informationA Matrix Representation of Panel Data
web Extensin 6 Appendix 6.A A Matrix Representatin f Panel Data Panel data mdels cme in tw brad varieties, distinct intercept DGPs and errr cmpnent DGPs. his appendix presents matrix algebra representatins
More informationPhysical Layer: Outline
18-: Intrductin t Telecmmunicatin Netwrks Lectures : Physical Layer Peter Steenkiste Spring 01 www.cs.cmu.edu/~prs/nets-ece Physical Layer: Outline Digital Representatin f Infrmatin Characterizatin f Cmmunicatin
More informationAgenda. What is Machine Learning? Learning Type of Learning: Supervised, Unsupervised and semi supervised Classification
Agenda Artificial Intelligence and its applicatins Lecture 6 Supervised Learning Prfessr Daniel Yeung danyeung@ieee.rg Dr. Patrick Chan patrickchan@ieee.rg Suth China University f Technlgy, China Learning
More informationWe say that y is a linear function of x if. Chapter 13: The Correlation Coefficient and the Regression Line
Chapter 13: The Crrelatin Cefficient and the Regressin Line We begin with a sme useful facts abut straight lines. Recall the x, y crdinate system, as pictured belw. 3 2 1 y = 2.5 y = 0.5x 3 2 1 1 2 3 1
More informationCity of Angels School Independent Study Los Angeles Unified School District
City f Angels Schl Independent Study Ls Angeles Unified Schl District INSTRUCTIONAL GUIDE Algebra 1B Curse ID #310302 (CCSS Versin- 06/15) This curse is the secnd semester f Algebra 1, fulfills ne half
More informationSection 5.8 Notes Page Exponential Growth and Decay Models; Newton s Law
Sectin 5.8 Ntes Page 1 5.8 Expnential Grwth and Decay Mdels; Newtn s Law There are many applicatins t expnential functins that we will fcus n in this sectin. First let s lk at the expnential mdel. Expnential
More informationMATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank
MATCHING TECHNIQUES Technical Track Sessin VI Emanuela Galass The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Emanuela Galass fr the purpse f this wrkshp When can we use
More informationWriting Guidelines. (Updated: November 25, 2009) Forwards
Writing Guidelines (Updated: Nvember 25, 2009) Frwards I have fund in my review f the manuscripts frm ur students and research assciates, as well as thse submitted t varius jurnals by thers that the majr
More informationDistributions, spatial statistics and a Bayesian perspective
Distributins, spatial statistics and a Bayesian perspective Dug Nychka Natinal Center fr Atmspheric Research Distributins and densities Cnditinal distributins and Bayes Thm Bivariate nrmal Spatial statistics
More information