Decision Tree Learning. Decision Tree Learning. Example. Decision Trees. Blue slides: Mitchell. Olive slides: Alpaydin Humidity
|
|
- Lizbeth Heath
- 6 years ago
- Views:
Transcription
1 Decision Tree Learning Decision Tree Learning Blue slides: Michell Oulook Olive slides: Alpaydin Huidiy Sunny Overcas Rain Wind High ral Srong Weak Learn o approxiae discree-valued arge funcions. Sep-y-sep decision aking: I can learn disjuncive expressions: Hypohesis space is copleely expressive, avoiding proles wih resriced hypohesis spaces. Inducive ias: sall rees over large rees. 2 Exaple Decision Trees Day Oulook Teperaure Huidiy Wind PlayTennis D Sunny Ho High Weak D2 Sunny Ho High Srong D3 Overcas Ho High Weak D4 Rain Mild High Weak D5 Rain Cool ral Weak D6 Rain Cool ral Srong A popular inducive inference algorih. Algorihs: ID3, ASSISTANT, C4.5, ec. Applicaions: edical diagnosis, assess credi risk of loan applicans, ec. D7 Overcas Cool ral Srong D8 Sunny Mild High Weak D9 Sunny Cool ral Weak D0 Rain Mild ral Weak D Sunny Mild ral Srong D2 Overcas Mild High Srong D3 Overcas Ho ral Weak D4 Rain Mild High Srong 3 4
2 Decision Trees: Operaion Oulook 3 Tree Uses des and Leaves Sunny Overcas Rain Huidiy Wind High ral Srong Weak Each insance holds ariue values. Insances are classified y filering he ariue values down he decision ree, down o a leaf which gives he final answer. Inernal nodes: ariue naes or ariue values. Branching occurs a ariue nodes. 5 4 Divide and Conquer Inernal decision nodes Univariae: Uses a single ariue, x i Nueric x i : Binary spli : x i > w Discree x i : n-way spli for n possile values Mulivariae: Uses all ariues, x Leaves Classificaion: Class laels, or proporions Regression: Nueric; r average, or local fi Learning is greedy; find he es spli recursively (Breian e al, 984; Quinlan, 986, 993) Decision Trees: Wha They Represen Sunny Oulook Overcas Rain Huidiy Wind High ral Srong Weak Each pah fro roo o leaf is a conjuncions of consrains on he ariue values. (Oulook = Sunny Huidiy = ral) (Oulook = Overcas) (Oulook = Rain W ind = W eak) 6
3 Appropriae Tasks for Decision Trees Consrucing Decision Trees fro Exaples Good a classificaion proles where: Insances are represened y ariue-value pairs. The arge funcion has discree oupu values. Disjuncive descripions ay e required. Given a se of exaples (raining se), oh posiive and negaive, he ask is o consruc a decision ree ha descries a concise decision pah. Using he resuling decision ree, we wan o classify new insances of exaples (eiher as yes or no). The raining daa ay conain errors. The raining daa ay conain issing ariue values. 7 8 Consrucing Decision Trees: Trivial Soluion Finding a Concise Decision Tree A rivial soluion is o explicily consruc pahs for each given exaple. In his case, you will ge a ree where he nuer of leaves is he sae as he nuer of raining exaples. The prole wih his approach is ha i is no ale o deal wih siuaions where, soe ariue values are issing or new kinds of siuaions arise. Consider ha soe ariues ay no coun uch oward he final classificaion. Meorizing all cases ay no e he es way. We wan o exrac a decision paern ha can descrie a large nuer of cases in a concise way. In ers of a decision ree, we wan o ake as few ess as possile efore reaching a decision, i.e. he deph of he ree should e shallow. 9 0
4 Finding a Concise Decision Tree (con d) Decision Tree Learning Algorih: ID3 Basic idea: pick up ariues ha can clearly separae posiive and negaive cases. These ariues are ore iporan han ohers: he final classificaion heavily depend on he value of hese ariues. Main loop:. A he es decision ariue for nex node 2. Assign A as decision ariue for node 3. For each value of A, creae new descendan of node 4. Sor raining exaples o leaf nodes 5. If raining exaples perfecly classified, Then STOP, Else ierae over new leaf nodes ID3 sands for Ieraive Dichooizer 3 2 Choosing he Bes Ariue A=? [29+,35-] [29+,35-] f A2=? f Choosing he Bes Ariue o Tes Firs Use Shannon s inforaion heory o choose he ariue ha give he axiu inforaion gain. A or A2? [2+,5-] [8+,30-] [8+,33-] [+,2-] Wih iniial and final nuer of posiive and negaive exaples ased on he ariue jus esed, we wan o decide which ariue is eer. Pick an ariue such ha he inforaion gain (or enropy reducion) is axiized. Enropy easures he average surprisal of evens. Less proale evens are ore surprising. How o quaniaively easure which one is eer? 3 4
5 Inforaion Theory (Inforal Inro) Given wo evens, H and T (Head and Tail): Inforaion Theory (Con d).0 Rare (uncerain) evens give ore surprise: H ore surprising han T if P (H) < P (T ) H ore uncerain han T if P (H) < P (T ) Enropy(S) 0.5 How o represen ore surprising, or ore uncerain? Surprise(H) > Surprise(T ) if P (H) < P (T ) P (H) > P (T ) ( ) ( ) log > log P (H) P (T ) log (P (H)) > log (P (T )) p + S is a saple of raining exaples p is he proporion of posiive exaples in S p is he proporion of negaive exaples in S Enropy easures he average uncerainy in S Enropy(S) p log 2 p p log 2 p 6 log (P (X)) as a easure of uncerainy. Uncerainy and Inforaion By perforing soe query, if you go fro sae S wih enropy E(S ) o sae S 2 wih enropy E(S 2 ), where E(S ) > E(S 2 ), your uncerainy has decreased. The aoun y which uncerainy decreased, i.e., E(S ) E(S 2 ), can e hough of as inforaion you gained (inforaion gain) hrough geing answers o your query. Enropy and Code Lengh Enropy(S) = expeced nuer of is needed o encode class ( or ) of randoly drawn eer of S (under he opial, shores-lengh code) Inforaion heory: opial lengh code assigns log 2 p is o essage having proailiy p. Encode wih shor sring for frequen essages (less surprising), and long sring for rarely occurring essages (ore surprising). So, expeced nuer of is o encode or of rando eer of S: p ( log 2 p ) + p ( log 2 p ) Enropy(S) p log 2 p p log 2 p 7 8
6 Enropy and Inforaion Gain Exaple Day Oulook Teperaure Huidiy Wind PlayTennis Enropy(S) = i C P i log 2 (P i ) Gain(S, A) = Enropy(S) C: caegories (classificaions) S: se of exaples A: a single ariue v V alues(a) S v : se of exaples where ariue A = v. X : cardinaliy of arirary se X. S v S Enropy(S v) D Sunny Ho High Weak D2 Sunny Ho High Srong D3 Overcas Ho High Weak D4 Rain Mild High Weak D5 Rain Cool ral Weak D6 Rain Cool ral Srong D7 Overcas Cool ral Srong D8 Sunny Mild High Weak D9 Sunny Cool ral Weak D0 Rain Mild ral Weak D Sunny Mild ral Srong D2 Overcas Mild High Srong D3 Overcas Ho ral Weak D4 Rain Mild High Srong Which ariue o es firs? 9 20 Choosing he Bes Ariue S: [9+,5-] Which ariue is he es classifier? E =0.940 E =0.940 Huidiy S: [9+,5-] Wind Parially Learned Tree {D, D2,..., D4} [9+,5 ] Oulook Sunny Overcas Rain High ral Weak Srong {D,D2,D8,D9,D} {D3,D7,D2,D3} {D4,D5,D6,D0,D4} [2+,3 ] [4+,0 ] [3+,2 ]?? [3+,4-] [6+,-] [6+,2-] [3+,3-] E =0.985 E =0.592 E =0.8 E =.00 Gain (S, Huidiy ) Gain (S, Wind) = (7/4) (7/4).592 =.5 = (8/4).8 - (6/4).0 =.048 +: # of posiive exaples; : # of negaive exaples Iniial enropy = 9 4 log log 5 4 = You can calculae he res. e: 0.0 log even hough log 0.0 is no defined. 2 Which ariue should e esed here? S sunny = {D,D2,D8,D9,D} Gain (Ssunny, Huidiy) =.970 (3/5) 0.0 (2/5) 0.0 =.970 Gain (S sunny, Teperaure) =.970 (2/5) 0.0 (2/5).0 (/5) 0.0 =.570 Gain (S sunny, Wind) =.970 (2/5).0 (3/5).98 =.09 Selec nex ariue, ased on he reaining exaples. 22
7 Hypohesis Space Search in ID3 Hypohesis Space Search in ID3 + + Hypohesis space is coplee! Targe funcion surely in here A A Oupus a single hypohesis (which one?) Can play 20 quesions... A2 A2 ack racking + + A A4 Local inia Saisically-ased search choices A each ranch, we ake a decision regarding a paricular ariue. Choice of an ariue direcs he search oward a cerain final hypohesis. Rous o noisy daa... Inducive ias: approx prefer shores ree Inducive Bias in ID3 Accuracy of Decision Trees ID3 is iased: ecause of he resricion on he hypohesis space, u Because of he preference for a paricular hypohesis. Such an inducive ias is called Occa s razor: The os likely hypohesis is he siples one ha is consisen wih all oservaions. Divide exaples ino raining and es ses. Train using he raining se. Measure accuracy of resuling decision ree on he es se
8 Issue: Overfiing Issue: ise Oulook Accuracy Huidiy Sunny Overcas Rain Wind On raining daa On es daa High ral Srong Weak Size of ree (nuer of nodes) Overfiing: Given a hypohesis space H, a hypohesis h H is said o overfi he raining daa if here exiss soe alernaive hypohesis h H such ha h is worse han h on he raining se u h is eer han h over he enire disriuion of insances. Can e due o noise in daa. 27 Overcoing Overfiing Sop early. Allow overfiing, hen pos-prune ree. Use separae se of exaples no used in raining o onior perforance on unoserved daa (validaion se). Use all availale daa, u perfor saisical es o esiae chance of iproving. Use explici easure of coplexiy of encoding, and pu a ound on ree size Wha if Oulook = Sunny, T ep = Ho, Huidiy = ral, W ind = Srong, P lay = was added as a raining exaple? Furher elaoraion of he aove ree ecoes necessary. The resuling ree will fi he raining daa plus he noise, u i ay perfor poorly on he rue insance disriuion. 28 Regression Trees Error a node : E x N Afer spliing: j E' x if xx : xreachesnode 0 oherwise N 2 r g x g 2 r g j j j j x g x r x if xxj : x reachesnode and ranch j 0 oherwis e j x r x j
9 Model Selecion in Trees 0 Pruning Trees Reove surees for eer generalizaion (decrease variance) Prepruning: Early sopping Pospruning: Grow he whole ree hen prune surees ha overfi on he pruning se Prepruning is faser, pospruning is ore accurae (requires a separae pruning se) 9 Rule Exracion fro Trees Learning Rules 2 C4.5Rules (Quinlan, 993) Rule inducion is siilar o ree inducion u ree inducion is readh-firs, rule inducion is deph-firs; one rule a a ie Rule se conains rules; rules are conjuncions of ers Rule covers an exaple if all ers of he rule evaluae o rue for he exaple Sequenial covering: Generae rules one a a ie unil all posiive exaples are covered IREP (Fürnkranz and Wider, 994), Ripper (Cohen, 995)
10 Oher Issues Coninuous-valued ariues: dynaically define new discree-valued ariues Muli-valued ariues wih large nuer of possile values: Use easures oher han inforaion gain. Training exaples wih issing ariue values: Assign os coon value, or assign wih he occurring frequency. Ariues wih differen cos/weighing: Scale using he cos. 30
Decision Tree Learning. Decision Tree Learning. Decision Trees. Decision Trees: Operation. Blue slides: Mitchell. Turquoise slides: Alpaydin Humidity
Decision Tree Learning Decision Tree Learning Blue slides: Michell Oulook Turquoise slides: Alpaydin Huidiy Sunny Overcas Rain ral Srong Learn o approxiae discree-valued arge funcions. Sep-y-sep decision
More informationDecision Tree Learning. Decision Tree Learning. Decision Trees. Decision Trees: Operation. Blue slides: Mitchell. Orange slides: Alpaydin Humidity
Decision Tree Learning Decision Tree Learning Blue slides: Michell Oulook Orange slides: Alpaydin Huidiy Sunny Overcas Rain ral Srong Learn o approxiae discree-valued arge funcions. Sep-by-sep decision
More informationEnsamble methods: Bagging and Boosting
Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par
More informationEnsamble methods: Boosting
Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room
More informationMachine Learning 2nd Edi7on
Lecture Slides for INTRODUCTION TO Machine Learning 2nd Edi7on CHAPTER 9: Decision Trees ETHEM ALPAYDIN The MIT Press, 2010 Edited and expanded for CS 4641 by Chris Simpkins alpaydin@boun.edu.tr h1p://www.cmpe.boun.edu.tr/~ethem/i2ml2e
More informationINTRODUCTION TO MACHINE LEARNING 3RD EDITION
ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class
More informationLearning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power
Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.
More information1 Widrow-Hoff Algorithm
COS 511: heoreical Machine Learning Lecurer: Rob Schapire Lecure # 18 Scribe: Shaoqing Yang April 10, 014 1 Widrow-Hoff Algorih Firs le s review he Widrow-Hoff algorih ha was covered fro las lecure: Algorih
More informationLearning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power
Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.
More informationLecture 18 GMM:IV, Nonlinear Models
Lecure 8 :IV, Nonlinear Models Le Z, be an rx funcion of a kx paraeer vecor, r > k, and a rando vecor Z, such ha he r populaion oen condiions also called esiain equaions EZ, hold for all, where is he rue
More informationIntroduction to Numerical Analysis. In this lesson you will be taken through a pair of techniques that will be used to solve the equations of.
Inroducion o Nuerical Analysis oion In his lesson you will be aen hrough a pair of echniques ha will be used o solve he equaions of and v dx d a F d for siuaions in which F is well nown, and he iniial
More informationAsymptotic Equipartition Property - Seminar 3, part 1
Asympoic Equipariion Propery - Seminar 3, par 1 Ocober 22, 2013 Problem 1 (Calculaion of ypical se) To clarify he noion of a ypical se A (n) ε and he smalles se of high probabiliy B (n), we will calculae
More informationLecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.
Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in
More information13.3 Term structure models
13.3 Term srucure models 13.3.1 Expecaions hypohesis model - Simples "model" a) shor rae b) expecaions o ge oher prices Resul: y () = 1 h +1 δ = φ( δ)+ε +1 f () = E (y +1) (1) =δ + φ( δ) f (3) = E (y +)
More informationA First Course on Kinetics and Reaction Engineering. Class 19 on Unit 18
A Firs ourse on Kineics and Reacion Engineering lass 19 on Uni 18 Par I - hemical Reacions Par II - hemical Reacion Kineics Where We re Going Par III - hemical Reacion Engineering A. Ideal Reacors B. Perfecly
More informationDecision Tree Learning
0. Decision Tree Learning Based on Machine Learning, T. Mitchell, McGRAW Hill, 1997, ch. 3 Acknowledgement: The present slides are an adaptation of slides drawn by T. Mitchell PLAN 1. Concept learning:
More informationViscous Damping Summary Sheet No Damping Case: Damped behaviour depends on the relative size of ω o and b/2m 3 Cases: 1.
Viscous Daping: && + & + ω Viscous Daping Suary Shee No Daping Case: & + ω solve A ( ω + α ) Daped ehaviour depends on he relaive size of ω o and / 3 Cases:. Criical Daping Wee 5 Lecure solve sae BC s
More informationLecture 28: Single Stage Frequency response. Context
Lecure 28: Single Sage Frequency response Prof J. S. Sih Conex In oday s lecure, we will coninue o look a he frequency response of single sage aplifiers, saring wih a ore coplee discussion of he CS aplifier,
More informationReading. Lecture 28: Single Stage Frequency response. Lecture Outline. Context
Reading Lecure 28: Single Sage Frequency response Prof J. S. Sih Reading: We are discussing he frequency response of single sage aplifiers, which isn reaed in he ex unil afer uli-sae aplifiers (beginning
More informationT L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB
Elecronic Companion EC.1. Proofs of Technical Lemmas and Theorems LEMMA 1. Le C(RB) be he oal cos incurred by he RB policy. Then we have, T L E[C(RB)] 3 E[Z RB ]. (EC.1) Proof of Lemma 1. Using he marginal
More informationACE 564 Spring Lecture 7. Extensions of The Multiple Regression Model: Dummy Independent Variables. by Professor Scott H.
ACE 564 Spring 2006 Lecure 7 Exensions of The Muliple Regression Model: Dumm Independen Variables b Professor Sco H. Irwin Readings: Griffihs, Hill and Judge. "Dumm Variables and Varing Coefficien Models
More informationLearning Decision Trees
Learning Decision Trees Machine Learning Spring 2018 1 This lecture: Learning Decision Trees 1. Representation: What are decision trees? 2. Algorithm: Learning decision trees The ID3 algorithm: A greedy
More informationMapping in Dynamic Environments
Mapping in Dynaic Environens Wolfra Burgard Universiy of Freiburg, Gerany Mapping is a Key Technology for Mobile Robos Robos can robusly navigae when hey have a ap. Robos have been shown o being able o
More informationY. Xiang, Learning Bayesian Networks 1
Learning Bayesian Neworks Objecives Acquisiion of BNs Technical conex of BN learning Crierion of sound srucure learning BN srucure learning in 2 seps BN CPT esimaion Reference R.E. Neapolian: Learning
More informationThus the force is proportional but opposite to the displacement away from equilibrium.
Chaper 3 : Siple Haronic Moion Hooe s law saes ha he force (F) eered by an ideal spring is proporional o is elongaion l F= l where is he spring consan. Consider a ass hanging on a he spring. In equilibriu
More informationBoosting MIT Course Notes Cynthia Rudin
Credi: Freund, Schapire, Daubechies Boosing MIT 5.097 Course Noes Cynhia Rudin Boosing sared wih a quesion of Michael Kearns, abou wheher a weak learning algorih can be ade ino a srong learning algorih.
More informationIntroduction. Decision Tree Learning. Outline. Decision Tree 9/7/2017. Decision Tree Definition
Introduction Decision Tree Learning Practical methods for inductive inference Approximating discrete-valued functions Robust to noisy data and capable of learning disjunctive expression ID3 earch a completely
More informationMath 10B: Mock Mid II. April 13, 2016
Name: Soluions Mah 10B: Mock Mid II April 13, 016 1. ( poins) Sae, wih jusificaion, wheher he following saemens are rue or false. (a) If a 3 3 marix A saisfies A 3 A = 0, hen i canno be inverible. True.
More informationNon-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important
on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LTU, decision
More informationEcon107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8)
I. Definiions and Problems A. Perfec Mulicollineariy Econ7 Applied Economerics Topic 7: Mulicollineariy (Sudenmund, Chaper 8) Definiion: Perfec mulicollineariy exiss in a following K-variable regression
More informationOutline. Training Examples for EnjoySport. 2 lecture slides for textbook Machine Learning, c Tom M. Mitchell, McGraw Hill, 1997
Outline Training Examples for EnjoySport Learning from examples General-to-specific ordering over hypotheses [read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] Version spaces and candidate elimination
More informationLearning Decision Trees
Learning Decision Trees Machine Learning Fall 2018 Some slides from Tom Mitchell, Dan Roth and others 1 Key issues in machine learning Modeling How to formulate your problem as a machine learning problem?
More informationSome Ramsey results for the n-cube
Some Ramsey resuls for he n-cube Ron Graham Universiy of California, San Diego Jozsef Solymosi Universiy of Briish Columbia, Vancouver, Canada Absrac In his noe we esablish a Ramsey-ype resul for cerain
More informationComparing Means: t-tests for One Sample & Two Related Samples
Comparing Means: -Tess for One Sample & Two Relaed Samples Using he z-tes: Assumpions -Tess for One Sample & Two Relaed Samples The z-es (of a sample mean agains a populaion mean) is based on he assumpion
More informationLecture 3: Decision Trees
Lecture 3: Decision Trees Cognitive Systems II - Machine Learning SS 2005 Part I: Basic Approaches of Concept Learning ID3, Information Gain, Overfitting, Pruning Lecture 3: Decision Trees p. Decision
More informationLecture 33: November 29
36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure
More informationPhys1112: DC and RC circuits
Name: Group Members: Dae: TA s Name: Phys1112: DC and RC circuis Objecives: 1. To undersand curren and volage characerisics of a DC RC discharging circui. 2. To undersand he effec of he RC ime consan.
More informationCHAPTER 12 DIRECT CURRENT CIRCUITS
CHAPTER 12 DIRECT CURRENT CIUITS DIRECT CURRENT CIUITS 257 12.1 RESISTORS IN SERIES AND IN PARALLEL When wo resisors are conneced ogeher as shown in Figure 12.1 we said ha hey are conneced in series. As
More informationDecision-Tree Learning. Chapter 3: Decision Tree Learning. Classification Learning. Decision Tree for PlayTennis
Decision-Tree Learning Chapter 3: Decision Tree Learning CS 536: Machine Learning Littman (Wu, TA) [read Chapter 3] [some of Chapter 2 might help ] [recommended exercises 3.1, 3.2] Decision tree representation
More informationLecture 12: Multiple Hypothesis Testing
ECE 830 Fall 00 Saisical Signal Processing insrucor: R. Nowak, scribe: Xinjue Yu Lecure : Muliple Hypohesis Tesing Inroducion In many applicaions we consider muliple hypohesis es a he same ime. Example
More informationMore Digital Logic. t p output. Low-to-high and high-to-low transitions could have different t p. V in (t)
EECS 4 Spring 23 Lecure 2 EECS 4 Spring 23 Lecure 2 More igial Logic Gae delay and signal propagaion Clocked circui elemens (flip-flop) Wriing a word o memory Simplifying digial circuis: Karnaugh maps
More informationRandom Walk with Anti-Correlated Steps
Random Walk wih Ani-Correlaed Seps John Noga Dirk Wagner 2 Absrac We conjecure he expeced value of random walks wih ani-correlaed seps o be exacly. We suppor his conjecure wih 2 plausibiliy argumens and
More informationExpert Advice for Amateurs
Exper Advice for Amaeurs Ernes K. Lai Online Appendix - Exisence of Equilibria The analysis in his secion is performed under more general payoff funcions. Wihou aking an explici form, he payoffs of he
More informationCHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK
175 CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 10.1 INTRODUCTION Amongs he research work performed, he bes resuls of experimenal work are validaed wih Arificial Neural Nework. From he
More informationLongest Common Prefixes
Longes Common Prefixes The sandard ordering for srings is he lexicographical order. I is induced by an order over he alphabe. We will use he same symbols (,
More informationWEEK-3 Recitation PHYS 131. of the projectile s velocity remains constant throughout the motion, since the acceleration a x
WEEK-3 Reciaion PHYS 131 Ch. 3: FOC 1, 3, 4, 6, 14. Problems 9, 37, 41 & 71 and Ch. 4: FOC 1, 3, 5, 8. Problems 3, 5 & 16. Feb 8, 018 Ch. 3: FOC 1, 3, 4, 6, 14. 1. (a) The horizonal componen of he projecile
More informationHomework 2 Solutions
Mah 308 Differenial Equaions Fall 2002 & 2. See he las page. Hoework 2 Soluions 3a). Newon s secon law of oion says ha a = F, an we know a =, so we have = F. One par of he force is graviy, g. However,
More informationBiol. 356 Lab 8. Mortality, Recruitment, and Migration Rates
Biol. 356 Lab 8. Moraliy, Recruimen, and Migraion Raes (modified from Cox, 00, General Ecology Lab Manual, McGraw Hill) Las week we esimaed populaion size hrough several mehods. One assumpion of all hese
More informationArticle from. Predictive Analytics and Futurism. July 2016 Issue 13
Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning
More informationMath 2142 Exam 1 Review Problems. x 2 + f (0) 3! for the 3rd Taylor polynomial at x = 0. To calculate the various quantities:
Mah 4 Eam Review Problems Problem. Calculae he 3rd Taylor polynomial for arcsin a =. Soluion. Le f() = arcsin. For his problem, we use he formula f() + f () + f ()! + f () 3! for he 3rd Taylor polynomial
More informationDecision Tree Learning Mitchell, Chapter 3. CptS 570 Machine Learning School of EECS Washington State University
Decision Tree Learning Mitchell, Chapter 3 CptS 570 Machine Learning School of EECS Washington State University Outline Decision tree representation ID3 learning algorithm Entropy and information gain
More informationMATHEMATICAL DESCRIPTION OF THEORETICAL METHODS OF RESERVE ECONOMY OF CONSIGNMENT STORES
MAHEMAICAL DESCIPION OF HEOEICAL MEHODS OF ESEVE ECONOMY OF CONSIGNMEN SOES Péer elek, József Cselényi, György Demeer Universiy of Miskolc, Deparmen of Maerials Handling and Logisics Absrac: Opimizaion
More informationLecture 3: Exponential Smoothing
NATCOR: Forecasing & Predicive Analyics Lecure 3: Exponenial Smoohing John Boylan Lancaser Cenre for Forecasing Deparmen of Managemen Science Mehods and Models Forecasing Mehod A (numerical) procedure
More informationTime series Decomposition method
Time series Decomposiion mehod A ime series is described using a mulifacor model such as = f (rend, cyclical, seasonal, error) = f (T, C, S, e) Long- Iner-mediaed Seasonal Irregular erm erm effec, effec,
More informationFinish reading Chapter 2 of Spivak, rereading earlier sections as necessary. handout and fill in some missing details!
MAT 257, Handou 6: Ocober 7-2, 20. I. Assignmen. Finish reading Chaper 2 of Spiva, rereading earlier secions as necessary. handou and fill in some missing deails! II. Higher derivaives. Also, read his
More informationThe Arcsine Distribution
The Arcsine Disribuion Chris H. Rycrof Ocober 6, 006 A common heme of he class has been ha he saisics of single walker are ofen very differen from hose of an ensemble of walkers. On he firs homework, we
More informationReading from Young & Freedman: For this topic, read sections 25.4 & 25.5, the introduction to chapter 26 and sections 26.1 to 26.2 & 26.4.
PHY1 Elecriciy Topic 7 (Lecures 1 & 11) Elecric Circuis n his opic, we will cover: 1) Elecromoive Force (EMF) ) Series and parallel resisor combinaions 3) Kirchhoff s rules for circuis 4) Time dependence
More informationLearning Objectives: Practice designing and simulating digital circuits including flip flops Experience state machine design procedure
Lab 4: Synchronous Sae Machine Design Summary: Design and implemen synchronous sae machine circuis and es hem wih simulaions in Cadence Viruoso. Learning Objecives: Pracice designing and simulaing digial
More informationCS6375: Machine Learning Gautam Kunapuli. Decision Trees
Gautam Kunapuli Example: Restaurant Recommendation Example: Develop a model to recommend restaurants to users depending on their past dining experiences. Here, the features are cost (x ) and the user s
More informationHidden Markov Models. Adapted from. Dr Catherine Sweeney-Reed s slides
Hidden Markov Models Adaped from Dr Caherine Sweeney-Reed s slides Summary Inroducion Descripion Cenral in HMM modelling Exensions Demonsraion Specificaion of an HMM Descripion N - number of saes Q = {q
More informationIntroduction to Mechanical Vibrations and Structural Dynamics
Inroducion o Mechanical Viraions and Srucural Dynaics The one seeser schedule :. Viraion - classificaion. ree undaped single DO iraion, equaion of oion, soluion, inegraional consans, iniial condiions..
More information23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes
Represening Periodic Funcions by Fourier Series 3. Inroducion In his Secion we show how a periodic funcion can be expressed as a series of sines and cosines. We begin by obaining some sandard inegrals
More informationChapter 3: Decision Tree Learning
Chapter 3: Decision Tree Learning CS 536: Machine Learning Littman (Wu, TA) Administration Books? New web page: http://www.cs.rutgers.edu/~mlittman/courses/ml03/ schedule lecture notes assignment info.
More informationCMU-Q Lecture 3: Search algorithms: Informed. Teacher: Gianni A. Di Caro
CMU-Q 5-38 Lecure 3: Search algorihms: Informed Teacher: Gianni A. Di Caro UNINFORMED VS. INFORMED SEARCH Sraegy How desirable is o be in a cerain inermediae sae for he sake of (effecively) reaching a
More informationSpring Ammar Abu-Hudrouss Islamic University Gaza
Chaper 7 Reed-Solomon Code Spring 9 Ammar Abu-Hudrouss Islamic Universiy Gaza ١ Inroducion A Reed Solomon code is a special case of a BCH code in which he lengh of he code is one less han he size of he
More informationNotes on Kalman Filtering
Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren
More informationMOMENTUM CONSERVATION LAW
1 AAST/AEDT AP PHYSICS B: Impulse and Momenum Le us run an experimen: The ball is moving wih a velociy of V o and a force of F is applied on i for he ime inerval of. As he resul he ball s velociy changes
More informationOnline Convex Optimization Example And Follow-The-Leader
CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion
More informationNon-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important
on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LDA, logisic
More informationSection 3.5 Nonhomogeneous Equations; Method of Undetermined Coefficients
Secion 3.5 Nonhomogeneous Equaions; Mehod of Undeermined Coefficiens Key Terms/Ideas: Linear Differenial operaor Nonlinear operaor Second order homogeneous DE Second order nonhomogeneous DE Soluion o homogeneous
More informationStability. Coefficients may change over time. Evolution of the economy Policy changes
Sabiliy Coefficiens may change over ime Evoluion of he economy Policy changes Time Varying Parameers y = α + x β + Coefficiens depend on he ime period If he coefficiens vary randomly and are unpredicable,
More informationAn introduction to the theory of SDDP algorithm
An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking
More informationNotes for Lecture 17-18
U.C. Berkeley CS278: Compuaional Complexiy Handou N7-8 Professor Luca Trevisan April 3-8, 2008 Noes for Lecure 7-8 In hese wo lecures we prove he firs half of he PCP Theorem, he Amplificaion Lemma, up
More informationA Hop Constrained Min-Sum Arborescence with Outage Costs
A Hop Consrained Min-Sum Arborescence wih Ouage Coss Rakesh Kawara Minnesoa Sae Universiy, Mankao, MN 56001 Email: Kawara@mnsu.edu Absrac The hop consrained min-sum arborescence wih ouage coss problem
More informationb denotes trend at time point t and it is sum of two
Inernaional Conference on Innovaive Applicaions in Engineering and Inforaion echnology(iciaei207) Inernaional Journal of Advanced Scienific echnologies,engineering and Manageen Sciences (IJASEMSISSN: 2454356X)
More information3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon
3..3 INRODUCION O DYNAMIC OPIMIZAION: DISCREE IME PROBLEMS A. he Hamilonian and Firs-Order Condiions in a Finie ime Horizon Define a new funcion, he Hamilonian funcion, H. H he change in he oal value of
More informationForecasting optimally
I) ile: Forecas Evaluaion II) Conens: Evaluaing forecass, properies of opimal forecass, esing properies of opimal forecass, saisical comparison of forecas accuracy III) Documenaion: - Diebold, Francis
More informationFourier Series & The Fourier Transform. Joseph Fourier, our hero. Lord Kelvin on Fourier s theorem. What do we want from the Fourier Transform?
ourier Series & The ourier Transfor Wha is he ourier Transfor? Wha do we wan fro he ourier Transfor? We desire a easure of he frequencies presen in a wave. This will lead o a definiion of he er, he specru.
More informationConservation of Momentum. The purpose of this experiment is to verify the conservation of momentum in two dimensions.
Conseraion of Moenu Purose The urose of his exerien is o erify he conseraion of oenu in wo diensions. Inroducion and Theory The oenu of a body ( ) is defined as he roduc of is ass () and elociy ( ): When
More information8. Basic RL and RC Circuits
8. Basic L and C Circuis This chaper deals wih he soluions of he responses of L and C circuis The analysis of C and L circuis leads o a linear differenial equaion This chaper covers he following opics
More informationThe equation to any straight line can be expressed in the form:
Sring Graphs Par 1 Answers 1 TI-Nspire Invesigaion Suden min Aims Deermine a series of equaions of sraigh lines o form a paern similar o ha formed by he cables on he Jerusalem Chords Bridge. Deermine he
More informationVehicle Arrival Models : Headway
Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where
More informationEXERCISES FOR SECTION 1.5
1.5 Exisence and Uniqueness of Soluions 43 20. 1 v c 21. 1 v c 1 2 4 6 8 10 1 2 2 4 6 8 10 Graph of approximae soluion obained using Euler s mehod wih = 0.1. Graph of approximae soluion obained using Euler
More informationEchocardiography Project and Finite Fourier Series
Echocardiography Projec and Finie Fourier Series 1 U M An echocardiagram is a plo of how a porion of he hear moves as he funcion of ime over he one or more hearbea cycles If he hearbea repeas iself every
More informationOperating Systems Exercise 3
Operaing Sysems SS 00 Universiy of Zurich Operaing Sysems Exercise 3 00-06-4 Dominique Emery, s97-60-056 Thomas Bocek, s99-706-39 Philip Iezzi, s99-74-354 Florian Caflisch, s96-90-55 Table page Table page
More informationAdministration. Chapter 3: Decision Tree Learning (part 2) Measuring Entropy. Entropy Function
Administration Chapter 3: Decision Tree Learning (part 2) Book on reserve in the math library. Questions? CS 536: Machine Learning Littman (Wu, TA) Measuring Entropy Entropy Function S is a sample of training
More informationProblem Set 5. Graduate Macro II, Spring 2017 The University of Notre Dame Professor Sims
Problem Se 5 Graduae Macro II, Spring 2017 The Universiy of Nore Dame Professor Sims Insrucions: You may consul wih oher members of he class, bu please make sure o urn in your own work. Where applicable,
More informationPattern Classification (VI) 杜俊
Paern lassificaion VI 杜俊 jundu@usc.edu.cn Ouline Bayesian Decision Theory How o make he oimal decision? Maximum a oserior MAP decision rule Generaive Models Join disribuion of observaion and label sequences
More informationEXPONENTIAL PROBABILITY DISTRIBUTION
MTH/STA 56 EXPONENTIAL PROBABILITY DISTRIBUTION As discussed in Exaple (of Secion of Unifor Probabili Disribuion), in a Poisson process, evens are occurring independenl a rando and a a unifor rae per uni
More informationGraphical Event Models and Causal Event Models. Chris Meek Microsoft Research
Graphical Even Models and Causal Even Models Chris Meek Microsof Research Graphical Models Defines a join disribuion P X over a se of variables X = X 1,, X n A graphical model M =< G, Θ > G =< X, E > is
More informationCptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1
CpS 570 Machine Learning School of EECS Washingon Sae Universiy CpS 570 - Machine Learning 1 Form of underlying disribuions unknown Bu sill wan o perform classificaion and regression Semi-parameric esimaion
More information11!Hí MATHEMATICS : ERDŐS AND ULAM PROC. N. A. S. of decomposiion, properly speaking) conradics he possibiliy of defining a counably addiive real-valu
ON EQUATIONS WITH SETS AS UNKNOWNS BY PAUL ERDŐS AND S. ULAM DEPARTMENT OF MATHEMATICS, UNIVERSITY OF COLORADO, BOULDER Communicaed May 27, 1968 We shall presen here a number of resuls in se heory concerning
More informationDynamic Econometric Models: Y t = + 0 X t + 1 X t X t k X t-k + e t. A. Autoregressive Model:
Dynamic Economeric Models: A. Auoregressive Model: Y = + 0 X 1 Y -1 + 2 Y -2 + k Y -k + e (Wih lagged dependen variable(s) on he RHS) B. Disribued-lag Model: Y = + 0 X + 1 X -1 + 2 X -2 + + k X -k + e
More informationLicenciatura de ADE y Licenciatura conjunta Derecho y ADE. Hoja de ejercicios 2 PARTE A
Licenciaura de ADE y Licenciaura conjuna Derecho y ADE Hoja de ejercicios PARTE A 1. Consider he following models Δy = 0.8 + ε (1 + 0.8L) Δ 1 y = ε where ε and ε are independen whie noise processes. In
More informationKINEMATICS IN ONE DIMENSION
KINEMATICS IN ONE DIMENSION PREVIEW Kinemaics is he sudy of how hings move how far (disance and displacemen), how fas (speed and velociy), and how fas ha how fas changes (acceleraion). We say ha an objec
More informationSolutions to the Exam Digital Communications I given on the 11th of June = 111 and g 2. c 2
Soluions o he Exam Digial Communicaions I given on he 11h of June 2007 Quesion 1 (14p) a) (2p) If X and Y are independen Gaussian variables, hen E [ XY ]=0 always. (Answer wih RUE or FALSE) ANSWER: False.
More information1. Calibration factor
Annex_C_MUBDandP_eng_.doc, p. of pages Annex C: Measureen uncerainy of he oal heigh of profile of a deph-seing sandard ih he sandard deviaion of he groove deph as opography er In his exaple, he uncerainy
More informationDeep Learning: Theory, Techniques & Applications - Recurrent Neural Networks -
Deep Learning: Theory, Techniques & Applicaions - Recurren Neural Neworks - Prof. Maeo Maeucci maeo.maeucci@polimi.i Deparmen of Elecronics, Informaion and Bioengineering Arificial Inelligence and Roboics
More informationSTA 114: Statistics. Notes 2. Statistical Models and the Likelihood Function
STA 114: Saisics Noes 2. Saisical Models and he Likelihood Funcion Describing Daa & Saisical Models A physicis has a heory ha makes a precise predicion of wha s o be observed in daa. If he daa doesn mach
More informationDecision Tree Learning
Decision Tree Learning Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University References: 1. Machine Learning, Chapter 3 2. Data Mining: Concepts, Models,
More information