Outline. Intro. to Machine Learning. Outline. Course Info. Course Info.: People, References, Resources
|
|
- Sharyl Hopkins
- 6 years ago
- Views:
Transcription
1 Course Info. Machine Learning Curve Fiing Decision Theory Probabiliy Theory Course Info. Machine Learning Curve Fiing Decision Theory Probabiliy Theory Ouline Course Info.: People, References, Resources Inroducion o Machine Learning Machine Learning: Wha, Why, and How? Bishop PRML Ch. Curve Fiing: (e.g.) Regression and Model Selecion Alireza Ghane Decision Theory: ML, Loss Funcion, MAP Probabiliy Theory: (e.g.) Probabiliies and Parameer Esimaion Inro. o Machine Learning Course Info. Machine Learning Curve Fiing Decision Theory Probabiliy Theory Course Info. Machine Learning Ouline Alireza Ghane / Torsen Mo ller Curve Fiing Decision Theory Probabiliy Theory Course Info. Course Info.: People, References, Resources Machine Learning: Wha, Why, and How? Curve Fiing: (e.g.) Regression and Model Selecion Dr. Torsen Mo ller Decision Theory: ML, Loss Funcion, MAP Probabiliy Theory: (e.g.) Probabiliies and Parameer Esimaion Home: hp://vda.univie.ac.a/teaching/ml/5s Inro. o Machine Learning Discussions: hps://moodle.univie.ac.a/ Alireza Ghane / Torsen Mo ller 2 Inro. o Machine Learning Alireza Ghane / Torsen Mo ller Alireza Ghane 3
2 Regisraion Course ma paricipans: 25 Course Regisered: 6 Number of Seas: 48 Ecess: 3 Sign your name on he shee If you miss he firs wo sessions, you will be auomaically SIGNED OFF he course! References Main Tebook: Paern Recogniion and Machine Learning, Chrisopher M. Bishop, Springer 26. Oher Useful Resources: The Elemens of Saisical Learning, Trevor Hasie, Rober Tibshirani, and Jerome Friedman. Machine Learning, Tom Michel. Paern Classificaion (2nd ed.), Richard O. Duda, Peer E. Har, and David G. Sork. Machine Learning, An Algorihmic Perspecive, Sephen Marsland. The Top Ten Algorihms in Daa Mining, X. Wu, V. Kumar. Learning from Daa, Cherkassky-Mulier. Online Courses: Andrew Ng: hp://ml-class.org/ Inro. o Machine Learning Alireza Ghane / Torsen Möller 4 Inro. o Machine Learning Alireza Ghane / Torsen Möller 5 Grading: Assignmens / Labs ( 5% ) 5 assignmens, % each Final Eam ( 4 % ) Class Feedback ( % ) Assignmen lae policy 5 grace days for all assignmens ogeher afer he grace days, 25% penaly for each day Grading Course Topics We will cover echniques in he sandard ML oolki maimum likelihood, regularizaion, suppor vecor machines (SVM), Fisher s linear discriminan (LDA), boosing, principal componens analysis (PCA), Markov random fields (MRF), neural neworks, graphical models, belief propagaion, epecaion-maimizaion (EM), miure models, miures of epers (MoE), hidden Markov models (HMM), paricle filers, Markov Chain Mone Carlo (MCMC), Gibbs sampling,... Programming wih: MATLAB (licensed): hp://de.mahworks.com/ Ocave (free): hps:// Inro. o Machine Learning Alireza Ghane / Torsen Möller 6 Inro. o Machine Learning Alireza Ghane / Torsen Möller 7
3 Calculus: Background E = mc 2 E c = 2mc Linear algebra (PRML Appendi C): Au i = λ i u i ; Probabiliy (PRML Ch..2): p(x) = p(x, Y ); p() = Y (T a) = a p(, y)dy; E [f] = I will be possible o refresh, bu if you ve never seen hese before his course will be very difficul. p()f()d Inro. o Machine Learning Alireza Ghane / Torsen Möller 8 Ouline Course Info.: People, References, Resources Machine Learning: Wha, Why, and How? Curve Fiing: (e.g.) Regression and Model Selecion Decision Theory: ML, Loss Funcion, MAP Probabiliy Theory: (e.g.) Probabiliies and Parameer Esimaion Inro. o Machine Learning Alireza Ghane / Greg Mori 9 Wha is Machine Learning (ML)? Why ML? Algorihms ha auomaically improve performance hrough eperience Ofen his means define a model by hand, and use daa o fi is parameers The real world is comple difficul o hand-craf soluions. ML is he preferred framework for applicaions in many fields: Compuer Vision Naural Language Processing, Speech Recogniion Roboics... Inro. o Machine Learning Alireza Ghane / Greg Mori Inro. o Machine Learning Alireza Ghane / Greg Mori
4 Hand-wrien Digi Recogniion 58 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 24, NO. 24, APRIL 22 Hand-wrien Digi Recogniion 58 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 24, NO. 24, APRIL 22 i = i = (,,,,,,,,, ) Represen inpu image as a vecor i R 784. Fig. 8. All of he misclassified MNIST es digis using our mehod 63 ou of,). The e above each digi indicaes he eample number followed by he rue label and he assigned label. Belongie e al. PAMI 22 sraighforward sum of squared differences SSD). SSD error rae wih an average of only four wo-dimensional performs very well on his easy daabase due o he lack of views for each hree-dimensional objec, hanks o he Difficul o hand-craf rules abou digis variaion in lighing [24] PCA jus makes i faser). fleibiliy provided by he maching algorihm. The prooype selecion algorihm is illusraed in Fig MPEG-7 Shape Silhouee Daabase As seen, views are allocaed mainly for more comple caegories wih high wihin class variabiliy. The curve Our ne eperimen involves he MPEG-7 shape silhouee Inro. o Machine Learning Alireza Ghane / Greg Mori 2 marked SC-proo in Fig. 9 shows he improved classificaion daabase, specifically Core Eperimen CE-Shape- par B, performance using his prooype selecion sraegy insead which measures performance of similariy-based rerieval of equally-spaced views. Noe ha we obain a 2.4 percen [25]. The daabase consiss of,4 images: 7 shape caegories, 2 images per caegory. The performance is measured using he so-called ªbullseye es,º in which each Suppose we have a arge vecor i This is supervised learning Discree, finie label se: perhaps i {, }, a classificaion problem Given a raining se {(, ),..., ( N, N )}, learning problem is o consruc a good funcion y() from hese. y : R 784 R Inro. o Machine Learning Alireza Ghane / Greg Mori 3 Fig. 9. 3D objec recogniion using he COIL-2 daa se. Comparison of es se error for SSD, Shape Disance SD), and Shape Disance wih k-medoids prooypes SD-proo) versus number of prooype views. For SSD and SD, we varied he number of prooypes uniformly for all objecs. For SD-proo, he number of prooypes per objec depended on he wihin-objec variaion as well as he beween-objec similariy. Fig.. Prooype views seleced for wo differen 3D objecs from he COIL daa se using he algorihm described in Secion 5.2. Wih his approach, views are allocaed adapively depending on he visual compleiy of an objec wih respec o viewing angle. Fig. 8. All of he misclassified MNIST es digis using our mehod 63 ou of,). The e above each digi indicaes he eample number followed by he rue label and he assigned label. sraighforward sum of squared differences SSD). SSD performs very well on his easy daabase due o he lack of variaion in lighing [24] PCA jus makes i faser). The prooype selecion algorihm is illusraed in Fig.. As seen, views are allocaed mainly for more comple caegories wih high wihin class variabiliy. The curve marked SC-proo in Face Fig. 9Deecion shows he improved classificaion performance using his prooype selecion sraegy insead of equally-spaced views. Noe ha we obain a 2.4 percen error rae wih an average of only four wo-dimensional views for each hree-dimensional objec, hanks o he fleibiliy provided by he maching algorihm. 6.3 MPEG-7 Shape Silhouee Daabase Our ne eperimen involves he MPEG-7 shape silhouee daabase, specifically Core Eperimen CE-Shape- par B, Spam Deecion which measures performance of similariy-based rerieval [25]. The daabase consiss of,4 images: 7 shape caegories, 2 images per caegory. The performance is measured using he so-called ªbullseye es,º in which each Classificaion problem Schneiderman and Kanade, IJCV 22 i {, }, non-spam, spam Classificaion problem i couns of words, e.g. Viagra, sock, ouperform, Fig. 9. 3D objec recogniion using he COIL-2 daa se. Comparison of i {,, 2}, non-face, fronal face, profile face. es se error for SSD, Shape Disance SD), and Shape Disance wih k-medoids prooypes SD-proo) versus number of prooype views. For SSD and SD, we varied he number of prooypes uniformly for all muli-bagger Fig.. Prooype views seleced for wo differen 3D objecs from he COIL daa se using he algorihm described in Secion 5.2. Wih his Inro. o Machine Learning Alireza Ghane / Greg Mori 4 Inro. o Machine Learning Alireza Ghane / Greg Mori objecs. For SD-proo, he number of prooypes per objec depended on approach, views are allocaed adapively depending on he visual 5 he wihin-objec variaion as well as he beween-objec similariy. compleiy of an objec wih respec o viewing angle.
5 Cavea - Horses (source?) Once upon a ime here were wo neighboring farmers, Jed and Ned. Each owned a horse, and he horses boh liked o jump he fence beween he wo farms. Clearly he farmers needed some means o ell whose horse was whose. So Jed and Ned go ogeher and agreed on a scheme for discriminaing beween horses. Jed would cu a small noch in one ear of his horse. No a big, painful noch, bu jus big enough o be seen. Well, wouldn you know i, he day afer Jed cu he noch in horse s ear, Ned s horse caugh on he barbed wire fence and ore his ear he eac same way! Somehing else had o be devised, so Jed ied a big blue bow on he ail of his horse. Bu he ne day, Jed s horse jumped he fence, ran ino he field where Ned s horse was grazing, and chewed he bow righ off he oher horse s ail. Ae he whole bow! Inro. o Machine Learning Alireza Ghane / Greg Mori 6 Cavea - Horses (source?) Finally, Jed suggesed, and Ned concurred, ha hey should pick a feaure ha was less ap o change. Heigh seemed like a good feaure o use. Bu were he heighs differen? Well, each farmer wen and measured his horse, and do you know wha? The brown horse was a full inch aller han he whie one! Moral of he sory: ML provides heory and ools for seing parameers. Make sure you have he righ model and feaures. Think abou your feaure vecor. Inro. o Machine Learning Alireza Ghane / Greg Mori 7 Sock Price Predicion Clusering Images Problems in which i is coninuous are called regression E.g. i is sock price, i conains company profi, deb, cash flow, gross sales, number of spam s sen,... Inro. o Machine Learning Alireza Ghane / Greg Mori 8 Wang e al., CVPR 26 Only i is defined: unsupervised learning E.g. i describes image, find groups of similar images Inro. o Machine Learning Alireza Ghane / Greg Mori 9
6 Types of Learning Problems Ouline Course Info.: People, References, Resources Supervised Learning Classificaion Regression Unsupervised Learning Densiy esimaion Clusering: k-means, miure models, hierarchical clusering Hidden Markov models Reinforcemen Learning Machine Learning: Wha, Why, and How? Curve Fiing: (e.g.) Regression and Model Selecion Decision Theory: ML, Loss Funcion, MAP Probabiliy Theory: (e.g.) Probabiliies and Parameer Esimaion Inro. o Machine Learning Alireza Ghane / Greg Mori 2 Inro. o Machine Learning Alireza Ghane / Greg Mori 2 An Eample - Polynomial Curve Fiing Polynomial Curve Fiing Suppose we are given raining se of N observaions (,..., N ) and (,..., N ), i, i R Regression problem, esimae y() from hese daa Wha form is y()? Le s ry polynomials of degree M: y(, w) = w +w +w w M M This is he hypohesis space. How do we measure success? Sum of squared errors: E(w) = 2 N {y( n, w) n } 2 n= Among funcions in he class, choose ha which minimizes his error n y(n, w) n Inro. o Machine Learning Alireza Ghane / Greg Mori 22 Inro. o Machine Learning Alireza Ghane / Greg Mori 23
7 Polynomial Curve Fiing Which Degree of Polynomial? Error funcion Bes coefficiens E(w) = 2 N {y( n, w) n } 2 n= w = arg min E(w) w Found using pseudo-inverse (more laer) Inro. o Machine Learning Alireza Ghane / Greg Mori 24 A model selecion problem M = 9 E(w ) = : This is over-fiing Inro. o Machine Learning Alireza Ghane / Greg Mori 25 Generalizaion Conrolling Over-fiing: Regularizaion Training Tes Generalizaion is he holy grail of ML Wan good performance for new daa Measure generalizaion using a separae se Use roo-mean-squared (RMS) error: E RMS = 2E(w )/N As order of polynomial M increases, so do coefficien magniudes Penalize large coefficiens in error funcion: Ẽ(w) = 2 N {y( n, w) n } 2 + λ 2 w 2 n= Inro. o Machine Learning Alireza Ghane / Greg Mori 26 Inro. o Machine Learning Alireza Ghane / Greg Mori 27
8 Conrolling Over-fiing: Regularizaion Conrolling Over-fiing: Regularizaion Training Tes Noe he E RMS for he raining se. Perfec mach of raining se wih he model is a resul of over-fiing Training and es error show similar rend Inro. o Machine Learning Alireza Ghane / Greg Mori 28 Inro. o Machine Learning Alireza Ghane / Greg Mori 29 Over-fiing: Daase size Validaion Se Spli raining daa ino raining se and validaion se Train differen models (e.g. diff. order polynomials) on raining se Choose model (e.g. order of polynomial) wih minimum error on validaion se Wih more daa, more comple model (M = 9) can be fi Rule of humb: daapoins for each parameer Inro. o Machine Learning Alireza Ghane / Greg Mori 3 Inro. o Machine Learning Alireza Ghane / Greg Mori 3
9 Cross-validaion Summary run run 2 run 3 run 4 Daa are ofen limied Cross-validaion creaes S groups of daa, use S o rain, oher o validae Ereme case leave-one-ou cross-validaion (LOO-CV): S is number of raining daa poins Cross-validaion is an effecive mehod for model selecion, bu can be slow Models wih muliple compleiy parameers: eponenial number of runs Wan models ha generalize o new daa Train model on raining se Measure performance on held-ou es se Performance on es se is good esimae of performance on new daa Inro. o Machine Learning Alireza Ghane / Greg Mori 32 Inro. o Machine Learning Alireza Ghane / Greg Mori 33 Summary - Model Selecion Summary - Soluions I Which model o use? E.g. which degree polynomial? Training se error is lower wih more comple model Can jus choose he model wih lowes raining error Peeking a es error is unfair. E.g. picking polynomial wih lowes es error Performance on es se is no longer good esimae of performance on new daa Use a validaion se Train models on raining se. E.g. differen degree polynomials Measure performance on held-ou validaion se Measure performance of ha model on held-ou es se Can use cross-validaion on raining se insead of a separae validaion se if lile daa and los of ime Choose model wih lowes error over all cross-validaion folds (e.g. polynomial degree) Rerain ha model using all raining daa (e.g. polynomial coefficiens) Inro. o Machine Learning Alireza Ghane / Greg Mori 34 Inro. o Machine Learning Alireza Ghane / Greg Mori 35
10 Summary - Soluions II Ouline Course Info.: People, References, Resources Use regularizaion Train comple model (e.g high order polynomial) bu penalize being oo comple (e.g. large weigh magniudes) Need o balance error vs. regularizaion (λ) Choose λ using cross-validaion Ge more daa Machine Learning: Wha, Why, and How? Curve Fiing: (e.g.) Regression and Model Selecion Decision Theory: ML, Loss Funcion, MAP Probabiliy Theory: (e.g.) Probabiliies and Parameer Esimaion Inro. o Machine Learning Alireza Ghane / Greg Mori 36 Decision Theory Decision: Maimum Likelihood For a sample, decide which class(c k ) i is from. Ideas: Maimum Likelihood Minimum Loss/Cos (e.g. misclassificaion rae) Maimum Aposeriori (MAP) Inference sep: Deermine saisics from raining daa. p(, ) OR p( C k ) Decision sep: Deermine opimal for es inpu : = arg ma{ p ( C k ) k }{{} Likelihood } Inro. o Machine Learning Alireza Ghane 38 Inro. o Machine Learning Alireza Ghane 39
11 Decision: Minimum Misclassificaion Rae p(misake) = p ( R, C 2 ) + p ( R 2, C ) = R p (, C 2 ) d + R 2 p (, C ) d p(, C) R p(misake) = k p(, C2) R2 j R j p (, C k ) d ˆ: decision boundary. : opimal decision boundary : arg min{p (misake)} R Decision: Minimum Loss/Cos Misclassificaion rae: R : arg min {R i i {,,K}} L (R j, C k ) Weighed loss/cos funcion: R : arg min W j,k L (R j, C k ) {R i i {,,K}} Is useful when: The populaion of he classes are differen The failure cos is non-symmeric k k j j Inro. o Machine Learning Alireza Ghane 4 Inro. o Machine Learning Alireza Ghane 4 Decision: Maimum Aposeriori (MAP) Ouline Bayes Theorem: P {A B} = P {B A}P {A} P {B} Course Info.: People, References, Resources Machine Learning: Wha, Why, and How? p(c k ) }{{} P oserior p( C k ) }{{} p(c k ) }{{} Likelihood P rior Provides an Aposeriori Belief for he esimaion, raher han a single poin esimae. Can uilize Apriori Informaion in he decision. Curve Fiing: (e.g.) Regression and Model Selecion Decision Theory: ML, Loss Funcion, MAP Probabiliy Theory: (e.g.) Probabiliies and Parameer Esimaion Inro. o Machine Learning Alireza Ghane 42
12 Coin Tossing Coin Tossing - Model Le s say you re given a coin, and you wan o find ou P (heads), he probabiliy ha if you flip i i lands as heads. Flip i a few imes: H H T P (heads) = 2/3 Hmm... is his rigorous? Does his make sense? Bernoulli disribuion P (heads) = µ, P (ails) = µ Assume coin flips are independen and idenically disribued (i.i.d.) i.e. All are separae samples from he Bernoulli disribuion Given daa D = {,..., N }, heads: i =, ails: i =, he likelihood of he daa is: N N p(d µ) = p( n µ) = µ n ( µ) n n= n= Inro. o Machine Learning Alireza Ghane / Greg Mori 44 Inro. o Machine Learning Alireza Ghane / Greg Mori 45 Maimum Likelihood Esimaion Given D wih h heads and ails Wha should µ be? Maimum Likelihood Esimaion (MLE): choose µ which maimizes he likelihood of he daa µ ML = arg ma µ p(d µ) Since ln( ) is monoone increasing: µ ML = arg ma ln p(d µ) µ Inro. o Machine Learning Alireza Ghane / Greg Mori 46 Likelihood: Maimum Likelihood Esimaion Log-likelihood: ln p(d µ) = p(d µ) = N µ n ( µ) n n= N n ln µ + ( n ) ln( µ) n= Take derivaive, se o : d N dµ ln p(d µ) = n µ ( n) µ = µ h µ n= µ = h + h Inro. o Machine Learning Alireza Ghane / Greg Mori 47
13 Bayesian Learning Wai, does his make sense? Wha if I flip ime, heads? Do I believe µ=? Learn µ he Bayesian way: P (µ D) = P (µ D) }{{} poserior P (D µ)p (µ) P (D) P (D µ) P (µ) }{{}}{{} prior likelihood Bea Disribuion We will use he Bea disribuion o epress our prior knowledge abou coins: Bea(µ a, b) = Γ(a + b) µ a ( µ) b Γ(a)Γ(b) }{{} normalizaion Parameers a and b conrol he shape of his disribuion Prior encodes knowledge ha mos coins are 5-5 Conjugae prior makes mah simpler, easy inerpreaion For Bernoulli, he bea disribuion is is conjugae Inro. o Machine Learning Alireza Ghane / Greg Mori 48 Inro. o Machine Learning Alireza Ghane / Greg Mori 49 Poserior Maimum A Poseriori P (µ D) P (D µ)p (µ) N µ n ( µ) n µ a ( µ) b }{{} n= }{{} prior likelihood µ h ( µ) µ a ( µ) b µ h+a ( µ) +b Simple form for poserior is due o use of conjugae prior Parameers a and b ac as era observaions Noe ha as N = h +, prior is ignored Given poserior P (µ D) we could compue a single value, known as he Maimum a Poseriori (MAP) esimae for µ: µ MAP = arg ma µ P (µ D) Known as poin esimaion However, correc Bayesian hing o do is o use he full disribuion over µ i.e. Compue E µ [f] = p(µ D)f(µ)dµ This inegral is usually hard o compue Inro. o Machine Learning Alireza Ghane / Greg Mori 5 Inro. o Machine Learning Alireza Ghane / Greg Mori 5
14 Polynomial Curve Fiing: Wha We Did Curve Fiing: Probabilisic Approach Wha form is y()? Le s ry polynomials of degree M: y(, w) y(, w) = w +w +w w M M This is he hypohesis space. How do we measure success? Sum of squared errors: n y(, w) p(, w, β) 2σ y(, w) p(, w, β) y(, w) 2σ E(w) = 2 N {y( n, w) n } 2 n= Among funcions in he class, choose ha which minimizes his error n y(n, w) Inro. o Machine Learning Alireza Ghane 52 p(, w, β) = N N ( n y( n, w), β ) n= ln (p(, w, β)) = β N {y( n, w) n } 2 + N 2 2 ln β N ln (2π) Inro. o Machine Learning Alireza Ghane n= }{{}} 2 {{ 53}}{{} cons. cons. βe(w) Maimize log-likelihood Minimize E(w). Can opimize for β as well. Curve Fiing: Bayesian Approach Curve Fiing: Bayesian y(, w) p (, w, β, α) = N ( P w ( ), Q w,β,α ( )) y(, w) y(, w) p(, w, β) = p(, w, β) 2σ y(, w) N N ( n y( n, w), β ) n= Poserior Dis.:p (w,, α, β) p (, w, β) p ( α) Minimize: β N {y( n, w) n } 2 + α 2 2 wt w n= }{{}}{{} regularizaion. βe(w) Inro. o Machine Learning Alireza Ghane 54 p(, w, β) y(, w) 2σ y(, w) p (,, ) = N ( m(), s 2 () ) y(, w) 2σ m() p(, = w, φ() T β) S N φ( n ) n n= s 2 () = β ( + φ() T Sφ() ) S = α N β I + φ( n )φ( n ) T y(, w) 2σ p(, w, β) n= Inro. o Machine Learning Alireza Ghane 55
15 Readings: Chaper.,.3,.5, 2. Types of learning problems Supervised: regression, classificaion Unsupervised Learning as opimizaion Squared error loss funcion Maimum likelihood (ML) Maimum a poseriori (MAP) Wan generalizaion, avoid over-fiing Cross-validaion Regularizaion Bayesian prior on model parameers Inro. o Machine Learning Alireza Ghane 56
Introduction to Machine Learning
Introduction to Machine Learning Bishop PRML Ch. 1 Alireza Ghane Outline Course Info.: People, References, Resources Machine Learning: What, Why, and How? Curve Fitting: (e.g.) Regression and Model Selection
More informationNon-parametric Methods
Non-parametric Methods Machine Learning Alireza Ghane Non-Parametric Methods Alireza Ghane / Torsten Möller 1 Outline Machine Learning: What, Why, and How? Curve Fitting: (e.g.) Regression and Model Selection
More informationNon-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important
on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LTU, decision
More informationNon-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important
on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LDA, logisic
More informationEnsamble methods: Bagging and Boosting
Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par
More informationEnsamble methods: Boosting
Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room
More informationPattern Classification (VI) 杜俊
Paern lassificaion VI 杜俊 jundu@usc.edu.cn Ouline Bayesian Decision Theory How o make he oimal decision? Maximum a oserior MAP decision rule Generaive Models Join disribuion of observaion and label sequences
More information1.1. Example: Polynomial Curve Fitting 4 1. INTRODUCTION
4. INTRODUCTION Figure.2 Plo of a raining daa se of N = poins, shown as blue circles, each comprising an observaion of he inpu variable along wih he corresponding arge variable. The green curve shows he
More information20. Applications of the Genetic-Drift Model
0. Applicaions of he Geneic-Drif Model 1) Deermining he probabiliy of forming any paricular combinaion of genoypes in he nex generaion: Example: If he parenal allele frequencies are p 0 = 0.35 and q 0
More informationZürich. ETH Master Course: L Autonomous Mobile Robots Localization II
Roland Siegwar Margaria Chli Paul Furgale Marco Huer Marin Rufli Davide Scaramuzza ETH Maser Course: 151-0854-00L Auonomous Mobile Robos Localizaion II ACT and SEE For all do, (predicion updae / ACT),
More informationCptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1
CpS 570 Machine Learning School of EECS Washingon Sae Universiy CpS 570 - Machine Learning 1 Form of underlying disribuions unknown Bu sill wan o perform classificaion and regression Semi-parameric esimaion
More informationSpeech and Language Processing
Speech and Language rocessing Lecure 4 Variaional inference and sampling Informaion and Communicaions Engineering Course Takahiro Shinozaki 08//5 Lecure lan (Shinozaki s par) I gives he firs 6 lecures
More informationSpeaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis
Speaker Adapaion Techniques For Coninuous Speech Using Medium and Small Adapaion Daa Ses Consaninos Boulis Ouline of he Presenaion Inroducion o he speaker adapaion problem Maximum Likelihood Sochasic Transformaions
More informationAnno accademico 2006/2007. Davide Migliore
Roboica Anno accademico 2006/2007 Davide Migliore migliore@ele.polimi.i Today Eercise session: An Off-side roblem Robo Vision Task Measuring NBA layers erformance robabilisic Roboics Inroducion The Bayesian
More informationSequential Importance Resampling (SIR) Particle Filter
Paricle Filers++ Pieer Abbeel UC Berkeley EECS Many slides adaped from Thrun, Burgard and Fox, Probabilisic Roboics 1. Algorihm paricle_filer( S -1, u, z ): 2. Sequenial Imporance Resampling (SIR) Paricle
More informationLecture 33: November 29
36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure
More informationGeorey E. Hinton. University oftoronto. Technical Report CRG-TR February 22, Abstract
Parameer Esimaion for Linear Dynamical Sysems Zoubin Ghahramani Georey E. Hinon Deparmen of Compuer Science Universiy oftorono 6 King's College Road Torono, Canada M5S A4 Email: zoubin@cs.orono.edu Technical
More informationUsing the Kalman filter Extended Kalman filter
Using he Kalman filer Eended Kalman filer Doz. G. Bleser Prof. Sricker Compuer Vision: Objec and People Tracking SA- Ouline Recap: Kalman filer algorihm Using Kalman filers Eended Kalman filer algorihm
More informationINTRODUCTION TO MACHINE LEARNING 3RD EDITION
ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class
More information2 1. INTRODUCTION. Figure 1.1. Examples of hand-written digits taken from US zip codes.
Inroducion The problem of searching for paerns in daa is a fundamenal one and has a long and successful hisory. For insance, he eensive asronomical observaions of Tycho Brahe in he 6 h cenury allowed Johannes
More informationRobust estimation based on the first- and third-moment restrictions of the power transformation model
h Inernaional Congress on Modelling and Simulaion, Adelaide, Ausralia, 6 December 3 www.mssanz.org.au/modsim3 Robus esimaion based on he firs- and hird-momen resricions of he power ransformaion Nawaa,
More informationVehicle Arrival Models : Headway
Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where
More information3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon
3..3 INRODUCION O DYNAMIC OPIMIZAION: DISCREE IME PROBLEMS A. he Hamilonian and Firs-Order Condiions in a Finie ime Horizon Define a new funcion, he Hamilonian funcion, H. H he change in he oal value of
More informationCHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK
175 CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 10.1 INTRODUCTION Amongs he research work performed, he bes resuls of experimenal work are validaed wih Arificial Neural Nework. From he
More information1 Review of Zero-Sum Games
COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any
More informationDiebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles
Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance
More informationWritten HW 9 Sol. CS 188 Fall Introduction to Artificial Intelligence
CS 188 Fall 2018 Inroducion o Arificial Inelligence Wrien HW 9 Sol. Self-assessmen due: Tuesday 11/13/2018 a 11:59pm (submi via Gradescope) For he self assessmen, fill in he self assessmen boxes in your
More informationTwo Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017
Two Popular Bayesian Esimaors: Paricle and Kalman Filers McGill COMP 765 Sep 14 h, 2017 1 1 1, dx x Bel x u x P x z P Recall: Bayes Filers,,,,,,, 1 1 1 1 u z u x P u z u x z P Bayes z = observaion u =
More informationLearning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power
Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.
More informationState-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter
Sae-Space Models Iniializaion, Esimaion and Smoohing of he Kalman Filer Iniializaion of he Kalman Filer The Kalman filer shows how o updae pas predicors and he corresponding predicion error variances when
More informationHidden Markov Models
Hidden Markov Models Probabilisic reasoning over ime So far, we ve mosly deal wih episodic environmens Excepions: games wih muliple moves, planning In paricular, he Bayesian neworks we ve seen so far describe
More informationChapter 4. Truncation Errors
Chaper 4. Truncaion Errors and he Taylor Series Truncaion Errors and he Taylor Series Non-elemenary funcions such as rigonomeric, eponenial, and ohers are epressed in an approimae fashion using Taylor
More information1.6. Slopes of Tangents and Instantaneous Rate of Change
1.6 Slopes of Tangens and Insananeous Rae of Change When you hi or kick a ball, he heigh, h, in meres, of he ball can be modelled by he equaion h() 4.9 2 v c. In his equaion, is he ime, in seconds; c represens
More informationBias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé
Bias in Condiional and Uncondiional Fixed Effecs Logi Esimaion: a Correcion * Tom Coupé Economics Educaion and Research Consorium, Naional Universiy of Kyiv Mohyla Academy Address: Vul Voloska 10, 04070
More informationComparing Means: t-tests for One Sample & Two Related Samples
Comparing Means: -Tess for One Sample & Two Relaed Samples Using he z-tes: Assumpions -Tess for One Sample & Two Relaed Samples The z-es (of a sample mean agains a populaion mean) is based on he assumpion
More informationChapter 2. First Order Scalar Equations
Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.
More information0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED
0.1 MAXIMUM LIKELIHOOD ESTIMATIO EXPLAIED Maximum likelihood esimaion is a bes-fi saisical mehod for he esimaion of he values of he parameers of a sysem, based on a se of observaions of a random variable
More informationLearning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power
Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.
More informationPENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD
PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.
More informationSelf assessment due: Monday 4/29/2019 at 11:59pm (submit via Gradescope)
CS 188 Spring 2019 Inroducion o Arificial Inelligence Wrien HW 10 Due: Monday 4/22/2019 a 11:59pm (submi via Gradescope). Leave self assessmen boxes blank for his due dae. Self assessmen due: Monday 4/29/2019
More informationMath 10B: Mock Mid II. April 13, 2016
Name: Soluions Mah 10B: Mock Mid II April 13, 016 1. ( poins) Sae, wih jusificaion, wheher he following saemens are rue or false. (a) If a 3 3 marix A saisfies A 3 A = 0, hen i canno be inverible. True.
More informationTasty Coffee example
Lecure Slides for (Binary) Classificaion: Learning a Class from labeled Examples ITRODUCTIO TO Machine Learning ETHEM ALPAYDI The MIT Press, 00 (modified by dph, 0000) CHAPTER : Supervised Learning Things
More informationUnit Root Time Series. Univariate random walk
Uni Roo ime Series Univariae random walk Consider he regression y y where ~ iid N 0, he leas squares esimae of is: ˆ yy y y yy Now wha if = If y y hen le y 0 =0 so ha y j j If ~ iid N 0, hen y ~ N 0, he
More informationBlock Diagram of a DCS in 411
Informaion source Forma A/D From oher sources Pulse modu. Muliplex Bandpass modu. X M h: channel impulse response m i g i s i Digial inpu Digial oupu iming and synchronizaion Digial baseband/ bandpass
More informationReliability of Technical Systems
eliabiliy of Technical Sysems Main Topics Inroducion, Key erms, framing he problem eliabiliy parameers: Failure ae, Failure Probabiliy, Availabiliy, ec. Some imporan reliabiliy disribuions Componen reliabiliy
More informationEnsemble Confidence Estimates Posterior Probability
Ensemble Esimaes Poserior Probabiliy Michael Muhlbaier, Aposolos Topalis, and Robi Polikar Rowan Universiy, Elecrical and Compuer Engineering, Mullica Hill Rd., Glassboro, NJ 88, USA {muhlba6, opali5}@sudens.rowan.edu
More informationMath 2142 Exam 1 Review Problems. x 2 + f (0) 3! for the 3rd Taylor polynomial at x = 0. To calculate the various quantities:
Mah 4 Eam Review Problems Problem. Calculae he 3rd Taylor polynomial for arcsin a =. Soluion. Le f() = arcsin. For his problem, we use he formula f() + f () + f ()! + f () 3! for he 3rd Taylor polynomial
More informationKriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Kriging Models Predicing Arazine Concenraions in Surface Waer Draining Agriculural Waersheds Paul L. Mosquin, Jeremy Aldworh, Wenlin Chen Supplemenal Maerial Number
More informationOnline Convex Optimization Example And Follow-The-Leader
CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion
More informationKEY. Math 334 Midterm I Fall 2008 sections 001 and 003 Instructor: Scott Glasgow
1 KEY Mah 4 Miderm I Fall 8 secions 1 and Insrucor: Sco Glasgow Please do NOT wrie on his eam. No credi will be given for such work. Raher wrie in a blue book, or on our own paper, preferabl engineering
More informationArticle from. Predictive Analytics and Futurism. July 2016 Issue 13
Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning
More informationData Fusion using Kalman Filter. Ioannis Rekleitis
Daa Fusion using Kalman Filer Ioannis Rekleiis Eample of a arameerized Baesian Filer: Kalman Filer Kalman filers (KF represen poserior belief b a Gaussian (normal disribuion A -d Gaussian disribuion is
More informationGMM - Generalized Method of Moments
GMM - Generalized Mehod of Momens Conens GMM esimaion, shor inroducion 2 GMM inuiion: Maching momens 2 3 General overview of GMM esimaion. 3 3. Weighing marix...........................................
More information3.1 More on model selection
3. More on Model selecion 3. Comparing models AIC, BIC, Adjused R squared. 3. Over Fiing problem. 3.3 Sample spliing. 3. More on model selecion crieria Ofen afer model fiing you are lef wih a handful of
More informationApplications in Industry (Extended) Kalman Filter. Week Date Lecture Title
hp://elec34.com Applicaions in Indusry (Eended) Kalman Filer 26 School of Informaion echnology and Elecrical Engineering a he Universiy of Queensland Lecure Schedule: Week Dae Lecure ile 29-Feb Inroducion
More informationACE 564 Spring Lecture 7. Extensions of The Multiple Regression Model: Dummy Independent Variables. by Professor Scott H.
ACE 564 Spring 2006 Lecure 7 Exensions of The Muliple Regression Model: Dumm Independen Variables b Professor Sco H. Irwin Readings: Griffihs, Hill and Judge. "Dumm Variables and Varing Coefficien Models
More informationt is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t...
Mah 228- Fri Mar 24 5.6 Marix exponenials and linear sysems: The analogy beween firs order sysems of linear differenial equaions (Chaper 5) and scalar linear differenial equaions (Chaper ) is much sronger
More informationWeek 1 Lecture 2 Problems 2, 5. What if something oscillates with no obvious spring? What is ω? (problem set problem)
Week 1 Lecure Problems, 5 Wha if somehing oscillaes wih no obvious spring? Wha is ω? (problem se problem) Sar wih Try and ge o SHM form E. Full beer can in lake, oscillaing F = m & = ge rearrange: F =
More informationTracking. Announcements
Tracking Tuesday, Nov 24 Krisen Grauman UT Ausin Announcemens Pse 5 ou onigh, due 12/4 Shorer assignmen Auo exension il 12/8 I will no hold office hours omorrow 5 6 pm due o Thanksgiving 1 Las ime: Moion
More informationIsolated-word speech recognition using hidden Markov models
Isolaed-word speech recogniion using hidden Markov models Håkon Sandsmark December 18, 21 1 Inroducion Speech recogniion is a challenging problem on which much work has been done he las decades. Some of
More informationEcon107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8)
I. Definiions and Problems A. Perfec Mulicollineariy Econ7 Applied Economerics Topic 7: Mulicollineariy (Sudenmund, Chaper 8) Definiion: Perfec mulicollineariy exiss in a following K-variable regression
More informationACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H.
ACE 56 Fall 005 Lecure 5: he Simple Linear Regression Model: Sampling Properies of he Leas Squares Esimaors by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Inference in he Simple
More informationSection 4.4 Logarithmic Properties
Secion. Logarihmic Properies 59 Secion. Logarihmic Properies In he previous secion, we derived wo imporan properies of arihms, which allowed us o solve some asic eponenial and arihmic equaions. Properies
More informationL07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms
L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS NA568 Mobile Roboics: Mehods & Algorihms Today s Topic Quick review on (Linear) Kalman Filer Kalman Filering for Non-Linear Sysems Exended Kalman Filer (EKF)
More informationSolutions from Chapter 9.1 and 9.2
Soluions from Chaper 9 and 92 Secion 9 Problem # This basically boils down o an exercise in he chain rule from calculus We are looking for soluions of he form: u( x) = f( k x c) where k x R 3 and k is
More informationSMT 2014 Calculus Test Solutions February 15, 2014 = 3 5 = 15.
SMT Calculus Tes Soluions February 5,. Le f() = and le g() =. Compue f ()g (). Answer: 5 Soluion: We noe ha f () = and g () = 6. Then f ()g () =. Plugging in = we ge f ()g () = 6 = 3 5 = 5.. There is a
More informationAn Introduction to Stochastic Programming: The Recourse Problem
An Inroducion o Sochasic Programming: he Recourse Problem George Danzig and Phil Wolfe Ellis Johnson, Roger Wes, Dick Cole, and Me John Birge Where o look in he ex pp. 6-7, Secion.2.: Inroducion o sochasic
More informationLecture 10 - Model Identification
Lecure - odel Idenificaion Wha is ssem idenificaion? Direc impulse response idenificaion Linear regression Regularizaion Parameric model ID nonlinear LS Conrol Engineering - Wha is Ssem Idenificaion? Experimen
More informationLecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.
Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in
More informationUnderstanding the asymptotic behaviour of empirical Bayes methods
Undersanding he asympoic behaviour of empirical Bayes mehods Boond Szabo, Aad van der Vaar and Harry van Zanen EURANDOM, 11.10.2011. Conens 2/20 Moivaion Nonparameric Bayesian saisics Signal in Whie noise
More informationUSP. Surplus-Production Models
USP Surplus-Producion Models 2 Overview Purpose of slides: Inroducion o he producion model Overview of differen mehods of fiing Go over some criique of he mehod Source: Haddon 2001, Chaper 10 Hilborn and
More informationES.1803 Topic 22 Notes Jeremy Orloff
ES.83 Topic Noes Jeremy Orloff Fourier series inroducion: coninued. Goals. Be able o compue he Fourier coefficiens of even or odd periodic funcion using he simplified formulas.. Be able o wrie and graph
More informationSection 4.4 Logarithmic Properties
Secion. Logarihmic Properies 5 Secion. Logarihmic Properies In he previous secion, we derived wo imporan properies of arihms, which allowed us o solve some asic eponenial and arihmic equaions. Properies
More informationA Bayesian Approach to Spectral Analysis
Chirped Signals A Bayesian Approach o Specral Analysis Chirped signals are oscillaing signals wih ime variable frequencies, usually wih a linear variaion of frequency wih ime. E.g. f() = A cos(ω + α 2
More informationCS 4495 Computer Vision Tracking 1- Kalman,Gaussian
CS 4495 Compuer Vision A. Bobick CS 4495 Compuer Vision - KalmanGaussian Aaron Bobick School of Ineracive Compuing CS 4495 Compuer Vision A. Bobick Adminisrivia S5 will be ou his Thurs Due Sun Nov h :55pm
More informationAn introduction to the theory of SDDP algorithm
An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking
More information13.3 Term structure models
13.3 Term srucure models 13.3.1 Expecaions hypohesis model - Simples "model" a) shor rae b) expecaions o ge oher prices Resul: y () = 1 h +1 δ = φ( δ)+ε +1 f () = E (y +1) (1) =δ + φ( δ) f (3) = E (y +)
More informationKalman filtering for maximum likelihood estimation given corrupted observations.
alman filering maimum likelihood esimaion given corruped observaions... Holmes Naional Marine isheries Service Inroducion he alman filer is used o eend likelihood esimaion o cases wih hidden saes such
More informationTwo Coupled Oscillators / Normal Modes
Lecure 3 Phys 3750 Two Coupled Oscillaors / Normal Modes Overview and Moivaion: Today we ake a small, bu significan, sep owards wave moion. We will no ye observe waves, bu his sep is imporan in is own
More informationAsymptotic Equipartition Property - Seminar 3, part 1
Asympoic Equipariion Propery - Seminar 3, par 1 Ocober 22, 2013 Problem 1 (Calculaion of ypical se) To clarify he noion of a ypical se A (n) ε and he smalles se of high probabiliy B (n), we will calculae
More informationAir Traffic Forecast Empirical Research Based on the MCMC Method
Compuer and Informaion Science; Vol. 5, No. 5; 0 ISSN 93-8989 E-ISSN 93-8997 Published by Canadian Cener of Science and Educaion Air Traffic Forecas Empirical Research Based on he MCMC Mehod Jian-bo Wang,
More information15. Vector Valued Functions
1. Vecor Valued Funcions Up o his poin, we have presened vecors wih consan componens, for example, 1, and,,4. However, we can allow he componens of a vecor o be funcions of a common variable. For example,
More informationTracking. Many slides adapted from Kristen Grauman, Deva Ramanan
Tracking Man slides adaped from Krisen Grauman Deva Ramanan Coures G. Hager Coures G. Hager J. Kosecka cs3b Adapive Human-Moion Tracking Acquisiion Decimaion b facor 5 Moion deecor Grascale convers. Image
More information5. Stochastic processes (1)
Lec05.pp S-38.45 - Inroducion o Teleraffic Theory Spring 2005 Conens Basic conceps Poisson process 2 Sochasic processes () Consider some quaniy in a eleraffic (or any) sysem I ypically evolves in ime randomly
More informationFinish reading Chapter 2 of Spivak, rereading earlier sections as necessary. handout and fill in some missing details!
MAT 257, Handou 6: Ocober 7-2, 20. I. Assignmen. Finish reading Chaper 2 of Spiva, rereading earlier secions as necessary. handou and fill in some missing deails! II. Higher derivaives. Also, read his
More informationProbabilistic Robotics SLAM
Probabilisic Roboics SLAM The SLAM Problem SLAM is he process by which a robo builds a map of he environmen and, a he same ime, uses his map o compue is locaion Localizaion: inferring locaion given a map
More informationNotes on Kalman Filtering
Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren
More informationPhysics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle
Chaper 2 Newonian Mechanics Single Paricle In his Chaper we will review wha Newon s laws of mechanics ell us abou he moion of a single paricle. Newon s laws are only valid in suiable reference frames,
More informationApplication of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing
Applicaion of a Sochasic-Fuzzy Approach o Modeling Opimal Discree Time Dynamical Sysems by Using Large Scale Daa Processing AA WALASZE-BABISZEWSA Deparmen of Compuer Engineering Opole Universiy of Technology
More information6. 6 v ; degree = 7; leading coefficient = 6; 7. The expression has 3 terms; t p no; subtracting x from 3x ( 3x x 2x)
70. a =, r = 0%, = 0. 7. a = 000, r = 0.%, = 00 7. a =, r = 00%, = 7. ( ) = 0,000 0., where = ears 7. ( ) = + 0.0, where = weeks 7 ( ) =,000,000 0., where = das 7 = 77. = 9 7 = 7 geomeric 0. geomeric arihmeic,
More informationAnnouncements. Recap: Filtering. Recap: Reasoning Over Time. Example: State Representations for Robot Localization. Particle Filtering
Inroducion o Arificial Inelligence V22.0472-001 Fall 2009 Lecure 18: aricle & Kalman Filering Announcemens Final exam will be a 7pm on Wednesday December 14 h Dae of las class 1.5 hrs long I won ask anyhing
More informationChapter 7: Solving Trig Equations
Haberman MTH Secion I: The Trigonomeric Funcions Chaper 7: Solving Trig Equaions Le s sar by solving a couple of equaions ha involve he sine funcion EXAMPLE a: Solve he equaion sin( ) The inverse funcions
More information4.1 Other Interpretations of Ridge Regression
CHAPTER 4 FURTHER RIDGE THEORY 4. Oher Inerpreaions of Ridge Regression In his secion we will presen hree inerpreaions for he use of ridge regression. The firs one is analogous o Hoerl and Kennard reasoning
More information12: AUTOREGRESSIVE AND MOVING AVERAGE PROCESSES IN DISCRETE TIME. Σ j =
1: AUTOREGRESSIVE AND MOVING AVERAGE PROCESSES IN DISCRETE TIME Moving Averages Recall ha a whie noise process is a series { } = having variance σ. The whie noise process has specral densiy f (λ) = of
More information!!"#"$%&#'()!"#&'(*%)+,&',-)./0)1-*23)
"#"$%&#'()"#&'(*%)+,&',-)./)1-*) #$%&'()*+,&',-.%,/)*+,-&1*#$)()5*6$+$%*,7&*-'-&1*(,-&*6&,7.$%$+*&%'(*8$&',-,%'-&1*(,-&*6&,79*(&,%: ;..,*&1$&$.$%&'()*1$$.,'&',-9*(&,%)?%*,('&5
More informationNature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time.
Supplemenary Figure 1 Spike-coun auocorrelaions in ime. Normalized auocorrelaion marices are shown for each area in a daase. The marix shows he mean correlaion of he spike coun in each ime bin wih he spike
More informationObject tracking: Using HMMs to estimate the geographical location of fish
Objec racking: Using HMMs o esimae he geographical locaion of fish 02433 - Hidden Markov Models Marin Wæver Pedersen, Henrik Madsen Course week 13 MWP, compiled June 8, 2011 Objecive: Locae fish from agging
More informationR t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t
Exercise 7 C P = α + β R P + u C = αp + βr + v (a) (b) C R = α P R + β + w (c) Assumpions abou he disurbances u, v, w : Classical assumions on he disurbance of one of he equaions, eg. on (b): E(v v s P,
More informationChapter 2. Models, Censoring, and Likelihood for Failure-Time Data
Chaper 2 Models, Censoring, and Likelihood for Failure-Time Daa William Q. Meeker and Luis A. Escobar Iowa Sae Universiy and Louisiana Sae Universiy Copyrigh 1998-2008 W. Q. Meeker and L. A. Escobar. Based
More informationDistribution of Estimates
Disribuion of Esimaes From Economerics (40) Linear Regression Model Assume (y,x ) is iid and E(x e )0 Esimaion Consisency y α + βx + he esimaes approach he rue values as he sample size increases Esimaion
More information