Speech and Language Processing

Size: px
Start display at page:

Download "Speech and Language Processing"

Transcription

1 Speech and Language rocessing Lecure 4 Variaional inference and sampling Informaion and Communicaions Engineering Course Takahiro Shinozaki 08//5

2 Lecure lan (Shinozaki s par) I gives he firs 6 lecures abou speech recogniion. Through hese lecures he backbone of he laes speech recogniion echniues is eplained.. 0/9 (remoe) Speech recogniion based on G and N gram. 0/6 (remoe) aimum likelihood esimaion and E algorihm 3. /5 (@TAIST) Bayesian nework and Bayesian inference 4. /5 (@TAIST) Variaional inference and sampling 5. /6 (@TAIST) Neural nework based acousic and language models 6. /6 (@TAIST) Weighed finie sae ransducer (WFST) and speech decoding

3 Today s Topic Answers for he previous eercises Variaional inference Sampling 3

4 Answers for he revious Eercises 4

5 Eercise 3. Is he direced graph a AG? Yes No Graph A Graph B 5

6 Eercise 3. Represen he following join probabiliy by a BN 6 ) ( ) ( E B B C A B A E C B A A B C E

7 Eercise 3.3 Fill he blanks so ha he following and he BN become euivalen Iniial sae Iniial sae probabiliy (a) (b) (S S ) S =a S =b S =a ( 0.3 ) ( 0.7 ) S =b ( 0.4 ) 0.6 a 0.4 b a a b b S S S 3 S T N X a a N X b b N X X X 3 X T X s s BN 7

8 Eercise 3.4 Assumes a probabilisic model (μ) a raining sample and a prior disribuion of a parameer (μ) are given as follows. Gaussian disribuion wih mean μ and variance ep μ ep Gaussian disribuion wih mean 0 and variance ) Esimae poserior disribuion ) Esimae predicive disribuion Noe: ep c d c d 8

9 Eercise 3.4 (Answer) 9 ep 4 ep ep ep ep ep ep N d d 3 3 ep 3 N d

10 Variaional inference 0

11 Variaional Bayes Evaluaing poserior disribuion is ofen no feasible Le s approimae i wih a simpler disribuion p p p p p d Simpler disribuion Too comple As he approimaion measure KL divergence can be used arg min KL p * (For simpliciy of noaion) The minimum of he KL divergence is found using variaional mehod

12 KL and Lower bound inimizing KL is eual o maimizing lower bound d p p d p p d p p KL d p p KL p odel evidence (Consan for ) Lower bound d p L

13 Approimaion by Facorizaion Assumes (groups of) hidden variables are condiionally independen I is called a mean filed approimaion by anay o physics No resricion on he funcional forms p i i i 3

14 oserior Inference wih ean Field Approimaion Suppose a probabiliy model consis of wo (groups of) hidden variables and a (group of) observaion variable We approimae p() as ()=()() arg ma L arg ma L d d The mehod of Lagrange muliplier arg ma F ] L d F[ d 4

15 aimizaion of F[] 5 ] [ d d dd p F 0 0 d p d p z d p C d p C ep ep C C : normalizaion consan Variaional mehod (See he appendi for he derivaion) The maimum is obained by alernaively updaing () and () saring from an iniial disribuion

16 Variaional G 6 X p X X X T Θ T π d d p p ep ep ean field approimaion:

17 Con. 7 d X d d d X X ep ep ep ep ep Cf. (compare wih he above resuls) X X X T Θ T π ep X ep X X X ep ep

18 Sampling ehods 8

19 seudo Random Generaor On digial compuer everyhing is deerminisically calculaed and here is no randomness owever someimes we wan random numbers os programing languages have a pseudo random generaor funcion yhon.6 > impor random > random.random() > random.random() > random.random()

20 Sampling From a Uniform isribuion Random numbers disribued uniformly over some region Eample: isogram of samples obained from a uniform disribuion over (0 ) (To make he graph scilab was used) 0 samples 000 samples Number of occurrence Number of occurrence hisplo(0rand(:0) normalizaion=%f) hisplo(0rand(:000) normalizaion=%f) 0

21 Sampling From a Gaussian isribuion Sandard normal (Gaussian) disribuion has a mean 0.0 and a variance.0 Eample: 00 samples 0000 samples Number of occurrence Number of occurrence

22 Transform of Random Variable Le be a random variable and f be a funcion y = f(). When follows p() y follows he following disribuion (y) y d p dy

23 Eample When p() and y = f() are given as follows obain disribuion (y) p( ) 0 Answer y 3 y d ( y) y 5 3 dy 3 3 p() () 0 /3 5 y 3

24 Eercise 4. When p() and y = f() are given as follows obain disribuion (y) p( ) 0 y 0 y 0 y ep d dy y ep y d ( y) p( ) ep( y) dy # of occurrence isogram of # of occurrence isogram of y y 4

25 Eercise 4. When p() and y = f() are given as follows obain disribuion (y) p( ) ep N 0 y 3 4 y 4 3 d dy 3 y 4 N 43 d ( y) p( ) ep y dy 3 3 isogram of isogram of y # of occurrence # of occurrence y 5

26 Sampling form Comple isribuions isribuions ha can be obained by a ransformaion from a simple disribuion (such as uniform disribuion) is limied We need sampling mehods ha do no reuire inegral and inverse of a funcion and can be applied o more comple disribuions Ancesral sampling Rejecion sampling arkov Chain one Carlo (CC) 6

27 Ancesral Sampling Assumpions: We wan samples from a join disribuion p( ) The join disribuion is given as a Bayesian nework Sampling from he condiional disribuions are easy Algorihm: Sample from he paren nodes o he child nodes in order For he child nodes use he already sampled paren value in he condiional par A B (A) (BA) (CB) (AB) (E) C E 7

28 Rejecion Sampling Assumpions: We wan samples from a disribuion p(). The normalizaion consan may be unknown. p ~ p p~ We have a disribuion () from which we can easily derive samples. We refer o as a proposal disracion 8

29 rocedure of Rejecion Sampling Algorihm:. Choose a consan k so ha he following holds. erive a sample s from () 3. erive a sample u from a uniform disribuion ranging [0 k(s)] 4. If u > p~ hen rejec he sample. Oherwise adop i p~ k ~ p u s k(s) k() 9

30 arkov Chain one Carlo General and powerful framework for sampling Scales well wih he dimensionaliy of he sample space ainains a sae ha forms a arkov chain. The se of he saes follows he desired disribuion eropolis algorihm eropolis asings algorihm Gibbs Sampling 30

31 eropolis Algorihm Assumpions: We wan samples from a disribuion p(x) p ~ X px The normalizaion consan may be unknown Iniializaion:. repare a symmeric proposal disribuion ( A B ) ha saisfy ( A B )=( B A ). repare an iniial sae 0 3

32 Eercise 4.3 Show ha N( A B ) = N( B A ) where N(mv) is he Gaussian disribuion wih mean m and variance v 3 ep ep A B A B B A B A N N

33 rocedure of eropolis Algorihm Algorihm:. Ge a candidae sample * from he proposal disribuion. ( ) based on he curren sae ~ p Accep he candidae wih probabiliy * A * min ~ or discard i p 3. If he candidae is acceped save i as he ne sae +. If i is discarded hen se + euals o 4. Goo sep * 33

34 eropolis asings Algorihm An eension of he eropolis algorihm No symmeric reuiremen for he proposal disribuion The accepance probabiliy of he candidae sae is defined as follows. Oher par is he same as he eropolis algorihm A k * min ~ p ~ p * k * k * 34

35 Gibbs Sampling roblem: We wan samples from a join disribuion p( ) Algorihm:. repare an iniial sae X 0 = < > 0. Selec one of he variables i in order or a random 3. Ge a sample from p i X \ i and updae i wih ha value 4. Goo sep. Afer enough ieraions he disribuion of follows p(x) Compared o he eropolis algorihm: Sampling from condiional disribuion of i given all oher variables need o be feasible There is no rejecion sep 35

36 Quick ome Work Q3. Q4. ue: 0:00 :00 Today Submission: Aach o an Tile: TAISTQ3Q4 Forma: Te file file name: your suden I (eg ) Your name Your I Q3. (A) (B) Q4. 36

37 Appendi 37

38 Variaional ehod (Ouline) Review of derivaives Funcion: a mapping from a value o a value f f Se of values Se of values f f

39 Funcional and Funcional erivaive Funcional Se of funcions Ff Se of values E. Enropy [p] akes a funcion p (probabiliy disribuion) and reurns a value F f If a funcional F akes a maimum/minimum a f 0 and f is close o f 0 Ff Ff f f f 0 f 39

40 F f Euler Lagrange Euaion f f F is a funcional of f Suppose F akes minimum/maimum a 0 f 0. Le η be an arbiral funcion of and ε is a scalar consan g F f g(ε) is a funcion of ε (akes and reurns a scalar) When ε is closed o 0 f is close o f 0. F f Ff Therefore. 0 0 This mus hold for arbiral η. 0 F f g erely a derivaive of a funcion 40

41 Con. C When F f h f d F f h h f f h f f d d d h d 0 f mus hold for arbiral η. h f 0 C.f. ow abou when Ff hf f d a b 0 b a 4

Anno accademico 2006/2007. Davide Migliore

Anno accademico 2006/2007. Davide Migliore Roboica Anno accademico 2006/2007 Davide Migliore migliore@ele.polimi.i Today Eercise session: An Off-side roblem Robo Vision Task Measuring NBA layers erformance robabilisic Roboics Inroducion The Bayesian

More information

Sequential Importance Resampling (SIR) Particle Filter

Sequential Importance Resampling (SIR) Particle Filter Paricle Filers++ Pieer Abbeel UC Berkeley EECS Many slides adaped from Thrun, Burgard and Fox, Probabilisic Roboics 1. Algorihm paricle_filer( S -1, u, z ): 2. Sequenial Imporance Resampling (SIR) Paricle

More information

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017 Two Popular Bayesian Esimaors: Paricle and Kalman Filers McGill COMP 765 Sep 14 h, 2017 1 1 1, dx x Bel x u x P x z P Recall: Bayes Filers,,,,,,, 1 1 1 1 u z u x P u z u x z P Bayes z = observaion u =

More information

L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms

L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS NA568 Mobile Roboics: Mehods & Algorihms Today s Topic Quick review on (Linear) Kalman Filer Kalman Filering for Non-Linear Sysems Exended Kalman Filer (EKF)

More information

Comparing Means: t-tests for One Sample & Two Related Samples

Comparing Means: t-tests for One Sample & Two Related Samples Comparing Means: -Tess for One Sample & Two Relaed Samples Using he z-tes: Assumpions -Tess for One Sample & Two Relaed Samples The z-es (of a sample mean agains a populaion mean) is based on he assumpion

More information

Kalman filtering for maximum likelihood estimation given corrupted observations.

Kalman filtering for maximum likelihood estimation given corrupted observations. alman filering maimum likelihood esimaion given corruped observaions... Holmes Naional Marine isheries Service Inroducion he alman filer is used o eend likelihood esimaion o cases wih hidden saes such

More information

Estimation of Poses with Particle Filters

Estimation of Poses with Particle Filters Esimaion of Poses wih Paricle Filers Dr.-Ing. Bernd Ludwig Chair for Arificial Inelligence Deparmen of Compuer Science Friedrich-Alexander-Universiä Erlangen-Nürnberg 12/05/2008 Dr.-Ing. Bernd Ludwig (FAU

More information

Machine Learning 4771

Machine Learning 4771 ony Jebara, Columbia Universiy achine Learning 4771 Insrucor: ony Jebara ony Jebara, Columbia Universiy opic 20 Hs wih Evidence H Collec H Evaluae H Disribue H Decode H Parameer Learning via JA & E ony

More information

Tom Heskes and Onno Zoeter. Presented by Mark Buller

Tom Heskes and Onno Zoeter. Presented by Mark Buller Tom Heskes and Onno Zoeer Presened by Mark Buller Dynamic Bayesian Neworks Direced graphical models of sochasic processes Represen hidden and observed variables wih differen dependencies Generalize Hidden

More information

CSE-571 Robotics. Sample-based Localization (sonar) Motivation. Bayes Filter Implementations. Particle filters. Density Approximation

CSE-571 Robotics. Sample-based Localization (sonar) Motivation. Bayes Filter Implementations. Particle filters. Density Approximation Moivaion CSE57 Roboics Bayes Filer Implemenaions Paricle filers So far, we discussed he Kalman filer: Gaussian, linearizaion problems Paricle filers are a way o efficienly represen nongaussian disribuions

More information

1 1 + x 2 dx. tan 1 (2) = ] ] x 3. Solution: Recall that the given integral is improper because. x 3. 1 x 3. dx = lim dx.

1 1 + x 2 dx. tan 1 (2) = ] ] x 3. Solution: Recall that the given integral is improper because. x 3. 1 x 3. dx = lim dx. . Use Simpson s rule wih n 4 o esimae an () +. Soluion: Since we are using 4 seps, 4 Thus we have [ ( ) f() + 4f + f() + 4f 3 [ + 4 4 6 5 + + 4 4 3 + ] 5 [ + 6 6 5 + + 6 3 + ]. 5. Our funcion is f() +.

More information

State-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter

State-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter Sae-Space Models Iniializaion, Esimaion and Smoohing of he Kalman Filer Iniializaion of he Kalman Filer The Kalman filer shows how o updae pas predicors and he corresponding predicion error variances when

More information

Outline. lse-logo. Outline. Outline. 1 Wald Test. 2 The Likelihood Ratio Test. 3 Lagrange Multiplier Tests

Outline. lse-logo. Outline. Outline. 1 Wald Test. 2 The Likelihood Ratio Test. 3 Lagrange Multiplier Tests Ouline Ouline Hypohesis Tes wihin he Maximum Likelihood Framework There are hree main frequenis approaches o inference wihin he Maximum Likelihood framework: he Wald es, he Likelihood Raio es and he Lagrange

More information

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon 3..3 INRODUCION O DYNAMIC OPIMIZAION: DISCREE IME PROBLEMS A. he Hamilonian and Firs-Order Condiions in a Finie ime Horizon Define a new funcion, he Hamilonian funcion, H. H he change in he oal value of

More information

GMM - Generalized Method of Moments

GMM - Generalized Method of Moments GMM - Generalized Mehod of Momens Conens GMM esimaion, shor inroducion 2 GMM inuiion: Maching momens 2 3 General overview of GMM esimaion. 3 3. Weighing marix...........................................

More information

Written HW 9 Sol. CS 188 Fall Introduction to Artificial Intelligence

Written HW 9 Sol. CS 188 Fall Introduction to Artificial Intelligence CS 188 Fall 2018 Inroducion o Arificial Inelligence Wrien HW 9 Sol. Self-assessmen due: Tuesday 11/13/2018 a 11:59pm (submi via Gradescope) For he self assessmen, fill in he self assessmen boxes in your

More information

Probabilistic Robotics

Probabilistic Robotics Probabilisic Roboics Bayes Filer Implemenaions Gaussian filers Bayes Filer Reminder Predicion bel p u bel d Correcion bel η p z bel Gaussians : ~ π e p N p - Univariae / / : ~ μ μ μ e p Ν p d π Mulivariae

More information

References are appeared in the last slide. Last update: (1393/08/19)

References are appeared in the last slide. Last update: (1393/08/19) SYSEM IDEIFICAIO Ali Karimpour Associae Professor Ferdowsi Universi of Mashhad References are appeared in he las slide. Las updae: 0..204 393/08/9 Lecure 5 lecure 5 Parameer Esimaion Mehods opics o be

More information

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t Exercise 7 C P = α + β R P + u C = αp + βr + v (a) (b) C R = α P R + β + w (c) Assumpions abou he disurbances u, v, w : Classical assumions on he disurbance of one of he equaions, eg. on (b): E(v v s P,

More information

Linear Cryptanalysis

Linear Cryptanalysis Linear Crypanalysis T-79.550 Crypology Lecure 5 February 6, 008 Kaisa Nyberg Linear Crypanalysis /36 SPN A Small Example Linear Crypanalysis /36 Linear Approximaion of S-boxes Linear Crypanalysis 3/36

More information

Y. Xiang, Learning Bayesian Networks 1

Y. Xiang, Learning Bayesian Networks 1 Learning Bayesian Neworks Objecives Acquisiion of BNs Technical conex of BN learning Crierion of sound srucure learning BN srucure learning in 2 seps BN CPT esimaion Reference R.E. Neapolian: Learning

More information

Using the Kalman filter Extended Kalman filter

Using the Kalman filter Extended Kalman filter Using he Kalman filer Eended Kalman filer Doz. G. Bleser Prof. Sricker Compuer Vision: Objec and People Tracking SA- Ouline Recap: Kalman filer algorihm Using Kalman filers Eended Kalman filer algorihm

More information

An EM algorithm for maximum likelihood estimation given corrupted observations. E. E. Holmes, National Marine Fisheries Service

An EM algorithm for maximum likelihood estimation given corrupted observations. E. E. Holmes, National Marine Fisheries Service An M algorihm maimum likelihood esimaion given corruped observaions... Holmes Naional Marine Fisheries Service Inroducion M algorihms e likelihood esimaion o cases wih hidden saes such as when observaions

More information

SMT 2014 Calculus Test Solutions February 15, 2014 = 3 5 = 15.

SMT 2014 Calculus Test Solutions February 15, 2014 = 3 5 = 15. SMT Calculus Tes Soluions February 5,. Le f() = and le g() =. Compue f ()g (). Answer: 5 Soluion: We noe ha f () = and g () = 6. Then f ()g () =. Plugging in = we ge f ()g () = 6 = 3 5 = 5.. There is a

More information

Zürich. ETH Master Course: L Autonomous Mobile Robots Localization II

Zürich. ETH Master Course: L Autonomous Mobile Robots Localization II Roland Siegwar Margaria Chli Paul Furgale Marco Huer Marin Rufli Davide Scaramuzza ETH Maser Course: 151-0854-00L Auonomous Mobile Robos Localizaion II ACT and SEE For all do, (predicion updae / ACT),

More information

Hidden Markov Models

Hidden Markov Models Hidden Markov Models Probabilisic reasoning over ime So far, we ve mosly deal wih episodic environmens Excepions: games wih muliple moves, planning In paricular, he Bayesian neworks we ve seen so far describe

More information

Chapter 14. (Supplementary) Bayesian Filtering for State Estimation of Dynamic Systems

Chapter 14. (Supplementary) Bayesian Filtering for State Estimation of Dynamic Systems Chaper 4. Supplemenary Bayesian Filering for Sae Esimaion of Dynamic Sysems Neural Neworks and Learning Machines Haykin Lecure Noes on Selflearning Neural Algorihms ByoungTak Zhang School of Compuer Science

More information

Data Fusion using Kalman Filter. Ioannis Rekleitis

Data Fusion using Kalman Filter. Ioannis Rekleitis Daa Fusion using Kalman Filer Ioannis Rekleiis Eample of a arameerized Baesian Filer: Kalman Filer Kalman filers (KF represen poserior belief b a Gaussian (normal disribuion A -d Gaussian disribuion is

More information

Robust estimation based on the first- and third-moment restrictions of the power transformation model

Robust estimation based on the first- and third-moment restrictions of the power transformation model h Inernaional Congress on Modelling and Simulaion, Adelaide, Ausralia, 6 December 3 www.mssanz.org.au/modsim3 Robus esimaion based on he firs- and hird-momen resricions of he power ransformaion Nawaa,

More information

Object tracking: Using HMMs to estimate the geographical location of fish

Object tracking: Using HMMs to estimate the geographical location of fish Objec racking: Using HMMs o esimae he geographical locaion of fish 02433 - Hidden Markov Models Marin Wæver Pedersen, Henrik Madsen Course week 13 MWP, compiled June 8, 2011 Objecive: Locae fish from agging

More information

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.

More information

Introduction to Mobile Robotics

Introduction to Mobile Robotics Inroducion o Mobile Roboics Bayes Filer Kalman Filer Wolfram Burgard Cyrill Sachniss Giorgio Grisei Maren Bennewiz Chrisian Plagemann Bayes Filer Reminder Predicion bel p u bel d Correcion bel η p z bel

More information

Graphical Event Models and Causal Event Models. Chris Meek Microsoft Research

Graphical Event Models and Causal Event Models. Chris Meek Microsoft Research Graphical Even Models and Causal Even Models Chris Meek Microsof Research Graphical Models Defines a join disribuion P X over a se of variables X = X 1,, X n A graphical model M =< G, Θ > G =< X, E > is

More information

Tracking. Many slides adapted from Kristen Grauman, Deva Ramanan

Tracking. Many slides adapted from Kristen Grauman, Deva Ramanan Tracking Man slides adaped from Krisen Grauman Deva Ramanan Coures G. Hager Coures G. Hager J. Kosecka cs3b Adapive Human-Moion Tracking Acquisiion Decimaion b facor 5 Moion deecor Grascale convers. Image

More information

Tracking. Many slides adapted from Kristen Grauman, Deva Ramanan

Tracking. Many slides adapted from Kristen Grauman, Deva Ramanan Tracking Man slides adaped from Krisen Grauman Deva Ramanan Coures G. Hager Coures G. Hager J. Kosecka cs3b Adapive Human-Moion Tracking Acquisiion Decimaion b facor 5 Moion deecor Grascale convers. Image

More information

Temporal probability models

Temporal probability models Temporal probabiliy models CS194-10 Fall 2011 Lecure 25 CS194-10 Fall 2011 Lecure 25 1 Ouline Hidden variables Inerence: ilering, predicion, smoohing Hidden Markov models Kalman ilers (a brie menion) Dynamic

More information

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance

More information

Two Coupled Oscillators / Normal Modes

Two Coupled Oscillators / Normal Modes Lecure 3 Phys 3750 Two Coupled Oscillaors / Normal Modes Overview and Moivaion: Today we ake a small, bu significan, sep owards wave moion. We will no ye observe waves, bu his sep is imporan in is own

More information

Ensamble methods: Boosting

Ensamble methods: Boosting Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room

More information

Maximum Likelihood Parameter Estimation in State-Space Models

Maximum Likelihood Parameter Estimation in State-Space Models Maximum Likelihood Parameer Esimaion in Sae-Space Models Arnaud Douce Deparmen of Saisics, Oxford Universiy Universiy College London 4 h Ocober 212 A. Douce (UCL Maserclass Oc. 212 4 h Ocober 212 1 / 32

More information

Self assessment due: Monday 4/29/2019 at 11:59pm (submit via Gradescope)

Self assessment due: Monday 4/29/2019 at 11:59pm (submit via Gradescope) CS 188 Spring 2019 Inroducion o Arificial Inelligence Wrien HW 10 Due: Monday 4/22/2019 a 11:59pm (submi via Gradescope). Leave self assessmen boxes blank for his due dae. Self assessmen due: Monday 4/29/2019

More information

Ensamble methods: Bagging and Boosting

Ensamble methods: Bagging and Boosting Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par

More information

Ordinary dierential equations

Ordinary dierential equations Chaper 5 Ordinary dierenial equaions Conens 5.1 Iniial value problem........................... 31 5. Forward Euler's mehod......................... 3 5.3 Runge-Kua mehods.......................... 36

More information

Temporal probability models. Chapter 15, Sections 1 5 1

Temporal probability models. Chapter 15, Sections 1 5 1 Temporal probabiliy models Chaper 15, Secions 1 5 Chaper 15, Secions 1 5 1 Ouline Time and uncerainy Inerence: ilering, predicion, smoohing Hidden Markov models Kalman ilers (a brie menion) Dynamic Bayesian

More information

Georey E. Hinton. University oftoronto. Technical Report CRG-TR February 22, Abstract

Georey E. Hinton. University oftoronto.   Technical Report CRG-TR February 22, Abstract Parameer Esimaion for Linear Dynamical Sysems Zoubin Ghahramani Georey E. Hinon Deparmen of Compuer Science Universiy oftorono 6 King's College Road Torono, Canada M5S A4 Email: zoubin@cs.orono.edu Technical

More information

Chapter 2. First Order Scalar Equations

Chapter 2. First Order Scalar Equations Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.

More information

WATER LEVEL TRACKING WITH CONDENSATION ALGORITHM

WATER LEVEL TRACKING WITH CONDENSATION ALGORITHM WATER LEVEL TRACKING WITH CONDENSATION ALGORITHM Shinsuke KOBAYASHI, Shogo MURAMATSU, Hisakazu KIKUCHI, Masahiro IWAHASHI Dep. of Elecrical and Elecronic Eng., Niigaa Universiy, 8050 2-no-cho Igarashi,

More information

Pattern Classification (VI) 杜俊

Pattern Classification (VI) 杜俊 Paern lassificaion VI 杜俊 jundu@usc.edu.cn Ouline Bayesian Decision Theory How o make he oimal decision? Maximum a oserior MAP decision rule Generaive Models Join disribuion of observaion and label sequences

More information

Exponential model. The Gibbs sampler is described for ecological examples by Clark et al. (2003, 2004) and

Exponential model. The Gibbs sampler is described for ecological examples by Clark et al. (2003, 2004) and APPEDIX: GIBBS SAMPLER FOR BAYESIA SAE-SPACE MODELS Eponenial model he Gis sampler is descried or ecological eamples Clark e al. (3 4) and Wikle (3). I involves alernael sampling rom each o he condiional

More information

THE SINE INTEGRAL. x dt t

THE SINE INTEGRAL. x dt t THE SINE INTEGRAL As one learns in elemenary calculus, he limi of sin(/ as vanishes is uniy. Furhermore he funcion is even and has an infinie number of zeros locaed a ±n for n1,,3 Is plo looks like his-

More information

CSE-473. A Gentle Introduction to Particle Filters

CSE-473. A Gentle Introduction to Particle Filters CSE-473 A Genle Inroducion o Paricle Filers Bayes Filers for Robo Localizaion Dieer Fo 2 Bayes Filers: Framework Given: Sream of observaions z and acion daa u: d Sensor model Pz. = { u, z2, u 1, z 1 Dynamics

More information

Lecture Notes 2. The Hilbert Space Approach to Time Series

Lecture Notes 2. The Hilbert Space Approach to Time Series Time Series Seven N. Durlauf Universiy of Wisconsin. Basic ideas Lecure Noes. The Hilber Space Approach o Time Series The Hilber space framework provides a very powerful language for discussing he relaionship

More information

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis Speaker Adapaion Techniques For Coninuous Speech Using Medium and Small Adapaion Daa Ses Consaninos Boulis Ouline of he Presenaion Inroducion o he speaker adapaion problem Maximum Likelihood Sochasic Transformaions

More information

Chapter 4. Truncation Errors

Chapter 4. Truncation Errors Chaper 4. Truncaion Errors and he Taylor Series Truncaion Errors and he Taylor Series Non-elemenary funcions such as rigonomeric, eponenial, and ohers are epressed in an approimae fashion using Taylor

More information

Lecture 4: November 13

Lecture 4: November 13 Compuaional Learning Theory Fall Semeser, 2017/18 Lecure 4: November 13 Lecurer: Yishay Mansour Scribe: Guy Dolinsky, Yogev Bar-On, Yuval Lewi 4.1 Fenchel-Conjugae 4.1.1 Moivaion Unil his lecure we saw

More information

20. Applications of the Genetic-Drift Model

20. Applications of the Genetic-Drift Model 0. Applicaions of he Geneic-Drif Model 1) Deermining he probabiliy of forming any paricular combinaion of genoypes in he nex generaion: Example: If he parenal allele frequencies are p 0 = 0.35 and q 0

More information

Then. 1 The eigenvalues of A are inside R = n i=1 R i. 2 Union of any k circles not intersecting the other (n k)

Then. 1 The eigenvalues of A are inside R = n i=1 R i. 2 Union of any k circles not intersecting the other (n k) Ger sgorin Circle Chaper 9 Approimaing Eigenvalues Per-Olof Persson persson@berkeley.edu Deparmen of Mahemaics Universiy of California, Berkeley Mah 128B Numerical Analysis (Ger sgorin Circle) Le A be

More information

Smoothing. Backward smoother: At any give T, replace the observation yt by a combination of observations at & before T

Smoothing. Backward smoother: At any give T, replace the observation yt by a combination of observations at & before T Smoohing Consan process Separae signal & noise Smooh he daa: Backward smooher: A an give, replace he observaion b a combinaion of observaions a & before Simple smooher : replace he curren observaion wih

More information

Testing the Random Walk Model. i.i.d. ( ) r

Testing the Random Walk Model. i.i.d. ( ) r he random walk heory saes: esing he Random Walk Model µ ε () np = + np + Momen Condiions where where ε ~ i.i.d he idea here is o es direcly he resricions imposed by momen condiions. lnp lnp µ ( lnp lnp

More information

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes Represening Periodic Funcions by Fourier Series 3. Inroducion In his Secion we show how a periodic funcion can be expressed as a series of sines and cosines. We begin by obaining some sandard inegrals

More information

Financial Econometrics Kalman Filter: some applications to Finance University of Evry - Master 2

Financial Econometrics Kalman Filter: some applications to Finance University of Evry - Master 2 Financial Economerics Kalman Filer: some applicaions o Finance Universiy of Evry - Maser 2 Eric Bouyé January 27, 2009 Conens 1 Sae-space models 2 2 The Scalar Kalman Filer 2 21 Presenaion 2 22 Summary

More information

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Simulaion-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Week Descripion Reading Maerial 2 Compuer Simulaion of Dynamic Models Finie Difference, coninuous saes, discree ime Simple Mehods Euler Trapezoid

More information

Probabilistic learning

Probabilistic learning Probabilisic learning Charles Elkan November 8, 2012 Imporan: These lecure noes are based closely on noes wrien by Lawrence Saul. Tex may be copied direcly from his noes, or paraphrased. Also, hese ypese

More information

CS 4495 Computer Vision Tracking 1- Kalman,Gaussian

CS 4495 Computer Vision Tracking 1- Kalman,Gaussian CS 4495 Compuer Vision A. Bobick CS 4495 Compuer Vision - KalmanGaussian Aaron Bobick School of Ineracive Compuing CS 4495 Compuer Vision A. Bobick Adminisrivia S5 will be ou his Thurs Due Sun Nov h :55pm

More information

Answers to QUIZ

Answers to QUIZ 18441 Answers o QUIZ 1 18441 1 Le P be he proporion of voers who will voe Yes Suppose he prior probabiliy disribuion of P is given by Pr(P < p) p for 0 < p < 1 You ake a poll by choosing nine voers a random,

More information

Testing for a Single Factor Model in the Multivariate State Space Framework

Testing for a Single Factor Model in the Multivariate State Space Framework esing for a Single Facor Model in he Mulivariae Sae Space Framework Chen C.-Y. M. Chiba and M. Kobayashi Inernaional Graduae School of Social Sciences Yokohama Naional Universiy Japan Faculy of Economics

More information

Computer-Aided Analysis of Electronic Circuits Course Notes 3

Computer-Aided Analysis of Electronic Circuits Course Notes 3 Gheorghe Asachi Technical Universiy of Iasi Faculy of Elecronics, Telecommunicaions and Informaion Technologies Compuer-Aided Analysis of Elecronic Circuis Course Noes 3 Bachelor: Telecommunicaion Technologies

More information

Recent Developments In Evolutionary Data Assimilation And Model Uncertainty Estimation For Hydrologic Forecasting Hamid Moradkhani

Recent Developments In Evolutionary Data Assimilation And Model Uncertainty Estimation For Hydrologic Forecasting Hamid Moradkhani Feb 6-8, 208 Recen Developmens In Evoluionary Daa Assimilaion And Model Uncerainy Esimaion For Hydrologic Forecasing Hamid Moradkhani Cener for Complex Hydrosysems Research Deparmen of Civil, Consrucion

More information

Review - Quiz # 1. 1 g(y) dy = f(x) dx. y x. = u, so that y = xu and dy. dx (Sometimes you may want to use the substitution x y

Review - Quiz # 1. 1 g(y) dy = f(x) dx. y x. = u, so that y = xu and dy. dx (Sometimes you may want to use the substitution x y Review - Quiz # 1 (1) Solving Special Tpes of Firs Order Equaions I. Separable Equaions (SE). d = f() g() Mehod of Soluion : 1 g() d = f() (The soluions ma be given implicil b he above formula. Remember,

More information

Ensemble Confidence Estimates Posterior Probability

Ensemble Confidence Estimates Posterior Probability Ensemble Esimaes Poserior Probabiliy Michael Muhlbaier, Aposolos Topalis, and Robi Polikar Rowan Universiy, Elecrical and Compuer Engineering, Mullica Hill Rd., Glassboro, NJ 88, USA {muhlba6, opali5}@sudens.rowan.edu

More information

Lecture 33: November 29

Lecture 33: November 29 36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure

More information

Probabilistic Robotics Sebastian Thrun-- Stanford

Probabilistic Robotics Sebastian Thrun-- Stanford robabilisic Roboics Sebasian Thrn-- Sanford Inrodcion robabiliies Baes rle Baes filers robabilisic Roboics Ke idea: Eplici represenaion of ncerain sing he calcls of probabili heor ercepion sae esimaion

More information

Math 10B: Mock Mid II. April 13, 2016

Math 10B: Mock Mid II. April 13, 2016 Name: Soluions Mah 10B: Mock Mid II April 13, 016 1. ( poins) Sae, wih jusificaion, wheher he following saemens are rue or false. (a) If a 3 3 marix A saisfies A 3 A = 0, hen i canno be inverible. True.

More information

1 Review of Zero-Sum Games

1 Review of Zero-Sum Games COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any

More information

The Optimal Stopping Time for Selling an Asset When It Is Uncertain Whether the Price Process Is Increasing or Decreasing When the Horizon Is Infinite

The Optimal Stopping Time for Selling an Asset When It Is Uncertain Whether the Price Process Is Increasing or Decreasing When the Horizon Is Infinite American Journal of Operaions Research, 08, 8, 8-9 hp://wwwscirporg/journal/ajor ISSN Online: 60-8849 ISSN Prin: 60-8830 The Opimal Sopping Time for Selling an Asse When I Is Uncerain Wheher he Price Process

More information

Differential Equations

Differential Equations Mah 21 (Fall 29) Differenial Equaions Soluion #3 1. Find he paricular soluion of he following differenial equaion by variaion of parameer (a) y + y = csc (b) 2 y + y y = ln, > Soluion: (a) The corresponding

More information

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED 0.1 MAXIMUM LIKELIHOOD ESTIMATIO EXPLAIED Maximum likelihood esimaion is a bes-fi saisical mehod for he esimaion of he values of he parameers of a sysem, based on a se of observaions of a random variable

More information

A Bayesian Approach to Spectral Analysis

A Bayesian Approach to Spectral Analysis Chirped Signals A Bayesian Approach o Specral Analysis Chirped signals are oscillaing signals wih ime variable frequencies, usually wih a linear variaion of frequency wih ime. E.g. f() = A cos(ω + α 2

More information

Laplace Transforms. Examples. Is this equation differential? y 2 2y + 1 = 0, y 2 2y + 1 = 0, (y ) 2 2y + 1 = cos x,

Laplace Transforms. Examples. Is this equation differential? y 2 2y + 1 = 0, y 2 2y + 1 = 0, (y ) 2 2y + 1 = cos x, Laplace Transforms Definiion. An ordinary differenial equaion is an equaion ha conains one or several derivaives of an unknown funcion which we call y and which we wan o deermine from he equaion. The equaion

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Noes for EE7C Spring 018: Convex Opimizaion and Approximaion Insrucor: Moriz Hard Email: hard+ee7c@berkeley.edu Graduae Insrucor: Max Simchowiz Email: msimchow+ee7c@berkeley.edu Ocober 15, 018 3

More information

7630 Autonomous Robotics Probabilistic Localisation

7630 Autonomous Robotics Probabilistic Localisation 7630 Auonomous Roboics Probabilisic Localisaion Principles of Probabilisic Localisaion Paricle Filers for Localisaion Kalman Filer for Localisaion Based on maerial from R. Triebel, R. Käsner, R. Siegwar,

More information

Description of the MS-Regress R package (Rmetrics)

Description of the MS-Regress R package (Rmetrics) Descriion of he MS-Regress R ackage (Rmerics) Auhor: Marcelo Perlin PhD Suden / ICMA Reading Universiy Email: marceloerlin@gmail.com / m.erlin@icmacenre.ac.uk The urose of his documen is o show he general

More information

CHAPTER 2: Mathematics for Microeconomics

CHAPTER 2: Mathematics for Microeconomics CHAPTER : Mahemaics for Microeconomics The problems in his chaper are primarily mahemaical. They are inended o give sudens some pracice wih he conceps inroduced in Chaper, bu he problems in hemselves offer

More information

ACE 564 Spring Lecture 7. Extensions of The Multiple Regression Model: Dummy Independent Variables. by Professor Scott H.

ACE 564 Spring Lecture 7. Extensions of The Multiple Regression Model: Dummy Independent Variables. by Professor Scott H. ACE 564 Spring 2006 Lecure 7 Exensions of The Muliple Regression Model: Dumm Independen Variables b Professor Sco H. Irwin Readings: Griffihs, Hill and Judge. "Dumm Variables and Varing Coefficien Models

More information

Linear Gaussian State Space Models

Linear Gaussian State Space Models Linear Gaussian Sae Space Models Srucural Time Series Models Level and Trend Models Basic Srucural Model (BSM Dynamic Linear Models Sae Space Model Represenaion Level, Trend, and Seasonal Models Time Varying

More information

OBJECTIVES OF TIME SERIES ANALYSIS

OBJECTIVES OF TIME SERIES ANALYSIS OBJECTIVES OF TIME SERIES ANALYSIS Undersanding he dynamic or imedependen srucure of he observaions of a single series (univariae analysis) Forecasing of fuure observaions Asceraining he leading, lagging

More information

Tracking. Announcements

Tracking. Announcements Tracking Tuesday, Nov 24 Krisen Grauman UT Ausin Announcemens Pse 5 ou onigh, due 12/4 Shorer assignmen Auo exension il 12/8 I will no hold office hours omorrow 5 6 pm due o Thanksgiving 1 Las ime: Moion

More information

Distribution of Estimates

Distribution of Estimates Disribuion of Esimaes From Economerics (40) Linear Regression Model Assume (y,x ) is iid and E(x e )0 Esimaion Consisency y α + βx + he esimaes approach he rue values as he sample size increases Esimaion

More information

ADDITIONAL PROBLEMS (a) Find the Fourier transform of the half-cosine pulse shown in Fig. 2.40(a). Additional Problems 91

ADDITIONAL PROBLEMS (a) Find the Fourier transform of the half-cosine pulse shown in Fig. 2.40(a). Additional Problems 91 ddiional Problems 9 n inverse relaionship exiss beween he ime-domain and freuency-domain descripions of a signal. Whenever an operaion is performed on he waveform of a signal in he ime domain, a corresponding

More information

Notes on Kalman Filtering

Notes on Kalman Filtering Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren

More information

Recursive Least-Squares Fixed-Interval Smoother Using Covariance Information based on Innovation Approach in Linear Continuous Stochastic Systems

Recursive Least-Squares Fixed-Interval Smoother Using Covariance Information based on Innovation Approach in Linear Continuous Stochastic Systems 8 Froniers in Signal Processing, Vol. 1, No. 1, July 217 hps://dx.doi.org/1.2266/fsp.217.112 Recursive Leas-Squares Fixed-Inerval Smooher Using Covariance Informaion based on Innovaion Approach in Linear

More information

Expectation- Maximization & Baum-Welch. Slides: Roded Sharan, Jan 15; revised by Ron Shamir, Nov 15

Expectation- Maximization & Baum-Welch. Slides: Roded Sharan, Jan 15; revised by Ron Shamir, Nov 15 Expecaion- Maximizaion & Baum-Welch Slides: Roded Sharan, Jan 15; revised by Ron Shamir, Nov 15 1 The goal Inpu: incomplee daa originaing from a probabiliy disribuion wih some unknown parameers Wan o find

More information

2.160 System Identification, Estimation, and Learning. Lecture Notes No. 8. March 6, 2006

2.160 System Identification, Estimation, and Learning. Lecture Notes No. 8. March 6, 2006 2.160 Sysem Idenificaion, Esimaion, and Learning Lecure Noes No. 8 March 6, 2006 4.9 Eended Kalman Filer In many pracical problems, he process dynamics are nonlinear. w Process Dynamics v y u Model (Linearized)

More information

Probabilistic Robotics SLAM

Probabilistic Robotics SLAM Probabilisic Roboics SLAM The SLAM Problem SLAM is he process by which a robo builds a map of he environmen and, a he same ime, uses his map o compue is locaion Localizaion: inferring locaion given a map

More information

Advanced FDTD Algorithms

Advanced FDTD Algorithms EE 5303 Elecromagneic Analsis Using Finie Difference Time Domain Lecure #5 Advanced FDTD Algorihms Lecure 5 These noes ma conain coprighed maerial obained under fair use rules. Disribuion of hese maerials

More information

מקורות לחומר בשיעור ספר הלימוד: Forsyth & Ponce מאמרים שונים חומר באינטרנט! פרק פרק 18

מקורות לחומר בשיעור ספר הלימוד: Forsyth & Ponce מאמרים שונים חומר באינטרנט! פרק פרק 18 עקיבה מקורות לחומר בשיעור ספר הלימוד: פרק 5..2 Forsh & once פרק 8 מאמרים שונים חומר באינטרנט! Toda Tracking wih Dnamics Deecion vs. Tracking Tracking as probabilisic inference redicion and Correcion Linear

More information

Recent Developments in the Unit Root Problem for Moving Averages

Recent Developments in the Unit Root Problem for Moving Averages Recen Developmens in he Uni Roo Problem for Moving Averages Richard A. Davis Colorado Sae Universiy Mei-Ching Chen Chaoyang Insiue of echnology homas Miosch Universiy of Groningen Non-inverible MA() Model

More information

Západočeská Univerzita v Plzni, Czech Republic and Groupe ESIEE Paris, France

Západočeská Univerzita v Plzni, Czech Republic and Groupe ESIEE Paris, France ADAPTIVE SIGNAL PROCESSING USING MAXIMUM ENTROPY ON THE MEAN METHOD AND MONTE CARLO ANALYSIS Pavla Holejšovsá, Ing. *), Z. Peroua, Ing. **), J.-F. Bercher, Prof. Assis. ***) Západočesá Univerzia v Plzni,

More information

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t...

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t... Mah 228- Fri Mar 24 5.6 Marix exponenials and linear sysems: The analogy beween firs order sysems of linear differenial equaions (Chaper 5) and scalar linear differenial equaions (Chaper ) is much sronger

More information

Announcements. Recap: Filtering. Recap: Reasoning Over Time. Example: State Representations for Robot Localization. Particle Filtering

Announcements. Recap: Filtering. Recap: Reasoning Over Time. Example: State Representations for Robot Localization. Particle Filtering Inroducion o Arificial Inelligence V22.0472-001 Fall 2009 Lecure 18: aricle & Kalman Filering Announcemens Final exam will be a 7pm on Wednesday December 14 h Dae of las class 1.5 hrs long I won ask anyhing

More information