Bias-Variance Error Bounds for Temporal Difference Updates

Size: px
Start display at page:

Download "Bias-Variance Error Bounds for Temporal Difference Updates"

Transcription

1 Bias-Variance Bounds for Temporal Difference Updaes Michael Kearns AT&T Labs Sainder Singh AT&T Labs Absrac We give he firs rigorous upper bounds on he error of emporal difference (d) algorihms for policy evaluaion as a funcion of he amoun of experience. These upper bounds prove exponenially fas convergence, wih boh he rae of convergence and he asympoe srongly dependen on he lengh of he backups k or he parameer. Our bounds give formal verificaion o he long-sanding inuiion ha d mehods are subjec o a bias-variance rade-off, and hey lead o schedules for k and ha are prediced o be beer han any fixed values for hese parameers. We give preliminary experimenal confirmaion of our heory for a version of he random walk problem. Inroducion In he policy evaluaion problem, we mus predic he expeced discouned reurn (or value) for a fixed policy, given only he abiliy o generae experience in an unknown Markov decision process (MDP) M. A family of well-sudied emporal difference (or d) [3] algorihms have been developed for his problem ha make use of repeaed rajecories under from he sae(s) of ineres, and perform ieraive updaes o he value funcion. The main difference beween he d varians lies in how far hey look ahead in he rajecories. The d(k) family of algorihms use he firs k rewards and he (curren) value predicion a he (k +)s sae reached in making is updae. The more commonly used d() family of algorihms use exponenially weighed sums of d(k) updaes (wih decay parameer ). The smaller he value for k or, he less he algorihm depends on he acual rewards received in he rajecory, and he more i depends on he curren predicions for he value funcion. Conversely, he larger he value for k or, he more he algorihm depends on he acual rewards obained, wih he curren value funcion playing a lessened role. The exreme cases of d(k = ) and d( = )become he Mone Carlo algorihm, which updaes each predicion o be he average of he discouned reurns in he rajecories. A long-sanding quesion is wheher i is beer o use large or small values of he parameers k and. Wakins [5] informally discusses he rade-off ha his decision gives rise o: larger values for he d parameers suffer larger variance in he updaes (since more sochasic reward erms appear), bu also enjoy lower bias (since he error in he curren value funcion predicions have less influence). This argumen has largely remained an inuiion. However, some conclusions arising from his inuiion for insance, ha inermediae values of k and ofen yield he bes performance in he shor erm have been borne ou experimenally [4, 2]. In his paper, we provide he firs rigorous upper bounds on he error in he value funcions of

2 he d algorihms as a funcion of he number of rajecories used. In oher words, we give bounds on he learning curves of d mehods ha hold for any MDP. These upper bounds decay exponenially fas, and are obained by firs deriving a one-sep recurrence relaing he errors before and afer a d updae, and hen ieraing his recurrence for he desired number of seps. Of paricular ineres is he form of our bounds, since i formalizes he rade-off discussed above he bounds consis of erms ha are monoonically growing wih k and (corresponding o he increased variance), and erms ha are monoonically shrinking wih k and (corresponding o he decreased influence of he curren error). Overall, our bounds provide he following conribuions and predicions:. A formal heoreical explanaion of he bias-variance rade-off in muli-sep d updaes; 2. A proof of exponenially fas raes of convergence for any fixed k or ; 3. A rigorous upper bound ha predics ha larger values of k and lead o faser convergence, buohigher asympoic errror; 4. Formal explanaion of he superioriy of inermediae values of k and (U-shaped curves) for any fixed number of ieraions; 5. Derivaion of a decreasing schedule of k and ha our bound predics should bea any fixed value of hese parameers. Furhermore, we provide some preliminary experimenal confirmaion of our heory for he random walk problem. We noe ha some of he findings above were conjecured by Singh and Dayan [2] hrough analysis of specific MDPs. 2 Technical Preliminaries Le M = (P; R) be an MDP consising of he ransiion probabiliies P (js; a) and he reward disribuions R(js). For any policy in M, and any sar sae s 0,arajecory generaed by saring from s 0 is a random variable ha is an infinie sequence of saes and rewards: =(s 0 ;r 0 )! (s ;r )! (s 2 ;r 2 )!. Here each random reward r i is disribued according o R(js i ), and each sae s i+ is disribued according o P (js i ;(s i )). For simpliciy we will assume ha he suppor of R(js i ) is [,; +]. However, all of our resuls easily generalize o he case of bounded variance. We now recall he sandard d(k) (also known as k-sep backup) and d() mehods for updaing an esimae of he value funcion. Given a rajecory generaed by from s 0,and given an esimae ^V () for he value funcion V (), for any naural number k we define d(k; ; ^V ()) = (, ) ^V (s 0 )+ r 0 + r + + k, r k, + k ^V (s k ) : The d(k) updae based on is simply ^V (s 0 ) d(k; ; ^V ()). I is implici ha he updae is always applied o he esimae a he iniial sae of he rajecory, and we regard he discoun facor and he learning rae as being fixed. For any 2 [0; ], hed() updae can now be easily expressed as an infinie linear combinaion of he d(k) updaes: d(; ; ^V ()) = X (, ) k, d(k; ; ^V ()): Given a sequence ; 2 ; 3 ;:::, we can simply apply eiher ype of d updae sequenially. In eiher case, as eiher k becomes large or approaches, he updaes approach a Mone Carlo mehod, in which we use each rajecory i enirely, and ignore our curren esimae ^V (). Ask becomes small or approaches 0, we rely heavily on he esimae ^V (), and

3 effecively use only a few seps of each i. The common inuiionis ha early in he sequence of udpaes, he esimae ^V () is poor, and we are beer off choosing k large or near. However, since he rajecories i do obey he saisics of, he value funcion esimaes will evenually improve, a which poin we may be beer off boosrapping by choosing small k or. In order o provide a rigorous analysis of his inuiion, we will sudy a framework which we call phased d updaes. This framework is inended o simplify he complexiies of he moving average inroduced by he learning rae. In each phase, we are given n rajecories under from every sae s, wheren is a parameer of he analysis. Thus, phase consiss of a se S() =f s ())g i s;i, wheres ranges over all saes, i ranges from o n, and s () i is an independen random rajecory generaed by saring from sae s. In phase, phased d averages all n of he rajecories in S() ha sar from sae s o obain is updae of he value funcion esimae for s. In oher words, he d(k) updaes become nx ^V + (s) (=n) i= r i 0 + ri + + k, r i k, + k ^V (si k ) where he r i j are he rewards along rajecory s (), i andsi is he kh sae reached along ha k rajecory. The d() updaes become nx X! ^V + (s) (=n) (, ) r k, i 0 + ri + + k, r i k, + k ^V (si ) k i= Phased d updaes wih a fixed value of n are analogous o sandard d updaes wih a consan learning rae []. In he ensuing secions, we provide a rigorous upper bound on he error in he value funcion esimaes of phased d updaes as a funcion of he number of phases. This upper bound clearly capures he inuiions expressed above. 3 Bounding he of d Updaes Theorem (Phased d(k) Recurrence) Le S() be he se of rajecories generaed by in phase (n rajecories from each sae), le ^V () be he value funcion esimae of phased d(k) afer phase, and le = max s fj ^V (s),v (s)jg. Then for any >>0, wih probabiliy a leas,, r, k 3 log(k=) + k,: (), n Here he error, afer phase, is fixed, and he probabiliy is aken over only he rajecories in S(). Proof:(Skech) We begin by wriing V (s) = E[r 0 + r + + k, r k, + k V (s k )] = E[r 0 ]+E[r ]++ k, E[r k,] + k E[V (s k )]: Here he expecaions are over a random rajecory under ; hus E[r`] (` k, ) denoes he expeced value of he `h reward received, while E[V (s k )] is he expeced value of he rue value P funcion a he kh sae reached. The phased d(k) updae sums he erms n `(=n) i= rì, whose expecaions are exacly he `E[r`] appearing above. By a sandard large deviaion p analysis (omied), he probabiliy ha any of hese erms deviae by more han = 3 log(k=)=n from heir expeced values is a mos. If no such deviaion occurs, he oal conribuion o he error in he value funcion esimae is bounded

4 by ((, k )=(, )), giving rise o he variance erm in P our overall bound above. The remainder of he phased d(k) updae is simply k n (=n) ^V i=, (si ). Bu since k j ^V, (si ), V (s i )j k k, by definiion, he conribuion o he error is a mos k,, which is he bias erm of he bound. We noe ha a similar argumen leads o bounds in expecaion raher han he PAC-syle bounds given here. 2 Le us ake a brief momen o analyze he qualiaive behavior of Equaion () as a funcion of k. For large values of k, he quaniy k becomes negligible, and he bound is approximaely (=(, )) p 3 log(k=)=n, giving almos all he weigh o he error incurred by variance in he firs k rewards, and negligible weigh o he error in our curren value funcion. A he oher exreme, when k =our reward variance conribues error only p 3 log(=)=n, bu he error in our curren value funcion has weigh. Thus, he firs erm increases wih k, while he second erm decreases wih k, in a manner ha formalizes he inuiive rade-off ha one faces when choosing beween longer or shorer backups. Equaion () describes he effec of a single phase of d(k) backups, bu we can ierae his recurrence over many phases o derive an upper bound on he full learning curve for any value of k. Assuming ha he recurrence holds for consecuive seps, and assuming 0 = wihou loss of generaliy, soluion of he recurrence (deails omied) yields, k, p 3 log(k=)=n + k : (2) This bound makes a number of predicions abou he effecs of differen values for k. Firs of all, as approaches infiniy, he bound on approaches he value (=(, ))p 3 log(k=)=n, which increases wih k. Thus, he bound predics ha he asympoic error of phased d(k) updaes is larger for larger k. On he oher hand, he rae of convergence o his asympoe is k, which is always exponenially fas, bu faser for larger k. Thus, in choosing a fixed value of k, we mus choose beween having eiher rapid convergence o a worse asympoe, or slower convergence o a beer asympoe. This predicion is illusraed graphically in Figure (a), where wih all of he parameers besides k and fixed (namely,,,andn),we have ploedheboundof Equaion (2)as a funcionof for several differen choices of k. Noe ha while he plos of Figure (a) were obained by choosing fixed values for k and ieraing he recurrence of Equaion (), a each phase we can insead use Equaion () o choose he value of k ha maximizes he prediced decrease in error, +.Inoher words, he recurrence immediaely yields a schedule for k, along wih an upper boundon he learning curve for his schedule ha ouperforms he upper bound on he learning curve for any fixed value of k. The learning curve for he schedule is also shown in Figure (a), and Figure (b) plos he schedule iself. Anoher ineresing se of plos is obained by fixing he number of phases, and compuing for each k he error afer phases using d(k) updaes ha is prediced by Equaion (2). Such plos are given in Figure (c), and hey clearly predic a unique minimum ha is, an opimal value of k for each fixed (his can also be verified analyically from equaion 2). For moderae values of, valuesofk ha are oo small suffer from heir overemphasis on a sill-inaccurae value funcion approximaion, while values of k ha are oo large suffer from heir refusal o boosrap. Of course, as increases, he opimal value of k decreases, since small values of k have ime o reach heir superior asympoes. We now go on o provide a similar analysis for he d() family of updaes, beginning wih he analogue o Theorem. Formally, we can apply Theorem by choosing = 0 =(N ), wheren is he number of saes in he MDP. Then wih probabiliy a leas, 0, he bound of Equaion () will hold a every sae for consecuive seps.

5 (a) (b) (c) 25 5 scheduled k k= k=25 0 k=2 Opimal k = =2 =5 =00 = k Figure : (a) Upper bounds on he learning curves of phased d(k) for several values of k, as a funcion of he number of phases (parameers n = 3000, = 0:9, = 0:). Noe ha larger values of k lead o more rapid convergence, bu o higher asympoic errors. Boh he heory and he curves sugges a (decreasing) schedule for k, inuiively obained by always jumping o he learning curve ha enjoys he greaes one-sep decrease from he curren error. This schedule can be efficienly compued from he analyical upper bounds, and leads o he bes (lowes) of he learning curves ploed, which is significanly beer han for any fixed k. (b) The schedulefor k derived from he heory as a funcion of he number of phases. (c) For several values of he number of phases, he upper bound on for d(k) as a funcion of k. These curves show he prediced rade-off, wih a unique opimal value for k idenified unil is sufficienly large o permi -sep backups o converge o heir opimal asympoes. Theorem 2 (Phased d() Recurrence) Le S() be he se of rajecories generaed by in phase (n rajecories from each sae), le ^V () be he value funcion esimae of phased d() afer phase, and le = max s fj ^V (s),v (s)jg. Then for any >>0, wih probabiliy a leas,, ( r ), () k 3 log(k=) min + ()k (, ) + k, n,,,: (3) Here he error, afer phase, is fixed, and he probabiliy is aken over only he rajecories in S(). We omi he proof of his heorem, bu i roughly follows ha of Theorem. Tha proof exploied he fac ha in d(k) updaes, we only need o apply large deviaion bounds o he rewards of a finie number (k) of averaged rajecory seps. In d(), all of he rewards conribue o he updae. However, we can always choose o bound he deviaions of he firs k seps, for any value of k, and assume maximum variance for he remainder (whose weigh diminishes rapidly as we increase k). This logic is he source of he min k fg erm of he bound. One can view Equaion (3) as a variaional upper bound, in he sense ha i provides a familyof upper bounds, one for each k, and hen minimizes over he variaional parameer k. The reader can verify ha he erms appearing in Equaion (3) exhibi a rade-off as a funcion of analogous o ha exhibied by Equaion () as a funcion of k. In he ineres of breviy, we move direcly o he d() analogue of Equaion (2). I will be noaionally convenien o define k = argmin k ff ()g, wheref () is he funcion appearing inside he min k fg in Equaion (3). (Here we regard all parameers oher han as fixed.) I can be shown ha for 0 =, repeaed ieraion of Equaion (3) yields he -phase inequaliy where, b a + b, b r a =, ()k 3 log(k =) + ()k, n, b = (, ), (4)

6 (a) (b) = =2 Scheduled λ λ = λ = λ = λ = 0.4 λ = 0.2 λ = 0.0 λ = 0. =5 =0 =25 = λ Figure 2: (a) Upper bounds on he learning curves of phased d() for several values of, asa funcion of he number of phases (parameers n = 3000, = 0:9, =0:). The predicions are analogous o hose for d(k) in Figure, and we have again ploed he prediced bes learning curve obained via a decreasing schedule of. (b) For several values of he number of phases, he upper bound on for d() as a funcion of. While Equaion (4) may be more difficul o parse han is d(k) counerpar, he basic predicions and inuiions remain inac. As approaches infiniy, he bound on asympoes a a =(, b ), and he rae of approach o his asympoe is simply b, which is again exponenially fas. Analysis of he derivaive of b wih respec o confirms ha for all <, b is a decreasing funcion of ha is, he larger he, he faser he convergence. Analyically verifying ha he asympoe a =(, b ) increases wih is more difficul due o he presence of k, which involves a minimizaion operaion. However, he learning curve plos of Figure 2(a) clearly show he prediced phenomena increasing yields faser convergence o a worseasympoe. As we didfor he d(k) case, we use our recurrence o derive a schedule for ; Figure 2(a) also shows he prediced improvemen in he learning curve by using such a schedule. Finally, Figure 2(b) again shows he non-monoonic prediced error as a funcion of for a fixed number of phases. 4 Some Experimenal Confirmaion In order o es he various predicions made by our heory, we have performed a number of experimens using phased d(k) on a version of he so-called random walk problem [4]. In his problem, we have a Markov process wih 5 saes arranged in a ring. A each sep, here is probabiliy 0.05 ha we remain in our curren sae, and probabiliy 5 ha we advance one sae clockwise around he ring. (Noe ha since we are only concerned wih he evaluaion of a fixed policy, we have simply defined a Markov process raher han a Markov decision process.) Two adjacen saes on he ring have reward + and, respecively, while he remaining saes have reward 0. The sandard random walk problem has a chain of saes, wih an absorbing sae a each end; here we chose a ring srucure simply o avoid asymmeries in he saes induced by he absorbing saes. To es he heory, we ran a series of simulaions compuing he d(k) esimae of he value funcion in his Markov process. For several differen values of k, we compued he error in he value funcion esimae as a funcion of he number of phases. ( is easily compued, since we can compue he rue value funcion for his simple problem.) The resuling plo in Figure 3(a) is he experimenal analogue of he heoreical predicions in Figure (a). We see ha hese predicions are qualiaively confirmed larger k leads o faser convergence o an inferior asympoe. Given hese empirical learning curves, we can hen compue he empirical schedule ha hey sugges. Namely, o deermine experimenally a schedule for k ha should ouperform (a leas) he values of k we esed in Figure 3(a), we used he empirical learning curves o deermine, for any given value of, which of he empirical curves enjoyed he greaes

7 (a) (b) k=2 00 k=50 k=20 0 k=5 scheduled k Opimal k Figure 3: (a) Empirical learning curves for d(k) for several values of k on he random walk problem (parameers n =40and =0:98). Each plo is averaged over 5000 runs of d(k). Also shown is he learning curve (averaged over 5000 runs) for he empirical schedule compued from he d(k) learning curves, which is beer han any of hese curves. (b) The empirical schedule. one-sep decrease in error when is curren error was (approximaely). This is simply he empirical counerpar of he schedule compuaion suggesed by he heory described above, and he resuling experimenal learning curve for his schedule is also shown in Figure 3(a), and he schedule iself in Figure 3(b). We see ha here are significan improvemens in he learning curve from using he schedule, and ha he form of he schedule is qualiaively similar o he heoreical schedule of Figure (b). 5 Conclusion We have given he firs provable upper bounds on he error of d mehods for policy evaluaion. These upper bounds have exponenial raes of convergence, and clearly ariculae he bias-variance rade-off ha such mehods obey. References [] M. Kearns and S. Singh Finie-Sample Convergence Raes for Q-Learning and Indirec Algorihms NIPS, 998. [2] S. Singh and P. Dayan Analyical Mean Squared Curves for Temporal Difference Learning. Machine Learning, 998. [3] R.S.SuonLearning o Predic by he Mehods of Temporal Differences. Machine Learning, 3, 9-44, 988. [4] R. S. Suon and A. G. Baro Reinforcemen Learning: An Inroducion. MIT Press, 998. [5] C.J.C.H. Wakins Learning from Delayed Rewards. Cambridge Univ., England, 989.

Bias-Variance Error Bounds for Temporal Difference Updates

Bias-Variance Error Bounds for Temporal Difference Updates Bias-Variance Bounds for Temporal Difference Updates Michael Kearns AT&T Labs mkearns@research.att.com Satinder Singh AT&T Labs baveja@research.att.com Abstract We give the first rigorous upper bounds

More information

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle Chaper 2 Newonian Mechanics Single Paricle In his Chaper we will review wha Newon s laws of mechanics ell us abou he moion of a single paricle. Newon s laws are only valid in suiable reference frames,

More information

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.

More information

Vehicle Arrival Models : Headway

Vehicle Arrival Models : Headway Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where

More information

5.1 - Logarithms and Their Properties

5.1 - Logarithms and Their Properties Chaper 5 Logarihmic Funcions 5.1 - Logarihms and Their Properies Suppose ha a populaion grows according o he formula P 10, where P is he colony size a ime, in hours. When will he populaion be 2500? We

More information

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB Elecronic Companion EC.1. Proofs of Technical Lemmas and Theorems LEMMA 1. Le C(RB) be he oal cos incurred by he RB policy. Then we have, T L E[C(RB)] 3 E[Z RB ]. (EC.1) Proof of Lemma 1. Using he marginal

More information

1 Review of Zero-Sum Games

1 Review of Zero-Sum Games COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any

More information

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1 RL Lecure 7: Eligibiliy Traces R. S. Suon and A. G. Baro: Reinforcemen Learning: An Inroducion 1 N-sep TD Predicion Idea: Look farher ino he fuure when you do TD backup (1, 2, 3,, n seps) R. S. Suon and

More information

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance

More information

Robust estimation based on the first- and third-moment restrictions of the power transformation model

Robust estimation based on the first- and third-moment restrictions of the power transformation model h Inernaional Congress on Modelling and Simulaion, Adelaide, Ausralia, 6 December 3 www.mssanz.org.au/modsim3 Robus esimaion based on he firs- and hird-momen resricions of he power ransformaion Nawaa,

More information

Notes for Lecture 17-18

Notes for Lecture 17-18 U.C. Berkeley CS278: Compuaional Complexiy Handou N7-8 Professor Luca Trevisan April 3-8, 2008 Noes for Lecure 7-8 In hese wo lecures we prove he firs half of he PCP Theorem, he Amplificaion Lemma, up

More information

Lecture Notes 2. The Hilbert Space Approach to Time Series

Lecture Notes 2. The Hilbert Space Approach to Time Series Time Series Seven N. Durlauf Universiy of Wisconsin. Basic ideas Lecure Noes. The Hilber Space Approach o Time Series The Hilber space framework provides a very powerful language for discussing he relaionship

More information

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits DOI: 0.545/mjis.07.5009 Exponenial Weighed Moving Average (EWMA) Char Under The Assumpion of Moderaeness And Is 3 Conrol Limis KALPESH S TAILOR Assisan Professor, Deparmen of Saisics, M. K. Bhavnagar Universiy,

More information

20. Applications of the Genetic-Drift Model

20. Applications of the Genetic-Drift Model 0. Applicaions of he Geneic-Drif Model 1) Deermining he probabiliy of forming any paricular combinaion of genoypes in he nex generaion: Example: If he parenal allele frequencies are p 0 = 0.35 and q 0

More information

Some Basic Information about M-S-D Systems

Some Basic Information about M-S-D Systems Some Basic Informaion abou M-S-D Sysems 1 Inroducion We wan o give some summary of he facs concerning unforced (homogeneous) and forced (non-homogeneous) models for linear oscillaors governed by second-order,

More information

Christos Papadimitriou & Luca Trevisan November 22, 2016

Christos Papadimitriou & Luca Trevisan November 22, 2016 U.C. Bereley CS170: Algorihms Handou LN-11-22 Chrisos Papadimiriou & Luca Trevisan November 22, 2016 Sreaming algorihms In his lecure and he nex one we sudy memory-efficien algorihms ha process a sream

More information

Random Walk with Anti-Correlated Steps

Random Walk with Anti-Correlated Steps Random Walk wih Ani-Correlaed Seps John Noga Dirk Wagner 2 Absrac We conjecure he expeced value of random walks wih ani-correlaed seps o be exacly. We suppor his conjecure wih 2 plausibiliy argumens and

More information

Econ107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8)

Econ107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8) I. Definiions and Problems A. Perfec Mulicollineariy Econ7 Applied Economerics Topic 7: Mulicollineariy (Sudenmund, Chaper 8) Definiion: Perfec mulicollineariy exiss in a following K-variable regression

More information

Lecture 33: November 29

Lecture 33: November 29 36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure

More information

4.1 - Logarithms and Their Properties

4.1 - Logarithms and Their Properties Chaper 4 Logarihmic Funcions 4.1 - Logarihms and Their Properies Wha is a Logarihm? We define he common logarihm funcion, simply he log funcion, wrien log 10 x log x, as follows: If x is a posiive number,

More information

Final Spring 2007

Final Spring 2007 .615 Final Spring 7 Overview The purpose of he final exam is o calculae he MHD β limi in a high-bea oroidal okamak agains he dangerous n = 1 exernal ballooning-kink mode. Effecively, his corresponds o

More information

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017 Two Popular Bayesian Esimaors: Paricle and Kalman Filers McGill COMP 765 Sep 14 h, 2017 1 1 1, dx x Bel x u x P x z P Recall: Bayes Filers,,,,,,, 1 1 1 1 u z u x P u z u x z P Bayes z = observaion u =

More information

13.3 Term structure models

13.3 Term structure models 13.3 Term srucure models 13.3.1 Expecaions hypohesis model - Simples "model" a) shor rae b) expecaions o ge oher prices Resul: y () = 1 h +1 δ = φ( δ)+ε +1 f () = E (y +1) (1) =δ + φ( δ) f (3) = E (y +)

More information

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H.

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H. ACE 56 Fall 005 Lecure 5: he Simple Linear Regression Model: Sampling Properies of he Leas Squares Esimaors by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Inference in he Simple

More information

Maintenance Models. Prof. Robert C. Leachman IEOR 130, Methods of Manufacturing Improvement Spring, 2011

Maintenance Models. Prof. Robert C. Leachman IEOR 130, Methods of Manufacturing Improvement Spring, 2011 Mainenance Models Prof Rober C Leachman IEOR 3, Mehods of Manufacuring Improvemen Spring, Inroducion The mainenance of complex equipmen ofen accouns for a large porion of he coss associaed wih ha equipmen

More information

Physics 127b: Statistical Mechanics. Fokker-Planck Equation. Time Evolution

Physics 127b: Statistical Mechanics. Fokker-Planck Equation. Time Evolution Physics 7b: Saisical Mechanics Fokker-Planck Equaion The Langevin equaion approach o he evoluion of he velociy disribuion for he Brownian paricle migh leave you uncomforable. A more formal reamen of his

More information

Ensamble methods: Bagging and Boosting

Ensamble methods: Bagging and Boosting Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par

More information

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Simulaion-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Week Descripion Reading Maerial 2 Compuer Simulaion of Dynamic Models Finie Difference, coninuous saes, discree ime Simple Mehods Euler Trapezoid

More information

GMM - Generalized Method of Moments

GMM - Generalized Method of Moments GMM - Generalized Mehod of Momens Conens GMM esimaion, shor inroducion 2 GMM inuiion: Maching momens 2 3 General overview of GMM esimaion. 3 3. Weighing marix...........................................

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Noes for EE7C Spring 018: Convex Opimizaion and Approximaion Insrucor: Moriz Hard Email: hard+ee7c@berkeley.edu Graduae Insrucor: Max Simchowiz Email: msimchow+ee7c@berkeley.edu Ocober 15, 018 3

More information

Nature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time.

Nature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time. Supplemenary Figure 1 Spike-coun auocorrelaions in ime. Normalized auocorrelaion marices are shown for each area in a daase. The marix shows he mean correlaion of he spike coun in each ime bin wih he spike

More information

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon 3..3 INRODUCION O DYNAMIC OPIMIZAION: DISCREE IME PROBLEMS A. he Hamilonian and Firs-Order Condiions in a Finie ime Horizon Define a new funcion, he Hamilonian funcion, H. H he change in he oal value of

More information

Air Traffic Forecast Empirical Research Based on the MCMC Method

Air Traffic Forecast Empirical Research Based on the MCMC Method Compuer and Informaion Science; Vol. 5, No. 5; 0 ISSN 93-8989 E-ISSN 93-8997 Published by Canadian Cener of Science and Educaion Air Traffic Forecas Empirical Research Based on he MCMC Mehod Jian-bo Wang,

More information

Handout: Power Laws and Preferential Attachment

Handout: Power Laws and Preferential Attachment CS224W: Analysis of Neworks Fall 2017 Handou: Power Laws and Preferenial Aachmen 1 Preferenial Aachmen Empirical sudies of real world neworks revealed ha degree disribuion ofen follows a heavyailed disribuion,

More information

Notes on Kalman Filtering

Notes on Kalman Filtering Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren

More information

An introduction to the theory of SDDP algorithm

An introduction to the theory of SDDP algorithm An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking

More information

A Dynamic Model of Economic Fluctuations

A Dynamic Model of Economic Fluctuations CHAPTER 15 A Dynamic Model of Economic Flucuaions Modified for ECON 2204 by Bob Murphy 2016 Worh Publishers, all righs reserved IN THIS CHAPTER, OU WILL LEARN: how o incorporae dynamics ino he AD-AS model

More information

18 Biological models with discrete time

18 Biological models with discrete time 8 Biological models wih discree ime The mos imporan applicaions, however, may be pedagogical. The elegan body of mahemaical heory peraining o linear sysems (Fourier analysis, orhogonal funcions, and so

More information

The Arcsine Distribution

The Arcsine Distribution The Arcsine Disribuion Chris H. Rycrof Ocober 6, 006 A common heme of he class has been ha he saisics of single walker are ofen very differen from hose of an ensemble of walkers. On he firs homework, we

More information

Hamilton- J acobi Equation: Explicit Formulas In this lecture we try to apply the method of characteristics to the Hamilton-Jacobi equation: u t

Hamilton- J acobi Equation: Explicit Formulas In this lecture we try to apply the method of characteristics to the Hamilton-Jacobi equation: u t M ah 5 2 7 Fall 2 0 0 9 L ecure 1 0 O c. 7, 2 0 0 9 Hamilon- J acobi Equaion: Explici Formulas In his lecure we ry o apply he mehod of characerisics o he Hamilon-Jacobi equaion: u + H D u, x = 0 in R n

More information

A Shooting Method for A Node Generation Algorithm

A Shooting Method for A Node Generation Algorithm A Shooing Mehod for A Node Generaion Algorihm Hiroaki Nishikawa W.M.Keck Foundaion Laboraory for Compuaional Fluid Dynamics Deparmen of Aerospace Engineering, Universiy of Michigan, Ann Arbor, Michigan

More information

Comparing Means: t-tests for One Sample & Two Related Samples

Comparing Means: t-tests for One Sample & Two Related Samples Comparing Means: -Tess for One Sample & Two Relaed Samples Using he z-tes: Assumpions -Tess for One Sample & Two Relaed Samples The z-es (of a sample mean agains a populaion mean) is based on he assumpion

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

Presentation Overview

Presentation Overview Acion Refinemen in Reinforcemen Learning by Probabiliy Smoohing By Thomas G. Dieerich & Didac Busques Speaer: Kai Xu Presenaion Overview Bacground The Probabiliy Smoohing Mehod Experimenal Sudy of Acion

More information

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Kriging Models Predicing Arazine Concenraions in Surface Waer Draining Agriculural Waersheds Paul L. Mosquin, Jeremy Aldworh, Wenlin Chen Supplemenal Maerial Number

More information

) were both constant and we brought them from under the integral.

) were both constant and we brought them from under the integral. YIELD-PER-RECRUIT (coninued The yield-per-recrui model applies o a cohor, bu we saw in he Age Disribuions lecure ha he properies of a cohor do no apply in general o a collecion of cohors, which is wha

More information

Matlab and Python programming: how to get started

Matlab and Python programming: how to get started Malab and Pyhon programming: how o ge sared Equipping readers he skills o wrie programs o explore complex sysems and discover ineresing paerns from big daa is one of he main goals of his book. In his chaper,

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Linear Response Theory: The connecion beween QFT and experimens 3.1. Basic conceps and ideas Q: How do we measure he conduciviy of a meal? A: we firs inroduce a weak elecric field E, and

More information

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still. Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in

More information

Guest Lectures for Dr. MacFarlane s EE3350 Part Deux

Guest Lectures for Dr. MacFarlane s EE3350 Part Deux Gues Lecures for Dr. MacFarlane s EE3350 Par Deux Michael Plane Mon., 08-30-2010 Wrie name in corner. Poin ou his is a review, so I will go faser. Remind hem o go lisen o online lecure abou geing an A

More information

On Boundedness of Q-Learning Iterates for Stochastic Shortest Path Problems

On Boundedness of Q-Learning Iterates for Stochastic Shortest Path Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. 38, No. 2, May 2013, pp. 209 227 ISSN 0364-765X (prin) ISSN 1526-5471 (online) hp://dx.doi.org/10.1287/moor.1120.0562 2013 INFORMS On Boundedness of Q-Learning Ieraes

More information

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t...

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t... Mah 228- Fri Mar 24 5.6 Marix exponenials and linear sysems: The analogy beween firs order sysems of linear differenial equaions (Chaper 5) and scalar linear differenial equaions (Chaper ) is much sronger

More information

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t Exercise 7 C P = α + β R P + u C = αp + βr + v (a) (b) C R = α P R + β + w (c) Assumpions abou he disurbances u, v, w : Classical assumions on he disurbance of one of he equaions, eg. on (b): E(v v s P,

More information

( ) is the stretch factor, and x the

( ) is the stretch factor, and x the (Lecures 7-8) Liddle, Chaper 5 Simple cosmological models (i) Hubble s Law revisied Self-similar srech of he universe All universe models have his characerisic v r ; v = Hr since only his conserves homogeneiy

More information

Distribution of Estimates

Distribution of Estimates Disribuion of Esimaes From Economerics (40) Linear Regression Model Assume (y,x ) is iid and E(x e )0 Esimaion Consisency y α + βx + he esimaes approach he rue values as he sample size increases Esimaion

More information

Assignment 6. Tyler Shendruk December 6, 2010

Assignment 6. Tyler Shendruk December 6, 2010 Assignmen 6 Tyler Shendruk December 6, 1 1 Harden Problem 1 Le K be he coupling and h he exernal field in a 1D Ising model. From he lecures hese can be ransformed ino effecive coupling and fields K and

More information

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé Bias in Condiional and Uncondiional Fixed Effecs Logi Esimaion: a Correcion * Tom Coupé Economics Educaion and Research Consorium, Naional Universiy of Kyiv Mohyla Academy Address: Vul Voloska 10, 04070

More information

MANAGEMENT SCIENCE doi /mnsc ec pp. ec1 ec20

MANAGEMENT SCIENCE doi /mnsc ec pp. ec1 ec20 MANAGEMENT SCIENCE doi.287/mnsc.7.82ec pp. ec ec2 e-companion ONLY AVAILABLE IN ELECTRONIC FORM informs 28 INFORMS Elecronic Companion Saffing of Time-Varying Queues o Achieve Time-Sable Performance by

More information

Ensamble methods: Boosting

Ensamble methods: Boosting Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room

More information

More Digital Logic. t p output. Low-to-high and high-to-low transitions could have different t p. V in (t)

More Digital Logic. t p output. Low-to-high and high-to-low transitions could have different t p. V in (t) EECS 4 Spring 23 Lecure 2 EECS 4 Spring 23 Lecure 2 More igial Logic Gae delay and signal propagaion Clocked circui elemens (flip-flop) Wriing a word o memory Simplifying digial circuis: Karnaugh maps

More information

5. Stochastic processes (1)

5. Stochastic processes (1) Lec05.pp S-38.45 - Inroducion o Teleraffic Theory Spring 2005 Conens Basic conceps Poisson process 2 Sochasic processes () Consider some quaniy in a eleraffic (or any) sysem I ypically evolves in ime randomly

More information

On Measuring Pro-Poor Growth. 1. On Various Ways of Measuring Pro-Poor Growth: A Short Review of the Literature

On Measuring Pro-Poor Growth. 1. On Various Ways of Measuring Pro-Poor Growth: A Short Review of the Literature On Measuring Pro-Poor Growh 1. On Various Ways of Measuring Pro-Poor Growh: A Shor eview of he Lieraure During he pas en years or so here have been various suggesions concerning he way one should check

More information

OBJECTIVES OF TIME SERIES ANALYSIS

OBJECTIVES OF TIME SERIES ANALYSIS OBJECTIVES OF TIME SERIES ANALYSIS Undersanding he dynamic or imedependen srucure of he observaions of a single series (univariae analysis) Forecasing of fuure observaions Asceraining he leading, lagging

More information

Planning in POMDPs. Dominik Schoenberger Abstract

Planning in POMDPs. Dominik Schoenberger Abstract Planning in POMDPs Dominik Schoenberger d.schoenberger@sud.u-darmsad.de Absrac This documen briefly explains wha a Parially Observable Markov Decision Process is. Furhermore i inroduces he differen approaches

More information

Chapter 2. First Order Scalar Equations

Chapter 2. First Order Scalar Equations Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.

More information

Lab 10: RC, RL, and RLC Circuits

Lab 10: RC, RL, and RLC Circuits Lab 10: RC, RL, and RLC Circuis In his experimen, we will invesigae he behavior of circuis conaining combinaions of resisors, capaciors, and inducors. We will sudy he way volages and currens change in

More information

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED 0.1 MAXIMUM LIKELIHOOD ESTIMATIO EXPLAIED Maximum likelihood esimaion is a bes-fi saisical mehod for he esimaion of he values of he parameers of a sysem, based on a se of observaions of a random variable

More information

Inventory Control of Perishable Items in a Two-Echelon Supply Chain

Inventory Control of Perishable Items in a Two-Echelon Supply Chain Journal of Indusrial Engineering, Universiy of ehran, Special Issue,, PP. 69-77 69 Invenory Conrol of Perishable Iems in a wo-echelon Supply Chain Fariborz Jolai *, Elmira Gheisariha and Farnaz Nojavan

More information

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Article from. Predictive Analytics and Futurism. July 2016 Issue 13 Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning

More information

Biol. 356 Lab 8. Mortality, Recruitment, and Migration Rates

Biol. 356 Lab 8. Mortality, Recruitment, and Migration Rates Biol. 356 Lab 8. Moraliy, Recruimen, and Migraion Raes (modified from Cox, 00, General Ecology Lab Manual, McGraw Hill) Las week we esimaed populaion size hrough several mehods. One assumpion of all hese

More information

A First Course on Kinetics and Reaction Engineering. Class 19 on Unit 18

A First Course on Kinetics and Reaction Engineering. Class 19 on Unit 18 A Firs ourse on Kineics and Reacion Engineering lass 19 on Uni 18 Par I - hemical Reacions Par II - hemical Reacion Kineics Where We re Going Par III - hemical Reacion Engineering A. Ideal Reacors B. Perfecly

More information

Lecture 2 October ε-approximation of 2-player zero-sum games

Lecture 2 October ε-approximation of 2-player zero-sum games Opimizaion II Winer 009/10 Lecurer: Khaled Elbassioni Lecure Ocober 19 1 ε-approximaion of -player zero-sum games In his lecure we give a randomized ficiious play algorihm for obaining an approximae soluion

More information

Numerical Dispersion

Numerical Dispersion eview of Linear Numerical Sabiliy Numerical Dispersion n he previous lecure, we considered he linear numerical sabiliy of boh advecion and diffusion erms when approimaed wih several spaial and emporal

More information

Policy regimes Theory

Policy regimes Theory Advanced Moneary Theory and Policy EPOS 2012/13 Policy regimes Theory Giovanni Di Barolomeo giovanni.dibarolomeo@uniroma1.i The moneary policy regime The simple model: x = - s (i - p e ) + x e + e D p

More information

EXERCISES FOR SECTION 1.5

EXERCISES FOR SECTION 1.5 1.5 Exisence and Uniqueness of Soluions 43 20. 1 v c 21. 1 v c 1 2 4 6 8 10 1 2 2 4 6 8 10 Graph of approximae soluion obained using Euler s mehod wih = 0.1. Graph of approximae soluion obained using Euler

More information

STATE-SPACE MODELLING. A mass balance across the tank gives:

STATE-SPACE MODELLING. A mass balance across the tank gives: B. Lennox and N.F. Thornhill, 9, Sae Space Modelling, IChemE Process Managemen and Conrol Subjec Group Newsleer STE-SPACE MODELLING Inroducion: Over he pas decade or so here has been an ever increasing

More information

Supplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence

Supplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence Supplemen for Sochasic Convex Opimizaion: Faser Local Growh Implies Faser Global Convergence Yi Xu Qihang Lin ianbao Yang Proof of heorem heorem Suppose Assumpion holds and F (w) obeys he LGC (6) Given

More information

Lecture 2 April 04, 2018

Lecture 2 April 04, 2018 Sas 300C: Theory of Saisics Spring 208 Lecure 2 April 04, 208 Prof. Emmanuel Candes Scribe: Paulo Orensein; edied by Sephen Baes, XY Han Ouline Agenda: Global esing. Needle in a Haysack Problem 2. Threshold

More information

True Online Temporal-Difference Learning. A. Rupam Mahmood Patrick M. Pilarski

True Online Temporal-Difference Learning. A. Rupam Mahmood Patrick M. Pilarski True Online Temporal-Difference Learning True Online Temporal-Difference Learning Harm van Seijen harm.vanseijen@ualbera.ca A. Rupam Mahmood ashique@ualbera.ca Parick M. Pilarski parick.pilarski@ualbera.ca

More information

Estimation of Poses with Particle Filters

Estimation of Poses with Particle Filters Esimaion of Poses wih Paricle Filers Dr.-Ing. Bernd Ludwig Chair for Arificial Inelligence Deparmen of Compuer Science Friedrich-Alexander-Universiä Erlangen-Nürnberg 12/05/2008 Dr.-Ing. Bernd Ludwig (FAU

More information

Tom Heskes and Onno Zoeter. Presented by Mark Buller

Tom Heskes and Onno Zoeter. Presented by Mark Buller Tom Heskes and Onno Zoeer Presened by Mark Buller Dynamic Bayesian Neworks Direced graphical models of sochasic processes Represen hidden and observed variables wih differen dependencies Generalize Hidden

More information

EXPLICIT TIME INTEGRATORS FOR NONLINEAR DYNAMICS DERIVED FROM THE MIDPOINT RULE

EXPLICIT TIME INTEGRATORS FOR NONLINEAR DYNAMICS DERIVED FROM THE MIDPOINT RULE Version April 30, 2004.Submied o CTU Repors. EXPLICIT TIME INTEGRATORS FOR NONLINEAR DYNAMICS DERIVED FROM THE MIDPOINT RULE Per Krysl Universiy of California, San Diego La Jolla, California 92093-0085,

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

Unit Root Time Series. Univariate random walk

Unit Root Time Series. Univariate random walk Uni Roo ime Series Univariae random walk Consider he regression y y where ~ iid N 0, he leas squares esimae of is: ˆ yy y y yy Now wha if = If y y hen le y 0 =0 so ha y j j If ~ iid N 0, hen y ~ N 0, he

More information

Intermediate Macro In-Class Problems

Intermediate Macro In-Class Problems Inermediae Macro In-Class Problems Exploring Romer Model June 14, 016 Today we will explore he mechanisms of he simply Romer model by exploring how economies described by his model would reac o exogenous

More information

Navneet Saini, Mayank Goyal, Vishal Bansal (2013); Term Project AML310; Indian Institute of Technology Delhi

Navneet Saini, Mayank Goyal, Vishal Bansal (2013); Term Project AML310; Indian Institute of Technology Delhi Creep in Viscoelasic Subsances Numerical mehods o calculae he coefficiens of he Prony equaion using creep es daa and Herediary Inegrals Mehod Navnee Saini, Mayank Goyal, Vishal Bansal (23); Term Projec

More information

ACE 562 Fall Lecture 8: The Simple Linear Regression Model: R 2, Reporting the Results and Prediction. by Professor Scott H.

ACE 562 Fall Lecture 8: The Simple Linear Regression Model: R 2, Reporting the Results and Prediction. by Professor Scott H. ACE 56 Fall 5 Lecure 8: The Simple Linear Regression Model: R, Reporing he Resuls and Predicion by Professor Sco H. Irwin Required Readings: Griffihs, Hill and Judge. "Explaining Variaion in he Dependen

More information

Introduction D P. r = constant discount rate, g = Gordon Model (1962): constant dividend growth rate.

Introduction D P. r = constant discount rate, g = Gordon Model (1962): constant dividend growth rate. Inroducion Gordon Model (1962): D P = r g r = consan discoun rae, g = consan dividend growh rae. If raional expecaions of fuure discoun raes and dividend growh vary over ime, so should he D/P raio. Since

More information

Online Convex Optimization Example And Follow-The-Leader

Online Convex Optimization Example And Follow-The-Leader CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion

More information

10. State Space Methods

10. State Space Methods . Sae Space Mehods. Inroducion Sae space modelling was briefly inroduced in chaper. Here more coverage is provided of sae space mehods before some of heir uses in conrol sysem design are covered in he

More information

DEPARTMENT OF STATISTICS

DEPARTMENT OF STATISTICS A Tes for Mulivariae ARCH Effecs R. Sco Hacker and Abdulnasser Haemi-J 004: DEPARTMENT OF STATISTICS S-0 07 LUND SWEDEN A Tes for Mulivariae ARCH Effecs R. Sco Hacker Jönköping Inernaional Business School

More information

A Reinforcement Learning Approach for Collaborative Filtering

A Reinforcement Learning Approach for Collaborative Filtering A Reinforcemen Learning Approach for Collaboraive Filering Jungkyu Lee, Byonghwa Oh 2, Jihoon Yang 2, and Sungyong Park 2 Cyram Inc, Seoul, Korea jklee@cyram.com 2 Sogang Universiy, Seoul, Korea {mrfive,yangjh,parksy}@sogang.ac.kr

More information

Time series Decomposition method

Time series Decomposition method Time series Decomposiion mehod A ime series is described using a mulifacor model such as = f (rend, cyclical, seasonal, error) = f (T, C, S, e) Long- Iner-mediaed Seasonal Irregular erm erm effec, effec,

More information

Math 2142 Exam 1 Review Problems. x 2 + f (0) 3! for the 3rd Taylor polynomial at x = 0. To calculate the various quantities:

Math 2142 Exam 1 Review Problems. x 2 + f (0) 3! for the 3rd Taylor polynomial at x = 0. To calculate the various quantities: Mah 4 Eam Review Problems Problem. Calculae he 3rd Taylor polynomial for arcsin a =. Soluion. Le f() = arcsin. For his problem, we use he formula f() + f () + f ()! + f () 3! for he 3rd Taylor polynomial

More information

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class

More information

Explaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015

Explaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015 Explaining Toal Facor Produciviy Ulrich Kohli Universiy of Geneva December 2015 Needed: A Theory of Toal Facor Produciviy Edward C. Presco (1998) 2 1. Inroducion Toal Facor Produciviy (TFP) has become

More information

Energy Storage Benchmark Problems

Energy Storage Benchmark Problems Energy Sorage Benchmark Problems Daniel F. Salas 1,3, Warren B. Powell 2,3 1 Deparmen of Chemical & Biological Engineering 2 Deparmen of Operaions Research & Financial Engineering 3 Princeon Laboraory

More information

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis Speaker Adapaion Techniques For Coninuous Speech Using Medium and Small Adapaion Daa Ses Consaninos Boulis Ouline of he Presenaion Inroducion o he speaker adapaion problem Maximum Likelihood Sochasic Transformaions

More information

3.1 More on model selection

3.1 More on model selection 3. More on Model selecion 3. Comparing models AIC, BIC, Adjused R squared. 3. Over Fiing problem. 3.3 Sample spliing. 3. More on model selecion crieria Ofen afer model fiing you are lef wih a handful of

More information

State-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter

State-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter Sae-Space Models Iniializaion, Esimaion and Smoohing of he Kalman Filer Iniializaion of he Kalman Filer The Kalman filer shows how o updae pas predicors and he corresponding predicion error variances when

More information