Reinforcement Learning

Size: px
Start display at page:

Download "Reinforcement Learning"

Transcription

1 Noe o oher eachers and users of hese sldes. Andrew would be delghed f you found hs source maeral useful n gvng your own lecures. Feel free o use hese sldes verbam, or o modfy hem o f your own needs. PowerPon orgnals are avalable. If you make use of a sgnfcan poron of hese sldes n your own lecure, please nclude hs message, or he followng lnk o he source reposory of Andrew s uorals: hp:// Commens and correcons graefully receved. Renforcemen Learnng Andrew W. Moore Assocae Professor School of Compuer Scence Carnege Mellon Unversy awm@cs.cmu.edu Copyrgh 2002, Andrew W. Moore Aprl 23rd, 2002 Predcng Delayed Rewards IN A 0.4 DISCOUNTED MARKOV SYSTEM R=-5 S R=0 S R=0 0.5 S R=3 R=0 S 0.2 S R= S 6 Prob(nex sae = S 5 hs sae = S 4 ) = 0.8 ec Wha s expeced sum of fuure rewards (dscouned)? Ε = 0 γ ( [ ] ) S[ 0] = S R S us Solve I! We use sandard Markov Sysem Theory Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 2

2 Learnng Delayed Rewards S? R=???? S 4? R=??? S 2? R=??? S 5? R=??? S 3? R=??? S 6? R=??? All you can see s a seres of saes and rewards: S (R=0) S 2 (R=0) S 3 (R=4) S 2 (R=0) S 4 (R=0) S 5 (R=0) Task: Based on hs sequence, mae *(S ),*(S 2 ) *(S 6 ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 3 Idea : Supervsed Learnng S (R=0) S 2 (R=0) Assume?=/2 S 3 (R=4) S 2 (R=0) S 4 (R=0) S 5 (R=0) A = we were n sae S and evenually go a long erm dscouned reward of 0+?0+? 2 4+? 3 0+? 4 0 = A =2 n sae S 2 ldr = 2 A =5 n sae S 4 ldr = 0 A =3 n sae S 3 ldr = 4 A =6 n sae S 5 ldr = 0 A =4 n sae S 2 ldr = 0 Sae Observaons of LTDR Mean LTDR S = (S ) S 2 2, 0 = (S 2 ) S = (S 3 ) S = (S 4 ) S = (S 5 ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 4 2

3 Supervsed Learnng ALG Wach a raecory S[0] r[0] S[] r[] S[T]r[T] For =0,, T, compue [ ] Compue ( S ) = =0 mean value of = among n sae S Le MATCHES ( ) γ r[ + ] [ ] all ranso ns begnnng on he raecor y ( S ) = { S[ ] [ ] MATCHES S You re done! S = MATCHES( S ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 5 ( ) = S }, hen defne Supervsed Learnng ALG for he md If you have an anxous personaly you may be worred abou edge effecs for some of he fnal ransons. Wh large raecores hese are neglgble. Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 6 3

4 Onlne Supervsed Learnng Inalze: Coun[S ] = 0 S Sum[S ] = 0 S Elgbly[S ] = 0 S Observe: When we experence S wh reward r do hs: Elg[S ]?Elg[S ] Elg[S ] Elg[S ] + Sum[S ] Sum[S ]+rxelg[s ] Coun[S ] Coun[S ] + Then a any me, (S )= Sum[S ]/Coun[S ] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 7 Onlne Supervsed Learnng Economcs Gven N saes S S N, OSL needs O(N) memory. Each updae needs O(N) work snce we mus updae all Elg[ ] array elemens Idea: Be sparse and only updae/process Elg[ ] elemens wh values >?for ny? There are only log such elemens ξ log γ Easy o prove: As T, ( S ) ( S) S Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 8 4

5 Onlne Supervsed Learnng Le s grab OSL off he sree, bundle no a black van, ake o a bunker and nerrogae under 600 Wa lghs. S (r=0) S 2 (r=0) Sae Observaons of LTDR ^ (S ) S S 2 2, 0 S S S COMPLAINT S 3 (r=4) S 2 (r=0) S 4 (r=0) S 5 (r=0) There s somehng a lle suspcous abou hs (effcency-wse) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 9 Cerany-Equvalen (CE) Learnng S (r=0) S 2 (r=0) Idea: Use your daa o mae he underlyng Markov sysem, nsead of ryng o mae drecly. Esmaed Markov Sysem: S 3 (r=4) S 2 (r=0) S 4 (r=0) S 5 (r=0) You draw n he ransons + probs S r = 0 S 2 r = 0 S 3 r = 4 S 4 r = 0 S 5 r = 0 Wha re he maed values? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 0 5

6 C.E. Mehod for Markov Sysems Inalze: Coun[S ] = 0 S #Tmes vsed S SumR[S ] = 0 Sum of rewards from S S Trans[S,S ] = 0 #Tmes ransoned from S S When we are n sae S, and we receve reward r, and we move o S Coun[S ] Coun[S ] + SumR[S ] SumR[S ] + r Trans[S,S ] Trans[S,S ] + Then a any me r (S ) = SumR[S ] / Coun[S ] P = Esmaed Prob(nex = S hs = S ) = Trans[S,S ] / Coun[S ] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde C.E. for Markov Sysems (connued) So a any me we have r (S ) and P (nex=s hs=s ) S S = P So a any me we can solve he se of lnear equaons ( S ) = r ( S ) + γ P ( S S ) ( S ) [In vecor noaon, = r +?P => = (I-?P ) - r where r are vecors of lengh N P s an NxN marx N = # saes ] S Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 2 6

7 C.E. Onlne Economcs Memory: O(N 2 ) Tme o updae couners: O() Tme o re-evaluae O(N 3 ) f use marx nverson O(N 2 k CRIT ) f use value eraon and we need k CRIT eraons o converge O(Nk CRIT ) f use value eraon, and k CRIT o converge, and M.S. s Sparse (.e. mean # successors s consan) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 3 Cerany Equvalen Learnng Memory use could be O(N 2 )! And me per updae could be O(Nk CRIT ) up o O(N 3 )! Too expensve for some people. COMPLAINT Prorzed sweepng wll help, (see laer), bu frs le s revew a very nexpensve approach Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 4 7

8 Why hs obsesson wh onlneness? I really care abou supplyng up-o-dae maes all he me. Can you guess why? If no, all wll be revealed n good me Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 5 Less Tme: More Daa Lmed Backups Do prevous C.E. algorhm. A each me mep we observe S (r) S and updae Coun[S ], SumR[S ], Trans[S,S ] And hus also updae maes r and P oucomes ( S ) Bu nsead of re-solvng for, do much less work. us do one backup of S [ ] [ S ] r + γ P [ S ] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 6 8

9 One Backup C.E. Economcs Space : O(N 2 ) NO IMPROVEMENT THERE! Tme o updae sascs : O() Tme o updae : O() Good News: Much cheaper per ranson Good News: Conracon Mappng proof (modfed) promses convergence o opmal Bad News: Wases daa Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 7 Prorzed Sweepng [Moore + Akeson, 93] Tres o be almos as daa-effcen as full CE bu no much more expensve han One Backup CE. On every ranson, some number (ß) of saes may have a backup appled. Whch ones? The mos deservng We keep a prory queue of whch saes have he bgg poenal for changng her (S) value Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 8 9

10 Where Are We? Tryng o do onlne predcon from sreams of ransons Space Updae Cos Daa Effcency: Supervsed Learnng Full C.E. Learnng One Backup C.E. Learnng Prorzed Sweepng 0(N s ) 0(N so ) 0(N so ) 0(N so ) 0( ) 0(N so N s ) 0(N so k CRIT ) 0() 0() log(/?) N so = # sae-oucomes (number of arrows on he M.S. dagram) N s = # saes Wha Nex? Sample Backups!!! Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 9 Temporal Dfference Learnng Only manan a array nohng else [Suon 988] So you ve go (S ) (S 2 ), (S N ) and you observe S r S wha should you do? Can You Guess? A ranson from ha receves an mmedae reward of r and umps o Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 20 0

11 S r S TD Learnng ( S ) We updae = We nudge o be closer o expeced fuure rewards α = ( S ) ( α) ( S ) + Expeced fuure α[ ] rewards ( α) ( S ) + α[ r + γ ( S )] WEIGHTED SUM s called a learnng rae parameer. (See? n he neural lecure) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 2 Smplfed TD Analyss P =? r =? TERMINATE S 0 P 2 =? r 2 =? TERMINATE r=0 P M =? : r M =? TERMINATE Suppose you always begn n S 0 You hen ranson a random o one of M places. You don know he ranson probs. You hen ge a place-dependen reward (unknown n advance). Then he ral ermnaes. Defne *(S 0 )= Expeced reward Le s mae wh TD Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 22

12 S 0 r=0 p () p (2) p (N) r () r (2) r (N) r (k) = reward of k h ermnal sae p (k) = prob of k h ermnal sae We ll do a seres of rals. Reward on h = Ε r n = p k= k r k ral s r [ ] ( ) ( ) [ ] [ Noe Ε r s ndependen of ] Defne *(S 0 ) = * = E[r ] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 23 Le s run TD-Learnng, where = Esmae (S 0 ) before he h ral. From defnon of TD-Learnng: + = (-a) + ar Useful quany: Defne σ 2 = Varance of reward= Ε M = P k= ( r ) 2 ( k ) ( k ) [ ] 2 ( r ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 24 2

13 Ε = Remember * = E[r ], s 2 = E[(r -*) 2 ] + = ar + (-a) [ + ] = [ αr + ( α ) ] = Ε ( α ) Ε[ ] Thus... lm Ε [ ] = Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 25 W H Y? Is hs mpressve?? Remember * = E[r ], s 2 = E[(r -*) 2 ] + = ar + (-a) Wre S = Expeced squared error beween and * before he h eraon S + = E[( + -*) 2 ] = E[(ar +(-a) - *) 2 ] = E[(a[r -*]+(-a)[ - *]) 2 ] = E[a 2 (r -*) 2 +a(-a)(r -*)( - *)+(-a) 2 ( - *) 2 ] = a 2 E[(r -*) 2 ]+a(-a)e[(r -*)( - *)]+(-a) 2 E[( - *) 2 ] = = a 2 s 2 +(-a) 2 S WHY? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 26 3

14 And s hus easy o show ha. lm S = lm Ε [( ) ] 2 2 ασ = (2 α) Wha do you hnk of TD learnng? How would you mprove? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 27 Decayng Learnng Rae [Dayan 99sh] showed ha for General TD learnng of a Markow Sysem (no us our smple model) ha f you use updae rule ( S ) α [ r + γ ( S )] + ( α ) ( S ) hen, as number of observaons goes o nfny S S PROVIDED All saes vsed 8 ly ofen α = = ( ) ( ) 2 T 2 α < k. T. α < k = = Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 28 k. T. Ths means T = Ths means α > k 4

15 Decayng Learnng Rae Ths Works: a = / Ths Doesn : a = a 0 Ths Works: a = ß/(ß+) [e.g. ß=000] Ths Doesn : a = ßa - (ß<) IN OUR EXAMPLE.USE a = / 2 2 Remember = Ε[ r ], σ = Ε[ (r ) ] + Wre = α r C C + ( α ) = r + ( ) = + ( ) = r + C and so you' ll + = r + = see ha 0 And Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 29 Decayng Learnng Rae con σ + [ ] ( 0 - ) = 2 lm Ε[ ( - ) ] = 0 Ε ( - ) so, ulmaely Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 30 5

16 A Fancer TD Wre S[] = sae a me Suppose a = /4? = /2 Assume (S 23 )=0 (S 7 )=0 (S 44 )=6 Assume = 405 and S[] = S 23 Observe (r=0) S 23 S 7 wh reward 0 Now = 406, S[] = S 7, S[-] = S 23 (S 23 )=, (S 7 )=, (S 44 )= (r=0) Observe S 7 S 44 Now = 407, S[] = S44 (S 23 )=, (S 7 )=, (S 44 )= INSIGHT: (S 23 ) mgh hnk I goa ge me some of ha!!! Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 3 TD(?) Commens TD(?=0) s he orgnal TD TD(?=) s almos he same as supervsed learnng (excep uses a learnng rae nsead of explc couns) TD(?=0.7) s ofen emprcally he b performer Dayan s proof holds for all 0=?= Updaes can be made more compuaonally effcen wh elgbly races (smlar o O.S.L.) Quon: Can you nven a problem ha would make TD(0) look bad and TD() look good? How abou TD(0) look good & TD() bad?? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 32 6

17 Learnng M.S. Summary MODEL-BASED Supervsed Learnng Full C.E. Learnng One Backup C.E. Learnng Prorzed Sweepng Space 0(N s ) 0(N so ) 0(N so ) 0(N so ) Updae Cos 0 log γ 0(N so N s ) 0(N so k CRIT ) 0() 0() Daa Effcency MODEL FREE TD(0) TD(?), 0<?= 0(N s ) 0(N s ) 0() 0 log γλ Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 33 Learnng Polces for MDPs See prevous lecure sldes for defnon of and compuaon wh MDPs. The Hear of REINFORCEMENT Learnng sae Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 34 7

18 The ask: World: You are n sae 34. Your mmedae reward s 3. You have 3 acons. Robo: I ll ake acon 2. World: You are n sae 77. Your mmedae reward s -7. You have 2 acons. Robo: I ll ake acon. World: You re n sae 34 (agan). Your mmedae reward s 3. You have 3 acons. The Markov propery means once you ve seleced an acon he P.D.F. of your nex sae s he same as he las me you red he acon n hs sae. Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 35 The Cred Assgnmen Problem I m n sae 43, 39, 22, 2, 2, 3, 54, 26, reward = 0, = 0, = 0, = 0, = 0, = 0, = 0, = 00, acon = 2 = 4 = = = = 2 = 2 Yppee! I go o a sae wh a bg reward! Bu whch of my acons along he way acually helped me ge here?? Ths s he Cred Assgnmen problem. I makes Supervsed Learnng approaches (e.g. Boxes [Mche & Chambers]) very, very slow. Usng he MDP assumpon helps avod hs problem. Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 36 8

19 Full C.E. Learnng One Backup C.E. Learnng Prorzed Sweepng MDP Polcy Learnng 0(N sao ) 0(N sao ) 0(N sao ) Space Updae Cos 0(N sao k CRIT ) 0(N?0 ) 0(ßN?0 ) We ll hnk abou Model-Free n a momen Daa Effcency The C.E. mehods are very smlar o he MS case, excep now do value-eraon-for-mdp backups = a S ( S ) max r + γ P ( S S, a) ( S ) SUCCS ( ) S Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 37 Choosng Acons We re n sae S We can mae r P (nex = S hs = S, acon a) (nex = S ) So wha acon should we choose? IDEA : IDEA 2 : a = arg max r + γ P a a = random Any problems wh hese deas? Any oher suggons? Could we be opmal? ( S S, a ) ( S ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 38 9

20 Model-Free R.L. Why no use T.D.? Observe updae r S a S ( S ) α ( r + γ ( S )) + ( α ) ( S ) Wha s wrong wh hs? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 39 Q-Learnng: Model-Free R.L. [Wakns, 988] Defne Q*(S,a)= Expeced sum of dscouned fuure rewards f I sar n sae S, f I hen ake acon a, and f I m subsequenly opmal Quons: Defne Q*(S,a) n erms of * Defne *(S ) n erms of Q* Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 40 20

21 Q Noe ha Q Q-Learnng Updae ( S, a) = r + ( ) Q ( a γ P S S, α max S, ) S SUCCS ( S ) In Q-learnng we manan a able of Q values nsead of values When you see S ( S, a) α r γ max Q ( S, a ) + ( α ) Q ( S, a) + reward acon a a S do Ths s even cleverer han looks: he Q values are no based by any parcular exploraon polcy. I avods he Cred Assgnmen problem. a Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 4 Q-Learnng: Choosng Acons Same ssues as for CE choosng acons Don always be greedy, so don always choose: arg max Q a Don always be random (oherwse wll ake a long me o reach somewhere excng) Bolzmann exploraon [Wakns] Q Prob(choose acon a) exp K ( s, a) Opmsm n he face of uncerany [Suon 90, Kaelblng 90] Inalze Q-values opmscally hgh o encourage exploraon Or ake no accoun how ofen each s,a par has been red ( ) s, a Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 42 2

22 Q-Learnng Commens [Wakns] proved ha Q-learnng wll evenually converge o an opmal polcy. Emprcally s cue Emprcally s very slow Why no do Q(?)? Would no make much sense [renroduce he cred assgnmen problem] Some people (e.g. Peng & Wllams) have red o work her way around hs. Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 43 If we had me Value funcon approxmaon Use a Neural Ne o represen [e.g. Tesauro] Use a Neural Ne o represen Q [e.g. Cres] Use a decson ree wh Q-learnng [Chapman + Kaelblng 9] wh C.E. learnng [Moore 9] How o spl up space? Sgnfcance on Q values [Chapman + Kaelblng] Execuon accuracy monorng [Moore 9] Game Theory [Moore + Akeson 95] New nfluence/varance crera [Munos 99] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 44 22

23 R.L. Theory If we had me Counerexamples [Boyan + Moore], [Bard] Value Funcon Approxmaors wh Averagng wll converge o somehng [Gordon] Neural Nes can fal [Bard] Neural Nes wh Resdual Graden updaes wll converge o somehng Lnear approxmaors for TD learnng wll converge o somehng useful [Tsskls + Van Roy] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 45 Wha You Should Know Supervsed learnng for predcng delayed rewards Cerany equvalen learnng for predcng delayed rewards Model free learnng (TD) for predcng delayed rewards Renforcemen Learnng wh MDPs: Wha s he ask? Why s hard o choose acons? Q-learnng (ncludng beng able o work hrough small smulaed examples of RL) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 46 23

Lecture 6: Learning for Control (Generalised Linear Regression)

Lecture 6: Learning for Control (Generalised Linear Regression) Lecure 6: Learnng for Conrol (Generalsed Lnear Regresson) Conens: Lnear Mehods for Regresson Leas Squares, Gauss Markov heorem Recursve Leas Squares Lecure 6: RLSC - Prof. Sehu Vjayakumar Lnear Regresson

More information

Hidden Markov Models Following a lecture by Andrew W. Moore Carnegie Mellon University

Hidden Markov Models Following a lecture by Andrew W. Moore Carnegie Mellon University Hdden Markov Models Followng a lecure by Andrew W. Moore Carnege Mellon Unversy www.cs.cmu.edu/~awm/uorals A Markov Sysem Has N saes, called s, s 2.. s N s 2 There are dscree meseps, 0,, s s 3 N 3 0 Hdden

More information

Lecture VI Regression

Lecture VI Regression Lecure VI Regresson (Lnear Mehods for Regresson) Conens: Lnear Mehods for Regresson Leas Squares, Gauss Markov heorem Recursve Leas Squares Lecure VI: MLSC - Dr. Sehu Vjayakumar Lnear Regresson Model M

More information

Clustering with Gaussian Mixtures

Clustering with Gaussian Mixtures Noe o oher eachers and users of hese sldes. Andrew would be delghed f you found hs source maeral useful n gvng your own lecures. Feel free o use hese sldes verbam, or o modfy hem o f your own needs. PowerPon

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 4

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 4 CS434a/54a: Paern Recognon Prof. Olga Veksler Lecure 4 Oulne Normal Random Varable Properes Dscrmnan funcons Why Normal Random Varables? Analycally racable Works well when observaon comes form a corruped

More information

John Geweke a and Gianni Amisano b a Departments of Economics and Statistics, University of Iowa, USA b European Central Bank, Frankfurt, Germany

John Geweke a and Gianni Amisano b a Departments of Economics and Statistics, University of Iowa, USA b European Central Bank, Frankfurt, Germany Herarchcal Markov Normal Mxure models wh Applcaons o Fnancal Asse Reurns Appendx: Proofs of Theorems and Condonal Poseror Dsrbuons John Geweke a and Gann Amsano b a Deparmens of Economcs and Sascs, Unversy

More information

Fall 2010 Graduate Course on Dynamic Learning

Fall 2010 Graduate Course on Dynamic Learning Fall 200 Graduae Course on Dynamc Learnng Chaper 4: Parcle Flers Sepember 27, 200 Byoung-Tak Zhang School of Compuer Scence and Engneerng & Cognve Scence and Bran Scence Programs Seoul aonal Unversy hp://b.snu.ac.kr/~bzhang/

More information

In the complete model, these slopes are ANALYSIS OF VARIANCE FOR THE COMPLETE TWO-WAY MODEL. (! i+1 -! i ) + [(!") i+1,q - [(!

In the complete model, these slopes are ANALYSIS OF VARIANCE FOR THE COMPLETE TWO-WAY MODEL. (! i+1 -! i ) + [(!) i+1,q - [(! ANALYSIS OF VARIANCE FOR THE COMPLETE TWO-WAY MODEL The frs hng o es n wo-way ANOVA: Is here neracon? "No neracon" means: The man effecs model would f. Ths n urn means: In he neracon plo (wh A on he horzonal

More information

Variants of Pegasos. December 11, 2009

Variants of Pegasos. December 11, 2009 Inroducon Varans of Pegasos SooWoong Ryu bshboy@sanford.edu December, 009 Youngsoo Cho yc344@sanford.edu Developng a new SVM algorhm s ongong research opc. Among many exng SVM algorhms, we wll focus on

More information

( ) () we define the interaction representation by the unitary transformation () = ()

( ) () we define the interaction representation by the unitary transformation () = () Hgher Order Perurbaon Theory Mchael Fowler 3/7/6 The neracon Represenaon Recall ha n he frs par of hs course sequence, we dscussed he chrödnger and Hesenberg represenaons of quanum mechancs here n he chrödnger

More information

Outline. Probabilistic Model Learning. Probabilistic Model Learning. Probabilistic Model for Time-series Data: Hidden Markov Model

Outline. Probabilistic Model Learning. Probabilistic Model Learning. Probabilistic Model for Time-series Data: Hidden Markov Model Probablsc Model for Tme-seres Daa: Hdden Markov Model Hrosh Mamsuka Bonformacs Cener Kyoo Unversy Oulne Three Problems for probablsc models n machne learnng. Compung lkelhood 2. Learnng 3. Parsng (predcon

More information

Solution in semi infinite diffusion couples (error function analysis)

Solution in semi infinite diffusion couples (error function analysis) Soluon n sem nfne dffuson couples (error funcon analyss) Le us consder now he sem nfne dffuson couple of wo blocks wh concenraon of and I means ha, n a A- bnary sysem, s bondng beween wo blocks made of

More information

Department of Economics University of Toronto

Department of Economics University of Toronto Deparmen of Economcs Unversy of Torono ECO408F M.A. Economercs Lecure Noes on Heeroskedascy Heeroskedascy o Ths lecure nvolves lookng a modfcaons we need o make o deal wh he regresson model when some of

More information

FTCS Solution to the Heat Equation

FTCS Solution to the Heat Equation FTCS Soluon o he Hea Equaon ME 448/548 Noes Gerald Reckenwald Porland Sae Unversy Deparmen of Mechancal Engneerng gerry@pdxedu ME 448/548: FTCS Soluon o he Hea Equaon Overvew Use he forward fne d erence

More information

Econ107 Applied Econometrics Topic 5: Specification: Choosing Independent Variables (Studenmund, Chapter 6)

Econ107 Applied Econometrics Topic 5: Specification: Choosing Independent Variables (Studenmund, Chapter 6) Econ7 Appled Economercs Topc 5: Specfcaon: Choosng Independen Varables (Sudenmund, Chaper 6 Specfcaon errors ha we wll deal wh: wrong ndependen varable; wrong funconal form. Ths lecure deals wh wrong ndependen

More information

Density Matrix Description of NMR BCMB/CHEM 8190

Density Matrix Description of NMR BCMB/CHEM 8190 Densy Marx Descrpon of NMR BCMBCHEM 89 Operaors n Marx Noaon Alernae approach o second order specra: ask abou x magnezaon nsead of energes and ranson probables. If we say wh one bass se, properes vary

More information

The Finite Element Method for the Analysis of Non-Linear and Dynamic Systems

The Finite Element Method for the Analysis of Non-Linear and Dynamic Systems Swss Federal Insue of Page 1 The Fne Elemen Mehod for he Analyss of Non-Lnear and Dynamc Sysems Prof. Dr. Mchael Havbro Faber Dr. Nebojsa Mojslovc Swss Federal Insue of ETH Zurch, Swzerland Mehod of Fne

More information

WiH Wei He

WiH Wei He Sysem Idenfcaon of onlnear Sae-Space Space Baery odels WH We He wehe@calce.umd.edu Advsor: Dr. Chaochao Chen Deparmen of echancal Engneerng Unversy of aryland, College Par 1 Unversy of aryland Bacground

More information

(,,, ) (,,, ). In addition, there are three other consumers, -2, -1, and 0. Consumer -2 has the utility function

(,,, ) (,,, ). In addition, there are three other consumers, -2, -1, and 0. Consumer -2 has the utility function MACROECONOMIC THEORY T J KEHOE ECON 87 SPRING 5 PROBLEM SET # Conder an overlappng generaon economy le ha n queon 5 on problem e n whch conumer lve for perod The uly funcon of he conumer born n perod,

More information

CHAPTER 10: LINEAR DISCRIMINATION

CHAPTER 10: LINEAR DISCRIMINATION CHAPER : LINEAR DISCRIMINAION Dscrmnan-based Classfcaon 3 In classfcaon h K classes (C,C,, C k ) We defned dscrmnan funcon g j (), j=,,,k hen gven an es eample, e chose (predced) s class label as C f g

More information

Lecture 11 SVM cont

Lecture 11 SVM cont Lecure SVM con. 0 008 Wha we have done so far We have esalshed ha we wan o fnd a lnear decson oundary whose margn s he larges We know how o measure he margn of a lnear decson oundary Tha s: he mnmum geomerc

More information

Let s treat the problem of the response of a system to an applied external force. Again,

Let s treat the problem of the response of a system to an applied external force. Again, Page 33 QUANTUM LNEAR RESPONSE FUNCTON Le s rea he problem of he response of a sysem o an appled exernal force. Agan, H() H f () A H + V () Exernal agen acng on nernal varable Hamlonan for equlbrum sysem

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecure 9 Hamlonan Equaons of Moon (Chaper 8) Wha We Dd Las Tme Consruced Hamlonan formalsm H ( q, p, ) = q p L( q, q, ) H p = q H q = p H = L Equvalen o Lagrangan formalsm Smpler, bu

More information

Appendix to Online Clustering with Experts

Appendix to Online Clustering with Experts A Appendx o Onlne Cluserng wh Expers Furher dscusson of expermens. Here we furher dscuss expermenal resuls repored n he paper. Ineresngly, we observe ha OCE (and n parcular Learn- ) racks he bes exper

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecure 9 Hamlonan Equaons of Moon (Chaper 8) Wha We Dd Las Tme Consruced Hamlonan formalsm Hqp (,,) = qp Lqq (,,) H p = q H q = p H L = Equvalen o Lagrangan formalsm Smpler, bu wce as

More information

Density Matrix Description of NMR BCMB/CHEM 8190

Density Matrix Description of NMR BCMB/CHEM 8190 Densy Marx Descrpon of NMR BCMBCHEM 89 Operaors n Marx Noaon If we say wh one bass se, properes vary only because of changes n he coeffcens weghng each bass se funcon x = h< Ix > - hs s how we calculae

More information

FI 3103 Quantum Physics

FI 3103 Quantum Physics /9/4 FI 33 Quanum Physcs Aleander A. Iskandar Physcs of Magnesm and Phooncs Research Grou Insu Teknolog Bandung Basc Conces n Quanum Physcs Probably and Eecaon Value Hesenberg Uncerany Prncle Wave Funcon

More information

[ ] 2. [ ]3 + (Δx i + Δx i 1 ) / 2. Δx i-1 Δx i Δx i+1. TPG4160 Reservoir Simulation 2018 Lecture note 3. page 1 of 5

[ ] 2. [ ]3 + (Δx i + Δx i 1 ) / 2. Δx i-1 Δx i Δx i+1. TPG4160 Reservoir Simulation 2018 Lecture note 3. page 1 of 5 TPG460 Reservor Smulaon 08 page of 5 DISCRETIZATIO OF THE FOW EQUATIOS As we already have seen, fne dfference appromaons of he paral dervaves appearng n he flow equaons may be obaned from Taylor seres

More information

THEORETICAL AUTOCORRELATIONS. ) if often denoted by γ. Note that

THEORETICAL AUTOCORRELATIONS. ) if often denoted by γ. Note that THEORETICAL AUTOCORRELATIONS Cov( y, y ) E( y E( y))( y E( y)) ρ = = Var( y) E( y E( y)) =,, L ρ = and Cov( y, y ) s ofen denoed by whle Var( y ) f ofen denoed by γ. Noe ha γ = γ and ρ = ρ and because

More information

Notes on the stability of dynamic systems and the use of Eigen Values.

Notes on the stability of dynamic systems and the use of Eigen Values. Noes on he sabl of dnamc ssems and he use of Egen Values. Source: Macro II course noes, Dr. Davd Bessler s Tme Seres course noes, zarads (999) Ineremporal Macroeconomcs chaper 4 & Techncal ppend, and Hamlon

More information

CS 268: Packet Scheduling

CS 268: Packet Scheduling Pace Schedulng Decde when and wha pace o send on oupu ln - Usually mplemened a oupu nerface CS 68: Pace Schedulng flow Ion Soca March 9, 004 Classfer flow flow n Buffer managemen Scheduler soca@cs.bereley.edu

More information

Discrete Markov Process. Introduction. Example: Balls and Urns. Stochastic Automaton. INTRODUCTION TO Machine Learning 3rd Edition

Discrete Markov Process. Introduction. Example: Balls and Urns. Stochastic Automaton. INTRODUCTION TO Machine Learning 3rd Edition EHEM ALPAYDI he MI Press, 04 Lecure Sldes for IRODUCIO O Machne Learnng 3rd Edon alpaydn@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/ml3e Sldes from exboo resource page. Slghly eded and wh addonal examples

More information

Robustness Experiments with Two Variance Components

Robustness Experiments with Two Variance Components Naonal Insue of Sandards and Technology (NIST) Informaon Technology Laboraory (ITL) Sascal Engneerng Dvson (SED) Robusness Expermens wh Two Varance Componens by Ana Ivelsse Avlés avles@ns.gov Conference

More information

Introduction ( Week 1-2) Course introduction A brief introduction to molecular biology A brief introduction to sequence comparison Part I: Algorithms

Introduction ( Week 1-2) Course introduction A brief introduction to molecular biology A brief introduction to sequence comparison Part I: Algorithms Course organzaon Inroducon Wee -2) Course nroducon A bref nroducon o molecular bology A bref nroducon o sequence comparson Par I: Algorhms for Sequence Analyss Wee 3-8) Chaper -3, Models and heores» Probably

More information

1 Constant Real Rate C 1

1 Constant Real Rate C 1 Consan Real Rae. Real Rae of Inees Suppose you ae equally happy wh uns of he consumpon good oday o 5 uns of he consumpon good n peod s me. C 5 Tha means you ll be pepaed o gve up uns oday n eun fo 5 uns

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecure 0 Canoncal Transformaons (Chaper 9) Wha We Dd Las Tme Hamlon s Prncple n he Hamlonan formalsm Dervaon was smple δi δ Addonal end-pon consrans pq H( q, p, ) d 0 δ q ( ) δq ( ) δ

More information

Lecture 2 M/G/1 queues. M/G/1-queue

Lecture 2 M/G/1 queues. M/G/1-queue Lecure M/G/ queues M/G/-queue Posson arrval process Arbrary servce me dsrbuon Sngle server To deermne he sae of he sysem a me, we mus now The number of cusomers n he sysems N() Tme ha he cusomer currenly

More information

Advanced time-series analysis (University of Lund, Economic History Department)

Advanced time-series analysis (University of Lund, Economic History Department) Advanced me-seres analss (Unvers of Lund, Economc Hsor Dearmen) 3 Jan-3 Februar and 6-3 March Lecure 4 Economerc echnues for saonar seres : Unvarae sochasc models wh Box- Jenns mehodolog, smle forecasng

More information

Hidden Markov Models

Hidden Markov Models 11-755 Machne Learnng for Sgnal Processng Hdden Markov Models Class 15. 12 Oc 2010 1 Admnsrva HW2 due Tuesday Is everyone on he projecs page? Where are your projec proposals? 2 Recap: Wha s an HMM Probablsc

More information

On One Analytic Method of. Constructing Program Controls

On One Analytic Method of. Constructing Program Controls Appled Mahemacal Scences, Vol. 9, 05, no. 8, 409-407 HIKARI Ld, www.m-hkar.com hp://dx.do.org/0.988/ams.05.54349 On One Analyc Mehod of Consrucng Program Conrols A. N. Kvko, S. V. Chsyakov and Yu. E. Balyna

More information

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1 RL Lecure 7: Eligibiliy Traces R. S. Suon and A. G. Baro: Reinforcemen Learning: An Inroducion 1 N-sep TD Predicion Idea: Look farher ino he fuure when you do TD backup (1, 2, 3,, n seps) R. S. Suon and

More information

V.Abramov - FURTHER ANALYSIS OF CONFIDENCE INTERVALS FOR LARGE CLIENT/SERVER COMPUTER NETWORKS

V.Abramov - FURTHER ANALYSIS OF CONFIDENCE INTERVALS FOR LARGE CLIENT/SERVER COMPUTER NETWORKS R&RATA # Vol.) 8, March FURTHER AALYSIS OF COFIDECE ITERVALS FOR LARGE CLIET/SERVER COMPUTER ETWORKS Vyacheslav Abramov School of Mahemacal Scences, Monash Unversy, Buldng 8, Level 4, Clayon Campus, Wellngon

More information

Displacement, Velocity, and Acceleration. (WHERE and WHEN?)

Displacement, Velocity, and Acceleration. (WHERE and WHEN?) Dsplacemen, Velocy, and Acceleraon (WHERE and WHEN?) Mah resources Append A n your book! Symbols and meanng Algebra Geomery (olumes, ec.) Trgonomery Append A Logarhms Remnder You wll do well n hs class

More information

Graduate Macroeconomics 2 Problem set 5. - Solutions

Graduate Macroeconomics 2 Problem set 5. - Solutions Graduae Macroeconomcs 2 Problem se. - Soluons Queson 1 To answer hs queson we need he frms frs order condons and he equaon ha deermnes he number of frms n equlbrum. The frms frs order condons are: F K

More information

January Examinations 2012

January Examinations 2012 Page of 5 EC79 January Examnaons No. of Pages: 5 No. of Quesons: 8 Subjec ECONOMICS (POSTGRADUATE) Tle of Paper EC79 QUANTITATIVE METHODS FOR BUSINESS AND FINANCE Tme Allowed Two Hours ( hours) Insrucons

More information

. The geometric multiplicity is dim[ker( λi. number of linearly independent eigenvectors associated with this eigenvalue.

. The geometric multiplicity is dim[ker( λi. number of linearly independent eigenvectors associated with this eigenvalue. Lnear Algebra Lecure # Noes We connue wh he dscusson of egenvalues, egenvecors, and dagonalzably of marces We wan o know, n parcular wha condons wll assure ha a marx can be dagonalzed and wha he obsrucons

More information

Reactive Methods to Solve the Berth AllocationProblem with Stochastic Arrival and Handling Times

Reactive Methods to Solve the Berth AllocationProblem with Stochastic Arrival and Handling Times Reacve Mehods o Solve he Berh AllocaonProblem wh Sochasc Arrval and Handlng Tmes Nsh Umang* Mchel Berlare* * TRANSP-OR, Ecole Polyechnque Fédérale de Lausanne Frs Workshop on Large Scale Opmzaon November

More information

Panel Data Regression Models

Panel Data Regression Models Panel Daa Regresson Models Wha s Panel Daa? () Mulple dmensoned Dmensons, e.g., cross-secon and me node-o-node (c) Pongsa Pornchawseskul, Faculy of Economcs, Chulalongkorn Unversy (c) Pongsa Pornchawseskul,

More information

Performance Analysis for a Network having Standby Redundant Unit with Waiting in Repair

Performance Analysis for a Network having Standby Redundant Unit with Waiting in Repair TECHNI Inernaonal Journal of Compung Scence Communcaon Technologes VOL.5 NO. July 22 (ISSN 974-3375 erformance nalyss for a Nework havng Sby edundan Un wh ang n epar Jendra Sngh 2 abns orwal 2 Deparmen

More information

We are estimating the density of long distant migrant (LDM) birds in wetlands along Lake Michigan.

We are estimating the density of long distant migrant (LDM) birds in wetlands along Lake Michigan. Ch 17 Random ffecs and Mxed Models 17. Random ffecs Models We are esmang he densy of long dsan mgran (LDM) brds n welands along Lake Mchgan. μ + = LDM per hecaren h weland ~ N(0, ) The varably of expeced

More information

J i-1 i. J i i+1. Numerical integration of the diffusion equation (I) Finite difference method. Spatial Discretization. Internal nodes.

J i-1 i. J i i+1. Numerical integration of the diffusion equation (I) Finite difference method. Spatial Discretization. Internal nodes. umercal negraon of he dffuson equaon (I) Fne dfference mehod. Spaal screaon. Inernal nodes. R L V For hermal conducon le s dscree he spaal doman no small fne spans, =,,: Balance of parcles for an nernal

More information

Chapters 2 Kinematics. Position, Distance, Displacement

Chapters 2 Kinematics. Position, Distance, Displacement Chapers Knemacs Poson, Dsance, Dsplacemen Mechancs: Knemacs and Dynamcs. Knemacs deals wh moon, bu s no concerned wh he cause o moon. Dynamcs deals wh he relaonshp beween orce and moon. The word dsplacemen

More information

. The geometric multiplicity is dim[ker( λi. A )], i.e. the number of linearly independent eigenvectors associated with this eigenvalue.

. The geometric multiplicity is dim[ker( λi. A )], i.e. the number of linearly independent eigenvectors associated with this eigenvalue. Mah E-b Lecure #0 Noes We connue wh he dscusson of egenvalues, egenvecors, and dagonalzably of marces We wan o know, n parcular wha condons wll assure ha a marx can be dagonalzed and wha he obsrucons are

More information

Fall 2009 Social Sciences 7418 University of Wisconsin-Madison. Problem Set 2 Answers (4) (6) di = D (10)

Fall 2009 Social Sciences 7418 University of Wisconsin-Madison. Problem Set 2 Answers (4) (6) di = D (10) Publc Affars 974 Menze D. Chnn Fall 2009 Socal Scences 7418 Unversy of Wsconsn-Madson Problem Se 2 Answers Due n lecure on Thursday, November 12. " Box n" your answers o he algebrac quesons. 1. Consder

More information

Dynamic Team Decision Theory. EECS 558 Project Shrutivandana Sharma and David Shuman December 10, 2005

Dynamic Team Decision Theory. EECS 558 Project Shrutivandana Sharma and David Shuman December 10, 2005 Dynamc Team Decson Theory EECS 558 Proec Shruvandana Sharma and Davd Shuman December 0, 005 Oulne Inroducon o Team Decson Theory Decomposon of he Dynamc Team Decson Problem Equvalence of Sac and Dynamc

More information

Online Appendix for. Strategic safety stocks in supply chains with evolving forecasts

Online Appendix for. Strategic safety stocks in supply chains with evolving forecasts Onlne Appendx for Sraegc safey socs n supply chans wh evolvng forecass Tor Schoenmeyr Sephen C. Graves Opsolar, Inc. 332 Hunwood Avenue Hayward, CA 94544 A. P. Sloan School of Managemen Massachuses Insue

More information

GMM parameter estimation. Xiaoye Lu CMPS290c Final Project

GMM parameter estimation. Xiaoye Lu CMPS290c Final Project GMM paraeer esaon Xaoye Lu M290c Fnal rojec GMM nroducon Gaussan ure Model obnaon of several gaussan coponens Noaon: For each Gaussan dsrbuon:, s he ean and covarance ar. A GMM h ures(coponens): p ( 2π

More information

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification Instance-Based earnng (a.k.a. memory-based learnng) Part I: Nearest Neghbor Classfcaton Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n

More information

( t) Outline of program: BGC1: Survival and event history analysis Oslo, March-May Recapitulation. The additive regression model

( t) Outline of program: BGC1: Survival and event history analysis Oslo, March-May Recapitulation. The additive regression model BGC1: Survval and even hsory analyss Oslo, March-May 212 Monday May 7h and Tuesday May 8h The addve regresson model Ørnulf Borgan Deparmen of Mahemacs Unversy of Oslo Oulne of program: Recapulaon Counng

More information

10. A.C CIRCUITS. Theoretically current grows to maximum value after infinite time. But practically it grows to maximum after 5τ. Decay of current :

10. A.C CIRCUITS. Theoretically current grows to maximum value after infinite time. But practically it grows to maximum after 5τ. Decay of current : . A. IUITS Synopss : GOWTH OF UNT IN IUIT : d. When swch S s closed a =; = d. A me, curren = e 3. The consan / has dmensons of me and s called he nducve me consan ( τ ) of he crcu. 4. = τ; =.63, n one

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Lnear Response Theory: The connecon beween QFT and expermens 3.1. Basc conceps and deas Q: ow do we measure he conducvy of a meal? A: we frs nroduce a weak elecrc feld E, and hen measure

More information

Machine Learning Linear Regression

Machine Learning Linear Regression Machne Learnng Lnear Regresson Lesson 3 Lnear Regresson Bascs of Regresson Leas Squares esmaon Polynomal Regresson Bass funcons Regresson model Regularzed Regresson Sascal Regresson Mamum Lkelhood (ML)

More information

Scattering at an Interface: Oblique Incidence

Scattering at an Interface: Oblique Incidence Course Insrucor Dr. Raymond C. Rumpf Offce: A 337 Phone: (915) 747 6958 E Mal: rcrumpf@uep.edu EE 4347 Appled Elecromagnecs Topc 3g Scaerng a an Inerface: Oblque Incdence Scaerng These Oblque noes may

More information

RELATIONSHIP BETWEEN VOLATILITY AND TRADING VOLUME: THE CASE OF HSI STOCK RETURNS DATA

RELATIONSHIP BETWEEN VOLATILITY AND TRADING VOLUME: THE CASE OF HSI STOCK RETURNS DATA RELATIONSHIP BETWEEN VOLATILITY AND TRADING VOLUME: THE CASE OF HSI STOCK RETURNS DATA Mchaela Chocholaá Unversy of Economcs Braslava, Slovaka Inroducon (1) one of he characersc feaures of sock reurns

More information

2. SPATIALLY LAGGED DEPENDENT VARIABLES

2. SPATIALLY LAGGED DEPENDENT VARIABLES 2. SPATIALLY LAGGED DEPENDENT VARIABLES In hs chaper, we descrbe a sascal model ha ncorporaes spaal dependence explcly by addng a spaally lagged dependen varable y on he rgh-hand sde of he regresson equaon.

More information

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore. Ths documen s downloaded from DR-NTU, Nanyang Technologcal Unversy Lbrary, Sngapore. Tle A smplfed verb machng algorhm for word paron n vsual speech processng( Acceped verson ) Auhor(s) Foo, Say We; Yong,

More information

Normal Random Variable and its discriminant functions

Normal Random Variable and its discriminant functions Noral Rando Varable and s dscrnan funcons Oulne Noral Rando Varable Properes Dscrnan funcons Why Noral Rando Varables? Analycally racable Works well when observaon coes for a corruped snle prooype 3 The

More information

Computing Relevance, Similarity: The Vector Space Model

Computing Relevance, Similarity: The Vector Space Model Compung Relevance, Smlary: The Vecor Space Model Based on Larson and Hears s sldes a UC-Bereley hp://.sms.bereley.edu/courses/s0/f00/ aabase Managemen Sysems, R. Ramarshnan ocumen Vecors v ocumens are

More information

Cubic Bezier Homotopy Function for Solving Exponential Equations

Cubic Bezier Homotopy Function for Solving Exponential Equations Penerb Journal of Advanced Research n Compung and Applcaons ISSN (onlne: 46-97 Vol. 4, No.. Pages -8, 6 omoopy Funcon for Solvng Eponenal Equaons S. S. Raml *,,. Mohamad Nor,a, N. S. Saharzan,b and M.

More information

2.1 Constitutive Theory

2.1 Constitutive Theory Secon.. Consuve Theory.. Consuve Equaons Governng Equaons The equaons governng he behavour of maerals are (n he spaal form) dρ v & ρ + ρdv v = + ρ = Conservaon of Mass (..a) d x σ j dv dvσ + b = ρ v& +

More information

Ordinary Differential Equations in Neuroscience with Matlab examples. Aim 1- Gain understanding of how to set up and solve ODE s

Ordinary Differential Equations in Neuroscience with Matlab examples. Aim 1- Gain understanding of how to set up and solve ODE s Ordnary Dfferenal Equaons n Neuroscence wh Malab eamples. Am - Gan undersandng of how o se up and solve ODE s Am Undersand how o se up an solve a smple eample of he Hebb rule n D Our goal a end of class

More information

Dual Approximate Dynamic Programming for Large Scale Hydro Valleys

Dual Approximate Dynamic Programming for Large Scale Hydro Valleys Dual Approxmae Dynamc Programmng for Large Scale Hydro Valleys Perre Carpener and Jean-Phlppe Chanceler 1 ENSTA ParsTech and ENPC ParsTech CMM Workshop, January 2016 1 Jon work wh J.-C. Alas, suppored

More information

Volatility Interpolation

Volatility Interpolation Volaly Inerpolaon Prelmnary Verson March 00 Jesper Andreasen and Bran Huge Danse Mares, Copenhagen wan.daddy@danseban.com brno@danseban.com Elecronc copy avalable a: hp://ssrn.com/absrac=69497 Inro Local

More information

Advanced Macroeconomics II: Exchange economy

Advanced Macroeconomics II: Exchange economy Advanced Macroeconomcs II: Exchange economy Krzyszof Makarsk 1 Smple deermnsc dynamc model. 1.1 Inroducon Inroducon Smple deermnsc dynamc model. Defnons of equlbrum: Arrow-Debreu Sequenal Recursve Equvalence

More information

Reinforcement learning

Reinforcement learning Renforcement learnng Nathanel Daw Gatsby Computatonal Neuroscence Unt daw @ gatsby.ucl.ac.uk http://www.gatsby.ucl.ac.uk/~daw Mostly adapted from Andrew Moore s tutorals, copyrght 2002, 2004 by Andrew

More information

Epistemic Game Theory: Online Appendix

Epistemic Game Theory: Online Appendix Epsemc Game Theory: Onlne Appendx Edde Dekel Lucano Pomao Marcano Snscalch July 18, 2014 Prelmnares Fx a fne ype srucure T I, S, T, β I and a probably µ S T. Le T µ I, S, T µ, βµ I be a ype srucure ha

More information

Attribute Reduction Algorithm Based on Discernibility Matrix with Algebraic Method GAO Jing1,a, Ma Hui1, Han Zhidong2,b

Attribute Reduction Algorithm Based on Discernibility Matrix with Algebraic Method GAO Jing1,a, Ma Hui1, Han Zhidong2,b Inernaonal Indusral Informacs and Compuer Engneerng Conference (IIICEC 05) Arbue educon Algorhm Based on Dscernbly Marx wh Algebrac Mehod GAO Jng,a, Ma Hu, Han Zhdong,b Informaon School, Capal Unversy

More information

Lecture Slides for INTRODUCTION TO. Machine Learning. ETHEM ALPAYDIN The MIT Press,

Lecture Slides for INTRODUCTION TO. Machine Learning. ETHEM ALPAYDIN The MIT Press, Lecure Sldes for INTRDUCTIN T Machne Learnng ETHEM ALAYDIN The MIT ress, 2004 alpaydn@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/2ml CHATER 3: Hdden Marov Models Inroducon Modelng dependences n npu; no

More information

, t 1. Transitions - this one was easy, but in general the hardest part is choosing the which variables are state and control variables

, t 1. Transitions - this one was easy, but in general the hardest part is choosing the which variables are state and control variables Opmal Conrol Why Use I - verss calcls of varaons, opmal conrol More generaly More convenen wh consrans (e.g., can p consrans on he dervaves More nsghs no problem (a leas more apparen han hrogh calcls of

More information

Lecture 2 L n i e n a e r a M od o e d l e s

Lecture 2 L n i e n a e r a M od o e d l e s Lecure Lnear Models Las lecure You have learned abou ha s machne learnng Supervsed learnng Unsupervsed learnng Renforcemen learnng You have seen an eample learnng problem and he general process ha one

More information

Time-interval analysis of β decay. V. Horvat and J. C. Hardy

Time-interval analysis of β decay. V. Horvat and J. C. Hardy Tme-nerval analyss of β decay V. Horva and J. C. Hardy Work on he even analyss of β decay [1] connued and resuled n he developmen of a novel mehod of bea-decay me-nerval analyss ha produces hghly accurae

More information

UNIVERSITAT AUTÒNOMA DE BARCELONA MARCH 2017 EXAMINATION

UNIVERSITAT AUTÒNOMA DE BARCELONA MARCH 2017 EXAMINATION INTERNATIONAL TRADE T. J. KEHOE UNIVERSITAT AUTÒNOMA DE BARCELONA MARCH 27 EXAMINATION Please answer wo of he hree quesons. You can consul class noes, workng papers, and arcles whle you are workng on he

More information

Biol. 356 Lab 8. Mortality, Recruitment, and Migration Rates

Biol. 356 Lab 8. Mortality, Recruitment, and Migration Rates Biol. 356 Lab 8. Moraliy, Recruimen, and Migraion Raes (modified from Cox, 00, General Ecology Lab Manual, McGraw Hill) Las week we esimaed populaion size hrough several mehods. One assumpion of all hese

More information

CS286.2 Lecture 14: Quantum de Finetti Theorems II

CS286.2 Lecture 14: Quantum de Finetti Theorems II CS286.2 Lecure 14: Quanum de Fne Theorems II Scrbe: Mara Okounkova 1 Saemen of he heorem Recall he las saemen of he quanum de Fne heorem from he prevous lecure. Theorem 1 Quanum de Fne). Le ρ Dens C 2

More information

F-Tests and Analysis of Variance (ANOVA) in the Simple Linear Regression Model. 1. Introduction

F-Tests and Analysis of Variance (ANOVA) in the Simple Linear Regression Model. 1. Introduction ECOOMICS 35* -- OTE 9 ECO 35* -- OTE 9 F-Tess and Analyss of Varance (AOVA n he Smple Lnear Regresson Model Inroducon The smple lnear regresson model s gven by he followng populaon regresson equaon, or

More information

THE PREDICTION OF COMPETITIVE ENVIRONMENT IN BUSINESS

THE PREDICTION OF COMPETITIVE ENVIRONMENT IN BUSINESS THE PREICTION OF COMPETITIVE ENVIRONMENT IN BUSINESS INTROUCTION The wo dmensonal paral dfferenal equaons of second order can be used for he smulaon of compeve envronmen n busness The arcle presens he

More information

OMXS30 Balance 20% Index Rules

OMXS30 Balance 20% Index Rules OMX30 Balance 0% ndex Rules Verson as of 30 March 009 Copyrgh 008, The NADAQ OMX Group, nc. All rghs reserved. NADAQ OMX, The NADAQ ock Marke and NADAQ are regsered servce/rademarks of The NADAQ OMX Group,

More information

Single-loop System Reliability-Based Design & Topology Optimization (SRBDO/SRBTO): A Matrix-based System Reliability (MSR) Method

Single-loop System Reliability-Based Design & Topology Optimization (SRBDO/SRBTO): A Matrix-based System Reliability (MSR) Method 10 h US Naonal Congress on Compuaonal Mechancs Columbus, Oho 16-19, 2009 Sngle-loop Sysem Relably-Based Desgn & Topology Opmzaon (SRBDO/SRBTO): A Marx-based Sysem Relably (MSR) Mehod Tam Nguyen, Junho

More information

Approximate Analytic Solution of (2+1) - Dimensional Zakharov-Kuznetsov(Zk) Equations Using Homotopy

Approximate Analytic Solution of (2+1) - Dimensional Zakharov-Kuznetsov(Zk) Equations Using Homotopy Arcle Inernaonal Journal of Modern Mahemacal Scences, 4, (): - Inernaonal Journal of Modern Mahemacal Scences Journal homepage: www.modernscenfcpress.com/journals/jmms.aspx ISSN: 66-86X Florda, USA Approxmae

More information

A Cell Decomposition Approach to Online Evasive Path Planning and the Video Game Ms. Pac-Man

A Cell Decomposition Approach to Online Evasive Path Planning and the Video Game Ms. Pac-Man Cell Decomoson roach o Onlne Evasve Pah Plannng and he Vdeo ame Ms. Pac-Man reg Foderaro Vram Raju Slva Ferrar Laboraory for Inellgen Sysems and Conrols LISC Dearmen of Mechancal Engneerng and Maerals

More information

A Deterministic Algorithm for Summarizing Asynchronous Streams over a Sliding Window

A Deterministic Algorithm for Summarizing Asynchronous Streams over a Sliding Window A Deermnsc Algorhm for Summarzng Asynchronous Sreams over a Sldng ndow Cosas Busch Rensselaer Polyechnc Insue Srkana Trhapura Iowa Sae Unversy Oulne of Talk Inroducon Algorhm Analyss Tme C Daa sream: 3

More information

Appendix H: Rarefaction and extrapolation of Hill numbers for incidence data

Appendix H: Rarefaction and extrapolation of Hill numbers for incidence data Anne Chao Ncholas J Goell C seh lzabeh L ander K Ma Rober K Colwell and Aaron M llson 03 Rarefacon and erapolaon wh ll numbers: a framewor for samplng and esmaon n speces dversy sudes cology Monographs

More information

An introduction to Support Vector Machine

An introduction to Support Vector Machine An nroducon o Suppor Vecor Machne 報告者 : 黃立德 References: Smon Haykn, "Neural Neworks: a comprehensve foundaon, second edon, 999, Chaper 2,6 Nello Chrsann, John Shawe-Tayer, An Inroducon o Suppor Vecor Machnes,

More information

Lecture 9: Dynamic Properties

Lecture 9: Dynamic Properties Shor Course on Molecular Dynamcs Smulaon Lecure 9: Dynamc Properes Professor A. Marn Purdue Unversy Hgh Level Course Oulne 1. MD Bascs. Poenal Energy Funcons 3. Inegraon Algorhms 4. Temperaure Conrol 5.

More information

Clustering (Bishop ch 9)

Clustering (Bishop ch 9) Cluserng (Bshop ch 9) Reference: Daa Mnng by Margare Dunham (a slde source) 1 Cluserng Cluserng s unsupervsed learnng, here are no class labels Wan o fnd groups of smlar nsances Ofen use a dsance measure

More information

Li An-Ping. Beijing , P.R.China

Li An-Ping. Beijing , P.R.China A New Type of Cpher: DICING_csb L An-Png Bejng 100085, P.R.Chna apl0001@sna.com Absrac: In hs paper, we wll propose a new ype of cpher named DICING_csb, whch s derved from our prevous sream cpher DICING.

More information

How about the more general "linear" scalar functions of scalars (i.e., a 1st degree polynomial of the following form with a constant term )?

How about the more general linear scalar functions of scalars (i.e., a 1st degree polynomial of the following form with a constant term )? lmcd Lnear ransformaon of a vecor he deas presened here are que general hey go beyond he radonal mar-vecor ype seen n lnear algebra Furhermore, hey do no deal wh bass and are equally vald for any se of

More information

Chapter Lagrangian Interpolation

Chapter Lagrangian Interpolation Chaper 5.4 agrangan Inerpolaon Afer readng hs chaper you should be able o:. dere agrangan mehod of nerpolaon. sole problems usng agrangan mehod of nerpolaon and. use agrangan nerpolans o fnd deraes and

More information

Bandlimited channel. Intersymbol interference (ISI) This non-ideal communication channel is also called dispersive channel

Bandlimited channel. Intersymbol interference (ISI) This non-ideal communication channel is also called dispersive channel Inersymol nererence ISI ISI s a sgnal-dependen orm o nererence ha arses ecause o devaons n he requency response o a channel rom he deal channel. Example: Bandlmed channel Tme Doman Bandlmed channel Frequency

More information

DEEP UNFOLDING FOR MULTICHANNEL SOURCE SEPARATION SUPPLEMENTARY MATERIAL

DEEP UNFOLDING FOR MULTICHANNEL SOURCE SEPARATION SUPPLEMENTARY MATERIAL DEEP UNFOLDING FOR MULTICHANNEL SOURCE SEPARATION SUPPLEMENTARY MATERIAL Sco Wsdom, John Hershey 2, Jonahan Le Roux 2, and Shnj Waanabe 2 Deparmen o Elecrcal Engneerng, Unversy o Washngon, Seale, WA, USA

More information