Reinforcement Learning
|
|
- Marjory Montgomery
- 6 years ago
- Views:
Transcription
1 Noe o oher eachers and users of hese sldes. Andrew would be delghed f you found hs source maeral useful n gvng your own lecures. Feel free o use hese sldes verbam, or o modfy hem o f your own needs. PowerPon orgnals are avalable. If you make use of a sgnfcan poron of hese sldes n your own lecure, please nclude hs message, or he followng lnk o he source reposory of Andrew s uorals: hp:// Commens and correcons graefully receved. Renforcemen Learnng Andrew W. Moore Assocae Professor School of Compuer Scence Carnege Mellon Unversy awm@cs.cmu.edu Copyrgh 2002, Andrew W. Moore Aprl 23rd, 2002 Predcng Delayed Rewards IN A 0.4 DISCOUNTED MARKOV SYSTEM R=-5 S R=0 S R=0 0.5 S R=3 R=0 S 0.2 S R= S 6 Prob(nex sae = S 5 hs sae = S 4 ) = 0.8 ec Wha s expeced sum of fuure rewards (dscouned)? Ε = 0 γ ( [ ] ) S[ 0] = S R S us Solve I! We use sandard Markov Sysem Theory Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 2
2 Learnng Delayed Rewards S? R=???? S 4? R=??? S 2? R=??? S 5? R=??? S 3? R=??? S 6? R=??? All you can see s a seres of saes and rewards: S (R=0) S 2 (R=0) S 3 (R=4) S 2 (R=0) S 4 (R=0) S 5 (R=0) Task: Based on hs sequence, mae *(S ),*(S 2 ) *(S 6 ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 3 Idea : Supervsed Learnng S (R=0) S 2 (R=0) Assume?=/2 S 3 (R=4) S 2 (R=0) S 4 (R=0) S 5 (R=0) A = we were n sae S and evenually go a long erm dscouned reward of 0+?0+? 2 4+? 3 0+? 4 0 = A =2 n sae S 2 ldr = 2 A =5 n sae S 4 ldr = 0 A =3 n sae S 3 ldr = 4 A =6 n sae S 5 ldr = 0 A =4 n sae S 2 ldr = 0 Sae Observaons of LTDR Mean LTDR S = (S ) S 2 2, 0 = (S 2 ) S = (S 3 ) S = (S 4 ) S = (S 5 ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 4 2
3 Supervsed Learnng ALG Wach a raecory S[0] r[0] S[] r[] S[T]r[T] For =0,, T, compue [ ] Compue ( S ) = =0 mean value of = among n sae S Le MATCHES ( ) γ r[ + ] [ ] all ranso ns begnnng on he raecor y ( S ) = { S[ ] [ ] MATCHES S You re done! S = MATCHES( S ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 5 ( ) = S }, hen defne Supervsed Learnng ALG for he md If you have an anxous personaly you may be worred abou edge effecs for some of he fnal ransons. Wh large raecores hese are neglgble. Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 6 3
4 Onlne Supervsed Learnng Inalze: Coun[S ] = 0 S Sum[S ] = 0 S Elgbly[S ] = 0 S Observe: When we experence S wh reward r do hs: Elg[S ]?Elg[S ] Elg[S ] Elg[S ] + Sum[S ] Sum[S ]+rxelg[s ] Coun[S ] Coun[S ] + Then a any me, (S )= Sum[S ]/Coun[S ] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 7 Onlne Supervsed Learnng Economcs Gven N saes S S N, OSL needs O(N) memory. Each updae needs O(N) work snce we mus updae all Elg[ ] array elemens Idea: Be sparse and only updae/process Elg[ ] elemens wh values >?for ny? There are only log such elemens ξ log γ Easy o prove: As T, ( S ) ( S) S Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 8 4
5 Onlne Supervsed Learnng Le s grab OSL off he sree, bundle no a black van, ake o a bunker and nerrogae under 600 Wa lghs. S (r=0) S 2 (r=0) Sae Observaons of LTDR ^ (S ) S S 2 2, 0 S S S COMPLAINT S 3 (r=4) S 2 (r=0) S 4 (r=0) S 5 (r=0) There s somehng a lle suspcous abou hs (effcency-wse) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 9 Cerany-Equvalen (CE) Learnng S (r=0) S 2 (r=0) Idea: Use your daa o mae he underlyng Markov sysem, nsead of ryng o mae drecly. Esmaed Markov Sysem: S 3 (r=4) S 2 (r=0) S 4 (r=0) S 5 (r=0) You draw n he ransons + probs S r = 0 S 2 r = 0 S 3 r = 4 S 4 r = 0 S 5 r = 0 Wha re he maed values? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 0 5
6 C.E. Mehod for Markov Sysems Inalze: Coun[S ] = 0 S #Tmes vsed S SumR[S ] = 0 Sum of rewards from S S Trans[S,S ] = 0 #Tmes ransoned from S S When we are n sae S, and we receve reward r, and we move o S Coun[S ] Coun[S ] + SumR[S ] SumR[S ] + r Trans[S,S ] Trans[S,S ] + Then a any me r (S ) = SumR[S ] / Coun[S ] P = Esmaed Prob(nex = S hs = S ) = Trans[S,S ] / Coun[S ] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde C.E. for Markov Sysems (connued) So a any me we have r (S ) and P (nex=s hs=s ) S S = P So a any me we can solve he se of lnear equaons ( S ) = r ( S ) + γ P ( S S ) ( S ) [In vecor noaon, = r +?P => = (I-?P ) - r where r are vecors of lengh N P s an NxN marx N = # saes ] S Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 2 6
7 C.E. Onlne Economcs Memory: O(N 2 ) Tme o updae couners: O() Tme o re-evaluae O(N 3 ) f use marx nverson O(N 2 k CRIT ) f use value eraon and we need k CRIT eraons o converge O(Nk CRIT ) f use value eraon, and k CRIT o converge, and M.S. s Sparse (.e. mean # successors s consan) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 3 Cerany Equvalen Learnng Memory use could be O(N 2 )! And me per updae could be O(Nk CRIT ) up o O(N 3 )! Too expensve for some people. COMPLAINT Prorzed sweepng wll help, (see laer), bu frs le s revew a very nexpensve approach Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 4 7
8 Why hs obsesson wh onlneness? I really care abou supplyng up-o-dae maes all he me. Can you guess why? If no, all wll be revealed n good me Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 5 Less Tme: More Daa Lmed Backups Do prevous C.E. algorhm. A each me mep we observe S (r) S and updae Coun[S ], SumR[S ], Trans[S,S ] And hus also updae maes r and P oucomes ( S ) Bu nsead of re-solvng for, do much less work. us do one backup of S [ ] [ S ] r + γ P [ S ] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 6 8
9 One Backup C.E. Economcs Space : O(N 2 ) NO IMPROVEMENT THERE! Tme o updae sascs : O() Tme o updae : O() Good News: Much cheaper per ranson Good News: Conracon Mappng proof (modfed) promses convergence o opmal Bad News: Wases daa Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 7 Prorzed Sweepng [Moore + Akeson, 93] Tres o be almos as daa-effcen as full CE bu no much more expensve han One Backup CE. On every ranson, some number (ß) of saes may have a backup appled. Whch ones? The mos deservng We keep a prory queue of whch saes have he bgg poenal for changng her (S) value Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 8 9
10 Where Are We? Tryng o do onlne predcon from sreams of ransons Space Updae Cos Daa Effcency: Supervsed Learnng Full C.E. Learnng One Backup C.E. Learnng Prorzed Sweepng 0(N s ) 0(N so ) 0(N so ) 0(N so ) 0( ) 0(N so N s ) 0(N so k CRIT ) 0() 0() log(/?) N so = # sae-oucomes (number of arrows on he M.S. dagram) N s = # saes Wha Nex? Sample Backups!!! Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 9 Temporal Dfference Learnng Only manan a array nohng else [Suon 988] So you ve go (S ) (S 2 ), (S N ) and you observe S r S wha should you do? Can You Guess? A ranson from ha receves an mmedae reward of r and umps o Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 20 0
11 S r S TD Learnng ( S ) We updae = We nudge o be closer o expeced fuure rewards α = ( S ) ( α) ( S ) + Expeced fuure α[ ] rewards ( α) ( S ) + α[ r + γ ( S )] WEIGHTED SUM s called a learnng rae parameer. (See? n he neural lecure) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 2 Smplfed TD Analyss P =? r =? TERMINATE S 0 P 2 =? r 2 =? TERMINATE r=0 P M =? : r M =? TERMINATE Suppose you always begn n S 0 You hen ranson a random o one of M places. You don know he ranson probs. You hen ge a place-dependen reward (unknown n advance). Then he ral ermnaes. Defne *(S 0 )= Expeced reward Le s mae wh TD Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 22
12 S 0 r=0 p () p (2) p (N) r () r (2) r (N) r (k) = reward of k h ermnal sae p (k) = prob of k h ermnal sae We ll do a seres of rals. Reward on h = Ε r n = p k= k r k ral s r [ ] ( ) ( ) [ ] [ Noe Ε r s ndependen of ] Defne *(S 0 ) = * = E[r ] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 23 Le s run TD-Learnng, where = Esmae (S 0 ) before he h ral. From defnon of TD-Learnng: + = (-a) + ar Useful quany: Defne σ 2 = Varance of reward= Ε M = P k= ( r ) 2 ( k ) ( k ) [ ] 2 ( r ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 24 2
13 Ε = Remember * = E[r ], s 2 = E[(r -*) 2 ] + = ar + (-a) [ + ] = [ αr + ( α ) ] = Ε ( α ) Ε[ ] Thus... lm Ε [ ] = Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 25 W H Y? Is hs mpressve?? Remember * = E[r ], s 2 = E[(r -*) 2 ] + = ar + (-a) Wre S = Expeced squared error beween and * before he h eraon S + = E[( + -*) 2 ] = E[(ar +(-a) - *) 2 ] = E[(a[r -*]+(-a)[ - *]) 2 ] = E[a 2 (r -*) 2 +a(-a)(r -*)( - *)+(-a) 2 ( - *) 2 ] = a 2 E[(r -*) 2 ]+a(-a)e[(r -*)( - *)]+(-a) 2 E[( - *) 2 ] = = a 2 s 2 +(-a) 2 S WHY? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 26 3
14 And s hus easy o show ha. lm S = lm Ε [( ) ] 2 2 ασ = (2 α) Wha do you hnk of TD learnng? How would you mprove? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 27 Decayng Learnng Rae [Dayan 99sh] showed ha for General TD learnng of a Markow Sysem (no us our smple model) ha f you use updae rule ( S ) α [ r + γ ( S )] + ( α ) ( S ) hen, as number of observaons goes o nfny S S PROVIDED All saes vsed 8 ly ofen α = = ( ) ( ) 2 T 2 α < k. T. α < k = = Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 28 k. T. Ths means T = Ths means α > k 4
15 Decayng Learnng Rae Ths Works: a = / Ths Doesn : a = a 0 Ths Works: a = ß/(ß+) [e.g. ß=000] Ths Doesn : a = ßa - (ß<) IN OUR EXAMPLE.USE a = / 2 2 Remember = Ε[ r ], σ = Ε[ (r ) ] + Wre = α r C C + ( α ) = r + ( ) = + ( ) = r + C and so you' ll + = r + = see ha 0 And Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 29 Decayng Learnng Rae con σ + [ ] ( 0 - ) = 2 lm Ε[ ( - ) ] = 0 Ε ( - ) so, ulmaely Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 30 5
16 A Fancer TD Wre S[] = sae a me Suppose a = /4? = /2 Assume (S 23 )=0 (S 7 )=0 (S 44 )=6 Assume = 405 and S[] = S 23 Observe (r=0) S 23 S 7 wh reward 0 Now = 406, S[] = S 7, S[-] = S 23 (S 23 )=, (S 7 )=, (S 44 )= (r=0) Observe S 7 S 44 Now = 407, S[] = S44 (S 23 )=, (S 7 )=, (S 44 )= INSIGHT: (S 23 ) mgh hnk I goa ge me some of ha!!! Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 3 TD(?) Commens TD(?=0) s he orgnal TD TD(?=) s almos he same as supervsed learnng (excep uses a learnng rae nsead of explc couns) TD(?=0.7) s ofen emprcally he b performer Dayan s proof holds for all 0=?= Updaes can be made more compuaonally effcen wh elgbly races (smlar o O.S.L.) Quon: Can you nven a problem ha would make TD(0) look bad and TD() look good? How abou TD(0) look good & TD() bad?? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 32 6
17 Learnng M.S. Summary MODEL-BASED Supervsed Learnng Full C.E. Learnng One Backup C.E. Learnng Prorzed Sweepng Space 0(N s ) 0(N so ) 0(N so ) 0(N so ) Updae Cos 0 log γ 0(N so N s ) 0(N so k CRIT ) 0() 0() Daa Effcency MODEL FREE TD(0) TD(?), 0<?= 0(N s ) 0(N s ) 0() 0 log γλ Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 33 Learnng Polces for MDPs See prevous lecure sldes for defnon of and compuaon wh MDPs. The Hear of REINFORCEMENT Learnng sae Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 34 7
18 The ask: World: You are n sae 34. Your mmedae reward s 3. You have 3 acons. Robo: I ll ake acon 2. World: You are n sae 77. Your mmedae reward s -7. You have 2 acons. Robo: I ll ake acon. World: You re n sae 34 (agan). Your mmedae reward s 3. You have 3 acons. The Markov propery means once you ve seleced an acon he P.D.F. of your nex sae s he same as he las me you red he acon n hs sae. Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 35 The Cred Assgnmen Problem I m n sae 43, 39, 22, 2, 2, 3, 54, 26, reward = 0, = 0, = 0, = 0, = 0, = 0, = 0, = 00, acon = 2 = 4 = = = = 2 = 2 Yppee! I go o a sae wh a bg reward! Bu whch of my acons along he way acually helped me ge here?? Ths s he Cred Assgnmen problem. I makes Supervsed Learnng approaches (e.g. Boxes [Mche & Chambers]) very, very slow. Usng he MDP assumpon helps avod hs problem. Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 36 8
19 Full C.E. Learnng One Backup C.E. Learnng Prorzed Sweepng MDP Polcy Learnng 0(N sao ) 0(N sao ) 0(N sao ) Space Updae Cos 0(N sao k CRIT ) 0(N?0 ) 0(ßN?0 ) We ll hnk abou Model-Free n a momen Daa Effcency The C.E. mehods are very smlar o he MS case, excep now do value-eraon-for-mdp backups = a S ( S ) max r + γ P ( S S, a) ( S ) SUCCS ( ) S Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 37 Choosng Acons We re n sae S We can mae r P (nex = S hs = S, acon a) (nex = S ) So wha acon should we choose? IDEA : IDEA 2 : a = arg max r + γ P a a = random Any problems wh hese deas? Any oher suggons? Could we be opmal? ( S S, a ) ( S ) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 38 9
20 Model-Free R.L. Why no use T.D.? Observe updae r S a S ( S ) α ( r + γ ( S )) + ( α ) ( S ) Wha s wrong wh hs? Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 39 Q-Learnng: Model-Free R.L. [Wakns, 988] Defne Q*(S,a)= Expeced sum of dscouned fuure rewards f I sar n sae S, f I hen ake acon a, and f I m subsequenly opmal Quons: Defne Q*(S,a) n erms of * Defne *(S ) n erms of Q* Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 40 20
21 Q Noe ha Q Q-Learnng Updae ( S, a) = r + ( ) Q ( a γ P S S, α max S, ) S SUCCS ( S ) In Q-learnng we manan a able of Q values nsead of values When you see S ( S, a) α r γ max Q ( S, a ) + ( α ) Q ( S, a) + reward acon a a S do Ths s even cleverer han looks: he Q values are no based by any parcular exploraon polcy. I avods he Cred Assgnmen problem. a Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 4 Q-Learnng: Choosng Acons Same ssues as for CE choosng acons Don always be greedy, so don always choose: arg max Q a Don always be random (oherwse wll ake a long me o reach somewhere excng) Bolzmann exploraon [Wakns] Q Prob(choose acon a) exp K ( s, a) Opmsm n he face of uncerany [Suon 90, Kaelblng 90] Inalze Q-values opmscally hgh o encourage exploraon Or ake no accoun how ofen each s,a par has been red ( ) s, a Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 42 2
22 Q-Learnng Commens [Wakns] proved ha Q-learnng wll evenually converge o an opmal polcy. Emprcally s cue Emprcally s very slow Why no do Q(?)? Would no make much sense [renroduce he cred assgnmen problem] Some people (e.g. Peng & Wllams) have red o work her way around hs. Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 43 If we had me Value funcon approxmaon Use a Neural Ne o represen [e.g. Tesauro] Use a Neural Ne o represen Q [e.g. Cres] Use a decson ree wh Q-learnng [Chapman + Kaelblng 9] wh C.E. learnng [Moore 9] How o spl up space? Sgnfcance on Q values [Chapman + Kaelblng] Execuon accuracy monorng [Moore 9] Game Theory [Moore + Akeson 95] New nfluence/varance crera [Munos 99] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 44 22
23 R.L. Theory If we had me Counerexamples [Boyan + Moore], [Bard] Value Funcon Approxmaors wh Averagng wll converge o somehng [Gordon] Neural Nes can fal [Bard] Neural Nes wh Resdual Graden updaes wll converge o somehng Lnear approxmaors for TD learnng wll converge o somehng useful [Tsskls + Van Roy] Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 45 Wha You Should Know Supervsed learnng for predcng delayed rewards Cerany equvalen learnng for predcng delayed rewards Model free learnng (TD) for predcng delayed rewards Renforcemen Learnng wh MDPs: Wha s he ask? Why s hard o choose acons? Q-learnng (ncludng beng able o work hrough small smulaed examples of RL) Copyrgh 2002, Andrew W. Moore Renforcemen Learnng: Slde 46 23
Lecture 6: Learning for Control (Generalised Linear Regression)
Lecure 6: Learnng for Conrol (Generalsed Lnear Regresson) Conens: Lnear Mehods for Regresson Leas Squares, Gauss Markov heorem Recursve Leas Squares Lecure 6: RLSC - Prof. Sehu Vjayakumar Lnear Regresson
More informationHidden Markov Models Following a lecture by Andrew W. Moore Carnegie Mellon University
Hdden Markov Models Followng a lecure by Andrew W. Moore Carnege Mellon Unversy www.cs.cmu.edu/~awm/uorals A Markov Sysem Has N saes, called s, s 2.. s N s 2 There are dscree meseps, 0,, s s 3 N 3 0 Hdden
More informationLecture VI Regression
Lecure VI Regresson (Lnear Mehods for Regresson) Conens: Lnear Mehods for Regresson Leas Squares, Gauss Markov heorem Recursve Leas Squares Lecure VI: MLSC - Dr. Sehu Vjayakumar Lnear Regresson Model M
More informationClustering with Gaussian Mixtures
Noe o oher eachers and users of hese sldes. Andrew would be delghed f you found hs source maeral useful n gvng your own lecures. Feel free o use hese sldes verbam, or o modfy hem o f your own needs. PowerPon
More informationCS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 4
CS434a/54a: Paern Recognon Prof. Olga Veksler Lecure 4 Oulne Normal Random Varable Properes Dscrmnan funcons Why Normal Random Varables? Analycally racable Works well when observaon comes form a corruped
More informationJohn Geweke a and Gianni Amisano b a Departments of Economics and Statistics, University of Iowa, USA b European Central Bank, Frankfurt, Germany
Herarchcal Markov Normal Mxure models wh Applcaons o Fnancal Asse Reurns Appendx: Proofs of Theorems and Condonal Poseror Dsrbuons John Geweke a and Gann Amsano b a Deparmens of Economcs and Sascs, Unversy
More informationFall 2010 Graduate Course on Dynamic Learning
Fall 200 Graduae Course on Dynamc Learnng Chaper 4: Parcle Flers Sepember 27, 200 Byoung-Tak Zhang School of Compuer Scence and Engneerng & Cognve Scence and Bran Scence Programs Seoul aonal Unversy hp://b.snu.ac.kr/~bzhang/
More informationIn the complete model, these slopes are ANALYSIS OF VARIANCE FOR THE COMPLETE TWO-WAY MODEL. (! i+1 -! i ) + [(!") i+1,q - [(!
ANALYSIS OF VARIANCE FOR THE COMPLETE TWO-WAY MODEL The frs hng o es n wo-way ANOVA: Is here neracon? "No neracon" means: The man effecs model would f. Ths n urn means: In he neracon plo (wh A on he horzonal
More informationVariants of Pegasos. December 11, 2009
Inroducon Varans of Pegasos SooWoong Ryu bshboy@sanford.edu December, 009 Youngsoo Cho yc344@sanford.edu Developng a new SVM algorhm s ongong research opc. Among many exng SVM algorhms, we wll focus on
More information( ) () we define the interaction representation by the unitary transformation () = ()
Hgher Order Perurbaon Theory Mchael Fowler 3/7/6 The neracon Represenaon Recall ha n he frs par of hs course sequence, we dscussed he chrödnger and Hesenberg represenaons of quanum mechancs here n he chrödnger
More informationOutline. Probabilistic Model Learning. Probabilistic Model Learning. Probabilistic Model for Time-series Data: Hidden Markov Model
Probablsc Model for Tme-seres Daa: Hdden Markov Model Hrosh Mamsuka Bonformacs Cener Kyoo Unversy Oulne Three Problems for probablsc models n machne learnng. Compung lkelhood 2. Learnng 3. Parsng (predcon
More informationSolution in semi infinite diffusion couples (error function analysis)
Soluon n sem nfne dffuson couples (error funcon analyss) Le us consder now he sem nfne dffuson couple of wo blocks wh concenraon of and I means ha, n a A- bnary sysem, s bondng beween wo blocks made of
More informationDepartment of Economics University of Toronto
Deparmen of Economcs Unversy of Torono ECO408F M.A. Economercs Lecure Noes on Heeroskedascy Heeroskedascy o Ths lecure nvolves lookng a modfcaons we need o make o deal wh he regresson model when some of
More informationFTCS Solution to the Heat Equation
FTCS Soluon o he Hea Equaon ME 448/548 Noes Gerald Reckenwald Porland Sae Unversy Deparmen of Mechancal Engneerng gerry@pdxedu ME 448/548: FTCS Soluon o he Hea Equaon Overvew Use he forward fne d erence
More informationEcon107 Applied Econometrics Topic 5: Specification: Choosing Independent Variables (Studenmund, Chapter 6)
Econ7 Appled Economercs Topc 5: Specfcaon: Choosng Independen Varables (Sudenmund, Chaper 6 Specfcaon errors ha we wll deal wh: wrong ndependen varable; wrong funconal form. Ths lecure deals wh wrong ndependen
More informationDensity Matrix Description of NMR BCMB/CHEM 8190
Densy Marx Descrpon of NMR BCMBCHEM 89 Operaors n Marx Noaon Alernae approach o second order specra: ask abou x magnezaon nsead of energes and ranson probables. If we say wh one bass se, properes vary
More informationThe Finite Element Method for the Analysis of Non-Linear and Dynamic Systems
Swss Federal Insue of Page 1 The Fne Elemen Mehod for he Analyss of Non-Lnear and Dynamc Sysems Prof. Dr. Mchael Havbro Faber Dr. Nebojsa Mojslovc Swss Federal Insue of ETH Zurch, Swzerland Mehod of Fne
More informationWiH Wei He
Sysem Idenfcaon of onlnear Sae-Space Space Baery odels WH We He wehe@calce.umd.edu Advsor: Dr. Chaochao Chen Deparmen of echancal Engneerng Unversy of aryland, College Par 1 Unversy of aryland Bacground
More information(,,, ) (,,, ). In addition, there are three other consumers, -2, -1, and 0. Consumer -2 has the utility function
MACROECONOMIC THEORY T J KEHOE ECON 87 SPRING 5 PROBLEM SET # Conder an overlappng generaon economy le ha n queon 5 on problem e n whch conumer lve for perod The uly funcon of he conumer born n perod,
More informationCHAPTER 10: LINEAR DISCRIMINATION
CHAPER : LINEAR DISCRIMINAION Dscrmnan-based Classfcaon 3 In classfcaon h K classes (C,C,, C k ) We defned dscrmnan funcon g j (), j=,,,k hen gven an es eample, e chose (predced) s class label as C f g
More informationLecture 11 SVM cont
Lecure SVM con. 0 008 Wha we have done so far We have esalshed ha we wan o fnd a lnear decson oundary whose margn s he larges We know how o measure he margn of a lnear decson oundary Tha s: he mnmum geomerc
More informationLet s treat the problem of the response of a system to an applied external force. Again,
Page 33 QUANTUM LNEAR RESPONSE FUNCTON Le s rea he problem of he response of a sysem o an appled exernal force. Agan, H() H f () A H + V () Exernal agen acng on nernal varable Hamlonan for equlbrum sysem
More informationMechanics Physics 151
Mechancs Physcs 5 Lecure 9 Hamlonan Equaons of Moon (Chaper 8) Wha We Dd Las Tme Consruced Hamlonan formalsm H ( q, p, ) = q p L( q, q, ) H p = q H q = p H = L Equvalen o Lagrangan formalsm Smpler, bu
More informationAppendix to Online Clustering with Experts
A Appendx o Onlne Cluserng wh Expers Furher dscusson of expermens. Here we furher dscuss expermenal resuls repored n he paper. Ineresngly, we observe ha OCE (and n parcular Learn- ) racks he bes exper
More informationMechanics Physics 151
Mechancs Physcs 5 Lecure 9 Hamlonan Equaons of Moon (Chaper 8) Wha We Dd Las Tme Consruced Hamlonan formalsm Hqp (,,) = qp Lqq (,,) H p = q H q = p H L = Equvalen o Lagrangan formalsm Smpler, bu wce as
More informationDensity Matrix Description of NMR BCMB/CHEM 8190
Densy Marx Descrpon of NMR BCMBCHEM 89 Operaors n Marx Noaon If we say wh one bass se, properes vary only because of changes n he coeffcens weghng each bass se funcon x = h< Ix > - hs s how we calculae
More informationFI 3103 Quantum Physics
/9/4 FI 33 Quanum Physcs Aleander A. Iskandar Physcs of Magnesm and Phooncs Research Grou Insu Teknolog Bandung Basc Conces n Quanum Physcs Probably and Eecaon Value Hesenberg Uncerany Prncle Wave Funcon
More information[ ] 2. [ ]3 + (Δx i + Δx i 1 ) / 2. Δx i-1 Δx i Δx i+1. TPG4160 Reservoir Simulation 2018 Lecture note 3. page 1 of 5
TPG460 Reservor Smulaon 08 page of 5 DISCRETIZATIO OF THE FOW EQUATIOS As we already have seen, fne dfference appromaons of he paral dervaves appearng n he flow equaons may be obaned from Taylor seres
More informationTHEORETICAL AUTOCORRELATIONS. ) if often denoted by γ. Note that
THEORETICAL AUTOCORRELATIONS Cov( y, y ) E( y E( y))( y E( y)) ρ = = Var( y) E( y E( y)) =,, L ρ = and Cov( y, y ) s ofen denoed by whle Var( y ) f ofen denoed by γ. Noe ha γ = γ and ρ = ρ and because
More informationNotes on the stability of dynamic systems and the use of Eigen Values.
Noes on he sabl of dnamc ssems and he use of Egen Values. Source: Macro II course noes, Dr. Davd Bessler s Tme Seres course noes, zarads (999) Ineremporal Macroeconomcs chaper 4 & Techncal ppend, and Hamlon
More informationCS 268: Packet Scheduling
Pace Schedulng Decde when and wha pace o send on oupu ln - Usually mplemened a oupu nerface CS 68: Pace Schedulng flow Ion Soca March 9, 004 Classfer flow flow n Buffer managemen Scheduler soca@cs.bereley.edu
More informationDiscrete Markov Process. Introduction. Example: Balls and Urns. Stochastic Automaton. INTRODUCTION TO Machine Learning 3rd Edition
EHEM ALPAYDI he MI Press, 04 Lecure Sldes for IRODUCIO O Machne Learnng 3rd Edon alpaydn@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/ml3e Sldes from exboo resource page. Slghly eded and wh addonal examples
More informationRobustness Experiments with Two Variance Components
Naonal Insue of Sandards and Technology (NIST) Informaon Technology Laboraory (ITL) Sascal Engneerng Dvson (SED) Robusness Expermens wh Two Varance Componens by Ana Ivelsse Avlés avles@ns.gov Conference
More informationIntroduction ( Week 1-2) Course introduction A brief introduction to molecular biology A brief introduction to sequence comparison Part I: Algorithms
Course organzaon Inroducon Wee -2) Course nroducon A bref nroducon o molecular bology A bref nroducon o sequence comparson Par I: Algorhms for Sequence Analyss Wee 3-8) Chaper -3, Models and heores» Probably
More information1 Constant Real Rate C 1
Consan Real Rae. Real Rae of Inees Suppose you ae equally happy wh uns of he consumpon good oday o 5 uns of he consumpon good n peod s me. C 5 Tha means you ll be pepaed o gve up uns oday n eun fo 5 uns
More informationMechanics Physics 151
Mechancs Physcs 5 Lecure 0 Canoncal Transformaons (Chaper 9) Wha We Dd Las Tme Hamlon s Prncple n he Hamlonan formalsm Dervaon was smple δi δ Addonal end-pon consrans pq H( q, p, ) d 0 δ q ( ) δq ( ) δ
More informationLecture 2 M/G/1 queues. M/G/1-queue
Lecure M/G/ queues M/G/-queue Posson arrval process Arbrary servce me dsrbuon Sngle server To deermne he sae of he sysem a me, we mus now The number of cusomers n he sysems N() Tme ha he cusomer currenly
More informationAdvanced time-series analysis (University of Lund, Economic History Department)
Advanced me-seres analss (Unvers of Lund, Economc Hsor Dearmen) 3 Jan-3 Februar and 6-3 March Lecure 4 Economerc echnues for saonar seres : Unvarae sochasc models wh Box- Jenns mehodolog, smle forecasng
More informationHidden Markov Models
11-755 Machne Learnng for Sgnal Processng Hdden Markov Models Class 15. 12 Oc 2010 1 Admnsrva HW2 due Tuesday Is everyone on he projecs page? Where are your projec proposals? 2 Recap: Wha s an HMM Probablsc
More informationOn One Analytic Method of. Constructing Program Controls
Appled Mahemacal Scences, Vol. 9, 05, no. 8, 409-407 HIKARI Ld, www.m-hkar.com hp://dx.do.org/0.988/ams.05.54349 On One Analyc Mehod of Consrucng Program Conrols A. N. Kvko, S. V. Chsyakov and Yu. E. Balyna
More informationRL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1
RL Lecure 7: Eligibiliy Traces R. S. Suon and A. G. Baro: Reinforcemen Learning: An Inroducion 1 N-sep TD Predicion Idea: Look farher ino he fuure when you do TD backup (1, 2, 3,, n seps) R. S. Suon and
More informationV.Abramov - FURTHER ANALYSIS OF CONFIDENCE INTERVALS FOR LARGE CLIENT/SERVER COMPUTER NETWORKS
R&RATA # Vol.) 8, March FURTHER AALYSIS OF COFIDECE ITERVALS FOR LARGE CLIET/SERVER COMPUTER ETWORKS Vyacheslav Abramov School of Mahemacal Scences, Monash Unversy, Buldng 8, Level 4, Clayon Campus, Wellngon
More informationDisplacement, Velocity, and Acceleration. (WHERE and WHEN?)
Dsplacemen, Velocy, and Acceleraon (WHERE and WHEN?) Mah resources Append A n your book! Symbols and meanng Algebra Geomery (olumes, ec.) Trgonomery Append A Logarhms Remnder You wll do well n hs class
More informationGraduate Macroeconomics 2 Problem set 5. - Solutions
Graduae Macroeconomcs 2 Problem se. - Soluons Queson 1 To answer hs queson we need he frms frs order condons and he equaon ha deermnes he number of frms n equlbrum. The frms frs order condons are: F K
More informationJanuary Examinations 2012
Page of 5 EC79 January Examnaons No. of Pages: 5 No. of Quesons: 8 Subjec ECONOMICS (POSTGRADUATE) Tle of Paper EC79 QUANTITATIVE METHODS FOR BUSINESS AND FINANCE Tme Allowed Two Hours ( hours) Insrucons
More information. The geometric multiplicity is dim[ker( λi. number of linearly independent eigenvectors associated with this eigenvalue.
Lnear Algebra Lecure # Noes We connue wh he dscusson of egenvalues, egenvecors, and dagonalzably of marces We wan o know, n parcular wha condons wll assure ha a marx can be dagonalzed and wha he obsrucons
More informationReactive Methods to Solve the Berth AllocationProblem with Stochastic Arrival and Handling Times
Reacve Mehods o Solve he Berh AllocaonProblem wh Sochasc Arrval and Handlng Tmes Nsh Umang* Mchel Berlare* * TRANSP-OR, Ecole Polyechnque Fédérale de Lausanne Frs Workshop on Large Scale Opmzaon November
More informationPanel Data Regression Models
Panel Daa Regresson Models Wha s Panel Daa? () Mulple dmensoned Dmensons, e.g., cross-secon and me node-o-node (c) Pongsa Pornchawseskul, Faculy of Economcs, Chulalongkorn Unversy (c) Pongsa Pornchawseskul,
More informationPerformance Analysis for a Network having Standby Redundant Unit with Waiting in Repair
TECHNI Inernaonal Journal of Compung Scence Communcaon Technologes VOL.5 NO. July 22 (ISSN 974-3375 erformance nalyss for a Nework havng Sby edundan Un wh ang n epar Jendra Sngh 2 abns orwal 2 Deparmen
More informationWe are estimating the density of long distant migrant (LDM) birds in wetlands along Lake Michigan.
Ch 17 Random ffecs and Mxed Models 17. Random ffecs Models We are esmang he densy of long dsan mgran (LDM) brds n welands along Lake Mchgan. μ + = LDM per hecaren h weland ~ N(0, ) The varably of expeced
More informationJ i-1 i. J i i+1. Numerical integration of the diffusion equation (I) Finite difference method. Spatial Discretization. Internal nodes.
umercal negraon of he dffuson equaon (I) Fne dfference mehod. Spaal screaon. Inernal nodes. R L V For hermal conducon le s dscree he spaal doman no small fne spans, =,,: Balance of parcles for an nernal
More informationChapters 2 Kinematics. Position, Distance, Displacement
Chapers Knemacs Poson, Dsance, Dsplacemen Mechancs: Knemacs and Dynamcs. Knemacs deals wh moon, bu s no concerned wh he cause o moon. Dynamcs deals wh he relaonshp beween orce and moon. The word dsplacemen
More information. The geometric multiplicity is dim[ker( λi. A )], i.e. the number of linearly independent eigenvectors associated with this eigenvalue.
Mah E-b Lecure #0 Noes We connue wh he dscusson of egenvalues, egenvecors, and dagonalzably of marces We wan o know, n parcular wha condons wll assure ha a marx can be dagonalzed and wha he obsrucons are
More informationFall 2009 Social Sciences 7418 University of Wisconsin-Madison. Problem Set 2 Answers (4) (6) di = D (10)
Publc Affars 974 Menze D. Chnn Fall 2009 Socal Scences 7418 Unversy of Wsconsn-Madson Problem Se 2 Answers Due n lecure on Thursday, November 12. " Box n" your answers o he algebrac quesons. 1. Consder
More informationDynamic Team Decision Theory. EECS 558 Project Shrutivandana Sharma and David Shuman December 10, 2005
Dynamc Team Decson Theory EECS 558 Proec Shruvandana Sharma and Davd Shuman December 0, 005 Oulne Inroducon o Team Decson Theory Decomposon of he Dynamc Team Decson Problem Equvalence of Sac and Dynamc
More informationOnline Appendix for. Strategic safety stocks in supply chains with evolving forecasts
Onlne Appendx for Sraegc safey socs n supply chans wh evolvng forecass Tor Schoenmeyr Sephen C. Graves Opsolar, Inc. 332 Hunwood Avenue Hayward, CA 94544 A. P. Sloan School of Managemen Massachuses Insue
More informationGMM parameter estimation. Xiaoye Lu CMPS290c Final Project
GMM paraeer esaon Xaoye Lu M290c Fnal rojec GMM nroducon Gaussan ure Model obnaon of several gaussan coponens Noaon: For each Gaussan dsrbuon:, s he ean and covarance ar. A GMM h ures(coponens): p ( 2π
More informationInstance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification
Instance-Based earnng (a.k.a. memory-based learnng) Part I: Nearest Neghbor Classfcaton Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n
More information( t) Outline of program: BGC1: Survival and event history analysis Oslo, March-May Recapitulation. The additive regression model
BGC1: Survval and even hsory analyss Oslo, March-May 212 Monday May 7h and Tuesday May 8h The addve regresson model Ørnulf Borgan Deparmen of Mahemacs Unversy of Oslo Oulne of program: Recapulaon Counng
More information10. A.C CIRCUITS. Theoretically current grows to maximum value after infinite time. But practically it grows to maximum after 5τ. Decay of current :
. A. IUITS Synopss : GOWTH OF UNT IN IUIT : d. When swch S s closed a =; = d. A me, curren = e 3. The consan / has dmensons of me and s called he nducve me consan ( τ ) of he crcu. 4. = τ; =.63, n one
More informationLinear Response Theory: The connection between QFT and experiments
Phys540.nb 39 3 Lnear Response Theory: The connecon beween QFT and expermens 3.1. Basc conceps and deas Q: ow do we measure he conducvy of a meal? A: we frs nroduce a weak elecrc feld E, and hen measure
More informationMachine Learning Linear Regression
Machne Learnng Lnear Regresson Lesson 3 Lnear Regresson Bascs of Regresson Leas Squares esmaon Polynomal Regresson Bass funcons Regresson model Regularzed Regresson Sascal Regresson Mamum Lkelhood (ML)
More informationScattering at an Interface: Oblique Incidence
Course Insrucor Dr. Raymond C. Rumpf Offce: A 337 Phone: (915) 747 6958 E Mal: rcrumpf@uep.edu EE 4347 Appled Elecromagnecs Topc 3g Scaerng a an Inerface: Oblque Incdence Scaerng These Oblque noes may
More informationRELATIONSHIP BETWEEN VOLATILITY AND TRADING VOLUME: THE CASE OF HSI STOCK RETURNS DATA
RELATIONSHIP BETWEEN VOLATILITY AND TRADING VOLUME: THE CASE OF HSI STOCK RETURNS DATA Mchaela Chocholaá Unversy of Economcs Braslava, Slovaka Inroducon (1) one of he characersc feaures of sock reurns
More information2. SPATIALLY LAGGED DEPENDENT VARIABLES
2. SPATIALLY LAGGED DEPENDENT VARIABLES In hs chaper, we descrbe a sascal model ha ncorporaes spaal dependence explcly by addng a spaally lagged dependen varable y on he rgh-hand sde of he regresson equaon.
More informationThis document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.
Ths documen s downloaded from DR-NTU, Nanyang Technologcal Unversy Lbrary, Sngapore. Tle A smplfed verb machng algorhm for word paron n vsual speech processng( Acceped verson ) Auhor(s) Foo, Say We; Yong,
More informationNormal Random Variable and its discriminant functions
Noral Rando Varable and s dscrnan funcons Oulne Noral Rando Varable Properes Dscrnan funcons Why Noral Rando Varables? Analycally racable Works well when observaon coes for a corruped snle prooype 3 The
More informationComputing Relevance, Similarity: The Vector Space Model
Compung Relevance, Smlary: The Vecor Space Model Based on Larson and Hears s sldes a UC-Bereley hp://.sms.bereley.edu/courses/s0/f00/ aabase Managemen Sysems, R. Ramarshnan ocumen Vecors v ocumens are
More informationCubic Bezier Homotopy Function for Solving Exponential Equations
Penerb Journal of Advanced Research n Compung and Applcaons ISSN (onlne: 46-97 Vol. 4, No.. Pages -8, 6 omoopy Funcon for Solvng Eponenal Equaons S. S. Raml *,,. Mohamad Nor,a, N. S. Saharzan,b and M.
More information2.1 Constitutive Theory
Secon.. Consuve Theory.. Consuve Equaons Governng Equaons The equaons governng he behavour of maerals are (n he spaal form) dρ v & ρ + ρdv v = + ρ = Conservaon of Mass (..a) d x σ j dv dvσ + b = ρ v& +
More informationOrdinary Differential Equations in Neuroscience with Matlab examples. Aim 1- Gain understanding of how to set up and solve ODE s
Ordnary Dfferenal Equaons n Neuroscence wh Malab eamples. Am - Gan undersandng of how o se up and solve ODE s Am Undersand how o se up an solve a smple eample of he Hebb rule n D Our goal a end of class
More informationDual Approximate Dynamic Programming for Large Scale Hydro Valleys
Dual Approxmae Dynamc Programmng for Large Scale Hydro Valleys Perre Carpener and Jean-Phlppe Chanceler 1 ENSTA ParsTech and ENPC ParsTech CMM Workshop, January 2016 1 Jon work wh J.-C. Alas, suppored
More informationVolatility Interpolation
Volaly Inerpolaon Prelmnary Verson March 00 Jesper Andreasen and Bran Huge Danse Mares, Copenhagen wan.daddy@danseban.com brno@danseban.com Elecronc copy avalable a: hp://ssrn.com/absrac=69497 Inro Local
More informationAdvanced Macroeconomics II: Exchange economy
Advanced Macroeconomcs II: Exchange economy Krzyszof Makarsk 1 Smple deermnsc dynamc model. 1.1 Inroducon Inroducon Smple deermnsc dynamc model. Defnons of equlbrum: Arrow-Debreu Sequenal Recursve Equvalence
More informationReinforcement learning
Renforcement learnng Nathanel Daw Gatsby Computatonal Neuroscence Unt daw @ gatsby.ucl.ac.uk http://www.gatsby.ucl.ac.uk/~daw Mostly adapted from Andrew Moore s tutorals, copyrght 2002, 2004 by Andrew
More informationEpistemic Game Theory: Online Appendix
Epsemc Game Theory: Onlne Appendx Edde Dekel Lucano Pomao Marcano Snscalch July 18, 2014 Prelmnares Fx a fne ype srucure T I, S, T, β I and a probably µ S T. Le T µ I, S, T µ, βµ I be a ype srucure ha
More informationAttribute Reduction Algorithm Based on Discernibility Matrix with Algebraic Method GAO Jing1,a, Ma Hui1, Han Zhidong2,b
Inernaonal Indusral Informacs and Compuer Engneerng Conference (IIICEC 05) Arbue educon Algorhm Based on Dscernbly Marx wh Algebrac Mehod GAO Jng,a, Ma Hu, Han Zhdong,b Informaon School, Capal Unversy
More informationLecture Slides for INTRODUCTION TO. Machine Learning. ETHEM ALPAYDIN The MIT Press,
Lecure Sldes for INTRDUCTIN T Machne Learnng ETHEM ALAYDIN The MIT ress, 2004 alpaydn@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/2ml CHATER 3: Hdden Marov Models Inroducon Modelng dependences n npu; no
More information, t 1. Transitions - this one was easy, but in general the hardest part is choosing the which variables are state and control variables
Opmal Conrol Why Use I - verss calcls of varaons, opmal conrol More generaly More convenen wh consrans (e.g., can p consrans on he dervaves More nsghs no problem (a leas more apparen han hrogh calcls of
More informationLecture 2 L n i e n a e r a M od o e d l e s
Lecure Lnear Models Las lecure You have learned abou ha s machne learnng Supervsed learnng Unsupervsed learnng Renforcemen learnng You have seen an eample learnng problem and he general process ha one
More informationTime-interval analysis of β decay. V. Horvat and J. C. Hardy
Tme-nerval analyss of β decay V. Horva and J. C. Hardy Work on he even analyss of β decay [1] connued and resuled n he developmen of a novel mehod of bea-decay me-nerval analyss ha produces hghly accurae
More informationUNIVERSITAT AUTÒNOMA DE BARCELONA MARCH 2017 EXAMINATION
INTERNATIONAL TRADE T. J. KEHOE UNIVERSITAT AUTÒNOMA DE BARCELONA MARCH 27 EXAMINATION Please answer wo of he hree quesons. You can consul class noes, workng papers, and arcles whle you are workng on he
More informationBiol. 356 Lab 8. Mortality, Recruitment, and Migration Rates
Biol. 356 Lab 8. Moraliy, Recruimen, and Migraion Raes (modified from Cox, 00, General Ecology Lab Manual, McGraw Hill) Las week we esimaed populaion size hrough several mehods. One assumpion of all hese
More informationCS286.2 Lecture 14: Quantum de Finetti Theorems II
CS286.2 Lecure 14: Quanum de Fne Theorems II Scrbe: Mara Okounkova 1 Saemen of he heorem Recall he las saemen of he quanum de Fne heorem from he prevous lecure. Theorem 1 Quanum de Fne). Le ρ Dens C 2
More informationF-Tests and Analysis of Variance (ANOVA) in the Simple Linear Regression Model. 1. Introduction
ECOOMICS 35* -- OTE 9 ECO 35* -- OTE 9 F-Tess and Analyss of Varance (AOVA n he Smple Lnear Regresson Model Inroducon The smple lnear regresson model s gven by he followng populaon regresson equaon, or
More informationTHE PREDICTION OF COMPETITIVE ENVIRONMENT IN BUSINESS
THE PREICTION OF COMPETITIVE ENVIRONMENT IN BUSINESS INTROUCTION The wo dmensonal paral dfferenal equaons of second order can be used for he smulaon of compeve envronmen n busness The arcle presens he
More informationOMXS30 Balance 20% Index Rules
OMX30 Balance 0% ndex Rules Verson as of 30 March 009 Copyrgh 008, The NADAQ OMX Group, nc. All rghs reserved. NADAQ OMX, The NADAQ ock Marke and NADAQ are regsered servce/rademarks of The NADAQ OMX Group,
More informationSingle-loop System Reliability-Based Design & Topology Optimization (SRBDO/SRBTO): A Matrix-based System Reliability (MSR) Method
10 h US Naonal Congress on Compuaonal Mechancs Columbus, Oho 16-19, 2009 Sngle-loop Sysem Relably-Based Desgn & Topology Opmzaon (SRBDO/SRBTO): A Marx-based Sysem Relably (MSR) Mehod Tam Nguyen, Junho
More informationApproximate Analytic Solution of (2+1) - Dimensional Zakharov-Kuznetsov(Zk) Equations Using Homotopy
Arcle Inernaonal Journal of Modern Mahemacal Scences, 4, (): - Inernaonal Journal of Modern Mahemacal Scences Journal homepage: www.modernscenfcpress.com/journals/jmms.aspx ISSN: 66-86X Florda, USA Approxmae
More informationA Cell Decomposition Approach to Online Evasive Path Planning and the Video Game Ms. Pac-Man
Cell Decomoson roach o Onlne Evasve Pah Plannng and he Vdeo ame Ms. Pac-Man reg Foderaro Vram Raju Slva Ferrar Laboraory for Inellgen Sysems and Conrols LISC Dearmen of Mechancal Engneerng and Maerals
More informationA Deterministic Algorithm for Summarizing Asynchronous Streams over a Sliding Window
A Deermnsc Algorhm for Summarzng Asynchronous Sreams over a Sldng ndow Cosas Busch Rensselaer Polyechnc Insue Srkana Trhapura Iowa Sae Unversy Oulne of Talk Inroducon Algorhm Analyss Tme C Daa sream: 3
More informationAppendix H: Rarefaction and extrapolation of Hill numbers for incidence data
Anne Chao Ncholas J Goell C seh lzabeh L ander K Ma Rober K Colwell and Aaron M llson 03 Rarefacon and erapolaon wh ll numbers: a framewor for samplng and esmaon n speces dversy sudes cology Monographs
More informationAn introduction to Support Vector Machine
An nroducon o Suppor Vecor Machne 報告者 : 黃立德 References: Smon Haykn, "Neural Neworks: a comprehensve foundaon, second edon, 999, Chaper 2,6 Nello Chrsann, John Shawe-Tayer, An Inroducon o Suppor Vecor Machnes,
More informationLecture 9: Dynamic Properties
Shor Course on Molecular Dynamcs Smulaon Lecure 9: Dynamc Properes Professor A. Marn Purdue Unversy Hgh Level Course Oulne 1. MD Bascs. Poenal Energy Funcons 3. Inegraon Algorhms 4. Temperaure Conrol 5.
More informationClustering (Bishop ch 9)
Cluserng (Bshop ch 9) Reference: Daa Mnng by Margare Dunham (a slde source) 1 Cluserng Cluserng s unsupervsed learnng, here are no class labels Wan o fnd groups of smlar nsances Ofen use a dsance measure
More informationLi An-Ping. Beijing , P.R.China
A New Type of Cpher: DICING_csb L An-Png Bejng 100085, P.R.Chna apl0001@sna.com Absrac: In hs paper, we wll propose a new ype of cpher named DICING_csb, whch s derved from our prevous sream cpher DICING.
More informationHow about the more general "linear" scalar functions of scalars (i.e., a 1st degree polynomial of the following form with a constant term )?
lmcd Lnear ransformaon of a vecor he deas presened here are que general hey go beyond he radonal mar-vecor ype seen n lnear algebra Furhermore, hey do no deal wh bass and are equally vald for any se of
More informationChapter Lagrangian Interpolation
Chaper 5.4 agrangan Inerpolaon Afer readng hs chaper you should be able o:. dere agrangan mehod of nerpolaon. sole problems usng agrangan mehod of nerpolaon and. use agrangan nerpolans o fnd deraes and
More informationBandlimited channel. Intersymbol interference (ISI) This non-ideal communication channel is also called dispersive channel
Inersymol nererence ISI ISI s a sgnal-dependen orm o nererence ha arses ecause o devaons n he requency response o a channel rom he deal channel. Example: Bandlmed channel Tme Doman Bandlmed channel Frequency
More informationDEEP UNFOLDING FOR MULTICHANNEL SOURCE SEPARATION SUPPLEMENTARY MATERIAL
DEEP UNFOLDING FOR MULTICHANNEL SOURCE SEPARATION SUPPLEMENTARY MATERIAL Sco Wsdom, John Hershey 2, Jonahan Le Roux 2, and Shnj Waanabe 2 Deparmen o Elecrcal Engneerng, Unversy o Washngon, Seale, WA, USA
More information