Temporal Abstraction in Temporal-difference Networks

Size: px
Start display at page:

Download "Temporal Abstraction in Temporal-difference Networks"

Transcription

1 Temporal Absracion in Temporal-difference Neworks Richard S. Suon, Eddie J. Rafols, Anna Koop Deparmen of Compuing Science Universiy of Albera Edmonon, AB, Canada T6G 2E8 Absrac We presen a generalizaion of emporal-difference neworks o include emporally absrac opions on he links of he quesion nework. Temporal-difference (TD) neworks have been proposed as a way of represening and learning a wide variey of predicions abou he ineracion beween an agen and is environmen. These predicions are composiional in ha heir arges are defined in erms of oher predicions, and subjuncive in ha ha hey are abou wha would happen if an acion or sequence of acions were aken. In convenional TD neworks, he iner-relaed predicions are a successive ime seps and coningen on a single acion; here we generalize hem o accommodae exended ime inervals and coningency on whole ways of behaving. Our generalizaion is based on he opions framework for emporal absracion. The primary conribuion of his paper is o inroduce a new algorihm for inra-opion learning in TD neworks wih funcion approximaion and eligibiliy races. We presen empirical examples of our algorihm s effeciveness and of he greaer represenaional expressiveness of emporallyabsrac TD neworks. The primary disinguishing feaure of emporal-difference (TD) neworks (Suon & Tanner, 2005) is ha hey permi a general composiional specificaion of he goals of learning. The goals of learning are hough of as predicive quesions being asked by he agen in he learning problem, such as Wha will I see if I sep forward and look righ? or If I open he fridge, will I see a bole of beer? Seeing a bole of beer is of course a complicaed percepual ac. I migh be hough of as obaining a se of predicions abou wha would happen if cerain reaching and grasping acions were aken, abou wha would happen if he bole were opened and urned upside down, and of wha he bole would look like if viewed from various angles. To predic seeing a bole of beer is hus o make a predicion abou a se of oher predicions. The arge for he overall predicion is a composiion in he mahemaical sense of he firs predicion wih each of he oher predicions. TD neworks are he firs framework for represening he goals of predicive learning in a composiional, machine-accessible form. Each node of a TD nework represens an individual quesion somehing o be prediced and has associaed wih i a value represening an answer o he quesion a predicion of ha somehing. The quesions are represened by a se of direced links beween nodes. If node 1 is linked o node 2, hen node 1 rep-

2 resens a quesion incorporaing node 2 s quesion; is value is a predicion abou node 2 s predicion. Higher-level predicions can be composed in several ways from lower ones, producing a powerful, srucured represenaion language for he arges of learning. The composiional srucure is no jus in a human designer s head; i is expressed in he links and hus is accessible o he agen and is learning algorihm. The nework of hese links is referred o as he quesion nework. An enirely separae se of direced links beween he nodes is used o compue he values (predicions, answers) associaed wih each node. These links collecively are referred o as he answer nework. The compuaion in he answer nework is composiional in a convenional way node values are compued from oher node values. The essenial insigh of TD neworks is ha he noion of composiionaliy should apply o quesions as well as o answers. A secondary disinguishing feaure of TD neworks is ha he predicions (node values) a each momen in ime can be used as a represenaion of he sae of he world a ha ime. In his way hey are an insance of he idea of predicive sae represenaions (PSRs) inroduced by Liman, Suon and Singh (2002), Jaeger (2000), and Rives and Schapire (1987). Represening a sae by is predicions is a poenially powerful sraegy for sae absracion (Rafols e al., 2005). We noe ha he quesions used in all previous work wih PSRs are defined in erms of concree acions and observaions, no oher predicions. They are no composiional in he sense ha TD-nework quesions are. The quesions we have discussed so far are subjuncive, meaning ha hey are condiional on a cerain way of behaving. We predic wha we would see if we were o sep forward and look righ, or if we were o open he fridge. The quesions in convenional TD neworks are subjuncive, bu hey are condiional only on primiive acions or open-loop sequences of primiive acions (as are convenional PSRs). I is naural o generalize his, as we have in he informal examples above, o quesions ha are condiional on closed-loop emporally exended ways of behaving. For example, opening he fridge is a complex, high-level acion. The arm mus be lifed o he door, he hand shaped for grasping he handle, ec. To ask quesions like if I were o go o he coffee room, would I see John? would require subsanial emporal absracion in addiion o sae absracion. The opions framework (Suon, Precup & Singh, 1999) is a sraighforward way of alking abou emporally exended ways of behaving and abou predicions of heir oucomes. In his paper we exend he opions framework so ha i can be applied o TD neworks. Significan exensions of he original opions framework are needed. Novel feaures of our opion-exended TD neworks are ha hey 1) predic componens of opion oucomes raher han full oucome probabiliy disribuions, 2) learn according o he firs inra-opion mehod o use eligibiliy races (see Suon & Baro, 1998), and 3) include he possibiliy of opions whose policies are indifferen o which of several acions are seleced. 1 The opions framework In his secion we presen he essenial elemens of he opions framework (Suon, Precup & Singh, 1999) ha we will need for our exension of TD neworks. In his framework, an agen and an environmen inerac a discree ime seps = 1, 2, 3... In each sae s S, he agen selecs an acion a A, deermining he nex sae s An acion is a way of behaving for one ime sep; he opions framework les us alk abou emporally exended ways of behaving. An individual opion consiss of hree pars. The firs is he iniiaion se, I S, he subse of saes in which he opion can be sared. The second componen of an opion is is policy, π : S A [0, 1], specifying how he agen behaves when 1 Alhough he opions framework includes rewards, we omi hem here because we are concerned only wih predicion, no conrol.

3 following he opion. Finally, a erminaion funcion, β : S A [0, 1], specifies how he opion ends: β(s) denoes he probabiliy of erminaing when in sae s. The opion is hus compleely and formally defined by he 3-uple (I, π, β). 2 Convenional TD neworks In his secion we briefly presen he deails of he srucure and he learning algorihm comprising TD neworks as inroduced by Suon and Tanner (2005). TD neworks address a predicion problem in which he agen may no have direc access o he sae of he environmen. Insead, a each ime sep he agen receives an observaion o O dependen on he sae. The experience sream hus consiss of a sequence of alernaing acions and observaions, o 1, a 1, o 2, a 2, o 3. The TD nework consiss of a se of nodes, each represening a single scalar predicion, inerlinked by he quesion and answer neworks as suggesed previously. For a nework of n nodes, he vecor of all predicions a ime sep is denoed y = (y 1,..., y n ) T. The predicions are esimaes of he expeced value of some scalar quaniy, ypically of a bi, in which case hey can be inerpreed as esimaes of probabiliies. The predicions are updaed a each ime sep according o a vecor-valued funcion u wih modifiable parameer W, which is ofen aken o be of a linear form: y = u(y 1, a 1, o, W ) = σ(w x ), (1) where x R m is an m-vecor of feaures creaed from (y 1, a 1, o ), W is an n m marix (whose elemens are someimes referred o as weighs), and σ is he n-vecor form of eiher he ideniy funcion or he S-shaped logisic funcion σ(s) = 1 1+e. The s feaure vecor is an arbirary vecor-valued funcion of y 1, a 1, and o. For example, in he simples case he feaure vecor is a uni basis vecor wih he locaion of he one communicaing he curren sae. In a parially observable environmen, he feaure vecor may be a combinaion of he agen s acion, observaions, and predicions from he previous ime sep. The overall updae u defines he answer nework. The quesion nework consiss of a se of arge funcions, z i : O R n R, and condiion funcions, c i : A R n [0, 1] n. We define z i = z i (o +1, ỹ +1 ) as he arge for predicion y. i 2 Similarly, we define c i = c i (a, y ) as he condiion a ime. The learning algorihm for each componen w ij of W can hen be wrien w ij +1 = wij + α ( z i y i ) c i y i w ij, (2) where α is a posiive sep-size parameer. Noe ha he arges here are funcions of he observaion and predicions exacly one ime sep laer, and ha he condiions are funcions of a single primiive acion. This is wha makes his algorihm suiable only for learning abou one-sep TD relaionships. By chaining ogeher muliple nodes, Suon and Tanner (2005) used i o predic k seps ahead, for various paricular values of k, and o predic he oucome of specific acion sequences (as in PSRs, e.g., Liman e al., 2002; Singh e al., 2004). Now we consider he exension o emporally absrac acions. 3 Opion-exended TD neworks In his secion we presen our inra-opion learning algorihm for TD neworks wih opions and eligibiliy races. As suggesed earlier, each node s ougoing link in he quesion 2 The quaniy ỹ is almos he same as y, and we encourage he reader o hink of hem as idenical here. The difference is ha ỹ is calculaed by weighs ha are one sep ou of dae as compared o y, i.e., ỹ = u(y 1, a 1, o, W 1) (cf. equaion 1).

4 nework will now correspond o an opion applying over possibly many seps. The policy of he ih node s opion corresponds o he condiion funcion c i, which we hink of as a recognizer for he opion. I inspecs each acion aken o assess wheher he opion is being followed: c i = 1 if he agen is acing consisenly wih he opion policy and c i = 0 oherwise (inermediae values are also possible). When an agen ceases o ac consisenly wih he opion policy, we say ha he opion has diverged. The possibiliy of recognizing more han one acion as consisen wih he opion is a significan generalizaion of he original idea of opions. If no acions are recognized as accepable in a sae, hen he opion canno be followed and hus canno be iniiaed. Here we ake he se of saes wih a leas one recognized acion o be he iniiaion se of he opion. The opion-erminaion funcion β generalizes naurally o TD neworks. Each node i is given a corresponding erminaion funcion, β i : O R n [0, 1], where β i = β i (o +1, y ) is he probabiliy of erminaing a ime. 3 β i = 1 indicaes ha he opion has erminaed a ime ; β i = 0 indicaes ha i has no, and inermediae values of β correspond o sof or sochasic erminaion condiions. If an opion erminaes, hen z i acs as he arge, bu if he opion is ongoing wihou erminaion, hen he node s own nex value, ỹ i +1, should be he arge. The erminaion funcion specifies which of he wo arges (or mixure of he wo arges) is used o produce a form of TD error for each node i: δ i = β i z i + (1 β i )ỹ i +1 y i. (3) Our opion-exended algorihm incorporaes eligibiliy races (see Suon & Baro, 1998) as shor-erm memory variables organized in an n m marix E, paralleling he weigh marix. The races are a record of he effec ha each weigh could have had on each node s predicion during he ime he agen has been acing consisenly wih he node s opion. The componens e ij of he eligibiliy marix are updaed by e ij = c i [ λe ij 1 (1 βi ) + yi w ij ], (4) where 0 λ 1 is he race-decay parameer familiar from he TD(λ) learning algorihm. Because of he c i facor, all of a node s races will be immediaely rese o zero whenever he agen deviaes from he node s opion s policy. If he agen follows he policy and he opion does no erminae, hen he race decays by λ and incremens by he gradien in he way ypical of eligibiliy races. If he policy is followed and he opion does erminae, hen he race will be rese o zero on he immediaely following ime sep, and a new race will sar building. Finally, our algorihm updaes he weighs on each ime sep by 4 Fully observable experimen w ij +1 = wij + α δ i e ij. (5) This experimen was designed o es he correcness of he algorihm in a simple gridworld where he environmenal sae is observable. We applied an opions-exended TD nework o he problem of learning o predic observaions from ineracion wih he gridworld environmen shown on he lef in Figure 1. Empy squares indicae spaces where he agen can move freely, and colored squares (shown shaded in he figure) indicae walls. The agen is egocenric. A each ime sep he agen receives from he environmen six bis represening he color i is facing (red, green, blue, orange, yellow, or whie). In his firs experimen we also provided = 144 oher bis direcly indicaing he complee sae of he environmen (square and orienaion). 3 The fac ha he opion depends only on he curren predicions, acion, and observaion means ha we are considering only Markov opions.

5 Figure 1: The es world (lef) and he quesion nework (righ) used in he experimens. The riangle in he world indicaes he locaion and orienaion of he agen. The walls are labeled R, O, Y, G, and B represening he colors red, orange, yellow, green and blue. Noe ha he lef wall is mosly blue bu parly green. The righ diagram shows in full he porion of he quesion nework corresponding o he red bi. This srucure is repeaed, bu no shown, for he oher four (non-whie) colors. L, R, and F are primiive acions, and Forward and Wander are opions. There are hree possible acions: A ={F, R, L}. Acions were seleced according o a fixed sochasic policy independen of he sae. The probabiliy of he F, L, and R acions were 0.5, 0.25, and 0.25 respecively. L and R cause he agen o roae 90 degrees o he lef or righ. F causes he agen o move ahead one square wih probabiliy 1 p and o say in he same square wih probabiliy p. The probabiliy p is called he slipping probabiliy. If he forward movemen would cause he agen o move ino a wall, hen he agen does no move. In his experimen, we used p = 0, p = 0.1, and p = 0.5. In addiion o hese primiive acions, we provided wo emporally absrac opions, Forward and Wander. The Forward opion akes he acion F in every sae and erminaes when he agen senses a wall (color) in fron of i. The policy of he Wander opion is he same as ha acually followed by he agen. Wander erminaes wih probabiliy 1 when a wall is sensed, and sponaneously wih probabiliy 0.5 oherwise. We used he quesion nework shown on he righ in Figure 1. The predicions of nodes 1, 2, and 3 are esimaes of he probabiliy ha he red bi would be observed if he corresponding primiive acion were aken. Node 4 is a predicion of wheher he agen will see he red bi upon erminaion of he Wander opion if i were aken. Node 5 predics he probabiliy of observing he red bi given ha he Forward opion is followed unil erminaion. Nodes 6 and 7 represen predicions of he oucome of a primiive acion followed by he Forward opion. Nodes 8 and 9 ake his one sep furher: hey represen predicions of he red bi if he Forward opion were followed o erminaion, hen a primiive acion were aken, and hen he Forward opion were followed again o erminaion. We applied our algorihm o learn he parameer W of he answer nework for his quesion nework. The sep-size parameer α was 1.0, and he race-decay parameer λ was 0.9. The iniial W 0, E 0, and y 0 were all 0. Each run began wih he agen in he sae indicaed in Figure 1 (lef). In his experimen σ( ) was he ideniy funcion. For each value of p, we ran 50 runs of 20,000 ime seps. On each ime sep, he roo-meansquared (RMS) error in each node s predicion was compued and hen averaged over all he nodes. The nodes corresponding o he Wander opion were no included in he average because of he difficuly of calculaing heir correc predicions. This average was hen

6 0.4 RMS Error Fully Observable 0.4 RMS Error Parially Observable p = 0.5 p = 0.1 p = Seps Seps Figure 2: Learning curves in he fully-observable experimen for each slippage probabiliy (lef) and in he parially-observable experimen (righ). iself averaged over he 50 runs and bins of 1,000 ime seps o produce he learning curves shown on he lef in Figure 2. For all slippage probabiliies, he error in all predicions fell almos o zero. Afer approximaely 12,000 rials, he agen made almos perfec predicions in all cases. No surprisingly, learning was slower a he higher slippage probabiliies. These resuls show ha our augmened TD nework is able o make a complee emporally-absrac model of his world. 5 Parially observable experimen In our second experimen, only he six color observaion bis were available o he agen. This experimen provides a more challenging es of our algorihm. To model he environmen well, he TD nework mus consruc a represenaion of sae from very sparse informaion. In fac, compleely accurae predicion is no possible in his problem wih our quesion nework. In his experimen he inpu vecor consised of hree groups of 46 componens each, 138 in oal. If he acion was R, he firs 46 componens were se o he 40 node values and he six observaion bis, and he oher componens were 0. If he acion was L, he nex group of 46 componens was filled in in he same way, and he firs and hird groups were zero. If he acion was F, he hird group was filled. This echnique enables he answer nework as funcion approximaor o represen a wider class of funcions in a linear form han would oherwise be possible. In his experimen, σ( ) was he S-shaped logisic funcion. The slippage probabiliy was p = 0.1. As our performance measure we used he RMS error, as in he firs experimen, excep ha he predicions for he primiive acions (nodes 1-3) were no included. These predicions can never become compleely accurae because he agen can ell in deail where i is locaed in he open space. As before, we averaged RMS error over 50 runs and 1,000 ime sep bins, o produce he learning curve shown on he righ in Figure 2. As before, he RMS error approached zero. Node 5 in Figure 1 holds he predicion of red if he agen were o march forward o he wall ahead of i. Corresponding nodes in he oher subneworks hold he predicions of he oher colors upon Forward. To make hese predicions accuraely, he agen mus keep rack of which wall i is facing, even if i is many seps away from i. I has o learn a sor of compass ha i can keep updaed as i urns in he middle of he space. Figure 3 is a demonsraion of he compass learned afer a represenaive run of 200,000 ime seps. A he end of he run, he agen was driven manually o he sae shown in he firs row (relaive

7 ime index = 1). On seps 1-25 he agen was spun clockwise in place. The hird column shows he predicion for node 5 in each porion of he quesion nework. Tha is, he predicions shown are for each color-observaion bi a erminaion of he Forward opion. A = 1, he agen is facing he orange wall and i predics ha he Forward opion would resul in seeing he orange bi and none oher. Over seps 2-5 we see ha he predicions are mainained accuraely as he agen spins despie he fac ha is observaion bis remain he same. Even afer spinning for 25 seps he agen knows exacly which way i is facing. While spinning, he agen correcly never predics seeing he green bi (afer Forward), bu if i is driven up and urned, as in he las row of he figure, he green bi is accuraely prediced. The fourh column shows he predicion for node 8 in each porion of he quesion nework. Recall ha hese nodes correspond o he sequence Forward, L, Forward. A ime = 1, he agen accuraely predics ha Forward will bring i o orange (hird column) and also predics ha Forward, L, Forward will bring i o green. The predicions made for node 8 a each subsequen sep of he sequence are also correc. These resuls show ha he agen is able o accuraely mainain is long erm predicions wihou direcly encounering sensory verificaion. How much larger would he TD nework have o be o handle a 100x100 gridworld? The answer is no a all. The same quesion nework applies o any size problem. If he layou of he colored walls remain he same, hen even he answer nework ransfers across worlds of widely varying sizes. In oher experimens, raining on successively larger problems, we have shown ha he same TD nework as used here can learn o make all he long-erm predicions correcly on a 100x100 version of he 6x6 gridworld used here s y 5 y 8 1 Figure 3: An illusraion of par of wha he agen learns in he parially observable environmen. The second column is a sequence of saes wih (relaive) ime index as given by he firs column. The sequence was generaed by conrolling he agen manually. On seps 1-25 he agen was spun clockwise in place, and he rajecory afer ha is shown by he line in he las sae diagram. The hird and fourh columns show he values of he nodes corresponding o 5 and 8 in Figure 1, one for each color-observaion bi.

8 6 Conclusion Our experimens show ha opion-exended TD neworks can learn effecively. They can learn facs abou heir environmens ha are no represenable in convenional TD neworks or in any oher mehod for learning models of he world. One concern is ha our inra-opion learning algorihm is an off-policy learning mehod incorporaing funcion approximaion and boosrapping (learning from predicions). The combinaion of hese hree is known o produce convergence problems for some mehods (see Suon & Baro, 1998), and hey may arise here. A sound soluion may require modificaions o incorporae imporance sampling (see Precup, Suon & Dasgupa, 2001). In his paper we have considered only inra-opion eligibiliy races races exending over he ime span wihin an opion bu no persising across opions. Tanner and Suon (2005) have proposed a mehod for iner-opion races ha could perhaps be combined wih our inra-opion races. The primary conribuion of his paper is he inroducion of a new learning algorihm for TD neworks ha incorporaes opions and eligibiliy races. Our experimens are small and do lile more han exercise he learning algorihm, showing ha i does no break immediaely. More significan is he greaer represenaional power of opion-exended TD neworks. Opions are a general framework for emporal absracion, predicive sae represenaions are a promising sraegy for sae absracion, and TD neworks are able o represen composiional quesions. The combinaion of hese hree is poenially very powerful and worhy of furher sudy. Acknowledgmens The auhors graefully acknowledge he ideas and encouragemen hey have received in his work from Mark Ring, Brian Tanner, Sainder Singh, Doina Precup, and all he members of he rlai.ne group. References Jaeger, H. (2000). Observable operaor models for discree sochasic ime series. Neural Compuaion, 12(6): MIT Press. Liman, M., Suon, R. S., & Singh, S. (2002). Predicive represenaions of sae. In T. G. Dieerich, S. Becker and Z. Ghahramani (eds.), Advances In Neural Informaion Processing Sysems 14, pp MIT Press. Precup, D., Suon, R. S., & Dasgupa, S. (2001). Off-policy emporal-difference learning wih funcion approximaion. In C. E. Brodley, A. P. Danyluk (eds.), Proceedings of he Eigheenh Inernaional Conference on Machine Learning, pp San Francisco, CA: Morgan Kaufmann. Rafols, E. J., Ring, M., Suon, R.S., & Tanner, B. (2005). Using predicive represenaions o improve generalizaion in reinforcemen learning. To appear in Proceedings of he Nineeenh Inernaional Join Conference on Arificial Inelligence. Rives, R. L., & Schapire, R. E. (1987). Diversiy-based inference of finie auomaa. In Proceedings of he Tweny Eighh Annual Symposium on Foundaions of Compuer Science, (pp ). IEEE Compuer Sociey. Singh, S., James, M. R., & Rudary, M. R. (2004). Predicive sae represenaions: A new heory for modeling dynamical sysems. In Uncerainy in Arificial Inelligence: Proceedings of he Twenieh Conference in Uncerainy in Arificial Inelligence, (pp ). AUAI Press. Suon, R. S., & Baro, A. G. (1998). Reinforcemen learning: An inroducion. Cambridge, MA: MIT Press. Suon, R. S., Precup, D., Singh, S. (1999). Beween MDPs and semi-mdps: A framework for emporal absracion in reinforcemen learning. Arificial Inelligence, 112, pp Suon, R. S., & Tanner, B. (2005). Temporal-difference neworks. To appear in Neural Informaion Processing Sysems Conference 17. Tanner, B., Suon, R. S. (2005) Temporal-difference neworks wih hisory. To appear in Proceedings of he Nineeenh Inernaional Join Conference on Arificial Inelligence.

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1 RL Lecure 7: Eligibiliy Traces R. S. Suon and A. G. Baro: Reinforcemen Learning: An Inroducion 1 N-sep TD Predicion Idea: Look farher ino he fuure when you do TD backup (1, 2, 3,, n seps) R. S. Suon and

More information

Vehicle Arrival Models : Headway

Vehicle Arrival Models : Headway Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where

More information

1 Review of Zero-Sum Games

1 Review of Zero-Sum Games COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any

More information

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB Elecronic Companion EC.1. Proofs of Technical Lemmas and Theorems LEMMA 1. Le C(RB) be he oal cos incurred by he RB policy. Then we have, T L E[C(RB)] 3 E[Z RB ]. (EC.1) Proof of Lemma 1. Using he marginal

More information

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes Represening Periodic Funcions by Fourier Series 3. Inroducion In his Secion we show how a periodic funcion can be expressed as a series of sines and cosines. We begin by obaining some sandard inegrals

More information

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still. Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in

More information

Overview. COMP14112: Artificial Intelligence Fundamentals. Lecture 0 Very Brief Overview. Structure of this course

Overview. COMP14112: Artificial Intelligence Fundamentals. Lecture 0 Very Brief Overview. Structure of this course OMP: Arificial Inelligence Fundamenals Lecure 0 Very Brief Overview Lecurer: Email: Xiao-Jun Zeng x.zeng@mancheser.ac.uk Overview This course will focus mainly on probabilisic mehods in AI We shall presen

More information

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED 0.1 MAXIMUM LIKELIHOOD ESTIMATIO EXPLAIED Maximum likelihood esimaion is a bes-fi saisical mehod for he esimaion of he values of he parameers of a sysem, based on a se of observaions of a random variable

More information

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 175 CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 10.1 INTRODUCTION Amongs he research work performed, he bes resuls of experimenal work are validaed wih Arificial Neural Nework. From he

More information

STATE-SPACE MODELLING. A mass balance across the tank gives:

STATE-SPACE MODELLING. A mass balance across the tank gives: B. Lennox and N.F. Thornhill, 9, Sae Space Modelling, IChemE Process Managemen and Conrol Subjec Group Newsleer STE-SPACE MODELLING Inroducion: Over he pas decade or so here has been an ever increasing

More information

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing Applicaion of a Sochasic-Fuzzy Approach o Modeling Opimal Discree Time Dynamical Sysems by Using Large Scale Daa Processing AA WALASZE-BABISZEWSA Deparmen of Compuer Engineering Opole Universiy of Technology

More information

10. State Space Methods

10. State Space Methods . Sae Space Mehods. Inroducion Sae space modelling was briefly inroduced in chaper. Here more coverage is provided of sae space mehods before some of heir uses in conrol sysem design are covered in he

More information

Lecture Notes 2. The Hilbert Space Approach to Time Series

Lecture Notes 2. The Hilbert Space Approach to Time Series Time Series Seven N. Durlauf Universiy of Wisconsin. Basic ideas Lecure Noes. The Hilber Space Approach o Time Series The Hilber space framework provides a very powerful language for discussing he relaionship

More information

Notes on Kalman Filtering

Notes on Kalman Filtering Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren

More information

Chapter 7: Solving Trig Equations

Chapter 7: Solving Trig Equations Haberman MTH Secion I: The Trigonomeric Funcions Chaper 7: Solving Trig Equaions Le s sar by solving a couple of equaions ha involve he sine funcion EXAMPLE a: Solve he equaion sin( ) The inverse funcions

More information

Designing Information Devices and Systems I Spring 2019 Lecture Notes Note 17

Designing Information Devices and Systems I Spring 2019 Lecture Notes Note 17 EES 16A Designing Informaion Devices and Sysems I Spring 019 Lecure Noes Noe 17 17.1 apaciive ouchscreen In he las noe, we saw ha a capacior consiss of wo pieces on conducive maerial separaed by a nonconducive

More information

Learning to Take Concurrent Actions

Learning to Take Concurrent Actions Learning o Take Concurren Acions Khashayar Rohanimanesh Deparmen of Compuer Science Universiy of Massachuses Amhers, MA 0003 khash@cs.umass.edu Sridhar Mahadevan Deparmen of Compuer Science Universiy of

More information

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance

More information

3.1 More on model selection

3.1 More on model selection 3. More on Model selecion 3. Comparing models AIC, BIC, Adjused R squared. 3. Over Fiing problem. 3.3 Sample spliing. 3. More on model selecion crieria Ofen afer model fiing you are lef wih a handful of

More information

Explaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015

Explaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015 Explaining Toal Facor Produciviy Ulrich Kohli Universiy of Geneva December 2015 Needed: A Theory of Toal Facor Produciviy Edward C. Presco (1998) 2 1. Inroducion Toal Facor Produciviy (TFP) has become

More information

CSE/NB 528 Lecture 14: From Supervised to Reinforcement Learning (Chapter 9) R. Rao, 528: Lecture 14

CSE/NB 528 Lecture 14: From Supervised to Reinforcement Learning (Chapter 9) R. Rao, 528: Lecture 14 CSE/NB 58 Lecure 14: From Supervised o Reinforcemen Learning Chaper 9 1 Recall from las ime: Sigmoid Neworks Oupu v T g w u g wiui w Inpu nodes u = u 1 u u 3 T i Sigmoid oupu funcion: 1 g a 1 a e 1 ga

More information

ODEs II, Lecture 1: Homogeneous Linear Systems - I. Mike Raugh 1. March 8, 2004

ODEs II, Lecture 1: Homogeneous Linear Systems - I. Mike Raugh 1. March 8, 2004 ODEs II, Lecure : Homogeneous Linear Sysems - I Mike Raugh March 8, 4 Inroducion. In he firs lecure we discussed a sysem of linear ODEs for modeling he excreion of lead from he human body, saw how o ransform

More information

CSE/NB 528 Lecture 14: Reinforcement Learning (Chapter 9)

CSE/NB 528 Lecture 14: Reinforcement Learning (Chapter 9) CSE/NB 528 Lecure 14: Reinforcemen Learning Chaper 9 Image from hp://clasdean.la.asu.edu/news/images/ubep2001/neuron3.jpg Lecure figures are from Dayan & Abbo s book hp://people.brandeis.edu/~abbo/book/index.hml

More information

Some Basic Information about M-S-D Systems

Some Basic Information about M-S-D Systems Some Basic Informaion abou M-S-D Sysems 1 Inroducion We wan o give some summary of he facs concerning unforced (homogeneous) and forced (non-homogeneous) models for linear oscillaors governed by second-order,

More information

Finish reading Chapter 2 of Spivak, rereading earlier sections as necessary. handout and fill in some missing details!

Finish reading Chapter 2 of Spivak, rereading earlier sections as necessary. handout and fill in some missing details! MAT 257, Handou 6: Ocober 7-2, 20. I. Assignmen. Finish reading Chaper 2 of Spiva, rereading earlier secions as necessary. handou and fill in some missing deails! II. Higher derivaives. Also, read his

More information

Online Convex Optimization Example And Follow-The-Leader

Online Convex Optimization Example And Follow-The-Leader CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion

More information

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Article from. Predictive Analytics and Futurism. July 2016 Issue 13 Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning

More information

Christos Papadimitriou & Luca Trevisan November 22, 2016

Christos Papadimitriou & Luca Trevisan November 22, 2016 U.C. Bereley CS170: Algorihms Handou LN-11-22 Chrisos Papadimiriou & Luca Trevisan November 22, 2016 Sreaming algorihms In his lecure and he nex one we sudy memory-efficien algorihms ha process a sream

More information

Bias-Variance Error Bounds for Temporal Difference Updates

Bias-Variance Error Bounds for Temporal Difference Updates Bias-Variance Bounds for Temporal Difference Updaes Michael Kearns AT&T Labs mkearns@research.a.com Sainder Singh AT&T Labs baveja@research.a.com Absrac We give he firs rigorous upper bounds on he error

More information

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017 Two Popular Bayesian Esimaors: Paricle and Kalman Filers McGill COMP 765 Sep 14 h, 2017 1 1 1, dx x Bel x u x P x z P Recall: Bayes Filers,,,,,,, 1 1 1 1 u z u x P u z u x z P Bayes z = observaion u =

More information

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t...

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t... Mah 228- Fri Mar 24 5.6 Marix exponenials and linear sysems: The analogy beween firs order sysems of linear differenial equaions (Chaper 5) and scalar linear differenial equaions (Chaper ) is much sronger

More information

Zürich. ETH Master Course: L Autonomous Mobile Robots Localization II

Zürich. ETH Master Course: L Autonomous Mobile Robots Localization II Roland Siegwar Margaria Chli Paul Furgale Marco Huer Marin Rufli Davide Scaramuzza ETH Maser Course: 151-0854-00L Auonomous Mobile Robos Localizaion II ACT and SEE For all do, (predicion updae / ACT),

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Linear Response Theory: The connecion beween QFT and experimens 3.1. Basic conceps and ideas Q: How do we measure he conduciviy of a meal? A: we firs inroduce a weak elecric field E, and

More information

Robust estimation based on the first- and third-moment restrictions of the power transformation model

Robust estimation based on the first- and third-moment restrictions of the power transformation model h Inernaional Congress on Modelling and Simulaion, Adelaide, Ausralia, 6 December 3 www.mssanz.org.au/modsim3 Robus esimaion based on he firs- and hird-momen resricions of he power ransformaion Nawaa,

More information

2. Nonlinear Conservation Law Equations

2. Nonlinear Conservation Law Equations . Nonlinear Conservaion Law Equaions One of he clear lessons learned over recen years in sudying nonlinear parial differenial equaions is ha i is generally no wise o ry o aack a general class of nonlinear

More information

Random Walk with Anti-Correlated Steps

Random Walk with Anti-Correlated Steps Random Walk wih Ani-Correlaed Seps John Noga Dirk Wagner 2 Absrac We conjecure he expeced value of random walks wih ani-correlaed seps o be exacly. We suppor his conjecure wih 2 plausibiliy argumens and

More information

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H.

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H. ACE 56 Fall 005 Lecure 5: he Simple Linear Regression Model: Sampling Properies of he Leas Squares Esimaors by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Inference in he Simple

More information

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle Chaper 2 Newonian Mechanics Single Paricle In his Chaper we will review wha Newon s laws of mechanics ell us abou he moion of a single paricle. Newon s laws are only valid in suiable reference frames,

More information

Planning in POMDPs. Dominik Schoenberger Abstract

Planning in POMDPs. Dominik Schoenberger Abstract Planning in POMDPs Dominik Schoenberger d.schoenberger@sud.u-darmsad.de Absrac This documen briefly explains wha a Parially Observable Markov Decision Process is. Furhermore i inroduces he differen approaches

More information

20. Applications of the Genetic-Drift Model

20. Applications of the Genetic-Drift Model 0. Applicaions of he Geneic-Drif Model 1) Deermining he probabiliy of forming any paricular combinaion of genoypes in he nex generaion: Example: If he parenal allele frequencies are p 0 = 0.35 and q 0

More information

Presentation Overview

Presentation Overview Acion Refinemen in Reinforcemen Learning by Probabiliy Smoohing By Thomas G. Dieerich & Didac Busques Speaer: Kai Xu Presenaion Overview Bacground The Probabiliy Smoohing Mehod Experimenal Sudy of Acion

More information

5. Stochastic processes (1)

5. Stochastic processes (1) Lec05.pp S-38.45 - Inroducion o Teleraffic Theory Spring 2005 Conens Basic conceps Poisson process 2 Sochasic processes () Consider some quaniy in a eleraffic (or any) sysem I ypically evolves in ime randomly

More information

Chapter 2. First Order Scalar Equations

Chapter 2. First Order Scalar Equations Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.

More information

The expectation value of the field operator.

The expectation value of the field operator. The expecaion value of he field operaor. Dan Solomon Universiy of Illinois Chicago, IL dsolom@uic.edu June, 04 Absrac. Much of he mahemaical developmen of quanum field heory has been in suppor of deermining

More information

Stability and Bifurcation in a Neural Network Model with Two Delays

Stability and Bifurcation in a Neural Network Model with Two Delays Inernaional Mahemaical Forum, Vol. 6, 11, no. 35, 175-1731 Sabiliy and Bifurcaion in a Neural Nework Model wih Two Delays GuangPing Hu and XiaoLing Li School of Mahemaics and Physics, Nanjing Universiy

More information

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Simulaion-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Week Descripion Reading Maerial 2 Compuer Simulaion of Dynamic Models Finie Difference, coninuous saes, discree ime Simple Mehods Euler Trapezoid

More information

Econ107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8)

Econ107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8) I. Definiions and Problems A. Perfec Mulicollineariy Econ7 Applied Economerics Topic 7: Mulicollineariy (Sudenmund, Chaper 8) Definiion: Perfec mulicollineariy exiss in a following K-variable regression

More information

Unit Root Time Series. Univariate random walk

Unit Root Time Series. Univariate random walk Uni Roo ime Series Univariae random walk Consider he regression y y where ~ iid N 0, he leas squares esimae of is: ˆ yy y y yy Now wha if = If y y hen le y 0 =0 so ha y j j If ~ iid N 0, hen y ~ N 0, he

More information

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t Exercise 7 C P = α + β R P + u C = αp + βr + v (a) (b) C R = α P R + β + w (c) Assumpions abou he disurbances u, v, w : Classical assumions on he disurbance of one of he equaions, eg. on (b): E(v v s P,

More information

d 1 = c 1 b 2 - b 1 c 2 d 2 = c 1 b 3 - b 1 c 3

d 1 = c 1 b 2 - b 1 c 2 d 2 = c 1 b 3 - b 1 c 3 and d = c b - b c c d = c b - b c c This process is coninued unil he nh row has been compleed. The complee array of coefficiens is riangular. Noe ha in developing he array an enire row may be divided or

More information

Deep Learning: Theory, Techniques & Applications - Recurrent Neural Networks -

Deep Learning: Theory, Techniques & Applications - Recurrent Neural Networks - Deep Learning: Theory, Techniques & Applicaions - Recurren Neural Neworks - Prof. Maeo Maeucci maeo.maeucci@polimi.i Deparmen of Elecronics, Informaion and Bioengineering Arificial Inelligence and Roboics

More information

Ensamble methods: Bagging and Boosting

Ensamble methods: Bagging and Boosting Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par

More information

Morning Time: 1 hour 30 minutes Additional materials (enclosed):

Morning Time: 1 hour 30 minutes Additional materials (enclosed): ADVANCED GCE 78/0 MATHEMATICS (MEI) Differenial Equaions THURSDAY JANUARY 008 Morning Time: hour 30 minues Addiional maerials (enclosed): None Addiional maerials (required): Answer Bookle (8 pages) Graph

More information

An introduction to the theory of SDDP algorithm

An introduction to the theory of SDDP algorithm An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking

More information

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.

More information

GENERALIZATION OF THE FORMULA OF FAA DI BRUNO FOR A COMPOSITE FUNCTION WITH A VECTOR ARGUMENT

GENERALIZATION OF THE FORMULA OF FAA DI BRUNO FOR A COMPOSITE FUNCTION WITH A VECTOR ARGUMENT Inerna J Mah & Mah Sci Vol 4, No 7 000) 48 49 S0670000970 Hindawi Publishing Corp GENERALIZATION OF THE FORMULA OF FAA DI BRUNO FOR A COMPOSITE FUNCTION WITH A VECTOR ARGUMENT RUMEN L MISHKOV Received

More information

Matrix Versions of Some Refinements of the Arithmetic-Geometric Mean Inequality

Matrix Versions of Some Refinements of the Arithmetic-Geometric Mean Inequality Marix Versions of Some Refinemens of he Arihmeic-Geomeric Mean Inequaliy Bao Qi Feng and Andrew Tonge Absrac. We esablish marix versions of refinemens due o Alzer ], Carwrigh and Field 4], and Mercer 5]

More information

Lecture 33: November 29

Lecture 33: November 29 36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure

More information

Ensamble methods: Boosting

Ensamble methods: Boosting Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room

More information

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits DOI: 0.545/mjis.07.5009 Exponenial Weighed Moving Average (EWMA) Char Under The Assumpion of Moderaeness And Is 3 Conrol Limis KALPESH S TAILOR Assisan Professor, Deparmen of Saisics, M. K. Bhavnagar Universiy,

More information

) were both constant and we brought them from under the integral.

) were both constant and we brought them from under the integral. YIELD-PER-RECRUIT (coninued The yield-per-recrui model applies o a cohor, bu we saw in he Age Disribuions lecure ha he properies of a cohor do no apply in general o a collecion of cohors, which is wha

More information

Off-policy TD(λ) with a true online equivalence

Off-policy TD(λ) with a true online equivalence Off-policy TD(λ) wih a rue online equivalence Hado van Hassel A Rupam Mahmood Richard S Suon Reinforcemen Learning and Arificial Inelligence Laboraory Universiy of Albera, Edmonon, AB T6G 2E8 Canada Absrac

More information

arxiv: v1 [cs.lg] 18 Jul 2018

arxiv: v1 [cs.lg] 18 Jul 2018 General Value Funcion Neworks Mahew Schlegel Universiy of Albera mkschleg@ualbera.ca Adam Whie Universiy of Albera amw8@ualbera.ca Andrew Paerson Indiana Universiy andnpa@indiana.edu arxiv:1807.06763v1

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

Math 2142 Exam 1 Review Problems. x 2 + f (0) 3! for the 3rd Taylor polynomial at x = 0. To calculate the various quantities:

Math 2142 Exam 1 Review Problems. x 2 + f (0) 3! for the 3rd Taylor polynomial at x = 0. To calculate the various quantities: Mah 4 Eam Review Problems Problem. Calculae he 3rd Taylor polynomial for arcsin a =. Soluion. Le f() = arcsin. For his problem, we use he formula f() + f () + f ()! + f () 3! for he 3rd Taylor polynomial

More information

Particle Swarm Optimization Combining Diversification and Intensification for Nonlinear Integer Programming Problems

Particle Swarm Optimization Combining Diversification and Intensification for Nonlinear Integer Programming Problems Paricle Swarm Opimizaion Combining Diversificaion and Inensificaion for Nonlinear Ineger Programming Problems Takeshi Masui, Masaoshi Sakawa, Kosuke Kao and Koichi Masumoo Hiroshima Universiy 1-4-1, Kagamiyama,

More information

On Measuring Pro-Poor Growth. 1. On Various Ways of Measuring Pro-Poor Growth: A Short Review of the Literature

On Measuring Pro-Poor Growth. 1. On Various Ways of Measuring Pro-Poor Growth: A Short Review of the Literature On Measuring Pro-Poor Growh 1. On Various Ways of Measuring Pro-Poor Growh: A Shor eview of he Lieraure During he pas en years or so here have been various suggesions concerning he way one should check

More information

Reading from Young & Freedman: For this topic, read sections 25.4 & 25.5, the introduction to chapter 26 and sections 26.1 to 26.2 & 26.4.

Reading from Young & Freedman: For this topic, read sections 25.4 & 25.5, the introduction to chapter 26 and sections 26.1 to 26.2 & 26.4. PHY1 Elecriciy Topic 7 (Lecures 1 & 11) Elecric Circuis n his opic, we will cover: 1) Elecromoive Force (EMF) ) Series and parallel resisor combinaions 3) Kirchhoff s rules for circuis 4) Time dependence

More information

STA 114: Statistics. Notes 2. Statistical Models and the Likelihood Function

STA 114: Statistics. Notes 2. Statistical Models and the Likelihood Function STA 114: Saisics Noes 2. Saisical Models and he Likelihood Funcion Describing Daa & Saisical Models A physicis has a heory ha makes a precise predicion of wha s o be observed in daa. If he daa doesn mach

More information

Georey E. Hinton. University oftoronto. Technical Report CRG-TR February 22, Abstract

Georey E. Hinton. University oftoronto.   Technical Report CRG-TR February 22, Abstract Parameer Esimaion for Linear Dynamical Sysems Zoubin Ghahramani Georey E. Hinon Deparmen of Compuer Science Universiy oftorono 6 King's College Road Torono, Canada M5S A4 Email: zoubin@cs.orono.edu Technical

More information

di Bernardo, M. (1995). A purely adaptive controller to synchronize and control chaotic systems.

di Bernardo, M. (1995). A purely adaptive controller to synchronize and control chaotic systems. di ernardo, M. (995). A purely adapive conroller o synchronize and conrol chaoic sysems. hps://doi.org/.6/375-96(96)8-x Early version, also known as pre-prin Link o published version (if available):.6/375-96(96)8-x

More information

Single-Pass-Based Heuristic Algorithms for Group Flexible Flow-shop Scheduling Problems

Single-Pass-Based Heuristic Algorithms for Group Flexible Flow-shop Scheduling Problems Single-Pass-Based Heurisic Algorihms for Group Flexible Flow-shop Scheduling Problems PEI-YING HUANG, TZUNG-PEI HONG 2 and CHENG-YAN KAO, 3 Deparmen of Compuer Science and Informaion Engineering Naional

More information

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Kriging Models Predicing Arazine Concenraions in Surface Waer Draining Agriculural Waersheds Paul L. Mosquin, Jeremy Aldworh, Wenlin Chen Supplemenal Maerial Number

More information

Final Spring 2007

Final Spring 2007 .615 Final Spring 7 Overview The purpose of he final exam is o calculae he MHD β limi in a high-bea oroidal okamak agains he dangerous n = 1 exernal ballooning-kink mode. Effecively, his corresponds o

More information

Notes for Lecture 17-18

Notes for Lecture 17-18 U.C. Berkeley CS278: Compuaional Complexiy Handou N7-8 Professor Luca Trevisan April 3-8, 2008 Noes for Lecure 7-8 In hese wo lecures we prove he firs half of he PCP Theorem, he Amplificaion Lemma, up

More information

On Multicomponent System Reliability with Microshocks - Microdamages Type of Components Interaction

On Multicomponent System Reliability with Microshocks - Microdamages Type of Components Interaction On Mulicomponen Sysem Reliabiliy wih Microshocks - Microdamages Type of Componens Ineracion Jerzy K. Filus, and Lidia Z. Filus Absrac Consider a wo componen parallel sysem. The defined new sochasic dependences

More information

KINEMATICS IN ONE DIMENSION

KINEMATICS IN ONE DIMENSION KINEMATICS IN ONE DIMENSION PREVIEW Kinemaics is he sudy of how hings move how far (disance and displacemen), how fas (speed and velociy), and how fas ha how fas changes (acceleraion). We say ha an objec

More information

EXERCISES FOR SECTION 1.5

EXERCISES FOR SECTION 1.5 1.5 Exisence and Uniqueness of Soluions 43 20. 1 v c 21. 1 v c 1 2 4 6 8 10 1 2 2 4 6 8 10 Graph of approximae soluion obained using Euler s mehod wih = 0.1. Graph of approximae soluion obained using Euler

More information

Physics 127b: Statistical Mechanics. Fokker-Planck Equation. Time Evolution

Physics 127b: Statistical Mechanics. Fokker-Planck Equation. Time Evolution Physics 7b: Saisical Mechanics Fokker-Planck Equaion The Langevin equaion approach o he evoluion of he velociy disribuion for he Brownian paricle migh leave you uncomforable. A more formal reamen of his

More information

Ensemble Confidence Estimates Posterior Probability

Ensemble Confidence Estimates Posterior Probability Ensemble Esimaes Poserior Probabiliy Michael Muhlbaier, Aposolos Topalis, and Robi Polikar Rowan Universiy, Elecrical and Compuer Engineering, Mullica Hill Rd., Glassboro, NJ 88, USA {muhlba6, opali5}@sudens.rowan.edu

More information

NCSS Statistical Software. , contains a periodic (cyclic) component. A natural model of the periodic component would be

NCSS Statistical Software. , contains a periodic (cyclic) component. A natural model of the periodic component would be NCSS Saisical Sofware Chaper 468 Specral Analysis Inroducion This program calculaes and displays he periodogram and specrum of a ime series. This is someimes nown as harmonic analysis or he frequency approach

More information

Rapid Termination Evaluation for Recursive Subdivision of Bezier Curves

Rapid Termination Evaluation for Recursive Subdivision of Bezier Curves Rapid Terminaion Evaluaion for Recursive Subdivision of Bezier Curves Thomas F. Hain School of Compuer and Informaion Sciences, Universiy of Souh Alabama, Mobile, AL, U.S.A. Absrac Bézier curve flaening

More information

EXPLICIT TIME INTEGRATORS FOR NONLINEAR DYNAMICS DERIVED FROM THE MIDPOINT RULE

EXPLICIT TIME INTEGRATORS FOR NONLINEAR DYNAMICS DERIVED FROM THE MIDPOINT RULE Version April 30, 2004.Submied o CTU Repors. EXPLICIT TIME INTEGRATORS FOR NONLINEAR DYNAMICS DERIVED FROM THE MIDPOINT RULE Per Krysl Universiy of California, San Diego La Jolla, California 92093-0085,

More information

Seminar 4: Hotelling 2

Seminar 4: Hotelling 2 Seminar 4: Hoelling 2 November 3, 211 1 Exercise Par 1 Iso-elasic demand A non renewable resource of a known sock S can be exraced a zero cos. Demand for he resource is of he form: D(p ) = p ε ε > A a

More information

Affine term structure models

Affine term structure models Affine erm srucure models A. Inro o Gaussian affine erm srucure models B. Esimaion by minimum chi square (Hamilon and Wu) C. Esimaion by OLS (Adrian, Moench, and Crump) D. Dynamic Nelson-Siegel model (Chrisensen,

More information

Object tracking: Using HMMs to estimate the geographical location of fish

Object tracking: Using HMMs to estimate the geographical location of fish Objec racking: Using HMMs o esimae he geographical locaion of fish 02433 - Hidden Markov Models Marin Wæver Pedersen, Henrik Madsen Course week 13 MWP, compiled June 8, 2011 Objecive: Locae fish from agging

More information

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé Bias in Condiional and Uncondiional Fixed Effecs Logi Esimaion: a Correcion * Tom Coupé Economics Educaion and Research Consorium, Naional Universiy of Kyiv Mohyla Academy Address: Vul Voloska 10, 04070

More information

Section 3.5 Nonhomogeneous Equations; Method of Undetermined Coefficients

Section 3.5 Nonhomogeneous Equations; Method of Undetermined Coefficients Secion 3.5 Nonhomogeneous Equaions; Mehod of Undeermined Coefficiens Key Terms/Ideas: Linear Differenial operaor Nonlinear operaor Second order homogeneous DE Second order nonhomogeneous DE Soluion o homogeneous

More information

Let us start with a two dimensional case. We consider a vector ( x,

Let us start with a two dimensional case. We consider a vector ( x, Roaion marices We consider now roaion marices in wo and hree dimensions. We sar wih wo dimensions since wo dimensions are easier han hree o undersand, and one dimension is a lile oo simple. However, our

More information

Inventory Control of Perishable Items in a Two-Echelon Supply Chain

Inventory Control of Perishable Items in a Two-Echelon Supply Chain Journal of Indusrial Engineering, Universiy of ehran, Special Issue,, PP. 69-77 69 Invenory Conrol of Perishable Iems in a wo-echelon Supply Chain Fariborz Jolai *, Elmira Gheisariha and Farnaz Nojavan

More information

Some Ramsey results for the n-cube

Some Ramsey results for the n-cube Some Ramsey resuls for he n-cube Ron Graham Universiy of California, San Diego Jozsef Solymosi Universiy of Briish Columbia, Vancouver, Canada Absrac In his noe we esablish a Ramsey-ype resul for cerain

More information

Navneet Saini, Mayank Goyal, Vishal Bansal (2013); Term Project AML310; Indian Institute of Technology Delhi

Navneet Saini, Mayank Goyal, Vishal Bansal (2013); Term Project AML310; Indian Institute of Technology Delhi Creep in Viscoelasic Subsances Numerical mehods o calculae he coefficiens of he Prony equaion using creep es daa and Herediary Inegrals Mehod Navnee Saini, Mayank Goyal, Vishal Bansal (23); Term Projec

More information

2.7. Some common engineering functions. Introduction. Prerequisites. Learning Outcomes

2.7. Some common engineering functions. Introduction. Prerequisites. Learning Outcomes Some common engineering funcions 2.7 Inroducion This secion provides a caalogue of some common funcions ofen used in Science and Engineering. These include polynomials, raional funcions, he modulus funcion

More information

Phys1112: DC and RC circuits

Phys1112: DC and RC circuits Name: Group Members: Dae: TA s Name: Phys1112: DC and RC circuis Objecives: 1. To undersand curren and volage characerisics of a DC RC discharging circui. 2. To undersand he effec of he RC ime consan.

More information

Single and Double Pendulum Models

Single and Double Pendulum Models Single and Double Pendulum Models Mah 596 Projec Summary Spring 2016 Jarod Har 1 Overview Differen ypes of pendulums are used o model many phenomena in various disciplines. In paricular, single and double

More information

In this chapter the model of free motion under gravity is extended to objects projected at an angle. When you have completed it, you should

In this chapter the model of free motion under gravity is extended to objects projected at an angle. When you have completed it, you should Cambridge Universiy Press 978--36-60033-7 Cambridge Inernaional AS and A Level Mahemaics: Mechanics Coursebook Excerp More Informaion Chaper The moion of projeciles In his chaper he model of free moion

More information

SUPPLEMENTARY INFORMATION

SUPPLEMENTARY INFORMATION SUPPLEMENTARY INFORMATION DOI: 0.038/NCLIMATE893 Temporal resoluion and DICE * Supplemenal Informaion Alex L. Maren and Sephen C. Newbold Naional Cener for Environmenal Economics, US Environmenal Proecion

More information

LAB # 2 - Equilibrium (static)

LAB # 2 - Equilibrium (static) AB # - Equilibrium (saic) Inroducion Isaac Newon's conribuion o physics was o recognize ha despie he seeming compleiy of he Unierse, he moion of is pars is guided by surprisingly simple aws. Newon's inspiraion

More information

A First Course on Kinetics and Reaction Engineering. Class 19 on Unit 18

A First Course on Kinetics and Reaction Engineering. Class 19 on Unit 18 A Firs ourse on Kineics and Reacion Engineering lass 19 on Uni 18 Par I - hemical Reacions Par II - hemical Reacion Kineics Where We re Going Par III - hemical Reacion Engineering A. Ideal Reacors B. Perfecly

More information

EKF SLAM vs. FastSLAM A Comparison

EKF SLAM vs. FastSLAM A Comparison vs. A Comparison Michael Calonder, Compuer Vision Lab Swiss Federal Insiue of Technology, Lausanne EPFL) michael.calonder@epfl.ch The wo algorihms are described wih a planar robo applicaion in mind. Generalizaion

More information