Classical Conditioning IV: TD learning in the brain

Size: px
Start display at page:

Download "Classical Conditioning IV: TD learning in the brain"

Transcription

1 Classical Condiioning IV: TD learning in he brain PSY/NEU338: Animal learning and decision making: Psychological, compuaional and neural perspecives recap: Marr s levels of analysis David Marr ( ) proposed hree levels of analysis: 1. he problem (Compuaional Level) 2. he sraegy (Algorihmic Level) 3. how is acually done by neworks of neurons (Implemenaional Level) 2

2 les sar over, his ime from he op... V = E V = E V = E i=+1 " 1 X i=+1 T i=+1 r i i 1 r i # r i wan o predic expeced sum of fuure reinforcemen wan o predic expeced sum of discouned fuure reinf. (0<γ<1) wan o predic expeced sum of fuure reinforcemen in a rial/episode 3 les sar over, his ime from he op... V = E [r +1 + r r T ] = E [r +1 ]+E [r r T ] (noe: indexes ime wihin a rial) = E [r +1 ]+V +1 V = E T i=+1 r i wan o predic expeced sum of fuure reinforcemen in a rial/episode 4

3 les sar over, his ime from he op... V = E [r +1 + r r T ] = E [r +1 ]+E [r r T ] (noe: indexes ime wihin a rial) = E [r +1 ]+V +1 Think fooball Wha would be a sensible learning rule here? How is his differen from Rescorla-Wagner? 5 Temporal Difference (TD) learning Marr s 3 levels: The algorihm: (noe: indexes ime wihin a rial, T indexes rials) V = E [r +1 ]+V +1 V new = V old + (r +1 + V old +1 V old ) emporal difference predicion error δ(+1) compare o: V T +1 = V T + r T V T Suon & Baro 1983,

4 Temporal Difference (TD) learning Marr s 3 levels: The algorihm: V new = V old + (r +1 + V old +1 V old ) predicion error δ(+1) beginning of rial r = 0 V -1 = 0 δ() = V middle of rial r = 0 δ()=v -V -1 end of rial V = 0 δ()=r - V -1 Suon & Baro 1983, dopamine δ() = r δ() = V δ() = r -V -1 δ() = V δ() = 0-V -1 Schulz e al,

5 simulaion wha would happen wih parial reinforcemen? wha would happen in second order condiioning? A noe on ime book-keeping V new = V old + (r +1 + V old +1 V old ) predicion error δ+1 he predicion error δ can be defined a any ime poin, bu can only be based on informaion he animal already has (no informaion ha i will ge in he fuure!) so, we can say δ+1 = r+1+v+1-v as above, because a ime (+1) informaion regarding r+1 and V+1 is already known we can also say δ = r+v-v-1 in prey much he same way we canno say δ = r+1+v+1-v as i jus would no make logical sense! imporanly, in all cases, δ is used o updae he preceding predicion, ha is, δ+1 is used o updae V and δ is used o updae V-1 10

6 wha does he heory explain? acquisiion exincion blocking overshadowing emporal relaionships overexpecaion 2 nd order cond. R-W X X TD 11 Summary so far... Temporal difference learning is a beer version of Rescorla-Wagner learning derived from firs principles (from definiion of problem) explains everyhing ha R-W does, and more (eg. 2 nd order condiioning) basically a generalizaion of R-W o real ime break! 12

7 Back o Marr s hree levels The algorihm: emporal difference learning Neural implemenaion: does he brain use TD learning? 13 we already saw his: δ() = r δ() = V δ() = r -V -1 δ() = V δ() = 0-V -1 Schulz e al,

8 predicion error hypohesis of dopamine The idea: Dopamine encodes a emporal difference reward predicion error (Monague, Dayan, Baro mid 90 s) Schulz e al, 1993 Tobler e al, predicion error hypohesis of dopamine Fiorillo e al, 2003 measured firing rae model predicion error Bayer & Glimcher (2005) 16

9 where does dopamine projec o? main arge: sriaum in basal ganglia (also prefronal corex) 17 he basal ganglia: afferens inpus o sriaum are from all over he corex (and hey are opographic) Voorn e al,

10 a precise microsrucure 19 dopamine and synapic plasiciy predicion errors are for learning corico-sriaal synapses show dopamine-dependen plasiciy hree-facor learning rule: need presynapic+possynapic+dopamine Wickens e al,

11 summary Classical condiioning can be viewed as predicion learning The problem: predicion of fuure reward The algorihm: emporal difference learning Neural implemenaion: dopamine dependen learning in BG A compuaional model of learning allows us o look in he brain for hidden variables posulaed by he model Precise (normaive!) heory for generaion of dopamine firing paerns Explains anicipaory dopaminergic responding, 2 nd order condiioning Compelling accoun for he role of dopamine in classical condiioning: predicion error drives predicion learning 21 if you are confused or inrigued: addiional reading Rescorla & Wagner (1972) - A heory of Pavlovian condiioning: Variaions in he effeciveness of reinforcemen and nonreinforcemen - he original chaper ha is so well cied (and well wrien!) Suon & Baro (1990) - Time derivaive models of Pavlovian reinforcemen - shows sep by sep why TD learning is a suiable rule for modeling classical condiioning Niv & Schoenbaum (2008) - Dialogues on predicion errors - a guide for he perplexed Baro (1995) - adapive criic and he basal ganglia - very clear exposiion o TD learning in he basal ganglia (all will be on BlackBoard) 22

CSE/NB 528 Lecture 14: Reinforcement Learning (Chapter 9)

CSE/NB 528 Lecture 14: Reinforcement Learning (Chapter 9) CSE/NB 528 Lecure 14: Reinforcemen Learning Chaper 9 Image from hp://clasdean.la.asu.edu/news/images/ubep2001/neuron3.jpg Lecure figures are from Dayan & Abbo s book hp://people.brandeis.edu/~abbo/book/index.hml

More information

CSE/NB 528 Lecture 14: From Supervised to Reinforcement Learning (Chapter 9) R. Rao, 528: Lecture 14

CSE/NB 528 Lecture 14: From Supervised to Reinforcement Learning (Chapter 9) R. Rao, 528: Lecture 14 CSE/NB 58 Lecure 14: From Supervised o Reinforcemen Learning Chaper 9 1 Recall from las ime: Sigmoid Neworks Oupu v T g w u g wiui w Inpu nodes u = u 1 u u 3 T i Sigmoid oupu funcion: 1 g a 1 a e 1 ga

More information

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1 RL Lecure 7: Eligibiliy Traces R. S. Suon and A. G. Baro: Reinforcemen Learning: An Inroducion 1 N-sep TD Predicion Idea: Look farher ino he fuure when you do TD backup (1, 2, 3,, n seps) R. S. Suon and

More information

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 175 CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 10.1 INTRODUCTION Amongs he research work performed, he bes resuls of experimenal work are validaed wih Arificial Neural Nework. From he

More information

Chapter 21. Reinforcement Learning. The Reinforcement Learning Agent

Chapter 21. Reinforcement Learning. The Reinforcement Learning Agent CSE 47 Chaper Reinforcemen Learning The Reinforcemen Learning Agen Agen Sae u Reward r Acion a Enironmen CSE AI Faculy Why reinforcemen learning Programming an agen o drie a car or fly a helicoper is ery

More information

1 Review of Zero-Sum Games

1 Review of Zero-Sum Games COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any

More information

CSE/NEURO 528 Lecture 13: Reinforcement Learning & Course Review (Chapter 9)

CSE/NEURO 528 Lecture 13: Reinforcement Learning & Course Review (Chapter 9) CSE/NEURO 528 Lecure 13: Reinforceen Learning & Course Review Chaper 9 Aniaion: To Creed, SJU 1 Early Resuls: Pavlov and his Dog F Classical Pavlovian condiioning experiens F Training: Bell Food F Afer:

More information

Code-specific policy gradient rules for spiking neurons

Code-specific policy gradient rules for spiking neurons Code-specific policy gradien rules for spiking neurons Henning Sprekeler Guillaume Hennequin Wulfram Gersner Laboraory for Compuaional Neuroscience École Polyechnique Fédérale de Lausanne 115 Lausanne

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

Final Spring 2007

Final Spring 2007 .615 Final Spring 7 Overview The purpose of he final exam is o calculae he MHD β limi in a high-bea oroidal okamak agains he dangerous n = 1 exernal ballooning-kink mode. Effecively, his corresponds o

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

Notes on online convex optimization

Notes on online convex optimization Noes on online convex opimizaion Karl Sraos Online convex opimizaion (OCO) is a principled framework for online learning: OnlineConvexOpimizaion Inpu: convex se S, number of seps T For =, 2,..., T : Selec

More information

Notes on Kalman Filtering

Notes on Kalman Filtering Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren

More information

Econ107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8)

Econ107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8) I. Definiions and Problems A. Perfec Mulicollineariy Econ7 Applied Economerics Topic 7: Mulicollineariy (Sudenmund, Chaper 8) Definiion: Perfec mulicollineariy exiss in a following K-variable regression

More information

Ensamble methods: Boosting

Ensamble methods: Boosting Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room

More information

Energy Storage and Renewables in New Jersey: Complementary Technologies for Reducing Our Carbon Footprint

Energy Storage and Renewables in New Jersey: Complementary Technologies for Reducing Our Carbon Footprint Energy Sorage and Renewables in New Jersey: Complemenary Technologies for Reducing Our Carbon Fooprin ACEE E-filliaes workshop November 14, 2014 Warren B. Powell Daniel Seingar Harvey Cheng Greg Davies

More information

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance

More information

Attention-Gated Reinforcement Learning in Neural Networks A Unified View

Attention-Gated Reinforcement Learning in Neural Networks A Unified View Aenion-Gaed Reinforcemen Learning in Neural Neworks A Unified View Tobias Brosch, Friedhelm Schwenker, and Heiko Neumann Insiue of Neural Informaion Processing, Universiy of Ulm, 89069 Ulm, Germany {obias.brosch,friedhelm.schwenker,heiko.neumann}@uni-ulm.de

More information

Online Convex Optimization Example And Follow-The-Leader

Online Convex Optimization Example And Follow-The-Leader CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion

More information

Ensamble methods: Bagging and Boosting

Ensamble methods: Bagging and Boosting Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par

More information

Scale invariant value computation for reinforcement learning in continuous time

Scale invariant value computation for reinforcement learning in continuous time Scale invarian value compuaion for reinforcemen learning in coninuous ime Zoran Tiganj, Karhi H. Shanar and Marc W. Howard Deparmen of Psychological and Brain Sciences Boson Universiy Absrac Naural learners

More information

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Simulaion-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Week Descripion Reading Maerial 2 Compuer Simulaion of Dynamic Models Finie Difference, coninuous saes, discree ime Simple Mehods Euler Trapezoid

More information

Bias-Variance Error Bounds for Temporal Difference Updates

Bias-Variance Error Bounds for Temporal Difference Updates Bias-Variance Bounds for Temporal Difference Updaes Michael Kearns AT&T Labs mkearns@research.a.com Sainder Singh AT&T Labs baveja@research.a.com Absrac We give he firs rigorous upper bounds on he error

More information

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle Chaper 2 Newonian Mechanics Single Paricle In his Chaper we will review wha Newon s laws of mechanics ell us abou he moion of a single paricle. Newon s laws are only valid in suiable reference frames,

More information

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class

More information

Introduction D P. r = constant discount rate, g = Gordon Model (1962): constant dividend growth rate.

Introduction D P. r = constant discount rate, g = Gordon Model (1962): constant dividend growth rate. Inroducion Gordon Model (1962): D P = r g r = consan discoun rae, g = consan dividend growh rae. If raional expecaions of fuure discoun raes and dividend growh vary over ime, so should he D/P raio. Since

More information

14 Autoregressive Moving Average Models

14 Autoregressive Moving Average Models 14 Auoregressive Moving Average Models In his chaper an imporan parameric family of saionary ime series is inroduced, he family of he auoregressive moving average, or ARMA, processes. For a large class

More information

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé Bias in Condiional and Uncondiional Fixed Effecs Logi Esimaion: a Correcion * Tom Coupé Economics Educaion and Research Consorium, Naional Universiy of Kyiv Mohyla Academy Address: Vul Voloska 10, 04070

More information

Lecture 9: September 25

Lecture 9: September 25 0-725: Opimizaion Fall 202 Lecure 9: Sepember 25 Lecurer: Geoff Gordon/Ryan Tibshirani Scribes: Xuezhi Wang, Subhodeep Moira, Abhimanu Kumar Noe: LaTeX emplae couresy of UC Berkeley EECS dep. Disclaimer:

More information

Financial Econometrics Jeffrey R. Russell Midterm Winter 2009 SOLUTIONS

Financial Econometrics Jeffrey R. Russell Midterm Winter 2009 SOLUTIONS Name SOLUTIONS Financial Economerics Jeffrey R. Russell Miderm Winer 009 SOLUTIONS You have 80 minues o complee he exam. Use can use a calculaor and noes. Try o fi all your work in he space provided. If

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Noes for EE7C Spring 018: Convex Opimizaion and Approximaion Insrucor: Moriz Hard Email: hard+ee7c@berkeley.edu Graduae Insrucor: Max Simchowiz Email: msimchow+ee7c@berkeley.edu Ocober 15, 018 3

More information

Chapter 4. Truncation Errors

Chapter 4. Truncation Errors Chaper 4. Truncaion Errors and he Taylor Series Truncaion Errors and he Taylor Series Non-elemenary funcions such as rigonomeric, eponenial, and ohers are epressed in an approimae fashion using Taylor

More information

Robust and Learning Control for Complex Systems

Robust and Learning Control for Complex Systems Robus and Learning Conrol for Complex Sysems Peer M. Young Sepember 13, 2007 & Talk Ouline Inroducion Robus Conroller Analysis and Design Theory Experimenal Applicaions Overview MIMO Robus HVAC Conrol

More information

Sophisticated Monetary Policies. Andrew Atkeson. V.V. Chari. Patrick Kehoe

Sophisticated Monetary Policies. Andrew Atkeson. V.V. Chari. Patrick Kehoe Sophisicaed Moneary Policies Andrew Akeson UCLA V.V. Chari Universiy of Minnesoa Parick Kehoe Federal Reserve Bank of Minneapolis and Universiy of Minnesoa Barro, Lucas-Sokey Approach o Policy Solve Ramsey

More information

Linear Time-invariant systems, Convolution, and Cross-correlation

Linear Time-invariant systems, Convolution, and Cross-correlation Linear Time-invarian sysems, Convoluion, and Cross-correlaion (1) Linear Time-invarian (LTI) sysem A sysem akes in an inpu funcion and reurns an oupu funcion. x() T y() Inpu Sysem Oupu y() = T[x()] An

More information

True Online Temporal-Difference Learning. A. Rupam Mahmood Patrick M. Pilarski

True Online Temporal-Difference Learning. A. Rupam Mahmood Patrick M. Pilarski True Online Temporal-Difference Learning True Online Temporal-Difference Learning Harm van Seijen harm.vanseijen@ualbera.ca A. Rupam Mahmood ashique@ualbera.ca Parick M. Pilarski parick.pilarski@ualbera.ca

More information

The Rosenblatt s LMS algorithm for Perceptron (1958) is built around a linear neuron (a neuron with a linear

The Rosenblatt s LMS algorithm for Perceptron (1958) is built around a linear neuron (a neuron with a linear In The name of God Lecure4: Percepron and AALIE r. Majid MjidGhoshunih Inroducion The Rosenbla s LMS algorihm for Percepron 958 is buil around a linear neuron a neuron ih a linear acivaion funcion. Hoever,

More information

Planning in POMDPs. Dominik Schoenberger Abstract

Planning in POMDPs. Dominik Schoenberger Abstract Planning in POMDPs Dominik Schoenberger d.schoenberger@sud.u-darmsad.de Absrac This documen briefly explains wha a Parially Observable Markov Decision Process is. Furhermore i inroduces he differen approaches

More information

Inventory Analysis and Management. Multi-Period Stochastic Models: Optimality of (s, S) Policy for K-Convex Objective Functions

Inventory Analysis and Management. Multi-Period Stochastic Models: Optimality of (s, S) Policy for K-Convex Objective Functions Muli-Period Sochasic Models: Opimali of (s, S) Polic for -Convex Objecive Funcions Consider a seing similar o he N-sage newsvendor problem excep ha now here is a fixed re-ordering cos (> 0) for each (re-)order.

More information

This document was generated at 1:04 PM, 09/10/13 Copyright 2013 Richard T. Woodward. 4. End points and transversality conditions AGEC

This document was generated at 1:04 PM, 09/10/13 Copyright 2013 Richard T. Woodward. 4. End points and transversality conditions AGEC his documen was generaed a 1:4 PM, 9/1/13 Copyrigh 213 Richard. Woodward 4. End poins and ransversaliy condiions AGEC 637-213 F z d Recall from Lecure 3 ha a ypical opimal conrol problem is o maimize (,,

More information

LAPLACE TRANSFORM AND TRANSFER FUNCTION

LAPLACE TRANSFORM AND TRANSFER FUNCTION CHBE320 LECTURE V LAPLACE TRANSFORM AND TRANSFER FUNCTION Professor Dae Ryook Yang Spring 2018 Dep. of Chemical and Biological Engineering 5-1 Road Map of he Lecure V Laplace Transform and Transfer funcions

More information

Introduction to choice over time

Introduction to choice over time Microeconomic Theory -- Choice over ime Inroducion o choice over ime Individual choice Income and subsiuion effecs 7 Walrasian equilibrium ineres rae 9 pages John Riley Ocober 9, 08 Microeconomic Theory

More information

The Brock-Mirman Stochastic Growth Model

The Brock-Mirman Stochastic Growth Model c December 3, 208, Chrisopher D. Carroll BrockMirman The Brock-Mirman Sochasic Growh Model Brock and Mirman (972) provided he firs opimizing growh model wih unpredicable (sochasic) shocks. The social planner

More information

Machine Learning 4771

Machine Learning 4771 ony Jebara, Columbia Universiy achine Learning 4771 Insrucor: ony Jebara ony Jebara, Columbia Universiy opic 20 Hs wih Evidence H Collec H Evaluae H Disribue H Decode H Parameer Learning via JA & E ony

More information

Math Week 14 April 16-20: sections first order systems of linear differential equations; 7.4 mass-spring systems.

Math Week 14 April 16-20: sections first order systems of linear differential equations; 7.4 mass-spring systems. Mah 2250-004 Week 4 April 6-20 secions 7.-7.3 firs order sysems of linear differenial equaions; 7.4 mass-spring sysems. Mon Apr 6 7.-7.2 Sysems of differenial equaions (7.), and he vecor Calculus we need

More information

EECE 301 Signals & Systems Prof. Mark Fowler

EECE 301 Signals & Systems Prof. Mark Fowler EECE 31 Signals & Sysems Prof. Mark Fowler Noe Se #1 C-T Sysems: Convoluion Represenaion Reading Assignmen: Secion 2.6 of Kamen and Heck 1/11 Course Flow Diagram The arrows here show concepual flow beween

More information

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still. Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in

More information

Lecture 4: November 13

Lecture 4: November 13 Compuaional Learning Theory Fall Semeser, 2017/18 Lecure 4: November 13 Lecurer: Yishay Mansour Scribe: Guy Dolinsky, Yogev Bar-On, Yuval Lewi 4.1 Fenchel-Conjugae 4.1.1 Moivaion Unil his lecure we saw

More information

Solutions Problem Set 3 Macro II (14.452)

Solutions Problem Set 3 Macro II (14.452) Soluions Problem Se 3 Macro II (14.452) Francisco A. Gallego 04/27/2005 1 Q heory of invesmen in coninuous ime and no uncerainy Consider he in nie horizon model of a rm facing adjusmen coss o invesmen.

More information

Some Basic Information about M-S-D Systems

Some Basic Information about M-S-D Systems Some Basic Informaion abou M-S-D Sysems 1 Inroducion We wan o give some summary of he facs concerning unforced (homogeneous) and forced (non-homogeneous) models for linear oscillaors governed by second-order,

More information

3.1 More on model selection

3.1 More on model selection 3. More on Model selecion 3. Comparing models AIC, BIC, Adjused R squared. 3. Over Fiing problem. 3.3 Sample spliing. 3. More on model selecion crieria Ofen afer model fiing you are lef wih a handful of

More information

A Reinforcement Learning Approach for Collaborative Filtering

A Reinforcement Learning Approach for Collaborative Filtering A Reinforcemen Learning Approach for Collaboraive Filering Jungkyu Lee, Byonghwa Oh 2, Jihoon Yang 2, and Sungyong Park 2 Cyram Inc, Seoul, Korea jklee@cyram.com 2 Sogang Universiy, Seoul, Korea {mrfive,yangjh,parksy}@sogang.ac.kr

More information

Affine term structure models

Affine term structure models Affine erm srucure models A. Inro o Gaussian affine erm srucure models B. Esimaion by minimum chi square (Hamilon and Wu) C. Esimaion by OLS (Adrian, Moench, and Crump) D. Dynamic Nelson-Siegel model (Chrisensen,

More information

Wednesday, November 7 Handout: Heteroskedasticity

Wednesday, November 7 Handout: Heteroskedasticity Amhers College Deparmen of Economics Economics 360 Fall 202 Wednesday, November 7 Handou: Heeroskedasiciy Preview Review o Regression Model o Sandard Ordinary Leas Squares (OLS) Premises o Esimaion Procedures

More information

An introduction to the theory of SDDP algorithm

An introduction to the theory of SDDP algorithm An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking

More information

( ) ( ) if t = t. It must satisfy the identity. So, bulkiness of the unit impulse (hyper)function is equal to 1. The defining characteristic is

( ) ( ) if t = t. It must satisfy the identity. So, bulkiness of the unit impulse (hyper)function is equal to 1. The defining characteristic is UNIT IMPULSE RESPONSE, UNIT STEP RESPONSE, STABILITY. Uni impulse funcion (Dirac dela funcion, dela funcion) rigorously defined is no sricly a funcion, bu disribuion (or measure), precise reamen requires

More information

Robust Learning Control with Application to HVAC Systems

Robust Learning Control with Application to HVAC Systems Robus Learning Conrol wih Applicaion o HVAC Sysems Naional Science Foundaion & Projec Invesigaors: Dr. Charles Anderson, CS Dr. Douglas Hile, ME Dr. Peer Young, ECE Mechanical Engineering Compuer Science

More information

Vehicle Arrival Models : Headway

Vehicle Arrival Models : Headway Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where

More information

Problem Set 5. Graduate Macro II, Spring 2017 The University of Notre Dame Professor Sims

Problem Set 5. Graduate Macro II, Spring 2017 The University of Notre Dame Professor Sims Problem Se 5 Graduae Macro II, Spring 2017 The Universiy of Nore Dame Professor Sims Insrucions: You may consul wih oher members of he class, bu please make sure o urn in your own work. Where applicable,

More information

Hamilton- J acobi Equation: Explicit Formulas In this lecture we try to apply the method of characteristics to the Hamilton-Jacobi equation: u t

Hamilton- J acobi Equation: Explicit Formulas In this lecture we try to apply the method of characteristics to the Hamilton-Jacobi equation: u t M ah 5 2 7 Fall 2 0 0 9 L ecure 1 0 O c. 7, 2 0 0 9 Hamilon- J acobi Equaion: Explici Formulas In his lecure we ry o apply he mehod of characerisics o he Hamilon-Jacobi equaion: u + H D u, x = 0 in R n

More information

Applications in Industry (Extended) Kalman Filter. Week Date Lecture Title

Applications in Industry (Extended) Kalman Filter. Week Date Lecture Title hp://elec34.com Applicaions in Indusry (Eended) Kalman Filer 26 School of Informaion echnology and Elecrical Engineering a he Universiy of Queensland Lecure Schedule: Week Dae Lecure ile 29-Feb Inroducion

More information

Air Quality Index Prediction Using Error Back Propagation Algorithm and Improved Particle Swarm Optimization

Air Quality Index Prediction Using Error Back Propagation Algorithm and Improved Particle Swarm Optimization Air Qualiy Index Predicion Using Error Back Propagaion Algorihm and Improved Paricle Swarm Opimizaion Jia Xu ( ) and Lang Pei College of Compuer Science, Wuhan Qinchuan Universiy, Wuhan, China 461406563@qq.com

More information

Biol. 356 Lab 8. Mortality, Recruitment, and Migration Rates

Biol. 356 Lab 8. Mortality, Recruitment, and Migration Rates Biol. 356 Lab 8. Moraliy, Recruimen, and Migraion Raes (modified from Cox, 00, General Ecology Lab Manual, McGraw Hill) Las week we esimaed populaion size hrough several mehods. One assumpion of all hese

More information

APPM 2360 Homework Solutions, Due June 10

APPM 2360 Homework Solutions, Due June 10 2.2.2: Find general soluions for he equaion APPM 2360 Homework Soluions, Due June 10 Soluion: Finding he inegraing facor, dy + 2y = 3e µ) = e 2) = e 2 Muliplying he differenial equaion by he inegraing

More information

18 Biological models with discrete time

18 Biological models with discrete time 8 Biological models wih discree ime The mos imporan applicaions, however, may be pedagogical. The elegan body of mahemaical heory peraining o linear sysems (Fourier analysis, orhogonal funcions, and so

More information

An Analysis of Actor/Critic Algorithms using Eligibility Traces: Reinforcement Learning with Imperfect Value Functions

An Analysis of Actor/Critic Algorithms using Eligibility Traces: Reinforcement Learning with Imperfect Value Functions An Analysis of Acor/Criic Algorihms using Eligibiliy Traces: Reinforcemen Learning wih Imperfec Value Funcions Hajime Kimura 3 Tokyo Insiue of Technology gen@fe.dis.iech.ac.jp Shigenobu Kobayashi Tokyo

More information

m = 41 members n = 27 (nonfounders), f = 14 (founders) 8 markers from chromosome 19

m = 41 members n = 27 (nonfounders), f = 14 (founders) 8 markers from chromosome 19 Sequenial Imporance Sampling (SIS) AKA Paricle Filering, Sequenial Impuaion (Kong, Liu, Wong, 994) For many problems, sampling direcly from he arge disribuion is difficul or impossible. One reason possible

More information

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017 Two Popular Bayesian Esimaors: Paricle and Kalman Filers McGill COMP 765 Sep 14 h, 2017 1 1 1, dx x Bel x u x P x z P Recall: Bayes Filers,,,,,,, 1 1 1 1 u z u x P u z u x z P Bayes z = observaion u =

More information

Announcements: Warm-up Exercise:

Announcements: Warm-up Exercise: Fri Apr 13 7.1 Sysems of differenial equaions - o model muli-componen sysems via comparmenal analysis hp//en.wikipedia.org/wiki/muli-comparmen_model Announcemens Warm-up Exercise Here's a relaively simple

More information

Lecture #11: Wavepacket Dynamics for Harmonic Oscillator

Lecture #11: Wavepacket Dynamics for Harmonic Oscillator Lecure #11: Wavepacke Dynamics for Harmonic Oscillaor and PIB Las ime: Time Dependen Schrödinger Equaion Ψ HHΨ = iħ Express Ψ in complee basis se of eigenfuncions of ime independen H H {ψ n (x), E n }

More information

AP Chemistry--Chapter 12: Chemical Kinetics

AP Chemistry--Chapter 12: Chemical Kinetics AP Chemisry--Chaper 12: Chemical Kineics I. Reacion Raes A. The area of chemisry ha deals wih reacion raes, or how fas a reacion occurs, is called chemical kineics. B. The rae of reacion depends on he

More information

Week 1 Lecture 2 Problems 2, 5. What if something oscillates with no obvious spring? What is ω? (problem set problem)

Week 1 Lecture 2 Problems 2, 5. What if something oscillates with no obvious spring? What is ω? (problem set problem) Week 1 Lecure Problems, 5 Wha if somehing oscillaes wih no obvious spring? Wha is ω? (problem se problem) Sar wih Try and ge o SHM form E. Full beer can in lake, oscillaing F = m & = ge rearrange: F =

More information

Online Appendix to Solution Methods for Models with Rare Disasters

Online Appendix to Solution Methods for Models with Rare Disasters Online Appendix o Soluion Mehods for Models wih Rare Disasers Jesús Fernández-Villaverde and Oren Levinal In his Online Appendix, we presen he Euler condiions of he model, we develop he pricing Calvo block,

More information

5.1 - Logarithms and Their Properties

5.1 - Logarithms and Their Properties Chaper 5 Logarihmic Funcions 5.1 - Logarihms and Their Properies Suppose ha a populaion grows according o he formula P 10, where P is he colony size a ime, in hours. When will he populaion be 2500? We

More information

Seminar 4: Hotelling 2

Seminar 4: Hotelling 2 Seminar 4: Hoelling 2 November 3, 211 1 Exercise Par 1 Iso-elasic demand A non renewable resource of a known sock S can be exraced a zero cos. Demand for he resource is of he form: D(p ) = p ε ε > A a

More information

Non-parametric directionality analyses of electrophysiological and neuroimaging signals.

Non-parametric directionality analyses of electrophysiological and neuroimaging signals. Non-parameric direcionaliy analyses of elecrophysiological and neuroimaging signals. David Halliday Deparmen of Elecronic Engineering Universiy of York Neural circuiry in C elegans (Newman, 010) Overview

More information

Differential Geometry: Numerical Integration and Surface Flow

Differential Geometry: Numerical Integration and Surface Flow Differenial Geomery: Numerical Inegraion and Surface Flow [Implici Fairing of Irregular Meshes using Diffusion and Curaure Flow. Desbrun e al., 1999] Energy Minimizaion Recall: We hae been considering

More information

Homework sheet Exercises done during the lecture of March 12, 2014

Homework sheet Exercises done during the lecture of March 12, 2014 EXERCISE SESSION 2A FOR THE COURSE GÉOMÉTRIE EUCLIDIENNE, NON EUCLIDIENNE ET PROJECTIVE MATTEO TOMMASINI Homework shee 3-4 - Exercises done during he lecure of March 2, 204 Exercise 2 Is i rue ha he parameerized

More information

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors We are InechOpen, he world s leading publisher of Open Access books Buil by scieniss, for scieniss 3,500 08,500.7 M Open access books available Inernaional auhors and ediors Downloads Our auhors are among

More information

Slide03 Historical Overview Haykin Chapter 3 (Chap 1, 3, 3rd Ed): Single-Layer Perceptrons Multiple Faces of a Single Neuron Part I: Adaptive Filter

Slide03 Historical Overview Haykin Chapter 3 (Chap 1, 3, 3rd Ed): Single-Layer Perceptrons Multiple Faces of a Single Neuron Part I: Adaptive Filter Slide3 Haykin Chaper 3 (Chap, 3, 3rd Ed): Single-Layer Perceprons CPSC 636-6 Insrucor: Yoonsuck Choe Hisorical Overview McCulloch and Pis (943): neural neworks as compuing machines. Hebb (949): posulaed

More information

Dynamic Econometric Models: Y t = + 0 X t + 1 X t X t k X t-k + e t. A. Autoregressive Model:

Dynamic Econometric Models: Y t = + 0 X t + 1 X t X t k X t-k + e t. A. Autoregressive Model: Dynamic Economeric Models: A. Auoregressive Model: Y = + 0 X 1 Y -1 + 2 Y -2 + k Y -k + e (Wih lagged dependen variable(s) on he RHS) B. Disribued-lag Model: Y = + 0 X + 1 X -1 + 2 X -2 + + k X -k + e

More information

EXERCISES FOR SECTION 1.5

EXERCISES FOR SECTION 1.5 1.5 Exisence and Uniqueness of Soluions 43 20. 1 v c 21. 1 v c 1 2 4 6 8 10 1 2 2 4 6 8 10 Graph of approximae soluion obained using Euler s mehod wih = 0.1. Graph of approximae soluion obained using Euler

More information

THE MYSTERY OF STOCHASTIC MECHANICS. Edward Nelson Department of Mathematics Princeton University

THE MYSTERY OF STOCHASTIC MECHANICS. Edward Nelson Department of Mathematics Princeton University THE MYSTERY OF STOCHASTIC MECHANICS Edward Nelson Deparmen of Mahemaics Princeon Universiy 1 Classical Hamilon-Jacobi heory N paricles of various masses on a Euclidean space. Incorporae he masses in he

More information

Numerical Dispersion

Numerical Dispersion eview of Linear Numerical Sabiliy Numerical Dispersion n he previous lecure, we considered he linear numerical sabiliy of boh advecion and diffusion erms when approimaed wih several spaial and emporal

More information

Chapter 2. First Order Scalar Equations

Chapter 2. First Order Scalar Equations Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Linear Response Theory: The connecion beween QFT and experimens 3.1. Basic conceps and ideas Q: How do we measure he conduciviy of a meal? A: we firs inroduce a weak elecric field E, and

More information

Subway stations energy and air quality management

Subway stations energy and air quality management Subway saions energy and air qualiy managemen wih sochasic opimizaion Trisan Rigau 1,2,4, Advisors: P. Carpenier 3, J.-Ph. Chancelier 2, M. De Lara 2 EFFICACITY 1 CERMICS, ENPC 2 UMA, ENSTA 3 LISIS, IFSTTAR

More information

Solutions from Chapter 9.1 and 9.2

Solutions from Chapter 9.1 and 9.2 Soluions from Chaper 9 and 92 Secion 9 Problem # This basically boils down o an exercise in he chain rule from calculus We are looking for soluions of he form: u( x) = f( k x c) where k x R 3 and k is

More information

Deep Learning: Theory, Techniques & Applications - Recurrent Neural Networks -

Deep Learning: Theory, Techniques & Applications - Recurrent Neural Networks - Deep Learning: Theory, Techniques & Applicaions - Recurren Neural Neworks - Prof. Maeo Maeucci maeo.maeucci@polimi.i Deparmen of Elecronics, Informaion and Bioengineering Arificial Inelligence and Roboics

More information

Unit 1 Test Review Physics Basics, Movement, and Vectors Chapters 1-3

Unit 1 Test Review Physics Basics, Movement, and Vectors Chapters 1-3 A.P. Physics B Uni 1 Tes Reiew Physics Basics, Moemen, and Vecors Chapers 1-3 * In sudying for your es, make sure o sudy his reiew shee along wih your quizzes and homework assignmens. Muliple Choice Reiew:

More information

Particle Swarm Optimization

Particle Swarm Optimization Paricle Swarm Opimizaion Speaker: Jeng-Shyang Pan Deparmen of Elecronic Engineering, Kaohsiung Universiy of Applied Science, Taiwan Email: jspan@cc.kuas.edu.w 7/26/2004 ppso 1 Wha is he Paricle Swarm Opimizaion

More information

Errata (1 st Edition)

Errata (1 st Edition) P Sandborn, os Analysis of Elecronic Sysems, s Ediion, orld Scienific, Singapore, 03 Erraa ( s Ediion) S K 05D Page 8 Equaion (7) should be, E 05D E Nu e S K he L appearing in he equaion in he book does

More information

Synapses with short-term plasticity are optimal estimators of presynaptic membrane potentials: supplementary note

Synapses with short-term plasticity are optimal estimators of presynaptic membrane potentials: supplementary note Synapses wih shor-erm plasiciy are opimal esimaors of presynapic membrane poenials: supplemenary noe Jean-Pascal Pfiser, Peer Dayan, Máé Lengyel Supplemenary Noe 1 The local possynapic poenial In he main

More information

Lecture 3: Exponential Smoothing

Lecture 3: Exponential Smoothing NATCOR: Forecasing & Predicive Analyics Lecure 3: Exponenial Smoohing John Boylan Lancaser Cenre for Forecasing Deparmen of Managemen Science Mehods and Models Forecasing Mehod A (numerical) procedure

More information

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t...

t is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t... Mah 228- Fri Mar 24 5.6 Marix exponenials and linear sysems: The analogy beween firs order sysems of linear differenial equaions (Chaper 5) and scalar linear differenial equaions (Chaper ) is much sronger

More information

This document was generated at 7:34 PM, 07/27/09 Copyright 2009 Richard T. Woodward

This document was generated at 7:34 PM, 07/27/09 Copyright 2009 Richard T. Woodward his documen was generaed a 7:34 PM, 07/27/09 Copyrigh 2009 Richard. Woodward 15. Bang-bang and mos rapid approach problems AGEC 637 - Summer 2009 here are some problems for which he opimal pah does no

More information

Unit Root Time Series. Univariate random walk

Unit Root Time Series. Univariate random walk Uni Roo ime Series Univariae random walk Consider he regression y y where ~ iid N 0, he leas squares esimae of is: ˆ yy y y yy Now wha if = If y y hen le y 0 =0 so ha y j j If ~ iid N 0, hen y ~ N 0, he

More information

L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms

L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS NA568 Mobile Roboics: Mehods & Algorihms Today s Topic Quick review on (Linear) Kalman Filer Kalman Filering for Non-Linear Sysems Exended Kalman Filer (EKF)

More information

RC, RL and RLC circuits

RC, RL and RLC circuits Name Dae Time o Complee h m Parner Course/ Secion / Grade RC, RL and RLC circuis Inroducion In his experimen we will invesigae he behavior of circuis conaining combinaions of resisors, capaciors, and inducors.

More information

Empirical Process Theory

Empirical Process Theory Empirical Process heory 4.384 ime Series Analysis, Fall 27 Reciaion by Paul Schrimpf Supplemenary o lecures given by Anna Mikusheva Ocober 7, 28 Reciaion 7 Empirical Process heory Le x be a real-valued

More information