Fundamentals of Computational Neuroscience 2e

Size: px
Start display at page:

Download "Fundamentals of Computational Neuroscience 2e"

Transcription

1 Fundamentals of Computational Neuroscience 2e Thomas Trappenberg March 21, 2009 Chapter 9: Modular networks, motor control, and reinforcement learning

2 Mixture of experts Expert 1 Input Expert 2 Integration network Output Expert n Gating network A. Absolute function f (x ) = abs (x ) B. Mixture of expert for absolute function X ΣΠ abs(x ) x

3 The what-and-where task B. Without bias towards short connections A. Model retina with sample image 5 4 Output node # Output node # Hidden node # C. With bias towards short connections Hidden node # Jacobs and Jordan (1992)

4 Coupled attractor networks A. Coupled attractor networks B. The left-right universe with letters Node group 1 Node group 2 Connections between groups

5 Limit on modularity Load capacity α : c m = 2 A. Load capacity B. Bounds on intermodular strength m = 4 m = g : Relative intermodular strength g : Relative intramodular strength m : Number of modules

6 Sequence learning A. Modular attractor model B. Time evolution of overlaps Input Pathway w AB w AA w w BB BA Overlap in A Overlap in B Module A Module B Time [τ] Lawrence, Trappenberg and Fine (2006); (Sommer and Wennekers (2005))

7 Working memory Prefrontal cortex Parietal motor cortex Hippocampus O Reilly, Braver, and Cohen 1999

8 Limit on working memory B. Two object C. Four object Time 20 Node number 120 Node number Node number A. One object Time Time 20

9 Motor learning and control Disturbance Desired state - Motor command generator Motor command Controlled object Actual state Sensory system Afferent Re-afferent

10 Forward model controller Disturbance Desired state - Motor command generator Motor command Controlled object Actual state Sensory system + Forward dynamic model Forward output model - - Afferent Re-afferent

11 Inverse model controller Inverse model Disturbance Desired state - Motor command generator - + Motor command Controlled object Actual state Sensory system Afferent Re-afferent

12 Cerebellum Stellate cell Parallel fibre Molecular layer Purkinje layer Granular layer { { { Climbing fibre Purkinje neuron Granule cell Basket cell Golgi cell Intracerebellar and vestibular nuclei Inferior olive Out Mossy fibre Excitatory synapse Inhibitory synapse Spinal cord External cuneate nucleus { Reticular nuclei Pontine nuclei

13 Reinforcement learning

14 Basal Ganglia A. Outline of basic BG anatomy Cerebral cortex C. Recordings of SNc neurons and simulations Caudate nucleus Putamen Thalamus Stimulus A No reward Globus pallidus Superior colliculus Subthalamic nucleus Substantia nigra pars compacta ( pars reticulata ) Stimulus B Stimulus A Reward rhat Pattern 4 Pattern 3 Pattern Episode

15 temporal difference learning A. Linear predictor node B. Temporal delta rule C. Temporal difference rule in in in r 1 (t) r 2 (t) r 3 (t) r 4 in (t) in in in r 1 (t) r 2 (t) r 3 (t) r 4 in (t) in in in r 1 (t) r 2 (t) r 3 (t) r 4 in (t) V(t) r j in (t-1) slow V(t) V(t 1) r (t) in (t-1) r j V(t) slow V(t 1) fast γ V(t) r (t) r (t) r (t)

16 Actor-critique and Q-learning (frontal) TH B. Actor-critic model of BG Cerebral cortex F C C C D. Q-learning model of BG state action Cerebral cortex state / action coding Matrix module ST (actor) SPm SPs PD DA Basal ganglia ST Striosomal module (critic) SNc Striatum reward prediction Pallidum action selection Thalamus Primary reinforcement Primary reinforcement

17 Actor-critique controller Critic Reinforcement signal Disturbance Desired state - Motor command generator (actor) Motor command Controlled object Actual state Sensory system Afferent Re-afferent

18 Further Readings Robert A. Jacobs, Michael I. Jordan, and Andrew G. Barto (1991), Task decomposition through competition in a modular connectionist architecture: the what and where tasks, in Cognitive Science 15: Geoffrey Hinton (1999), Products of experts, in Proceedings of the Ninth International Conference on Artificial Neural Networks, ICANN 99, 1:1 6. Yaneer Bar-Yam (1997), Dynamics of complex systems, Addison-Wesley. Edmund T. Rolls and Simon M. Stringer (1999), A model of the interaction between mood and memory, in Networks: Comptutation in neural systems 12: N. J. Nilsson (1965), Learning machines: foundations of trainable pattern-classifying systems, McGraw-Hill. O. G. Selfridge (1958), Pandemonium: a paradigm of learning, in the mechanization of thought processes, in Proceedings of a Symposium Held at the National Physical Laboratory, November 1958, , London HMSO. Marvin Minsky (1986), The society of mind, Simon & Schuster. Akira Miyake and Priti Shah (eds.) (1999), Models of working memory, Cambridge University Press. Daniel M. Wolpert, R. Chris Miall, and Mitsuo Kawato (1998), Internal models in the cerebellum, in Trends Cognitive Science 2: Edmund T. Rolls and Alessandro Treves (1998), Neural networks and brain function, Oxford University Press. James C. Houk, Joel L. Davis, and David G. Beiser (eds.) (1995), Models of information processing in the basal ganglia, MIT Press. Richard S. Sutton and Andrew G. Barto (1998), Reinforcement learning: an introduction, MIT Press. Peter Dayan and Laurence F. Abbott (2001), Theoretical Neuroscience, MIT Press.

Fundamentals of Computational Neuroscience 2e

Fundamentals of Computational Neuroscience 2e Fundamentals of Computational Neuroscience 2e January 1, 2010 Chapter 10: The cognitive brain Hierarchical maps and attentive vision A. Ventral visual pathway B. Layered cortical maps Receptive field size

More information

Reinforcement learning

Reinforcement learning einforcement learning How to learn to make decisions in sequential problems (like: chess, a maze) Why is this difficult? Temporal credit assignment Prediction can help Further reading For modeling: Chapter

More information

Basal Ganglia. Input: Cortex Output: Thalamus motor and sensory gajng. Striatum Caudate Nucleus Putamen Globus Pallidus

Basal Ganglia. Input: Cortex Output: Thalamus motor and sensory gajng. Striatum Caudate Nucleus Putamen Globus Pallidus An Interac*ve Channel Model of the Basal Ganglia: Bifurca*on Analysis Under Healthy and Parkinsonian Condi*ons Robert Merrison- Hort, Nada Yousif, Felix Njap, Ulrich G. Hofmann, Oleksandr Burylko, Roman

More information

SUPPLEMENTARY MATERIAL. Patient Group Hem. Lesion Description Etiology FRONTAL LOBE GROUP 1 FL-LAT right SFG (pos), SFS, Central sulcus

SUPPLEMENTARY MATERIAL. Patient Group Hem. Lesion Description Etiology FRONTAL LOBE GROUP 1 FL-LAT right SFG (pos), SFS, Central sulcus 1 SUPPLEMENTARY MATERIAL Table S1: Lesion locations of the Frontal Lobe (FL), Basal Ganglia (BG) and Parietal-temporal (PTL) clinical groups. Lesion site was determined by (T1 and T2 weighted) anatomical

More information

Supervised Learning in Neural Networks

Supervised Learning in Neural Networks The Norwegian University of Science and Technology (NTNU Trondheim, Norway keithd@idi.ntnu.no March 7, 2011 Supervised Learning Constant feedback from an instructor, indicating not only right/wrong, but

More information

Reinforcement Learning. Odelia Schwartz 2016

Reinforcement Learning. Odelia Schwartz 2016 Reinforcement Learning Odelia Schwartz 2016 Forms of learning? Forms of learning Unsupervised learning Supervised learning Reinforcement learning Forms of learning Unsupervised learning Supervised learning

More information

I D I A P. Online Policy Adaptation for Ensemble Classifiers R E S E A R C H R E P O R T. Samy Bengio b. Christos Dimitrakakis a IDIAP RR 03-69

I D I A P. Online Policy Adaptation for Ensemble Classifiers R E S E A R C H R E P O R T. Samy Bengio b. Christos Dimitrakakis a IDIAP RR 03-69 R E S E A R C H R E P O R T Online Policy Adaptation for Ensemble Classifiers Christos Dimitrakakis a IDIAP RR 03-69 Samy Bengio b I D I A P December 2003 D a l l e M o l l e I n s t i t u t e for Perceptual

More information

Nervous Tissue. Neurons Electrochemical Gradient Propagation & Transduction Neurotransmitters Temporal & Spatial Summation

Nervous Tissue. Neurons Electrochemical Gradient Propagation & Transduction Neurotransmitters Temporal & Spatial Summation Nervous Tissue Neurons Electrochemical Gradient Propagation & Transduction Neurotransmitters Temporal & Spatial Summation What is the function of nervous tissue? Maintain homeostasis & respond to stimuli

More information

Computational Neuroscience

Computational Neuroscience Computational Neuroscience Zoltán Somogyvári senior research fellow Wigner Research Institute for Physics, Theoretical Department Supporting materials: http://www.kfki.hu/~soma/bscs/ BSCS 2012 Lengyel

More information

COMPUTATIONAL MODELLING OF THE HUMAN MOTOR CONTROL SYSTEM

COMPUTATIONAL MODELLING OF THE HUMAN MOTOR CONTROL SYSTEM COMPUTATIONAL MODELLING OF THE HUMAN MOTOR CONTROL SYSTEM Nonlinear Enhancement of the Adaptive Model Theory through Simulation and Experiment A thesis submitted in partial fulftlment of the requirements

More information

Nervous Tissue. Neurons Neural communication Nervous Systems

Nervous Tissue. Neurons Neural communication Nervous Systems Nervous Tissue Neurons Neural communication Nervous Systems What is the function of nervous tissue? Maintain homeostasis & respond to stimuli Sense & transmit information rapidly, to specific cells and

More information

Neural Systems and Artificial Life Group, Institute of Psychology, National Research Council, Rome. Evolving Modular Architectures for Neural Networks

Neural Systems and Artificial Life Group, Institute of Psychology, National Research Council, Rome. Evolving Modular Architectures for Neural Networks Neural Systems and Artificial Life Group, Institute of Psychology, National Research Council, Rome Evolving Modular Architectures for Neural Networks Andrea Di Ferdinando, Raffaele Calabretta and Domenico

More information

Intro and Homeostasis

Intro and Homeostasis Intro and Homeostasis Physiology - how the body works. Homeostasis - staying the same. Functional Types of Neurons Sensory (afferent - coming in) neurons: Detects the changes in the body. Informations

More information

Modelling stochastic neural learning

Modelling stochastic neural learning Modelling stochastic neural learning Computational Neuroscience András Telcs telcs.andras@wigner.mta.hu www.cs.bme.hu/~telcs http://pattern.wigner.mta.hu/participants/andras-telcs Compiled from lectures

More information

to appear in Frank Eeckman, ed., Proceedings of the Computational Neurosciences Conference CNS*93, Kluwer Academic Press.

to appear in Frank Eeckman, ed., Proceedings of the Computational Neurosciences Conference CNS*93, Kluwer Academic Press. to appear in Frank Eeckman, ed., Proceedings of the Computational Neurosciences Conference CNS*93, Kluwer Academic Press. The Sinusoidal Array: A Theory of Representation for Spatial Vectors A. David Redish,

More information

Artificial Neural Network and Fuzzy Logic

Artificial Neural Network and Fuzzy Logic Artificial Neural Network and Fuzzy Logic 1 Syllabus 2 Syllabus 3 Books 1. Artificial Neural Networks by B. Yagnanarayan, PHI - (Cover Topologies part of unit 1 and All part of Unit 2) 2. Neural Networks

More information

How to build a brain. Cognitive Modelling. Terry & Chris Centre for Theoretical Neuroscience University of Waterloo

How to build a brain. Cognitive Modelling. Terry & Chris Centre for Theoretical Neuroscience University of Waterloo How to build a brain Cognitive Modelling Terry & Chris Centre for Theoretical Neuroscience University of Waterloo So far... We have learned how to implement high-dimensional vector representations linear

More information

Brains and Computation

Brains and Computation 15-883: Computational Models of Neural Systems Lecture 1.1: Brains and Computation David S. Touretzky Computer Science Department Carnegie Mellon University 1 Models of the Nervous System Hydraulic network

More information

Supplementary Information. Brain networks involved in tactile speed classification of moving dot patterns: the. effects of speed and dot periodicity

Supplementary Information. Brain networks involved in tactile speed classification of moving dot patterns: the. effects of speed and dot periodicity Supplementary Information Brain networks involved in tactile speed classification of moving dot patterns: the effects of speed and dot periodicity Jiajia Yang, Ryo Kitada *, Takanori Kochiyama, Yinghua

More information

Rewards and Value Systems: Conditioning and Reinforcement Learning

Rewards and Value Systems: Conditioning and Reinforcement Learning Rewards and Value Systems: Conditioning and Reinforcement Learning Acknowledgements: Many thanks to P. Dayan and L. Abbott for making the figures of their book available online. Also thanks to R. Sutton

More information

Leo Kadanoff and 2d XY Models with Symmetry-Breaking Fields. renormalization group study of higher order gradients, cosines and vortices

Leo Kadanoff and 2d XY Models with Symmetry-Breaking Fields. renormalization group study of higher order gradients, cosines and vortices Leo Kadanoff and d XY Models with Symmetry-Breaking Fields renormalization group study of higher order gradients, cosines and vortices Leo Kadanoff and Random Matrix Theory Non-Hermitian Localization in

More information

COMPUTATIONAL ROLE OF DISINHIBITION IN BRAIN FUNCTION. A Dissertation YINGWEI YU

COMPUTATIONAL ROLE OF DISINHIBITION IN BRAIN FUNCTION. A Dissertation YINGWEI YU COMPUTATIONAL ROLE OF DISINHIBITION IN BRAIN FUNCTION A Dissertation by YINGWEI YU Submitted to the Office of Graduate Studies of Texas A&M University in partial fulfillment of the requirements for the

More information

CMSC 421: Neural Computation. Applications of Neural Networks

CMSC 421: Neural Computation. Applications of Neural Networks CMSC 42: Neural Computation definition synonyms neural networks artificial neural networks neural modeling connectionist models parallel distributed processing AI perspective Applications of Neural Networks

More information

Neurochemistry 1. Nervous system is made of neurons & glia, as well as other cells. Santiago Ramon y Cajal Nobel Prize 1906

Neurochemistry 1. Nervous system is made of neurons & glia, as well as other cells. Santiago Ramon y Cajal Nobel Prize 1906 Neurochemistry 1 Nervous system is made of neurons & glia, as well as other cells. Santiago Ramon y Cajal Nobel Prize 1906 How Many Neurons Do We Have? The human brain contains ~86 billion neurons and

More information

Introduction and Perceptron Learning

Introduction and Perceptron Learning Artificial Neural Networks Introduction and Perceptron Learning CPSC 565 Winter 2003 Christian Jacob Department of Computer Science University of Calgary Canada CPSC 565 - Winter 2003 - Emergent Computing

More information

Decision-making and Weber s law: a neurophysiological model

Decision-making and Weber s law: a neurophysiological model European Journal of Neuroscience, Vol. 24, pp. 901 916, 2006 doi:10.1111/j.14-9568.2006.04940.x Decision-making and Weber s law: a neurophysiological model Gustavo Deco 1 and Edmund T. Rolls 2 1 Institucio

More information

Processing of Time Series by Neural Circuits with Biologically Realistic Synaptic Dynamics

Processing of Time Series by Neural Circuits with Biologically Realistic Synaptic Dynamics Processing of Time Series by Neural Circuits with iologically Realistic Synaptic Dynamics Thomas Natschläger & Wolfgang Maass Institute for Theoretical Computer Science Technische Universität Graz, ustria

More information

Christopher Watkins and Peter Dayan. Noga Zaslavsky. The Hebrew University of Jerusalem Advanced Seminar in Deep Learning (67679) November 1, 2015

Christopher Watkins and Peter Dayan. Noga Zaslavsky. The Hebrew University of Jerusalem Advanced Seminar in Deep Learning (67679) November 1, 2015 Q-Learning Christopher Watkins and Peter Dayan Noga Zaslavsky The Hebrew University of Jerusalem Advanced Seminar in Deep Learning (67679) November 1, 2015 Noga Zaslavsky Q-Learning (Watkins & Dayan, 1992)

More information

The Variance of Covariance Rules for Associative Matrix Memories and Reinforcement Learning

The Variance of Covariance Rules for Associative Matrix Memories and Reinforcement Learning NOTE Communicated by David Willshaw The Variance of Covariance Rules for Associative Matrix Memories and Reinforcement Learning Peter Dayan Terrence J. Sejnowski Computational Neurobiology Laboratory,

More information

The wake-sleep algorithm for unsupervised neural networks

The wake-sleep algorithm for unsupervised neural networks The wake-sleep algorithm for unsupervised neural networks Geoffrey E Hinton Peter Dayan Brendan J Frey Radford M Neal Department of Computer Science University of Toronto 6 King s College Road Toronto

More information

Biologically Plausible Local Learning Rules for the Adaptation of the Vestibulo-Ocular Reflex

Biologically Plausible Local Learning Rules for the Adaptation of the Vestibulo-Ocular Reflex Biologically Plausible Local Learning Rules for the Adaptation of the Vestibulo-Ocular Reflex Olivier Coenen* Terrence J. Sejnowski Computational Neurobiology Laboratory Howard Hughes Medical Institute

More information

Tuning tuning curves. So far: Receptive fields Representation of stimuli Population vectors. Today: Contrast enhancment, cortical processing

Tuning tuning curves. So far: Receptive fields Representation of stimuli Population vectors. Today: Contrast enhancment, cortical processing Tuning tuning curves So far: Receptive fields Representation of stimuli Population vectors Today: Contrast enhancment, cortical processing Firing frequency N 3 s max (N 1 ) = 40 o N4 N 1 N N 5 2 s max

More information

Replacing eligibility trace for action-value learning with function approximation

Replacing eligibility trace for action-value learning with function approximation Replacing eligibility trace for action-value learning with function approximation Kary FRÄMLING Helsinki University of Technology PL 5500, FI-02015 TKK - Finland Abstract. The eligibility trace is one

More information

Using Expectation-Maximization for Reinforcement Learning

Using Expectation-Maximization for Reinforcement Learning NOTE Communicated by Andrew Barto and Michael Jordan Using Expectation-Maximization for Reinforcement Learning Peter Dayan Department of Brain and Cognitive Sciences, Center for Biological and Computational

More information

Marr's Theory of the Hippocampus: Part I

Marr's Theory of the Hippocampus: Part I Marr's Theory of the Hippocampus: Part I Computational Models of Neural Systems Lecture 3.3 David S. Touretzky October, 2015 David Marr: 1945-1980 10/05/15 Computational Models of Neural Systems 2 Marr

More information

Backpropagation: The Good, the Bad and the Ugly

Backpropagation: The Good, the Bad and the Ugly Backpropagation: The Good, the Bad and the Ugly The Norwegian University of Science and Technology (NTNU Trondheim, Norway keithd@idi.ntnu.no October 3, 2017 Supervised Learning Constant feedback from

More information

Animal learning theory

Animal learning theory Animal learning theory Based on [Sutton and Barto, 1990, Dayan and Abbott, 2001] Bert Kappen [Sutton and Barto, 1990] Classical conditioning: - A conditioned stimulus (CS) and unconditioned stimulus (US)

More information

Input areas Cortex/thalamus/limbic. Striatum SNc STN. GPe. BG output. Excitatory Inhibitory. Dopaminergic

Input areas Cortex/thalamus/limbic. Striatum SNc STN. GPe. BG output. Excitatory Inhibitory. Dopaminergic Analysis and simulation of a model of intrinsic processing in the Basal Ganglia K. Gurney, P. Redgrave, A. Prescott Dept Psychology, University of Sheffield, Sheffield S10 2TP Abstract We have recently

More information

How to make computers work like the brain

How to make computers work like the brain How to make computers work like the brain (without really solving the brain) Dileep George a single special machine can be made to do the work of all. It could in fact be made to work as a model of any

More information

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5 Artificial Neural Networks Q550: Models in Cognitive Science Lecture 5 "Intelligence is 10 million rules." --Doug Lenat The human brain has about 100 billion neurons. With an estimated average of one thousand

More information

Temporal Difference Learning & Policy Iteration

Temporal Difference Learning & Policy Iteration Temporal Difference Learning & Policy Iteration Advanced Topics in Reinforcement Learning Seminar WS 15/16 ±0 ±0 +1 by Tobias Joppen 03.11.2015 Fachbereich Informatik Knowledge Engineering Group Prof.

More information

Effect of number of hidden neurons on learning in large-scale layered neural networks

Effect of number of hidden neurons on learning in large-scale layered neural networks ICROS-SICE International Joint Conference 009 August 18-1, 009, Fukuoka International Congress Center, Japan Effect of on learning in large-scale layered neural networks Katsunari Shibata (Oita Univ.;

More information

MOTOR CONTROL OF A LIMB SEGMENT ACTUATED BY ARTIFICIAL MUSCLES

MOTOR CONTROL OF A LIMB SEGMENT ACTUATED BY ARTIFICIAL MUSCLES Proceedings 3rd nnual Conference IEEE/EMBS Oct.5-8, 00, Istanbul, TURKEY of 4 MOTOR CONTROL OF LIMB SEGMENT CTUTE BY RTIFICIL MUSCLES S. Eskiizmirliler, 3, B. Tondu, C. arlot 3 SS, epartment of Technology,

More information

In the Name of God. Lecture 9: ANN Architectures

In the Name of God. Lecture 9: ANN Architectures In the Name of God Lecture 9: ANN Architectures Biological Neuron Organization of Levels in Brains Central Nervous sys Interregional circuits Local circuits Neurons Dendrite tree map into cerebral cortex,

More information

Neuroinformatics. Marcus Kaiser. Week 10: Cortical maps and competitive population coding (textbook chapter 7)!

Neuroinformatics. Marcus Kaiser. Week 10: Cortical maps and competitive population coding (textbook chapter 7)! 0 Neuroinformatics Marcus Kaiser Week 10: Cortical maps and competitive population coding (textbook chapter 7)! Outline Topographic maps Self-organizing maps Willshaw & von der Malsburg Kohonen Dynamic

More information

Reciprocity Between the Cerebellum and the Cerebral Cortex: Nonlinear Dynamics in Microscopic Modules for Generating Voluntary Motor Commands

Reciprocity Between the Cerebellum and the Cerebral Cortex: Nonlinear Dynamics in Microscopic Modules for Generating Voluntary Motor Commands Reciprocity Between the Cerebellum and the Cerebral Cortex: Nonlinear Dynamics in Microscopic Modules for Generating Voluntary Motor Commands JUN WANG, 1 GREGORY DAM, 2 SULE YILDIRIM, 3 WILLIAM RAND, 4

More information

MULTISCALE MODULARITY IN BRAIN SYSTEMS

MULTISCALE MODULARITY IN BRAIN SYSTEMS MULTISCALE MODULARITY IN BRAIN SYSTEMS Danielle S. Bassett University of California Santa Barbara Department of Physics The Brain: A Multiscale System Spatial Hierarchy: Temporal-Spatial Hierarchy: http://www.idac.tohoku.ac.jp/en/frontiers/column_070327/figi-i.gif

More information

Artificial Neural Networks

Artificial Neural Networks Artificial Neural Networks CPSC 533 Winter 2 Christian Jacob Neural Networks in the Context of AI Systems Neural Networks as Mediators between Symbolic AI and Statistical Methods 2 5.-NeuralNets-2.nb Neural

More information

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA   1/ 21 Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural

More information

Lecture 4: Feed Forward Neural Networks

Lecture 4: Feed Forward Neural Networks Lecture 4: Feed Forward Neural Networks Dr. Roman V Belavkin Middlesex University BIS4435 Biological neurons and the brain A Model of A Single Neuron Neurons as data-driven models Neural Networks Training

More information

Learning a world model and planning with a self-organizing, dynamic neural system

Learning a world model and planning with a self-organizing, dynamic neural system Learning a world model and planning with a self-organizing, dynamic neural system Marc Toussaint Institut für Neuroinformatik Ruhr-Universität Bochum, ND 04 44780 Bochum Germany mt@neuroinformatik.rub.de

More information

Neurons, Synapses, and Signaling

Neurons, Synapses, and Signaling LECTURE PRESENTATIONS For CAMPBELL BIOLOGY, NINTH EDITION Jane B. Reece, Lisa A. Urry, Michael L. Cain, Steven A. Wasserman, Peter V. Minorsky, Robert B. Jackson Chapter 48 Neurons, Synapses, and Signaling

More information

CORRELATION TRANSFER FROM BASAL GANGLIA TO THALAMUS IN PARKINSON S DISEASE. by Pamela Reitsma. B.S., University of Maine, 2007

CORRELATION TRANSFER FROM BASAL GANGLIA TO THALAMUS IN PARKINSON S DISEASE. by Pamela Reitsma. B.S., University of Maine, 2007 CORRELATION TRANSFER FROM BASAL GANGLIA TO THALAMUS IN PARKINSON S DISEASE by Pamela Reitsma B.S., University of Maine, 27 Submitted to the Graduate Faculty of the Department of Mathematics in partial

More information

Ângelo Cardoso 27 May, Symbolic and Sub-Symbolic Learning Course Instituto Superior Técnico

Ângelo Cardoso 27 May, Symbolic and Sub-Symbolic Learning Course Instituto Superior Técnico BIOLOGICALLY INSPIRED COMPUTER MODELS FOR VISUAL RECOGNITION Ângelo Cardoso 27 May, 2010 Symbolic and Sub-Symbolic Learning Course Instituto Superior Técnico Index Human Vision Retinal Ganglion Cells Simple

More information

Hierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References

Hierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References 24th March 2011 Update Hierarchical Model Rao and Ballard (1999) presented a hierarchical model of visual cortex to show how classical and extra-classical Receptive Field (RF) effects could be explained

More information

Signal, donnée, information dans les circuits de nos cerveaux

Signal, donnée, information dans les circuits de nos cerveaux NeuroSTIC Brest 5 octobre 2017 Signal, donnée, information dans les circuits de nos cerveaux Claude Berrou Signal, data, information: in the field of telecommunication, everything is clear It is much less

More information

COMP9444 Neural Networks and Deep Learning 2. Perceptrons. COMP9444 c Alan Blair, 2017

COMP9444 Neural Networks and Deep Learning 2. Perceptrons. COMP9444 c Alan Blair, 2017 COMP9444 Neural Networks and Deep Learning 2. Perceptrons COMP9444 17s2 Perceptrons 1 Outline Neurons Biological and Artificial Perceptron Learning Linear Separability Multi-Layer Networks COMP9444 17s2

More information

Computational Neuroscience. Structure Dynamics Implementation Algorithm Computation - Function

Computational Neuroscience. Structure Dynamics Implementation Algorithm Computation - Function Computational Neuroscience Structure Dynamics Implementation Algorithm Computation - Function Learning at psychological level Classical conditioning Hebb's rule When an axon of cell A is near enough to

More information

Finding a Basis for the Neural State

Finding a Basis for the Neural State Finding a Basis for the Neural State Chris Cueva ccueva@stanford.edu I. INTRODUCTION How is information represented in the brain? For example, consider arm movement. Neurons in dorsal premotor cortex (PMd)

More information

arxiv:physics/ v1 [physics.bio-ph] 19 Feb 1999

arxiv:physics/ v1 [physics.bio-ph] 19 Feb 1999 Odor recognition and segmentation by coupled olfactory bulb and cortical networks arxiv:physics/9902052v1 [physics.bioph] 19 Feb 1999 Abstract Zhaoping Li a,1 John Hertz b a CBCL, MIT, Cambridge MA 02139

More information

Nervous System Organization

Nervous System Organization The Nervous System Nervous System Organization Receptors respond to stimuli Sensory receptors detect the stimulus Motor effectors respond to stimulus Nervous system divisions Central nervous system Command

More information

Outline. NIP: Hebbian Learning. Overview. Types of Learning. Neural Information Processing. Amos Storkey

Outline. NIP: Hebbian Learning. Overview. Types of Learning. Neural Information Processing. Amos Storkey Outline NIP: Hebbian Learning Neural Information Processing Amos Storkey 1/36 Overview 2/36 Types of Learning Types of learning, learning strategies Neurophysiology, LTP/LTD Basic Hebb rule, covariance

More information

McGill University > Schulich School of Music > MUMT 611 > Presentation III. Neural Networks. artificial. jason a. hockman

McGill University > Schulich School of Music > MUMT 611 > Presentation III. Neural Networks. artificial. jason a. hockman jason a. hockman Overvie hat is a neural netork? basics and architecture learning applications in music History 1940s: William McCulloch defines neuron 1960s: Perceptron 1970s: limitations presented (Minsky)

More information

Does the Wake-sleep Algorithm Produce Good Density Estimators?

Does the Wake-sleep Algorithm Produce Good Density Estimators? Does the Wake-sleep Algorithm Produce Good Density Estimators? Brendan J. Frey, Geoffrey E. Hinton Peter Dayan Department of Computer Science Department of Brain and Cognitive Sciences University of Toronto

More information

A Spiking Independent Accumulator Model for Winner-Take-All Computation

A Spiking Independent Accumulator Model for Winner-Take-All Computation A Spiking Independent Accumulator Model for Winner-Take-All Computation Jan Gosmann (jgosmann@uwaterloo.ca) Aaron R. Voelker (arvoelke@uwaterloo.ca) Chris Eliasmith (celiasmith@uwaterloo.ca) Centre for

More information

Sections 18.6 and 18.7 Artificial Neural Networks

Sections 18.6 and 18.7 Artificial Neural Networks Sections 18.6 and 18.7 Artificial Neural Networks CS4811 - Artificial Intelligence Nilufer Onder Department of Computer Science Michigan Technological University Outline The brain vs. artifical neural

More information

Back-propagation as reinforcement in prediction tasks

Back-propagation as reinforcement in prediction tasks Back-propagation as reinforcement in prediction tasks André Grüning Cognitive Neuroscience Sector S.I.S.S.A. via Beirut 4 34014 Trieste Italy gruening@sissa.it Abstract. The back-propagation (BP) training

More information

Neural networks. Chapter 20, Section 5 1

Neural networks. Chapter 20, Section 5 1 Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of

More information

A biologically realistic cleanup memory: Autoassociation in spiking neurons

A biologically realistic cleanup memory: Autoassociation in spiking neurons Available online at www.sciencedirect.com Cognitive Systems Research 12 (2011) 84 92 www.elsevier.com/locate/cogsys A biologically realistic cleanup memory: Autoassociation in spiking neurons Action editor:

More information

Do neural models scale up to a human brain?

Do neural models scale up to a human brain? Do neural models scale up to a human brain? Roman V. Belavkin (r.belavkin@mdx.ac.uk) School of Computing Science, Middlesex University, London NW4 4BT, UK August 27, 2007 oman Belavkin, Middlesex University,

More information

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 Multi-layer networks Steve Renals Machine Learning Practical MLP Lecture 3 7 October 2015 MLP Lecture 3 Multi-layer networks 2 What Do Single

More information

Instituto Tecnológico y de Estudios Superiores de Occidente Departamento de Electrónica, Sistemas e Informática. Introductory Notes on Neural Networks

Instituto Tecnológico y de Estudios Superiores de Occidente Departamento de Electrónica, Sistemas e Informática. Introductory Notes on Neural Networks Introductory Notes on Neural Networs Dr. José Ernesto Rayas Sánche April Introductory Notes on Neural Networs Dr. José Ernesto Rayas Sánche BIOLOGICAL NEURAL NETWORKS The brain can be seen as a highly

More information

Introduction to Neural Networks U. Minn. Psy 5038 Spring, 1999 Daniel Kersten. Lecture 2a. The Neuron - overview of structure. From Anderson (1995)

Introduction to Neural Networks U. Minn. Psy 5038 Spring, 1999 Daniel Kersten. Lecture 2a. The Neuron - overview of structure. From Anderson (1995) Introduction to Neural Networks U. Minn. Psy 5038 Spring, 1999 Daniel Kersten Lecture 2a The Neuron - overview of structure From Anderson (1995) 2 Lect_2a_Mathematica.nb Basic Structure Information flow:

More information

Spatial Representations in the Parietal Cortex May Use Basis Functions

Spatial Representations in the Parietal Cortex May Use Basis Functions Spatial Representations in the Parietal Cortex May Use Basis Functions Alexandre Pouget alex@salk.edu Terrence J. Sejnowski terry@salk.edu Howard Hughes Medical Institute The Salk Institute La Jolla, CA

More information

Introduction to Artificial Neural Networks

Introduction to Artificial Neural Networks Facultés Universitaires Notre-Dame de la Paix 27 March 2007 Outline 1 Introduction 2 Fundamentals Biological neuron Artificial neuron Artificial Neural Network Outline 3 Single-layer ANN Perceptron Adaline

More information

SUPPLEMENTARY INFORMATION

SUPPLEMENTARY INFORMATION SUPPLEMENTARY INFORMATION 1. Supplementary Tables 2. Supplementary Figures 1/12 Supplementary tables TABLE S1 Response to Expected Value: Anatomical locations of regions correlating with the expected value

More information

EE04 804(B) Soft Computing Ver. 1.2 Class 2. Neural Networks - I Feb 23, Sasidharan Sreedharan

EE04 804(B) Soft Computing Ver. 1.2 Class 2. Neural Networks - I Feb 23, Sasidharan Sreedharan EE04 804(B) Soft Computing Ver. 1.2 Class 2. Neural Networks - I Feb 23, 2012 Sasidharan Sreedharan www.sasidharan.webs.com 3/1/2012 1 Syllabus Artificial Intelligence Systems- Neural Networks, fuzzy logic,

More information

Introduction Biologically Motivated Crude Model Backpropagation

Introduction Biologically Motivated Crude Model Backpropagation Introduction Biologically Motivated Crude Model Backpropagation 1 McCulloch-Pitts Neurons In 1943 Warren S. McCulloch, a neuroscientist, and Walter Pitts, a logician, published A logical calculus of the

More information

CSE/NB 528 Final Lecture: All Good Things Must. CSE/NB 528: Final Lecture

CSE/NB 528 Final Lecture: All Good Things Must. CSE/NB 528: Final Lecture CSE/NB 528 Final Lecture: All Good Things Must 1 Course Summary Where have we been? Course Highlights Where do we go from here? Challenges and Open Problems Further Reading 2 What is the neural code? What

More information

Neural Networks: Introduction

Neural Networks: Introduction Neural Networks: Introduction Machine Learning Fall 2017 Based on slides and material from Geoffrey Hinton, Richard Socher, Dan Roth, Yoav Goldberg, Shai Shalev-Shwartz and Shai Ben-David, and others 1

More information

Response Selection Using Neural Phase Oscillators

Response Selection Using Neural Phase Oscillators Response Selection Using Neural Phase Oscillators J. Acacio de Barros 1 A symposium on the occasion of Patrick Suppes 90th birthday March 10, 2012 1 With Pat Suppes and Gary Oas (Suppes et al., 2012).

More information

CISC 3250 Systems Neuroscience

CISC 3250 Systems Neuroscience CISC 3250 Systems Neuroscience Systems Neuroscience How the nervous system performs computations How groups of neurons work together to achieve intelligence Professor Daniel Leeds dleeds@fordham.edu JMH

More information

Neural Nets in PR. Pattern Recognition XII. Michal Haindl. Outline. Neural Nets in PR 2

Neural Nets in PR. Pattern Recognition XII. Michal Haindl. Outline. Neural Nets in PR 2 Neural Nets in PR NM P F Outline Motivation: Pattern Recognition XII human brain study complex cognitive tasks Michal Haindl Faculty of Information Technology, KTI Czech Technical University in Prague

More information

Will Penny. 21st April The Macroscopic Brain. Will Penny. Cortical Unit. Spectral Responses. Macroscopic Models. Steady-State Responses

Will Penny. 21st April The Macroscopic Brain. Will Penny. Cortical Unit. Spectral Responses. Macroscopic Models. Steady-State Responses The The 21st April 2011 Jansen and Rit (1995), building on the work of Lopes Da Sliva and others, developed a biologically inspired model of EEG activity. It was originally developed to explain alpha activity

More information

Neurons and Nervous Systems

Neurons and Nervous Systems 34 Neurons and Nervous Systems Concept 34.1 Nervous Systems Consist of Neurons and Glia Nervous systems have two categories of cells: Neurons, or nerve cells, are excitable they generate and transmit electrical

More information

Model of a Biological Neuron as a Temporal Neural Network

Model of a Biological Neuron as a Temporal Neural Network Model of a Biological Neuron as a Temporal Neural Network Sean D. Murphy and Edward W. Kairiss Interdepartmental Neuroscience Program, Department of Psychology, and The Center for Theoretical and Applied

More information

Homeostatic plasticity improves signal propagation in. continuous-time recurrent neural networks

Homeostatic plasticity improves signal propagation in. continuous-time recurrent neural networks Homeostatic plasticity improves signal propagation in continuous-time recurrent neural networks Hywel Williams and Jason Noble {hywelw jasonn}@comp.leeds.ac.uk School of Computing, University of Leeds,

More information

An Introductory Course in Computational Neuroscience

An Introductory Course in Computational Neuroscience An Introductory Course in Computational Neuroscience Contents Series Foreword Acknowledgments Preface 1 Preliminary Material 1.1. Introduction 1.1.1 The Cell, the Circuit, and the Brain 1.1.2 Physics of

More information

Reinforcement Learning for Continuous. Action using Stochastic Gradient Ascent. Hajime KIMURA, Shigenobu KOBAYASHI JAPAN

Reinforcement Learning for Continuous. Action using Stochastic Gradient Ascent. Hajime KIMURA, Shigenobu KOBAYASHI JAPAN Reinforcement Learning for Continuous Action using Stochastic Gradient Ascent Hajime KIMURA, Shigenobu KOBAYASHI Tokyo Institute of Technology, 4259 Nagatsuda, Midori-ku Yokohama 226-852 JAPAN Abstract:

More information

Fast neural network simulations with population density methods

Fast neural network simulations with population density methods Fast neural network simulations with population density methods Duane Q. Nykamp a,1 Daniel Tranchina b,a,c,2 a Courant Institute of Mathematical Science b Department of Biology c Center for Neural Science

More information

Collecting Aligned Activity & Connectomic Data Example: Mouse Vibrissal Touch Barrel Cortex Exploiting Coherence to Reduce Dimensionality Example: C.

Collecting Aligned Activity & Connectomic Data Example: Mouse Vibrissal Touch Barrel Cortex Exploiting Coherence to Reduce Dimensionality Example: C. Collecting Aligned Activity & Connectomic Data Example: Mouse Vibrissal Touch Barrel Cortex Exploiting Coherence to Reduce Dimensionality Example: C. elegans Motor Control Sequence Spatially & Temporally

More information

Time Indexed Hierarchical Relative Entropy Policy Search

Time Indexed Hierarchical Relative Entropy Policy Search Time Indexed Hierarchical Relative Entropy Policy Search Florentin Mehlbeer June 19, 2013 1 / 15 Structure Introduction Reinforcement Learning Relative Entropy Policy Search Hierarchical Relative Entropy

More information

Introduction Principles of Signaling and Organization p. 3 Signaling in Simple Neuronal Circuits p. 4 Organization of the Retina p.

Introduction Principles of Signaling and Organization p. 3 Signaling in Simple Neuronal Circuits p. 4 Organization of the Retina p. Introduction Principles of Signaling and Organization p. 3 Signaling in Simple Neuronal Circuits p. 4 Organization of the Retina p. 5 Signaling in Nerve Cells p. 9 Cellular and Molecular Biology of Neurons

More information

REPORT NUMIUR Molecular & Cell Biology Af U49R Drivision of Neurobiology Berkeley, CA 94720

REPORT NUMIUR Molecular & Cell Biology Af U49R Drivision of Neurobiology Berkeley, CA 94720 AD-A247 860 TION PAGE 1 A &0f~a Oat SAA010N91OSU 4.74 INVESTIGATION OF DYNAMIC ALGORIYHMS FOR PATTERN RECOGNITION IDENTIFIED IN CEREBRAL CORTEXPE-612 AFOSR-88-0268 ILAUTNOA(S)TA-B Dr Walter J. Freeman.

More information

Nervous System Organization

Nervous System Organization The Nervous System Chapter 44 Nervous System Organization All animals must be able to respond to environmental stimuli -Sensory receptors = Detect stimulus -Motor effectors = Respond to it -The nervous

More information

Neural Networks and the Back-propagation Algorithm

Neural Networks and the Back-propagation Algorithm Neural Networks and the Back-propagation Algorithm Francisco S. Melo In these notes, we provide a brief overview of the main concepts concerning neural networks and the back-propagation algorithm. We closely

More information

A recurrent model of transformation invariance by association

A recurrent model of transformation invariance by association NN 1381 PERGAMON Neural Networks 13 (2000) 225 237 Contributed article A recurrent model of transformation invariance by association Martin C.M. Elliffe a, Edmund T. Rolls a,1,2, *,Néstor Parga b,2,3,

More information

PETER SCHUSTER IN THIS ISSUE

PETER SCHUSTER IN THIS ISSUE Vol 14-2-CPLX_Cover1 10/25/08 10:30 AM Page 1 Vol. 14/No. 2, November/December 2008 $15.95 PETER SCHUSTER IN THIS ISSUE ALFRED HÜBLER RECIPROCITY BETWEEN THE CEREBELLUM AND THE CEREBRAL CORTEX: NONLINEAR

More information

Sections 18.6 and 18.7 Analysis of Artificial Neural Networks

Sections 18.6 and 18.7 Analysis of Artificial Neural Networks Sections 18.6 and 18.7 Analysis of Artificial Neural Networks CS4811 - Artificial Intelligence Nilufer Onder Department of Computer Science Michigan Technological University Outline Univariate regression

More information

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28 1 / 28 Neural Networks Mark van Rossum School of Informatics, University of Edinburgh January 15, 2018 2 / 28 Goals: Understand how (recurrent) networks behave Find a way to teach networks to do a certain

More information