Synaptic plasticity in neuromorphic hardware. Stefano Fusi Columbia University
|
|
- Emerald Preston
- 5 years ago
- Views:
Transcription
1 Synaptic plasticity in neuromorphic hardware Stefano Fusi Columbia University
2 The memory problem Several efficient memory models assume that the synaptic dynamic variables are unbounded, or can be modified with arbitrary precision Neuromorphic hardware implementations impose constraints on the dynamics of the synaptic variables Formalize the problem and analyze different strategies
3 Formalizing the memory problem time (for example )
4 time w ij 1 w ij Memory strength
5 Example: the Hopfield model
6 w ij =0 p time w ij 1
7 w ij =0 p time w ij 1
8 w ij =0 w ij p
9 w ij =0 w ij p Memories are random and uncorrelated
10 w ij =0 w ij p Memories are random and uncorrelated
11 w ij =0 w ij p Memories are random and uncorrelated 1
12 w ij =0 w ij p Memories are random and uncorrelated 1 0
13 Memory noise
14 Memory noise
15 Memory noise Noise increases with the number of memories
16 Number of memories (or time) Hopfield 1982, Amit Gutfreund Sompolinsky, 1985
17 Blackout Catastrophe Hopfield 1982, Amit Gutfreund Sompolinsky 1985
18 toward more realistic synapses Unbounded
19 toward more realistic synapses Unbounded Bounded (binary), offline learning Sompolinsky 1986
20 toward more realistic synapses Unbounded Bounded (binary), offline learning Sompolinsky 1986 Bounded (binary), online learning
21 w time time time
22 A learning rule for binary synapses pre post w= w=-1 with probability q w= w=+1
23 Initial signal
24 Signal after p memories
25 Signal after p memories Noise (at equilibrium) Amit Fusi 1992, Amit Fusi 1994
26
27
28 Learning rate SNR(0) Number of memories FAST ~ 1 SLOW 0
29 Sparse representations =1 with probability f =0 with probability 1-f Learning rule pre post w=1 with probability q + q + = w=-1 with probability q - q - = f Tsodyks Feigelman 1989, Amit Fusi 1994
30 Learning rate SPARSE SNR(0) Number of memories SLOW Amit Fusi, Neural Computation, 1994
31 A significant improvement, but 1) Not robust to noise ( must be exactly 0). In the presence of noise: p ~ N syn 2) The amount of information per memory is significantly smaller (it scales like f) 3) Not scalable (for large N syn it is very difficult to readout the relevant info) Ben Dayan Rubin, Fusi, Frontiers in Comp. Neuroscience 2007
32 Learning systems with multiple timescales Fast SNR Slow p
33 FAST SLOW N syn /m N syn /m N syn /m N syn /m q 1 q 2 q m-1 m groups q m =q s q k =(q s ) (k-1)/(m-1)
34
35 SNR Learning rate SNR(0) Number of memories FAST SLOW HETEROGENEOUS Fusi, Abbott, Neuron 2005; Roxin Fusi, PLoS Comp Biol. 2013
36 SNR Learning rate SNR(0) Number of memories FAST SLOW HETEROGENEOUS Fusi, Abbott, Neuron 2005; Roxin Fusi, PLoS Comp Biol. 2013
37 SNR
38 N log(1/q s ) q s N q s N m e-q st q s N e -q st
39 Heterogeneous systems with memory transfer A. Roxin, S. Fusi, PLoS Comp. Biology 2013
40 FAST SLOW N/m N/m N/m N/m q 1 q 2 q m-1 m groups q m =q s q k =(q s ) k/m
41 FAST INPUT SLOW Memory transfer q 1 q 2 q m-1 q m =q s
42 FAST SLOW q 1 q 2 q m-1 q m =q s S 1 S 2 S m-1 S m Readout: S = max {S 1,,S m }
43 N log(1/q s ) N m q s N ~m 1/4 ~1/q s ~m/q s
44 Learning rate SNR(0) Number of memories FAST SLOW HETEROGENEOUS HETEROGENEOUS with MEMORY TRANSFER
45 Fusi, Annunziato, Badoni, Salamon, Amit, Neural Computation (2000)
46 Giacomo Indiveri+ Fabio Stefanini
47 Autonomous real-time associative learning of visual stimuli on chip Poster M. Giulioni M. Giulioni, F. Corradi, V. Dante, P. Del Giudice, in preparation Two simple visual stimuli are acquired by a silicon retina (INI-Zurich) Input to a recurrent network of spiking IF neurons and plastic Hebbian, stochastic synapses on two chips Mean-field theory (effective response function) to navigate the parameter space Autonomous Hebbian learning generates stimulus-selective attractor states with error correction properties
48 Conclusions Neuromorphic synapses that are bistable require special machinery for preventing catastrophic forgetting Two important principles to improve performance: 1) Heterogeneity (multiple timescales) 2) Efficient memory transfer
49 People Theory of memory Daniel Amit Nicolas Brunel Francesco Battaglia Francesco Carusi Walter Senn Larry Abbott Daniel Ben Dayan Rubin Alex Roxin Srdjan Ostojic Marcus Benna Synaptic dynamics Hardware implementation Roberto Riccardi Gaetano Salina Mario Annunziato Paolo Del Giudice Maurizio Mattia Davide Badoni Stefano Buglioni Vittorio Dante Giacomo Indiveri Srinjoy Mitra Elisabetta Chicca Fabio Stefanini Mario Annunziato Joseph Brader Walter Senn Mattia Rigotti Dani Marti Kyo Iigaya
50 Teacher Teacher LTP Long-Term Potentiation Teacher Teacher LTD Long-Term Depression
51 Fusi, Annunziato, Badoni, Salamon, Amit, Neural Computation (2000)
52 No modification Long term modification Teacher Teacher
53 Brader, Senn, Fusi, Neural Computation (2007)
A brain-inspired neuromorphic architecture for robust neural computation
A brain-inspired neuromorphic architecture for robust neural computation Fabio Stefanini and Giacomo Indiveri Institute of Neuroinformatics University of Zurich and ETH Zurich BIC Workshop @ ISCA40 June
More informationNeurophysiology of a VLSI spiking neural network: LANN21
Neurophysiology of a VLSI spiking neural network: LANN21 Stefano Fusi INFN, Sezione Roma I Università di Roma La Sapienza Pza Aldo Moro 2, I-185, Roma fusi@jupiter.roma1.infn.it Paolo Del Giudice Physics
More informationOptimizing One-Shot Learning with Binary Synapses
ARTICLE Communicated by Mark Goldman Optimizing One-Shot Learning with Binary Synapses Sandro Romani sandro.romani@gmail.com Human Physiology, UniversitàdiRomaLaSapienza,Rome00185,Italy Daniel J. Amit
More informationCharacterizing the firing properties of an adaptive
Characterizing the firing properties of an adaptive analog VLSI neuron Daniel Ben Dayan Rubin 1,2, Elisabetta Chicca 2, and Giacomo Indiveri 2 1 Department of Bioengineering, Politecnico di Milano, P.zza
More informationMemory capacity of networks with stochastic binary synapses
Memory capacity of networks with stochastic binary synapses Alexis M. Dubreuil 1,, Yali Amit 3 and Nicolas Brunel 1, 1 UMR 8118, CNRS, Université Paris Descartes, Paris, France Departments of Statistics
More informationContraction Properties of VLSI Cooperative Competitive Neural Networks of Spiking Neurons
Contraction Properties of VLSI Cooperative Competitive Neural Networks of Spiking s Emre Neftci 1, Elisabetta Chicca 1, Giacomo Indiveri 1, Jean-Jacques Slotine 2, Rodney Douglas 1 1 Institute of Neuroinformatics,
More informationContraction Properties of VLSI Cooperative Competitive Neural Networks of Spiking Neurons
Contraction Properties of VLSI Cooperative Competitive Neural Networks of Spiking s Emre Neftci 1, Elisabetta Chicca 1, Giacomo Indiveri 1, Jean-Jacques Slotine 2, Rodney Douglas 1 1 Institute of Neuroinformatics,
More informationNovel VLSI Implementation for Triplet-based Spike-Timing Dependent Plasticity
Novel LSI Implementation for Triplet-based Spike-Timing Dependent Plasticity Mostafa Rahimi Azghadi, Omid Kavehei, Said Al-Sarawi, Nicolangelo Iannella, and Derek Abbott Centre for Biomedical Engineering,
More informationThe Variance of Covariance Rules for Associative Matrix Memories and Reinforcement Learning
NOTE Communicated by David Willshaw The Variance of Covariance Rules for Associative Matrix Memories and Reinforcement Learning Peter Dayan Terrence J. Sejnowski Computational Neurobiology Laboratory,
More informationLinking connectivity, dynamics and computations in low-rank recurrent neural networks
Linking connectivity, dynamics and computations in low-rank recurrent neural networks Francesca Mastrogiuseppe 1,2, Srdjan Ostojic 1 * 1 Laboratoire de Neurosciences Cognitives, INSERM U960 and 2 Laboratoire
More informationMarkov Transitions between Attractor States in a Recurrent Neural Network
Markov Transitions between Attractor States in a Recurrent Neural Network Jeremy Bernstein Computation and Neural Systems California Institute of Technology, USA bernstein@caltech.edu Ishita Dasgupta Department
More informationOutline. NIP: Hebbian Learning. Overview. Types of Learning. Neural Information Processing. Amos Storkey
Outline NIP: Hebbian Learning Neural Information Processing Amos Storkey 1/36 Overview 2/36 Types of Learning Types of learning, learning strategies Neurophysiology, LTP/LTD Basic Hebb rule, covariance
More informationAn Introductory Course in Computational Neuroscience
An Introductory Course in Computational Neuroscience Contents Series Foreword Acknowledgments Preface 1 Preliminary Material 1.1. Introduction 1.1.1 The Cell, the Circuit, and the Brain 1.1.2 Physics of
More informationF.P. Battaglia 1. Istituto di Fisica. Universita di Roma, La Sapienza, Ple Aldo Moro, Roma and. S. Fusi
partially structured synaptic transitions F.P. Battaglia 1 Istituto di Fisica Universita di Roma, La Sapienza, Ple Aldo oro, Roma and S. Fusi INFN, Sezione dell'istituto Superiore di Sanita, Viale Regina
More informationStorage Capacity of Letter Recognition in Hopfield Networks
Storage Capacity of Letter Recognition in Hopfield Networks Gang Wei (gwei@cs.dal.ca) Zheyuan Yu (zyu@cs.dal.ca) Faculty of Computer Science, Dalhousie University, Halifax, N.S., Canada B3H 1W5 Abstract:
More informationArtificial Intelligence Hopfield Networks
Artificial Intelligence Hopfield Networks Andrea Torsello Network Topologies Single Layer Recurrent Network Bidirectional Symmetric Connection Binary / Continuous Units Associative Memory Optimization
More informationProbabilistic Models in Theoretical Neuroscience
Probabilistic Models in Theoretical Neuroscience visible unit Boltzmann machine semi-restricted Boltzmann machine restricted Boltzmann machine hidden unit Neural models of probabilistic sampling: introduction
More informationNeural networks: Unsupervised learning
Neural networks: Unsupervised learning 1 Previously The supervised learning paradigm: given example inputs x and target outputs t learning the mapping between them the trained network is supposed to give
More informationSupporting Online Material for
www.sciencemag.org/cgi/content/full/319/5869/1543/dc1 Supporting Online Material for Synaptic Theory of Working Memory Gianluigi Mongillo, Omri Barak, Misha Tsodyks* *To whom correspondence should be addressed.
More informationLearning at the edge of chaos : Temporal Coupling of Spiking Neurons Controller for Autonomous Robotic
Learning at the edge of chaos : Temporal Coupling of Spiking Neurons Controller for Autonomous Robotic Hédi Soula and Aravind Alwan and Guillaume Beslon ALAB Team of Prisma Lab Computer Science Dept National
More informationBiological Modeling of Neural Networks:
Week 14 Dynamics and Plasticity 14.1 Reservoir computing - Review:Random Networks - Computing with rich dynamics Biological Modeling of Neural Networks: 14.2 Random Networks - stationary state - chaos
More informationConvergence of stochastic learning in perceptrons with binary synapses
PhysRevE Convergence of stochastic learning in perceptrons with binary synapses Walter Senn and Stefano Fusi Department of Physiology, University of Bern (Dated: February 18, 2005) The efficacy of a biological
More informationDecoding. How well can we learn what the stimulus is by looking at the neural responses?
Decoding How well can we learn what the stimulus is by looking at the neural responses? Two approaches: devise explicit algorithms for extracting a stimulus estimate directly quantify the relationship
More informationPlasticity and learning in a network of coupled phase oscillators
PHYSICAL REVIEW E, VOLUME 65, 041906 Plasticity and learning in a network of coupled phase oscillators Philip Seliger, Stephen C. Young, and Lev S. Tsimring Institute for onlinear Science, University of
More informationIntroduction. Previous work has shown that AER can also be used to construct largescale networks with arbitrary, configurable synaptic connectivity.
Introduction The goal of neuromorphic engineering is to design and implement microelectronic systems that emulate the structure and function of the brain. Address-event representation (AER) is a communication
More informationDynamic State and Parameter Estimation Applied to Neuromorphic Systems
LETTER Communicated by Paolo Del Giudice Dynamic State and Parameter Estimation Applied to Neuromorphic Systems Emre Ozgur Neftci emre@ini.phys.ethz.ch Institute of Informatics, University of Zurich and
More informationSystems Biology: A Personal View IX. Landscapes. Sitabhra Sinha IMSc Chennai
Systems Biology: A Personal View IX. Landscapes Sitabhra Sinha IMSc Chennai Fitness Landscapes Sewall Wright pioneered the description of how genotype or phenotypic fitness are related in terms of a fitness
More informationNeural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28
1 / 28 Neural Networks Mark van Rossum School of Informatics, University of Edinburgh January 15, 2018 2 / 28 Goals: Understand how (recurrent) networks behave Find a way to teach networks to do a certain
More informationCorrelations strike back (again): the case of associative memory retrieval
Correlations strike back (again): the case of associative memory retrieval Cristina Savin 1 cs664@cam.ac.uk Peter Dayan 2 dayan@gatsby.ucl.ac.uk Máté Lengyel 1 m.lengyel@eng.cam.ac.uk 1 Computational &
More informationUnsupervised learning of an efficient short-term memory network
Unsupervised learning of an efficient short-term memory network Pietro Vertechi Wieland Brendel Christian K. Machens Champalimaud Neuroscience Programme Champalimaud Centre for the Unknown Lisbon, Portugal
More informationHow can ideas from quantum computing improve or speed up neuromorphic models of computation?
Neuromorphic Computation: Architectures, Models, Applications Associative Memory Models with Adiabatic Quantum Optimization Kathleen Hamilton, Alexander McCaskey, Jonathan Schrock, Neena Imam and Travis
More informationEffects of refractory periods in the dynamics of a diluted neural network
Effects of refractory periods in the dynamics of a diluted neural network F. A. Tamarit, 1, * D. A. Stariolo, 2, * S. A. Cannas, 2, *, and P. Serra 2, 1 Facultad de Matemática, Astronomía yfísica, Universidad
More information( ) T. Reading. Lecture 22. Definition of Covariance. Imprinting Multiple Patterns. Characteristics of Hopfield Memory
Part 3: Autonomous Agents /8/07 Reading Lecture 22 Flake, ch. 20 ( Genetics and Evolution ) /8/07 /8/07 2 Imprinting Multiple Patterns Let x, x 2,, x p be patterns to be imprinted Define the sum-of-outer-products
More informationAutoassociative Memory Retrieval and Spontaneous Activity Bumps in Small-World Networks of Integrate-and-Fire Neurons
Autoassociative Memory Retrieval and Spontaneous Activity Bumps in Small-World Networks of Integrate-and-Fire Neurons Anastasia Anishchenko Department of Physics and Brain Science Program Brown University,
More informationDesign and Implementation of BCM Rule Based on Spike-Timing Dependent Plasticity
Design and Implementation of BCM Rule Based on Spike-Timing Dependent Plasticity Mostafa Rahimi Azghadi, Said Al-Sarawi, Nicolangelo Iannella, and Derek Abbott Centre for Biomedical Engineering, School
More information(a) (b) (c) Time Time. Time
Baltzer Journals Stochastic Neurodynamics and the System Size Expansion Toru Ohira and Jack D. Cowan 2 Sony Computer Science Laboratory 3-4-3 Higashi-gotanda, Shinagawa, Tokyo 4, Japan E-mail: ohiracsl.sony.co.jp
More informationHopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5
Hopfield Neural Network and Associative Memory Typical Myelinated Vertebrate Motoneuron (Wikipedia) PHY 411-506 Computational Physics 2 1 Wednesday, March 5 1906 Nobel Prize in Physiology or Medicine.
More informationHow to read a burst duration code
Neurocomputing 58 60 (2004) 1 6 www.elsevier.com/locate/neucom How to read a burst duration code Adam Kepecs a;, John Lisman b a Cold Spring Harbor Laboratory, Marks Building, 1 Bungtown Road, Cold Spring
More informationStationary Bumps in Networks of Spiking Neurons
LETTER Communicated by Misha Tsodyks Stationary Bumps in Networks of Spiking Neurons Carlo R. Laing Carson C. Chow Department of Mathematics, University of Pittsburgh, Pittsburgh PA 1526, U.S.A. We examine
More informationApplied Physics and by courtesy, Neurobiology and EE
Information theoretic limits on the memory capacity of neuronal and synaptic networks. Surya Ganguli Applied Physics and by courtesy, Neurobiology and EE Stanford The Nature and Scope of Theoretical Neuroscience
More informationBinary synapses: better than expected
October 13, 2008. version 8. 1 Binary synapses: better than expected 1 Introduction Memory is an absolutely critical component in any biological system. Without it, learning would be impossible, and behavior
More informationSynaptic Plasticity. Introduction. Biophysics of Synaptic Plasticity. Functional Modes of Synaptic Plasticity. Activity-dependent synaptic plasticity:
Synaptic Plasticity Introduction Dayan and Abbott (2001) Chapter 8 Instructor: Yoonsuck Choe; CPSC 644 Cortical Networks Activity-dependent synaptic plasticity: underlies learning and memory, and plays
More informationLeo Kadanoff and 2d XY Models with Symmetry-Breaking Fields. renormalization group study of higher order gradients, cosines and vortices
Leo Kadanoff and d XY Models with Symmetry-Breaking Fields renormalization group study of higher order gradients, cosines and vortices Leo Kadanoff and Random Matrix Theory Non-Hermitian Localization in
More informationWhat Can a Neuron Learn with Spike-Timing-Dependent Plasticity?
LETTER Communicated by Wulfram Gerstner What Can a Neuron Learn with Spike-Timing-Dependent Plasticity? Robert Legenstein legi@igi.tugraz.at Christian Naeger naeger@gmx.de Wolfgang Maass maass@igi.tugraz.at
More informationLearning and Memory in Neural Networks
Learning and Memory in Neural Networks Guy Billings, Neuroinformatics Doctoral Training Centre, The School of Informatics, The University of Edinburgh, UK. Neural networks consist of computational units
More informationEnergy-Efficient Neuromorphic Classifiers
ARTICLE Communicated by Gert Cauwenberghs Energy-Efficient Neuromorphic Classifiers Daniel Martí daniel.marti@ens.fr Département d Études Cognitives, École Normale Supérieure PSL Research University, 75005
More informationA nanoparticle-organic memory field-effect transistor behaving as a programmable spiking synapse
A nanoparticle-organic memory field-effect transistor behaving as a programmable spiking synapse F. Alibart,. Pleutin, D. Guerin, K. Lmimouni, D. Vuillaume Molecular Nanostructures & Devices group, Institute
More informationNeural Coding: Integrate-and-Fire Models of Single and Multi-Neuron Responses
Neural Coding: Integrate-and-Fire Models of Single and Multi-Neuron Responses Jonathan Pillow HHMI and NYU http://www.cns.nyu.edu/~pillow Oct 5, Course lecture: Computational Modeling of Neuronal Systems
More informationProperties of Associative Memory Model with the β-th-order Synaptic Decay
Regular Paper Properties of Associative Memory Model with the β-th-order Synaptic Decay Ryota Miyata 1,2 Toru Aonishi 1 Jun Tsuzurugi 3 Koji Kurata 4,a) Received: January 30, 2013, Revised: March 20, 2013/June
More informationA recurrent model of transformation invariance by association
NN 1381 PERGAMON Neural Networks 13 (2000) 225 237 Contributed article A recurrent model of transformation invariance by association Martin C.M. Elliffe a, Edmund T. Rolls a,1,2, *,Néstor Parga b,2,3,
More informationIntroduction to Neural Networks
Introduction to Neural Networks What are (Artificial) Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning
More informationLarge-scale neural modeling
Large-scale neural modeling We re acquiring brain data at an unprecedented rate Dendritic recording Serial Scanning EM Ca ++ imaging Kwabena Boahen Stanford Bioengineering boahen@stanford.edu Goal: Link
More informationNeural Nets and Symbolic Reasoning Hopfield Networks
Neural Nets and Symbolic Reasoning Hopfield Networks Outline The idea of pattern completion The fast dynamics of Hopfield networks Learning with Hopfield networks Emerging properties of Hopfield networks
More informationREAL-TIME COMPUTING WITHOUT STABLE
REAL-TIME COMPUTING WITHOUT STABLE STATES: A NEW FRAMEWORK FOR NEURAL COMPUTATION BASED ON PERTURBATIONS Wolfgang Maass Thomas Natschlager Henry Markram Presented by Qiong Zhao April 28 th, 2010 OUTLINE
More informationStochastic Learning in a Neural Network with Adapting. Synapses. Istituto Nazionale di Fisica Nucleare, Sezione di Bari
Stochastic Learning in a Neural Network with Adapting Synapses. G. Lattanzi 1, G. Nardulli 1, G. Pasquariello and S. Stramaglia 1 Dipartimento di Fisica dell'universita di Bari and Istituto Nazionale di
More informationPlasticity and Learning
Chapter 8 Plasticity and Learning 8.1 Introduction Activity-dependent synaptic plasticity is widely believed to be the basic phenomenon underlying learning and memory, and it is also thought to play a
More informationHow do synapses transform inputs?
Neurons to networks How do synapses transform inputs? Excitatory synapse Input spike! Neurotransmitter release binds to/opens Na channels Change in synaptic conductance! Na+ influx E.g. AMA synapse! Depolarization
More information7 Rate-Based Recurrent Networks of Threshold Neurons: Basis for Associative Memory
Physics 178/278 - David Kleinfeld - Fall 2005; Revised for Winter 2017 7 Rate-Based Recurrent etworks of Threshold eurons: Basis for Associative Memory 7.1 A recurrent network with threshold elements The
More informationarxiv: v1 [cs.ne] 30 Mar 2013
A Neuromorphic VLSI Design for Spike Timing and Rate Based Synaptic Plasticity Mostafa Rahimi Azghadi a,, Said Al-Sarawi a,, Derek Abbott a, Nicolangelo Iannella a, a School of Electrical and Electronic
More informationRate- and Phase-coded Autoassociative Memory
Rate- and Phase-coded Autoassociative Memory Máté Lengyel Peter Dayan Gatsby Computational Neuroscience Unit, University College London 7 Queen Square, London WCN 3AR, United Kingdom {lmate,dayan}@gatsby.ucl.ac.uk
More informationCovariance Learning of Correlated Patterns in Competitive. Networks. Ali A. Minai. University of Cincinnati. (To appearinneural Computation)
Covariance Learning of Correlated Patterns in Competitive Networks Ali A. Minai Department of Electrical & Computer Engineering and Computer Science University of Cincinnati Cincinnati, OH 45221 (To appearinneural
More informationEvent-Driven Random Backpropagation: Enabling Neuromorphic Deep Learning Machines
Event-Driven Random Backpropagation: Enabling Neuromorphic Deep Learning Machines Emre Neftci Department of Cognitive Sciences, UC Irvine, Department of Computer Science, UC Irvine, March 7, 2017 Scalable
More informationMachine Learning. Neural Networks. (slides from Domingos, Pardo, others)
Machine Learning Neural Networks (slides from Domingos, Pardo, others) For this week, Reading Chapter 4: Neural Networks (Mitchell, 1997) See Canvas For subsequent weeks: Scaling Learning Algorithms toward
More informationSCRAM: Statistically Converging Recurrent Associative Memory
SCRAM: Statistically Converging Recurrent Associative Memory Sylvain Chartier Sébastien Hélie Mounir Boukadoum Robert Proulx Department of computer Department of computer science, UQAM, Montreal, science,
More informationNeuron, Volume 63. Supplemental Data. Generating Coherent Patterns of Activity. from Chaotic Neural Networks. David Sussillo and L.F.
Neuron, Volume 63 Supplemental Data Generating Coherent Patterns of Activity from Chaotic Neural Networks David Sussillo and L.F. Abbott Supplementary Material FORCE Learning without RLS It is possible
More informationDynamical systems in neuroscience. Pacific Northwest Computational Neuroscience Connection October 1-2, 2010
Dynamical systems in neuroscience Pacific Northwest Computational Neuroscience Connection October 1-2, 2010 What do I mean by a dynamical system? Set of state variables Law that governs evolution of state
More informationImplementation of a Restricted Boltzmann Machine in a Spiking Neural Network
Implementation of a Restricted Boltzmann Machine in a Spiking Neural Network Srinjoy Das Department of Electrical and Computer Engineering University of California, San Diego srinjoyd@gmail.com Bruno Umbria
More informationLiquid Computing. Wolfgang Maass. Institut für Grundlagen der Informationsverarbeitung Technische Universität Graz, Austria
NNB SS10 1 Liquid Computing Wolfgang Maass Institut für Grundlagen der Informationsverarbeitung Technische Universität Graz, Austria Institute for Theoretical Computer Science http://www.igi.tugraz.at/maass/
More informationCorrelations and neural information coding Shlens et al. 09
Correlations and neural information coding Shlens et al. 09 Joel Zylberberg www.jzlab.org The neural code is not one-to-one ρ = 0.52 ρ = 0.80 d 3s [Max Turner, UW] b # of trials trial 1!! trial 2!!.!.!.
More informationMachine Learning. Neural Networks. (slides from Domingos, Pardo, others)
Machine Learning Neural Networks (slides from Domingos, Pardo, others) Human Brain Neurons Input-Output Transformation Input Spikes Output Spike Spike (= a brief pulse) (Excitatory Post-Synaptic Potential)
More informationRE-ENGINEERING COMPUTING WITH NEURO- MIMETIC DEVICES, CIRCUITS, AND ALGORITHMS
RE-ENGINEERING COMPUTING WITH NEURO- MIMETIC DEVICES, CIRCUITS, AND ALGORITHMS Kaushik Roy Abhronil Sengupta, Gopal Srinivasan, Aayush Ankit, Priya Panda, Xuanyao Fong, Deliang Fan, Jason Allred School
More informationBasic elements of neuroelectronics -- membranes -- ion channels -- wiring. Elementary neuron models -- conductance based -- modelers alternatives
Computing in carbon Basic elements of neuroelectronics -- membranes -- ion channels -- wiring Elementary neuron models -- conductance based -- modelers alternatives Wiring neurons together -- synapses
More informationConsider the way we are able to retrieve a pattern from a partial key as in Figure 10 1.
CompNeuroSci Ch 10 September 8, 2004 10 Associative Memory Networks 101 Introductory concepts Consider the way we are able to retrieve a pattern from a partial key as in Figure 10 1 Figure 10 1: A key
More informationA MEAN FIELD THEORY OF LAYER IV OF VISUAL CORTEX AND ITS APPLICATION TO ARTIFICIAL NEURAL NETWORKS*
683 A MEAN FIELD THEORY OF LAYER IV OF VISUAL CORTEX AND ITS APPLICATION TO ARTIFICIAL NEURAL NETWORKS* Christopher L. Scofield Center for Neural Science and Physics Department Brown University Providence,
More informationHardware-Amenable Structural Learning for Spikebased Pattern Classification using a Simple Model of Active Dendrites
1 arxiv:1411.5881v2 [cs.ne] 25 Nov 2014 Hardware-Amenable Structural Learning for Spikebased Pattern Classification using a Simple Model of Active Dendrites Shaista Hussain School of Electrical and Electronic
More informationNeural information often passes through many different
Transmission of population coded information Alfonso Renart, and Mark C. W. van Rossum Instituto de Neurociencias de Alicante. Universidad Miguel Hernndez - CSIC 03550 Sant Joan d Alacant, Spain, Center
More informationarxiv: v1 [cs.et] 20 Nov 2014
Liquid State Machine with Dendritically Enhanced Readout for Low-power, Neuromorphic VLSI Implementations Subhrajit Roy, Student Member, IEEE, Amitava Banerjee and Arindam Basu, Member, IEEE arxiv:4.5458v
More informationMagnetic tunnel junction beyond memory from logic to neuromorphic computing WANJUN PARK DEPT. OF ELECTRONIC ENGINEERING, HANYANG UNIVERSITY
Magnetic tunnel junction beyond memory from logic to neuromorphic computing WANJUN PARK DEPT. OF ELECTRONIC ENGINEERING, HANYANG UNIVERSITY Magnetic Tunnel Junctions (MTJs) Structure High density memory
More informationUC Irvine UC Irvine Previously Published Works
UC Irvine UC Irvine Previously Published Works Title Neuromorphic Deep Learning Machines Permalink https://escholarship.org/uc/item/335376xb Authors Neftci, E Augustine, C Paul, S et al. Publication Date
More informationMemory Capacity for Sequences in a Recurrent Network with Biological Constraints
Memory Capacity for Sequences in a Recurrent Network with Biological Constraints Christian Leibold,3 and Richard Kempter,2,3 Institute for Theoretical Biology, Humboldt-Universität zu Berlin, Germany 2
More informationBeyond Hebbian plasticity: Effective learning with ineffective Hebbian learning rules
Beyond Hebbian plasticity: Effective learning with ineffective Hebbian learning rules Gal Chechik The center for neural computation Hebrew University in Jerusalem, Israel and the School of Mathematical
More informationLateral organization & computation
Lateral organization & computation review Population encoding & decoding lateral organization Efficient representations that reduce or exploit redundancy Fixation task 1rst order Retinotopic maps Log-polar
More informationMachine Learning. Neural Networks. (slides from Domingos, Pardo, others)
Machine Learning Neural Networks (slides from Domingos, Pardo, others) For this week, Reading Chapter 4: Neural Networks (Mitchell, 1997) See Canvas For subsequent weeks: Scaling Learning Algorithms toward
More information7 Recurrent Networks of Threshold (Binary) Neurons: Basis for Associative Memory
Physics 178/278 - David Kleinfeld - Winter 2019 7 Recurrent etworks of Threshold (Binary) eurons: Basis for Associative Memory 7.1 The network The basic challenge in associative networks, also referred
More informationInformation Theory. Mark van Rossum. January 24, School of Informatics, University of Edinburgh 1 / 35
1 / 35 Information Theory Mark van Rossum School of Informatics, University of Edinburgh January 24, 2018 0 Version: January 24, 2018 Why information theory 2 / 35 Understanding the neural code. Encoding
More informationarxiv: v1 [cond-mat.dis-nn] 9 May 2008
Functional Optimization in Complex Excitable Networks Samuel Johnson, J. Marro, and Joaquín J. Torres Departamento de Electromagnetismo y Física de la Materia, and Institute Carlos I for Theoretical and
More informationAnalysis of Neural Networks with Chaotic Dynamics
Chaos, Solitonr & Fructals Vol. 3, No. 2, pp. 133-139, 1993 Printed in Great Britain @60-0779/93$6.00 + 40 0 1993 Pergamon Press Ltd Analysis of Neural Networks with Chaotic Dynamics FRANCOIS CHAPEAU-BLONDEAU
More informationBiosciences in the 21st century
Biosciences in the 21st century Lecture 1: Neurons, Synapses, and Signaling Dr. Michael Burger Outline: 1. Why neuroscience? 2. The neuron 3. Action potentials 4. Synapses 5. Organization of the nervous
More informationAssociative Memory with Dynamic Synapses
LETTER Communicated by Alessandro Treves Associative Memory with Dynamic Synapses Lovorka Pantic lovorka@mbfys.kun.nl JoaquÂõ n J. Torres torres@onsager.ugr.es Hilbert J. Kappen bert@mbfys.kun.nl Stan
More informationneural networks Balázs B Ujfalussy 17 october, 2016 idegrendszeri modellezés 2016 október 17.
neural networks Balázs B Ujfalussy 17 october, 2016 Hierarchy of the nervous system behaviour idegrendszeri modellezés 1m CNS 10 cm systems 1 cm maps 1 mm networks 100 μm neurons 1 μm synapses 10 nm molecules
More informationModel neurons!!poisson neurons!
Model neurons!!poisson neurons! Suggested reading:! Chapter 1.4 in Dayan, P. & Abbott, L., heoretical Neuroscience, MI Press, 2001.! Model neurons: Poisson neurons! Contents: Probability of a spike sequence
More informationAdaptation in the Neural Code of the Retina
Adaptation in the Neural Code of the Retina Lens Retina Fovea Optic Nerve Optic Nerve Bottleneck Neurons Information Receptors: 108 95% Optic Nerve 106 5% After Polyak 1941 Visual Cortex ~1010 Mean Intensity
More informationSampling-based probabilistic inference through neural and synaptic dynamics
Sampling-based probabilistic inference through neural and synaptic dynamics Wolfgang Maass for Robert Legenstein Institute for Theoretical Computer Science Graz University of Technology, Austria Institute
More informationThis document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.
This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore. Title Liquid state machine with dendritically enhanced readout for low-power, neuromorphic VLSI implementations
More informationBiological Modeling of Neural Networks
Week 4 part 2: More Detail compartmental models Biological Modeling of Neural Networks Week 4 Reducing detail - Adding detail 4.2. Adding detail - apse -cable equat Wulfram Gerstner EPFL, Lausanne, Switzerland
More informationDivisive Inhibition in Recurrent Networks
Divisive Inhibition in Recurrent Networks Frances S. Chance and L. F. Abbott Volen Center for Complex Systems and Department of Biology Brandeis University Waltham MA 2454-911 Abstract Models of visual
More informationThe N3XT Technology for. Brain-Inspired Computing
The N3XT Technology for Brain-Inspired Computing SystemX Alliance 27..8 Department of Electrical Engineering 25.4.5 2 25.4.5 Source: Google 3 25.4.5 Source: vrworld.com 4 25.4.5 Source: BDC Stanford Magazine
More informationMemories Associated with Single Neurons and Proximity Matrices
Memories Associated with Single Neurons and Proximity Matrices Subhash Kak Oklahoma State University, Stillwater Abstract: This paper extends the treatment of single-neuron memories obtained by the use
More informationNeural variability and Poisson statistics
Neural variability and Poisson statistics January 15, 2014 1 Introduction We are in the process of deriving the Hodgkin-Huxley model. That model describes how an action potential is generated by ion specic
More informationSynaptic Devices and Neuron Circuits for Neuron-Inspired NanoElectronics
Synaptic Devices and Neuron Circuits for Neuron-Inspired NanoElectronics Byung-Gook Park Inter-university Semiconductor Research Center & Department of Electrical and Computer Engineering Seoul National
More information