Fast and exact simulation methods applied on a broad range of neuron models

Similar documents
Synaptic dynamics. John D. Murray. Synaptic currents. Simple model of the synaptic gating variable. First-order kinetics

DISCRETE EVENT SIMULATION IN THE NEURON ENVIRONMENT

Event-driven simulations of nonlinear integrate-and-fire neurons

Neural Modeling and Computational Neuroscience. Claudio Gallicchio

Biological Modeling of Neural Networks

Consider the following spike trains from two different neurons N1 and N2:

DEVS Simulation of Spiking Neural Networks

The Spike Response Model: A Framework to Predict Neuronal Spike Trains

Basic elements of neuroelectronics -- membranes -- ion channels -- wiring

Dynamical Constraints on Computing with Spike Timing in the Cortex

A Generalized Linear Integrate-And-Fire Neural. Model Produces Diverse Spiking Behaviors

NIH Public Access Author Manuscript Neural Comput. Author manuscript; available in PMC 2010 October 13.

An analytical model for the large, fluctuating synaptic conductance state typical of neocortical neurons in vivo

An Efficient Method for Computing Synaptic Conductances Based on a Kinetic Model of Receptor Binding

A FINITE STATE AUTOMATON MODEL FOR MULTI-NEURON SIMULATIONS

How do synapses transform inputs?

Kinetic models of spike-timing dependent plasticity and their functional consequences in detecting correlations

Liquid Computing in a Simplified Model of Cortical Layer IV: Learning to Balance a Ball

Conductance-Based Integrate-and-Fire Models

We observe the model neuron s response to constant input current, studying the dependence of:

Mathematical Foundations of Neuroscience - Lecture 3. Electrophysiology of neurons - continued

Subthreshold cross-correlations between cortical neurons: Areference model with static synapses

Bursting and Chaotic Activities in the Nonlinear Dynamics of FitzHugh-Rinzel Neuron Model

An Introductory Course in Computational Neuroscience

A gradient descent rule for spiking neurons emitting multiple spikes

Action Potential Initiation in the Hodgkin-Huxley Model

A General Framework for Neurobiological Modeling: An Application to the Vestibular System 1

Probabilistic Models in Theoretical Neuroscience

Effects of synaptic conductance on the voltage distribution and firing rate of spiking neurons

CSE/NB 528 Final Lecture: All Good Things Must. CSE/NB 528: Final Lecture

Dendritic computation

Linear Regression, Neural Networks, etc.

9 Generation of Action Potential Hodgkin-Huxley Model

INVERSE PROBLEMS FOR THE HODGKIN-HUXLEY MODEL

Mathematical Models of Dynamic Behavior of Individual Neural Networks of Central Nervous System

Activity Driven Adaptive Stochastic. Resonance. Gregor Wenning and Klaus Obermayer. Technical University of Berlin.

Lecture 11 : Simple Neuron Models. Dr Eileen Nugent

Mathematical Models of Dynamic Behavior of Individual Neural Networks of Central Nervous System

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

On the performance of voltage-stepping for the simulation of adaptive, nonlinear, integrate-and-fire neuronal networks.

Analytical Integrate-and-Fire Neuron Models with Conductance-Based Dynamics for Event-Driven Simulation Strategies

Topics in Neurophysics

Dynamical systems in neuroscience. Pacific Northwest Computational Neuroscience Connection October 1-2, 2010

Fast neural network simulations with population density methods

Phase Response. 1 of of 11. Synaptic input advances (excitatory) or delays (inhibitory) spiking

Neuronal Dynamics: Computational Neuroscience of Single Neurons

Spike-Frequency Adaptation: Phenomenological Model and Experimental Tests

80% of all excitatory synapses - at the dendritic spines.

Neuron. Detector Model. Understanding Neural Components in Detector Model. Detector vs. Computer. Detector. Neuron. output. axon

In#uence of dendritic topologyon "ring patterns in model neurons

Hybrid Integrate-and-Fire Model of a Bursting Neuron

Computational Explorations in Cognitive Neuroscience Chapter 2

STDP Learning of Image Patches with Convolutional Spiking Neural Networks

MANY scientists believe that pulse-coupled neural networks

Action Potentials and Synaptic Transmission Physics 171/271

A Three-dimensional Physiologically Realistic Model of the Retina

Model of a Biological Neuron as a Temporal Neural Network

Neurophysiology of a VLSI spiking neural network: LANN21

Structure and Measurement of the brain lecture notes

Discrete and Indiscrete Models of Biological Networks

3 Detector vs. Computer

Linking non-binned spike train kernels to several existing spike train metrics

How and why neurons fire

Computing with Inter-spike Interval Codes in Networks of Integrate and Fire Neurons

Biological Modeling of Neural Networks:

Visual Selection and Attention Shifting Based on FitzHugh-Nagumo Equations

Compartmental Modelling

A new method to infer higher-order spike correlations from membrane potentials

Processing of Time Series by Neural Circuits with Biologically Realistic Synaptic Dynamics

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.

Supervisor: Prof. Stefano Spaccapietra Dr. Fabio Porto Student: Yuanjian Wang Zufferey. EPFL - Computer Science - LBD 1

Artificial Neural Network

Analyzing Neuroscience Signals using Information Theory and Complexity

MATH 3104: THE HODGKIN-HUXLEY EQUATIONS

Integration of synaptic inputs in dendritic trees

Emergence of resonances in neural systems: the interplay between adaptive threshold and short-term synaptic plasticity

Computational Neuroscience. Lubica Benuskova Lecture 1

Oscillations and Synchrony in Large-scale Cortical Network Models

IN THIS turorial paper we exploit the relationship between

A First Attempt of Reservoir Pruning for Classification Problems

arxiv: v1 [q-bio.nc] 30 Apr 2012

System Identification for the Hodgkin-Huxley Model using Artificial Neural Networks

Mathematical Foundations of Neuroscience - Lecture 9. Simple Models of Neurons and Synapses.

Principles of DCM. Will Penny. 26th May Principles of DCM. Will Penny. Introduction. Differential Equations. Bayesian Estimation.

This script will produce a series of pulses of amplitude 40 na, duration 1ms, recurring every 50 ms.

Neurophysiology. Danil Hammoudi.MD

arxiv: v1 [q-bio.nc] 16 May 2017

Methods for Estimating the Computational Power and Generalization Capability of Neural Microcircuits

Outline. Neural dynamics with log-domain integrator circuits. Where it began Biophysics of membrane channels

Balance of Electric and Diffusion Forces

Implementing Synaptic Plasticity in a VLSI Spiking Neural Network Model

Novel VLSI Implementation for Triplet-based Spike-Timing Dependent Plasticity

Causality and communities in neural networks

Electrophysiology of the neuron

How to read a burst duration code

Single neuron models. L. Pezard Aix-Marseille University

Effects of Interactive Function Forms and Refractoryperiod in a Self-Organized Critical Model Based on Neural Networks

A Model for Real-Time Computation in Generic Neural Microcircuits

A dynamical system-markov model for active postsynaptic responses of muscle spindle afferent nerve

Neural Nets in PR. Pattern Recognition XII. Michal Haindl. Outline. Neural Nets in PR 2

Transcription:

Fast and exact simulation methods applied on a broad range of neuron models Michiel D Haene michiel.dhaene@ugent.be Benjamin Schrauwen benjamin.schrauwen@ugent.be Ghent University, Electronics and Information Systems Department, 9000 Ghent, Belgium Abstract Recently, van Elburg and van Ooyen (2009) published a generalization of the Event- Based Integration Scheme for an Integrate-and-Fire neuron model with exponentially decaying excitatory currents and double exponential inhibitory synaptic currents, introduced by Carnevale and Hines. In the paper, it was shown that the constraints on the synaptic time constants imposed by the Newton-Raphson iteration scheme, can be relaxed. In this note, we show that according to the results published in D Haene et al. (2009), a further generalization is possible eliminating any constraint on the time constants. We also demonstrate that in fact a wide range of linear neuron models can be efficiently simulated with this computation scheme, including neuron models mimicking complex neuronal behavior. These results can change the way complex neuronal spiking behavior is modeled: instead of highly non linear neuron models with few state variables, it is possible to efficiently simulate linear models with a large number of state variables. 1

In Carnevale and Hines (2002) and Hines and Carnevale (2004), an event driven approach was described for three types of Integrate-and-Fire Models. The most general model in these contributions, IntFire4, uses four time constants, τ e for fast excitatory currents, τ j and τ i for a slower inhibitory current with rise time, and τ m for the membrane. The model was described by the following set of differential equations: dv de dj di = 1 τ m V (t) + a e e + a i i = 1 τ e e = 1 τ j j = 1 τ i i + a j j. In order to assure correct convergence of a Newton-Raphson iteration scheme, the following restrictions were imposed: τ e < τ j < τ i < τ m. These constraints imply a significant limitation on the biological realism of the simulations, e.g. the decay times of excitatory AMPA synapses are typically larger than the rise times of inhibitory GABA currents (Bier et al., 1996; Häusser and Roth, 1997), or the simulation of high conductance states involves the use of small membrane time constants (Destexhe et al., 2003). Recently, van Elburg and van Ooyen (2009) showed that these non-physiological constraints on the time constants can be relaxed. It was shown that multiple time constants can be used with the only restriction that the slowest decaying excitatory synaptic current should decay faster than the fastest decaying inhibitory current, i.e. τ eν < τ iµ for any combination of ν and µ. The authors proved that when these constraints are applied, a normal Newton-Raphson iteration scheme can be used in order to find the firing time. Moreover, since there are no overestimations of the exact firing time, the overhead of firing time predictions can be reduced for each incoming spike by performing only a single Newton-Raphson iteration at a time. After each iteration, an internal self -event with the current estimated time is rescheduled in the event-driven environment. The generalizations discussed in van Elburg and van Ooyen (2009) already allow to simulate more general neuron models in an event driven way. The remaining constraint that all excitatory currents should decay faster than any inhibitory current is also assumed in many related publications on integrate-and-fire models with multiple synap- 2

tic time constants, since this significantly simplifies the threshold finding mechanism, e.g. Brette (2007), Rudolph and Destexhe (2006). However, in some circumstances, this is still a (physiologically) unacceptable restriction. For example, the simulation of NMDA-like currents, which can be modeled by an excitatory double exponential current (when ignoring non-linear effects such as the magnesium block), requires slower excitatory time constants (van Elburg and van Ooyen, 2009). A similar approach for finding the firing time was proposed in D Haene et al. (2009). This paper presents an efficient integration scheme which can be used for arbitrary integrate-and-fire neuron models with multiple exponentially decaying synaptic time constants, with no limitations on the time constants or the number of different time constants. The analytical solution of the membrane function is a sum of exponential functions. It is shown that the convergence problems with Newton-Raphson based iterative methods can only occur when there are negative contributions in the membrane function (which is a sum of exponential functions) with a time constant larger than any of the positive terms. In this case, the first derivative of the membrane function, used to perform a Newton-Raphson iteration, can underestimate the membrane function and converge to the wrong firing time or even not converge at all. To prevent this, all contributions to the first derivative which can cause convergence problems are replaced by a virtual contribution in which the time constant of the exponential function is set to the smallest time-constant that occur in all positive terms. The applicability of the algorithm of D Haene et al. (2009) is not limited to integrateand-fire neuron models with multiple exponentially decaying synapses. Since the solution of the double exponential inhibitory currents used in the IntFire4-model also takes the form of the sum of two exponential functions, the algorithm in D Haene et al. (2009) can be applied without modifications. In general, arbitrary synaptic functions can be efficiently simulated when they can be approximated as a sum of exponential functions. From the simulators point of view, these synapse functions are represented by multiple interconnections between two neurons, each implementing a simple exponential decaying synapse. Using also the propagation delay of the interconnections, synaptic responses with multiple peaks can be easily approximated. It can be shown that if the constraints from van Elburg and van Ooyen (2009) are applied to the IntFire4-model, the membrane function will never contain negative 3

exponential terms with larger time constants. In this case, the algorithm of D Haene et al. (2009) will not need to replace any negative contributions during the computation of the first derivative of the membrane function. On this level, both methods would perform the same computations. However, it was also shown in D Haene et al. (2009) that the computationally costly Newton-Raphson iterations can be completely eliminated most of the time, as are the update of all state variables after each incoming spike. This accounts for the largest part of the high speedups that were reported. One of the main ideas behind many of the recently presented neuron models is limiting the number of state variables in order to decrease the simulation time, while maximizing the number of biologically observed neuron properties such as frequency adaptation and bursting (e.g. Izhikevich (2004), Brette and Gerstner (2005)). These models are generally characterized by strongly non-linear differential equations and thereby a lack of an analytical solution for the firing times. Although the limitation of the number of state variables simplifies the model and allows a faster time-step based simulation compared to biologically realistic neuron models (e.g. compartmental Hodgkin-Huxley neurons), the lack of an analytical solution as well as the complex non-linearities makes the simulation of these models hard to optimize computationally using event-based methods. This stands in contrast to the high speedup-ratios that can be reached by the event-based simulation of linear models (D Haene et al., 2009). An important side-effect of the optimizations presented in that work is that the simulation time is almost not affected by the number of state variables that are used to describe the neuron model. Therefore, there is an increasing interest in powerful linear models which try to mimic more complex behavior by adding more state variables. An excellent example of this is the model recently presented by Mihalaş and Niebur (2009). This model starts with a simple Integrate-and-Fire neuron model with exponential synapses. But the paper shows that the addition of membrane voltage dependent threshold dynamics can mimic much of the rich behaviors that are normally only possible in non-linear models. The model is described by the following system: di j = 1 I j (t) ; j = 1,..., N τ j dv = 1 ( I e + ) j C I j (t) G (V (t) E L ) dθ = a (V (t) E L ) b (θ (t) θ ). 4

This system of differential equations is analytically solvable, and moreover, the solution for both the membrane V (t) and the threshold θ(t) takes the form of a sum of exponential functions. The firing time can be easily found for V (t) θ(t) = 0, which is also a sum of exponential functions. Therefore, the optimization methods presented in D Haene et al. (2009) can be directly applied to this model, giving it an important computational benefit over non-linear neuron models. We conclude that there exists an interesting alternative to the recent trend of using highly non-linear neuron models with the least possible number of state variables (and which cannot well be accelerated using event-based techniques): neuron models where the membrane potential can be written as a sum of (possible many) exponentials can be very efficiently simulated using the techniques discussed in this note, and this both in simple single cell simulations and in large scale biologically plausible networks (D Haene et al., 2009). Moreover, these algorithms do not require any additional constraint on the time constants of the different state variables. The introduction of these novel event-based simulation techniques could lead to a paradigm shift in the way complex neuronal spiking behavior is modeled, moving from low-order non-linear to high-order linear differential equations. This was already demonstrated in Mihalaş and Niebur (2009) where a complex neuron model was introduced which can exhibit rich neuronal behavior, usually only observed in non-linear models. Acknowledgments This work is partially funded by the Institute for the Promotion of Innovation through Science and Technology in Flanders (IWT-Vlaanderen) and by the European Community s Seventh Framework Programme (EU FP7) under grant agreement n. 231267 Self-organized recurrent neural learning for language processing (ORGANIC). References Bier, M., Kits, K., and Borst, G. (1996). Relation between rise times and amplitude of GABAergic postsynaptic currents. Journal of Neurophysiology, 75:1008 1012. 5

Brette, R. (2007). Exact simulation of integrate-and-fire models with exponential currents. Neural Computation, 19:2604 2609. Brette, R. and Gerstner, W. (2005). Adaptive exponential integrate-and-fire model as an effective description of neuronal activity. Journal of Neurophysiology, 94:3637 3642. Carnevale, N. and Hines, M. (2002). Efficient discrete event simulation of spiking neurons in NEURON. In 2002 SFN Meeting, Yale University, New Haven. SfN abstract 312.10. Destexhe, A., Rudolph, M., and Paré, D. (2003). The high-conductance state of neocortical neurons in vivo. Neuroscience, 4:739 751. D Haene, M., Schrauwen, B., Van Campenhout, J., and Strooban, D. (2009). Accelerating event-driven simulation of spiking neurons with multiple synaptic time constants. Neural Computation, 21(4):1068 1099. Häusser, M. and Roth, A. (1997). Estimating the time course of the excitatory synaptic conductance in neocortical pyramidal cells using a novel voltage jump method. Journal of Neuroscience, 17:7606 7625. Hines, M. and Carnevale, N. (2004). Discrete event simulation in the NEURON environment. Neurocomputing, 58:1117 1122. Izhikevich, E. (2004). Which model to use for cortical spiking neurons? IEEE Transactions on Neural Networks, 15:1063 1070. Mihalaş, Ş. and Niebur, E. (2009). A generalized linear integrate-and-fire neural model produces diverse spiking behaviors. Neural Computation, 21:704 718. Rudolph, M. and Destexhe, A. (2006). Analytical integrate-and-fire neuron models with conductance-based dynamics for event-driven simulation strategies. Neural Computation, 18:2146 2210. van Elburg, R. A. J. and van Ooyen, A. (2009). Generalization of the event-based Carnevale-Hines integration scheme for Integrate-and-Fire models. Neural Computation, 21:1913 1930. 6