Computing with Inter-spike Interval Codes in Networks of Integrate and Fire Neurons

Similar documents
Computing with inter-spike interval codes in networks ofintegrate and fire neurons

Consider the following spike trains from two different neurons N1 and N2:

When is an Integrate-and-fire Neuron like a Poisson Neuron?

Membrane equation. VCl. dv dt + V = V Na G Na + V K G K + V Cl G Cl. G total. C m. G total = G Na + G K + G Cl

Fast neural network simulations with population density methods

Neurophysiology of a VLSI spiking neural network: LANN21

DEVS Simulation of Spiking Neural Networks

The homogeneous Poisson process

Methods for Estimating the Computational Power and Generalization Capability of Neural Microcircuits

An Introductory Course in Computational Neuroscience

Model neurons!!poisson neurons!

Introduction to Neural Networks U. Minn. Psy 5038 Spring, 1999 Daniel Kersten. Lecture 2a. The Neuron - overview of structure. From Anderson (1995)

REAL-TIME COMPUTING WITHOUT STABLE

A Model for Real-Time Computation in Generic Neural Microcircuits

Activity Driven Adaptive Stochastic. Resonance. Gregor Wenning and Klaus Obermayer. Technical University of Berlin.

Lecture IV: LTI models of physical systems

Decoding Input Signals in Time Domain A Model Approach

Synaptic dynamics. John D. Murray. Synaptic currents. Simple model of the synaptic gating variable. First-order kinetics

Roles of Fluctuations. in Pulsed Neural Networks

The Spike Response Model: A Framework to Predict Neuronal Spike Trains

On the Dynamics of Delayed Neural Feedback Loops. Sebastian Brandt Department of Physics, Washington University in St. Louis

A gradient descent rule for spiking neurons emitting multiple spikes

arxiv:physics/ v1 [physics.bio-ph] 19 Feb 1999

Dendritic cable with active spines: a modelling study in the spike-diffuse-spike framework

Liquid Computing in a Simplified Model of Cortical Layer IV: Learning to Balance a Ball

Characterizing the firing properties of an adaptive

Spiking Neural Networks as Timed Automata

Phase Response Properties and Phase-Locking in Neural Systems with Delayed Negative-Feedback. Carter L. Johnson

Synfire Waves in Small Balanced Networks

Mathematical Models of Dynamic Behavior of Individual Neural Networks of Central Nervous System

Emergence of resonances in neural systems: the interplay between adaptive threshold and short-term synaptic plasticity

Distinguishing the Causes of Firing with the Membrane Potential Slope

Spike-Frequency Adaptation of a Generalized Leaky Integrate-and-Fire Model Neuron

NON-MARKOVIAN SPIKING STATISTICS OF A NEURON WITH DELAYED FEEDBACK IN PRESENCE OF REFRACTORINESS. Kseniia Kravchuk and Alexander Vidybida

Introduction to Neural Networks. Daniel Kersten. Lecture 2. Getting started with Mathematica. Review this section in Lecture 1

Probabilistic Models in Theoretical Neuroscience

Supporting Online Material for

CSE/NB 528 Final Lecture: All Good Things Must. CSE/NB 528: Final Lecture

Decoding. How well can we learn what the stimulus is by looking at the neural responses?

DISCRETE EVENT SIMULATION IN THE NEURON ENVIRONMENT

Localized Excitations in Networks of Spiking Neurons

Hierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References

Activity of any neuron with delayed feedback stimulated with Poisson stream is non-markov

Frequency Adaptation and Bursting

How to do backpropagation in a brain

Dynamical Constraints on Computing with Spike Timing in the Cortex

Infinite systems of interacting chains with memory of variable length - a stochastic model for biological neural nets

Mathematical Models of Dynamic Behavior of Individual Neural Networks of Central Nervous System

AT2 Neuromodeling: Problem set #3 SPIKE TRAINS

+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing

Circular symmetry of solutions of the neural field equation

Analytical Integrate-and-Fire Neuron Models with Conductance-Based Dynamics for Event-Driven Simulation Strategies

Coherence detection in a spiking neuron via Hebbian learning

Event-driven simulations of nonlinear integrate-and-fire neurons

Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5

High-conductance states in a mean-eld cortical network model

Fast and exact simulation methods applied on a broad range of neuron models

Effects of synaptic conductance on the voltage distribution and firing rate of spiking neurons

Phase Response. 1 of of 11. Synaptic input advances (excitatory) or delays (inhibitory) spiking

Fast classification using sparsely active spiking networks. Hesham Mostafa Institute of neural computation, UCSD

STDP Learning of Image Patches with Convolutional Spiking Neural Networks

Basic elements of neuroelectronics -- membranes -- ion channels -- wiring

Lecture 11 : Simple Neuron Models. Dr Eileen Nugent

Efficient Digital Neurons for Large Scale Cortical Architectures

Interspike interval distributions of spiking neurons driven by fluctuating inputs

Neuronal Dynamics: Computational Neuroscience of Single Neurons

Figure 1: Equivalent electric (RC) circuit of a neurons membrane

Dendritic computation

Dynamic Stochastic Synapses as Computational Units

Compartmental Modelling

THE TRANSFER AND PROPAGATION OF CORRELATED NEURONAL ACTIVITY

Sampling-based probabilistic inference through neural and synaptic dynamics

Exercises. Chapter 1. of τ approx that produces the most accurate estimate for this firing pattern.

CISC 3250 Systems Neuroscience

Synchronization, oscillations, and 1/ f noise in networks of spiking neurons

Population Coding. Maneesh Sahani Gatsby Computational Neuroscience Unit University College London

Neural Modeling and Computational Neuroscience. Claudio Gallicchio

Different Types of Noise in Leaky Integrate-and-Fire Model of Neuronal Dynamics with Discrete Periodical Input

NE 204 mini-syllabus (weeks 4 8)

Ranking Neurons for Mining Structure-Activity Relations in Biological Neural Networks: NeuronRank

Simple neuron model Components of simple neuron

Balance of Electric and Diffusion Forces

How to read a burst duration code

Spike-Frequency Adaptation: Phenomenological Model and Experimental Tests

Using Variable Threshold to Increase Capacity in a Feedback Neural Network

When do Correlations Increase with Firing Rates? Abstract. Author Summary. Andrea K. Barreiro 1* and Cheng Ly 2

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.

Causality and communities in neural networks

Decoding Synchronized Oscillations within the Brain: Phase-Delayed Inhibition Provides a Robust Mechanism for Creating a Sharp Synchrony Filter

Extracting Oscillations: Neuronal Coincidence Detection with Noisy Periodic Spike Input

Structure and Measurement of the brain lecture notes

Maximum Likelihood Estimation of a Stochastic Integrate-and-Fire Neural Model

Communicated by John Rinzel

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

At the Edge of Chaos: Real-time Computations and Self-Organized Criticality in Recurrent Neural Networks

How Spike Generation Mechanisms Determine the Neuronal Response to Fluctuating Inputs

Fitting a Stochastic Neural Network Model to Real Data

Modelling stochastic neural learning

!) + log(t) # n i. The last two terms on the right hand side (RHS) are clearly independent of θ and can be

System Identification of Adapting Neurons

Transcription:

Computing with Inter-spike Interval Codes in Networks of Integrate and Fire Neurons Dileep George a,b Friedrich T. Sommer b a Dept. of Electrical Engineering, Stanford University 350 Serra Mall, Stanford, CA 94305, USA email: dil@stanford.edu b Redwood Neuroscience Institute 1010 El Camino Real, Menlo Park, CA 94025, USA email: {dgeorge, fsommer}@rni.org Abstract Information encoding in spikes and computations performed by spiking neurons are two sides of the same coin and should be consistent with each other. This study uses this consistency requirement to derive some new results for Inter-spike Interval (ISI) coding in networks of Integrate and Fire (IF) neurons. Our analysis shows that such a model can carry out useful computations and that it does also account for variability in spike timing as observed in cortical neurons. Our general result is that IF type neurons, though highly non-linear, perform a simple linear weighted sum operation of ISI coded quantities. Further we can derive bounds on the variation of ISI s that occur in the model although the neurons are deterministic. We also derive useful estimates of the maximum processing speed in a hierarchical network. Key words: ISI Coding, Integrate and Fire Neuron, Spike Timing, Cortical Variability. 1 Introduction It is well known that neurons communicate using spikes. But how is information encoded in the spikes that travel from one neuron to another? Since outputs of neurons are fed to other neurons for further processing, a generic information encoding scheme will need to have the same encoding mechanism at the input to a neuron as at the output. Given such an encoding scheme, neurons of the same type can be organized in multiple layers to perform useful computations [2]. A computational model for a neuron should, therefore, perform computations on its input spike trains and produce the result of the Preprint submitted to Elsevier Preprint 26 January 2004

computation at its output with the same encoding as at its inputs. Thus, information encoding, computation and neuron models are part of the same problem and should be consistent with each other. As an example, consider the Poisson rate coding hypothesis where the spike trains are modelled as Poisson processes, with the rate of the process carrying the relevant information [4]. Softky and Koch showed that the rate coding hypothesis is inconsistent with Integrate and Fire neuron models [6]. A neuron model which is consistent with the rate coding hypothesis would require a random number generator inside the neuron so that the outputs are Poisson aswell. This example illustrates that it is important to consider neuron models together with an information encoding and processing mechanism. What is a consistent kind of coding mechanism for the widely used Integrate and Fire (IF) neurons? What computation do IF neurons perform with its inputs and its outputs coded in that fashion? In this paper we show that Inter-spike Interval (ISI) coding is a viable coding mechanism that can produce useful computations in networks of IF neurons while being consistent with the above requirements. In addition, we show that the ISI coding and computation mechanisms as suggested here could account for spike timing variability as observed in cortical spike trains. The rest of this paper is organized as follows. Section 2 establishes the general relationship between ISI coding and IF neurons. Section 3 describes the nature of computations performed on ISI coded streams by IF neurons. This section also describes the source of variability arising from these computations. Section 4 compares these results with previous work and suggests topics for further investigation. 2 ISI Coding and Integrate-and-Fire Neurons One way to encode a stream of non-negative numbers s 1, s 2, s N based on ISIs is to use these values to control the time elapsed between successive spikes as given by t n+1 t n = f(s n+1 ) (1) where t n s are the spike instants and f is a positive valued function. Consider an integrate and fire (IF) neuron without leak driven by a constant input current I. Such a neuron would produce output spikes at regular intervals T out given by T out = V θ C/I (2) 2

where V θ is the threshold of firing and C is the capacitance of the neuron. We assume that after firing, the membrane potential is reset to zero. The corresponding expression for a leaky integrate and fire neuron (LIF) with leakage resistance R is given by [7], 1 T out = RCln( 1 (V θ /IR) ) (3) Equations 2 and 3 satisfy the general form of ISI encoding given in equation 1 1 with the f(s) = V θ C/s for equation 2 and and f(s) = RCln( ) for 1 (V θ /sr) equation 3. In both cases, T out decreases with increasing I, i.e, f is a decreasing function of the input magnitude. Inputs to cortical neurons are spikes rather than constant currents. Moreover, any given neuron will have many input synapses carrying different spike streams. Therefore, it is important to know how the output spike intervals of a neuron is related to the the input spike intervals on its synapses. In the next section we derive this relationship for an IF neuron with two synaptic inputs. 3 Computations with ISI encoded input streams Consider ISI streams encoding static values s 1 and s 2 applied as inputs to an IF neuron with synaptic weights w 1 and w 2. We assume that streams s 1 and s 2 were produced by IF neurons with the same threshold and capacitance. Initially we consider neurons with no leakage. We then know that values s 1 and s 2 correspond to input interspike intervals T 1 and T 2 where T 1 = V θ C/s 1 and T 2 = V θ C/s 2. Let T out be an output inter-spike interval for these inputs. We assume that the firing threshold is such that T out > 2max{T 1, T 2 }. Then during the interval T out, synapse 1 received approximately T out /T 1 spikes and synapse 2 received approximately T out /T 2 spikes. These input spikes increment the membrane potential of the neuron eventually making it reach the threshold and fire T out units after its previous firing. If each spike causes an increment V δ in the membrane potential, on an average, the neuron fires when the following condition is met w 1 (T out /T 1 ) + w 2 (T out /T 2 ) = V θ /V δ (4) Thus, using equation 2, the quantity s out encoded by the output ISI T out can be derived as s out = (V δ /C)( w 1 T 1 + w 2 T 2 ) = (V δ /V θ )(w 1 s 1 + w 2 s 2 ) (5) 3

If we assume the ratio V δ /V θ to be the same for all neurons producing the input spike streams, a general approximate input-output relationship holds for a neuron with multiple inputs: s out = Σ i w i s i, where i runs over all synapses of this neuron. Thus, even though the integrate and fire neuron is highly nonlinear, the input-output relation between the spike intervals remains a linear function. The above derivations of output ISIs are only approximate since we did not consider the cases where two simultaneous input spikes fire an output spike. The maximum output ISI T max will occur subsequent to a spike that was fired by two coincident spikes on the input streams. The next firing, which terminates the ISI corresponding to T max would then occur synchronously with spikes on either of the synapses. Without loss of generality, we assume that this happens synchronously at synapse 1. Then T max = n 1 T 1, where n 1 is the number of spikes received at synapse 1. Further, the number of spikes received on synapse 2 during this interval is given by n 2 = T max /T 2. Equating the weighted number of spikes to the firing threshold we get an upper bound on the maximum output inter-spike interval T max. Similarly, a lower bound for the minimum ISI T min can be derived. The two bounds are as given below by T max (V θ/v δ ) + max{w 1, w 2 } (w 1 /T 1 ) + (w 2 /T 2 ) T min (V θ/v δ ) max{w 1, w 2 } (w 1 /T 1 ) + (w 2 /T 2 ) (6) where V θ is the firing threshold and V δ is the increment in membrane potential caused by an incoming spike. Although the exact probability density of the ISI variation was not derived, from symmetry arguments we conclude that the average output ISI is the one given by equation 4. This implies a new way of interpreting the output of an IF neuron with static ISI coded inputs. The output ISIs, on an average, represent the weighted sum of the quantities represented by the input ISIs. The output ISIs vary about this average and the variance depends on the relative magnitude of the voltage increments caused by an input spike. This jitter could account for part of the variability of cortical spike trains. The stated results can be extended to time varying input streams by imposing appropriate constraints on the maximum rate of change of the input streams. Since an IF neuron loses all the history of inputs after firing, a step change at any of the input streams is reflected at the output within a time t 2Tout max where Tout max is the maximum ISI of the output spike stream. Let Tout max = KTin max where Tin max is the maximum ISI at the input and K 2 is a constant that depends on the time constant of the neuron and its synaptic strengths. If we assume that the input spike trains are faithful samples of a bandlimited input current, then by the sampling theorem, Tin max 1/(2f max ), where f max is the maximum frequency component of the input current. Then Tout max 4

(K/2)f max. This means that the maximum frequency that can be represented by the spike trains drops by a factor of K 2 for every layer of IF neurons in a layered architecture. If this condition is violated for any layer, the output spike stream can no longer be interpreted as ISI coded as described in this paper. Knowing the minimum rate at which neurons at the highest level of the hierarchy should respond would give us an upper bound on the input frequencies that can be faithfully represented and processed using the ISI coding scheme described here. The result that the basic type of computation is of the sort s out = w 1 s 1 +w 2 s 2 is not only valid for IF neurons, but also holds for LIF neurons. For a LIF neuron with leakage resistance R and capacitance C, the membrane voltage decays exponentially with a time constant τ = RC as given in equation 3. Therefore, on an average, the neuron fires when the following condition is met w 1 (1 + e T 1 τ + e 2 T 1 τ +w 2 (1 + e T 2 τ + e 2 T 2 τ + + e ( T out T 1 1) T 1 τ ) + + e ( T out 1) T 2 T 2 τ )= w 1 (e Tout/τ 1)( e T 1/τ 1 + w 2 e T 2/τ 1 )=V θ/v δ By comparing this with equation 3, the quantity s out encoded by the output ISI T out becomes w 1 s out = (V δ /R)( e T 1/τ 1 + w 2 e T 2/τ 1 ) = (V δ/v θ )(w 1 s 1 + w 2 s 2 ) (7) 4 Discussion The question of how spike timing is used for computations has been under debate for a long time [1 3]. The challenge lies in finding coding and computational mechanisms that are consistent with neuron models. Rate coding with Poisson distributed spikes has been suggested as one mechanism for encoding information in spikes [4]. But the traditional view of Poisson spikes cannot be implemented with IF neurons [6]. The ISI coding mechanism we suggested for IF neurons can be interpreted as an instantaneous rate code. However, in our scheme the spikes are more precisely timed and hence can carry information at a higher rate compared to Poisson rate coding [5]. Rank Order Coding(ROC) [2] is another spike timing based computation suggested in the literature. However, ROC has the disadvantage that a slight delay in the onset of one of the inputs can disrupt the entire chain. Poisson rate coding and ROC can be considered to be at the opposite extremes of a continuum of spike coding mechanisms. The ISI coding we suggested here can be positioned 5

somewhere between these extremes, thus giving a tradeoff between information rate and synchronicity requirements. Since ISI codes are nothing but rate codes with precise timing, all the results we derived in section 3 are valid in general for all rate codes. There has been extensive discussion about the nature of variability of spike trains in cortex [6,4]. Suggested mechanisms for this variability include random firing thresholds and stochastic synapses. The ISI coding mechanisms we suggested here demonstrate that the variability of the spike trains could, at least in part, arise as a natural consequence of computations based on ISI coded spike stream inputs, even in entirely deterministic IF neurons. More studies are required to verify what percentage of the observed cortical variability of spike trains can be explained using this approach. We know that neurons in the lower levels of a cortical hierarchy need to respond faster compared to neurons higher up in the hierarchy. Our model predicts a drop in the maximum frequency as more computational layers are stacked like in a hierarchical architecture. Clearly, there is a tradeoff between speed of processing and accuracy of representation at each stage. Further studies are required to quantify this tradeoff. Also, the question of how learning could occur in ISI coded networks is left as a topic for further study. References [1] Carlos. D. Brody and J. J. Hopfield. Simple networks for spike-timing-based computation, with application to olfactory processing. Neuron, 37(5):843 852, Mar 2003. [2] A. Delorme and S. J. Thorpe. Face identification using one spike per neuron: resistance to image degradations. Neural Netw, 14(6-7):795 803, Jul 2001. [3] Wolfgang Maass. On the relevance of time in neural computation and learning. Theoretical Computer Science, 261(1):157 178, 2001. [4] M. N. Shadlen and W. T. Newsome. Noise, neural codes and cortical organization. Curr Opin Neurobiol, 4(4):569 579, Aug 1994. [5] W. R. Softky. Simple codes versus efficient codes. Curr Opin Neurobiol, 5(2):239 247, Apr 1995. [6] W. R. Softky and C. Koch. The highly irregular firing of cortical cells is inconsistent with temporal integration of random EPSPs. J Neurosci, 13(1):334 350, Jan 1993. [7] R. B. Stein. The frequency of nerve action potentials generated by applied currents. Proc R Soc Lond B Biol Sci, 167(6):64 86, Jan 1967. 6