Adaptive contrast gain control and information maximization $

Similar documents
Design principles for contrast gain control from an information theoretic perspective

ScholarOne, 375 Greenbrier Drive, Charlottesville, VA, 22901

Dynamical mechanisms underlying contrast gain control in single neurons

Information Theory. Mark van Rossum. January 24, School of Informatics, University of Edinburgh 1 / 35

1/12/2017. Computational neuroscience. Neurotechnology.

encoding and estimation bottleneck and limits to visual fidelity

What do V1 neurons like about natural signals

Efficient Spike-Coding with Multiplicative Adaptation in a Spike Response Model

Estimation of information-theoretic quantities

The Spike Response Model: A Framework to Predict Neuronal Spike Trains

Membrane equation. VCl. dv dt + V = V Na G Na + V K G K + V Cl G Cl. G total. C m. G total = G Na + G K + G Cl

Do Neurons Process Information Efficiently?

Efficient coding of natural images with a population of noisy Linear-Nonlinear neurons

How to read a burst duration code

Exercises. Chapter 1. of τ approx that produces the most accurate estimate for this firing pattern.

!) + log(t) # n i. The last two terms on the right hand side (RHS) are clearly independent of θ and can be

Limulus. The Neural Code. Response of Visual Neurons 9/21/2011

Fisher Information Quantifies Task-Specific Performance in the Blowfly Photoreceptor

SPIKE TRIGGERED APPROACHES. Odelia Schwartz Computational Neuroscience Course 2017

Efficient coding of natural images with a population of noisy Linear-Nonlinear neurons

Nonlinear reverse-correlation with synthesized naturalistic noise

Modeling of Retinal Ganglion Cell Responses to Electrical Stimulation with Multiple Electrodes L.A. Hruby Salk Institute for Biological Studies

SUPPLEMENTARY INFORMATION

Statistical models for neural encoding

What is the neural code? Sekuler lab, Brandeis

Efficient and direct estimation of a neural subunit model for sensory coding

Analyzing Neuroscience Signals using Information Theory and Complexity

Statistical models for neural encoding, decoding, information estimation, and optimal on-line stimulus design

arxiv: v1 [q-bio.nc] 9 Mar 2016

Spike Count Correlation Increases with Length of Time Interval in the Presence of Trial-to-Trial Variation

Dimensionality reduction in neural models: An information-theoretic generalization of spike-triggered average and covariance analysis

arxiv:cond-mat/ v2 27 Jun 1997

Dimensionality reduction in neural models: an information-theoretic generalization of spiketriggered average and covariance analysis

arxiv:physics/ v1 [physics.data-an] 7 Jun 2003

Spike-Frequency Adaptation: Phenomenological Model and Experimental Tests

Encoding or decoding

Combining biophysical and statistical methods for understanding neural codes

Natural Image Statistics

High-conductance states in a mean-eld cortical network model

Comparing integrate-and-fire models estimated using intracellular and extracellular data 1

Lecture 11: Continuous-valued signals and differential entropy

Neural Coding: Integrate-and-Fire Models of Single and Multi-Neuron Responses

Natural Image Statistics and Neural Representations

Linearization of F-I Curves by Adaptation

Effects of Betaxolol on Hodgkin-Huxley Model of Tiger Salamander Retinal Ganglion Cell

Neural coding Ecological approach to sensory coding: efficient adaptation to the natural environment

Maximum Likelihood Estimation of a Stochastic Integrate-and-Fire Neural Model

Activity Driven Adaptive Stochastic. Resonance. Gregor Wenning and Klaus Obermayer. Technical University of Berlin.

Adaptation in the Neural Code of the Retina

Optimal quantization for energy-efficient information transfer in a population of neuron-like devices

CSE/NB 528 Final Lecture: All Good Things Must. CSE/NB 528: Final Lecture

Searching for simple models

Finding a Basis for the Neural State

1 Balanced networks: Trading speed for noise

Efficient Coding. Odelia Schwartz 2017

When is an Integrate-and-fire Neuron like a Poisson Neuron?

Emergence of resonances in neural systems: the interplay between adaptive threshold and short-term synaptic plasticity

Another look at the retina as an image dithered scalar quantizer

Neural Modeling and Computational Neuroscience. Claudio Gallicchio

Decoding Poisson Spike Trains by Gaussian Filtering

Second Order Dimensionality Reduction Using Minimum and Maximum Mutual Information Models

Computing with inter-spike interval codes in networks ofintegrate and fire neurons

Modeling and Characterization of Neural Gain Control. Odelia Schwartz. A dissertation submitted in partial fulfillment

THE retina in general consists of three layers: photoreceptors

Introduction to neural spike train data for phase-amplitude analysis

Information maximization in a network of linear neurons

Data assimilation with and without a model

Synaptic Input. Linear Model of Synaptic Transmission. Professor David Heeger. September 5, 2000

Neural Encoding. Mark van Rossum. January School of Informatics, University of Edinburgh 1 / 58

Bayesian probability theory and generative models

Comparison of objective functions for estimating linear-nonlinear models

Pacific Symposium on Biocomputing 6: (2001)

Biological Modeling of Neural Networks

Computer Vision & Digital Image Processing

Intrinsic gain modulation and adaptive neural coding. Abstract. Sungho Hong 1,2, Brian Nils Lundstrom 1 and Adrienne L. Fairhall 1

Mid Year Project Report: Statistical models of visual neurons

Modeling Convergent ON and OFF Pathways in the Early Visual System

Inferring synaptic conductances from spike trains under a biophysically inspired point process model

Subthreshold cross-correlations between cortical neurons: Areference model with static synapses

Visual motion processing and perceptual decision making

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Neural characterization in partially observed populations of spiking neurons

Transformation of stimulus correlations by the retina

Introduction to Neural Networks U. Minn. Psy 5038 Spring, 1999 Daniel Kersten. Lecture 2a. The Neuron - overview of structure. From Anderson (1995)

Action Potential Initiation in the Hodgkin-Huxley Model

Lateral organization & computation

Modeling Surround Suppression in V1 Neurons with a Statistically-Derived Normalization Model

Computation in a single neuron: Hodgkin and Huxley revisited

Analyzing Neural Responses to Natural Signals: Maximally Informative Dimensions

Information Maximization in Single Neurons

Broadband coding with dynamic synapses

SUPPLEMENTARY INFORMATION

System Identification of Adapting Neurons

Neural variability and Poisson statistics

Neural Encoding: Firing Rates and Spike Statistics

Gatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II

Learning quadratic receptive fields from neural responses to natural signals: information theoretic and likelihood methods

Entropy and information in neural spike trains: Progress on the sampling problem

Comparison of receptive fields to polar and Cartesian stimuli computed with two kinds of models

Transcription:

Neurocomputing 65 66 (2005) 6 www.elsevier.com/locate/neucom Adaptive contrast gain control and information maximization $ Yuguo Yu a,, Tai Sing Lee b a Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh PA 523, USA b Computer Science Department, Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh PA 523, USA Available online 2 December 2004 Abstract Contrast gain control is an important mechanism underlying the visual system s adaptation to contrast of luminance in varying visual environments. Our previous work showed that the threshold and saturation determine the preferred contrast sensitivity as well as the maximum information coding capacity of the neuronal model. In this report, we investigated the design principles underlying adaptation behavior in contrast gain control by an adaptive linear nonlinear model. We found that an adaptive rescaling mechanism predicted by information transmission maximization can explain a variety of observed contrast gain control phenomena in neurophysiological experiments, including the divisive input output relations, and the inverse power-law relation between response gain and input contrast. Our results suggest that contrast gain control in visual systems might be designed for information maximization. r 2004 Elsevier B.V. All rights reserved. Keywords: Adaptation; Contrast gain control; Information maximization $ This research was supported by NSF CAREER 9984706, NIH Vision Research core Grant EY08098, and a NIH grant on the Pittsburgh Supercomputing Center. Corresponding author. E-mail address: yug@cnbc.cmu.edu (Y. Yu). 0925-232/$ - see front matter r 2004 Elsevier B.V. All rights reserved. doi:0.06/j.neucom.2004.0.062

2 Y. Yu, T. Sing Lee / Neurocomputing 65 66 (2005) 6. Introduction The visual systems exhibit great flexibility in adapting their input output functions to the variance or the contrast [6,9,0,2] of light signal in the visual environment. The amplitude gain of the transfer function of visual neurons was found to decrease with input variance [9], and even displays an inverse power-law relation [2] (see Fig. 2). In addition, the contrast response function in visual cortical neurons was found to shift along the log contrast axis to adaptively match the prevailing input signals [6]. These contrast adaptation phenomena have been widely observed in the retina [9], striate [6], and extrastriate visual cortex [5] of mammals, and fly H neurons [,3] suggesting the existence of an contrast gain control mechanism. However, the biophysical factors and computational rules governing its operation remain elusive. How to understand these phenomena? Why do visual neurons behave in this way? We have shown that classical neuronal model with nonlinearities of threshold and saturation can display an apparent adaptation in the recovered transfer function to the variance of the input signal [3,4]. However, we found that the mutual information in those system, as well as the kernel gain, exhibit a bell-shape as a function of the input variance [6]. There is only one optimal variance that can induce the maximal information transmission of the system. That means nonlinear dynamics due to static nonlinearity alone cannot maintain a neuron s information transmission rate at the highest level with the variation in signal variance. Here, we explored theoretically a possible extra adaptation mechanism that might restore information maximization for any input variance in neurons. We introduce an additional gain rescaling to the static model, while keeping all other parameters (thresholds, saturation levels, membrane time constants) constant. We found it sufficient to produce information maximization that is consistent with some neurophysiological findings [,3]. 2. Model and results We use the similar standard linear nonlinear (LN) cascade model as in Ref. [4] except that we introduce an additional scaling factor bðsþ: See Fig.. b scales the Fig.. The adaptive LN model consists of a linear filter hðtþ followed by a nonlinearity gð:þ: The amplitude of the linear filter hðtþ is controlled by b; which acts as an adaptive mechanism. xðtþ is the convolution of the input signal sðtþ and the filter hðtþ: The nonlinearity gð:þ operates on xðtþ to generate the output yðtþ:

amplitude of the linear kernel according to the variance of the input signal as an adaptive mechanism. Linear kernel is given by hðtþ ¼bðsÞA sinðpt=t a Þ expð t=t b Þ with A ¼ ; t a ¼ 80 ms and t b ¼ 00 ms: The output of linear kernel xðtþ is given by xðtþ ¼ R þ 0 hðtþsðt tþ dt: The nonlinearity gð:þ is specified by 8 >< 0 if xoy; gðxþ ¼ x y if ypxoz; () >: Z y if xxz; where y is the threshold and Z is the saturation level. Without loss of generality, we use a Gaussian white noise stimulus sðtþ with zero mean and SD s as the input signal. Its probability density function (PDF) is given by pðsþ ¼pffiffiffiffiffiffiffiffiffiffi e s2 =2s 2 : 2ps 2 The linear response xðtþ also has a Gaussian distribution with PDF pðxþ ¼pffiffiffiffiffiffiffiffiffiffi e x2 =2s 2 x ; 2ps 2 x where s x is given by Z þ s 2 x ¼hx2 ðtþi ¼ s 2 h 2 ðtþ dt; 0 ARTICLE IN PRESS Y. Yu, T. Sing Lee / Neurocomputing 65 66 (2005) 6 3 where h i denotes time average. b is determined computationally by maximizing the mutual information between sðtþ and yðtþ with respect to b; for signals of a given s: This is the adaptive LN model. We have shown the amplitude gain of the recovered linear kernel of the static model is a function of the input signal variance s 2 [3]. The ratio a between the amplitude gain of the recovered linear kernel h 0 ðtþ and the original linear kernel hðtþ is given by R Z y a ¼ xðx yþpðxþ dx þðz yþ R þ Z xpðxþ dx s R 2 þ 0 h 2 : (2) ðtþ dt That is, for an input signal with Gaussian white distribution, we have h 0 ðtþ ¼a:hðtÞ; where gain factor a quantifies how the performance of linear function hðtþ is affected by the static nonlinearity (threshold y; saturation Z) and stimulus standard deviation s: Mutual information [8] quantifies the ability of a system or a communication channel to convey information and is given by I m ¼ HðyÞ Hðy j sþ ¼ X PðyÞ log 2 PðyÞþ X PðsÞPðy j sþ log 2 Pðy j sþ; ð3þ y s;y where HðyÞ is the total entropy of the response, Hðy j sþ is noise entropy, accounting for the variability in the response that is not due to variations in the stimulus, but comes from other noise sources. For simplicity, we consider the noiseless case, where

4 Y. Yu, T. Sing Lee / Neurocomputing 65 66 (2005) 6 Hðy j sþ ¼0: In this case, the mutual information is simply equal to the output entropy I m ¼ HðyÞ: We are interested in computing the entropy of the output y. The distribution function of y (also refereed to as the cumulative density function) is 8 0 if yo0; >< R yþy DðyÞ ¼P½Y pyš ¼ pðxþ dx if 0pyoZ y; (4) >: if yxz y: In order to compute the entropy of y, we approximate y by a discrete random variable ^y; which we define by P½ ^y ¼ idyš ¼P½y 2½ði ÞDy; idyšš; where Dy is the quantization resolution of y, and i is an integer from 0 to dðz yþ=ðdyþe: Note that P½y 2ða; bšš can be computed analytically as DðbÞ DðaÞ: The mutual information, which is equal to entropy for the noiseless case, is then computed as in Eq. (3). It can be shown that changes in Dy affect the entropy of ^y only by an additive constant [2]. Since we are only interested in comparing the values of mutual information I m as we change s; y; and Z; our choice of Dy does not affect our results. Here, we set Dy ¼ : What is the effect of different amplitude of b on the information transmission? Fig. 2a shows that there is a non-monotonous relation between b and I m for each given s: There exists an optimal value of b; named b opt where information I m is Im (Bits) 4 3 2 = = 2 = 5 = 0 = 5 β opt = 20 Im (Bits) 4 3 Adaptive Model Static Model (a) 0 5 0 β (b) 0 0 20 Im (Bits) 4 3 θ = 0; η = 00 θ = 0; η = 50 θ = 0; η = 50 θ = 20; η = 50 β opt 0 θ = 0 θ = 5 θ = 0 θ = 20 (c) 2 0 0 20 (d) 0 Fig. 2. (a) For an adaptive LN model with threshold y ¼ 0 and saturation Z ¼ 50; mutual information I as a function of rescaling factor b for signals of different s s. (b) The mutual information I as a function of s for the adaptive LN model and the static LN model. (c) Mutual information I as a function of s for various model parameters with b chosen to maximize I for each s: (d) Optimal b as a function of s for four given values of threshold in a log log plot exhibits an inverse power law.

Y. Yu, T. Sing Lee / Neurocomputing 65 66 (2005) 6 5 Gain γ θ = 0 θ = 5 θ = 0 θ = 20 Monocular Gain 0 Cell Cell 2 0. (a) 0 (b) 0 Contrast [%] Fig. 3. (a) Gain g as a function of s for four given values of threshold with b optimized for information transmission. (b) Monocular gain in two cat simple cells as a function of stimulus contrast (from 2:5% to 50%) in the log log axes; for comparison, a line with slope of is plotted in dashed line. Adapted from Fig. 5. of Ref. [2]. maximized. Let s now assume that the biological goal of an adaptive neuron is to maximize information transmission. That is, for any input s; the neuron will adjust its linear function amplitude to get b opt : Fig. 2b shows that the information transmission Imax for such an adaptive LN model is maintained at the highest level for any given variance. Note that for the static LN model with b ¼ ; I m varies with s; with only one global maximum I max at a particular s (see Fig. 2b). The signal with nonoptimal value cannot be efficiently processed by the system with high transmission rate. However, the adaptive model can ensure that signal with any distribution s can always be processed with highest information transmission rate. The highest value of information transmission is constrained only by the threshold and the saturation of the model itself [8,9]. The lower (or higher) is the threshold (saturation), the higher the maximum information rate can be reached (Fig. 2c). The adaptation factor b opt decreases monotonously with an increase in s: Interestingly, in the log-log plot, b opt and s displays a inverse power law relationship, the fitting slope is (Fig. 2d). The response gain g; i.e., the amplitude of the recovered kernel, now equal to a b (see Eq. (2)), also displays a scale invariant power-law relationship with a scale with s (Fig. 3a). Note that without gain rescaling, i.e., b ¼ is fixed, the gain g varies with s; displaying a bell-like tuning curve, which emerges simply from staticnonlinearity. Gain rescaling is crucial for maintaining mutual information constant with variation in s: This power law finding matches the experimental observation (see Fig. 3b) very well, thus implying that the contrast gain adaptation observed in experiments may reflect an underlying efficient encoding process. Fig. 3b is adapted from Fig. 5a in Ref. [2], which shows that response gain of cat simple cell varying with the contrast of monocular input also in power-law relationship in the log-contrast plot. A contrast gain slope of indicates that the gain control system is 00% effective, whereas a slope of 0 indicates that the system is completely ineffective. The cells in Fig. 3b showed a slope around 0:75 suggesting the neurons might be trying to re-scale gain adaptively in order maximize information transmission.

6 Y. Yu, T. Sing Lee / Neurocomputing 65 66 (2005) 6 3. Discussion The power-law relationship observed experimentally and the theoretical findings of this paper suggests that there might exist an intrinsic adaptation mechanism in single neurons or network to maximize the information encoding. Such an adaptation, which is independent of static nonlinearity, involves mainly changing of the amplitude of the linear transfer function. Biophysically, a single neuron itself is a circuit, whose transfer function is controlled by the ionic conductances and capacitances. Recent experiments showed that the activation of Na þ -activated and Ca 2þ -activated K þ currents inside single cells [4,7] or the adjustment of voltagedependent conductance might play an important role in contrast adaptation []. These mechanisms might serve to control the gain of the linear kernel part of the LN cascade. In summary, the contrast adaptation phenomena observed in neurophysiological experiments can potentially be factored into a component due simply to the static nonlinearities of the neurons [6,9] and another component due to an adaptive rescaling of the gain in the linear transfer function in the LN cascade for the purpose of information maximization. References [] N. Brenner, W. Bialek, R. de Ruyter van Steveninck, Adaptive rescaling maximizes information transmission, Neuron 26 (2000) 695 702. [2] P. Dayan, L. F. Abbott, Theoretical Neuroscience, MIT Press, Cambridge, 200, pp. 29 (Chapter 4). [3] A.L. Fairhall, G.D. Lewen, W. Bialek, R. de Ruyter van Steveninck, Efficiency and ambiguity in an adaptive neural code, Nature 42 (200) 787 792. [4] K.J. Kim, F. Rieke, Temporal contrast adaptation in the input and output signals of salamander retinal ganglion cells, J. Neurosci. 2 (200) 287 299. [5] A. Kohn, J.A. Movshon, Neuronal adaptation to visual motion in area MTof the macaque, Neuron 39 (2003) 68 69. [6] I. Ohzawa, G. Sclar, R.D. Freeman, Contrast gain control in the cat s visual system, J. Neurophysiol. 54 (985) 65 667. [7] M.V. Sanchez-Vives, L.G. Nowak, D.A. McCormick, Cellular mechanisms of long lasting adaptation in visual cortical neurons in vitro, J. Neurosci. 20 (2000) 4286 C4299. [8] C.E. Shannon, W. Weaver, The Mathematical Theory of Communication, University of Illinois Press, Urbana, IL, 949. [9] R.M. Shapley, J.D. Victor, The contrast gain control of the cat retina, Vision Res. 9 (979) 43 434. [0] S.M. Smirnakis, M.J. Berry, D.K. Warland, W. Bialek, M. Meister, Adaptation of retinal processing to image contrast and spatial scale, Nature 386 (997) 69 73. [] M. Stemmler, C. Koch, How voltage-dependent conductances can adapt to maximize the information encoded by neuronal firing rate, Nature Neurosci. 2 (999) 52 527. [2] A.M. Truchard, I. Ohzawa, R.D. Freeman, Contrast gain control in the visual cortex: monocular versus binocular mechanisms, J. Neurosci. 2 (2000) 287 299. [3] Y.G. Yu, T.S. Lee, Dynamical mechanisms underlying contrast gain control in single neurons, Phys. Rev. E 68 (2003) 90 907. [4] Y.G. Yu, B. Potetz, T.S. Lee, The role of spiking nonlinearity in contrast gain control and information transmission, Vision Res. 2004, in press.