Synaptic Plasticity. Introduction. Biophysics of Synaptic Plasticity. Functional Modes of Synaptic Plasticity. Activity-dependent synaptic plasticity:

Similar documents
Plasticity and Learning

Outline. NIP: Hebbian Learning. Overview. Types of Learning. Neural Information Processing. Amos Storkey

How do biological neurons learn? Insights from computational modelling of

Learning and Meta-learning

Hebb rule book: 'The Organization of Behavior' Theory about the neural bases of learning

Neural networks: Unsupervised learning

The Hebb rule Neurons that fire together wire together.

Iterative face image feature extraction with Generalized Hebbian Algorithm and a Sanger-like BCM rule

Financial Informatics XVII:

Novel VLSI Implementation for Triplet-based Spike-Timing Dependent Plasticity

arxiv: v1 [cs.ne] 30 Mar 2013

Introduction to Neural Networks

Triplets of Spikes in a Model of Spike Timing-Dependent Plasticity

Introduction to Machine Learning

An Introductory Course in Computational Neuroscience

Covariance and Correlation Matrix

Effect of Correlated LGN Firing Rates on Predictions for Monocular Eye Closure vs Monocular Retinal Inactivation

CSE/NB 528 Final Lecture: All Good Things Must. CSE/NB 528: Final Lecture

Slide10. Haykin Chapter 8: Principal Components Analysis. Motivation. Principal Component Analysis: Variance Probe

Slide02 Haykin Chapter 2: Learning Processes

Hebbian Learning II. Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester. July 20, 2017

Beyond Hebbian plasticity: Effective learning with ineffective Hebbian learning rules

The Variance of Covariance Rules for Associative Matrix Memories and Reinforcement Learning

Design and Implementation of BCM Rule Based on Spike-Timing Dependent Plasticity

Training and spontaneous reinforcement of neuronal assemblies by spike timing

Computational Neuroscience. Structure Dynamics Implementation Algorithm Computation - Function

+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing

AS Elementary Cybernetics. Lecture 4: Neocybernetic Basic Models

Effective Neuronal Learning with Ineffective Hebbian Learning Rules

Copyright by. Changan Liu. May, 2017

Consider the following spike trains from two different neurons N1 and N2:

Machine Learning and Adaptive Systems. Lectures 5 & 6

Systems Biology: A Personal View IX. Landscapes. Sitabhra Sinha IMSc Chennai

Synergies Between Intrinsic and Synaptic Plasticity Mechanisms

Neuroscience Introduction

Branch-Specific Plasticity Enables Self-Organization of Nonlinear Computation in Single Neurons

Time-Skew Hebb Rule in a Nonisopotential Neuron

Math in systems neuroscience. Quan Wen

PCA, Kernel PCA, ICA

Biological Modeling of Neural Networks:

Introduction. Previous work has shown that AER can also be used to construct largescale networks with arbitrary, configurable synaptic connectivity.

Spike-Timing Dependent Plasticity and Relevant Mutual Information Maximization

Bayesian synaptic plasticity makes predictions about plasticity experiments in vivo

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.

Bayesian Computation Emerges in Generic Cortical Microcircuits through Spike-Timing-Dependent Plasticity

Processing of Time Series by Neural Circuits with Biologically Realistic Synaptic Dynamics

How do synapses transform inputs?

This script will produce a series of pulses of amplitude 40 na, duration 1ms, recurring every 50 ms.

Synaptic Devices and Neuron Circuits for Neuron-Inspired NanoElectronics

Perceptron. (c) Marcin Sydow. Summary. Perceptron

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

An objective function for self-limiting neural plasticity rules.

Linearization of F-I Curves by Adaptation

Artificial Neural Networks Examination, March 2004

Data Mining Part 5. Prediction

In this section, we review some basics of modeling via linear algebra: finding a line of best fit, Hebbian learning, pattern classification.

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

Memories Associated with Single Neurons and Proximity Matrices

arxiv: v1 [q-bio.nc] 2 Oct 2014 Frontiers in Robotics and AI 1 INTRODUCTION

Neuron, Volume 63. Supplemental Data. Generating Coherent Patterns of Activity. from Chaotic Neural Networks. David Sussillo and L.F.

Linear Regression, Neural Networks, etc.

arxiv: v1 [q-bio.nc] 4 Jan 2016

Bayesian Computation Emerges in Generic Cortical Microcircuits through Spike-Timing-Dependent Plasticity

Supporting Online Material for

Abstract. Author Summary

Using Variable Threshold to Increase Capacity in a Feedback Neural Network

Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5

Computational Explorations in Cognitive Neuroscience Chapter 4: Hebbian Model Learning

Sampling-based probabilistic inference through neural and synaptic dynamics

Tuning tuning curves. So far: Receptive fields Representation of stimuli Population vectors. Today: Contrast enhancment, cortical processing

CS281 Section 4: Factor Analysis and PCA

PCA and admixture models

Neural Networks Introduction

Spike timing dependent plasticity - STDP

Immediate Reward Reinforcement Learning for Projective Kernel Methods

Introduction to Artificial Neural Networks

Emergence of resonances in neural systems: the interplay between adaptive threshold and short-term synaptic plasticity

Plasticity Kernels and Temporal Statistics

Connecting Epilepsy and Alzheimer s Disease: A Computational Modeling Framework

Investigations of long-term synaptic plasticity have revealed

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 28 Jul 2003

Neural Networks. Textbook. Other Textbooks and Books. Course Info. (click on

Artificial Neural Network and Fuzzy Logic

Decoding. How well can we learn what the stimulus is by looking at the neural responses?

Exercises. Chapter 1. of τ approx that produces the most accurate estimate for this firing pattern.

Marr's Theory of the Hippocampus: Part I

Basic Principles of Unsupervised and Unsupervised

Artificial Neural Networks Examination, June 2005

Linear Dimensionality Reduction

Rate- and Phase-coded Autoassociative Memory

Neuron Grids as seen as. Elastic Systems. Heikki Hyötyniemi TKK / Control Engineering Presentation at NeuroCafé March 31, 2006

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

Adaptation in the Neural Code of the Retina

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Plasticity and learning in a network of coupled phase oscillators

Face Recognition and Biometric Systems

Reducing Spike Train Variability: A Computational Theory Of Spike-Timing Dependent Plasticity

DYNAMICAL SYSTEMS OF THE BCM LEARNING RULE: by Lawrence C. Udeigwe

What Can a Neuron Learn with Spike-Timing-Dependent Plasticity?

Synaptic dynamics. John D. Murray. Synaptic currents. Simple model of the synaptic gating variable. First-order kinetics

Transcription:

Synaptic Plasticity Introduction Dayan and Abbott (2001) Chapter 8 Instructor: Yoonsuck Choe; CPSC 644 Cortical Networks Activity-dependent synaptic plasticity: underlies learning and memory, and plays a crucial role in neural circuit development. Donald Hebb: Hebb rule for synaptic plasticity (1949) neuron A contributes to firing of neuron B, then synapse between A and B should be strengthened. Subsequent activation of A will lead to stronger activation of B. Hebb s rule only increases synaptic strength. It can be generalized to weaken strength if neuron A repeatedly fails to activate B. 1 2 Biophysics of Synaptic Plasticity Functional Modes of Synaptic Plasticity Types of learning: Unsupervised learning Supervised learning Plasticity is found in many brain regions: hippocampus, cortex, cerebellum, etc. Plot above shows field potential recordings from CA1 region in rat hippocampus. High-frequency stimulation leads to long-term potentiation (LTP). Low-frequency stimulation leads to long-term depression (LTD). Consistent with Hebb rule. Postsynaptic concentration of Ca 2+ ions play a role in LTP and LTD. 3 Reinforcement learning Types of synaptic plasticity: Hebbian synaptic plasticity Non-Hebbian synaptic plasticity: e.g., anti-hebbian (decrease strength when co-activated). 4

Stability and Competition Network Model with Firing Rate Neurons Increasing synaptic plasticity is a positive feedback process: Uncontrolled growth possible if unchecked. Dealing with unbounded growth: Impose a saturation constraint: 0 w w max : Possible problem of every weight turning w max. Synaptic competition: Some weaken while some strengthen. 5 Input vector u Weight vector w from Chapter 7 Output (postsynaptic activity) v τ r dv X = v + w u = v + w b v b, b=1 or, after reaching steady state (set the above to 0): v = w u 6 Basic Hebb Rule: Correlation-Based Simplest form has: = vu, where is the time constant that controls the rate of change in w (learning rate). Ensemble averaging ( ) over the inputs: = vu = Q w or τ b w = X b =1 Q = uu or Q bb = u b u b. Q bb w b, where Unbounded Growth in Basic Hebb Rule Length of weight vector: Dot product of and w gives: w 2 = w w = X b = vu = 2w w 2 b. = 2v 2, given and w u = v. Note v 0, so the above always increases (unless v = 0). 7 8

Discrete Updating Rule for Hebbian Learning Commonly used discrete update rule is: w w + ɛq w, where ɛ is analogous to 1 in the continuous version. Even simpler implementation is: i.e., no ensemble averaging. w w + ɛvu, Covariance Rule We want to allow a single rule to allow both increase and decrease in synaptic weight. = (v θ v)u, where θ v is a threshold. Synaptic weight will decrease if v < θ v and increase if v > θ v. An alternative is to put the threshold on the input side: = v(u θ u), If θ u = u, we get = C w, where 9 C = (u u )(u u ) = uu u 2 = (u u )u. 10 Depression under Covariance Rule Instability of the Covariance Rule Homosynaptic depression: depression when nonzero input and v < θ v = (v θ v)u. Heterosynaptic depression: depression when input is inactive and v > 0 = v(u θ u) The covariance rule is unstable despite the threshold: = 2v(v v ), where the time average of RHS is proportional to v 2 v 2, which is positive (it s the variance of v). Implicit point: No input or output activity is required for LTD to happen. 11 12

BCM Rule Bienenstock, Cooper, and Munro (1982). Synaptic plasticity requires both pre- and postsynaptic activity: = vu(v θ v). Unstable like Hebb rule if θ v is kept fixed. Condition for stability is: τ θ dθ v = v 2 θ v, where threshold adaptation rate τ θ is typically smaller than. Preventing Unbounded Growth: Normalization Directly work on the weights rather than altering the threshold. Assumption is that increase in one synaptic weight should be balanced by the decrease in other synaptic weights. Thus, global constraints are needed: Hold total sum of weights constant. Constrain the sum of squares of the weights. Sliding threshold implements synaptic competition: Increase in one synaptic weight will increase output v, thus it will increase threshold, making other synpases hard to adapt. 13 14 Synaptic Normalization: Subtractive Add a subtractive term in weight update: = vu v(n u)n, where is the length of u, and n = (1, 1, 1,...1), so P wb = n w. This is a rigid constraint, since the sum of weights n w does not change: dn w Biological basis is unclear. = vn u 1 n n «= 0. Synaptic Normalization: Multiplicative Oja s rule (Oja, 1982) with a positive constant α. = vu αv2 w, It is based more on a theoretical argument than biological. Stability can be analyzed as before: = 2v 2 (1 α w 2 ). The steady state value of w 2 becomes 1/α (set the RHS to 0 and solve for w 2 ). In other words, the length of the weight vector is held constant. 15 16

Timing-Based Rules Timing-Based Rule STDP applied to firing rate models: Z = 0 dτ(h(τ)v(t)u(t τ)+h( τ)v(t τ)u(t)), where H(τ) takes a shape similar to the plot B in the previous page, depending on the sign of τ (sign(h(τ)) = sign(τ)). Plasticity is time-dependent: Spike Timing Dep. Plast. (STDP) Presynaptic spike time t pre and postsynaptic spike time t post : If post fires first then pre, t post t pre < 0: pre did not cause post to fire. If pre fires first then post, t post t pre > 0: pre did cause post to fire. 17 STDP is more naturally applied to spiking neuron models. 18 Unsupervised Learning Variants of Hebbian learning can be understood in the context of unsupervised learning. Major area of research: cortical map formation. Orientation map, ocular dominance map, spatial frequencey map, etc. etc. Activity-dependent (learning) and/or activity-independent (genetically determined)? Hebbian Learning and Principal Eigenvector Diagonalize the correlation matrix Q: Q e µ = λ µ e µ where e µ is an eigenvector (mutually orthogonal) and λ µ is an eigenvalue (µ = 1, 2,..., ). For correlation and covariance matrices, all eigenvalues are real and nonnegative. We can express any -dimensional vector as a linear combination of the eigenvectors e µ. So, w(t) = where c µ are the coefficients. X c µ (t)e µ, 19 20

From we get Principal Eigenvector (cont d) w(t) = 2 w = [c 1, c 2,...c µ...c ] 6 4 c µ (t)e µ, c µ (t) = w(t) e µ [ e 1 ] [ e 2 ]... [ e µ ]... [ e ] 3 = ce 7 5 Plugging Principal Eigenvector (cont d) w(t) = cµ(t)eµ into τw = Q w we get dcµ(t) τw eµ = Q cµ(t)eµ. Multiply both sides with eν, and eµ eν = δµν : dcµ dcµ τw = cµ(t)(q eµ) eµ becomes τw = cµ(t)λµeµ eµ. Since Qe = λe dcµ τw = cµ(t)λµ, so, we get cµ(t) = c exp λ! µt, where w(0) eµ = cµ(0), so c = w(0) eµ. τw since we 1 = c, and E 1 = E T. 21 w(t) = exp λ! µt `w(0) e µ eµ. τw 22 Principal Eigenvector (cont d) Principal Eigenvector: Issues w(t) = X λµ t exp «(w(0) e µ ) e µ. Thus, for large t, the vector term with the highest λ µ factor (µ = 1 if eigenvalues have been sorted) will dominate, so Finally, we get w e 1. v e 1 u Principal eigenvector e 1 = (1, 1)/ 2 The proportionality relation v e 1 u conceals the large exponential factor, which can grow without bound. which is the projection of the input vector along the principal eigenvector of the correlation/covariance matrix. 23 Saturation constraint can help, but it can prevent the weight update to converge to the principal eigenvector (see figure above), depending on the initial condition. 24

Principal Eigenvector: Use of Oja s Rule Oja s rule (Oja, 1982) can be used to prevent unlimited growth: = vu αv2 w. The rule gives w = e 1 / α as t. Principal Eigenvector: Use of Subtractive Averaging Normalization over input samples gives: v(n u)n = vu, (w Q n)n = Q w. Growth of w is unaffected by the second term if e µ n = 0. If e µ n 0 weight will grow without bound. 25 If principal eigenvector of Q is proportional to n, Q e 1 (e 1 Q n)n/n = 0, so principal eigenvector is unaffected by the learning rule. Also, e µ n = 0 for µ 2, so w(t) = (w(0) e 1 )e 1 + µ=2 26 λµ t exp «(w(0) e µ )e µ. Hebbian Learning for PCA A: correlation rule, zero mean B: correlation rule, non-zero mean C: covariance rule, non-zero mean. 27