J. Marro Institute Carlos I, University of Granada. Net-Works 08, Pamplona, 9-11 June 2008

Size: px
Start display at page:

Download "J. Marro Institute Carlos I, University of Granada. Net-Works 08, Pamplona, 9-11 June 2008"

Transcription

1 J. Marro Institute Carlos I, University of Granada Net-Works 08, Pamplona, 9-11 June 2008

2 Networks? mathematical objects Königsberg bridges + Euler graphs (s. XVIII),... models of natural systems social and economic nets, www and Internet, food webs, metabolic networks, communication and transport nets,... models of interacting particles capture essence of cooperation, importance of topology of interactions, synchronization,... antecedents: mathematics, neuro- and computer science,..., physics neuro- and computer science Santiago Ramón y Cajal (Nobel 1906), Warren McCulloch and Walter Pitts (1943), Donald Hebb (1949), Shun-Ichi Amary (1972), John Hopfield (1982) ( 5,000 cites), Daniel Amit (1989),... good (biological, hardware,...) data for (statistical) physicists (Positron Emission Tomography, Functional Magnetic Resonance Imaging, Electrophysiological Recording, Magnetoencefalography,... Computers Clusters, Farms, Parallelization,...) 2

3 Amari Hopfield Attractor Network inhomogeneous connections, w ji j each node receives a signal: i w ji h i = j w ji σ i state at time t is characterized by node activities: σ {σ i } (e.g. σ i =±1) connection weights: w {w ij є R} (i,j = 1,,N ) 3

4 Amari Hopfield Attractor Network synaptic connection weights are choosen (e.g.) by the Hebb (synaptic learning) prescription: w ji = N 1 ν ξ iν ξ j ν ( N neurons; P sets of activity patterns ξ ν = {ξ iν = ±1}, ν = 1,,P ) node activities are stochastic variables, e.g., assuming neuron at i fires (changes state) if σ i ( t + 1 ) = sign { h i (t) θ i } in practice, make change with probability depending on an inverse temperature, β, namely, p = { 1 + exp [ 2β ( h i θ i ) ] } 1 4

5 Dynamic Attractors, Associative Memory

6 This picture lacks two essential facts: Activity of systems in nature does not stay in one memory irregular wandering among stored patterns, e.g., insects olfactory processes associated with heteroclinic paths in patterns space One needs irregular time behavior & instabilities! Connections constant but fast time changes (fluctuations on the 10-3 s scale, short-term plasticity, synaptic fatigue including depression and/or facilitation,...) these changes during operation may induce dynamic instabilities and even chaos 6

7 Many different context in which connections are not only homogeneous but with fast time changes! Weighted, and time dependent communication lines transport nets: connections differ in capacity, number of flights and passengers condensed matter: spin glasses and other systems with impurities, microscopic disorder and local fields, reaction diffusion systems,... diffusion of ions, chemical reactions, local rearrangements,... constantly vary the effective interactions between the units fluxes of different and varying intensities in food webs, ecological and metabolic nets agents interchange different amounts of information or money in the internet, www, economic and other social nets complex and varying patterns of synapses in the central nervous system and the brain

8 I shall assume that the observed time variation of connections is because the weights depend on the system activity due to two main general features of nodes: nodes are excitable elements, i.e., respond to perturbation if above threshold, and then show time lag (to next response) in which they behave as refractory to further excitation effective excitability would cause a non-equilibrium steady condition in a set of interacting elements which would, e.g., impede the damping by friction of signals in forest fires, the nervous system and autocatalytic reactions in surfaces nodes may be quiet certain nodes are more active, and some may even not be engaged at a given time in a given cooperative task. there is no need for a network to maintain all the nodes informed of the activity of all the others at all times and, in fact, it has often been reported non-uniform distribution of activity, e.g., associated to working memories 8

9 Implementing these features in a model network: SILENT NODES At each time, only the state of a fraction ρ of (e.g.) randomly chosen nodes will be updated: ρ 1: all the nodes are updated at once (parallel or Little updating, as in cellular automata); ρ 0: just one node at a time (sequential or Glauber updating); Note: this mimics an observable property of excitable elements and, in any case, why not study all possibilities between these two limits? Question: what are the consequences of having nodes that do not fire at a given time (even if they receive action above threshold), or of eventually turning off a fraction of nodes or communication lines

10 Implementing these features in a model network: EFFECTIVE EXCITABILITY OF NODES Replace the constant (e.g., Hebb s) weights for noisy ones: ω ij ω ij x j where x j is stochastic variable varying on time scale much smaller than one for nodes changes, so that one considers stationary distribution, either P1 x q x, 1 q x 1 or else j j j j P2 x x 1 x 1 j j j (local versus global order influence on the weights)

11 Implementing other interesting features in a model network: WIRING TOPOLOGY Let the topology matrix: ij ij 1 if link does exist ij, where ij, ki j ij 0 if link does not exist Therefore, each node feels an effective field: h, where depends on P x eff eff eff i j ij ij j ij j which allows consider different connectivity distributions, e.g.: pk k k1 k k2, k2 k p k k, k k, k for finite N 0 m

12 Instabilities stored patterns are attractors of dynamics but noise : destabilizes the attractors and, consequently: susceptibility to external stimulus increases an efficient search in attractors space is induced search may be chaotic for ρ ρ C, where slight variations of ρ may switch search from regular to irregular J. Marro, J. Torres, J. Cortes, J. Stat. Mech. P02017 (2008) J. Torres et al., TBP, arxiv: v1 [cond-mat.dis-nn] 12

13 Instabilities Φ = 1 Bifurcation diagram with evidence of chaos and periodic windows in MC simulations (P 2, M = 20 random patterns, N = 3600 nodes, Φ = 0.5, β = 20) stationary order = sum of squares of overlaps (α=m/n) J. Marro, J. Torres, J. Cortes, J. Stat. Mech. P02017 (2008) 13

14 Instabilities Lyapunov surface λ as a function of Φ (β=25) and β (Φ=0.01) complex landscape: λ > 0 (chaos) if region is black. (there is chaos also for ρ = 1) (P 2, M = 1 + anti-pattern) For Φ = 0.005, β = 50, M = 1, analytically. J. Marro, J. Torres, J. Cortes, J. Stat. Mech. P02017 (2008) 14

15 Instabilities overlap versus time (N = 1600, M = 3 uncorrelated patterns, Φ = 0.4, β = 20) convergence and stability of attractor in Fully fact, irregular anti-pattern (positive (zero Lyapunov overlap with others) exponent) behavior ρ = 0.08 < for ρ C ρ = Regular oscillation between 0.50 attractor > ρ C and its negative for ρ = 0.65 > ρ C Onset of chaos as ρ is incresed somewhat; ρ = 0.92 Rapid and ordered periodic oscillations between pattern and its antipattern (all nodes synchronized) J. Marro, J. Torres, J. Cortes, J. Stat. Mech. P02017 (2008)] 15

16 Phase (attractors) space Irregular switching among attractors states of attention, possible role of chaos # attractors visited (and correlation distance between them) increases with ρ mean firing rate versus t n and phase space trajectories (Φ = 0.05, N = 1600, β = 167, ρ C = 0.38, and three patterns, ξ μ μ =1,2,3.) J. Marro, J. Torres, J. Cortes, J. Stat. Mech. P02017 (2008)] 16

17 IEEE Computational Intelligence Magazine, August 2007, pp Walter J. Freeman (2003, 2005) interpreted EEG signals from cortical neural activity in animals and humans over the gamma frequency band (20-80 Hz) which shows sustained quasi-stable patterns of activities for several 100 ms, and extends over spatial scales comparable to the size of the hemisphere using dynamic systems theory. Freeman explicitly assumes: the brain s basal state is a high-dimensional chaotic attractor. Under the influence of external stimuli, the dynamics is constrained to a lower-dimensional wing. The system stays in this wing intermittently and produces an amplitude modulation (AM) activity pattern. Ultimately, the system jumps to another wing as it explores the complex attractor landscape. This is in support of the chaotic itinerancy theory (Tsuda 2001) that successfully interprets EEG measurements in terms of trajectories of a dynamical system, which intermittently visits the attractor ruins as it traverses across the landscape. Behavioral and Brain Sciences 24, (2001) 17

18 Phase (attractors) space: experiment vs model Stimuli of same intensity and duration (green and red). Stimulus destabilization in the absence of chaos. (Top involves falseneighbor method with d = 5 and τ = 20.) Response to odor stimuli of certain neurons in the locust antennal lobe; Mazor & Laurent, Neuron 48 (2005) Trajectories and t variation of mean firing rate during two MC simulations for N = 1600, β = 4, Φ = 0.45, ρ = 3/64 < ρ C, M =6. J. Torres et al., TBP, arxiv: v1 [cond-mat.dis-nn] 18

19 Criticality overlap (same cases as in main graph: ρ increases from top to bottom) power spectra is rather flat (but for large fequency peaks) when ρ = 0.35, 0.55 regular behavior while becomes power lawed during the irregular, critical cases ρ = 0.4 and 0.45 power spectra of the local field h i for N = 1600, P = 5 and Φ = 0.5 S. de Franciscis et al., report. 19

20 Criticality Typical Local field h i time series just before (ρ=0.38) and at criticality (ρ=0.40) for the same system S. de Franciscis et al., report. 20

21 Criticality Probability that h i = const. (except for small fluctuations) during time interval Δτ (data for all i and several time series) Evidence for criticality when 0.4 ρ 0.5 T = 0.01, Φ = 0.5, N = 1600, P = 10. P 2 with ζ ~ μ (m μ ) 2 S. de Franciscis et al., report. 21

22 Criticality Same in the critical case ρ = 0.45 as one varies P and N but not the load α = P / N The observed critical behavior is robust S. de Franciscis et al., report. 22

23 Criticality Typical Local field h i time series just before (ρ=0.35) and at and inside criticality (ρ=0.40, 0.45, 0.50 and 0.55) for the system which is analysed next. S. de Franciscis et al., report. 23

24 Criticality The correlation function, defined as the average (over i and t but only one system in this case) of [h i (t+δτ) h i (t)] 2 showing criticality at 0.4 and, to some extent, also at 0.45 and 0.5 P 2, ζ ~ μ m μ S. de Franciscis et al., report. 24

25 Topology M = 1, large N, any connectivity p(k) and mean-field assumptions (substitute ε ij by its mean value), phase transitions: from high-t disorder to associative memory at T c = k 2 / kn at T=0, for fatigue parameter Φ Φ 0 = 1 k +1 / k +1 [ > 0 monitors wiring in j (σ,φ) ], phase with periodic hopping among attractors at finite T, the hopping becomes irregular below Φ c (T) ~ Φ 0, and no irregular behavior is observed above T ~ 0.35 T c. Confirmed in MC simulations for M > 1 as hopping among different patterns Φ c () the more heterogeneous p(k) is, the lower the amount of fatigue, and the higher T c, needed to destabilize dynamics: S. Johnson, J. Marro, J. Torres, Europhys. Lett., to appear 25

26 Topology Edge of chaos transition curve between associative memory and irregular hopping showing critical fatigue (theories and MC data, T = 2/N, k = 20, = 2) for bimodal (top) and power law p(h) s with varying Δ and. 2 : minimum fatigue (Φ ~ 1, slight depression; Φ = 1 means no change) needed to destabilize. (Functional not structural connectivity in brain is power lawed with 2 ; Eguiluz et al. PRL 2005; same with many other systems in nature) S. Johnson, J. Marro, J. Torres, EPL, arxiv: v1 [cond-mat.dis-nn] 26

27 Topology System is showed one of the M stored patterns at equal time intervals: h i > h i + δ ξ ν i Ideally (standard model) it d remain there unless strong next stimulus estimate performance from time average of overlap between current state and input pattern: better the closer to the edge of chaos (and Φ ~ 1, 2), where both retrieval and destabilization are more efficient S. Johnson, J. Marro, J. Torres, TBP, arxiv: v1 [cond-mat.dis-nn] 27

28 Generalization & applications instead of making synapses noisy by ω ij ω ij x j assume state at t (σ,μ), σ = {σ i } and μ = {ω ij αξ iμ ξ jμ }, evolves in discrete time: P t+1 (σ,μ) = T μ [σ σ] T σ [μ μ] P t (σ,μ ) i.e., according to two stochastic processes driven by independent temperatures and by the energy cost ΔH of the corresponding change, as in familiar Glauber dynamics.. J. Marro et al., TBP & AIP Proceedings Series 779, (2005) 28

29 Generalization & applications Phase Diagram (approx.) (T 0 neurons; T 1 synapses ) Phases: Ferromagnetic (better assoc. memory than standard case, e.g., no mixture states); Paramagnetic (P); Oscillatory (dynamic memory; relate ± correlated patterns) J. Marro et al., TBP & AIP Proceedings Series 779, (2005) 29

30

arxiv: v1 [cond-mat.dis-nn] 9 May 2008

arxiv: v1 [cond-mat.dis-nn] 9 May 2008 Functional Optimization in Complex Excitable Networks Samuel Johnson, J. Marro, and Joaquín J. Torres Departamento de Electromagnetismo y Física de la Materia, and Institute Carlos I for Theoretical and

More information

Functional optimization in complex excitable networks

Functional optimization in complex excitable networks August 28 EPL, 83 (28) 466 doi:.29/295-575/83/466 www.epljournal.org S. Johnson, J. Marro and J. J. Torres Departamento de Electromagnetismo y Física de la Materia, and Institute Carlos I for Theoretical

More information

Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5

Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5 Hopfield Neural Network and Associative Memory Typical Myelinated Vertebrate Motoneuron (Wikipedia) PHY 411-506 Computational Physics 2 1 Wednesday, March 5 1906 Nobel Prize in Physiology or Medicine.

More information

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 28 Apr 2006

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 28 Apr 2006 arxiv:cond-mat/0604662v1 [cond-mat.dis-nn] 28 Apr 2006 Sensitivity, Itinerancy and Chaos in Partly Synchronized Weighted Networks J. Marro +, J.J. Torres +, J.M. Cortes +, and B. Wemmenhove + Institute

More information

Artificial Intelligence Hopfield Networks

Artificial Intelligence Hopfield Networks Artificial Intelligence Hopfield Networks Andrea Torsello Network Topologies Single Layer Recurrent Network Bidirectional Symmetric Connection Binary / Continuous Units Associative Memory Optimization

More information

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network. III. Recurrent Neural Networks A. The Hopfield Network 2/9/15 1 2/9/15 2 Typical Artificial Neuron Typical Artificial Neuron connection weights linear combination activation function inputs output net

More information

Stochastic Model for Adaptation Using Basin Hopping Dynamics

Stochastic Model for Adaptation Using Basin Hopping Dynamics Stochastic Model for Adaptation Using Basin Hopping Dynamics Peter Davis NTT Communication Science Laboratories 2-4 Hikaridai, Keihanna Science City, Kyoto, Japan 619-0237 davis@cslab.kecl.ntt.co.jp Abstract

More information

7 Rate-Based Recurrent Networks of Threshold Neurons: Basis for Associative Memory

7 Rate-Based Recurrent Networks of Threshold Neurons: Basis for Associative Memory Physics 178/278 - David Kleinfeld - Fall 2005; Revised for Winter 2017 7 Rate-Based Recurrent etworks of Threshold eurons: Basis for Associative Memory 7.1 A recurrent network with threshold elements The

More information

Influence of Criticality on 1/f α Spectral Characteristics of Cortical Neuron Populations

Influence of Criticality on 1/f α Spectral Characteristics of Cortical Neuron Populations Influence of Criticality on 1/f α Spectral Characteristics of Cortical Neuron Populations Robert Kozma rkozma@memphis.edu Computational Neurodynamics Laboratory, Department of Computer Science 373 Dunn

More information

7 Recurrent Networks of Threshold (Binary) Neurons: Basis for Associative Memory

7 Recurrent Networks of Threshold (Binary) Neurons: Basis for Associative Memory Physics 178/278 - David Kleinfeld - Winter 2019 7 Recurrent etworks of Threshold (Binary) eurons: Basis for Associative Memory 7.1 The network The basic challenge in associative networks, also referred

More information

Using a Hopfield Network: A Nuts and Bolts Approach

Using a Hopfield Network: A Nuts and Bolts Approach Using a Hopfield Network: A Nuts and Bolts Approach November 4, 2013 Gershon Wolfe, Ph.D. Hopfield Model as Applied to Classification Hopfield network Training the network Updating nodes Sequencing of

More information

Effects of Interactive Function Forms in a Self-Organized Critical Model Based on Neural Networks

Effects of Interactive Function Forms in a Self-Organized Critical Model Based on Neural Networks Commun. Theor. Phys. (Beijing, China) 40 (2003) pp. 607 613 c International Academic Publishers Vol. 40, No. 5, November 15, 2003 Effects of Interactive Function Forms in a Self-Organized Critical Model

More information

Neural Nets and Symbolic Reasoning Hopfield Networks

Neural Nets and Symbolic Reasoning Hopfield Networks Neural Nets and Symbolic Reasoning Hopfield Networks Outline The idea of pattern completion The fast dynamics of Hopfield networks Learning with Hopfield networks Emerging properties of Hopfield networks

More information

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network. Part 3A: Hopfield Network III. Recurrent Neural Networks A. The Hopfield Network 1 2 Typical Artificial Neuron Typical Artificial Neuron connection weights linear combination activation function inputs

More information

Chaotic Neurodynamics for Autonomous Agents

Chaotic Neurodynamics for Autonomous Agents CHAOTIC NEURODYNAMICS FOR AUTONOMOUS AGENTS 1 Chaotic Neurodynamics for Autonomous Agents Derek Harter Member, Robert Kozma Senior Member Division of Computer Science, University of Memphis, TN, USA Abstract

More information

Hopfield Networks. (Excerpt from a Basic Course at IK 2008) Herbert Jaeger. Jacobs University Bremen

Hopfield Networks. (Excerpt from a Basic Course at IK 2008) Herbert Jaeger. Jacobs University Bremen Hopfield Networks (Excerpt from a Basic Course at IK 2008) Herbert Jaeger Jacobs University Bremen Building a model of associative memory should be simple enough... Our brain is a neural network Individual

More information

An Introductory Course in Computational Neuroscience

An Introductory Course in Computational Neuroscience An Introductory Course in Computational Neuroscience Contents Series Foreword Acknowledgments Preface 1 Preliminary Material 1.1. Introduction 1.1.1 The Cell, the Circuit, and the Brain 1.1.2 Physics of

More information

Analysis of Neural Networks with Chaotic Dynamics

Analysis of Neural Networks with Chaotic Dynamics Chaos, Solitonr & Fructals Vol. 3, No. 2, pp. 133-139, 1993 Printed in Great Britain @60-0779/93$6.00 + 40 0 1993 Pergamon Press Ltd Analysis of Neural Networks with Chaotic Dynamics FRANCOIS CHAPEAU-BLONDEAU

More information

Effects of refractory periods in the dynamics of a diluted neural network

Effects of refractory periods in the dynamics of a diluted neural network Effects of refractory periods in the dynamics of a diluted neural network F. A. Tamarit, 1, * D. A. Stariolo, 2, * S. A. Cannas, 2, *, and P. Serra 2, 1 Facultad de Matemática, Astronomía yfísica, Universidad

More information

Consider the following spike trains from two different neurons N1 and N2:

Consider the following spike trains from two different neurons N1 and N2: About synchrony and oscillations So far, our discussions have assumed that we are either observing a single neuron at a, or that neurons fire independent of each other. This assumption may be correct in

More information

The Mixed States of Associative Memories Realize Unimodal Distribution of Dominance Durations in Multistable Perception

The Mixed States of Associative Memories Realize Unimodal Distribution of Dominance Durations in Multistable Perception The Mixed States of Associative Memories Realize Unimodal Distribution of Dominance Durations in Multistable Perception Takashi Kanamaru Department of Mechanical Science and ngineering, School of Advanced

More information

Introduction to Neural Networks

Introduction to Neural Networks Introduction to Neural Networks What are (Artificial) Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning

More information

From Last Time. Gravitational forces are apparent at a wide range of scales. Obeys

From Last Time. Gravitational forces are apparent at a wide range of scales. Obeys From Last Time Gravitational forces are apparent at a wide range of scales. Obeys F gravity (Mass of object 1) (Mass of object 2) square of distance between them F = 6.7 10-11 m 1 m 2 d 2 Gravitational

More information

( ) T. Reading. Lecture 22. Definition of Covariance. Imprinting Multiple Patterns. Characteristics of Hopfield Memory

( ) T. Reading. Lecture 22. Definition of Covariance. Imprinting Multiple Patterns. Characteristics of Hopfield Memory Part 3: Autonomous Agents /8/07 Reading Lecture 22 Flake, ch. 20 ( Genetics and Evolution ) /8/07 /8/07 2 Imprinting Multiple Patterns Let x, x 2,, x p be patterns to be imprinted Define the sum-of-outer-products

More information

Neural Nets in PR. Pattern Recognition XII. Michal Haindl. Outline. Neural Nets in PR 2

Neural Nets in PR. Pattern Recognition XII. Michal Haindl. Outline. Neural Nets in PR 2 Neural Nets in PR NM P F Outline Motivation: Pattern Recognition XII human brain study complex cognitive tasks Michal Haindl Faculty of Information Technology, KTI Czech Technical University in Prague

More information

Hopfield and Potts-Hopfield Networks

Hopfield and Potts-Hopfield Networks Hopfield and Potts-Hopfield Networks Andy Somogyi Dec 10, 2010 1 Introduction Many early models of neural networks lacked mathematical rigor or analysis. They were, by and large, ad hoc models. John Hopfield

More information

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required. In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required. In humans, association is known to be a prominent feature of memory.

More information

THE LOCUST OLFACTORY SYSTEM AS A CASE STUDY FOR MODELING DYNAMICS OF NEUROBIOLOGICAL NETWORKS: FROM DISCRETE TIME NEURONS TO CONTINUOUS TIME NEURONS

THE LOCUST OLFACTORY SYSTEM AS A CASE STUDY FOR MODELING DYNAMICS OF NEUROBIOLOGICAL NETWORKS: FROM DISCRETE TIME NEURONS TO CONTINUOUS TIME NEURONS 1 THE LOCUST OLFACTORY SYSTEM AS A CASE STUDY FOR MODELING DYNAMICS OF NEUROBIOLOGICAL NETWORKS: FROM DISCRETE TIME NEURONS TO CONTINUOUS TIME NEURONS B. QUENET 1 AND G. HORCHOLLE-BOSSAVIT 2 1 Equipe de

More information

How can ideas from quantum computing improve or speed up neuromorphic models of computation?

How can ideas from quantum computing improve or speed up neuromorphic models of computation? Neuromorphic Computation: Architectures, Models, Applications Associative Memory Models with Adiabatic Quantum Optimization Kathleen Hamilton, Alexander McCaskey, Jonathan Schrock, Neena Imam and Travis

More information

Group Formation: Fragmentation Transitions in Network Coevolution Dynamics

Group Formation: Fragmentation Transitions in Network Coevolution Dynamics - Mallorca - Spain Workshop on Challenges and Visions in the Social Sciences, ETH August 08 Group Formation: Fragmentation Transitions in Network Coevolution Dynamics MAXI SAN MIGUEL CO-EVOLUTION Dynamics

More information

Lecture 4: Feed Forward Neural Networks

Lecture 4: Feed Forward Neural Networks Lecture 4: Feed Forward Neural Networks Dr. Roman V Belavkin Middlesex University BIS4435 Biological neurons and the brain A Model of A Single Neuron Neurons as data-driven models Neural Networks Training

More information

The homogeneous Poisson process

The homogeneous Poisson process The homogeneous Poisson process during very short time interval Δt there is a fixed probability of an event (spike) occurring independent of what happened previously if r is the rate of the Poisson process,

More information

Estimation of Propagating Phase Transients in EEG Data - Application of Dynamic Logic Neural Modeling Approach

Estimation of Propagating Phase Transients in EEG Data - Application of Dynamic Logic Neural Modeling Approach Proceedings of International Joint Conference on Neural Networks, Orlando, Florida, USA, August 12-17, 2007 Estimation of Propagating Phase Transients in EEG Data - Application of Dynamic Logic Neural

More information

3.3 Discrete Hopfield Net An iterative autoassociative net similar to the nets described in the previous sections has been developed by Hopfield

3.3 Discrete Hopfield Net An iterative autoassociative net similar to the nets described in the previous sections has been developed by Hopfield 3.3 Discrete Hopfield Net An iterative autoassociative net similar to the nets described in the previous sections has been developed by Hopfield (1982, 1984). - The net is a fully interconnected neural

More information

Hopfield Neural Network

Hopfield Neural Network Lecture 4 Hopfield Neural Network Hopfield Neural Network A Hopfield net is a form of recurrent artificial neural network invented by John Hopfield. Hopfield nets serve as content-addressable memory systems

More information

Artificial Neural Network and Fuzzy Logic

Artificial Neural Network and Fuzzy Logic Artificial Neural Network and Fuzzy Logic 1 Syllabus 2 Syllabus 3 Books 1. Artificial Neural Networks by B. Yagnanarayan, PHI - (Cover Topologies part of unit 1 and All part of Unit 2) 2. Neural Networks

More information

Supporting Online Material for

Supporting Online Material for www.sciencemag.org/cgi/content/full/319/5869/1543/dc1 Supporting Online Material for Synaptic Theory of Working Memory Gianluigi Mongillo, Omri Barak, Misha Tsodyks* *To whom correspondence should be addressed.

More information

Emergence of resonances in neural systems: the interplay between adaptive threshold and short-term synaptic plasticity

Emergence of resonances in neural systems: the interplay between adaptive threshold and short-term synaptic plasticity Emergence of resonances in neural systems: the interplay between adaptive threshold and short-term synaptic plasticity Jorge F. Mejias 1,2 and Joaquín J. Torres 2 1 Department of Physics and Center for

More information

Memory capacity of networks with stochastic binary synapses

Memory capacity of networks with stochastic binary synapses Memory capacity of networks with stochastic binary synapses Alexis M. Dubreuil 1,, Yali Amit 3 and Nicolas Brunel 1, 1 UMR 8118, CNRS, Université Paris Descartes, Paris, France Departments of Statistics

More information

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science Neural Networks Prof. Dr. Rudolf Kruse Computational Intelligence Group Faculty for Computer Science kruse@iws.cs.uni-magdeburg.de Rudolf Kruse Neural Networks 1 Hopfield Networks Rudolf Kruse Neural Networks

More information

Reducing neuronal networks to discrete dynamics

Reducing neuronal networks to discrete dynamics Physica D 237 (2008) 324 338 www.elsevier.com/locate/physd Reducing neuronal networks to discrete dynamics David Terman a,b,, Sungwoo Ahn a, Xueying Wang a, Winfried Just c a Department of Mathematics,

More information

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i )

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i ) Symmetric Networks Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). How can we model an associative memory? Let M = {v 1,..., v m } be a

More information

CHALMERS, GÖTEBORGS UNIVERSITET. EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

CHALMERS, GÖTEBORGS UNIVERSITET. EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD CHALMERS, GÖTEBORGS UNIVERSITET EXAM for ARTIFICIAL NEURAL NETWORKS COURSE CODES: FFR 135, FIM 72 GU, PhD Time: Place: Teachers: Allowed material: Not allowed: October 23, 217, at 8 3 12 3 Lindholmen-salar

More information

Effects of Interactive Function Forms and Refractoryperiod in a Self-Organized Critical Model Based on Neural Networks

Effects of Interactive Function Forms and Refractoryperiod in a Self-Organized Critical Model Based on Neural Networks Commun. Theor. Phys. (Beijing, China) 42 (2004) pp. 121 125 c International Academic Publishers Vol. 42, No. 1, July 15, 2004 Effects of Interactive Function Forms and Refractoryperiod in a Self-Organized

More information

Neural networks: Unsupervised learning

Neural networks: Unsupervised learning Neural networks: Unsupervised learning 1 Previously The supervised learning paradigm: given example inputs x and target outputs t learning the mapping between them the trained network is supposed to give

More information

Learning and Memory in Neural Networks

Learning and Memory in Neural Networks Learning and Memory in Neural Networks Guy Billings, Neuroinformatics Doctoral Training Centre, The School of Informatics, The University of Edinburgh, UK. Neural networks consist of computational units

More information

Systems Biology: A Personal View IX. Landscapes. Sitabhra Sinha IMSc Chennai

Systems Biology: A Personal View IX. Landscapes. Sitabhra Sinha IMSc Chennai Systems Biology: A Personal View IX. Landscapes Sitabhra Sinha IMSc Chennai Fitness Landscapes Sewall Wright pioneered the description of how genotype or phenotypic fitness are related in terms of a fitness

More information

arxiv:physics/ v1 [physics.bio-ph] 19 Feb 1999

arxiv:physics/ v1 [physics.bio-ph] 19 Feb 1999 Odor recognition and segmentation by coupled olfactory bulb and cortical networks arxiv:physics/9902052v1 [physics.bioph] 19 Feb 1999 Abstract Zhaoping Li a,1 John Hertz b a CBCL, MIT, Cambridge MA 02139

More information

Self-organized Criticality and Synchronization in a Pulse-coupled Integrate-and-Fire Neuron Model Based on Small World Networks

Self-organized Criticality and Synchronization in a Pulse-coupled Integrate-and-Fire Neuron Model Based on Small World Networks Commun. Theor. Phys. (Beijing, China) 43 (2005) pp. 466 470 c International Academic Publishers Vol. 43, No. 3, March 15, 2005 Self-organized Criticality and Synchronization in a Pulse-coupled Integrate-and-Fire

More information

By Nadha CHAOS THEORY

By Nadha CHAOS THEORY By Nadha CHAOS THEORY What is Chaos Theory? It is a field of study within applied mathematics It studies the behavior of dynamical systems that are highly sensitive to initial conditions It deals with

More information

Lecture 4: Importance of Noise and Fluctuations

Lecture 4: Importance of Noise and Fluctuations Lecture 4: Importance of Noise and Fluctuations Jordi Soriano Fradera Dept. Física de la Matèria Condensada, Universitat de Barcelona UB Institute of Complex Systems September 2016 1. Noise in biological

More information

arxiv:quant-ph/ v1 17 Oct 1995

arxiv:quant-ph/ v1 17 Oct 1995 PHYSICS AND CONSCIOUSNESS Patricio Pérez arxiv:quant-ph/9510017v1 17 Oct 1995 Departamento de Física, Universidad de Santiago de Chile Casilla 307, Correo 2, Santiago, Chile ABSTRACT Some contributions

More information

Hopfield Network for Associative Memory

Hopfield Network for Associative Memory CSE 5526: Introduction to Neural Networks Hopfield Network for Associative Memory 1 The next few units cover unsupervised models Goal: learn the distribution of a set of observations Some observations

More information

LFFs in Vertical Cavity Lasers

LFFs in Vertical Cavity Lasers LFFs in Vertical Cavity Lasers A. Torcini, G. Giacomelli, F. Marin torcini@inoa.it, S. Barland ISC - CNR - Firenze (I) Physics Dept - Firenze (I) INLN - Nice (F) FNOES - Nice, 14.9.5 p.1/21 Plan of the

More information

The effect of emigration and immigration on the dynamics of a discrete-generation population

The effect of emigration and immigration on the dynamics of a discrete-generation population J. Biosci., Vol. 20. Number 3, June 1995, pp 397 407. Printed in India. The effect of emigration and immigration on the dynamics of a discrete-generation population G D RUXTON Biomathematics and Statistics

More information

Synchronization of a General Delayed Complex Dynamical Network via Adaptive Feedback

Synchronization of a General Delayed Complex Dynamical Network via Adaptive Feedback Synchronization of a General Delayed Complex Dynamical Network via Adaptive Feedback Qunjiao Zhang and Junan Lu College of Mathematics and Statistics State Key Laboratory of Software Engineering Wuhan

More information

Abstract: Complex responses observed in an experimental, nonlinear, moored structural

Abstract: Complex responses observed in an experimental, nonlinear, moored structural AN INDEPENDENT-FLOW-FIELD MODEL FOR A SDOF NONLINEAR STRUCTURAL SYSTEM, PART II: ANALYSIS OF COMPLEX RESPONSES Huan Lin e-mail: linh@engr.orst.edu Solomon C.S. Yim e-mail: solomon.yim@oregonstate.edu Ocean

More information

(a) (b) (c) Time Time. Time

(a) (b) (c) Time Time. Time Baltzer Journals Stochastic Neurodynamics and the System Size Expansion Toru Ohira and Jack D. Cowan 2 Sony Computer Science Laboratory 3-4-3 Higashi-gotanda, Shinagawa, Tokyo 4, Japan E-mail: ohiracsl.sony.co.jp

More information

Neural Networks. Hopfield Nets and Auto Associators Fall 2017

Neural Networks. Hopfield Nets and Auto Associators Fall 2017 Neural Networks Hopfield Nets and Auto Associators Fall 2017 1 Story so far Neural networks for computation All feedforward structures But what about.. 2 Loopy network Θ z = ቊ +1 if z > 0 1 if z 0 y i

More information

Machine Learning. Neural Networks

Machine Learning. Neural Networks Machine Learning Neural Networks Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 Biological Analogy Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 THE

More information

Neurophysiology of a VLSI spiking neural network: LANN21

Neurophysiology of a VLSI spiking neural network: LANN21 Neurophysiology of a VLSI spiking neural network: LANN21 Stefano Fusi INFN, Sezione Roma I Università di Roma La Sapienza Pza Aldo Moro 2, I-185, Roma fusi@jupiter.roma1.infn.it Paolo Del Giudice Physics

More information

Neural Network Essentials 1

Neural Network Essentials 1 Neural Network Essentials Draft: Associative Memory Chapter Anthony S. Maida University of Louisiana at Lafayette, USA June 6, 205 Copyright c 2000 2003 by Anthony S. Maida Contents 0. Associative memory.............................

More information

How to do backpropagation in a brain

How to do backpropagation in a brain How to do backpropagation in a brain Geoffrey Hinton Canadian Institute for Advanced Research & University of Toronto & Google Inc. Prelude I will start with three slides explaining a popular type of deep

More information

Probabilistic Models in Theoretical Neuroscience

Probabilistic Models in Theoretical Neuroscience Probabilistic Models in Theoretical Neuroscience visible unit Boltzmann machine semi-restricted Boltzmann machine restricted Boltzmann machine hidden unit Neural models of probabilistic sampling: introduction

More information

Oscillatory Motion. Simple pendulum: linear Hooke s Law restoring force for small angular deviations. Oscillatory solution

Oscillatory Motion. Simple pendulum: linear Hooke s Law restoring force for small angular deviations. Oscillatory solution Oscillatory Motion Simple pendulum: linear Hooke s Law restoring force for small angular deviations d 2 θ dt 2 = g l θ θ l Oscillatory solution θ(t) =θ 0 sin(ωt + φ) F with characteristic angular frequency

More information

Renormalization Group: non perturbative aspects and applications in statistical and solid state physics.

Renormalization Group: non perturbative aspects and applications in statistical and solid state physics. Renormalization Group: non perturbative aspects and applications in statistical and solid state physics. Bertrand Delamotte Saclay, march 3, 2009 Introduction Field theory: - infinitely many degrees of

More information

Oscillatory Motion. Simple pendulum: linear Hooke s Law restoring force for small angular deviations. small angle approximation. Oscillatory solution

Oscillatory Motion. Simple pendulum: linear Hooke s Law restoring force for small angular deviations. small angle approximation. Oscillatory solution Oscillatory Motion Simple pendulum: linear Hooke s Law restoring force for small angular deviations d 2 θ dt 2 = g l θ small angle approximation θ l Oscillatory solution θ(t) =θ 0 sin(ωt + φ) F with characteristic

More information

Bifurcations and multistability in turbulence

Bifurcations and multistability in turbulence Bifurcations and multistability in turbulence VKS team A. Chiffaudel L. Marié F. Ravelet F.Daviaud CEA/Saclay, France O. Dauchot A. Prigent Conceptual Aspects of turbulence, Vienna, February 28 1 Bifurcations

More information

Hopfield networks. Lluís A. Belanche Soft Computing Research Group

Hopfield networks. Lluís A. Belanche Soft Computing Research Group Lluís A. Belanche belanche@lsi.upc.edu Soft Computing Research Group Dept. de Llenguatges i Sistemes Informàtics (Software department) Universitat Politècnica de Catalunya 2010-2011 Introduction Content-addressable

More information

Synchronization of Limit Cycle Oscillators by Telegraph Noise. arxiv: v1 [cond-mat.stat-mech] 5 Aug 2014

Synchronization of Limit Cycle Oscillators by Telegraph Noise. arxiv: v1 [cond-mat.stat-mech] 5 Aug 2014 Synchronization of Limit Cycle Oscillators by Telegraph Noise Denis S. Goldobin arxiv:148.135v1 [cond-mat.stat-mech] 5 Aug 214 Department of Physics, University of Potsdam, Postfach 61553, D-14415 Potsdam,

More information

Synchronous state transition graph

Synchronous state transition graph Heike Siebert, FU Berlin, Molecular Networks WS10/11 2-1 Synchronous state transition graph (0, 2) (1, 2) vertex set X (state space) edges (x,f(x)) every state has only one successor attractors are fixed

More information

Neural Modeling and Computational Neuroscience. Claudio Gallicchio

Neural Modeling and Computational Neuroscience. Claudio Gallicchio Neural Modeling and Computational Neuroscience Claudio Gallicchio 1 Neuroscience modeling 2 Introduction to basic aspects of brain computation Introduction to neurophysiology Neural modeling: Elements

More information

A Novel Chaotic Neural Network Architecture

A Novel Chaotic Neural Network Architecture ESANN' proceedings - European Symposium on Artificial Neural Networks Bruges (Belgium), - April, D-Facto public., ISBN ---, pp. - A Novel Neural Network Architecture Nigel Crook and Tjeerd olde Scheper

More information

When do Correlations Increase with Firing Rates? Abstract. Author Summary. Andrea K. Barreiro 1* and Cheng Ly 2

When do Correlations Increase with Firing Rates? Abstract. Author Summary. Andrea K. Barreiro 1* and Cheng Ly 2 When do Correlations Increase with Firing Rates? Andrea K. Barreiro 1* and Cheng Ly 2 1 Department of Mathematics, Southern Methodist University, Dallas, TX 75275 U.S.A. 2 Department of Statistical Sciences

More information

Exercises. Chapter 1. of τ approx that produces the most accurate estimate for this firing pattern.

Exercises. Chapter 1. of τ approx that produces the most accurate estimate for this firing pattern. 1 Exercises Chapter 1 1. Generate spike sequences with a constant firing rate r 0 using a Poisson spike generator. Then, add a refractory period to the model by allowing the firing rate r(t) to depend

More information

arxiv: v1 [cond-mat.stat-mech] 6 Mar 2008

arxiv: v1 [cond-mat.stat-mech] 6 Mar 2008 CD2dBS-v2 Convergence dynamics of 2-dimensional isotropic and anisotropic Bak-Sneppen models Burhan Bakar and Ugur Tirnakli Department of Physics, Faculty of Science, Ege University, 35100 Izmir, Turkey

More information

Chaotic motion. Phys 750 Lecture 9

Chaotic motion. Phys 750 Lecture 9 Chaotic motion Phys 750 Lecture 9 Finite-difference equations Finite difference equation approximates a differential equation as an iterative map (x n+1,v n+1 )=M[(x n,v n )] Evolution from time t =0to

More information

Mathematical Models of Dynamic Behavior of Individual Neural Networks of Central Nervous System

Mathematical Models of Dynamic Behavior of Individual Neural Networks of Central Nervous System Mathematical Models of Dynamic Behavior of Individual Neural Networks of Central Nervous System Dimitra Despoina Pagania 1, Adam Adamopoulos 1,2 and Spiridon D. Likothanassis 1 1 Pattern Recognition Laboratory,

More information

Chaotic Pattern Alternations Can Reproduce Properties Of Dominance Durations In Multistable Perception

Chaotic Pattern Alternations Can Reproduce Properties Of Dominance Durations In Multistable Perception Chaotic Pattern Alternations Can Reproduce Properties Of Dominance Durations In Multistable Perception Takashi Kanamaru Department of Mechanical Science and ngineering, School of Advanced ngineering, Kogakuin

More information

Measures for information propagation in Boolean networks

Measures for information propagation in Boolean networks Physica D 227 (2007) 100 104 www.elsevier.com/locate/physd Measures for information propagation in Boolean networks Pauli Rämö a,, Stuart Kauffman b, Juha Kesseli a, Olli Yli-Harja a a Institute of Signal

More information

CHAPTER 3. Pattern Association. Neural Networks

CHAPTER 3. Pattern Association. Neural Networks CHAPTER 3 Pattern Association Neural Networks Pattern Association learning is the process of forming associations between related patterns. The patterns we associate together may be of the same type or

More information

Self-organization of network structure in coupled-map systems

Self-organization of network structure in coupled-map systems Self-organization of network structure in coupled-map systems Junji Ito and Kunihiko Kaneko Abstract Coupled map models with variable connection weights between the units are studied. A generally observed

More information

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 10 Aug 2004

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 10 Aug 2004 arxiv:cond-mat/4829v [cond-mat.dis-nn] Aug 24 Competition between ferro-retrieval and anti-ferro orders in a Hopfield-like network model for plant intelligence Jun-ichi Inoue a Bikas K. Chakrabarti b a

More information

Chaotic motion. Phys 420/580 Lecture 10

Chaotic motion. Phys 420/580 Lecture 10 Chaotic motion Phys 420/580 Lecture 10 Finite-difference equations Finite difference equation approximates a differential equation as an iterative map (x n+1,v n+1 )=M[(x n,v n )] Evolution from time t

More information

Dynamical Systems and Deep Learning: Overview. Abbas Edalat

Dynamical Systems and Deep Learning: Overview. Abbas Edalat Dynamical Systems and Deep Learning: Overview Abbas Edalat Dynamical Systems The notion of a dynamical system includes the following: A phase or state space, which may be continuous, e.g. the real line,

More information

Using Variable Threshold to Increase Capacity in a Feedback Neural Network

Using Variable Threshold to Increase Capacity in a Feedback Neural Network Using Variable Threshold to Increase Capacity in a Feedback Neural Network Praveen Kuruvada Abstract: The article presents new results on the use of variable thresholds to increase the capacity of a feedback

More information

Biological Modeling of Neural Networks:

Biological Modeling of Neural Networks: Week 14 Dynamics and Plasticity 14.1 Reservoir computing - Review:Random Networks - Computing with rich dynamics Biological Modeling of Neural Networks: 14.2 Random Networks - stationary state - chaos

More information

STUDY OF SYNCHRONIZED MOTIONS IN A ONE-DIMENSIONAL ARRAY OF COUPLED CHAOTIC CIRCUITS

STUDY OF SYNCHRONIZED MOTIONS IN A ONE-DIMENSIONAL ARRAY OF COUPLED CHAOTIC CIRCUITS International Journal of Bifurcation and Chaos, Vol 9, No 11 (1999) 19 4 c World Scientific Publishing Company STUDY OF SYNCHRONIZED MOTIONS IN A ONE-DIMENSIONAL ARRAY OF COUPLED CHAOTIC CIRCUITS ZBIGNIEW

More information

Neural Networks for Machine Learning. Lecture 11a Hopfield Nets

Neural Networks for Machine Learning. Lecture 11a Hopfield Nets Neural Networks for Machine Learning Lecture 11a Hopfield Nets Geoffrey Hinton Nitish Srivastava, Kevin Swersky Tijmen Tieleman Abdel-rahman Mohamed Hopfield Nets A Hopfield net is composed of binary threshold

More information

Switching in Self-Controlled Chaotic Neuromodules

Switching in Self-Controlled Chaotic Neuromodules Switching in Self-Controlled Chaotic Neuromodules Nico Stollenwerk & Frank Pasemann Forschungszentrum Jülich, IBI/MOD, D-52425 Jülich n.stollenwerk@kfa-juelich.de & f.pasemann@kfa-juelich.de Abstract The

More information

Phase Locking. 1 of of 10. The PRC's amplitude determines which frequencies a neuron locks to. The PRC's slope determines if locking is stable

Phase Locking. 1 of of 10. The PRC's amplitude determines which frequencies a neuron locks to. The PRC's slope determines if locking is stable Printed from the Mathematica Help Browser 1 1 of 10 Phase Locking A neuron phase-locks to a periodic input it spikes at a fixed delay [Izhikevich07]. The PRC's amplitude determines which frequencies a

More information

Géza Ódor, MTA-EK-MFA Budapest Ronald Dickman, UFMG Brazil 08/04/2015

Géza Ódor, MTA-EK-MFA Budapest Ronald Dickman, UFMG Brazil 08/04/2015 Localization, Griffiths phases and burstyness in neural network models Géza Ódor, MTA-EK-MFA Budapest Ronald Dickman, UFMG Brazil 08/04/2015 Partners: Gergely Ódor R. Juhász Infocommunication technologies

More information

Canonical Neural Models 1

Canonical Neural Models 1 Canonical Neural Models 1 Frank Hoppensteadt 1 and Eugene zhikevich 2 ntroduction Mathematical modeling is a powerful tool in studying fundamental principles of information processing in the brain. Unfortunately,

More information

P The Entropy Trajectory: A Perspective to Classify Complex Systems. Tomoaki SUZUDO Japan Atomic Energy Research Institute, JAERI

P The Entropy Trajectory: A Perspective to Classify Complex Systems. Tomoaki SUZUDO Japan Atomic Energy Research Institute, JAERI P 0 0 8 The Entropy Trajectory: A Perspective to Classify Complex Systems Tomoaki SUZUDO Japan Atomic Energy Research Institute, JAERI What are complex systems? Having macroscopic properties which are

More information

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 30 Sep 1999

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 30 Sep 1999 arxiv:cond-mat/9909443v1 [cond-mat.dis-nn] 30 Sep 1999 Thresholds in layered neural networks with variable activity 1. Introduction D Bollé and G Massolo Instituut voor Theoretische Fysica, K.U. Leuven,

More information

Vertex Routing Models and Polyhomeostatic Optimization. Claudius Gros. Institute for Theoretical Physics Goethe University Frankfurt, Germany

Vertex Routing Models and Polyhomeostatic Optimization. Claudius Gros. Institute for Theoretical Physics Goethe University Frankfurt, Germany Vertex Routing Models and Polyhomeostatic Optimization Claudius Gros Institute for Theoretical Physics Goethe University Frankfurt, Germany 1 topics in complex system theory Vertex Routing Models modelling

More information

Storage Capacity of Letter Recognition in Hopfield Networks

Storage Capacity of Letter Recognition in Hopfield Networks Storage Capacity of Letter Recognition in Hopfield Networks Gang Wei (gwei@cs.dal.ca) Zheyuan Yu (zyu@cs.dal.ca) Faculty of Computer Science, Dalhousie University, Halifax, N.S., Canada B3H 1W5 Abstract:

More information

Synaptic Plasticity. Introduction. Biophysics of Synaptic Plasticity. Functional Modes of Synaptic Plasticity. Activity-dependent synaptic plasticity:

Synaptic Plasticity. Introduction. Biophysics of Synaptic Plasticity. Functional Modes of Synaptic Plasticity. Activity-dependent synaptic plasticity: Synaptic Plasticity Introduction Dayan and Abbott (2001) Chapter 8 Instructor: Yoonsuck Choe; CPSC 644 Cortical Networks Activity-dependent synaptic plasticity: underlies learning and memory, and plays

More information

Reaction diffusion processes on random and scale-free networks

Reaction diffusion processes on random and scale-free networks Reaction diffusion processes on random and scale-free networks arxiv:cond-mat/4464v1 [cond-mat.stat-mech] 27 Apr 24 Subhasis Banerjee, Shrestha Basu Mallik and Indrani Bose Department of Physics, Bose

More information

Christian Mohr

Christian Mohr Christian Mohr 20.12.2011 Recurrent Networks Networks in which units may have connections to units in the same or preceding layers Also connections to the unit itself possible Already covered: Hopfield

More information