Motivation. Lecture 23. The Stochastic Neuron ( ) Properties of Logistic Sigmoid. Logistic Sigmoid With Varying T. Logistic Sigmoid T = 0.

Similar documents
( ) T. Reading. Lecture 22. Definition of Covariance. Imprinting Multiple Patterns. Characteristics of Hopfield Memory

VI: Genetics & Evolution 11/12/03 1

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i )

Neural Nets and Symbolic Reasoning Hopfield Networks

7 Rate-Based Recurrent Networks of Threshold Neurons: Basis for Associative Memory

Dynamical Systems and Deep Learning: Overview. Abbas Edalat

Microcanonical Mean Field Annealing: A New Algorithm for Increasing the Convergence Speed of Mean Field Annealing.

7 Recurrent Networks of Threshold (Binary) Neurons: Basis for Associative Memory

COMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines. COMP9444 c Alan Blair, 2017

Hopfield Network for Associative Memory

( ) ( ) ( ) ( ) Simulated Annealing. Introduction. Pseudotemperature, Free Energy and Entropy. A Short Detour into Statistical Mechanics.

SIMU L TED ATED ANNEA L NG ING

Markov Chains and MCMC

Neural Networks for Machine Learning. Lecture 11a Hopfield Nets

Optimization Methods via Simulation

Stochastic Networks Variations of the Hopfield model

Energy-Decreasing Dynamics in Mean-Field Spin Models

5. Simulated Annealing 5.1 Basic Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini

Hopfield Networks. (Excerpt from a Basic Course at IK 2008) Herbert Jaeger. Jacobs University Bremen

Branislav K. Nikolić

Markov Chain Monte Carlo. Simulated Annealing.

Learning and Memory in Neural Networks

A.I.: Beyond Classical Search

immune response, spin glasses and, 174, 228 Ising model, definition of, 79-80

Stochastic Model for Adaptation Using Basin Hopping Dynamics

Lecture 4: Simulated Annealing. An Introduction to Meta-Heuristics, Produced by Qiangfu Zhao (Since 2012), All rights reserved

3.012 PS 6 THERMODYANMICS SOLUTIONS Issued: Fall 2005 Due:

Nonequilibrium dynamics in Coulomb glasses near the metal-insulator transition

5. Simulated Annealing 5.2 Advanced Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini

Physics of disordered materials. Gunnar A. Niklasson Solid State Physics Department of Engineering Sciences Uppsala University

Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms

Good vibrations: the issue of optimizing dynamical reservoirs

7.1 Basis for Boltzmann machine. 7. Boltzmann machines

CSC321 Lecture 8: Optimization

AI Programming CS F-20 Neural Networks

Spin Glas Dynamics and Stochastic Optimization Schemes. Karl Heinz Hoffmann TU Chemnitz

Entropy. Governing Rules Series. Instructor s Guide. Table of Contents

ENTROPY

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Associative Memories (I) Hopfield Networks

Onsager theory: overview

Effects of Interactive Function Forms in a Self-Organized Critical Model Based on Neural Networks

Emergence of complex structure through co-evolution:

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 18 Nov 1996

Introduction to Fluctuation Theorems

Relaxation in Glass: Review of Thermodynamics. Lecture 11: Thermodynamics in the Glass Transition Region

Systems Biology: A Personal View IX. Landscapes. Sitabhra Sinha IMSc Chennai

T sg. α c (0)= T=1/β. α c (T ) α=p/n

Logic Learning in Hopfield Networks

Chemical thermodynamics the area of chemistry that deals with energy relationships

Lecture 5: Logistic Regression. Neural Networks

Energy Landscapes of Deep Networks

Energy Barriers and Rates - Transition State Theory for Physicists

Boltzmann Machine and Hyperbolic Activation Function in Higher Order Network

Collective Effects. Equilibrium and Nonequilibrium Physics

SYSTEMS SCIENCE AND CYBERNETICS Vol. III Simulated Annealing: From Statistical Thermodynamics to Combinatory Problems Solving - D.

Slightly off-equilibrium dynamics

Self-Organization in Nonequilibrium Systems

CSC321 Lecture 7: Optimization

Part II. Combinatorial Models. 5 A Sketch of Basic Statistical Physics. 5.1 Thermodynamics

Through Kaleidoscope Eyes Spin Glasses Experimental Results and Theoretical Concepts

HOPFIELD neural networks (HNNs) are a class of nonlinear

Mineral Stability and Phase Diagrams Introduction

Statistical Thermodynamics

Hill climbing: Simulated annealing and Tabu search

Neural Networks. Textbook. Other Textbooks and Books. Course Info. (click on

Single Solution-based Metaheuristics

Casting Polymer Nets To Optimize Molecular Codes

Simulated Annealing. 2.1 Introduction

Is there a de Almeida-Thouless line in finite-dimensional spin glasses? (and why you should care)

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE

Analysis of Neural Networks with Chaotic Dynamics

Influence of Criticality on 1/f α Spectral Characteristics of Cortical Neuron Populations

Second Law of Thermodynamics

Deriving Thermodynamics from Linear Dissipativity Theory

Case study: molecular dynamics of solvent diffusion in polymers

Storage Capacity of Letter Recognition in Hopfield Networks

Advanced computational methods X Selected Topics: SGD

6 Markov Chain Monte Carlo (MCMC)

Chapter 11. Stochastic Methods Rooted in Statistical Mechanics

Chapter 19. Chemical Thermodynamics

Markov Chain Monte Carlo Lecture 4

IN THIS PAPER, we consider a class of continuous-time recurrent

Chapter 11 Spontaneous Change and Equilibrium

OPTIMIZATION BY SIMULATED ANNEALING: A NECESSARY AND SUFFICIENT CONDITION FOR CONVERGENCE. Bruce Hajek* University of Illinois at Champaign-Urbana

SAMPLE ANSWERS TO HW SET 3B

Zebo Peng Embedded Systems Laboratory IDA, Linköping University

Terminal attractor optical associative memory with adaptive control parameter

Landscape Approach to Glass Transition and Relaxation. Lecture # 4, April 1 (last of four lectures)

Neural networks. Chapter 19, Sections 1 5 1

Vertex Routing Models and Polyhomeostatic Optimization. Claudius Gros. Institute for Theoretical Physics Goethe University Frankfurt, Germany

4. References. words, the pseudo Laplacian is given by max {! 2 x,! 2 y}.

Fundamentals of Metaheuristics

Web Course Physical Properties of Glass. Range Behavior

Lecture 2: Learning from Evaluative Feedback. or Bandit Problems

Artificial Neural Network Simulation of Battery Performance

Chapter 19 Chemical Thermodynamics

Transcription:

Motivation Lecture 23 Idea: with low probability, go against the local field move up the energy surface make the wrong microdecision Potential value for optimization: escape from local optima Potential value for associative memory: escape from spurious states because they have higher energy than imprinted states /2/07 /2/07 2 The Stochastic Neuron Deterministic neuron : s " i = sgn h i Pr { s " i = +} = # h i Pr { s " i = $} =$ # h i Stochastic neuron : Pr { s " i = +} = #( h i ) Pr { s " i = $} =$# h i Logistic sigmoid : " h = + exp #2 h T σ(h) h Properties of As h +, σ(h) As h, σ(h) 0 σ(0) = /2 = " h + e #2h T /2/07 3 /2/07 4 With Varying T T = 0.5 T varying from 0.05 to (/T = β = 0,, 2,, 20) Slope at origin = / 2T /2/07 5 /2/07 6

T = 0.0 T = 0. /2/07 7 /2/07 8 T = T = 0 /2/07 9 /2/07 0 T = 00 Pseudo-Temperature Temperature = measure of thermal energy (heat) Thermal energy = vibrational energy of molecules A source of random motion Pseudo-temperature = a measure of nondirected (random) change Logistic sigmoid gives same equilibrium probabilities as Boltzmann-Gibbs distribution /2/07 /2/07 2 2

Transition Probability Recall, change in energy "E = #"s k h k = 2s k h k Pr { s " k = ±s k = m} = #( ±h k ) = #( $s k h k ) Pr{ s k " #s k } = + exp 2s k h k T = + exp $E T /2/07 3 Stability Are stochastic Hopfield nets stable? Thermal noise prevents absolute stability But with symmetric weights: average values s i become time - invariant /2/07 4 Does Thermal Noise Improve Memory Performance? Experiments by Bar-Yam (pp. 36-20): n = 00 p = 8 Random initial state To allow convergence, after 20 cycles set T = 0 How often does it converge to an imprinted pattern? /2/07 5 Probability of Random State Converging on Imprinted State (n=00, p=8) T = / β /2/07 (fig. from Bar-Yam) 6 Probability of Random State Converging on Imprinted State (n=00, p=8) Analysis of Stochastic Hopfield Network Complete analysis by Daniel J. Amit & colleagues in mid-80s See D. J. Amit, Modeling Brain Function: The World of Attractor Neural Networks, Cambridge Univ. Press, 989. The analysis is beyond the scope of this course /2/07 (fig. from Bar-Yam) 7 /2/07 8 3

Phase Diagram (D) all states melt Conceptual Diagrams of Energy Landscape (C) spin-glass states (A) imprinted = minima (B) imprinted, but s.g. = min. /2/07 (fig. from Domany & al. 99) 9 /2/07 (fig. from Hertz & al. Intr. Theory Neur. Comp.) 20 Phase Diagram Detail Simulated Annealing (Kirkpatrick, Gelatt & Vecchi, 983) /2/07 (fig. from Domany & al. 99) 2 /2/07 22 Dilemma In the early stages of search, we want a high temperature, so that we will explore the space and find the basins of the global minimum In the later stages we want a low temperature, so that we will relax into the global minimum and not wander away from it Solution: decrease the temperature gradually during search /2/07 23 Quenching vs. Annealing Quenching: rapid cooling of a hot material may result in defects & brittleness local order but global disorder locally low-energy, globally frustrated Annealing: slow cooling (or alternate heating & cooling) reaches equilibrium at each temperature allows global order to emerge achieves global low-energy state /2/07 24 4

Multiple Domains Moving Domain Boundaries global incoherence local coherence /2/07 25 /2/07 26 Effect of Moderate Temperature Effect of High Temperature /2/07 (fig. from Anderson Intr. Neur. Comp.) 27 /2/07 (fig. from Anderson Intr. Neur. Comp.) 28 Effect of Low Temperature Annealing Schedule Controlled decrease of temperature Should be sufficiently slow to allow equilibrium to be reached at each temperature With sufficiently slow annealing, the global minimum will be found with probability Design of schedules is a topic of research /2/07 (fig. from Anderson Intr. Neur. Comp.) 29 /2/07 30 5

Typical Practical Annealing Schedule Initial temperature T 0 sufficiently high so all transitions allowed Exponential cooling: T k+ = αt k typical 0.8 < α < 0.99 at least 0 accepted transitions at each temp. Final temperature: three successive temperatures without required number of accepted transitions Summary Non-directed change (random motion) permits escape from local optima and spurious states Pseudo-temperature can be controlled to adjust relative degree of exploration and exploitation /2/07 3 /2/07 32 Additional Bibliography. Anderson, J.A. An Introduction to Neural Networks, MIT, 995. 2. Arbib, M. (ed.) Handbook of Brain Theory & Neural Networks, MIT, 995. 3. Hertz, J., Krogh, A., & Palmer, R. G. Introduction to the Theory of Neural Computation, Addison-Wesley, 99. V. Genetics & Evolution /2/07 33 /2/07 34 The Second Law of Thermodynamics The Second Law and Open Systems closed system entropy H H open system energy concentration H waste /2/07 35 /2/07 36 6

Nonequilibrium Thermodynamics Classical thermodynamics limited to systems in equilibrium Extended by thermodynamics of transport processes i.e. accounting for entropy changes when matter/energy transported into or out of an open system Flow of matter/energy can maintain a dissipative system far from equilibrium for long periods Hence, nonequilibrium thermodynamics An Energy Flow Can Create Structure /2/07 37 /2/07 (photo from Camazine & al. Self-Org. Bio. Sys.) 38 Bénard Convection Cells Persistent Nonequilibrium Systems If flow creates system so structured to maintain flow then positive feedback causes nonequilibrium system to persist indefinitely but not forever (2 nd law) Systems we tend to see are those most successful at maintaining nonequil. state Applies to species as well as organisms /2/07 (photo from Camazine & al. Self-Org. Bio. Sys.) 39 /2/07 40 Order Through Fluctuations Stabilization at Successively Higher Levels of Organization Fluctuations (esp. when system forced out of ordinary operating range) test boundaries & nonlinear effects May lead to stabilization of new structures /2/07 4 fig. < Hart & Gregor, Inf. Sys. Found. /2/07 42 fig. < Hart & Gregor, Inf. Sys. Found. 7

Order for Free Relatively simple sets of rules or equations can generate rich structures & behaviors Small changes can lead to qualitatively different structures & behaviors A diverse resource for selection A basis for later fine tuning (microevolution) See Kaufmann (At Home in the Universe, etc.) and Wolfram (A New Kind of Science) Evolution in Broad Sense Evolution in the broadest terms: blind variation selective retention Has been applied to nonbiological evolution evolutionary epistemology creativity memes /2/07 43 /2/07 44 Evolution atoms & molecules replicating molecules living things prebiotic evolution biotic evolution /2/07 45 /2/07 (from NECSI) 46 8