Ingo Ahrns, Jorg Bruske, Gerald Sommer. Christian Albrechts University of Kiel - Cognitive Systems Group. Preusserstr Kiel - Germany

Similar documents
Lyapunov Stability Analysis of the Quantization Error for DCS Neural Networks

Abstract: Q-learning as well as other learning paradigms depend strongly on the representation

y(n) Time Series Data

Abstract This report has the purpose of describing several algorithms from the literature all related to competitive learning. A uniform terminology i

Introduction to Neural Networks

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

University of Genova - DITEN. Smart Patrolling. video and SIgnal Processing for Telecommunications ISIP40

Learning Vector Quantization

Artificial Neural Networks Examination, March 2004

Artificial Neural Networks. Edward Gatt

Learning Vector Quantization (LVQ)

Neural Networks. Fundamentals Framework for distributed processing Network topologies Training of ANN s Notation Perceptron Back Propagation

Supervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir

Generalization and Function Approximation

Lyapunov Analysis of Neural Network Stability in an Adaptive Flight Control System

Extending the Associative Rule Chaining Architecture for Multiple Arity Rules

Neural Network to Control Output of Hidden Node According to Input Patterns

Artificial Neural Networks Examination, March 2002

Artificial Neural Networks Examination, June 2005

A Novel Approach to the 2D Analytic Signal? Thomas Bulow and Gerald Sommer. Christian{Albrechts{Universitat zu Kiel

Neural Networks for Two-Group Classification Problems with Monotonicity Hints

Artificial Neural Network : Training

1 Introduction Tasks like voice or face recognition are quite dicult to realize with conventional computer systems, even for the most powerful of them

Short Communication EVOLUTIONARY NEURAL GAS: A SCALE-FREE SELF- ORGANIZING NEURAL NET WITHOUT TOPOLOGICAL CONSTRAINT

ESANN'2001 proceedings - European Symposium on Artificial Neural Networks Bruges (Belgium), April 2001, D-Facto public., ISBN ,

Artificial Neural Networks Examination, June 2004

A Dynamical Implementation of Self-organizing Maps. Wolfgang Banzhaf. Department of Computer Science, University of Dortmund, Germany.

Sample Exam COMP 9444 NEURAL NETWORKS Solutions

Using a Hopfield Network: A Nuts and Bolts Approach

Dynamic Classification of Power Systems via Kohonen Neural Network

Gaussian process for nonstationary time series prediction

Week 4: Hopfield Network

CHALMERS, GÖTEBORGS UNIVERSITET. EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

Remaining energy on log scale Number of linear PCA components

Iterative Laplacian Score for Feature Selection

Feedforward Neural Nets and Backpropagation

Multi-Robotic Systems

On Computational Limitations of Neural Network Architectures

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28

2- AUTOASSOCIATIVE NET - The feedforward autoassociative net considered in this section is a special case of the heteroassociative net.

CHAPTER IX Radial Basis Function Networks

Neural Networks Lecture 7: Self Organizing Maps

This is a repository copy of Improving the associative rule chaining architecture.

Self-organized criticality and the self-organizing map

6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

PDF hosted at the Radboud Repository of the Radboud University Nijmegen

Data Mining Part 5. Prediction

8. Lecture Neural Networks

Chapter 9: The Perceptron

Explaining Results of Neural Networks by Contextual Importance and Utility

Introduction to Neural Networks

Artificial Neural Network and Fuzzy Logic

Learning and Extracting Initial Mealy Automata With a. Modular Neural Network Model. Peter Tino. Department of Informatics and Computer Systems

Covariance Learning of Correlated Patterns in Competitive. Networks. Ali A. Minai. University of Cincinnati. (To appearinneural Computation)

Effects of Interactive Function Forms in a Self-Organized Critical Model Based on Neural Networks

Evaluation of Spectra in Chemistry and Physics. with Kohonen's Selforganizing Feature Map. Lehrstuhl fuer technische Informatik

1. A discrete-time recurrent network is described by the following equation: y(n + 1) = A y(n) + B x(n)

Matching the dimensionality of maps with that of the data

Classic K -means clustering. Classic K -means example (K = 2) Finding the optimal w k. Finding the optimal s n J =

Christian Mohr

Hopfield Network Recurrent Netorks

Φ 12 Φ 21 Φ 22 Φ L2 Φ LL Φ L1

Analysis of Interest Rate Curves Clustering Using Self-Organising Maps

Introduction to Neural Networks: Structure and Training

Perceptron. (c) Marcin Sydow. Summary. Perceptron

Several ways to solve the MSO problem

Neural Networks. Nethra Sambamoorthi, Ph.D. Jan CRMportals Inc., Nethra Sambamoorthi, Ph.D. Phone:

7 Rate-Based Recurrent Networks of Threshold Neurons: Basis for Associative Memory

Ch.8 Neural Networks

SEPARATION OF ACOUSTIC SIGNALS USING SELF-ORGANIZING NEURAL NETWORKS. Temujin Gautama & Marc M. Van Hulle

Integer weight training by differential evolution algorithms

9 Competitive Neural Networks

ESANN'2001 proceedings - European Symposium on Artificial Neural Networks Bruges (Belgium), April 2001, D-Facto public., ISBN ,

A Study on Energy Artificial Neuron Model and Its Applications in Self-growing and Self-organizing Neural Network

Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

7 Recurrent Networks of Threshold (Binary) Neurons: Basis for Associative Memory

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

Neural networks: Unsupervised learning

Administration. Registration Hw3 is out. Lecture Captioning (Extra-Credit) Scribing lectures. Questions. Due on Thursday 10/6

Unit 8: Introduction to neural networks. Perceptrons

Carnegie Mellon University Forbes Ave. Pittsburgh, PA 15213, USA. fmunos, leemon, V (x)ln + max. cost functional [3].

Training Multi-Layer Neural Networks. - the Back-Propagation Method. (c) Marcin Sydow

Scuola di Calcolo Scientifico con MATLAB (SCSM) 2017 Palermo 31 Luglio - 4 Agosto 2017

Computational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification

Neural Networks Lecture 2:Single Layer Classifiers

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

RBF Neural Network Combined with Knowledge Mining Based on Environment Simulation Applied for Photovoltaic Generation Forecasting

Information Capacity of Binary Weights Associative. Memories. Mathematical Sciences. University at Bualo. Bualo NY

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

5. KOHONEN S NETWORK FOR MODELING THE AUDITORY CORTEX OF A BAT

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

Radial Basis Function (RBF) Networks

ANALYSIS OF CONSENSUS AND COLLISION AVOIDANCE USING THE COLLISION CONE APPROACH IN THE PRESENCE OF TIME DELAYS. A Thesis by. Dipendra Khatiwada

memory networks, have been proposed by Hopeld (1982), Lapedes and Farber (1986), Almeida (1987), Pineda (1988), and Rohwer and Forrest (1987). Other r

Neural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET

Simple Neural Nets For Pattern Classification

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Transcription:

On-line Learning with Dynamic Cell Structures Ingo Ahrns, Jorg Bruske, Gerald Sommer Christian Albrechts University of Kiel - Cognitive Systems Group Preusserstr. 1-9 - 24118 Kiel - Germany Phone: ++49 431 560480 Fax: ++49 431 560481 e-mail: jbr@informatik.uni-kiel.de Abstract Dynamic Cell Structures (DCS) belong to the class of topology representing networks recently introduced by [Ma94]. DCS networks rest upon radial basis function networks (RBF) and an additional lateral connection structure between the neural units. This connection structure reects the topology of the input manifold. Initial simulations on CMU benchmarks presented in [Br95] are very encouraging but were all involved with o-line learning tasks. In this article we argue that DCS networks are also capable of performing on-line learning tasks where the training examples are generated by a dynamic system (i.e. not a xed training set). The proposed modications to the original learning scheme in order to cope with the dynamics bear relevance not only for DCS but for Kohonen type networks and TRN in general. Our test environment is pole-balancing. 1 Introduction Like Kohonen's famous SOM Dynamic Cell Structures (DCS) are capable of adaptive vector quantization, clustering, dimensionality reduction and have been extended for supervised learning. Moreover, DCS-networks form \topology preserving" feature maps [Ma94]. In this section we briey summarize the foundations of DCS. Section 2 presents our extensions for better coping with dynamic systems and in section 3 we report simulation results. Consider an input manifold I IR n and an output manifold O IR m. A neural DCS-unit is a point u = (c u ; w u ; u ; r u ) 2 I O [0; 1] I IR 0, where c u is called the center, w u the weight, u the activation function and r u the resource of the neural unit. A DCS-graph is an undirected graph N = (U; L) where U I O [0; 1] I IR 0 is a nite set of neural units and L ffu; vgju; v 2 U ^ u 6= vg is a set of lateral connections between dierent neural units. Published in: Pro c. of ICANN'95, Vol.2 (1995) 141-146

The lateral connection-strength is a function : L! IR 0 which is usually given by a symmetrical adjacency matrix C 2 IR ju jju j for which C ii = 0 for all i 2 f1; ; jujg. Moreover, C ij = 0 if unit i and unit j are not connected, and 0 < C ij 1 if unit i and unit j are connected with strength C ij. In general we have (fu i ; u j g) = C ij. A DCS-network is a triple (U; L; ) of a DCS-graph (U; L) and a lateral connectionstrength. In order to perform perfect topology learning, a simple competitive Hebbian learning rule is used ([Br95]). It was proven in [Ma94] that a lateral connection structure of neural units formed by the competitive Hebbian learning rule yields a perfect topology preserving map. In [Br95] and [Ma92] this learning rule was extended by introducing a forgetting constant in case of a non-stationary input distribution or changing neighborhood relations. GDCS is a growing DCS network that employs ideas from Fritzke's Growing Cell Structures (GCS) [Fr95]. Like in GCS each neuron contains a local error variable which is called the resource of a neural unit. Such a resource variable accumulates error values if its neuron wins the competition for a given input stimulus and becomes the best matching unit. In case of supervised learning one can e.g. use the squared output error as an error measure. The reason for introducing resources is to guide the insertion of new neural units by the local error in order to get a uniform distribution of all resource values. In GCS and GDCS centers are adapted in a Kohonen-like fashion, such that the centers of the best matching unit (bmu) and its neighbors are moved into the direction of the last input. In this article we argue that an improved adaptation rule can be obtained by taking into account the error variables? much like the classication performance is taken into account in LVQ-algorithms [Ko92]. 2 DCS for dynamic system control The GDCS-networks mentioned above has demonstrated eectiveness in supervised and unsupervised learning tasks [Br95]. Training sets were static and have been mixed to remove the inuence of the order of sample presentation. This kind of neural net training is legitimate if all training data are directly attainable. The focus on this article is on demonstrating how GDCS are able to learn control functions on-line. 2.1 The Task of Pole-Balancing Pole-balancing is a famous example of a dynamic system which has often been reported to be a dicult control problem. In this article we consider the pole-balancing problem in its simplest version adopted from Ritter [Ri91]. System dynamics are then comprised by the following dierential equation: (m + sin 2 ) + 1 2 _ 2 sin(2)? (m + 1) sin =?f cos : (1) Here, represents the angle of inclination of the pole, m is the mass of the base point of the pole, and the force applied to the base point is denoted as f. As an appropriate function for balancing the inverse pendulum, the force f L (; _ ) = 5 sin + _ was taken

as a teacher signal for supervised learning. In our simulations we released the pole at a randomly chosen angle?40 40 with zero velocity. The force for controlling the pole can be obtained by f = (t k )y(; _ ) + (1? (t k ))f L (; _ ); (2) where : [0; T ]! [0; 1] represents a monotonically increasing function which expresses the reliability of the control-function y of the network. 2.2 Improvements on DCS Let us now discuss some modications of DCS with respect to on-line learning in control problems. 2.2.1 Mixed insertion strategy In the classical GDCS approach, a new neural unit is inserted after teaching steps between the units with the largest resource values. We call this method error driven unit insertion. It is useful, if one has a xed training set T. In our control problem we don't have any xed training sets. Rather training data are obtained according to the dynamics of the controlled system which can cause severe problems with unfolding in GCS as well as GDCS. The solution is to augment GDCS with a second insertion strategy operating in parallel to the error driven insertion strategy: We introduce a threshold 2 IR 0 which indicates the maximum distance of an input stimulus to the center of the best matching unit. If the distance exceeds and the number of neural units is below a given maximum unit number, we insert a new neural unit at the position of the given training pair in I O. That new unit can be connected to the last best matching unit. Such an additional distance driven unit insertion is quite common, see e.g. [Ha93],[Pl91], where insertion takes place if the stimulus distance and the approximation error both exceed a threshold. Besides coping with unfolding problems, the distance driven insertion yields a coarse scale approximation which will be rened by the error driven insertion strategy. 2.2.2 Non-Symmetric Lateral Connection Structure As pointed out in [Ma92], it is possible to update the lateral connections of the best matching unit only. If we use a competitive Hebb rule as proposed in [Br95] for topology learning, including the deletion of lateral connections in the case of C ij <, unconnected neural units can result. Unconnected neural units become superuous if they lie in regions of the input manifold where input stimuli have low probability to stem from. Such neural units are called dead neurons. In order to avoid such neurons we now give up the demand of symmetry and adapt the lateral connections of the bmu non-symmetrically. Consequently, lateral connections will only be removed, if the connection strengths of both connected

units fall below. We obtain: C ij := 8 > <>: 1 : fi; jg = fbmu; secondg 0 : i = bmu ^ j 6= second ^ ( C ij < ^ C ji < ) 0 : j = bmu ^ i 6= second ^ ( C ji < ^ C ij < ) C ij : i = bmu ^ j 6= second ^ ( C ij _ C ji ) C ij : otherwise ; (3) where the best matching unit and the second best matching unit for a given input stimulus x 2 I are dened by 8u 2 U : kc bmu? xk kc u? xk; 8u 2 U n fbmug : kc second? xk kc u? xk (4) 2.2.3 Error Modulated Kohonen Rule for Center Adaptation The next modication of GDCS concerns the adaptation rule for center updating. If training data are generated by a dymamical system like the inverse pendulum tendencies will result which usually infers with traditional training methods. Another well known problem with Kohonen-type feature maps is to avoid that regions where inputs have a high temporary probability density attract all neural units. Instead of reecting the input probability density the distribution of neural units should result in a uniform distribution of local approximation errors. This is the motivation for our error modulated Kohonenrule: Let the resource values accumulate a kind of mean value of error signals (independent of frequency). We employ: r bmu := r bmu + (1? ) err; 2 [0; 1]: (5) Now let c : IR 0! [0; 1] be a monotonically decreasing function with c(0) = 1 and c(1) = 0. Our modied update rule is c bmu := c bmu + c r " bmu (x? c bmu ); (6) r bmu 8u 2 Nh(bmu) : c u := c u + c ru " N h (x? c u ); (7) r bmu where r = 1 jnh(bmu)j X u2n h(bmu) r u (8) and N h(bmu) := fvjfbmu; vg 2 Lg: (9) Hence the bmu is only attracted by the input stimulus if r bmu > r and neighbors are only attracted if their resource is below the bmu's resource, aiming at a uniform distribution of resource values. 3 Simulation results We applied our improved GDCS-network to the task of pole balancing mentioned above. The GDCS-network was trained for a duration of k max training steps. After a balancing

failure (jj > 45 ) or after k try successful balancing steps the simulation was started again. For our experiments we used a DCS-network with the mixed unit insertion strategy, the non-symmetric lateral connection structure and the error modulated Kohonen rule for center adaptation. After 3000 learning steps our network, consisting of 70 neural units, was able to balance the pole, see Fig. 1. a) b) c) d) Figure 1: a), b) DCS-network. c) Approximate function y(; ). _ d) (t) during polebalancing after training-phase. The parameter set of our simulation was: k max = 3000, k try = 100, time step width t = 0:3 s, pendulum mass m = 1:0 kg, learning constants " bmu = 0.1, " N h = 0.001, connection threshold = 0.01, insertion constant = 0.075, insertion step width = 100, resource averaging = 0.8, connection forgetting rate = 0:88, reliability function (t k ) = and c(x) = 2x 3? 3x 2 + 1 if x 2 [0; 1] and c(x) = 0 otherwise. Neural activations were computed by Gaussian basis-functions with a standard deviation of = 0:2. Output weights were adapted by using a simple delta-rule with a learning constant of = 0:001. k k max

4 Conclusion On-line learning with respect to control of dynamic systems is dicult due to the non stationary probability density of input stimuli and resulting tendencies in the training data. In this article we proposed three modications to the original GDCS learning scheme to cope with this problem without \freezing" of learning constants. First, our error modulated Kohonen type learning rule is no longer dominated by the input probability density but aims at a uniform approximation error (uniform distribution of resources) independent of the probability density. This idea applies to GCS and extensions of the Kohonen map as well. Second it was found that tendencies in the data can cause severe unfolding problems in GCS or GDCS if error driven insertion is employed only. We therefore decided to augment GDCS with a distance driven insertion scheme for approximation on a coarse scale. Error driven insertion remains responsible for error driven renement. The idea applies to GCS as well. Finally, TRN may suer from dead neurons if lateral connections are allowed to be forgotten. Our solution to this problem is to introduce unsymmetric lateral connections. Pole balancing in its simplest version was used for demonstration. Despite of its simplicity this simulation nevertheless had to cope with all the afore mentioned diculties in on-line learning of dynamic systems. Currently we work on using GDCS for obstacle avoidance on a TRC Labmate mobile robot. References [Br95] Jorg Bruske, Gerald Sommer: Dynamic Cell Structure learns Perfectly Topology Preserving Map, Neural Computation, Vol. 7, No. 4, 1995. [Ma92] Thomas Martinez: Selbstorganisierende neuronale Netzwerke zur Bewegungssteuerung, Dissertation, DIFKI-Verlag, 1992. [Ma94] Thomas Martinez: Topology Representing Networks, Neural Networks, Vol.7, pp. 505-522, 1994. [Ri91] H. Ritter, T. Martinetz, K. Schulten: Neuronale Netze, Addison Wesley, 1991. [Fr95] B. Fritzke: Growing Cell Structures - a self organizing network for unsupervised and supervised learning, Neural Networks, Vol.7, pp.1441-1460, 1995. [Ha93] J. Hakala and R. Eckmiller: Node Allocation and Topographical Encoding (NATE) for Inverse Kinematics of a Redundant Robot Arm for Inverse Kinematics of a 6- DOF Robot Arm, ICANN 93, pp. 309-312, 1993. [Ko88] Teuvo Kohonen: Adaptive, associative, and self-organizing functions in neural computing, Applied Optics, Vol. 26, pp.4910-4918, 1987. [Ko92] T. Kohonen et. al.: LVQ-PAK, The Learning Vector Quantization Program Package, Version 2.0, Helsinki University of Technology, 1992. [Pl91] J. Platt: A Resource-Allocating Network for Function Interpolation, Neural Computation, Vol.2, No. 3, pp.213-225, 1991.