Harnessing Nonlinearity: Predicting Chaotic Systems and Saving

Similar documents
International University Bremen Guided Research Proposal Improve on chaotic time series prediction using MLPs for output training

Reservoir Computing and Echo State Networks

Reservoir Computing Methods for Prognostics and Health Management (PHM) Piero Baraldi Energy Department Politecnico di Milano Italy

Echo State Networks with Filter Neurons and a Delay&Sum Readout

Advanced Methods for Recurrent Neural Networks Design

Refutation of Second Reviewer's Objections

MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION

Error Entropy Criterion in Echo State Network Training

A quick introduction to reservoir computing

Good vibrations: the issue of optimizing dynamical reservoirs

Several ways to solve the MSO problem

Reservoir Computing with Stochastic Bitstream Neurons

Reservoir Computing in Forecasting Financial Markets

Memory Capacity of Input-Driven Echo State NetworksattheEdgeofChaos

Neural Networks. Volker Tresp Summer 2015

Modelling Time Series with Neural Networks. Volker Tresp Summer 2017

Using reservoir computing in a decomposition approach for time series prediction.

Machine Learning for Computer Vision 8. Neural Networks and Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Analyzing the weight dynamics of recurrent learning algorithms

A Practical Guide to Applying Echo State Networks

Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions

( t) Identification and Control of a Nonlinear Bioreactor Plant Using Classical and Dynamical Neural Networks

By Nadha CHAOS THEORY

Recurrent Neural Networks

From Neural Networks To Reservoir Computing...

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

Short Term Memory and Pattern Matching with Simple Echo State Networks

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Using Artificial Neural Networks (ANN) to Control Chaos

Neural Network Training

Neuro-Fuzzy Comp. Ch. 4 March 24, R p

Chaos & Recursive. Ehsan Tahami. (Properties, Dynamics, and Applications ) PHD student of biomedical engineering

Lecture 5: Recurrent Neural Networks

HALF HOURLY ELECTRICITY LOAD PREDICTION USING ECHO STATE NETWORK

From Neural Networks To Reservoir Computing...

Introduction to Neural Networks

Automatic Structure and Parameter Training Methods for Modeling of Mechanical System by Recurrent Neural Networks

REAL-TIME COMPUTING WITHOUT STABLE

Analyzing the State Space Property of Echo

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28

An Introductory Course in Computational Neuroscience

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Short Term Memory Quantifications in Input-Driven Linear Dynamical Systems

Artificial Neural Networks. Edward Gatt

Implicitly and Explicitly Constrained Optimization Problems for Training of Recurrent Neural Networks

Lecture 17: Neural Networks and Deep Learning

Linear Models for Regression

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Chapter 15. Dynamically Driven Recurrent Networks

Optoelectronic Systems Trained With Backpropagation Through Time

Negatively Correlated Echo State Networks

Artificial Neural Networks

1. A discrete-time recurrent network is described by the following equation: y(n + 1) = A y(n) + B x(n)

Real Randomness with Noise and Chaos

ECE 471/571 - Lecture 17. Types of NN. History. Back Propagation. Recurrent (feedback during operation) Feedforward

CSC 578 Neural Networks and Deep Learning

y(n) Time Series Data

Topic 3: Neural Networks

Feed-forward Network Functions

1 What a Neural Network Computes

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Recurrent neural networks with trainable amplitude of activation functions

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

COMPUTATIONAL INTELLIGENCE (INTRODUCTION TO MACHINE LEARNING) SS16

Lecture 3 Feedforward Networks and Backpropagation

Recurrence Enhances the Spatial Encoding of Static Inputs in Reservoir Networks

Introduction to Deep Learning CMPT 733. Steven Bergner

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Artificial Neural Networks

Artificial Neuron (Perceptron)

Deep Learning & Artificial Intelligence WS 2018/2019

At the Edge of Chaos: Real-time Computations and Self-Organized Criticality in Recurrent Neural Networks

Introduction to Convolutional Neural Networks 2018 / 02 / 23

POWER SYSTEM DYNAMIC SECURITY ASSESSMENT CLASSICAL TO MODERN APPROACH

input unit N internal units output unit

Artificial Intelligence

Neural Network Control of Robot Manipulators and Nonlinear Systems

Adaptive Filtering Part II

Short Term Load Forecasting by Using ESN Neural Network Hamedan Province Case Study

Unit 8: Introduction to neural networks. Perceptrons

Edge of Chaos Computation in Mixed-Mode VLSI - A Hard Liquid

ARCHITECTURAL DESIGNS OF ECHO STATE NETWORK

ECE521 Lectures 9 Fully Connected Neural Networks

Backpropagation Introduction to Machine Learning. Matt Gormley Lecture 12 Feb 23, 2018

Information Dynamics Foundations and Applications

Deep Feedforward Networks. Seung-Hoon Na Chonbuk National University

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

A SELF-ORGANIZING RECURRENT NEURAL NETWORK

Midterm, Fall 2003

Based on the original slides of Hung-yi Lee

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

Lecture 3 Feedforward Networks and Backpropagation

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Artificial Neural Networks

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Lecture 2: Learning with neural networks

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Artificial Neural Network

Transcription:

Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication Publishde in Science Magazine, 2004 Siamak Saliminejad

Overview Eco State Networks How to build ESNs Chaotic Systems Equalization Conclusions

Authors Department of Electrical Engineering and Computer Science Jacobs UniversityBremen Dr. Herbert Jaeger (Computer Science) http://www.faculty.jacobs university.de/hjaeger/ Dr. Harald Haas (Electrical Engineering) https://www.jacobs university.de/directory/hhaas

Echo state Networks (ESN) ESN Is a Recurrent Neural Network (RNNs) RNN: is a class of NN where connections between units form a directed d cycle. Win is a K X N matrix Wfb is a L X N matrix W is a N X N matrix W N Wout is a (K+N) X N matrix K Win Wout L Wfb

ESN-Units Unit activation functions are typically sigmoid

ESN System System Equations X(n+1) = f(winu(n+1)+wx(n)+wfby(n)) X: N dimensional Reservoir state n: time U U: K dimensional input signal y: L dimensional output signal f :Sigmoid function Win W N Wout L K Wfb

ESN System System Equations y(n) ( ) = g( ( Wout Z Z(n)) ) where Z X Z(n)=[X(n); u(n)] y: L dimensional output tsignal U: K dimensional input signal X: N dimensional Reservoir state g: Output activation function (typically the identity or a sigmoid) Z: extended system state Win W N Wout L K Wfb

Echo State Property This condition in essence states that the effect of a previous state x(n) and a previous input u(n) on a future state x(n + k) should vanish ihgradually as time passes (i.e., k ( )), and not persist or even get amplified. Having echo states (or not having them) is a property of the network prior to training, that is, a property of the weight matrices Win, W, and (optionally, if they exist) Wfb Thepropertyisalsorelativetothetype of training data: the same untrained network may have echo states for certain training data but not for others. Unfortunately, there is no known necessary and sufficient algebraic condition which allows one to decide, given (Win,, W, Wfb), whether the network has the echo state property.

Building ESNs Step by step approach

Step1: Create Reservoir (W) There are many Reservoir recepies. However, weights (W) and topology are selected randomly, in allof them: I. Classic 1. Big: sufficiently large, with order ranging from tens to thousands. 2. Sparsely: the weight ihmatrix W is sparse, with ihup to 20 % of possible connections. 3. Randomly: the weights of the connections are usually generated randomly from a uniform distribution ib ti symmetric ti around the zero value. II. Different topologies of the reservoir from sparsely randomly connected ones. III. Modular reservoirs: dividing the reservoir into sub reservoirs IV. Layered Reservoir and Optimizing reservoirs for a particular task or a class of tasks in an automated Optimizing reservoirs for a particular task or a class of tasks in an automated fashion is currently the most important field of ESN research.

Step 2: Attach input units to the reservoir (Win) Absolute size of in nput weig hts Random all to all connections Small: the network state is only slightly excited around the DR s resting (zero) state. the network units operate around the linear central part of the sigmoid, i.e. one obtains a network with an almost linear dynamics. Large: network is strongly driven by input and the internal units go closer to the saturation of the sigmoid, which results in a more nonlinear behavior of the resulting model. Very large: the internal units will be driven into an almost pure 1 / +1 valued, binary dynamics. Manual adjustment and repeated learning trials will often be required to find the task appropriate scaling.

Step 3:Wfb If the task requires output feedback install randomly generated output to reservoir connections (all to all). Wback

Step 4: Training Given: A training input/output sequence (u(1), d(1)),..., (u(t), d(t)). Wanted: A trained ESN (Wout) whose output y(n) approximates the teacher output d(n), when the ESN is driven by the training input u(n). In other words, Minimize error. N K Wout L

Step 4: Training Minimize While n:time N: reservoir size T: # of Training pairs y: L dimensional output signal X: N dimensional Reservoir state Any method for computing linear regressions can be used to obtain w

Previous Works Minimize error (d(n) y(n)) Recurrent adapt all connections (input, recurrent, output)by some version of gradient descent. Reservoirs have 5 10 neurons The learning process is slow, may find suboptimal solutions, and is prone to become disrupted by bifurcations A bifurcation of a dynamical system is a qualitative ti change in its dynamics produced by varying parameters.

Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication

Chaotic Systems Chaoticsystems: Time dependent (dynamical) systems thatarehighly are highly sensitive to initial conditions. Arbitrarily small perturbation of the current trajectory may lead to significantly different future behavior (Butterfly Effect) Does the Flap of a Butterfly s Wings in Brazil set off a Tornado in Texas? By Edward Lorenz (1970) Chaotic systems are not periodic, not forever increasing and not approaching a fixed point. But, they are not random! Long term prediction of chaotic systems is impossible (weather) Lasers, oscillating chemical reactions, and fluid dynamics are other examples of chaotic systems

Other Methods wavelet networks (Liangyue Cao, et al., 1995) Likelihood and Bayesian (Berliner, 1991) Regression using Support Vector Machines (Mukherjee, 1997) and RBF (Rosipal etal, 1998) feedforward network trained with backpropagation Chakroporty et al (1992), Elsner et al (1991), Andresia etal(2000) Recurrent Neural Networks Echo State Networks (ESN) real time recurrent learning (Williams and Zipser 1989) Previous Works backpropagation through time (Werbos 1990) extended Kalman filteringbased methods (Puskoriusetal 2004), Atiya Parlos algorithm (Atiya and Parlos 2000)

Mackey Glass System It is a standard benchmark system Reservoir size (N) = 1000 The connectivity was 1% A 1000 X 1000 matrix was constructed Random weights drawn from a uniform distribution over ( 1,1) Uses feedback connections First 1000 steps were discarded to wash out initial transient. Echo signals x(n) were sampled from remaining 2000 steps and the network trained by them The results show a jumpinmodelingaccuracy modeling with respect to previous models

Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication

High frequency signal Transmission i High Distortion Demodulation d(n) Modulation High High Efficiency u(n) Equalization y(n)

Equalization of wireless channels A 46 neuron reservoir and W were randomly generated with ih a connectivity ii of 20%, and nonzero connection weights drawn from a uniform distribution over ( 1,1). Output neuron was a linear neuron Results Showed an improvement of two magnitudes for high signal to noise ratios.

Summary and conclusion The mathematical properties p of large RNNs such that they can be used with a linear, trainable readout mechanism for general blackbox modeling are elucidated. ESN is faster and more applicable than previous methods that try to train all the connections. ESNs can be applied to all basic tasks of signal processing and control ESNs have been developed from a mathematical and engineering perspective, but exhibit typical features of biological i l RNNs