A quick introduction to reservoir computing

Size: px
Start display at page:

Download "A quick introduction to reservoir computing"

Transcription

1 A quick introduction to reservoir computing Herbert Jaeger Jacobs University Bremen

2 1 Recurrent neural networks

3 Feedforward and recurrent ANNs A. feedforward B. recurrent Characteristics: Has at least one feedback loop of connections Thus can maintain dynamic activation even without input input time series output time series can approximate any dynamical system (universal approximation property, Funahashi & Nakamura 1993) mathematical analysis difficult learning algorithms computationally expensive and difficult to master few application-oriented publications, little research

4 Hopfield networks, Boltzmann machines, Deep Belief Networks, ART networks, etc. These are also recurrent. Differences: Symmetric connections Energy-based models with equilibrium / stochastic sampling dynamics For pattern recognition and associative memory tasks, -- not for time series processing (only if forced to...)

5 Problems with "classical" gradientbased training methods for RNNs Besides the generic difficulties of gradient-descent methods, in RNNs there are some more: passing through bifurcations error gradient information shrinks exponentially over time long memory effects difficult to train ("long short-term memory" (LSTM) RNNs are an exception, Gers et al 2000) algorithms and maths are difficult experienced professionals needed But: powerful in the hands of "deep learning" experts who have learnt how to tame them

6 2 Echo State Networks Basics

7 Echo state networks Developed independently and simultaneously with Liquid State Machines (LSM) (Wolfgang Maass, TU Graz) Basic idea of ESNs and LSMs: Use a large, random, fixed RNN (called dynamical reservoir, is not changed by training!) Train only the connections from reservoir to output units! traditional, e.g. BPTT: all connections are trained. ESN training: train only output weights

8 Illustrating the principle: training a tone generator Goal: train a network to work as a tuneable tone generator input: frequency setting output: sines of desired frequency

9 Phase 1: internal state collection Drive fixed "reservoir" network with teacher input and output, and save reservoir units time signals. Observation: internal states of dynamical reservoir reflect and modify both input and output teacher signals

10 Phase 2 (and finish): compute output weights Determine reservoir-to-output weights such that training output is optimally reconstituted from internal "echo" signals in a least mean square error sense. This is a linear regression on the teacher output signal (red) on the reservoir signals.

11 Tone generator: exploitation With new output weights in place, drive trained network with input. Observation: network continues to function as in training. internal states reflect input and output output is reconstituted from internal states internal states and output constitute each other

12 Dynamical reservoir large recurrent network ( units) works as "dynamical reservoir" output units combine different internal dynamics into desired dynamics i n p u t u n i t r e c u r r e n t " d y n a m i c a l r e s e r v o i r " o u t p u t u n i t s

13 3 Basic Examples Predicting chaotic timeseries Event detection for robots Speaker identification ("Japanese Vouwel" benchmark) Financial timeseries prediction Learning a fairytale (rather, its text statistics) Nonlinear control (pendulum, robot drive)

14 x ( t) The Mackey-Glass attractor = 0.2 x( t τ) / (1 + x( t τ) 10 ) 0.1x( t) delay differential equation delay τ > 16.8: chaotic widely used benchmark for time series prediction

15 Learning setup network size 1000 training sequence N = 3000 sampling rate 1

16 Prediction with model visible discrepancy after about 1500 steps

17 Comparison: NRMSE for 84-step prediction -5,1-4,2-1,7-1,7-1,7 log10(nrmse) -1,3-1,3-1,2-1,2 ESN (refined) -5,1 ESN (1+2 K) -4,2 PCR Local Model (McNames 99, 2 K) -1,7 SOM (Vesanto 97, 3K) -1,7 DCS-LLM (Chudy & Farkas 98, 3K) -1,7 AMB (Bersini et al 98,? K) * -1,3 Neural Gaz (Martinez et al 93, ~4K) -1,3 EPNet (Yao & Liu 97, 0.5 K) -1,2 BPNN (Lapedes & Farber 87,? K) * -1,2 (Jaeger & Haas 2004)

18 Event detection for robots (joint work with J.Hertzberg & F. Schönherr) Robot drives through office environment, experiences data streams (27 channels) like... infrared distance sensor left motor speed activation of "gothrudoor" external teacher signal, marking event category 10 sec

19 Learning setup 27 (raw) data channels unlimited number of event 100 unit RNN detector channels simulated robot (rich simulation) training run spans 15 simulated minutes event categories like pass through door pass by 90 corner pass by smooth corner

20 Results easy to train event hypothesis signals "boolean" categories possible single-shot learning possible

21 Japanese vowels: task benchmark data set from UCI KDD repository 1) 9 male Japanese speakers, samples of two-vowel utterance /ae/, represented by 12 LPC cepstrum coefficients task: discriminate speakers (learn on 9 x 30 samples, test on 370 samples in total, unevenly distributed) Figure: exemplary samples from the 9 speakers 1)

22 Japanese vowels: training setup Training No. 2 speaks input: 12 speech signal coefficients 100 unit augmented ESN basic idea: output: 9 speaker indicators No. 9 wins Testing

23 Japanese vowels: results discrimination learning original result (Kudo et al 1999) [5-unit continuous HMM]: 96.2 % correct (14 errors) previous best result (Strickert 2004): [self-organizing "merge neural gas", 1000 neurons]: 98.4 % correct (6 errors) ESN (Jaeger, Lukosevicius and Popovici 2007, refined architecture): zero test errors

24 Financial time series prediction NN3 financial time series forecasting competition, 2007 ( 111 short financial time series, origin withheld intentionally of varied nature Took part with machine learning graduate seminar at Jacobs University ( hjaeger/ TimeSeriesCompetitionOverview.html, using ESNs... and we won -!

25 Little Red Riding Hood training data: 3412-symbol sequence, shown here: first and last 500 once_upon_a_time_there_was_a_little_village_girl,_the_prettiest_ever_seen_her_mot her_doted_upon_her,_and_so_did_her_grandmother._she,_good_woman,_made_for_h er_a_little_red_hood_which_suited_her_so_well,_that_everyone_called_her_little_red _riding_hood._one_day_her_mother,_who_had_just_made_some_cakes,_said_to_her _my_dear,_you_shall_go_and_see_how_your_grandmother_is,_for_i_have_heard_she _is_ailing,_take_her_this_cake_and_this_little_pot_of_butter._little_red_riding_hood_ started_off_at_once_for_he oh,_grandmamma,_grandmamma,_what_great_arms_you_have_got_all_the_better_to _hug_you_with,_my_dear_oh,_grandmamma,_grandmamma,_what_great_legs_you_h ave_got_all_the_better_to_run_with,_my_dear_oh,_grandmamma,_grandmamma,_wh at_great_eyes_you_have_got_all_the_better_to_see_with,_my_dear_oh,_grandmamm a,_grandmamma,_what_great_teeth_you_have_got_all_the_better_to_gobble_you_up _so_saying,_the_wicked_wolf_leaped_on_little_red_riding_hood_and_gobbled_her_u p._here_endeth_the_tale_of_little_red_riding_hood.

26 Network setup in training 29 input channels code symbols 29 output channels for next symbol hypotheses _ a z 400 units

27 Trained network in "text" generation winning symbol is next input......!! decision mechanism, e.g. winner-take-all

28 Results Selection by random draw according to output yth_upsghteyshhfakeofw_io,l_yodoinglle_d_upeiuttytyr_hsymua_doey_sa mmusos_trll,t.krpuflvek_hwiblhooslolyoe,_wtheble_ft_a_gimllveteud_... Winner-take-all selection sdear_oh,_grandmamma,_who_will_go_and_the_wolf_said_the_wolf_said _the_wolf_said_the_wolf_said_the_wolf_said_the_wolf_said_the_wolf...

29 Results, continued Selection by nonlinearly weighted random draw (namely, out^3.5) d_wolf_said_the_better_to_her_the_wood_the_wolf_she urter_that_of_butter_to_he r_grandmother,_the_door_grandmamma,_who_was_bed,_she_the_better_the_wolf_sa_ and_she_little_red_her_grandmatm_aa_grandmother_mother_grandmother_mother_go od_wolf,_and_the_wolf_so_said_the_she_wolf,_and_i_have_gs_at_the_wolf,_wolf_bu tter_to_her_come_neard_the_bobbled_her_grandmamma,_grandmamma,_who_her_the _do_wolf_cake_her_grandmother_mother_to_her_to_she_me_the_better_to_her_the_b ettle_red_riding_hood_see_her_the_pot_of_butter_...

30 Direct / feedback tracking control training exploitation y( t) y( t-d) z -d u( t-d) pendulum pendulum y( t) z -d u( t) yref ( t+d) u( t) network "observes" torque u(t-d), plant state y(t-d), y(t) network input: reference y ref (t+d) and state feedback y(t) network learns how u(t) depends on y(t), y(t+d) network computes control input u(t)

31 Simulated pendulum example training data torque u exploitation control signal u issued by network [ ] disturb. [ ] pendulum coord. y y actual [ ] reference [ ] plus coord. x, plus (in state feedback) actual [ ] reference [ ]

32 Robot motor control: training and using the controller simulated robot wheel speeds Desired robot wheel speeds y( t-2) z -2 y( t) y( t) y ( ref t+2 ) u( t) z -2 u( t) u( t-2) previously trained ESN model of robot training data points (about 5 min robot run) 100 unit network original PWM robot data 1) (left & right) Controller network with output feedback NMSE about ) data courtesy Paul Plöger et al.

33 Results Original controller hand-designed PID (feedback) controller controlled trajectory lags behind target Target Controlled Control input Trained controller run on target generated by reference model Target Controlled Control input

34 5.2 Stable short-term memory

35 Stable short-term memory Motivation. Reservoir memory is transient, but we need stable dynamical STM as well In single-word speech/writing recognition: good to remember letter from beginning of word at end In phrase processing: good to remember initial words / grammatical categories

36 Example and architecture Task: count opening and closing brackets in text Text: random sequence of distorted {, } and 82 distractor symbols, in 12-pixel-high grayscale image Idea: for each opening "{", flip one further output neuron from -0.5 to dim grayscale vector y 1 y 2 # of open { y 1 y neuron reservoir trained fixed

37 Performance Red lines: output values Green lines: targets 0.19 % errors/character with 6 bracket levels (Pascanu & Jaeger 2010) Current Jacobs: mathematics of attractors in input-driven systems (Work and figures by Razvan Pascanu)

38 References Atiya, A. F., Parlos, A. G. (2000), New results on recurrent network training: Unifying the algorithms and accelerating convergence. IEEE Transactions on Neural Networks, 11(3): Funahashi, K.-I., Nakamura, Y. (1993), Approximation of dynamical systems by continuous time recurrent neural networks. Neural Networks 6, 1993, Gers, F.A., Schmidhuber, J., Cummins, F. (2000), Learning to forget: continual prediction with LSTM. Neural Computation 12(10), Hermans, M., Schrauwen, B. (2009): Memory in linear recurrent neural networks in continuous time. Neural Networks Jaeger, H. (2001): The "echo state" approach to analysing and training recurrent neural networks. GMD Report 148, German National Research Center for Information Technology, 2001Jaeger, H. (2001): Short term memory in echo state networks. GMD Report 152, German National Research Center for Information Technology, 2001 H. Jaeger (2006): Generating exponentially many periodic attractors with linearly growing Echo State Networks. IUB technical report Nr. 3 Jaeger, H. (2007), Discovering multiscale dynamical features with hierarchical Echo State Networks. Technical Report 10, School of Engineering and Science, Jacobs University. H., Haas, H. (2004), Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication. Science 304, 2004, Jaeger, H. (2010): Reservoir Self-Control for Achieving Invariance Against Slow Input Distortions. Jacobs University technical report Nr. 23 Jaeger, H., Haas, H. (2004), Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication. Science 304, 2004, Jaeger, H., Lukosevicius, M. Popovici, D. (2007): Optimization and Applications of Echo State Networks with Leaky Integrator Neurons. Neural Networks 20(3), , 2007 Jaeger, H., Maass, W.,Principe, J. (2007), Special issue on echo state networks and liquid state machines. Neural Networks 20(3), 2007, Kudo, M., Toyama, J., Shimbo, M. (1999), Multidimensional Curve Classification Using Passing-Through Regions. Pattern Recognition Letters, Vol. 20, No , pages Lukosecicius, M., Jaeger, H. (2009, to appear): Reservoir Computing Approaches to Recurrent Neural Network Training. Computer Science Review Maass, W., Natschlaeger, T., Markram, H. (2002), Real-Time Computing Without Stable States: A New Framework for Neural Computation Based on Perturbations. Neural Computation 14(11), 2002, Maass, W., Joshi, P., Sontag, E. (2007), Computational aspects of feedback in neural circuits. PLOS Computational Biology 3(1), 2007, 1-20 Pascanu, R. and Jaeger, H. (2010): A Neurodynamical Model for Working Memory. Neural Networks; DOI /j.neunet Strickert, M. (2004). Self-organizing neural networks for sequence processing. Ph.D. thesis. Univ. of Osnabrück, Dpt. of Computer Science. Diss384 thesis.pdf. White, O. L., Lee, D. D., Sompolinsky, H. S. (2004), Short-term Memory in Orthogonal Neural Networks. Phys. Rev. Lett. 92(14), 2004,

International University Bremen Guided Research Proposal Improve on chaotic time series prediction using MLPs for output training

International University Bremen Guided Research Proposal Improve on chaotic time series prediction using MLPs for output training International University Bremen Guided Research Proposal Improve on chaotic time series prediction using MLPs for output training Aakash Jain a.jain@iu-bremen.de Spring Semester 2004 1 Executive Summary

More information

Good vibrations: the issue of optimizing dynamical reservoirs

Good vibrations: the issue of optimizing dynamical reservoirs Good vibrations: the issue of optimizing dynamical reservoirs Workshop on ESNs / LSMs, NIPS 2006 Herbert Jaeger International University Bremen (Jacobs University Bremen, as of Spring 2007) The basic idea:

More information

Harnessing Nonlinearity: Predicting Chaotic Systems and Saving

Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication Publishde in Science Magazine, 2004 Siamak Saliminejad Overview Eco State Networks How to build ESNs Chaotic

More information

Reservoir Computing and Echo State Networks

Reservoir Computing and Echo State Networks An Introduction to: Reservoir Computing and Echo State Networks Claudio Gallicchio gallicch@di.unipi.it Outline Focus: Supervised learning in domain of sequences Recurrent Neural networks for supervised

More information

Using reservoir computing in a decomposition approach for time series prediction.

Using reservoir computing in a decomposition approach for time series prediction. Using reservoir computing in a decomposition approach for time series prediction. Francis wyffels, Benjamin Schrauwen and Dirk Stroobandt Ghent University - Electronics and Information Systems Department

More information

Reservoir Computing with Stochastic Bitstream Neurons

Reservoir Computing with Stochastic Bitstream Neurons Reservoir Computing with Stochastic Bitstream Neurons David Verstraeten, Benjamin Schrauwen and Dirk Stroobandt Department of Electronics and Information Systems (ELIS), Ugent {david.verstraeten, benjamin.schrauwen,

More information

Negatively Correlated Echo State Networks

Negatively Correlated Echo State Networks Negatively Correlated Echo State Networks Ali Rodan and Peter Tiňo School of Computer Science, The University of Birmingham Birmingham B15 2TT, United Kingdom E-mail: {a.a.rodan, P.Tino}@cs.bham.ac.uk

More information

Short Term Memory Quantifications in Input-Driven Linear Dynamical Systems

Short Term Memory Quantifications in Input-Driven Linear Dynamical Systems Short Term Memory Quantifications in Input-Driven Linear Dynamical Systems Peter Tiňo and Ali Rodan School of Computer Science, The University of Birmingham Birmingham B15 2TT, United Kingdom E-mail: {P.Tino,

More information

Modelling Time Series with Neural Networks. Volker Tresp Summer 2017

Modelling Time Series with Neural Networks. Volker Tresp Summer 2017 Modelling Time Series with Neural Networks Volker Tresp Summer 2017 1 Modelling of Time Series The next figure shows a time series (DAX) Other interesting time-series: energy prize, energy consumption,

More information

Memory Capacity of Input-Driven Echo State NetworksattheEdgeofChaos

Memory Capacity of Input-Driven Echo State NetworksattheEdgeofChaos Memory Capacity of Input-Driven Echo State NetworksattheEdgeofChaos Peter Barančok and Igor Farkaš Faculty of Mathematics, Physics and Informatics Comenius University in Bratislava, Slovakia farkas@fmph.uniba.sk

More information

Lecture 5: Recurrent Neural Networks

Lecture 5: Recurrent Neural Networks 1/25 Lecture 5: Recurrent Neural Networks Nima Mohajerin University of Waterloo WAVE Lab nima.mohajerin@uwaterloo.ca July 4, 2017 2/25 Overview 1 Recap 2 RNN Architectures for Learning Long Term Dependencies

More information

Refutation of Second Reviewer's Objections

Refutation of Second Reviewer's Objections Re: Submission to Science, "Harnessing nonlinearity: predicting chaotic systems and boosting wireless communication." (Ref: 1091277) Refutation of Second Reviewer's Objections Herbert Jaeger, Dec. 23,

More information

REAL-TIME COMPUTING WITHOUT STABLE

REAL-TIME COMPUTING WITHOUT STABLE REAL-TIME COMPUTING WITHOUT STABLE STATES: A NEW FRAMEWORK FOR NEURAL COMPUTATION BASED ON PERTURBATIONS Wolfgang Maass Thomas Natschlager Henry Markram Presented by Qiong Zhao April 28 th, 2010 OUTLINE

More information

Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions

Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions Artem Chernodub, Institute of Mathematical Machines and Systems NASU, Neurotechnologies

More information

Echo State Networks with Filter Neurons and a Delay&Sum Readout

Echo State Networks with Filter Neurons and a Delay&Sum Readout Echo State Networks with Filter Neurons and a Delay&Sum Readout Georg Holzmann 2,1 (Corresponding Author) http://grh.mur.at grh@mur.at Helmut Hauser 1 helmut.hauser@igi.tugraz.at 1 Institute for Theoretical

More information

From Neural Networks To Reservoir Computing...

From Neural Networks To Reservoir Computing... From Neural Networks To Reservoir Computing... An Introduction Naveen Kuppuswamy, PhD Candidate, A.I Lab, University of Zurich 1 Disclaimer : I am only an egg 2 Part I From Classical Computation to ANNs

More information

arxiv: v1 [cs.lg] 2 Feb 2018

arxiv: v1 [cs.lg] 2 Feb 2018 Short-term Memory of Deep RNN Claudio Gallicchio arxiv:1802.00748v1 [cs.lg] 2 Feb 2018 Department of Computer Science, University of Pisa Largo Bruno Pontecorvo 3-56127 Pisa, Italy Abstract. The extension

More information

Towards a Calculus of Echo State Networks arxiv: v1 [cs.ne] 1 Sep 2014

Towards a Calculus of Echo State Networks arxiv: v1 [cs.ne] 1 Sep 2014 his space is reserved for the Procedia header, do not use it owards a Calculus of Echo State Networks arxiv:49.28v [cs.ne] Sep 24 Alireza Goudarzi and Darko Stefanovic,2 Department of Computer Science,

More information

Deep Recurrent Neural Networks

Deep Recurrent Neural Networks Deep Recurrent Neural Networks Artem Chernodub e-mail: a.chernodub@gmail.com web: http://zzphoto.me ZZ Photo IMMSP NASU 2 / 28 Neuroscience Biological-inspired models Machine Learning p x y = p y x p(x)/p(y)

More information

Artificial Neural Networks

Artificial Neural Networks Artificial Neural Networks 鮑興國 Ph.D. National Taiwan University of Science and Technology Outline Perceptrons Gradient descent Multi-layer networks Backpropagation Hidden layer representations Examples

More information

Lecture 11 Recurrent Neural Networks I

Lecture 11 Recurrent Neural Networks I Lecture 11 Recurrent Neural Networks I CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 01, 2017 Introduction Sequence Learning with Neural Networks Some Sequence Tasks

More information

Artificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen

Artificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen Artificial Neural Networks Introduction to Computational Neuroscience Tambet Matiisen 2.04.2018 Artificial neural network NB! Inspired by biology, not based on biology! Applications Automatic speech recognition

More information

Short Term Memory and Pattern Matching with Simple Echo State Networks

Short Term Memory and Pattern Matching with Simple Echo State Networks Short Term Memory and Pattern Matching with Simple Echo State Networks Georg Fette (fette@in.tum.de), Julian Eggert (julian.eggert@honda-ri.de) Technische Universität München; Boltzmannstr. 3, 85748 Garching/München,

More information

Recurrent Neural Networks

Recurrent Neural Networks Recurrent Neural Networks Datamining Seminar Kaspar Märtens Karl-Oskar Masing Today's Topics Modeling sequences: a brief overview Training RNNs with back propagation A toy example of training an RNN Why

More information

Reservoir Computing in Forecasting Financial Markets

Reservoir Computing in Forecasting Financial Markets April 9, 2015 Reservoir Computing in Forecasting Financial Markets Jenny Su Committee Members: Professor Daniel Gauthier, Adviser Professor Kate Scholberg Professor Joshua Socolar Defense held on Wednesday,

More information

Lecture 11 Recurrent Neural Networks I

Lecture 11 Recurrent Neural Networks I Lecture 11 Recurrent Neural Networks I CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor niversity of Chicago May 01, 2017 Introduction Sequence Learning with Neural Networks Some Sequence Tasks

More information

Liquid Computing. Wolfgang Maass

Liquid Computing. Wolfgang Maass Liquid Computing Wolfgang Maass Institute for Theoretical Computer Science Technische Universitaet Graz A-8010 Graz, Austria maass@igi.tugraz.at http://www.igi.tugraz.at/ Abstract. This review addresses

More information

At the Edge of Chaos: Real-time Computations and Self-Organized Criticality in Recurrent Neural Networks

At the Edge of Chaos: Real-time Computations and Self-Organized Criticality in Recurrent Neural Networks At the Edge of Chaos: Real-time Computations and Self-Organized Criticality in Recurrent Neural Networks Thomas Natschläger Software Competence Center Hagenberg A-4232 Hagenberg, Austria Thomas.Natschlaeger@scch.at

More information

Sequence Modeling with Neural Networks

Sequence Modeling with Neural Networks Sequence Modeling with Neural Networks Harini Suresh y 0 y 1 y 2 s 0 s 1 s 2... x 0 x 1 x 2 hat is a sequence? This morning I took the dog for a walk. sentence medical signals speech waveform Successes

More information

MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION

MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION Computing and Informatics, Vol. 30, 2011, 321 334 MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION Štefan Babinec, Jiří Pospíchal Department of Mathematics Faculty of Chemical and Food Technology

More information

Recurrence Enhances the Spatial Encoding of Static Inputs in Reservoir Networks

Recurrence Enhances the Spatial Encoding of Static Inputs in Reservoir Networks Recurrence Enhances the Spatial Encoding of Static Inputs in Reservoir Networks Christian Emmerich, R. Felix Reinhart, and Jochen J. Steil Research Institute for Cognition and Robotics (CoR-Lab), Bielefeld

More information

Long-Short Term Memory and Other Gated RNNs

Long-Short Term Memory and Other Gated RNNs Long-Short Term Memory and Other Gated RNNs Sargur Srihari srihari@buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics in Sequence Modeling

More information

Advanced Methods for Recurrent Neural Networks Design

Advanced Methods for Recurrent Neural Networks Design Universidad Autónoma de Madrid Escuela Politécnica Superior Departamento de Ingeniería Informática Advanced Methods for Recurrent Neural Networks Design Master s thesis presented to apply for the Master

More information

From Neural Networks To Reservoir Computing...

From Neural Networks To Reservoir Computing... From Neural Networks To Reservoir Computing... An Introduction Naveen Kuppuswamy, PhD Candidate, A.I Lab, University of Zurich Disclaimer : I am only an egg 2 Part I From Classical Computation to ANNs

More information

Edge of Chaos Computation in Mixed-Mode VLSI - A Hard Liquid

Edge of Chaos Computation in Mixed-Mode VLSI - A Hard Liquid Edge of Chaos Computation in Mixed-Mode VLSI - A Hard Liquid Felix Schürmann, Karlheinz Meier, Johannes Schemmel Kirchhoff Institute for Physics University of Heidelberg Im Neuenheimer Feld 227, 6912 Heidelberg,

More information

Sample Exam COMP 9444 NEURAL NETWORKS Solutions

Sample Exam COMP 9444 NEURAL NETWORKS Solutions FAMILY NAME OTHER NAMES STUDENT ID SIGNATURE Sample Exam COMP 9444 NEURAL NETWORKS Solutions (1) TIME ALLOWED 3 HOURS (2) TOTAL NUMBER OF QUESTIONS 12 (3) STUDENTS SHOULD ANSWER ALL QUESTIONS (4) QUESTIONS

More information

On the use of Long-Short Term Memory neural networks for time series prediction

On the use of Long-Short Term Memory neural networks for time series prediction On the use of Long-Short Term Memory neural networks for time series prediction Pilar Gómez-Gil National Institute of Astrophysics, Optics and Electronics ccc.inaoep.mx/~pgomez In collaboration with: J.

More information

Analyzing the weight dynamics of recurrent learning algorithms

Analyzing the weight dynamics of recurrent learning algorithms Analyzing the weight dynamics of recurrent learning algorithms Ulf D. Schiller and Jochen J. Steil Neuroinformatics Group, Faculty of Technology, Bielefeld University Abstract We provide insights into

More information

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau Last update: October 26, 207 Neural networks CMSC 42: Section 8.7 Dana Nau Outline Applications of neural networks Brains Neural network units Perceptrons Multilayer perceptrons 2 Example Applications

More information

Neural Networks. Volker Tresp Summer 2015

Neural Networks. Volker Tresp Summer 2015 Neural Networks Volker Tresp Summer 2015 1 Introduction The performance of a classifier or a regression model critically depends on the choice of appropriate basis functions The problem with generic basis

More information

A Model for Real-Time Computation in Generic Neural Microcircuits

A Model for Real-Time Computation in Generic Neural Microcircuits A Model for Real-Time Computation in Generic Neural Microcircuits Wolfgang Maass, Thomas Natschläger Institute for Theoretical Computer Science Technische Universitaet Graz A-81 Graz, Austria maass, tnatschl

More information

Grundlagen der Künstlichen Intelligenz

Grundlagen der Künstlichen Intelligenz Grundlagen der Künstlichen Intelligenz Neural networks Daniel Hennes 21.01.2018 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Logistic regression Neural networks Perceptron

More information

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28 1 / 28 Neural Networks Mark van Rossum School of Informatics, University of Edinburgh January 15, 2018 2 / 28 Goals: Understand how (recurrent) networks behave Find a way to teach networks to do a certain

More information

arxiv: v3 [cs.lg] 14 Jan 2018

arxiv: v3 [cs.lg] 14 Jan 2018 A Gentle Tutorial of Recurrent Neural Network with Error Backpropagation Gang Chen Department of Computer Science and Engineering, SUNY at Buffalo arxiv:1610.02583v3 [cs.lg] 14 Jan 2018 1 abstract We describe

More information

An overview of reservoir computing: theory, applications and implementations

An overview of reservoir computing: theory, applications and implementations An overview of reservoir computing: theory, applications and implementations Benjamin Schrauwen David Verstraeten Jan Van Campenhout Electronics and Information Systems Department, Ghent University, Belgium

More information

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks INFOB2KI 2017-2018 Utrecht University The Netherlands ARTIFICIAL INTELLIGENCE Artificial Neural Networks Lecturer: Silja Renooij These slides are part of the INFOB2KI Course Notes available from www.cs.uu.nl/docs/vakken/b2ki/schema.html

More information

Christian Mohr

Christian Mohr Christian Mohr 20.12.2011 Recurrent Networks Networks in which units may have connections to units in the same or preceding layers Also connections to the unit itself possible Already covered: Hopfield

More information

Fisher Memory of Linear Wigner Echo State Networks

Fisher Memory of Linear Wigner Echo State Networks ESA 07 proceedings, European Symposium on Artificial eural etworks, Computational Intelligence and Machine Learning. Bruges (Belgium), 6-8 April 07, i6doc.com publ., ISB 978-87587039-. Fisher Memory of

More information

Neural networks. Chapter 19, Sections 1 5 1

Neural networks. Chapter 19, Sections 1 5 1 Neural networks Chapter 19, Sections 1 5 Chapter 19, Sections 1 5 1 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 19, Sections 1 5 2 Brains 10

More information

Neural networks. Chapter 20, Section 5 1

Neural networks. Chapter 20, Section 5 1 Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of

More information

A First Attempt of Reservoir Pruning for Classification Problems

A First Attempt of Reservoir Pruning for Classification Problems A First Attempt of Reservoir Pruning for Classification Problems Xavier Dutoit, Hendrik Van Brussel, Marnix Nuttin Katholieke Universiteit Leuven - P.M.A. Celestijnenlaan 300b, 3000 Leuven - Belgium Abstract.

More information

Initialization and Self-Organized Optimization of Recurrent Neural Network Connectivity

Initialization and Self-Organized Optimization of Recurrent Neural Network Connectivity Initialization and Self-Organized Optimization of Recurrent Neural Network Connectivity Joschka Boedecker Dep. of Adaptive Machine Systems, Osaka University, Suita, Osaka, Japan Oliver Obst CSIRO ICT Centre,

More information

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA   1/ 21 Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural

More information

Controlling Recurrent Neural Networks by Conceptors

Controlling Recurrent Neural Networks by Conceptors arxiv:143.3369v3 [cs.ne] 22 Apr 217 Herbert Jaeger Controlling Recurrent Neural Networks by Conceptors Technical Report No. 31 March 214 School of Engineering and Science Controlling Recurrent Neural Networks

More information

Machine Learning. Neural Networks

Machine Learning. Neural Networks Machine Learning Neural Networks Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 Biological Analogy Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 THE

More information

Artificial Neural Networks

Artificial Neural Networks Artificial Neural Networks Threshold units Gradient descent Multilayer networks Backpropagation Hidden layer representations Example: Face Recognition Advanced topics 1 Connectionist Models Consider humans:

More information

Reservoir Computing Methods for Prognostics and Health Management (PHM) Piero Baraldi Energy Department Politecnico di Milano Italy

Reservoir Computing Methods for Prognostics and Health Management (PHM) Piero Baraldi Energy Department Politecnico di Milano Italy Reservoir Computing Methods for Prognostics and Health Management (PHM) Piero Baraldi Energy Department Politecnico di Milano Italy 2 Piero Baraldi Data Industry 4.0 2 Digitalization 2.8 Trillion GD (ZD)

More information

Hopfield Networks. (Excerpt from a Basic Course at IK 2008) Herbert Jaeger. Jacobs University Bremen

Hopfield Networks. (Excerpt from a Basic Course at IK 2008) Herbert Jaeger. Jacobs University Bremen Hopfield Networks (Excerpt from a Basic Course at IK 2008) Herbert Jaeger Jacobs University Bremen Building a model of associative memory should be simple enough... Our brain is a neural network Individual

More information

Deep Learning Recurrent Networks 2/28/2018

Deep Learning Recurrent Networks 2/28/2018 Deep Learning Recurrent Networks /8/8 Recap: Recurrent networks can be incredibly effective Story so far Y(t+) Stock vector X(t) X(t+) X(t+) X(t+) X(t+) X(t+5) X(t+) X(t+7) Iterated structures are good

More information

Liquid Computing. Wolfgang Maass. Institut für Grundlagen der Informationsverarbeitung Technische Universität Graz, Austria

Liquid Computing. Wolfgang Maass. Institut für Grundlagen der Informationsverarbeitung Technische Universität Graz, Austria NNB SS10 1 Liquid Computing Wolfgang Maass Institut für Grundlagen der Informationsverarbeitung Technische Universität Graz, Austria Institute for Theoretical Computer Science http://www.igi.tugraz.at/maass/

More information

Neural Nets and Symbolic Reasoning Hopfield Networks

Neural Nets and Symbolic Reasoning Hopfield Networks Neural Nets and Symbolic Reasoning Hopfield Networks Outline The idea of pattern completion The fast dynamics of Hopfield networks Learning with Hopfield networks Emerging properties of Hopfield networks

More information

input unit N internal units output unit

input unit N internal units output unit Adaptive nonlinear system identification with echo state networks Herbert Jaeger Fraunhofer Institute for Autonomous Intelligent Systems AIS.INDY Schloss Birlinghoven, D-53754 Sankt Augustin, Germany herbert.jaeger@ais.fraunhofer.de

More information

arxiv: v1 [cs.lg] 8 Jul 2017

arxiv: v1 [cs.lg] 8 Jul 2017 Tailoring Artificial Neural Networks for Optimal Learning Pau Vilimelis Aceituno 1,2, Gang Yan 3, Yang-Yu Liu 1,4 1 Channing Division of Network Medicine, Brigham and Women s Hospital, Harvard Medical

More information

Several ways to solve the MSO problem

Several ways to solve the MSO problem Several ways to solve the MSO problem J. J. Steil - Bielefeld University - Neuroinformatics Group P.O.-Box 0 0 3, D-3350 Bielefeld - Germany Abstract. The so called MSO-problem, a simple superposition

More information

CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning

CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning Lei Lei Ruoxuan Xiong December 16, 2017 1 Introduction Deep Neural Network

More information

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, 2017 Spis treści Website Acknowledgments Notation xiii xv xix 1 Introduction 1 1.1 Who Should Read This Book?

More information

Long-Short Term Memory

Long-Short Term Memory Long-Short Term Memory Sepp Hochreiter, Jürgen Schmidhuber Presented by Derek Jones Table of Contents 1. Introduction 2. Previous Work 3. Issues in Learning Long-Term Dependencies 4. Constant Error Flow

More information

Lecture 17: Neural Networks and Deep Learning

Lecture 17: Neural Networks and Deep Learning UVA CS 6316 / CS 4501-004 Machine Learning Fall 2016 Lecture 17: Neural Networks and Deep Learning Jack Lanchantin Dr. Yanjun Qi 1 Neurons 1-Layer Neural Network Multi-layer Neural Network Loss Functions

More information

Optoelectronic Systems Trained With Backpropagation Through Time

Optoelectronic Systems Trained With Backpropagation Through Time IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 26, NO. 7, JULY 2015 1545 Optoelectronic Systems Trained With Backpropagation Through Time Michiel Hermans, Joni Dambre, and Peter Bienstman

More information

A New Look at Nonlinear Time Series Prediction with NARX Recurrent Neural Network. José Maria P. Menezes Jr. and Guilherme A.

A New Look at Nonlinear Time Series Prediction with NARX Recurrent Neural Network. José Maria P. Menezes Jr. and Guilherme A. A New Look at Nonlinear Time Series Prediction with NARX Recurrent Neural Network José Maria P. Menezes Jr. and Guilherme A. Barreto Department of Teleinformatics Engineering Federal University of Ceará,

More information

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann (Feed-Forward) Neural Networks 2016-12-06 Dr. Hajira Jabeen, Prof. Jens Lehmann Outline In the previous lectures we have learned about tensors and factorization methods. RESCAL is a bilinear model for

More information

Neural Network Control of Robot Manipulators and Nonlinear Systems

Neural Network Control of Robot Manipulators and Nonlinear Systems Neural Network Control of Robot Manipulators and Nonlinear Systems F.L. LEWIS Automation and Robotics Research Institute The University of Texas at Arlington S. JAG ANNATHAN Systems and Controls Research

More information

Course 395: Machine Learning - Lectures

Course 395: Machine Learning - Lectures Course 395: Machine Learning - Lectures Lecture 1-2: Concept Learning (M. Pantic) Lecture 3-4: Decision Trees & CBC Intro (M. Pantic & S. Petridis) Lecture 5-6: Evaluating Hypotheses (S. Petridis) Lecture

More information

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6 Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)

More information

Using a Hopfield Network: A Nuts and Bolts Approach

Using a Hopfield Network: A Nuts and Bolts Approach Using a Hopfield Network: A Nuts and Bolts Approach November 4, 2013 Gershon Wolfe, Ph.D. Hopfield Model as Applied to Classification Hopfield network Training the network Updating nodes Sequencing of

More information

Neural Networks 2. 2 Receptive fields and dealing with image inputs

Neural Networks 2. 2 Receptive fields and dealing with image inputs CS 446 Machine Learning Fall 2016 Oct 04, 2016 Neural Networks 2 Professor: Dan Roth Scribe: C. Cheng, C. Cervantes Overview Convolutional Neural Networks Recurrent Neural Networks 1 Introduction There

More information

Towards logics for neural conceptors

Towards logics for neural conceptors Towards logics for neural conceptors Till Mossakowski joint work with Razvan Diaconescu and Martin Glauer Otto-von-Guericke-Universität Magdeburg AITP 2018, Aussois, March 30, 2018 Overview 1 Conceptors

More information

Multimodal context analysis and prediction

Multimodal context analysis and prediction Multimodal context analysis and prediction Valeria Tomaselli (valeria.tomaselli@st.com) Sebastiano Battiato Giovanni Maria Farinella Tiziana Rotondo (PhD student) Outline 2 Context analysis vs prediction

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement

More information

100 inference steps doesn't seem like enough. Many neuron-like threshold switching units. Many weighted interconnections among units

100 inference steps doesn't seem like enough. Many neuron-like threshold switching units. Many weighted interconnections among units Connectionist Models Consider humans: Neuron switching time ~ :001 second Number of neurons ~ 10 10 Connections per neuron ~ 10 4 5 Scene recognition time ~ :1 second 100 inference steps doesn't seem like

More information

Structured reservoir computing with spatiotemporal chaotic attractors

Structured reservoir computing with spatiotemporal chaotic attractors Structured reservoir computing with spatiotemporal chaotic attractors Carlos Lourenço 1,2 1- Faculty of Sciences of the University of Lisbon - Informatics Department Campo Grande, 1749-016 Lisboa - Portugal

More information

Deep Learning. Recurrent Neural Network (RNNs) Ali Ghodsi. October 23, Slides are partially based on Book in preparation, Deep Learning

Deep Learning. Recurrent Neural Network (RNNs) Ali Ghodsi. October 23, Slides are partially based on Book in preparation, Deep Learning Recurrent Neural Network (RNNs) University of Waterloo October 23, 2015 Slides are partially based on Book in preparation, by Bengio, Goodfellow, and Aaron Courville, 2015 Sequential data Recurrent neural

More information

Recurrent neural networks

Recurrent neural networks 12-1: Recurrent neural networks Prof. J.C. Kao, UCLA Recurrent neural networks Motivation Network unrollwing Backpropagation through time Vanishing and exploding gradients LSTMs GRUs 12-2: Recurrent neural

More information

Artificial Neural Networks Examination, June 2004

Artificial Neural Networks Examination, June 2004 Artificial Neural Networks Examination, June 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum

More information

Analysis of Multilayer Neural Network Modeling and Long Short-Term Memory

Analysis of Multilayer Neural Network Modeling and Long Short-Term Memory Analysis of Multilayer Neural Network Modeling and Long Short-Term Memory Danilo López, Nelson Vera, Luis Pedraza International Science Index, Mathematical and Computational Sciences waset.org/publication/10006216

More information

ARCHITECTURAL DESIGNS OF ECHO STATE NETWORK

ARCHITECTURAL DESIGNS OF ECHO STATE NETWORK ARCHITECTURAL DESIGNS OF ECHO STATE NETWORK by ALI ABDALLAH ALI AL RODAN A thesis submitted to The University of Birmingham for the degree of DOCTOR OF PHILOSOPHY School of Computer Science College of

More information

Lecture 5 Neural models for NLP

Lecture 5 Neural models for NLP CS546: Machine Learning in NLP (Spring 2018) http://courses.engr.illinois.edu/cs546/ Lecture 5 Neural models for NLP Julia Hockenmaier juliahmr@illinois.edu 3324 Siebel Center Office hours: Tue/Thu 2pm-3pm

More information

Chapter 1. Liquid State Machines: Motivation, Theory, and Applications

Chapter 1. Liquid State Machines: Motivation, Theory, and Applications Chapter 1 Liquid State Machines: Motivation, Theory, and Applications Wolfgang Maass Institute for Theoretical Computer Science Graz University of Technology A-8010 Graz, Austria maass@igi.tugraz.at The

More information

Development of a Deep Recurrent Neural Network Controller for Flight Applications

Development of a Deep Recurrent Neural Network Controller for Flight Applications Development of a Deep Recurrent Neural Network Controller for Flight Applications American Control Conference (ACC) May 26, 2017 Scott A. Nivison Pramod P. Khargonekar Department of Electrical and Computer

More information

An Introductory Course in Computational Neuroscience

An Introductory Course in Computational Neuroscience An Introductory Course in Computational Neuroscience Contents Series Foreword Acknowledgments Preface 1 Preliminary Material 1.1. Introduction 1.1.1 The Cell, the Circuit, and the Brain 1.1.2 Physics of

More information

Simple Deterministically Constructed Cycle Reservoirs with Regular Jumps

Simple Deterministically Constructed Cycle Reservoirs with Regular Jumps 1 Simple Deterministically Constructed Cycle Reservoirs with Regular Jumps Ali Rodan 1 and Peter Tiňo 1 1 School of computer Science, The University of Birmingham, Birmingham B15 2TT, United Kingdom, (email:

More information

Tutorial on Machine Learning for Advanced Electronics

Tutorial on Machine Learning for Advanced Electronics Tutorial on Machine Learning for Advanced Electronics Maxim Raginsky March 2017 Part I (Some) Theory and Principles Machine Learning: estimation of dependencies from empirical data (V. Vapnik) enabling

More information

Introduction to RNNs!

Introduction to RNNs! Introduction to RNNs Arun Mallya Best viewed with Computer Modern fonts installed Outline Why Recurrent Neural Networks (RNNs)? The Vanilla RNN unit The RNN forward pass Backpropagation refresher The RNN

More information

1 What makes a dynamical system computationally powerful?

1 What makes a dynamical system computationally powerful? 1 What makes a dynamical system computationally powerful? Robert Legenstein and Wolfgang Maass Institute for Theoretical Computer Science Technische Universität Graz Austria, 8010 Graz {legi, maass}@igi.tugraz.at

More information

Explorations in Echo State Networks

Explorations in Echo State Networks Explorations in Echo State Networks Adrian Millea Master s Thesis Supervised by Dr. Marco Wiering (Department of Artificial Intelligence, University of Groningen, Groningen, The Netherlands) Prof. Mark

More information

arxiv: v1 [cs.lg] 9 Mar 2018

arxiv: v1 [cs.lg] 9 Mar 2018 Hybrid Forecasting of Chaotic Processes: Using Machine Learning in Conjunction with a Knowledge-Based Model Jaideep Pathak, 1 Alexander Wikner, 2 Rebeckah Fussell, 3 Sarthak Chandra, 1 Brian R. Hunt, 1

More information

Artificial Neural Networks Examination, June 2005

Artificial Neural Networks Examination, June 2005 Artificial Neural Networks Examination, June 2005 Instructions There are SIXTY questions. (The pass mark is 30 out of 60). For each question, please select a maximum of ONE of the given answers (either

More information

Introduction to Neural Networks

Introduction to Neural Networks Introduction to Neural Networks What are (Artificial) Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning

More information

Learning and Memory in Neural Networks

Learning and Memory in Neural Networks Learning and Memory in Neural Networks Guy Billings, Neuroinformatics Doctoral Training Centre, The School of Informatics, The University of Edinburgh, UK. Neural networks consist of computational units

More information

Analyzing the State Space Property of Echo

Analyzing the State Space Property of Echo Proceedings of International Joint Conference on Neural Networks, Montreal, Canada, July 31 August 4, 25 Analyzing the State Space Property of Echo State Networks for Chaotic System Prediction AbstractFor

More information