A quick introduction to reservoir computing
|
|
- Lee Doyle
- 6 years ago
- Views:
Transcription
1 A quick introduction to reservoir computing Herbert Jaeger Jacobs University Bremen
2 1 Recurrent neural networks
3 Feedforward and recurrent ANNs A. feedforward B. recurrent Characteristics: Has at least one feedback loop of connections Thus can maintain dynamic activation even without input input time series output time series can approximate any dynamical system (universal approximation property, Funahashi & Nakamura 1993) mathematical analysis difficult learning algorithms computationally expensive and difficult to master few application-oriented publications, little research
4 Hopfield networks, Boltzmann machines, Deep Belief Networks, ART networks, etc. These are also recurrent. Differences: Symmetric connections Energy-based models with equilibrium / stochastic sampling dynamics For pattern recognition and associative memory tasks, -- not for time series processing (only if forced to...)
5 Problems with "classical" gradientbased training methods for RNNs Besides the generic difficulties of gradient-descent methods, in RNNs there are some more: passing through bifurcations error gradient information shrinks exponentially over time long memory effects difficult to train ("long short-term memory" (LSTM) RNNs are an exception, Gers et al 2000) algorithms and maths are difficult experienced professionals needed But: powerful in the hands of "deep learning" experts who have learnt how to tame them
6 2 Echo State Networks Basics
7 Echo state networks Developed independently and simultaneously with Liquid State Machines (LSM) (Wolfgang Maass, TU Graz) Basic idea of ESNs and LSMs: Use a large, random, fixed RNN (called dynamical reservoir, is not changed by training!) Train only the connections from reservoir to output units! traditional, e.g. BPTT: all connections are trained. ESN training: train only output weights
8 Illustrating the principle: training a tone generator Goal: train a network to work as a tuneable tone generator input: frequency setting output: sines of desired frequency
9 Phase 1: internal state collection Drive fixed "reservoir" network with teacher input and output, and save reservoir units time signals. Observation: internal states of dynamical reservoir reflect and modify both input and output teacher signals
10 Phase 2 (and finish): compute output weights Determine reservoir-to-output weights such that training output is optimally reconstituted from internal "echo" signals in a least mean square error sense. This is a linear regression on the teacher output signal (red) on the reservoir signals.
11 Tone generator: exploitation With new output weights in place, drive trained network with input. Observation: network continues to function as in training. internal states reflect input and output output is reconstituted from internal states internal states and output constitute each other
12 Dynamical reservoir large recurrent network ( units) works as "dynamical reservoir" output units combine different internal dynamics into desired dynamics i n p u t u n i t r e c u r r e n t " d y n a m i c a l r e s e r v o i r " o u t p u t u n i t s
13 3 Basic Examples Predicting chaotic timeseries Event detection for robots Speaker identification ("Japanese Vouwel" benchmark) Financial timeseries prediction Learning a fairytale (rather, its text statistics) Nonlinear control (pendulum, robot drive)
14 x ( t) The Mackey-Glass attractor = 0.2 x( t τ) / (1 + x( t τ) 10 ) 0.1x( t) delay differential equation delay τ > 16.8: chaotic widely used benchmark for time series prediction
15 Learning setup network size 1000 training sequence N = 3000 sampling rate 1
16 Prediction with model visible discrepancy after about 1500 steps
17 Comparison: NRMSE for 84-step prediction -5,1-4,2-1,7-1,7-1,7 log10(nrmse) -1,3-1,3-1,2-1,2 ESN (refined) -5,1 ESN (1+2 K) -4,2 PCR Local Model (McNames 99, 2 K) -1,7 SOM (Vesanto 97, 3K) -1,7 DCS-LLM (Chudy & Farkas 98, 3K) -1,7 AMB (Bersini et al 98,? K) * -1,3 Neural Gaz (Martinez et al 93, ~4K) -1,3 EPNet (Yao & Liu 97, 0.5 K) -1,2 BPNN (Lapedes & Farber 87,? K) * -1,2 (Jaeger & Haas 2004)
18 Event detection for robots (joint work with J.Hertzberg & F. Schönherr) Robot drives through office environment, experiences data streams (27 channels) like... infrared distance sensor left motor speed activation of "gothrudoor" external teacher signal, marking event category 10 sec
19 Learning setup 27 (raw) data channels unlimited number of event 100 unit RNN detector channels simulated robot (rich simulation) training run spans 15 simulated minutes event categories like pass through door pass by 90 corner pass by smooth corner
20 Results easy to train event hypothesis signals "boolean" categories possible single-shot learning possible
21 Japanese vowels: task benchmark data set from UCI KDD repository 1) 9 male Japanese speakers, samples of two-vowel utterance /ae/, represented by 12 LPC cepstrum coefficients task: discriminate speakers (learn on 9 x 30 samples, test on 370 samples in total, unevenly distributed) Figure: exemplary samples from the 9 speakers 1)
22 Japanese vowels: training setup Training No. 2 speaks input: 12 speech signal coefficients 100 unit augmented ESN basic idea: output: 9 speaker indicators No. 9 wins Testing
23 Japanese vowels: results discrimination learning original result (Kudo et al 1999) [5-unit continuous HMM]: 96.2 % correct (14 errors) previous best result (Strickert 2004): [self-organizing "merge neural gas", 1000 neurons]: 98.4 % correct (6 errors) ESN (Jaeger, Lukosevicius and Popovici 2007, refined architecture): zero test errors
24 Financial time series prediction NN3 financial time series forecasting competition, 2007 ( 111 short financial time series, origin withheld intentionally of varied nature Took part with machine learning graduate seminar at Jacobs University ( hjaeger/ TimeSeriesCompetitionOverview.html, using ESNs... and we won -!
25 Little Red Riding Hood training data: 3412-symbol sequence, shown here: first and last 500 once_upon_a_time_there_was_a_little_village_girl,_the_prettiest_ever_seen_her_mot her_doted_upon_her,_and_so_did_her_grandmother._she,_good_woman,_made_for_h er_a_little_red_hood_which_suited_her_so_well,_that_everyone_called_her_little_red _riding_hood._one_day_her_mother,_who_had_just_made_some_cakes,_said_to_her _my_dear,_you_shall_go_and_see_how_your_grandmother_is,_for_i_have_heard_she _is_ailing,_take_her_this_cake_and_this_little_pot_of_butter._little_red_riding_hood_ started_off_at_once_for_he oh,_grandmamma,_grandmamma,_what_great_arms_you_have_got_all_the_better_to _hug_you_with,_my_dear_oh,_grandmamma,_grandmamma,_what_great_legs_you_h ave_got_all_the_better_to_run_with,_my_dear_oh,_grandmamma,_grandmamma,_wh at_great_eyes_you_have_got_all_the_better_to_see_with,_my_dear_oh,_grandmamm a,_grandmamma,_what_great_teeth_you_have_got_all_the_better_to_gobble_you_up _so_saying,_the_wicked_wolf_leaped_on_little_red_riding_hood_and_gobbled_her_u p._here_endeth_the_tale_of_little_red_riding_hood.
26 Network setup in training 29 input channels code symbols 29 output channels for next symbol hypotheses _ a z 400 units
27 Trained network in "text" generation winning symbol is next input......!! decision mechanism, e.g. winner-take-all
28 Results Selection by random draw according to output yth_upsghteyshhfakeofw_io,l_yodoinglle_d_upeiuttytyr_hsymua_doey_sa mmusos_trll,t.krpuflvek_hwiblhooslolyoe,_wtheble_ft_a_gimllveteud_... Winner-take-all selection sdear_oh,_grandmamma,_who_will_go_and_the_wolf_said_the_wolf_said _the_wolf_said_the_wolf_said_the_wolf_said_the_wolf_said_the_wolf...
29 Results, continued Selection by nonlinearly weighted random draw (namely, out^3.5) d_wolf_said_the_better_to_her_the_wood_the_wolf_she urter_that_of_butter_to_he r_grandmother,_the_door_grandmamma,_who_was_bed,_she_the_better_the_wolf_sa_ and_she_little_red_her_grandmatm_aa_grandmother_mother_grandmother_mother_go od_wolf,_and_the_wolf_so_said_the_she_wolf,_and_i_have_gs_at_the_wolf,_wolf_bu tter_to_her_come_neard_the_bobbled_her_grandmamma,_grandmamma,_who_her_the _do_wolf_cake_her_grandmother_mother_to_her_to_she_me_the_better_to_her_the_b ettle_red_riding_hood_see_her_the_pot_of_butter_...
30 Direct / feedback tracking control training exploitation y( t) y( t-d) z -d u( t-d) pendulum pendulum y( t) z -d u( t) yref ( t+d) u( t) network "observes" torque u(t-d), plant state y(t-d), y(t) network input: reference y ref (t+d) and state feedback y(t) network learns how u(t) depends on y(t), y(t+d) network computes control input u(t)
31 Simulated pendulum example training data torque u exploitation control signal u issued by network [ ] disturb. [ ] pendulum coord. y y actual [ ] reference [ ] plus coord. x, plus (in state feedback) actual [ ] reference [ ]
32 Robot motor control: training and using the controller simulated robot wheel speeds Desired robot wheel speeds y( t-2) z -2 y( t) y( t) y ( ref t+2 ) u( t) z -2 u( t) u( t-2) previously trained ESN model of robot training data points (about 5 min robot run) 100 unit network original PWM robot data 1) (left & right) Controller network with output feedback NMSE about ) data courtesy Paul Plöger et al.
33 Results Original controller hand-designed PID (feedback) controller controlled trajectory lags behind target Target Controlled Control input Trained controller run on target generated by reference model Target Controlled Control input
34 5.2 Stable short-term memory
35 Stable short-term memory Motivation. Reservoir memory is transient, but we need stable dynamical STM as well In single-word speech/writing recognition: good to remember letter from beginning of word at end In phrase processing: good to remember initial words / grammatical categories
36 Example and architecture Task: count opening and closing brackets in text Text: random sequence of distorted {, } and 82 distractor symbols, in 12-pixel-high grayscale image Idea: for each opening "{", flip one further output neuron from -0.5 to dim grayscale vector y 1 y 2 # of open { y 1 y neuron reservoir trained fixed
37 Performance Red lines: output values Green lines: targets 0.19 % errors/character with 6 bracket levels (Pascanu & Jaeger 2010) Current Jacobs: mathematics of attractors in input-driven systems (Work and figures by Razvan Pascanu)
38 References Atiya, A. F., Parlos, A. G. (2000), New results on recurrent network training: Unifying the algorithms and accelerating convergence. IEEE Transactions on Neural Networks, 11(3): Funahashi, K.-I., Nakamura, Y. (1993), Approximation of dynamical systems by continuous time recurrent neural networks. Neural Networks 6, 1993, Gers, F.A., Schmidhuber, J., Cummins, F. (2000), Learning to forget: continual prediction with LSTM. Neural Computation 12(10), Hermans, M., Schrauwen, B. (2009): Memory in linear recurrent neural networks in continuous time. Neural Networks Jaeger, H. (2001): The "echo state" approach to analysing and training recurrent neural networks. GMD Report 148, German National Research Center for Information Technology, 2001Jaeger, H. (2001): Short term memory in echo state networks. GMD Report 152, German National Research Center for Information Technology, 2001 H. Jaeger (2006): Generating exponentially many periodic attractors with linearly growing Echo State Networks. IUB technical report Nr. 3 Jaeger, H. (2007), Discovering multiscale dynamical features with hierarchical Echo State Networks. Technical Report 10, School of Engineering and Science, Jacobs University. H., Haas, H. (2004), Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication. Science 304, 2004, Jaeger, H. (2010): Reservoir Self-Control for Achieving Invariance Against Slow Input Distortions. Jacobs University technical report Nr. 23 Jaeger, H., Haas, H. (2004), Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication. Science 304, 2004, Jaeger, H., Lukosevicius, M. Popovici, D. (2007): Optimization and Applications of Echo State Networks with Leaky Integrator Neurons. Neural Networks 20(3), , 2007 Jaeger, H., Maass, W.,Principe, J. (2007), Special issue on echo state networks and liquid state machines. Neural Networks 20(3), 2007, Kudo, M., Toyama, J., Shimbo, M. (1999), Multidimensional Curve Classification Using Passing-Through Regions. Pattern Recognition Letters, Vol. 20, No , pages Lukosecicius, M., Jaeger, H. (2009, to appear): Reservoir Computing Approaches to Recurrent Neural Network Training. Computer Science Review Maass, W., Natschlaeger, T., Markram, H. (2002), Real-Time Computing Without Stable States: A New Framework for Neural Computation Based on Perturbations. Neural Computation 14(11), 2002, Maass, W., Joshi, P., Sontag, E. (2007), Computational aspects of feedback in neural circuits. PLOS Computational Biology 3(1), 2007, 1-20 Pascanu, R. and Jaeger, H. (2010): A Neurodynamical Model for Working Memory. Neural Networks; DOI /j.neunet Strickert, M. (2004). Self-organizing neural networks for sequence processing. Ph.D. thesis. Univ. of Osnabrück, Dpt. of Computer Science. Diss384 thesis.pdf. White, O. L., Lee, D. D., Sompolinsky, H. S. (2004), Short-term Memory in Orthogonal Neural Networks. Phys. Rev. Lett. 92(14), 2004,
International University Bremen Guided Research Proposal Improve on chaotic time series prediction using MLPs for output training
International University Bremen Guided Research Proposal Improve on chaotic time series prediction using MLPs for output training Aakash Jain a.jain@iu-bremen.de Spring Semester 2004 1 Executive Summary
More informationGood vibrations: the issue of optimizing dynamical reservoirs
Good vibrations: the issue of optimizing dynamical reservoirs Workshop on ESNs / LSMs, NIPS 2006 Herbert Jaeger International University Bremen (Jacobs University Bremen, as of Spring 2007) The basic idea:
More informationHarnessing Nonlinearity: Predicting Chaotic Systems and Saving
Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication Publishde in Science Magazine, 2004 Siamak Saliminejad Overview Eco State Networks How to build ESNs Chaotic
More informationReservoir Computing and Echo State Networks
An Introduction to: Reservoir Computing and Echo State Networks Claudio Gallicchio gallicch@di.unipi.it Outline Focus: Supervised learning in domain of sequences Recurrent Neural networks for supervised
More informationUsing reservoir computing in a decomposition approach for time series prediction.
Using reservoir computing in a decomposition approach for time series prediction. Francis wyffels, Benjamin Schrauwen and Dirk Stroobandt Ghent University - Electronics and Information Systems Department
More informationReservoir Computing with Stochastic Bitstream Neurons
Reservoir Computing with Stochastic Bitstream Neurons David Verstraeten, Benjamin Schrauwen and Dirk Stroobandt Department of Electronics and Information Systems (ELIS), Ugent {david.verstraeten, benjamin.schrauwen,
More informationNegatively Correlated Echo State Networks
Negatively Correlated Echo State Networks Ali Rodan and Peter Tiňo School of Computer Science, The University of Birmingham Birmingham B15 2TT, United Kingdom E-mail: {a.a.rodan, P.Tino}@cs.bham.ac.uk
More informationShort Term Memory Quantifications in Input-Driven Linear Dynamical Systems
Short Term Memory Quantifications in Input-Driven Linear Dynamical Systems Peter Tiňo and Ali Rodan School of Computer Science, The University of Birmingham Birmingham B15 2TT, United Kingdom E-mail: {P.Tino,
More informationModelling Time Series with Neural Networks. Volker Tresp Summer 2017
Modelling Time Series with Neural Networks Volker Tresp Summer 2017 1 Modelling of Time Series The next figure shows a time series (DAX) Other interesting time-series: energy prize, energy consumption,
More informationMemory Capacity of Input-Driven Echo State NetworksattheEdgeofChaos
Memory Capacity of Input-Driven Echo State NetworksattheEdgeofChaos Peter Barančok and Igor Farkaš Faculty of Mathematics, Physics and Informatics Comenius University in Bratislava, Slovakia farkas@fmph.uniba.sk
More informationLecture 5: Recurrent Neural Networks
1/25 Lecture 5: Recurrent Neural Networks Nima Mohajerin University of Waterloo WAVE Lab nima.mohajerin@uwaterloo.ca July 4, 2017 2/25 Overview 1 Recap 2 RNN Architectures for Learning Long Term Dependencies
More informationRefutation of Second Reviewer's Objections
Re: Submission to Science, "Harnessing nonlinearity: predicting chaotic systems and boosting wireless communication." (Ref: 1091277) Refutation of Second Reviewer's Objections Herbert Jaeger, Dec. 23,
More informationREAL-TIME COMPUTING WITHOUT STABLE
REAL-TIME COMPUTING WITHOUT STABLE STATES: A NEW FRAMEWORK FOR NEURAL COMPUTATION BASED ON PERTURBATIONS Wolfgang Maass Thomas Natschlager Henry Markram Presented by Qiong Zhao April 28 th, 2010 OUTLINE
More informationDirect Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions
Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions Artem Chernodub, Institute of Mathematical Machines and Systems NASU, Neurotechnologies
More informationEcho State Networks with Filter Neurons and a Delay&Sum Readout
Echo State Networks with Filter Neurons and a Delay&Sum Readout Georg Holzmann 2,1 (Corresponding Author) http://grh.mur.at grh@mur.at Helmut Hauser 1 helmut.hauser@igi.tugraz.at 1 Institute for Theoretical
More informationFrom Neural Networks To Reservoir Computing...
From Neural Networks To Reservoir Computing... An Introduction Naveen Kuppuswamy, PhD Candidate, A.I Lab, University of Zurich 1 Disclaimer : I am only an egg 2 Part I From Classical Computation to ANNs
More informationarxiv: v1 [cs.lg] 2 Feb 2018
Short-term Memory of Deep RNN Claudio Gallicchio arxiv:1802.00748v1 [cs.lg] 2 Feb 2018 Department of Computer Science, University of Pisa Largo Bruno Pontecorvo 3-56127 Pisa, Italy Abstract. The extension
More informationTowards a Calculus of Echo State Networks arxiv: v1 [cs.ne] 1 Sep 2014
his space is reserved for the Procedia header, do not use it owards a Calculus of Echo State Networks arxiv:49.28v [cs.ne] Sep 24 Alireza Goudarzi and Darko Stefanovic,2 Department of Computer Science,
More informationDeep Recurrent Neural Networks
Deep Recurrent Neural Networks Artem Chernodub e-mail: a.chernodub@gmail.com web: http://zzphoto.me ZZ Photo IMMSP NASU 2 / 28 Neuroscience Biological-inspired models Machine Learning p x y = p y x p(x)/p(y)
More informationArtificial Neural Networks
Artificial Neural Networks 鮑興國 Ph.D. National Taiwan University of Science and Technology Outline Perceptrons Gradient descent Multi-layer networks Backpropagation Hidden layer representations Examples
More informationLecture 11 Recurrent Neural Networks I
Lecture 11 Recurrent Neural Networks I CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 01, 2017 Introduction Sequence Learning with Neural Networks Some Sequence Tasks
More informationArtificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen
Artificial Neural Networks Introduction to Computational Neuroscience Tambet Matiisen 2.04.2018 Artificial neural network NB! Inspired by biology, not based on biology! Applications Automatic speech recognition
More informationShort Term Memory and Pattern Matching with Simple Echo State Networks
Short Term Memory and Pattern Matching with Simple Echo State Networks Georg Fette (fette@in.tum.de), Julian Eggert (julian.eggert@honda-ri.de) Technische Universität München; Boltzmannstr. 3, 85748 Garching/München,
More informationRecurrent Neural Networks
Recurrent Neural Networks Datamining Seminar Kaspar Märtens Karl-Oskar Masing Today's Topics Modeling sequences: a brief overview Training RNNs with back propagation A toy example of training an RNN Why
More informationReservoir Computing in Forecasting Financial Markets
April 9, 2015 Reservoir Computing in Forecasting Financial Markets Jenny Su Committee Members: Professor Daniel Gauthier, Adviser Professor Kate Scholberg Professor Joshua Socolar Defense held on Wednesday,
More informationLecture 11 Recurrent Neural Networks I
Lecture 11 Recurrent Neural Networks I CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor niversity of Chicago May 01, 2017 Introduction Sequence Learning with Neural Networks Some Sequence Tasks
More informationLiquid Computing. Wolfgang Maass
Liquid Computing Wolfgang Maass Institute for Theoretical Computer Science Technische Universitaet Graz A-8010 Graz, Austria maass@igi.tugraz.at http://www.igi.tugraz.at/ Abstract. This review addresses
More informationAt the Edge of Chaos: Real-time Computations and Self-Organized Criticality in Recurrent Neural Networks
At the Edge of Chaos: Real-time Computations and Self-Organized Criticality in Recurrent Neural Networks Thomas Natschläger Software Competence Center Hagenberg A-4232 Hagenberg, Austria Thomas.Natschlaeger@scch.at
More informationSequence Modeling with Neural Networks
Sequence Modeling with Neural Networks Harini Suresh y 0 y 1 y 2 s 0 s 1 s 2... x 0 x 1 x 2 hat is a sequence? This morning I took the dog for a walk. sentence medical signals speech waveform Successes
More informationMODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION
Computing and Informatics, Vol. 30, 2011, 321 334 MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION Štefan Babinec, Jiří Pospíchal Department of Mathematics Faculty of Chemical and Food Technology
More informationRecurrence Enhances the Spatial Encoding of Static Inputs in Reservoir Networks
Recurrence Enhances the Spatial Encoding of Static Inputs in Reservoir Networks Christian Emmerich, R. Felix Reinhart, and Jochen J. Steil Research Institute for Cognition and Robotics (CoR-Lab), Bielefeld
More informationLong-Short Term Memory and Other Gated RNNs
Long-Short Term Memory and Other Gated RNNs Sargur Srihari srihari@buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics in Sequence Modeling
More informationAdvanced Methods for Recurrent Neural Networks Design
Universidad Autónoma de Madrid Escuela Politécnica Superior Departamento de Ingeniería Informática Advanced Methods for Recurrent Neural Networks Design Master s thesis presented to apply for the Master
More informationFrom Neural Networks To Reservoir Computing...
From Neural Networks To Reservoir Computing... An Introduction Naveen Kuppuswamy, PhD Candidate, A.I Lab, University of Zurich Disclaimer : I am only an egg 2 Part I From Classical Computation to ANNs
More informationEdge of Chaos Computation in Mixed-Mode VLSI - A Hard Liquid
Edge of Chaos Computation in Mixed-Mode VLSI - A Hard Liquid Felix Schürmann, Karlheinz Meier, Johannes Schemmel Kirchhoff Institute for Physics University of Heidelberg Im Neuenheimer Feld 227, 6912 Heidelberg,
More informationSample Exam COMP 9444 NEURAL NETWORKS Solutions
FAMILY NAME OTHER NAMES STUDENT ID SIGNATURE Sample Exam COMP 9444 NEURAL NETWORKS Solutions (1) TIME ALLOWED 3 HOURS (2) TOTAL NUMBER OF QUESTIONS 12 (3) STUDENTS SHOULD ANSWER ALL QUESTIONS (4) QUESTIONS
More informationOn the use of Long-Short Term Memory neural networks for time series prediction
On the use of Long-Short Term Memory neural networks for time series prediction Pilar Gómez-Gil National Institute of Astrophysics, Optics and Electronics ccc.inaoep.mx/~pgomez In collaboration with: J.
More informationAnalyzing the weight dynamics of recurrent learning algorithms
Analyzing the weight dynamics of recurrent learning algorithms Ulf D. Schiller and Jochen J. Steil Neuroinformatics Group, Faculty of Technology, Bielefeld University Abstract We provide insights into
More informationLast update: October 26, Neural networks. CMSC 421: Section Dana Nau
Last update: October 26, 207 Neural networks CMSC 42: Section 8.7 Dana Nau Outline Applications of neural networks Brains Neural network units Perceptrons Multilayer perceptrons 2 Example Applications
More informationNeural Networks. Volker Tresp Summer 2015
Neural Networks Volker Tresp Summer 2015 1 Introduction The performance of a classifier or a regression model critically depends on the choice of appropriate basis functions The problem with generic basis
More informationA Model for Real-Time Computation in Generic Neural Microcircuits
A Model for Real-Time Computation in Generic Neural Microcircuits Wolfgang Maass, Thomas Natschläger Institute for Theoretical Computer Science Technische Universitaet Graz A-81 Graz, Austria maass, tnatschl
More informationGrundlagen der Künstlichen Intelligenz
Grundlagen der Künstlichen Intelligenz Neural networks Daniel Hennes 21.01.2018 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Logistic regression Neural networks Perceptron
More informationNeural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28
1 / 28 Neural Networks Mark van Rossum School of Informatics, University of Edinburgh January 15, 2018 2 / 28 Goals: Understand how (recurrent) networks behave Find a way to teach networks to do a certain
More informationarxiv: v3 [cs.lg] 14 Jan 2018
A Gentle Tutorial of Recurrent Neural Network with Error Backpropagation Gang Chen Department of Computer Science and Engineering, SUNY at Buffalo arxiv:1610.02583v3 [cs.lg] 14 Jan 2018 1 abstract We describe
More informationAn overview of reservoir computing: theory, applications and implementations
An overview of reservoir computing: theory, applications and implementations Benjamin Schrauwen David Verstraeten Jan Van Campenhout Electronics and Information Systems Department, Ghent University, Belgium
More informationARTIFICIAL INTELLIGENCE. Artificial Neural Networks
INFOB2KI 2017-2018 Utrecht University The Netherlands ARTIFICIAL INTELLIGENCE Artificial Neural Networks Lecturer: Silja Renooij These slides are part of the INFOB2KI Course Notes available from www.cs.uu.nl/docs/vakken/b2ki/schema.html
More informationChristian Mohr
Christian Mohr 20.12.2011 Recurrent Networks Networks in which units may have connections to units in the same or preceding layers Also connections to the unit itself possible Already covered: Hopfield
More informationFisher Memory of Linear Wigner Echo State Networks
ESA 07 proceedings, European Symposium on Artificial eural etworks, Computational Intelligence and Machine Learning. Bruges (Belgium), 6-8 April 07, i6doc.com publ., ISB 978-87587039-. Fisher Memory of
More informationNeural networks. Chapter 19, Sections 1 5 1
Neural networks Chapter 19, Sections 1 5 Chapter 19, Sections 1 5 1 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 19, Sections 1 5 2 Brains 10
More informationNeural networks. Chapter 20, Section 5 1
Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of
More informationA First Attempt of Reservoir Pruning for Classification Problems
A First Attempt of Reservoir Pruning for Classification Problems Xavier Dutoit, Hendrik Van Brussel, Marnix Nuttin Katholieke Universiteit Leuven - P.M.A. Celestijnenlaan 300b, 3000 Leuven - Belgium Abstract.
More informationInitialization and Self-Organized Optimization of Recurrent Neural Network Connectivity
Initialization and Self-Organized Optimization of Recurrent Neural Network Connectivity Joschka Boedecker Dep. of Adaptive Machine Systems, Osaka University, Suita, Osaka, Japan Oliver Obst CSIRO ICT Centre,
More informationNeural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21
Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural
More informationControlling Recurrent Neural Networks by Conceptors
arxiv:143.3369v3 [cs.ne] 22 Apr 217 Herbert Jaeger Controlling Recurrent Neural Networks by Conceptors Technical Report No. 31 March 214 School of Engineering and Science Controlling Recurrent Neural Networks
More informationMachine Learning. Neural Networks
Machine Learning Neural Networks Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 Biological Analogy Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 THE
More informationArtificial Neural Networks
Artificial Neural Networks Threshold units Gradient descent Multilayer networks Backpropagation Hidden layer representations Example: Face Recognition Advanced topics 1 Connectionist Models Consider humans:
More informationReservoir Computing Methods for Prognostics and Health Management (PHM) Piero Baraldi Energy Department Politecnico di Milano Italy
Reservoir Computing Methods for Prognostics and Health Management (PHM) Piero Baraldi Energy Department Politecnico di Milano Italy 2 Piero Baraldi Data Industry 4.0 2 Digitalization 2.8 Trillion GD (ZD)
More informationHopfield Networks. (Excerpt from a Basic Course at IK 2008) Herbert Jaeger. Jacobs University Bremen
Hopfield Networks (Excerpt from a Basic Course at IK 2008) Herbert Jaeger Jacobs University Bremen Building a model of associative memory should be simple enough... Our brain is a neural network Individual
More informationDeep Learning Recurrent Networks 2/28/2018
Deep Learning Recurrent Networks /8/8 Recap: Recurrent networks can be incredibly effective Story so far Y(t+) Stock vector X(t) X(t+) X(t+) X(t+) X(t+) X(t+5) X(t+) X(t+7) Iterated structures are good
More informationLiquid Computing. Wolfgang Maass. Institut für Grundlagen der Informationsverarbeitung Technische Universität Graz, Austria
NNB SS10 1 Liquid Computing Wolfgang Maass Institut für Grundlagen der Informationsverarbeitung Technische Universität Graz, Austria Institute for Theoretical Computer Science http://www.igi.tugraz.at/maass/
More informationNeural Nets and Symbolic Reasoning Hopfield Networks
Neural Nets and Symbolic Reasoning Hopfield Networks Outline The idea of pattern completion The fast dynamics of Hopfield networks Learning with Hopfield networks Emerging properties of Hopfield networks
More informationinput unit N internal units output unit
Adaptive nonlinear system identification with echo state networks Herbert Jaeger Fraunhofer Institute for Autonomous Intelligent Systems AIS.INDY Schloss Birlinghoven, D-53754 Sankt Augustin, Germany herbert.jaeger@ais.fraunhofer.de
More informationarxiv: v1 [cs.lg] 8 Jul 2017
Tailoring Artificial Neural Networks for Optimal Learning Pau Vilimelis Aceituno 1,2, Gang Yan 3, Yang-Yu Liu 1,4 1 Channing Division of Network Medicine, Brigham and Women s Hospital, Harvard Medical
More informationSeveral ways to solve the MSO problem
Several ways to solve the MSO problem J. J. Steil - Bielefeld University - Neuroinformatics Group P.O.-Box 0 0 3, D-3350 Bielefeld - Germany Abstract. The so called MSO-problem, a simple superposition
More informationCS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning
CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning Lei Lei Ruoxuan Xiong December 16, 2017 1 Introduction Deep Neural Network
More informationDeep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści
Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, 2017 Spis treści Website Acknowledgments Notation xiii xv xix 1 Introduction 1 1.1 Who Should Read This Book?
More informationLong-Short Term Memory
Long-Short Term Memory Sepp Hochreiter, Jürgen Schmidhuber Presented by Derek Jones Table of Contents 1. Introduction 2. Previous Work 3. Issues in Learning Long-Term Dependencies 4. Constant Error Flow
More informationLecture 17: Neural Networks and Deep Learning
UVA CS 6316 / CS 4501-004 Machine Learning Fall 2016 Lecture 17: Neural Networks and Deep Learning Jack Lanchantin Dr. Yanjun Qi 1 Neurons 1-Layer Neural Network Multi-layer Neural Network Loss Functions
More informationOptoelectronic Systems Trained With Backpropagation Through Time
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 26, NO. 7, JULY 2015 1545 Optoelectronic Systems Trained With Backpropagation Through Time Michiel Hermans, Joni Dambre, and Peter Bienstman
More informationA New Look at Nonlinear Time Series Prediction with NARX Recurrent Neural Network. José Maria P. Menezes Jr. and Guilherme A.
A New Look at Nonlinear Time Series Prediction with NARX Recurrent Neural Network José Maria P. Menezes Jr. and Guilherme A. Barreto Department of Teleinformatics Engineering Federal University of Ceará,
More information(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann
(Feed-Forward) Neural Networks 2016-12-06 Dr. Hajira Jabeen, Prof. Jens Lehmann Outline In the previous lectures we have learned about tensors and factorization methods. RESCAL is a bilinear model for
More informationNeural Network Control of Robot Manipulators and Nonlinear Systems
Neural Network Control of Robot Manipulators and Nonlinear Systems F.L. LEWIS Automation and Robotics Research Institute The University of Texas at Arlington S. JAG ANNATHAN Systems and Controls Research
More informationCourse 395: Machine Learning - Lectures
Course 395: Machine Learning - Lectures Lecture 1-2: Concept Learning (M. Pantic) Lecture 3-4: Decision Trees & CBC Intro (M. Pantic & S. Petridis) Lecture 5-6: Evaluating Hypotheses (S. Petridis) Lecture
More informationMachine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6
Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)
More informationUsing a Hopfield Network: A Nuts and Bolts Approach
Using a Hopfield Network: A Nuts and Bolts Approach November 4, 2013 Gershon Wolfe, Ph.D. Hopfield Model as Applied to Classification Hopfield network Training the network Updating nodes Sequencing of
More informationNeural Networks 2. 2 Receptive fields and dealing with image inputs
CS 446 Machine Learning Fall 2016 Oct 04, 2016 Neural Networks 2 Professor: Dan Roth Scribe: C. Cheng, C. Cervantes Overview Convolutional Neural Networks Recurrent Neural Networks 1 Introduction There
More informationTowards logics for neural conceptors
Towards logics for neural conceptors Till Mossakowski joint work with Razvan Diaconescu and Martin Glauer Otto-von-Guericke-Universität Magdeburg AITP 2018, Aussois, March 30, 2018 Overview 1 Conceptors
More informationMultimodal context analysis and prediction
Multimodal context analysis and prediction Valeria Tomaselli (valeria.tomaselli@st.com) Sebastiano Battiato Giovanni Maria Farinella Tiziana Rotondo (PhD student) Outline 2 Context analysis vs prediction
More informationArtificial Intelligence
Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement
More information100 inference steps doesn't seem like enough. Many neuron-like threshold switching units. Many weighted interconnections among units
Connectionist Models Consider humans: Neuron switching time ~ :001 second Number of neurons ~ 10 10 Connections per neuron ~ 10 4 5 Scene recognition time ~ :1 second 100 inference steps doesn't seem like
More informationStructured reservoir computing with spatiotemporal chaotic attractors
Structured reservoir computing with spatiotemporal chaotic attractors Carlos Lourenço 1,2 1- Faculty of Sciences of the University of Lisbon - Informatics Department Campo Grande, 1749-016 Lisboa - Portugal
More informationDeep Learning. Recurrent Neural Network (RNNs) Ali Ghodsi. October 23, Slides are partially based on Book in preparation, Deep Learning
Recurrent Neural Network (RNNs) University of Waterloo October 23, 2015 Slides are partially based on Book in preparation, by Bengio, Goodfellow, and Aaron Courville, 2015 Sequential data Recurrent neural
More informationRecurrent neural networks
12-1: Recurrent neural networks Prof. J.C. Kao, UCLA Recurrent neural networks Motivation Network unrollwing Backpropagation through time Vanishing and exploding gradients LSTMs GRUs 12-2: Recurrent neural
More informationArtificial Neural Networks Examination, June 2004
Artificial Neural Networks Examination, June 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More informationAnalysis of Multilayer Neural Network Modeling and Long Short-Term Memory
Analysis of Multilayer Neural Network Modeling and Long Short-Term Memory Danilo López, Nelson Vera, Luis Pedraza International Science Index, Mathematical and Computational Sciences waset.org/publication/10006216
More informationARCHITECTURAL DESIGNS OF ECHO STATE NETWORK
ARCHITECTURAL DESIGNS OF ECHO STATE NETWORK by ALI ABDALLAH ALI AL RODAN A thesis submitted to The University of Birmingham for the degree of DOCTOR OF PHILOSOPHY School of Computer Science College of
More informationLecture 5 Neural models for NLP
CS546: Machine Learning in NLP (Spring 2018) http://courses.engr.illinois.edu/cs546/ Lecture 5 Neural models for NLP Julia Hockenmaier juliahmr@illinois.edu 3324 Siebel Center Office hours: Tue/Thu 2pm-3pm
More informationChapter 1. Liquid State Machines: Motivation, Theory, and Applications
Chapter 1 Liquid State Machines: Motivation, Theory, and Applications Wolfgang Maass Institute for Theoretical Computer Science Graz University of Technology A-8010 Graz, Austria maass@igi.tugraz.at The
More informationDevelopment of a Deep Recurrent Neural Network Controller for Flight Applications
Development of a Deep Recurrent Neural Network Controller for Flight Applications American Control Conference (ACC) May 26, 2017 Scott A. Nivison Pramod P. Khargonekar Department of Electrical and Computer
More informationAn Introductory Course in Computational Neuroscience
An Introductory Course in Computational Neuroscience Contents Series Foreword Acknowledgments Preface 1 Preliminary Material 1.1. Introduction 1.1.1 The Cell, the Circuit, and the Brain 1.1.2 Physics of
More informationSimple Deterministically Constructed Cycle Reservoirs with Regular Jumps
1 Simple Deterministically Constructed Cycle Reservoirs with Regular Jumps Ali Rodan 1 and Peter Tiňo 1 1 School of computer Science, The University of Birmingham, Birmingham B15 2TT, United Kingdom, (email:
More informationTutorial on Machine Learning for Advanced Electronics
Tutorial on Machine Learning for Advanced Electronics Maxim Raginsky March 2017 Part I (Some) Theory and Principles Machine Learning: estimation of dependencies from empirical data (V. Vapnik) enabling
More informationIntroduction to RNNs!
Introduction to RNNs Arun Mallya Best viewed with Computer Modern fonts installed Outline Why Recurrent Neural Networks (RNNs)? The Vanilla RNN unit The RNN forward pass Backpropagation refresher The RNN
More information1 What makes a dynamical system computationally powerful?
1 What makes a dynamical system computationally powerful? Robert Legenstein and Wolfgang Maass Institute for Theoretical Computer Science Technische Universität Graz Austria, 8010 Graz {legi, maass}@igi.tugraz.at
More informationExplorations in Echo State Networks
Explorations in Echo State Networks Adrian Millea Master s Thesis Supervised by Dr. Marco Wiering (Department of Artificial Intelligence, University of Groningen, Groningen, The Netherlands) Prof. Mark
More informationarxiv: v1 [cs.lg] 9 Mar 2018
Hybrid Forecasting of Chaotic Processes: Using Machine Learning in Conjunction with a Knowledge-Based Model Jaideep Pathak, 1 Alexander Wikner, 2 Rebeckah Fussell, 3 Sarthak Chandra, 1 Brian R. Hunt, 1
More informationArtificial Neural Networks Examination, June 2005
Artificial Neural Networks Examination, June 2005 Instructions There are SIXTY questions. (The pass mark is 30 out of 60). For each question, please select a maximum of ONE of the given answers (either
More informationIntroduction to Neural Networks
Introduction to Neural Networks What are (Artificial) Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning
More informationLearning and Memory in Neural Networks
Learning and Memory in Neural Networks Guy Billings, Neuroinformatics Doctoral Training Centre, The School of Informatics, The University of Edinburgh, UK. Neural networks consist of computational units
More informationAnalyzing the State Space Property of Echo
Proceedings of International Joint Conference on Neural Networks, Montreal, Canada, July 31 August 4, 25 Analyzing the State Space Property of Echo State Networks for Chaotic System Prediction AbstractFor
More information