Evolutionay Wavelet Networks for Statistical Time Series Analysis

Similar documents
Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine

A Wavelet Neural Network Forecasting Model Based On ARIMA

ARTIFICIAL NEURAL NETWORK WITH HYBRID TAGUCHI-GENETIC ALGORITHM FOR NONLINEAR MIMO MODEL OF MACHINING PROCESSES

Choosing Variables with a Genetic Algorithm for Econometric models based on Neural Networks learning and adaptation.

Evolutionary Functional Link Interval Type-2 Fuzzy Neural System for Exchange Rate Prediction

Cluster based pruning and survival selection using soft computing

A Hybrid Model of Wavelet and Neural Network for Short Term Load Forecasting

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Integer weight training by differential evolution algorithms

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

USING WAVELET NEURAL NETWORK FOR THE IDENTIFICATION OF A BUILDING STRUCTURE FROM EXPERIMENTAL DATA

IMPROVING TIME SERIES PREDICTION

Distributed Genetic Algorithm for feature selection in Gaia RVS spectra. Application to ANN parameterization

An Improved Differential Evolution Trained Neural Network Scheme for Nonlinear System Identification

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

A Novel 2-D Model Approach for the Prediction of Hourly Solar Radiation

Wavelet Neural Networks for Nonlinear Time Series Analysis

Incremental Stochastic Gradient Descent

Artificial Neural Networks

Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions

4. Multilayer Perceptrons

A Hybrid Time-delay Prediction Method for Networked Control System

Model Identification in Wavelet Neural Networks Framework

Introduction to Signal Processing

Functional Radial Basis Function Networks (FRBFN)

Nonlinear System Identification Based on a Novel Adaptive Fuzzy Wavelet Neural Network

Lecture 4: Perceptrons and Multilayer Perceptrons

APPLICATION OF RADIAL BASIS FUNCTION NEURAL NETWORK, TO ESTIMATE THE STATE OF HEALTH FOR LFP BATTERY

Convolutional Neural Network Architecture

Forecasting & Futurism

A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE

A Bayesian Local Linear Wavelet Neural Network

Feedforward Neural Nets and Backpropagation

A Training-time Analysis of Robustness in Feed-Forward Neural Networks

Classification of Ordinal Data Using Neural Networks

The connection of dropout and Bayesian statistics

NONLINEAR PLANT IDENTIFICATION BY WAVELETS

ECE521 Lectures 9 Fully Connected Neural Networks

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

Reservoir Computing and Echo State Networks

An Evolutionary Programming Based Algorithm for HMM training

Artificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence

Classification of Hand-Written Digits Using Scattering Convolutional Network

Multilayer Neural Networks

Data Mining Part 5. Prediction

AN INTERACTIVE WAVELET ARTIFICIAL NEURAL NETWORK IN TIME SERIES PREDICTION

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION Vol. VI - System Identification Using Wavelets - Daniel Coca and Stephen A. Billings

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Artificial Neural Networks Examination, March 2004

A multiple testing procedure for input variable selection in neural networks

Neuroevolution for sound event detection in real life audio: A pilot study

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

Radial Basis Function (RBF) Networks

Lab 5: 16 th April Exercises on Neural Networks

Multiple Change Point Detection by Sparse Parameter Estimation

COMBINATION OF TAGUCHI METHOD AND ARTIFICIAL INTELLIGENCE TECHNIQUES FOR THE OPTIMAL DESIGN OF FLAT-PLATE COLLECTORS

An Improved Quantum Evolutionary Algorithm with 2-Crossovers

From perceptrons to word embeddings. Simon Šuster University of Groningen

Symmetric Wavelet Tight Frames with Two Generators

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Chapter 4 Neural Networks in System Identification

Comparison of the Complex Valued and Real Valued Neural Networks Trained with Gradient Descent and Random Search Algorithms

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Data-driven methods in application to flood defence systems monitoring and analysis Pyayt, A.

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Neural Network Weight Space Symmetries Can Speed up Genetic Learning

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

Introduction to Neural Networks: Structure and Training

22c145-Fall 01: Neural Networks. Neural Networks. Readings: Chapter 19 of Russell & Norvig. Cesare Tinelli 1

ECE Introduction to Artificial Neural Network and Fuzzy Systems

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Neural Networks. Bishop PRML Ch. 5. Alireza Ghane. Feed-forward Networks Network Training Error Backpropagation Applications

Forecasting River Flow in the USA: A Comparison between Auto-Regression and Neural Network Non-Parametric Models

Bearing fault diagnosis based on EMD-KPCA and ELM

Neural Networks biological neuron artificial neuron 1

Condensed Table of Contents for Introduction to Stochastic Search and Optimization: Estimation, Simulation, and Control by J. C.

ADAPTIVE NEURO-FUZZY INFERENCE SYSTEMS

Neural Networks: Backpropagation

Approximation of Functions by Multivariable Hermite Basis: A Hybrid Method

COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

Recurrent Neural Networks

Modelling Time Series with Neural Networks. Volker Tresp Summer 2017

Cascade Neural Networks with Node-Decoupled Extended Kalman Filtering

Forecasting Chaotic time series by a Neural Network

Predict Time Series with Multiple Artificial Neural Networks

Lecture 4: Feed Forward Neural Networks

Active Robot Calibration Algorithm

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS

A Hybrid Method of CART and Artificial Neural Network for Short-term term Load Forecasting in Power Systems

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

Introduction to Neural Networks

Data Mining. Preamble: Control Application. Industrial Researcher s Approach. Practitioner s Approach. Example. Example. Goal: Maintain T ~Td

A new method for short-term load forecasting based on chaotic time series and neural network

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Neural Network Training

OPTIMAL SURE PARAMETERS FOR SIGMOIDAL WAVELET SHRINKAGE

Transcription:

Evolutionay Wavelet Networks for Statistical Time Series Analysis TOMMASO MINERVA Department of Social Sciences University of Modena and Reggio Emilia Via Allegri 9, Reggio Emilia, I-42100 ITALY tommaso.minerva@unimore.it Abstract: In this paper I describe an evolutionary wavelet network to optimize the filtering of a statistical time series into separate contributions. The wavelet base is regarded as a neural network where the network nodes are discrete wavelet transforms, the wavelon, and the network structure and parameters are selected through evolutionary techniques. With this combined approach I can separate stochastic from structural components within an optimized framework and finally I can perform optimized predictive analysis on the time series components. Key-words: Statistical Time Series, Evolutionary Algorithms, Signal Processing, Wavelet, Filtering, Stochastic Decomposition, Wavelet Networks 1 Introduction In the early 90s, Zhang (Zhang and Benveniste, 1992; Zhang 1994; et al.) introduced wavelet networks as one hidden-layer net where the neurons in the hidden layer are wavelets (wavelon). There are some advantages in using Wavelet Networks instead of feed forward neural nets mainly related to the problem that standard feed forward networks can bring to a lack of efficiency for determining the net parameters and the net topology. Moreover, the issue of the optimal network selection is still an open problem. As shown further Wavelet Networks can be efficiently used to filter non stationary and complex time series (such as financial t.s., environmental t.s., dynamical control systems, etc ). If we consider the non parametric regression model involving an unknown function: (1) where X (input) and Y (output) are random variables in R d and R respectively, is a noise of zero mean, f is supposed to be non linear in some functional space and d is the input dimension. n If O 1 = {(x 1,y 1 ), (x 2,y 2 ),.(x n,y n )} is a sample of input-output pair (training data set) the problem of regression is to find a non-parametric estimator f* of f based on the observations. In this case non parametric means that f* belongs to some parameterized class of functions but flexible and with an unknown number of parameters in f*. This kind of non-parametric estimators are useful for black-box non-linear modelling (such as Neural Networks). The design of the estimator, usually, consists in two main steps: 1) Setting a class of possibile candidate f* functions (usually very large); 2) Developing an algorithm to determine the parameters in f* based on the observed data. In the proposed design the two points are addressed at the same footing. The first, by using the scheme of wavelet networks while the second by evolutionary computational approaches. The introduction of evolutionary wavelet networks in time series analysis can introduce some advantages respect to standard wavelet transform. Infact we should keep into full account that in signal filtering we must deal with border effects and this is a crucial effect in predictive models on Time Series where the border, i.e. the more recent observations, is the most significant area of data. We should also consider that it is not always possible to build orthonormal basis so the difference between ISSN: 1790-2769 95 ISBN: 978-960-474-189-2

The network structure (schetched in figure 1) has a wavelet on each neuron and can be generalized to M inputs and N outputs. standard wavelet transforms and wavelet networks could be also based on the properties of the expansion basis. In this could be usefull to consider that we can achieve more freedom if we give up the idea that the wavelet family should be a basis requiring only that the family is a redundant frame of localized functions. We can also give up the idea of a fixed regular lattice for the dilations and translations because this introduce some distorsions on the natural wavelet spectra. Moreover we prefer to use observed data to build the wavelet family (frame). The relation: (2) can be regarded as a one-hidden-layer feedforward neural network where each of the N neurons in the intermediate layer (wavelon) is a dilated and translated wavelet and u i represents the weight of the i- th wavelon. In this relation the parameters u i, a i, b i represent the wavelet expansion parameters and they have to be estimated numerically on the training data while the number of wavelon (hidden nodes), N, represents another parameter and another parameter is represented by the mother-wavelet, to be used in the wavelet expansion. This 3 layers (input-wavelon-output) structure is called Wavelet Network. We introduce also a bias parameter on each neuron to obtain eq. (3): (3) Figure 1. The Wavelet Network structure This network structure can be considered as another universal function approximator but with some kindly properties: Can efficiently deal with local properties and high non-linearity respect to Feed Forward Networks or Radial Basis Networks; Can be demonstrated that Radial Basis Networks are a special class of Wavelet Networks; Can efficiently deal with a high-order input space respect to standard wavelet transforms; Can be more flexible respect to standard ortho-normal discrete wavelet transform; Can be efficiently used in wavelet filtering without border effects because one has not to evaluate the coefficient as an integral over the wavelet. If we reconsider the eq. (3) we can identify the network parameters. The model, i.e. the wavelet decomposition, is established only when we are able to estimate the complete and optimal set of parameters: u i, a i, b i,, N and for each i=1,n. With this constrains the evaluation of the model, i.e. the set of parameters in equation (3) could be not so trivial and complex computational techniques have to be used. This is not a trivial task and several approaches has been proposed in recent literature. In this paper I introduce an approach based on evolutionary algorithms. 2 Evolutionary Wavelet Networks Zhang (1992, 1994) proposed an approach starting from a discrete candidate set of wavelet and then selecting the appropriate ones by stepwise-like algorithms, where the neural weights were evaluated by usual adaptive gradient-like algorithms. Y.He, F.Chu and B. Zhong (2002) and K. Kobayashi and T. Torioka (1998) have already proposed some evolutionary approaches to overcome the strong sensitivity to initial condition and path dependency in the stepwise approaches. Here I generalize this evolutionary approaches to evaluate on the some footing all the parameters in a ISSN: 1790-2769 96 ISBN: 978-960-474-189-2

Figure 2. The evolutionary code Wavelet Network without lacking the property to separate the signal (the time series) into different frequency domain to obtain an efficient filtering. Evolutionary Computation refers to a class of population based stochatistic global search algorithms developed from the idea and principles of natural evolution. Individuals in a population encode a candidate solution and compete and exchange information with each other in order to perform a task, exactly as in natural evolution. Evolutionary operators (selection, crossover, mutation, elitism, ) evolve the population respect to a fitness function measuring the goodness of each individual (solution) respect to the optimization problem. Evolutionary approaches can be useful for dealing with large complex problems with many local optimal. They can perform global search without being trapped in local optima and do not use gradient-like operators so do not require usual regularities of the functions. All these properties indicate that evolutionary approaches can be useful in tackling our problem. An Evolutionary scheme is represented by the following pseudo-code of a general Evolutionary Schema: set i=0; randomly generate the initial population P(0); do { evaluate the fitness function of each individual in P(i); select parents from P(i) based on their fitness value; apply evolutionary operators (crossover, mutation, elitism, reinsertion) to parents to produce offsprings in P(i+1); i++; } while (convergence criterion is satisfied) optimal solution == best individual; where P represent the population (set of candidate solutions) and i is the generation (searching iteration). The main components in this Evolutionary Schema are represented by: The Individual Encoding, i.e. how the optimization/selection problem is coded in terms of Evolutionary Chromosome; The Fitness Function, i.e., how to evaluate the goodness of an individual respect to the optimization problem; Parameters of the evolutionary operators: mutation rate, selection rate, selection rule, crossover, etc ; The main difficulty in applying evolutionary schema is to set up the genetic codification of the search/selection problem and to establish an efficient fitness function able to drive algorithm to the optimal solution. In our problem the genetic code has to be able to determine the optimal combination of the parameters u i, a i, b i,, N and while the fitness function has to measure the predictive goodness of the solutions. If we fix N, the number of wavelons (wavelet neurons), the genetic code is composed by N + 1 fragments. Each one of the N fragments represents the parameters related to each wavelon: u i, a i, b i, while the N+1 th fragment encodes the mother-wavelet. So if we have N node in the wavelon layer than we have N fragments encoding the wavelon parameters (u i, a i, b i and ) and a fragment encoding the wavelet family (see figure 2). Fixed N, the parameters related to each wavelon: u i, a i, ISSN: 1790-2769 97 ISBN: 978-960-474-189-2

Model Number of Nodes Predictive RMS EvoWaveNet 21 18.45 WaveNet 29 38.35 Wavelet --- 42.56 BP-NN 29 103.56 AR 5 168.35 (order) Table 1. Results from the Evolutionary Wavelet Network on predictive analysis performed on a benchmark time series dataset: sun spot. The predictive RMS has been obtained on a test set of data out-of-the-sample. b i, are obtained by evolutionary minimization of the root mean square (RMS) error between target and output of the wavelet network. The number of wavelons, N, is selected by an iterative loop minimizing the Akaike Final Prediction Error Criterion (FPE) that compare the goodness of the approximation and the number of parameters in the wavelet networks. Usually 200 individuals and 2000 generation are used in the evolutionary computation for each iteration but the algorithm stops after few hundred of generations (depending on the net complexity) and requires from few seconds (low N) to more than 4 hours (N>40) of CPU-time on a Mac Os X Server 2x2,8 Ghz quad-core Processor with 16 GB of RAM. Firstly we tested the Evolutionary Wavelet Networks on several sample of non linear functions obtaining better approximations respect to other Wavelet Network reported in literature and respect to backpropagation Neural Nets. We investigate also the forecasting capabilities on the well known (used as a benchmark) sun-spot time series. The results are summarized in the table where the RMS is evaluated on a disjoint test set. In this calculation the data were disjointed into a training set, a validation set and a test set to perform out-of-the-sample validation analysis. The weights were obtained on the training set while the fitness function was evaluated on the validation set and, finally, the predictive analysis was performed on a disjoint test set. In table 1 I show that the Evolutionary Wavelet Network (EWN) better performs respect to standard Wavelet Networks, Wavelet Transform decomposition, Back Propagation feed-forward Neural Networks (BP-NN) and linear AR models. In the table is shown that the EWN results simpler respect to other network structure with only 21 hidden nodes while maintaining the best predictive RMS. The better of EWN is mainly obtained in treating border effects on the recent data. The optimized Evolutionary Wavelet Network can be used as a filtering engine in stochastic time series filtering where noise and signal are separated by a wavelet decomposition [5]. 3 Conclusions In this paper I reported an evolutionary approach to build efficient wavelet expansion to perform predictive time series analysis based on stochastic wavelet filtering. The evolutionary approach can be used to break some regularity assumption of standard wavelet decomposition as, for example, regular wavelet lattice and regular wavelet basis. The wavelet filtering allowed us to separate stochastic from deterministic components for time series coming from complex phenomena (such as financial markets) ISSN: 1790-2769 98 ISBN: 978-960-474-189-2

as reported in a previous work (Minerva, 2010). For the deterministic component we perform a predictive analysis while the stochastic components were included on the basis of stochastic simulation techniques and, finally, we obtain a predictive distribution. The approach has been applied on benchmark datasets (sun spot) and on financial time series obtaining encouraging results also for long predictive horizon (t+20). The Evolutionary Wavelet Network has been compared to other linear (AR) and non linear models (Wavelet, Wavelet Network, Back Propagation Network) obtaining a simpler but more efficient predictive model. References in Wavelet Analysis and Multirate System. Minerva, T., Poli, I., Paterlini S., (1999), Algoritmi Ibridi per l Analisi di Serie Storiche Finanziarie. Integrazione di Reti Neurali e Algoritmi Genetici per la previsione dell andamento del contratto Future FIB30, Scienza e Business, 6. Minerva, T., and Galeotti, G., (2000), Algoritmi ibridi per l ottimizzazione di un Portafoglio Azionario.Simulazione filtrata mediante wavelet decomposition, Materiali di Discussione del Dipartimento di Economia Politica, Università di Modena e Reggio Emilia, n. 312. Ogden, R.T. (1997), Essential Wavelets for Statistical Applications and Data Analysis, Birkhauser, Berlin. Reyneri, L.M., (1999),Unification of Neural and Wavelet Networks and Fuzzy Systems, IEEE Transactions on Neural Networks, 10, 4. Antoniadis, A., and Oppenheim, G., (1995), Wavelet and Statistics, Lecture Notes in Statistics 103, Springer-Verlag. Cao L., et al., (1995), Predicting chaotic Time Series with Wavelet Networks, Physica D, 85. Chui, C. K. (1997), Wavelet: a Mathematical Tool for Signal Processing, SIAM monographs on mathematical modeling and computation, Philadelphia. Coifman, R. R., Mayer, Y., and Wickerhauser, M.V., (1992), Wavelet Analysis and Signal Processing, in M.B. Ruskai et al Eds., Wavelets and their applications, pp. 153-178, Jones and Bartlett. Donoho, D.L., and Johnstone I.M. (1994), Ideal Spatial Adaptation by Wavelet Shrinkage, Biometrika, 81, 425-455. Donoho, D.L., and Johnstone I.M. (1995), Adapting to Unknown Smoothness via Wavelet Shrinkage, Journal of American Statistical Association, 90, 1200-1224. Donoho, D.L., and Johnstone I.M. (1998), Minimax Estimation via Wavelet Shrinkage, The Annals of Statistics, 26, 879-921. Holmes, C.C., and Mallick, B.K. (2000), Bayesian Wavelet Networks for Non Parametric Regression, IEEE Transaction on Neural Networks, 11, 1, 2000. Huang, H.C., and Cressie, N., (2000), Deterministic/Stochastic Wavelet Decomposition for Recovery of Signal From Nisy Data, Technometrics, August 2000, Vol. 42, N. 3. Kahane, J.P., and Lamarié, P.G., (1995), Fourier Series and Wavelets, Studies in Development of Modern Mathematics, vol. 3, Gordon and Research Publishers. Minerva, T. (2010), Wavelet Filtering for Prediction in Statistical Time Series Analysis, to be published ISSN: 1790-2769 99 ISBN: 978-960-474-189-2