DETECTING CHANGE IN NONLINEAR DYNAMIC PROCESS SYSTEMS

Size: px
Start display at page:

Download "DETECTING CHANGE IN NONLINEAR DYNAMIC PROCESS SYSTEMS"

Transcription

1 DETECTING CHANGE IN NONLINEAR DYNAMIC PROCESS SYSTEMS - Leon Christo Bezuidenhout - Thesis submitted in partial fulfilment of the requirements for the degree Master of Science in Engineering (Chemical Engineering) in the Department of Process Engineering at the University of Stellenbosch Study Leader: Prof Chris Aldrich -March 4-

2 DECLARATION I, the undersigned, hereby declare that the work contained in this dissertation is my own original work and that I have not previously in its entirety or in part submitted it at any university for a degree. Signature: Date:

3 SUMMARY As result of the increasingly competitive performance in today s industrial environment, it has become necessary for production facilities to increase their efficiency. An essential step towards increasing the efficiency of these production facilities is through tighter processes control. Process control is a monitoring and modelling problem, and improvements in these areas will also lead to better process control. Given the difficulties of obtaining theoretical process models, it has become important to identify models from process data. The irregular behaviour of many chemical processes, which do not seem to be inherently stochastic, can be explained by analysing time series data from these systems in terms of their nonlinear dynamics. Since the discovery of time delay embedding for state space analysis of time series, a lot of time has been devoted to the development of techniques to extract information through analysis of the geometrical structure of the attractor underlying the time series. Nearly all of these techniques assume that the dynamical process under question is stationary, i.e. the dynamics of the process did not change during the observation period. The ability to detect dynamic changes in processes, from process data, is crucial to the reliability of these state space techniques. Detecting dynamic changes in processes is also important when using advanced control systems. Process characteristics are always changing, so that model parameters have to be recalibrated, models have to be updated and control settings have to be maintained. More reliable detection of changes in processes will improve the performance and adaptability of process models used in these control systems. This will lead to better automation and enormous cost savings. This work investigates and assesses techniques for detecting dynamical changes in processes, from process data. These measures include the use of multilayer perceptron (MLP) neural networks, nonlinear cross predictions and the correlation dimension statistic. I

4 The change detection techniques are evaluated by applying them to three case studies that exhibit (possible) nonstationary behaviour. From the research, it is evident that the performance of process models suffers when there are nonstationarities in the data. This can serve as an indication of changes in the process parameters. The nonlinear cross prediction algorithm gives a better indication of possible nonstationarities in the process data; except for instances where the data series is very short. Exploiting the correlation dimension statistic proved to be the most accurate method of detecting dynamic changes. Apart from positively identifying nonstationary in each of the case studies, it was also able to detect the parameter changes sooner than any other method tested. The way in which this technique is applied, also makes it ideal for online detection of dynamic changes in chemical processes. II

5 OPSOMMING Dit is belangrik om produksie aanlegte so effektief moontlik te bedryf. Indien nie, staar hulle die moontlikheid van finansiële ondergang in die gesig veral as gevolg van toenemende mededinging die industrie. Die effektiwiteit van produksie aanlegte kan verhoog word deur verbeterde prosesbeheer. Prosesbeheer is n moniterings en modellerings probleem, en vooruitgang in hierdie areas sal noodwendig ook lei tot beter prosesbeheer. Omdat dit moeilik is om teoretiese proses modelle af te lei, word dit al hoe belangriker om modelle vanuit proses data te identifiseer. Die ongewone optrede van baie chemiese prosesse, wat nie inherent stogasties blyk te wees nie, kan meestal verklaar word deur tydreeks data vanaf hierdie prosesse te analiseer in terme van hul nie-liniêre dinamika. Sedert die ontdekking van tydreeksontvouing vir toestandveranderlike stelsels, is baie tyd daaraan spandeer om tegnieke te ontwikkel wat inligting uit tydreekse kan onttrek deur die onderliggende geometriese struktuur van die attraktor te bestudeer. Byna al hierdie tegnieke aanvaar dat die dinamiese proses stationêr is, m.a.w dat die dinamika van die proses nie verander het tydens die observasie periode nie. Die vermoë om hierdie dinamiese proses veranderinge te kan identifiseer, is daarom baie belangrik. Ook in gevorderde beheerstelsels is vroegtydige identifisering van dinamiese veranderinge in prosesse belangrik. Proses karakteristieke is altyd besig om te verander, sodat model parameters herkalibreer moet word, modelle opgedateer moet word en beheer setpunte onderhou moet word. Meer betroubare tegnieke om veranderinge in prosesse te identifiseer sal die aanpasbaarheid van proses modelle in hierdie beheerstelsels verbeter. Dit sal lei tot beter outomatisering en sodoende lei tot enorme kostebesparings. Hierdie werk ondersoek tegnieke om dinamiese veranderinge in prosesse te identifiseer, deur die analise van proses data. Die tegnieke wat gebruik word sluit die volgende in: III

6 multilaag-perseptron neurale netwerke, nie-liniêre kruisvoorspelling statistieke en die korrelasie dimensie statistiek. Die tegnieke is op drie gevallestudies toegepas om te sien of hulle die dinamiese veranderinge in die data kan identifiseer. Vanuit die navorsing is dit duidelik dat proses modelle nadelig beinvloed word deur niestationêre data. Dit kan dien as n indikasie van veranderinge in die proses parameters. Die nie-liniêre kruisvoorspellings algoritme gee n beter indikasie van dinamiese veranderinge in die proses data, behalwe waar die tydreeks baie kort is. Toepassings van die korrelasie dimensie statistiek gee die beste resultate. Hierdie tegniek kon dinamiese veranderinge vinniger as enige ander tegniek identifiseer, en die manier waarop dit gebruik word maak dit ideaal vir die identifisering van dinamiese veranderinge in chemiese prosesse. IV

7 TABLE OF CONTENTS SUMMARY... I OPSOMMING... III LIST OF FIGURES...VII LIST OF TABLES...XII ACKNOWLEDGEMENTS... XIII 1 INTRODUCTION MOTIVATION GOALS, SCOPE AND APPROACH THESIS LAYOUT STATE SPACE ANALYSIS OF TIME SERIES DYNAMICAL SYSTEMS LINEAR SYSTEMS NONLINEAR SYSTEMS AND CHAOS STATE-SPACE RECONSTRUCTION Embedding Theorems Estimating Suitable Reconstruction Parameters DIMENSION ESTIMATES: INVARIANTS OF THE DYNAMICS The Box-counting Dimension The Information Dimension The Correlation Dimension SURROGATE DATA ANALYSIS Hypothesis testing The test statistic DETECTING DYNAMIC CHANGE MEASURES FOR DETECTING NONSTATIONARITY Model-Based Change Detection Probing Nonstationarity Using Nonlinear Cross Prediction Exploiting the Correlation Dimension as a Test for Nonstationarity CHANGE DETECTION METHODOLOGY...3-

8 4 CASE STUDIES AUTOCATALYTIC REACTOR State Space Reconstruction of the Autocatalytic System Surrogate Data Analysis for the Autocatalytic System Modelling the Autocatalytic System Detecting Dynamic Change in Autocatalytic Reactor Nonlinear Cross Prediction Detecting Dynamic Change in the Autocatalytic Reactor Correlation Dimension The Effect of Noise in the Autocatalytic Process Data THE BAKER S MAP State Space Reconstruction of the Baker s Map Surrogate Data Analysis for the Baker s Map Modelling the Baker s Map Detecting Dynamic Change in the Baker s Map Using Nonlinear Cross Prediction Using the Correlation Dimension to Detect Dynamic Change in the Baker s Map REAL DATA FROM A METAL LEACHING PLANT State Space Reconstruction of Metal Leaching Data Surrogate Data Analysis of the Metal Leaching Data Modelling the Metal Leaching Data Detecting Dynamic Change in Metal Leaching Data - Nonlinear Cross Prediction Detect Dynamic Change in Metal Leaching Data Correlation Dimension CONCLUSIONS REFERENCES APPENDIX... A A. STATE SPACE RECONSTRUCTION USING TIME DELAY EMBEDDING - A NUMERICAL EXAMPLE.A

9 LIST OF FIGURES Figure 1.1: Simplified Schematic of the Structure of Adaptive Predictive Controllers Figure.1: Illustration of a three dimensional attractor, where x,x,x 1 3 are the three variables governing the system. Figure.: Plot of x and y vs. time showing periodic nature of the solution Figure.4: Plot of x vs. y showing the closed circular orbit of the attractor. Figure.4: a) Conventional concept of system behaviour; b) Actual system behaviour Figure.5: Chaotic x,y signals of the Hênon-map. Figure.6. Chaotic attractor of Hênon-map Figure.7: Example of a chemical reactor to explain the reasoning behind state space reconstruction. Figure.8: Illustration of a) too small, b) too large and c) optimal time delay. Figure.9: Probing hypersphere on the attractor Figure.1: The log(ε)-log(c N ) plot for determining the correlation dimension via the Grassberger-Procaccia algorithm. Figure.11(a): A reconstructed attractor from Chua s circuit that seem low dimensional from a distance. Figure.11(b): Zooming in on part of the attractor reveals the high dimensional nature of the object. Figure.1: A typical graph illustrating Judd s method where d c is a function of ε. Figure.13: Irregular output from a simple linear process Figure.14: Irregular output from linear process observed through a nonlinear measurement function Figure.15: Two data sets that need to be classified. Figure.16: Correlation dimension plot of surrogates and original data for data set A. Figure.17: Correlation dimension plot of surrogates and original data for data set B. VII

10 Figure 3.1: Lorenz attractor for t = :1.7 Figure 3.: Lorenz attractor for t = : Figure 3.3: Architectural graph of a multiple-layer perceptron neural network with two hidden layers. Figure 3.4: Illustration of the directions of two basic signal flows in a multiple-layer perceptron: forward propagation of function signals and back propagation of error signals. Figure 3.5: Illustration of the model-based approach to detect nonstationarity Figure 3.6: Example of a typical 3-D surface plot for mutual cross-prediction errors. Figure 3.7: Example of a typical -D colour-coded mutual prediction map Figure 3.8: Illustration d ( ) c ε -curves from two halves of a time series. Figure 3.9: Illustration of d ( ) c ε -curves for segments of a time series. From the figure it is clear that there was a change in process parameters between segment and segment 3. It is suggested by the shift of the curves, as indicated by the green arrows. Figure 3.1: Illustration of the moving window approach. This approach is ideal for online monitoring of changing parameters. Figure 3.11: A flow diagram of the approach to detect parameter change. Figure 4.1: Schematic illustration of the autocatalytic reactor. Figure 4.(a): All 3 data points from the nonstationary time series generated by the autocatalytic reaction. Figure 4.(b): Data points 95 to 15 from the nonstationary time series generated by the autocatalytic reaction. Figure 4.3: The average mutual information, as function of the time delay, for the autocatalytic reaction. Figure 4.4: Autocorrelation function as statistic to determine time delay for the autocatalytic reaction. Figure 4.5: Fraction of FNN as function of embedding dimension for the autocatalytic reaction. Figure 4.6: Reconstructed attractor for the autocatalytic reaction system VIII

11 Figure 4.7: Correlation dimension curves for surrogates and actual data for autocatalytic reaction. Figure 4.8: History of the moving global minimum of the Schwartz Information Criterion versus the number of hidden nodes for the autocatalytic system Figure 4.9: Free-run prediction of data points 5-5. Figure 4.1: Plot of predicted values versus actual values (5-5), and the residuals. Figure 4.11: Free-run prediction of data points Figure 4.1: Plot of predicted values versus actual values (15-15), and the residuals. Figure 4.13: Free-run prediction of data points 5-5. Figure 4.14: Plot of predicted values versus actual values (5-5), and the residuals. Figure 4.15: 3-D surface plot of mutual cross-prediction errors for the autocatalytic reaction. Figure 4.16: -D colour-coded mutual prediction map for autocatalytic reaction. Figure 4.17: Attractors form by the two different parameters sets are embedded into each other. Figure 4.18: d ( ) c ε -curves from the two halves of the autocatalytic time series. Figure 4.19: d ( ) c ε -curves from six segments, each containing 5 data points, of the autocatalytic time series. Figure 4.: Calculating d ( ) c ε -curves from a moving window for the autocatalytic reaction time series. Figure 4.1: Data points 975 to 15 from the nonstationary time series generated by the autocatalytic reaction noisy data. Figure 4.: The average mutual information, as function of the time delay, for the autocatalytic reaction noisy data. Figure 4.3: Fraction of FNN as function of embedding dimension for the autocatalytic reaction noisy data. Figure 4.4: Reconstructed attractor for the autocatalytic reaction system, projected onto the first three principal components noisy data. IX

12 Figure 4.5: Correlation dimension curves for surrogates and actual data for autocatalytic reaction noisy data. Figure 4.6: Free run prediction on the first part of the time series where the process parameters were unchanged noisy data Figure 4.7: Free run prediction on the last part of the time series where the process parameters had already changed noisy data Figure 4.8: 3-D surface plot of mutual cross-prediction errors for the autocatalytic reaction noisy data. Figure 4.9: -D colour-coded mutual prediction map for autocatalytic reaction noisy data. Figure 4.3: Calculating d ( ) c ε -curves from a moving window for the autocatalytic reaction time series noisy data. Figure 4.31(a): Time series data for the nonstationary baker s map all 4 data points Figure 4.31(b): Time series data for the nonstationary baker s map data points 45 to 55. Figure 4.3: Reconstructed attractor for the baker s map. Figure 4.33: d ( ) c ε -curves for surrogate and actual data when embedding into 1 dimensions. Figure 4.34: History of the moving global minimum of the Schwartz Information Criterion versus the number of hidden nodes for the baker s map. Figure 4.35: Free-run prediction of model training data for baker s map. Figure 4.36: One step prediction for data points 77 to 775 to illustrate the model fitness. Figure 4.37: 3-D surface plot of mutual cross-prediction errors for the baker s map Figure 4.38: -D colour-coded mutual prediction map for the baker s map Figure 4.39: d ( ) c ε -curves from the two halves of the baker s map time series. Figure 4.4: d ( ) c ε -curves calculated from eight consecutive segments, each containing 5 data points, of the baker s map time series X

13 Figure 4.41: Calculating d ( ) c ε -curves from a 5 point moving window for the baker s map. Figure 4.4: Calculating d ( ) c ε -curves from a point moving window for the baker s map. Figure 4.43: Data from a metal leaching plant. Figure 4.44: Linearly adjusted data from the metal leaching plant Figure 4.45: Autocorrelation function of the metal leaching data Figure 4.46: Eigenvalues of the covariance matrix of the metal leaching data Figure 4.47: Reconstructed attractor for the metal leaching data, projected onto the first three principal components. The variance explained by each principal component is given in brackets. Figure 4.48: Correlation dimension curves of surrogate and actual metal leaching data Figure 4.49: SIC for modelling the metal leaching data. Figure 4.5: Predicted and observed values for on-step prediction of metal leaching data. Figure 4.51: d ( ) c ε -curves for the two halves of the metal leaching data. Figure 4.5: d ( ) c ε -curves from moving window of size 8 for metal leaching data. XI

14 LIST OF TABLES Table 4.1: R - statistic results for one-step prediction of different segments from the baker s map time series. Table 4.: R values for one-step prediction of metal leaching data. Table 4.3: Cross prediction errors from the four segments of the metal leaching data. XII

15 ACKNOWLEDGEMENTS 1) My study leader, Prof Chris Aldrich. ) Juliana Steyl, the secretary of Prof Chris Aldrich, for all the administrative work. 3) My fellow post graduate students, especially Gorden, for helping me to understand some of the theory behind nonlinear time series analysis. 4) JP Barnard and Prof C. Aldrich for using the Quickident Toolbox (Barnard & Aldrich, ) to do most of the data classification and state space reconstruction calculations. 5) Kevin Judd for using his code to do the correlation dimension calculations. 6) Rainer Hegger, Holger Kantz and Thomas Schreiber for using the nonlinear cross prediction error algorithm from the TISEAN package. 7) Macia, for always being supportive and understanding. XIII

16 1 INTRODUCTION There is no doubt that quality has become a major feature in the survival plan of companies. With diminishing markets resulting from the improved competitive performance in today s industrial environment, it is clear that unless there is a definite commitment towards increasing the efficiency of production facilities, they will loose their competitive edge. This will ultimately lead to elimination and the resultant harsh realities that come with it for the employees. Improving the efficiency of production facilities can have a positive impact on chemical processes in several ways: Improvement in the quality of the final product Increase in production Decrease in the generation of hazardous wastes. The first two issues are probably most important for a company and could mean the difference between success and failure. In previous times, decreasing hazardous waste was usually not considered a main objective. Recently, however, production facilities have come under pressure to comply with increasingly stringent environmental regulations, which makes this an import consideration. For the production facility, these pressures translate into a continuous effort to reduce process variability and maintain stability. This is usually accomplished through tighter process control, and is the reason why so much time and effort goes into developing advanced control systems. One of the major problems with the use of advanced control systems is that process characteristics are always changing, so that model parameters have to be recalibrated, models have to be updated and control settings have to be maintained. More reliable detection of changes in processes will not only improve the performance and adaptability of process models used in control systems, but will also lead to faster

17 INTRODUCTION 1- detection of dynamic changes and/or process upsets by monitoring these processes. This will lead to better automation and will result enormous cost savings. The overall goal of this work is to investigate and develop suitable measures for detecting dynamic changes in processes from process data, and evaluate their performance. 1.1 Motivation The degree of automation in chemical process control plants is still low. Many of the processes involve human operators. These operators are able to cope with the complexities of systems by applying their know-how skills acquired over years of hands-on operation of the plant. Given the complexity of the system and vagueness of situations, human operators are not always able to cope with the problem and provide good control performance. Control of manufacturing processes that are based on chemical reactions is often difficult in practice. Among the reasons are the nonlinear behaviour of such systems, large dead-times, and sometimes there are many conflicting goals within the system. The control of chemical processes is a process monitoring and modelling problem. Improving monitoring and modelling techniques, will necessarily also lead to improved process control. Process monitoring: Through process monitoring it is possible to detect failing sensors and process upsets, which are important for reasons of safety and process efficiency. Process control based on faulty sensors is inefficient and can lead to unsafe operating conditions. Process upsets or disturbances can also lead to operating inefficiencies. An integral aspect of the overall process control problem is therefore timely identification of failed sensors and upsets. Apart from detecting failing sensors and process upsets, one also want to detect less trivial changes in the process dynamics. Doing so will further improve process monitoring, as it allows the identification of different dynamic operating regions in the process which can each be exploited separately.

18 INTRODUCTION 1-3 Recent advances in process instrumentation and data collection techniques resulted in an increase in the amount of data recorded from chemical processes. Processes typically get more heavily instrumented, and measurements are done more extensively and more frequently. The result is a huge amount of process data. An increase in process data alone do not guarantee a better understanding of the process, and can be rather overwhelming and confusing. Much of the data may be redundant due to the high correlation of the measured variables. The difficulty lies in extracting only the most significant information from the plethora of the data. Unless this is done, monitoring and modelling the process can become problematic as result of the magnitude of data. This is also known as information overload. Ironically, most of the information is usually about less important process variables, and for the vital process variables (the ones that govern the process) there may be a lack of data. Part of the research involves the use of advanced data analysis techniques to overcome these common problems. Assessment of the validity of process models: Modelling has become an increasingly important aspect of chemical process control. This is evident by the progression of controller design methods. The modern era of process control started by applying PID (Proportional + Integral + Derivative) controllers, also known as conventional controllers. PID controllers account for more than 8% of installed automatic feedback control devices in the process industries (Willis & Tham, 1994). There are several drawbacks using PID controllers, as they tend to operate within a limited operating range and are only optimal for a second order linear process without time delays. In practise, process characteristics are usually nonlinear and can change with time. PID controllers therefore do not cope well with nonlinearity in processes or changes in process parameters. Recent advances in process control algorithms, particularly the model based controller design methods, have further increased the reliance on process models. Controller design advanced to using process models in the actual setting of the PID tuning parameters by implementing the controller within an adaptive framework (Willis & Tham, 1994). In this setup (Figure 1.1), the parameters of a model are updated regularly to reflect the current process characteristics. The controller settings can be updated continuously according to changes in the process characteristics. To ensure

19 INTRODUCTION 1-4 optimal process control, rapid and reliable detection of changes in the process characteristics is vital. This is one area where improved change detection techniques will be particularly valuable. DESIRED OUTPUT Σ CONTROLLER PROCESS PROCESS OUTPUT MODEL BUILDER CALCULATE CONTROLLER PARAMETERS PROCESS OUTPUT PREDICTER PREDICTED OUTPUT ADAPTIVE PART PREDICTIVE PART Figure 1.1: Simplified Schematic of the Structure of Adaptive Predictive Controllers The model-based control strategy that has been most widely applied in the process industries is model predictive control (MPC) (Perry & Green, 1997). One major advantage of MPC is that it can accommodate difficult or unusual dynamic behaviour such as large time delays, nonlinearities and inverse responses. It is also well-suited for difficult multi-input, multi-output control problems where there are significant interactions between the manipulated inputs and the controlled outputs. A key feature of MPC is that future process behaviour is predicted using a dynamic model and available measurements from the process. The controller outputs are calculated to minimise the difference between the predicted process response and the desired response (Figure 1.1). At each sampling instance, the control calculations are repeated and the predictions updated based on current measurements from the process. MPC has the potential to provide perfect automatic control that is if an ideal model of the process exists (Willis & Tham, 1994). This is why the critical factor in the successful

20 INTRODUCTION 1-5 application of MPC (or any model-based technique for that matter) is the availability of a suitable dynamic model. A model is nothing more than a mathematical abstraction of a real process (Seborg et al., 1989). The equations that comprise the model are at best an approximation to the true process. The model, therefore, cannot incorporate all of the features, both macroscopic and microscopic, of the real process. Depending on how they are derived, there are three different model classifications: Theoretical models developed using the principles of chemistry and physics. Empirical models obtained from a statistical analysis of process data Semi-empirical models that are a compromise between theoretical and empirical models Theoretical models have several advantages over statistical/empirical models. They can often be extrapolated over a wider range of operating conditions than purely statistical/empirical models, which are only accurate over a limited range. Theoretical models also provide the capability to infer how unmeasured or unmeasurable process variables vary as the process operating conditions change, which gives a deeper understanding of the process. The engineer normally has to seek a compromise involving the cost of obtaining the model, that is, the time and effort required to obtain and verify it. To establish a useful theoretical model, specialized knowledge about the system under study is needed. Theoretical models have parameters and fundamental relations that need to be evaluated from physical experiments. This can be time consuming and expensive. Furthermore, theoretical models are often compromised because they require so many simplifying assumptions in order to be tractable that they are often biased. Given all the difficulties of obtaining theoretical models, it has become increasingly important to identify models from process data. Such models, which simply describe the functional relationships between the system inputs (input space) and system outputs (output space), are referred to as black box models (Juditsky et al., 1995). Although the parameters of these models do not have any physical significance in terms of equivalence to process parameters, such as heat or mass transfer coefficients and

21 INTRODUCTION 1-6 reaction kinetics, the aim is merely to represent trends in the process behaviour faithfully. A problem frequently encountered when attempting to build black box models, is the availability of process data for all the variables in the input space. All the relevant input variables necessary to build the model are either just not recorded (probably as result of the costs involved), or are impossible to record. Most of the time, the engineer s only source of process data is that single output variable that needs to be controlled. Although such a modelling exercise might seem fundamentally flawed, a technique, called state space reconstruction 1, does exist to reconstruct the input space from this one-dimensional output. In this approach, the process data are viewed as a time series. Unlike the analyses of random samples of observations that are discussed in the context of most other statistics, the analysis of time series data is based on the assumption that successive values in the process data represent consecutive measurements taken at equally spaced time intervals. The main assumption underlying time series analysis is that the properties or parameters describing the data, i.e. the dynamics governing the process, are constant (Schreiber, 1997). In other words, the dataset must be stationary. In practice, this is rarely true. Process parameters are often subject to changes at unknown time instants. Process parameters can also slowly drift with time, which is a signature of the existence of nonstationarity. Nonstationarities in the data will therefore complicate time series analysis and can result in a process model that is inferior or unreliable. The overall process model can be improved by detecting dynamic changes in the process data and dividing it into stationary segments that can each be modelled separately. These submodels can be incorporated into a control strategy where the process is continuously being monitored for dynamic changes and where, depending on the process parameters, a different model is used for control. Once again, reliable methods that are sensitive to dynamic changes are vital for such an implementation. 1 The theory behind state space reconstruction is thoroughly explained in Chapter.

22 INTRODUCTION Goals, Scope and Approach The problem of detecting changes in properties of signals and systems has received increasing attention in the last twenty or so years. Most of the work, however, has been done on linear stochastic systems (Basseville & Nikiforov, 1993). Change detection for nonlinear dynamical systems has not been investigated in any depth. This was largely due to a historically poor understanding of the subject. In the last few years, with computers being more accessible and processing power becoming increasingly cheaper, methods for analyzing nonlinear time series have seen dramatic progress. These advances now make it possible to investigate more complex areas of interest, such as detecting dynamic changes in nonlinear systems, in much more detail. The main objective of this work is to detect changes (or nonstationarities) in dynamic process systems from process data. It includes changes that can occur both relatively fast or relatively slow with respect to the sampling period of the measurements. These changes by no means imply changes which are large in magnitude. The key difficulty is to detect intrinsic changes that are not necessarily directly observed and that are measured together with other types of perturbations. A key part of the research is to identify suitable methods for detecting changes in process systems, and evaluate their ability to detect changes quickly and reliably. The research is focused on nonlinear dynamic systems that exhibit deterministic behaviour, or at least mixed systems that have a dominant deterministic part. The theory behind nonlinear time series analysis forms an integral part of the research, and most of the work is done within this context. The change detection problem is approached by first classifying the data. It is important to determine whether the time series data are linear or nonlinear, and whether it exhibits stochastic or deterministic behaviour. This classification is a complicated part of time series analysis. Usually the data do not belong purely to a specific class and are therefore classified according to the degree of its behaviour. Surrogate data analysis techniques will be used to classify the system. The outcome of this classification will indicate whether dynamic change detection techniques, as discussed within the context of this work, need to be applied, or whether traditional linear statistics would be sufficient to detect changes in the process. The next step will be to apply the identified

23 INTRODUCTION 1-8 change detection techniques to various case studies consisting of simulated, as well as actual data from chemical processes. The techniques will then be evaluated on their ability to detect different kinds of dynamic change, as well as the time in which they could detect the dynamic change. 1.3 Thesis Layout All the relevant issues concerning the state space analysis of time series are discussed in Chapter. This includes the theory behind state space reconstruction, surrogate data analysis, as well as various other advanced analysis techniques and statistics that are needed to quantify dynamic behaviour in processes from time series data. This theory is fundamental to the research and receives a great deal of attention. Chapter 3 focuses on the issue of dynamic changes in processes, and measures for detecting these changes are identified and discussed. In Chapter 4, the change detection methodology is put into practice and applied to three case studies. The selected change detection techniques are investigated in depth in this section. The conclusions and limitations of the change detection methodology, based on these case studies, are discussed in Chapter 5.

24 STATE SPACE ANALYSIS OF TIME SERIES Time series data are sequences of measurements that follow non-random orders. Unlike the analyses of random samples of observations that are discussed in the context of most other statistics, the analysis of time series is based on the assumption that successive values in the data represent consecutive measurements, usually taken at equally spaced time intervals. There are two main goals for analyzing time series data: Identifying the nature of the phenomenon represented by the sequence of observations Forecasting predicting future values of the time series variables. Both these goals require that the pattern of the observed time series data is identified and formally described. Once the pattern is established, it can be interpreted and integrated with other data (Statsoft, 1984)..1 Dynamical Systems Time series analysis, and especially nonlinear time series analysis, are motivated and based on the theory of dynamical systems; that is, the time evolution is defined in some phase space (Kantz & Schreiber, 1997). Since these systems can exhibit deterministic chaos 1, this is a natural starting point when irregularity is present in a signal. A purely deterministic system is defined as a system where, once its present state is fixed, the states at all future times are determined as well. It is therefore essential to establish a vector space for the system. Such a state space or phase space specifies the 1 Chaos Irregular but deterministic motion which is characterized by a continuous, broadband Fourier spectrum. Possible only in a three-or-more dimensional nonlinear system of differential equations or a two-or-more dimensional nonlinear discrete time map. This nonlinear motion is slightly predictable, non-periodic and sensitive to changes in initial conditions. (Abarbanel, 1998)

25 STATE SPACE ANALYSIS OF TIME SERIES state of the system by specifying a point in the system. The construction of a state space makes it possible to study the dynamics of the system by studying the dynamics of the corresponding state space points. In theory, dynamical systems are generally defined by a set of first order ordinary differential equations acting on a state space. If certain conditions are met, the mathematical theory of ordinary differential equations ensures the existence and uniqueness of the trajectories in the state space. In a deterministic dynamical system, the state space is assumed to be a finite- d d dimensional vector space, R, where a state is specified by a vector x R general form for an autonomous discrete-time dynamical system is the map. The x n+ 1 = F( xn), n (.1) where x d, x 1 and : d d F R R is a diffeomorphism 3. In the second case, n n R time is a continuous variable: d x() t = f(, t x ()) t, t (.) dt This form is normally referred to as flow. The flow is called autonomous if f does not explicitly depend on time (t ). A sequence of points ( x or x () t ) solving the above equations is called a trajectory of n the dynamical system, where the initial condition is x() = x. Depending on the initial condition and the form of F (or f ), the trajectory will either run away to infinity as time proceeds or stay in a bounded area forever. In a dissipative 4 system, the points visited by the system after transient behaviour has died out will be concentrated on a subset of state space. The geometric object to which the trajectory orbits go in time is called the system attractor 5. Figure.1 gives an illustration of a typical attractor. In this particular case, the attractor is three dimensional and each axis represents one of the variables by which the system is described. Trajectory The path that a signal follows through state space 3 Diffeomorphism A smooth mapping with a smooth inverse. 4 Dissipative system A system with sources and sinks of energy 5 Attractor The set of points in state space visited by a signal trajectory after the transients are gone

26 STATE SPACE ANALYSIS OF TIME SERIES 3 It is important to note that even for non-deterministic systems the concept of the state of a system is still very powerful. x 3 trajectory attractor x 1 x Figure.1: Illustration of a three dimensional attractor, where x,x,x 1 3 are the three variables governing the system.. Linear Systems Consider the case where the flow in Equation (.) is linear. An autonomous linear system of f degrees of freedom, x ( t) = [ x1( t), x( t),..., xf ( t)], would yield the following equation (Abarbanel, 1993): d(t) x = A. x ( t) (.3) dt

27 STATE SPACE ANALYSIS OF TIME SERIES 4 Where A is a constant f f matrix. The possible courses of evolution of Equation (.3) are characterized by the eigenvalues of the matrix A. The trajectories of the solution behave in one of the following ways: 1) Directions in f-space along which the orbits shrink to zero namely, directions along which the real part of the eigenvalues of A are negative, or ) Directions along which the orbits unstably grow to infinity namely, directions along which the real part of the eigenvalues are positive, or 3) Directions where the eigenvalues occur in complex-conjugate pairs along with zero or negative real part. When the eigenvalues have an absolute value of unity, the trajectories will follow closed circular or elliptical orbits (Honkela, 1). However, in any situation where the eigenvalues has a positive real part, it is an indication that the linear dynamics are incorrect, and one must go to the nonlinear equations that govern the process to make a better approximation to the dynamics (Abarbanel, 1993). The linear case can be illustrated by means of a simple example, using the following two coupled linear differential equations: dx dt = ω y, dy dt = ωx (.4) The equations have a periodic solution in the form (Kantz & Schreiber, 1997): x() t a.cos[ ( t t )] = ω, yt a ω t t () =.sin[ ( )] (.5) When inspecting Equation (.5), as well as the visual illustration thereof (Figure. and Figure.3), it is clear that the solution will stay finite forever. This demonstrates the fact that autonomous linear dynamic systems are too straightforward to describe any interesting phenomena. In practice, stable linear systems either exponentially converge to a constant value, or exhibit periodic behaviour. Linear time series analysis is therefore well defined (Statsoft, 1984), and established linear mathematical techniques sufficiently meet the requirements.

28 STATE SPACE ANALYSIS OF TIME SERIES 5 x Time y Time Figure.: Plot of x and y vs. time showing periodic nature of the solution y x Figure.4: Plot of x vs. y showing the closed circular orbit of the attractor.

29 STATE SPACE ANALYSIS OF TIME SERIES 6.3 Nonlinear Systems and Chaos While the possible dynamics of linear systems are rather restricted, even very simple nonlinear systems can have very complex dynamical behaviour. Previously people believed that only a stochastic 6 or noisy input to the system could create a stochastic output, and that only a deterministic input to a deterministic system created wellbehaved deterministic outputs. In addition, it was believed that a small change in the initial conditions of the dynamic equations created only a small change at any future time. It has now become common knowledge that this is not true for nonlinear dynamical systems (see Figure.4). Deterministic input to a deterministic dynamical system can create a stochastic or irregular noise-like chaotic output, and a small change in the initial conditions can lead to an entirely different output after some lapse of time. (Chang & Lee, 1996). stochastic input stochastic output SYSTEM deterministic input (a) SYSTEM Stochastic or deterministic output (b) Figure.4: a) Conventional concept of system behaviour; b) Actual system behaviour Not all irregular motions are chaos. Irregularity may be caused by some other underlying reasons. For a process to be chaotic it must satisfy a series of tests. The most important test is that the signal has a sensitive dependence on initial conditions. It 6 Stochastic - Stochastic is synonymous with random. Opposite of deterministic

30 STATE SPACE ANALYSIS OF TIME SERIES 7 should also be indecomposable or ergodic 7 ; and have an element of regularity (Eckmann & Ruelle, 1985). Chaos seems to be the rule, rather than the exception, in nature. It also frequently occurs in process systems engineering, where nonlinear complexity and a great number of describing equations govern the dynamics. This is especially true for chemical engineering, where chemical processes are inherently nonlinear and often have structurally unstable dynamics. Although chaotic behaviour often occurs in chemical engineering processes, it is usually attributed to process noise and treated as such. The statistical approach normally used in such situations is without doubt a powerful tool. However, in process systems engineering there has to be dealt with interactions of highly complex nonlinear phenomena with multiple units involving chemical reactions, heat and mass transfer, separations, fluid flow, etc. Since chaos can occur without any single noise input to any single unit, the complicated chaos resulting from all the individual units, integrated as a system, is just imaginable (Chang & Lee, 1996). Chaotic behaviour can be illustrated visually using the Hênon-map as an example: x = a + x + by, y = n+ 1 xn (.6) n 1 n n The map yields irregular solutions for many choices of a and b. Setting a = 1.4 and b =.3, for example, generates a typical sequence of x n that will be chaotic (Kantz & Schreiber, 1997). The irregular signals are shown in Figure.5. Especially note the interesting shape of the chaotic attractor (Figure.6). This illustration of the chaotic behaviour of the Hênon-map, verifies the earlier claim that even simple nonlinear systems can have very complex dynamical behaviour. It is therefore understandable that traditional linear analysis tools can usually not be applied to nonlinear, chaotic systems with great success. Doing so would be much harder and conclusion drawn from the results would be fundamentally limited in range. 7 Ergodic theory says that a time average equals a space average, where the weight with which the space average has to be taken is an invariant measure

31 STATE SPACE ANALYSIS OF TIME SERIES 8 1 X Y Time (steps) Figure.5: Chaotic x,y signals of the Hênon-map Y Figure.6: Chaotic attractor of Hênon-map X

32 STATE SPACE ANALYSIS OF TIME SERIES 9 This predicament started the era of nonlinear time series analysis. Although nonlinear data analysis techniques are somewhat unconventional and often more difficult to use, it definitely is a big improvement on traditional methods. Even though long-term prediction of chaotic systems is impossible, it is often still possible to predict statistical features of its behaviour and find certain invariant 8 features that provide a qualitative description of the system..4 State-Space Reconstruction Apart from nonlinear and chaotic behaviour, intelligent data analysis problems are often complicated by the following, seemingly contradictorily, situation (Bradly, 1996): simultaneous overabundance and lack of data. Take for example the following chemical reactor setup:? T A Reactor A? B? C? Figure.7: Example of a chemical reactor to explain the reasoning behind state space reconstruction. The dynamics of this system are governed by, amongst others, the temperature inside the reactor, the pressure inside the reactor and the concentration of the species A, B and C. If there were only a temperature sensor installed on the reactor, one would have loads of information available about the temperature inside the reactor, but no data about other important quantities such as species concentration or pressure. This is 8 Invariant - Unchanged by a specified transformation or operation

33 STATE SPACE ANALYSIS OF TIME SERIES 1 usually the case when the other system properties are not sensor-accessible or hard to measure with available sensors. When doing intelligent data analysis, the analyst is therefore often required to extract meaningful conclusions about a complicated system using data from a single sensor. At first glance, the data analysis procedure would appear fundamentally limited. Probably the most powerful, and widely used, way of getting around this problem is a technique called time delay embedding. Time delay embedding is a method of reconstructing the internal dynamics of a complicated (nonlinear) system from a single time series. That is, time delay embedding can often be used to infer important information about unmeasurable variables, such as the species concentration or reactor pressure in Figure.7, from a single measured variable, e.g. the reactor temperature. The reconstruction produced by time delay embedding is not completely equivalent to the internal dynamics of the system in all situations. However, a properly done single sensor reconstruction can be extremely useful, because the results are guaranteed to be topologically (i.e. qualitatively) identical to the true internal dynamics of the system, and therefore the dynamical invariants must also be similar. This means that conclusions drawn from the reconstructed dynamics are also true of the internal unmeasured dynamics inside the black box..4.1 Embedding Theorems The solution to the problem of how to go from scalar observations to multivariate state space is contained in the geometric theorem, called the embedding theorem, attributed to Takens (1981): Let M be a smooth ( C ) m-dimensional manifold that constitutes the original state t space of the dynamical system under investigation, and let φ : M M the corresponding flow. Suppose that it is possible to measure some scalar quantity s() t = h( x ()) t that is given by the measurement function h: M R t where x() t = φ ( x ()). It is then possible to construct a delay coordinate map that maps a state x from the original state space M to a reconstructed state space de R :

34 STATE SPACE ANALYSIS OF TIME SERIES 11 de F : M R (.7) x y = F( x ) = ( st ( ), st ( + t), st ( + t),..., st ( + ( d 1) t) l l e l Here d e is the embedding dimension and t l is the lag or time delay used. Takens proved that for d m+ 1 it is a generic property of F to be an embedding of M in e F M F M R is a de e R, that is : ( ) d C diffeomorphism. Generic implies that the subset of pairs ( ht, l ) is an open and dense subset in the set of all pairs( ht, l ). The theorem was later generalized by Sauer, York and Casdagli (1991) in two ways: 1) They replaced the condition d m+ 1 by d d( A), where d ( ) A is the box-counting dimension of the attractor A M. e e ) They replaced the term generic by the term prevalent, which means that almost all ( ht, l ) will give an embedding. The first improvement is great progress for experimental systems that have a lowdimensional attractor (e.g. d ( ) 5 A < ) in a very high dimensional (e.g. m=1) space. In this case Takens theorem says that only for very large embedding dimensions (e.g. de 1) diffeomorphic equivalence can be expected, whereas Sauer et al. (1991) says that a small ( d e > 1) embedding dimension will be adequate. The second modification was necessary because examples of open and dense (i.e. generic) sets were found that were rather thin (Sauer et al., 1991). They also showed that for dimension estimation an embedding dimension d > d ( ) A should be enough. e Assume that the scalar signal s() t = h( x ()) t is sampled with a sampling time, t s. The resulting time series { s n } with sn = snt ( s) is used to reconstruct the states y = ( s, s, s,..., s ) (.8) n n n+ l n+ l n+ ( d 1) l e

35 STATE SPACE ANALYSIS OF TIME SERIES 1 for n= 1,..., N. The symbol l represents the delay time (lag) in units of the sampling time, t l = lt (Parlitz, 1995). s The y n replace the scalar data measurements s n with data vectors in a Euclidian d- dimensional space in which the invariant aspects of the sequence of points x n are captured with no loss of information regarding to the properties of the original system. The newly reconstructed space is related to the original x n - space by smooth, differential transformations. The smoothness is necessary in allowing the demonstration that all the invariants of the motion in the reconstructed space are the same as if they were evaluated in the original space. This suggests that just as much can be learned about a system studying the reconstructed state space, than studying the original (true) state space (Abarbanel, 1996). And herein lies the significance of state space reconstruction. A numerical example explaining the theory behind time delay embedding is given in Appendix A..4. Estimating Suitable Reconstruction Parameters State space reconstruction depends on two parameters: the lag (l) and the embedding dimension (d e ). For the reconstruction to be useful (e.g. for dynamic modelling), it is important to choose suitable (optimal) values for these parameters. The following sections discuss methods to estimate the embedding parameters. A. Choosing time delays The embedding theorem states that any time lag is acceptable when reconstructing a state space. This, however, is not very useful when extracting dynamical information from the data. Different lags lead to reconstructions of the attractor that are diffeomorphically equivalent but geometrically different. If the lag (l) is too small, the coordinates y n = s n + ( i 1) l will be so close to each other numerically, that they will be indistinguishable. This, from any practical point, will not provide two independent coordinates. On the other hand, when the lag is too large, the coordinates will be

36 STATE SPACE ANALYSIS OF TIME SERIES 13 statistically completely independent of each other and the projection of an orbit on the attractor is into two totally unrelated directions (Abarbanel, 1993) a.3. b.1.1 X X X X c.1 X X1 Figure.8: Illustration of a) too small, b) too large and c) optimal time delay. The fundamental issue is that there must be a balance between values of l that are too small, where each component of the vector does not add significant new information about the system dynamics, and values of l that are too large (Figure.8). Large values of l create uncorrelated elements in y n because of the instabilities in the nonlinear system that becomes noticeable over time. This will result in the components of y n becoming independent and convey no knowledge about the systems dynamics (Abarbanel, 1998).

37 STATE SPACE ANALYSIS OF TIME SERIES 14 It is therefore essential to determine an intermediate value for the lag that will give an optimal embedding. The linear autocorrelation function and the average mutual information (AMI) statistic are well-accepted methods for such calculations. Linear Autocorrelation Function: The linear autocorrelation function is defined as where C L () l = 1 N N k= 1 1 N [ s s][ s s] N k= 1 k+ l k [ s s] k (.9) s 1 N sk N k = 1 = The function gives information about the linear dependency of coordinates on each other. Higher values of C L indicate a higher average linear correlation between coordinates at that specific lag (l). Determining the time lag where C () l first passes through zero will give a good L estimate of l. Choosing l to be the first zero of the function C () l would, on average over the observations, make the coordinates [ sn, sn+ l, sn+ l,..., sn+ ( d 1) l] linearly independent. It is important to note that this independency may have no relation to their nonlinear independence or their usefulness as coordinates of a nonlinear system. The approach is not a perfect solution to the problem, but it at least gives an indication of what lag (l) to choose. L e Average Mutual Information: While it is helpful to use linear dependence as criteria to determine optimal lag (l), it is preferable to use a measure that takes the aspect of chaotic behaviour into consideration.

38 STATE SPACE ANALYSIS OF TIME SERIES 15 Average mutual information (Frasier & Swinney, 1986) is such an approach and uses the concept of information theory to determine an optimal embedding lag. Mutual information is a way of identifying how much information one can learn about a measurement at one time, from a measurement taken at another time. Say there exist two sets of measurements, set A = { a i } and set B = { b j }, with a probability distribution associated with each system governing the possible outcomes of observations on them. The mutual information between measurement a i drawn from set A = { a i } and measurement b j drawn from set B = { b j }, is the amount of information (in bits) about measurement a i that is acquired by observing b j. Mathematically this is represented by I PAB( ai, b ) j ( a, b ) = log PA( ai) PB( bj) AB i j (.1) where P ( a, b ) is the joint probability density for measurements from set A and B AB resulting in values a and b. PA ( a ) is the probability of observing a out of the set of all A and PB ( b ) the probability of finding b in a measurement from set B. The quantity IAB( ai, b j) is called the mutual information of the two measurements a i and b j, and is symmetric in how much is learned about b j from measuring a i. In a deterministic system these probabilities are evaluated by constructing a histogram of the variations of the a i or b j seen in their measurements. If the measurement of a value from set A (resulting in a i ) is completely independent of the measurement of a value from set B (resulting in b j ), then P ( a, b ) factorizes to P ( a) P ( b ) and the amount of information between the measurements, I ( a, b ), is A B AB AB i j zero. The average mutual information between measurements of any value a i from set A and b j from B is the average over all possible measurements of IAB( ai, b j) : I = P ( a, b ) I ( a, b ) (.11) AB AB i j AB i j a, b i j

39 STATE SPACE ANALYSIS OF TIME SERIES 16 This quantity is not related to the linear or nonlinear evolution rules of the quantities measured. It is strictly a set theoretical idea which connects two sets of measurements with each other and establishes a criterion for their mutual dependence based on the concept of information connection between them (Abarbanel, 1996). To place this definition in the context of observations from a physical system, the set of measurements s n is considered set A; and the set of measurements s n+ 1, a time lag l later, as set B. The average amount of information about s n+ 1 that is acquired when making an observation of s n, or in other words, the average mutual information between observations s n and n 1, is then s + and Il (). N Ps ( n, s ) n+ l Il () = Ps ( n, sn+ l)log n= 1 Ps ( n) Ps ( n+ l) (.1) As is the case with the linear autocorrelation function, the average mutual information function is used to determine a time lag (l) when the values of s n n l and s + are independent enough of each other to be useful as coordinates in a time delay vector, but not so independent as to have no connection with each other at all. Fraser and Swinney (1986) suggest, as a prescription, to choose the lag where the first minimum of Iloccurs, () as the lag (l) for the time delay reconstruction of the state space. This is a simple rule derived from their more detailed suggestion, but it serves quite well. The average mutual information is in fact a kind of generalization from the correlation function in the linear world to the nonlinear world (Abarbanel, 1993 & 1996). B. Choosing the embedding dimension Choosing optimal values for the embedding dimensions is just as important as choosing optimum time lag (l) values. If the chosen embedding dimension is too small, the conditions given in the embedding theorems are not satisfied. On the other hand, if the

40 STATE SPACE ANALYSIS OF TIME SERIES 17 dimension is too large, practical problems occur due to the fixed amount of points that constitute thinner and thinner sets in d R as d is increased (Parlitz, 1995). When considering possible values for embedding dimension, emphasis is placed on determining the integer global dimension where the number of coordinates chosen is just enough to unfold observed orbits from self-overlaps arising as result of projection of the attractor into a lower dimensional space. This means that if two points (or trajectory orbits) of a particular observation set lie close to each other in some dimension d they should do so because it is a property of the set of observations and not because of the small value of d in which the observations is viewed. The lowest dimension that unfolds the attractor so that no overlaps remain is called the embedding dimension (d e ). In practice, it is possible to guess a suitable value for d e by successively embedding into higher dimensions and looking for consistency in the results. (Abarbanel, 1993) From the previous, it should be clear that embedding data into a higher than necessary dimension does not create any ambiguity. The question now arises: why not always embed the observations into an arbitrary large dimension to be sure the attracter is totally unfolded? Two problems arise when working with dimensions that are larger than really required by the data and time delay embedding: 1) Computational time for extracting interesting properties from the embedded attractor increases exponentially with increasing d. ) In the presence of noise or other high dimensional contamination of the data, the extra dimensions are not populated by dynamics already captured by a smaller dimension, but entirely by the contaminated signal. In an embedding space that is too large, unnecessary time is spend working around aspects of badly represented observations which are solely filled with noise. This realization motivated the search for techniques that can identify an optimal embedding dimension from the data itself.

41 STATE SPACE ANALYSIS OF TIME SERIES 18 False Nearest Neighbours: The false nearest neighbours technique developed by Kennel et al. (199) is based on the following idea. For any point, y n, on an attractor one can ask whether its nearest neighbour, NN y n, in a state space of dimension d is there for dynamical reasons, or if it is instead projected into the neighbourhood because the dimension is too low. That neighbour is then examined in dimension d+1 by simply adding another coordinate to y n using the time delay reconstruction. If the nearest neighbour under examination remains a neighbour in the larger space, it is a true neighbour which arrived there dynamically. If the neighbour moves away from the point y n as dimensions are added, it was a false nearest neighbour and only a neighbour because the dimension d was too low. Once the number of false nearest neighbours becomes zero, he attractor has been ambiguously unfolded because all crossings of the orbits have been eliminated. The closeness of two points in the state space is determined by the Euclidean distance between them. In a d R - embedding, each point y = ( s, s, s,..., s ), n n n+ l n+ l n+ ( d 1) l has a nearest neighbour y = ( s, s, s,..., s ). NN n NN NN NN NN n n+ l n+ l n+ ( d 1) l If there is a large amount of data, the distance between them is relatively small. Expanding the state space to d +1 R yields the same points, y ˆ = ( s, s, s,..., s ) n n n+ l n+ l n+ (( d + 1) 1) l and NN y ˆ = ( s, s, s,..., s ), n NN NN NN NN n n+ l n+ l n+ (( d+ 1) 1) l in a d+1 dimensional space where the distance between them may or may not still be small. If the points where true neighbours they should separate relatively slow with successive increases in embedding dimension, and vice versa.

42 STATE SPACE ANALYSIS OF TIME SERIES 19 The increase in distance between the points is given only by the difference between the last components: yˆ yˆ y y = ( s s ). (.13) NN NN NN n n n n n+ (( d+ 1) 1) l n+ (( d+ 1) 1) l The normalized increase in distance between these two points is calculated and the points are considered to be false nearest neighbours if s s R. (.14) NN n+ (( d+ 1) 1) l n+ (( d+ 1) 1) l NN yn yn A suitable value for R T depends on the spatial distribution of the embedded data. If R T is too large, some false nearest neighbours will be missed, and if R T is too small, some true neighbours will be counted as false. It is therefore important to choose a value for R T that is suitable for the spatial distribution of the data under consideration. Typical values are 1 R T 3 (Small, 1998). T The dimension that corresponds to zero false nearest neighbours (or close to zero) is considered to be the minimum embedding dimension of the data; and is less than or equal to the sufficient dimension (remember that de d( A) ) found from the embedding theorem. C. Embedding by Singular Spectrum Analysis (SSA) An alternative approach to the method of delays (MOD) embedding is the use of singular spectrum analysis (SSA). Both methods are theoretically equivalent, but may differ in practice with limited amounts of possibly noisy data. Singular spectrum analysis (SSA) is based upon the theory of singular value decomposition and does not rely on the explicit calculation of embedding delay. Since the calculation of embedding delay is suspect in the presence of noise, the implicit calculation thereof makes SSA a particular attractive option (Barnard & Aldrich, ). The first step in the SSA method of reconstruction is to derive a large p-dimensional p state vector, R, from the scalar observations. This is achieved by embedding the

43 STATE SPACE ANALYSIS OF TIME SERIES observations using a time lag of one, and an embedding dimension that is equal (at least) to the point of linear decorrelation of the observed time series. This point is the smaller of either the first minimum or the first zero crossing of the linear autocorrelation function. The value of the preliminary embedding dimension is the same as the delay value (l) used in a conventional, autocorrelation function determined, MOD reconstruction. Applying these parameters (l=1 and d=l) to the generalized reconstructed state space vector (Equation.8), y = ( s, s, s,..., s ), n n n+ l n+ l n+ ( d 1) l results in y = ( s, s, s,..., s ) (.15) n,p n n+ 1 n+ n+ l 1 as the preliminary, p(=l)-dimensional, reconstruction. The covariance matrix of the embedding is calculated, and the eigenvalues of the covariance matrix are determined and ordered to decreasing rank. This is also known as Singular Value Decomposition (SVD). The final d-dimensional state vector is a projection onto the first d principle p components defined by the data inr, that is y = P y (.16) nd, n, p where y nd, = final d-dimensional reconstruction P = first chosen principle components y np, = preliminary p (=l)-dimensional reconstruction The difference between SSA and MOD is that in MOD the d-coordinates are samples separated by a fixed delay (l) and cover a specific time window length, l = ( d 1) l; while in SSA all the available samples in l w are initially used and then further processed with SVD so that the final d-coordinates are linear combinations of these measurements. The free parameter, d, in SSA is not critical at all and any choice over a lower limit would give essentially the same reconstruction, because the additional coordinates correspond to less significant singular values and give negligible variance assuming the time window length is sufficiently large (Kugiumtzis & Christophersen, 1997). To keep computational cost low it is sensible to choose smaller values for d, but w

44 STATE SPACE ANALYSIS OF TIME SERIES 1 d the reconstructed space, R, should still explain at least 95% of the variance of the original data. The benefits of singular spectrum analysis are (Barnard & Aldrich, ): a) Robustness against noise in the observed data b) Simplicity of the calculations c) Low numerical cost of the calculations d) Significant reduction in measurement and dynamic noise.5 Dimension Estimates: Invariants of the Dynamics Attractor dimension has been the most intensely studied invariant quantity for dynamical systems (Abarbanel, 1993). Generally, people think of real world objects as being one, two or three-dimensional. There, however, exist complex mathematical objects, known as fractals, which have non-integer dimensions. Attractors associated with regular (or integrable) systems always have an integer dimension, whereas attractors associated with chaotic dynamics usually have a fractal dimension. Evidence of fractal dimensions in experimental time series data demonstrates the existence of chaos in the real world, and this is what stimulates research in the area. Fractal dimensions are characteristic of the geometry of the attractor and relate to the way points on the attractor are distributed in d-dimensional space. It is invariant under the evolution operator of the system and is therefore independent of changes in the initial conditions of the attractor orbit. It is also independent of the coordinate system in which the attractor is observed, which means that it can be determined from the reconstructed state space. In other words, the fractal dimension of a dynamic system can be estimated from experimental data, which makes it an extremely valuable statistical quantity for classifying dynamical systems from experimental data.

45 STATE SPACE ANALYSIS OF TIME SERIES.5.1 The Box-counting Dimension The simplest concept of dimension is the number of coordinates needed to specify a state. This idea can be taken a step further by using the geometrically related concept of dimension, whereby volume scale as a function of a characteristic length (ε ) parameter: d V ( ε) ε (.17) Here (V ) is a measure of the volume, which can be length (one-dimensional), area (two-dimensional), volume (three-dimensional) or hyper-volume (multi-dimensional), depending on the geometry. The length scale is denoted byε (e.g. the length of a cube s side or the radius of a sphere) and the dimension of the object by d. If assumed that the relation in Equation (.17) holds true for a general fractal, a universal relation to define dimension can be obtained: log V ( ε) d = (.18) logε First attempts to estimate dimension involved calculating the volume of the attractor. This is done by partitioning the d-dimensional attractor in phase space with a ε -sized grid and counting the amount of partitions that contain at least one data point. The amount of partition elements that conform to this criterion is used as a measure of the volume at that resolutionε. The procedure is repeated, each time with a smaller grid size, untilε has spanned a large enough range (theoreticallyε ). This approach to estimate the dimension is known as the box-counting algorithm and the calculated dimension is called the box-counting dimension (d ). The largest possible value of d that can be obtained through these calculations is the dimension used to do the initial embedding (d e ). If the embedding is two dimensional, but the attractor has not unfolded completely, one will get a (incorrect) value of d =. Embedding the data in increasingly higher dimensions will yield d = d e, until the attractor s geometrical structure is fully unfolded. At this point higher dimensional embeddings will not increase the box-counting dimension (d ) significantly. This saturated fractal dimension is the proper value for d.

46 STATE SPACE ANALYSIS OF TIME SERIES 3 In Equation (.16) the volume is related to the amount of points in the space. Theiler (1987) proposed a more general form of Equation (.16): which results in bulk dim ension = size (.19) log( bulk) log( bulk) d = = (.) log( size) log( ε) The past few years researchers has spend quite some time developing techniques to improve the measuring of bulk, resulting in various dimension estimators. The generalized dimension ( d q ) is defined by d q q log( bulk) 1 log ρ = lim = lim (.1) size log( size) ε q 1 logε where ρ is the natural density of points in the phase space. This definition of d q provides a whole spectrum of invariant quantities for < q <. A more detailed explanation of how d q is derived from basic principles can be found in the article by Abarbanel (1993). Letting q = in Equation (.1) results in d log N = logε (.), which is simply the definition of the box-counting dimension as previously discussed..5. The Information Dimension The information dimension ( d 1 ) is another frequently cited dimension estimate. As with the box-counting dimension, the attracter is partitioned by a ε -sized grid in the phase space. Instead of simply counting the number of partitions containing data points, the volume of the attractor contained within each partition is calculated. This measure seeks to account for differences in the distribution density of points covering the attractor.

47 STATE SPACE ANALYSIS OF TIME SERIES 4 The information dimension ( d 1 ) is defined by letting q = 1 in the generalized dimension ( d q ) (Equation.1). Using L Hopital s rule: where d 1 = lim ε i p log p i log ε i (.3) p i is the probability of data points from the attractor occurring within the i th element of the partition: nx ( i ) px ( i ) = nx ( ) i i The difference between the measures d and d 1 is in the way the bulk (Equation.19) is calculated. The d statistic measures bulk by counting the number of nonempty partitions in the phase space. This means that regardless of how many data points a ε -sized partition contains, it contributes the same amount to the bulk. Contrary to this, d 1 is calculated so that each partition element s contribution to the bulk is proportional to how many data points it contains. Bulk in d is simply the volume the attractor covers, whereas in d 1 bulk is equivalent to the mass of the attractor..5.3 The Correlation Dimension The calculation of the box-counting dimension ( d ) and information dimension ( d 1 ), in practice, usually requires a prohibitive amount of computation time. This is why the most commonly used dimension estimate for attractors is the correlation dimension ( d c ). Calculation of the correlation dimension ( d c ) is computationally efficient and relatively fast when implemented as an algorithm for dimension estimation. Consider the case where q = in the generalized dimension (Equation.1): d log pi i = lim ε o log ε (.4)

48 STATE SPACE ANALYSIS OF TIME SERIES 5 Here d refers to the measure known as the correlation dimension ( d c ). The numerator in Equation (.4) constitutes a two-point correlation function which measures the probability of finding a pair of random data points within a specific partition element (Compare this to the significance of the numerator in d 1 : Equation.3). The numerical value of points have a separation distance less than some value ε. pi can be estimated by counting how many pairs of i In 1983 Grassberger and Procaccia suggested a simple and computationally efficient way of estimating the pi - term in the correlation dimension equation. Let { y } N i n= 1 i be an embedding of a time series in by: de R. The correlation function, C ( ) N ε, is defined C 1 ( ε ) N ( ε) = I y y < (.5) N i j i j N Here I( X ) is a Heavyside function whose value is one (1) if condition X is satisfied and zero () otherwise. The -term is the distance function in de R. The points, y i, are points on the reference trajectory and y j are other points on the attractor in the vicinity of y i. If the distance between point y i and y j is within ε, I( X ) takes a value of one; and if the distance between the points is outside ε it takes a value of zero. i j I y y < ε is the number of points within a distance This means that the sum ( ) ε of point y i. i The calculation and interpretation of the correlation dimension estimates forms an integral part of this thesis, so it is worth spending time understanding the physical meaning of Equation (.5). The idea is illustrated by Figure.9 where a hypersphere of radius ε is centred on one of the points defining the trajectory of the attractor.

49 STATE SPACE ANALYSIS OF TIME SERIES 6 ε y i y j Figure.9: Probing hypersphere on the attractor The number of other points on the attractor within the hypersphere is counted. The hypersphere is then moved from point to point along the trajectory and each time the number of points within is counted. The cumulative sum of points is then divided by N( N 1) to give the correlation sum ( C N ) for a hypersphere of a particular size. For each value of the hypersphere size (ε ) there will be a different value for the correlation sum ( C N ). The maximum value of C N is unity and this happens when the radius of the probing hypersphere is large enough to include the whole attractor, i.e. all the points of the attracter are counted. The minimum value of C N is [ N( N 1)] and occurs when only the closest two points on the attractor are counted. One should notice that the correlation sum counts the closest near neighbour twice as the probing hypersphere visits both points on its journey around the attractor. This is the reason for the in the numerator (Addison, 1997)

50 STATE SPACE ANALYSIS OF TIME SERIES 7 A. The Grassberger-Procaccia Algorithm The method traditionally most often employed to estimate the correlation dimension is the Grassberger-Procaccia algorithm (Grassberger & Procaccia, 1983). They used the assumption that the correlation sum ( C ) scaled with the hypersphere radius ( ε ) according to a power law in the form N CN d ε c, (.6) where d c is the correlation dimension of the attractor. In this method, the correlation sum ( C N ) is calculated for different sizes of the hypersphere radius (ε ). A graph is constructed that plots log C ( ε ) versuslogε. The gradient of this graph in the limit as ε should approach the correlation dimension. The problem, however, is that the graph will jump about irregularly for small values of ε when using a finite amount of data. This can be overcome by looking at the graph when using moderately small ε -values. A typical curve (Figure.1) from this plot will have three distinct features: N 1) For large hypersphere sizes (ε ) the graph will flatten. The flattening happens because the attractor has a finite size. ) At moderate ε -sizes the graph has a scaling region which is approximately linear. 3) For small ε -sizes the graph jumps about irregularly. The irregularity is a result of having only a finite amount of data, so that the contribution to the correlation function by a specific pair of points is seen as a jump in the graph. The extremes of the graph coincide with very large and very small scales that tend to be nonlinear. The correlation dimension is calculated from the slope of the linear part of the graph.

51 STATE SPACE ANALYSIS OF TIME SERIES 8 log(ε) 1 Largescale effects Smallscale effects Slope = d c log(c N ) 3 Figure.1: The log(ε)-log(c N ) plot for determining the correlation dimension via the Grassberger-Procaccia algorithm.

52 STATE SPACE ANALYSIS OF TIME SERIES 9 B. Judd s Algorithm Conventional implementations of the Grassberger-Procaccia method assume that most of the information about the dimension is contained in the linear (scaling region) of the log(ε)-log(c N ) plot and proceed by fitting a straight line to the scaling region. As Judd (199) points out there are deficiencies in the Grassberger-Procaccia algorithm: 1) From a sample trajectory of length n there are nn ( 1) interpoint distances, but these distances are not all independent. The triangle inequality states that the distance between two points is no more than the sum of the distances from a third point. This problem is inherent to all methods that are based on calculations of interpoint distances. ) The smoothness in the correlation function is misleading because it contains a lot of statistically correlated information. This is because the value of C ( ε ) is based on the number of points in the ε -sized hypersphere, which becomes more statistically correlated as ε is increased. Any fit of this function should take this into account by giving the larger values of ε less weight. N 3) It is possible that the scaling (linear) region may not reflect information about the dimension of an attractor. According to Judd (199) examples can be constructed were the scaling region reflects only large-scale properties of an attractor, and not the dimension. The scaling region is also a subjective, but critical choice, and sometimes a small change in the scaling region significantly changes the correlation dimension estimate. It often happens that the analyst chooses a scaling region that is not straight, but curved. When using a certain scaling region, information about small distances between points (which is still information about scaling) is ignored. There is no basis on which this information can be thrown away. He also states in a later paper (Judd, 1994) that linear correlation in the data set misleads the algorithm to falsely show convergence to some low dimension, which could then be misinterpreted for inherent low-dimensional dynamics.

53 STATE SPACE ANALYSIS OF TIME SERIES 3 Judd (199) proposes a different algorithm for calculating the correlation dimension. His algorithm is a modification of the Grassberger-Procaccia algorithm that, in contrast to conventional implementations, considers the contribution of the interpoint distances directly and not the correlation function. The significance of this is that problems () and (3) are completely avoided. He demonstrates that the following equation is a better description of the correlation function, valid forε < ε : CN d ( ε ) ε c q( ε) (.7) where q( ε ) is a polynomial of order t, the topological dimension of the set. The topological dimension is the smallest dimension of embedding space in which the attractor fully unfolds. The correlation dimension ( d c ) is a function of ε and written as d ( ) c ε. Where the Grassberger-Procaccia method assumes a linear scaling region by fitting CN d ε c d, the new method of Judd fits C ( ε ) ε c q( ε) N which allows for the presence of an additional polynomial term that takes into account variations of the slope within and outside of the scaling region. Judd s method dispenses with the need for a scaling region and substitutes a single scale parameter, ε. This has an interesting benefit, which is especially useful for the purpose of detecting dynamic changes (which will become evident in Chapter 3). Many natural objects do not have a constant dimension at all length scales. Consider a large river stone as an example. If one examines the surface at its largest length scale, it is almost two-dimensional. However, at smaller length scales (when the stone is examined closely and in more detail) one can discern the details of grains, which add to the complexity and increase the dimension at smaller scales. This is the same with attractors reconstructed from time series data. Have a look at the reconstructed attractor in Figure.11. At its largest length scale (Figure.11(a)) the attractor seems to be low dimensional. However, when zooming in on a smaller part of the attractor (Figure.11(b)), i.e. examining the smaller length scales, it is obvious that the complexity and therefore the dimension of the attractor increases.

54 STATE SPACE ANALYSIS OF TIME SERIES 31 Reconstructed attractor in 3D X Figure.11(a): A reconstructed attractor from Chua s circuit that seem low dimensional from a distance X X Figure.11(b): Zooming in on part of the attractor reveals the high dimensional nature of the object It is therefore natural to consider the dimension (in this instance the correlation dimension) as a function of ε. Figure.1 illustrates how the correlation dimension for the attractor in Figure.11 change with different length scales. As expected the correlation dimension ( d c ) has smaller values at larger length scales, and vice versa.

55 STATE SPACE ANALYSIS OF TIME SERIES 3 d c (ε ) 1.5 Small scale effects Large scale effects log(ε) Figure.1: A typical graph illustrating Judd s method where d c is a function of ε. By allowing the correlation dimension to be a function of length scale, estimates that are both more accurate and informative can be produced. Some of the approximation necessary to define correlation dimension as a single number can be avoided when using Judd s method, but most important, more detailed information on the topology of the attractor can be extracted..6 Surrogate Data Analysis Data from nonlinear deterministic processes require more complex and time-consuming analysis techniques (i.e. state space reconstruction) that are much less understood than the classical methods used for linear stochastic process data. Before blindly applying these advanced data analysis and modelling techniques, it is important to check whether the data exhibits such nonlinear behaviour that would deserve more advanced treatment.

56 STATE SPACE ANALYSIS OF TIME SERIES 33 It would be particularly useful to have answers to the following questions (Kantz & Schreiber, 1997): Is the apparent structure in the data most likely owing to nonlinearity or rather due to linear correlations? Is the irregularity of the data most likely due to nonlinear determinism or rather due to random inputs to the system or fluctuations in the parameters? The analyst should be aware that linear stochastic processes can also create very complicated looking signals and that not all irregularities in a data set are due to nonlinear dynamics within the system. To confirm this statement, consider the following simple linear stochastic process: x =.99x + η (.8) n n 1 n Here ηn are independent Gaussian random numbers. Successive values in this linear time series are strongly correlated and already yield some structures. Figure.13 illustrates the irregular behaviour of this process X Time index Figure.13: Irregular output from a simple linear process

57 STATE SPACE ANALYSIS OF TIME SERIES 34 The output can be made even more complex by observing the values from the process through a nonlinear measurement function: s n = x (.9) Note that although the measurement function is nonlinear, there are no nonlinear dynamics in this process. All the dynamics is contained in the linear AR(1) 9 part and the nonlinearity is purely static. Figure.14 shows the spiky irregular output from such a linear system. 3 n S Time index Figure.14: Irregular output from linear process observed through a nonlinear measurement function A time series can be anything between purely random and strictly deterministic. When looking at the previous examples it is understandable that linear structures can easily be mistaken for (nonlinear) determinism. It is therefore important to have a reliable test to classify the given time series with confidence. 9 Autoregressive process of order one.

58 STATE SPACE ANALYSIS OF TIME SERIES 35 The test most successfully applied by researchers and analysts, makes use of the method of surrogate data (Theiler et al., 199). This test involves a null hypothesis against which the data are tested, using a discriminating statistic. The data under investigation are first assumed to belong to a specific class of dynamic processes, e.g. linear stochastic. Surrogate data sets are generated from the original data, using this assumption. An appropriate statistic then is calculated for both the original data and the surrogate data sets. If the resulting statistic is significantly different for the surrogate data than for the original data, the null hypothesis is rejected. This means that the process (linear stochastic in this instance) that generated the surrogate data sets is not from the same class than that of the original data. By repeating this procedure, and through trail-and-error elimination, it is possible to get a good idea of the characteristics of the original process (Barnard et al., ). One should always progress from simple and specific assumptions to broader and more sophisticated models. More formally, the test can be explained as follow. Let ψ be a specific hypothesis and F ψ be the set of all process systems consistent with that hypothesis. The time series ( s n ) under investigation, which consists of n scalar measurements, can be expressed as n n s R. Let T : R U be a statistic which will be used to test the hypothesis that s was generated by some process F F ψ. Generally, U will be R, and it will be possible to discriminate between the original data ( s ) and the surrogates ( s i ) consistent with the hypothesis given by the approximate probability density p, () t, i.e. the probability density of T given F (Small & Judd, 1998). TF.6.1 Hypothesis Testing It is important to realise that it is only possible to reject a hypothesis; it is not possible show that it is correct. Data can be consistent with a hypothesis even if the hypothesis is not correct. However, if data are inconsistent with the hypothesis, the data show that the hypothesis is not correct. So, although data can definitely reject a hypothesis, it can only suggest in an informal way that a hypothesis is correct. One should keep in mind that a hypothesis can contain many hidden assumptions and if a hypothesis is rejected, one may actually be rejecting these hidden assumptions.

59 STATE SPACE ANALYSIS OF TIME SERIES 36 Classes of hypothesis Different types of surrogate data are generated to test membership of specific dynamical system classes, referred to as hypotheses. Three commonly used classes of hypothesis are discussed here. They are equivalent to the following assumptions: 1) The data are identically, independently distributed noise (type ) ) The data are linearly filtered noise (type 1) 3) The data are a monotonic non-linear transformation of linearly filtered noise (type ) To test if the data are consistent with a particular hypothesis, a model is build that is consistent with that hypotheses and has the same properties as the original data. Surrogate data are then generated from the model: 1) Type Shuffles (randomises) the data ) Type 1 Randomizes the phases of the Fourier transform of the data 3) Type Applies a phase randomising procedure to amplitude adjusted Gaussian noise. The procedure for generating type surrogates can be seen as a combination of generating type and type 1 surrogates: i. A normally distributed data set z is generated from the original data set s, and reordered so that z has the same rank distribution as s. ii. A type 1 surrogate set z i is generated from z by randomising the phases of the Fourier transform of z. iii. Finally, the original data s is reordered to create a surrogate set si which has the same rank distribution as z i. These type surrogates are also referred to as amplitude adjusted Fourier transformed (AAFT) surrogates.

60 STATE SPACE ANALYSIS OF TIME SERIES 37 The original data are now compared to the different surrogate data sets to if it is typical under the hypothesis. The surrogates generated by the algorithm, 1, or uses a linear model. Each of these surrogate tests addresses a hypothesis that the data are either linear, or some (linear or monotonic nonlinear) transformation of a linear process..6. The Test Statistic To compare the original data to its surrogate data, a appropriate test statistic must be selected. Such a statistic should measure a non-trivial invariant of a dynamical system that is independent of the way surrogates are generated. Theiler and Prichard (1996) suggested that there are two, fundamentally different types of test statistics: pivotal and non-pivotal. A statistic (T ) is considered pivotal if the probability distribution of T ( p TF, ) is the same for all processes ( F ) consistent with the hypothesis (ψ ). The probability distribution ( p TF, ) is therefore invariant for all F F ψ. One should also make the distinction between a simple hypothesis and a composite hypothesis. If the set of processes consistent with the hypothesis ( F ψ ) is singleton, the hypothesis is simple. Otherwise, the hypothesis is composite and the problem is twofold: generating surrogates consistent with a particular process ( F ), as well as estimating F F ψ. Theiler and Prichard (1996) argues that F has to be specified when the hypothesis is composite, unless the test statistic (T ) is pivotal in which case the distribution of T is the same for all F F ψ. They suggest that a constrained realization scheme should be employed when non-pivotal statistics are to be applied to composite hypotheses. This implies one must ensure that the surrogates are realizations of a system consistent with the hypothesis that gives identical estimates of the parameters of that system when compared to the estimates of those parameters obtained from the original data, instead of generating surrogates that are typical realizations of the model. So if ˆF F ψ is the process estimated from the data ( s ) and s i is a surrogate data set generated from Fi F ψ, then the process ( F ˆi the same as ˆF. F ψ ) estimated from the surrogate data set ( s i ) must be

61 STATE SPACE ANALYSIS OF TIME SERIES 38 For example, let ψ be the hypothesis that the original data ( s ) is generated by linearly filtered i.i.d. 1 noise. Non-constrained surrogate data ( s ' ) can be generated from a Monte Carlo simulation based on the best linear model estimated from s. The nonconstrained surrogate data ( s ' ) can be constrained by shuffling the phases of the i Fourier transform. This will produce a set of random data ( s i ") with the same power spectra and autocorrelation as the original data ( s ). Autocorrelation, nonlinear prediction error or rank distribution statistics would all be non-pivotal test statistics, as the distributions of these statistics would all depend on the form of the noise source and type of the linear filter. However, the correlation dimension (as well as other measures derived from dynamical system theory that are invariant under diffeomorphisms) would be pivotal statistics. The probability distributions of these quantities would be the same for all processes, despite the source of noise or the estimated model. i The use of the correlation dimension ( d c ) as pivotal statistic of choice, has gained a great deal of favour over the years. One of the reasons being that there exists a very powerful and reliable method of estimating d c (as described in the previous sections); which gives a quick, effective and informative method for classifying time series data (Small & Judd, 1998). The power of correlation dimension ( d c ) as a test statistic can be illustrated with an example. Consider the two data sets in Figure.15. It is impossible to tell with certainty whether they exhibit nonlinear deterministic behaviour and if both are simply stochastic. Fifteen surrogate data sets (between 15 and 3 sets are needed for statistical significance) are created from each of the two data sets. To compare the surrogate and actual data sets with each other, both are embedded into a higher dimensional space than was calculated from the original time series data. The assumption is that a lower dimensional attractor from a deterministic system will occupy a subspace of the higher dimensional space and should not extent into new coordinates. Usually an embedding dimension of 1 is enough to ensure that the attractor has fully unfolded. The original data ( s ) is then compared with the surrogate data sets ( s i ) by calculating the correlation dimension ( d c ) from these reconstructions. 1 Independently and identically distributed

62 STATE SPACE ANALYSIS OF TIME SERIES s A Time index s B Time index Figure.15: Two data sets that need to be classified. From Figure.16 it can be seen that there is a distinct separation between the cluster of surrogates and the curve of the original data. The type hypothesis can therefore be rejected with confidence, which indicates that there is a strong possibility for determinism in the data. The next graph, Figure.17, shows quite the opposite. The curve of the original data is part of the cluster of surrogate data. It is therefore clear that this data came from a linear stochastic process

63 STATE SPACE ANALYSIS OF TIME SERIES surrogates original data d c (ε ) 4 dc scale log(e) ε Figure.16: Correlation dimension plot of surrogates and original data for data set A surrogates original d c (ε ) dc scale log(e) ε Figure.17: Correlation dimension plot of surrogates and original data for data set B.

64 STATE SPACE ANALYSIS OF TIME SERIES 41 Dataset A comes from a nonlinear deterministic system, namely the Lorenz system (Lorenz, 1963). The dataset B was generated by Equation (.8), which generated linear stochastic data. The results from these two examples show that surrogate data analysis, with correlation dimension as test statistic, is a powerful data classification technique. A more detailed discussion on surrogate data analysis can be found in the articles by Schreiber & Schmitz () and Kaplan (1999).

65 3 DETECTING DYNAMIC CHANGE A problem frequently encountered when analyzing chemical process systems, is appropriate characterization of changes in the system dynamics. Since the discovery of time delay embedding for state space reconstruction, a considerable amount of work has been devoted to the development of techniques to extract information in observed time series data through analysis of the geometrical structure of the attractor underlying the time series. Underlying nearly all of these techniques (of which some are discussed in CHAPTER ) is the assumption that the dynamical process under question is stationary. This means that the dynamical process, and therefore the geometrical attractor, has not changed on long time scales i.e. time scales in the order of the length of the data set. If the geometrical shape of the attractor does change during this time, denoting a nonstationary system, there are two possible reasons: 1) There may be significant behaviour on time scales longer than may be reliably resolved with the given data; meaning that the given data set (time series) is too short to capture the total dynamic behaviour of the process. ) Process parameters, presumed to be fixed, have changed during the observation period. In short, this implies that if there are deterministic rules governing the dynamics, these rules must not change during the time covered by the given time series data. The first instance of nonstationarity can be illustrated using the Lorenz system (Lorenz, 1963) as an example. This three-dimensional system is well known for its nonlinear chaotic behaviour, and is described by the following coupled differential equations: dx = 1Y X dt dy 8X Y ZX dt =

66 DETECTING DYNAMIC CHANGE 3- dz 8 XY 3 Z dt = (3.1) Initially, the equations are solved by varying the time (t ) from zero to 1.7; generating 5 uniformly sampled data points. The resulting attractor, when plotting the variables X,Y and Z against each other, is shown in Figure Z 5 4 Y X Figure 3.1: Lorenz attractor for t = :1.7 A totally different attractor unfolds from the system when the number of time steps is increased to t = (Figure 3.). The resulting attractor is much fuller and display more dimensional detail.

67 DETECTING DYNAMIC CHANGE Z Y -5 - X Figure 3.: Lorenz attractor for t = : It is obvious that trying to model the system using the first, shorter time series will be disastrous. The time series is clearly nonstationary, since it is not sufficiently long 1 to trace out a good approximation of the invariant measures. The second instance of nonstationarity is more complicated. Changes in the process dynamics can be as result of either a slow drift in the system s parameters during the measurement period, or a near instantaneous change under the influence of the environment. Sometimes it is possible to handle a simple change of parameters once it is noticed. If the calibration of the measurement apparatus drifts, one can try to rescale the data continuously in order to keep the mean and variance constant. This is potentially dangerous, unless one is sure that it is the measurement scale, and not the process dynamics, that is drifting. In some cases where a strictly periodic modulation of 1 Long in this context does not necessarily mean the number of data points, but rather the time-evolution.

68 DETECTING DYNAMIC CHANGE 3-4 a parameter is established, it can be treated as a dynamic variable rather than a parameter and does not necessarily destroy stationarity (Kantz & Schreiber, 1997). More formally, a time series is called stationary if all transition probabilities from one state of the system to another are independent of time within the observation period, i.e. when calculated from the data. Detecting nonstationarity can be viewed as determining whether the dynamics underlying the system is autonomous or not. In the framework of deterministic systems, autonomous dynamical systems can be represented as being generated by time independent flow: x = F( x ) (3.) n+ 1 n Here x is a set of dynamic variables, and the reconstruction of a time series generated by such a system is sufficient to uniquely determine the flow that generated it. A nonautonomous dynamical system, however, is generated by time dependent flow: x = F( x, t) (3.3) n+ 1 n In this case the reconstruction of a time series is insufficient to determine the flow that generated it, and therefore the time series is considered nonstationary. For a closed, deterministic description of the system it is necessary to add time as a reconstruction variable. The problem is that such a reconstruction (with time as a variable) provides little useful information. With only one point in the phase space for each time, there is little predictive power for the outcome of a repeated experiment. That is unless the resultant time series is identical. For a variable to provide more useful information, its dynamics should be recurrent in the phase space. Equation (3.3) can also be rewritten in the form: x = Gx (, y), y = H( y, t) (3.4) n+ 1 n n n+ 1 n Here x carries information about the internal dynamics of the system, and y is the external driving force seeing as its dynamics is independent of x. If y is nonstationary, the entire process is nonstationary, and vice versa. Examples where y is stationary include cases where y is periodic, or is itself generated by another ergodic flow. One should keep in mind that ergodic systems are stationary, and that stationary systems are autonomous. Also, note that stationarity is a statistical concept, while

69 DETECTING DYNAMIC CHANGE 3-5 autonomy is a term that refers to the generating algorithm of the process (Manuca & Savit, 1996). 3.1 Measures for detecting nonstationarity After explaining the stationarity problem in the former part of the chapter, the problem at hand is to come up with solutions on how to, for a given time series, detect nonstationarity. What makes this analysis so complicated is that stationarity is a property which can never be positively established, and because stationarity requirements differ depending on the application, the task is even more complex. When testing for stationarity, the first solution one thinks of is to calculate a number of simple statistics, e.g. the mean and variance, from different parts of the time series and use a standard statistical hypothesis test based on their presumed equality. Conventional statistical process control (SPC) techniques, such as the standard Shewhart control charts, cumulative sum (CUSUM) control charts and various moving average (MA) control charts, are widely used to monitor the output of industrial processes (Basseville & Nikiforov, 1993). The main problem underlying these techniques is that they are designed to operate on constant mean processes, i.e. processes that are considered to be stationary within the linear framework (Nembhard and Kao, ). In the context of detecting dynamical nonstationarity, it is important to note that the notion of weak stationarity, found in the literature on linear time series analysis, is not considered. Weak stationarity only requires statistical quantities up to the second order (such as the standard deviation and mean) to stay constant, which in a nonlinear setting, is certainly inadequate (Kantz & Schreiber, 1997). Unlike the detection of changes in nonlinear dynamical systems, change detection for linear systems has been studied extensively. The most complete reference is probably the work done by Basseville and Nikiforov (1993). Their work covers various change detection algorithms. This includes elementary control charts (e.g. Shewhart, geometrical moving average, finite moving average, etc.), filtered derivative algorithms, CUSUM algorithms, Bayes-type algorithms and generalized likelihood ratio (GLR) algorithms, to name a few. To apply these techniques to nonlinear systems, the data have to be linearized using nonlinear filtering methods. These linearization techniques

70 DETECTING DYNAMIC CHANGE 3-6 are not always applicable and often complicates the change detection exercise significantly (Azimi-Sadjadi & Krishnaprasad, ) There are several reasons why these linear approaches are poor. Firstly, the statistics used is arbitrary and not related to any geometrical properties of the attractor, which is the interesting element when analyzing dynamical time series data. Such arbitrary choices are not particularly informative, unless the particular statistic estimates a parameter considered physically or dynamically important, and their power against various sorts of nonstationarity may vary greatly. Secondly, by simply applying such statistics, the significance of differences may be greatly overestimated. This is because observed dynamical time series data are far from uncorrelated, while the simple, classical estimations of confidence rely heavily on the concept of independent observations. If one would measure the empirical means for different parts of a chaotic data set and do a classical t - test (Vining, 1998) for their equality, the null hypothesis of stationarity will quite often be falsely rejected even when the data come from noise free stationary experiments or well-known systems such as the Lorenz attractor. Such methods do not reliably diagnose the intuitive concept of dynamical stationarity that one would typically imagine (Kennel, 1996). Once a suitable statistic has been identified, it must be implemented in a way as to determine the exact time at which the process parameters changed. This will enable one to divide the data into different segments, each with similar dynamical behaviour (signal segmentation), resulting in better modelling and ultimately better control of the process. Rapid detection of parameter changes will also allow online change detection, which is important for effective monitoring and control of chemical processes. A number of statistical tests for nonstationarity in time series data have been proposed in the literature. Most of the tests are based on the following idea: 1) The time series under investigation is divided into a number of segments. ) A certain parameter or statistic is calculated for each segment of the time series. Continuous monitoring of data from a process.

71 DETECTING DYNAMIC CHANGE 3-7 3) The statistic(s) from each segment are compared and if the observed variations are found to be significant, the time series is regarded as nonstationary. In such an approach, the choice of statistic is very important. As stated earlier, linear statistics are not very useful. It is vital to use a statistic that characterizes the geometrical structure of the attractor. In this work, emphasis is placed on the application of techniques (algorithms, statistics). Only techniques and/or statistics that have been shown to be reliable, and which is accessible (e.g. open source code) or can easily be implemented are considered. The following methods fall into this description and will be used to detect nonstationarity: 1) Evaluating the performance of process models ) The method of nonlinear cross predictions 3) Analytical techniques with correlation dimension as test statistic The first method uses model deterioration as an informal way to determine whether a time series is nonstationary. The second method is a more specialised change detection technique, whereas the correlation dimension statistic is even more precise and flexible, and will be the main focus when evaluating these techniques Model-Based Change Detection Model-based detection of changing process parameters is probably one of the more basic methods of determining nonstationarity. In this approach a model is fitted to a segment of the time series data. The model is then used to predict other parts of the observed time series, while its performance is monitored. A good model should produce acceptable small prediction errors when applied to other process data with similar dynamic behaviour. However, if the prediction errors for the other segments of the time series differ significantly from the first segment, the data can be considered nonstationary.

72 DETECTING DYNAMIC CHANGE 3-8 When attempting to model process data, there only exists methods to optimally exploit either the linear correlations or nonlinear determinism. Hence, the decision is whether to model the process as linear stochastic or nonlinear deterministic (Kantz & Schreiber, 1997). The most prominent linear stochastic modelling techniques include: autoregressive (AR) models, moving average (MA) models and Markov models, as well as the use of nonlinear filters (Kantz & Schreiber, 1997; Gershenfeld, N.A. & Weigend, A.S., 1993). In this work, all models to be considered are of the following form, Y = F ( X ) + ε, (3.5) t t where X is the input variables and Y is the output variables of the system., and ε is random variables which is independent of X and Y. t As stated earlier, the focus of this work is on process systems that have deterministic dynamics, or are at least mixed systems with a dominant deterministic part. When modelling nonlinear deterministic data, the observed time series is first embedded to form a d -dimensional reconstructed attractor. These reconstructed vectors are then used as inputs to the model building algorithms. Since time series data are discretely sampled over time, a deterministic model is always a map. In a time delay reconstruction it reads s = F( s, s, s,..., s ) (3.6) n+ 1 n n l n l n ( d 1) l where F() is the function that needs to be determined in order to predict a future values n +. In the above equation 1 d is the embedding dimension and l the time delay. Several nonlinear empirical model classes have been proposed to approximate F, including local linear model fitting, polynomial regression, radial basis function (RBF) neural networks, multiple-layer perceptron (MLP) neural networks, etc. (Kantz & Schreiber, 1997). e In this work, multiple-layer perceptron (MLP) neural networks are used for modelling purposes. It has been shown that they can successfully simulate or predict multidimensional chaotic data (De Oliveira et al., ). The reason for their success is

73 DETECTING DYNAMIC CHANGE 3-9 that MLP neural networks are strong functional approximations 3, which is consistent with the class of dynamical systems under discussion. MPL neural network modelling is relatively easy to use and is well supported by a number of large commercial software packages such as MATLAB. Multiple-Layer Perceptron (MLP) Neural Networks: Artificial neural networks, of which MLP neural networks is a special case, are computer algorithms which simulate, in a very simplified form, the ability of brain neurons to process information. A neural network is composed of interconnected units, organised in layers, which serve as model neurons. Each node in the network acts as a numerical processor with a specific activation (transfer) function. The role of the brain synapse 4 is modelled by a modifiable weight ( w ) which is associated with each connection between neurons in adjacent layers (Figure 3.3). In each neuron all the input weight signals are summed up, and either an excitatory or an inhibitory signal sent to the next layer of neurons. The decision of which signal is sent depends on whether or not the result of the sum has reached a certain threshold value, according to the specific activation function chosen. The transfer function is usually a smoothed step function: k ij 1 φ = (3.7) b c (1 + e X ) The whole network thus becomes a problem of solving the function, F( X) = k a i i= 1 ( 1+ exp( bx c ) where the parameters a i, b i and c i have to be determined by a fit. i i, (3.8) Topologically, a MLP neural network (Figure 3.3) consists of an input layer, one or more hidden layers and an output layer. The number of nodes in the input layer is equal to the dimension of the input space (X), or in this case the dimension of the embedding. 3 In strong functional approximations the optimal approximation error grows more slowly with dimension than for weak functional, e.g. linear and polynomial, approximations 4 The synapse is the structure in the brain which is responsible for storing information

74 DETECTING DYNAMIC CHANGE 3-1 The number of nodes in the output layer is equal to the dimension of the output space (Y), i.e. the number of variables that needs to be predicted. Input Hidden Output layer layers layer 1 1 w 11 1 w11 3 w X Y d w 1 ds, 1 S1 w S 1, S S w 3 S, p p DATA FLOW Figure 3.3: Architectural graph of a multiple-layer perceptron neural network with two hidden layers. Two kinds of signals (Figure 3.4) are identified in a MPL neural network: Functional signals propagate forward through the network. At each neuron through which a functional signal passes, the signal is calculated as a function of the inputs and associated weights applied to that neuron. Error signals propagates backward through the network. Their computation by every neuron involves an error-dependent function in one form or another.

75 DETECTING DYNAMIC CHANGE 3-11 Node Node Node Node Function signals Error signals Figure 3.4: Illustration of the directions of two basic signal flows in a multiple-layer perceptron: forward propagation of function signals and back propagation of error signals. The standard way to determine the model parameters, in other words the optimised weights ( w ), is by training the neural network. The network is trained by feeding it k ij with known time series data and minimising the prediction error with the error backpropagation algorithm. Supervised feedforward MLP neural networks learn from examples of actual data. The weights of the neuron connections are determined by presenting the network with an adequate (depending on the process system) set of actual input-output values, called the training data set. The output values from the network is compared with actual unseen 5 data from the same system by means of some error or cost function. In this work, the sum square error of the model output, defined by 5 Data that were not part of the data set used to train the network. Comparing the network output with seen data would yield illusory low error results

76 DETECTING DYNAMIC CHANGE 3-1 E 1 1 = n, r θ Z t, θ t= 1 n (3.9) is used; where θ denotes the particular model parameters and Z= [ X Y ] the input/output space of the particular data set. The symbol r t, θ is the prediction error when the output from the neural net model, yˆt, θ, is compared with the actual observed values from the system data, y. It is calculated as follow: t r = y yˆ (3.1) t, θ t t, θ The error is minimised by adjusting the weights according to the backpropagation algorithm, which corresponds to the gradient decent procedure with the inclusion of an inertial term to accelerate the convergence (see De Oliveira et al., ). The Levenberg-Marquart algorithm (Levenberg, 1944; Marquart, 1963) is used in this work, as it converges faster than most other training algorithms (Barnard & Aldrich, ). The number of model parameters in a model structure is known as the model order ( d ). For a MLP network the model order depends on the dimension of the input (d ) M and output ( p ) spaces, and also the number hidden nodes (S ): d = S( d + p) (3.11) M To determine the number of hidden nodes (S ) in a nonlinear model, an approach, which starts with a subclass of model structures and then optimises the model order against Rissanen s minimum description length (MDL) calculated for each model structure, is used (Judd & Mees, 1995). This technique encodes the model parameters and model error as a bit stream of information. Under the assumption that a more complex model and a larger modelling error will need more bits to encode, the model structure with the lowest description length (MDL) will be the optimal one. The advantage of this approach is that it presents a formalised structure to determine the model order; in contrast with other methods were overfitting 6 usually occurs. The minimum description length (MDL) can be approximated by the simpler Schwartz Information Criterion (SIC), 6 Implementing a model of too high an order which may fit the training data well, but generalise poor as result of the noise component in the data.

77 DETECTING DYNAMIC CHANGE 3-13 E SIC = N + d N N p θ, Zi log log( ) M i= 1, (3.1) where the summation spans the set of mean square error of each component of the multi dimensional time series. Solving for Equations (3.11) and (3.1) yields the model order ( d ) and the number of hidden nodes (S ) in the network. M Testing for nonstationarity by validating the model: Model validation is a crucial step in any model building procedure, as it ensures the reliable application of the model on new observations from the same data set or process. In this approach, model validation is used to test for nonstationarity. For a time series (i.e. process) to be stationary, the estimated model, trained on data from a specific segment of the recorded time series, should be able to do a faithful simulation (prediction) of observations from any other segment of the time series (Figure 3.5). In other words, the prediction error of the model should statistically be the same for all segments of the time series. The fitness of the model can be tested with the R -statistic, R = 1 ( y yˆ ), (3.13) ( n 1) σ y where y is the actual observations, ŷ is the model simulated (predicted) values, σ is the standard deviation of y and n is the length of the predicted data set. Another way of testing the model fitness is to plot the actual observations ( y ) and the predicted values ( ŷ ) on the same graph, and visually inspect the result. Model validation is usually based on a one-step 7 prediction, but additional insight into the model fitness can be gained by doing the more rigorous free-run 8 prediction (Small and Judd, 1998). 7 The model is only expected to predict one time step into the future. 8 In contrast with one-step prediction, the model is expected to predict multiple time steps into the future, which is more challenging.

78 DETECTING DYNAMIC CHANGE 3-14 Segment used to train model Segments used to test model fitness.8.6 X Time Figure 3.5: Illustration of the model-based approach to detect nonstationarity 3.1. Probing Nonstationarity Using Nonlinear Cross Prediction A more formal test for stationarity is proposed by Schreiber (1997). The test checks for the compatibility of nonlinear approximations to the dynamics made in different segments of the time series. The segments are compared directly and not by way of statistical parameters. He claims that this approach provides detailed information about episodes with similar dynamics during the measurement period and allows for a detailed analysis of physically relevant changes in the dynamics. The approach is based on the similarity between parts of the time series themselves, rather than the similarity of parameters derived from the time series. The similarity is

79 DETECTING DYNAMIC CHANGE 3-15 quantified by evaluating the nonlinear cross-prediction error 9 of different segments of the time series. Schreiber (1997) claims that the concept is particularly useful in cases were nonstationarity causes the shape of the dynamic attractor to change, while most dynamical invariants only shows slight change. Say one has a time series { s ; n 1,... N} n = which is split into contiguous segments of r length r and with the i-th segment being Si = { s( i 1) r+ 1,..., sir}. Traditionally, a statistic γ is now calculated for each segment and compared with each other, or i alternatively with γ calculated from the full sequence, to see if there is any significant change in the value to suggest nonstationarity. Schreiber (1997), however, follows a different approach and use statistics defined on pairs of segments, r r γ = γ( S, S ). (3.14) ij i j By using the statistic, γ ij, on pairs of segments, the number of parameters calculated for a fixed number of segments increase from N r to ( N r ). It can be argued that a lot of redundant information is gained for the purpose of statistical testing, since γ ij for different segment pairs ( i, j) are not expected to be independent. Schreiber (1997), however, have shown that this will allow the detection of different and more hidden kinds of nonstationarities. It will reveal a more detailed picture about the nature of the changes and, in particular, will enable one to locate segments of a nonstationary time series which are similar enough and which can be analysed together. The test exploits the information contained in the relative statistics, γ ( S, S ), in addition to the information contained in the diagonal terms, γ ( S, S ). The diagonal entries are expected to be systematically smaller. They represent in-sample errors since the training and test sets are the same. i i i j In principle, the statistic γ ( S, S ) can be any quantity which is sensitive to the dynamical differences in Si and i j S j respectively. It is important to use a statistic that is 9 The nonlinear cross prediction error quantifies the predictability of one segment using another segment as database.

80 DETECTING DYNAMIC CHANGE 3-16 robust, not too sensitive to the noise level in the data and from which stable estimates can be obtained on rather short segments S, S. A statistic that meets these criteria is the error of the nonlinear prediction algorithm. Usually stable results can be obtained from a few hundred points. The main advantage of this statistic, however, is its ability to calculate cross-correlations. In this case the statistic is known as the nonlinear crossprediction error, γ ij. i j Let X { x, n= 1, N } and Y { y, n 1, N } n X = be two time series, or segments of a n time series. If d is specified as the embedding dimension, the embedding vectors, x, n= d,...,( N 1) y, n= d,...,( N 1), can be formed from the { } n X and { } n respective time series. The idea is to, for each y n, predict one step into the future, i.e. y +, but by using X as a database. A locally constant approximation to the dynamics n 1 relating x n and x n + 1 yields the estimate Y Y yˆ X Here Uε ( yn) = { xn' : xn' yn < ε} X within the set X. The term U ( y n ) 1 = x. (3.15) X n + 1 X n' 1 U ( y ) + X ε n xn' Uε ( yn) ε is an ε -sized neighbourhood of y n, formed denotes the number of elements in that neighbourhood. For isolated points with empty neighbourhoods the sample mean of the ˆ X segment X is used as an estimate for y n + 1. The root mean squared prediction error, γ ( X, Y ) of the sequence Y, given X, is defined by N 1 1 Y X γ ( XY, ) = ( y n+ 1 yn+ 1). (3.16) N d Y n= d Test methodology: r r It is expected that for a stationary time series γ ( S, S ) would be independent of i and j, unless the coherent time of the process is longer than the length of the segments( r ). So if there is variability in the time series on scales that is longer than r, which can be i j

81 DETECTING DYNAMIC CHANGE 3-17 caused by either slow parameter drift or changing parameters, the diagonal terms for the r r nonlinear cross-prediction error γ ( S, S ) will be typically smaller than those with i i i j. The results are viewed as a mutual error prediction map in the form of a 3-D surface plot, Figure 3.6, or a -D colour coded surface, Figure 3.7. Figure 3.6: Example of a typical 3-D surface plot for mutual cross-prediction errors. Figure 3.7: Example of a typical -D colour-coded mutual prediction map

82 DETECTING DYNAMIC CHANGE 3-18 The change in height (Figure 3.6), or change in colour (Figure 3.7), suggests that there is an explicit change in the prediction errors round about segment 1. Segments 1 to 1 are useless for predictions in segments to 3, and vice versa. This indicates a change in process parameters at the time corresponding with segment 1. An element of confusion when interpreting the results is the possible asymmetry ofγ ij. In general γ ( X, Y) γ ( Y, X). If the attractor of Y is embedded within the attractor of X, data points in Y can be predicted with confidence using X as a database. However, Y does not contain enough information to predict all data points in X. This asymmetry, although sometimes confusing, can provide valuable insights into different kinds of nonstationarity. To simplify matters, it is possible to use a symmetrized statistic such as γ ij + γ, but then this advantage is lost. ji Exploiting the Correlation Dimension as a Test for Nonstationarity The change detection technique preferred by the author, involves the correlation dimension ( d c ) as test statistic. The reason for this is because the correlation dimension ( d c ) is a quantity that suffers considerably from nonstationarities in the data (Kantz & Schreiber, 1997). The theory behind correlation dimension ( d c ), as well as its merit as a test statistic, were thoroughly discussed in Section.5.3. This section only revisits some of the main issues concerning correlation dimension ( d c ) calculations, and explain the approach to detect dynamic change in time series data. Remember that for a process to be stationary there must be no significant changes in the geometrical shape of the attractor. Judd s (199) correlation dimension algorithm produces a curve that provides information about the attractor dimension (i.e. geometrical shape) on different length scales. It therefore characterizes the topology of the attractor, which allows the detection of geometrical changes in the attractor, and therefore dynamical changes in the process. This is in sharp contrast with the Grassberger-Procaccia algorithm (Grassberger & Procaccia, 1983) which only gives a binary result. Many researchers (such as Hively et al., 1999) have attempted to use this binary result to detect dynamical change, without much success, and therefore

83 DETECTING DYNAMIC CHANGE 3-19 discredited correlation dimension as a useful statistic. In this work, it will be shown that the d ( ) c ε -curves calculated through Judd s algorithm is especially useful for the detection of dynamical change. Test Methodology: There are a number of ways to use the d ( ) c ε -curves to detect nonstationarities in a time series. A fairly obvious method is to calculate the d ( ) c ε -curves for two halves of the time series. If the d ( ) c ε -curves differ from each other, with respect to either the region they occupy on the d ( ) c ε -ε chart or their shape, the process that generated the time series can be regarded as nonstationary (Figure 3.8)..5 1st halve nd halve d c (ε ) logε Figure 3.8: Illustration d ( ) c ε -curves from two halves of a time series.

84 DETECTING DYNAMIC CHANGE 3- While the above approach assesses stationarity, it does not give an indication of the time at which the process parameter(s) changed, or whether there is a continuous slow parameter drift in the process. An improvement on this approach is to divide the time series into a number of segments and calculate a d ( ) c ε -curve for each segment (Figure 3.9). An increase in the amount of segments the time series is divided into, increases the ability to identify the time at which the parameter change took place. There is a catch though. As the number of segments increase, the amount of data points in each segment decrease, and is there a growing danger that the individual segments are too short to capture the total dynamic behaviour of the process (1 st case of nonstationarity). Furthermore, Judd s algorithm needs at least 1 points to make a reliable estimation for d ( ) c ε. So apart from keeping a segment long enough to still be considered stationary, it seems that the ultimate lower limit of a segment s size is 1 points (This limit will be stretched somewhat in one of the case studies). d c (ε ). 1.8 seg. 1 seg. seg. 3 seg. 4 seg. 5 seg Dynamical change ε Figure 3.9: Illustration of d ( ) c ε -curves for segments of a time series. From the figure it is clear that there was a change in process parameters between segment and segment 3. This is suggested by the shift of the curves, as indicated by the green arrows.

85 DETECTING DYNAMIC CHANGE 3-1 A further improvement on the previous approach is to calculate the d ( ) c ε -curves for a fixed size moving window (Figure 3.1). The moving window approach is useful for shorter data sets, which cannot create enough independent segments to make reliable conclusions. Even when used on longer datasets it should improve on the sensitivity 1 of the segmentation approach, as it allows for a larger number of realizations within the same timeframe..8 1 st dc-curve nd dc-curve 3 rd dc-curve.6 X Time Figure 3.1: Illustration of the moving window approach. This approach is ideal for online monitoring of changing parameters. The moving window approach can be particularly useful in the online monitoring of chemical processes. Here the d ( ) c ε -curves are calculated continuously as new data from the process become available. Say for example that a process is being monitored where 1 new data points become available every minute, and it has been determined 1 Sensitivity in this context means the time it takes to detect a change in parameters.

86 DETECTING DYNAMIC CHANGE 3- that a window size of data points will meet the requirements. It is then possible to calculate a new d ( ) c ε -curve every minute, using the 1 new data points together with the last 19 data points from the previous window. In this manner dynamic changes in chemical processes can be monitored continuously online. 3. Change Detection Methodology The change detection algorithms are evaluated by applying them to three case studies. The case studies were selected not only to reflect a broad range of dynamical systems, but also out of practical considerations. Each case study is approached as if the underlying dynamic behaviour of the system is unknown which will be the case in most real world applications (Figure 3.11): 1) Time series data from the process is first classified through surrogate data analysis. This to acquire more information on the dynamical behaviour of the system. Knowing whether the system under investigation is nonlinear deterministic or stochastic will indicate whether more advanced change detection techniques are needed (as discussed in this work), or if traditional SPC methods will be sufficient. ) If the system prove to be nonlinear deterministic, the optimal embedding parameters (time delay and dimension) for the data set are determined and the state space reconstruction done. 3) The change detection techniques are applied to the data to determine whether (and when) dynamic changes occurred in the process. The change detection techniques will be evaluated on their ability to detect intrinsic changes in the dynamics of the processes, as well as the time necessary to detect these changes.

87 DETECTING DYNAMIC CHANGE 3-3 Process Data Deterministic Data classification via Surrogate Data Analysis. Stochastic Phase Space Reconstruction Traditional SPC techniques Stationary Dynamic Change Detection Algorithms Nonstationary Evaluation of Individual Change Detection algorithms Figure 3.11: A flow diagram of the approach to detect parameter change.

88 4 CASE STUDIES Evaluating change detection techniques using actual process data is not ideal. The major problem is that one usually do not know beforehand whether the time series data are stationary or not (This is the reason for developing nonstationarity tests!). Reliable conclusions about the performance of these techniques cannot be made if the behaviour of the data is unknown. It is for this reason that two of the case studies are based on simulated data with manually induced parameter changes. Nevertheless, because the whole point of this research is to improve the control of real-life chemical processes, it is important to include a case study based on observations from an actual process. The third case study is based on actual data from a metal leaching plant. 4.1 Autocatalytic Reactor The first case study is a classic chemical engineering problem. The system is consists of two parallel, isothermal autocatalytic reactions taking place in a continuous stirred tank reactor (CSTR) (Lee & Chang, 1996). A B D A B D C Figure 4.1: Schematic illustration of the autocatalytic reactor.

89 CASE STUDIES 4- The kinetics of the system proceeds according to the following steps: A + B 3B (4.1) B C (4.) D + B 3B (4.3) The kinetics is governed by the following rate equations: γ = kc C (4.4) A 1 A B γ = kc (4.5) C B γ = kc C (4.6) D 3 D B When the reaction is carried out in a CSTR, the system can be described by three ordinary differential equations: dx1 1 x1 D1 x1 x3 dτ = (4.7) dx 1 x D x x3 dτ = (4.8) dx3 1 (1 D3) x3 γ1 D1 x1 x3 γ D x x3 dτ (4.9) The parameters in differential equations are explained by the following relationships: x C A 1 =, x CA C C D =, x3 D C C B =, (4.1) B kc 1 BV D1 =, Q D kc 3 BV kv =, D3 =, (4.11) Q Q C A γ 1 =, CB C D γ 1 =, (4.1) CB t Q τ =. (4.13) V

90 CASE STUDIES 4-3 where x i is the dimensionless concentration of C A, C D and C B ; D i the Damkohler numbers for the species A, D and B ; γ i the ratios of the species in the feed; and τ the dimensionless time. With the parameters D 1 = 18, D = 4, D 3 = 8, γ 1 = 1.5 and γ = 4., the continuous autonomous system represented by Equations (4.9) (4.11) exhibits chaotic behaviour (Lee & Chang, 1996). The system is solved by generating 1 observations using the ODE45 subroutine in MATLAB. The time is varied from to 5 using a sampling rate of.5. At τ = 5, a parameter change is induced whereby the parameters γ 1, γ slowly change (at a constant rate) from γ 1 = 1.5, γ = 4. to γ 1 = 1.55, γ = 4.5 over the next 1 observations. At the point where τ = 1, the two parameters are kept fixed at γ 1 = 1.55 and γ = 4.5 for the next 1 observations, until τ = 15. The result is a time series consisting of three regions: 1) Observations 1 1 A stationary part were all the system parameters are kept constant. D = 18 D D = 4 = 8 γ = 1.5 γ = 4. ) Observations 11 A nonstationary part with a slow, but constant, parameter drift. D = 18 D D = 4 = 8 γ = 1.5 γ = 4. Slow Parameter change D = 18 D D = 4 = 8 γ = 1.55 γ = 4.5

91 CASE STUDIES 4-4 3) Observations 1 3 Parameters stabilize at their new values to form a second stationary part. D = 18 D D = 4 = 8 γ = 1.55 γ = 4.5 What in fact happened to the process, from a real-life perspective, is that the feed concentrations C A and C D, after being constant for a period of time, gradually increased over time and stabilized at a new higher value; thus creating a nonstationary time series. All the calculations in this case study are based on the x 1 observations (shown in Figure 4.), as it is assumed that only the x 1 - variable could be measured..6.4 X Time index 4 Figure 4.(a): All 3 data points from the nonstationary time series generated by the autocatalytic reaction.

92 CASE STUDIES Parameter change.5.4 X Time index x 1 4 Figure 4.(b): Data points 95 to 15 from the nonstationary time series generated by the autocatalytic reaction. It is impossible to detect the hidden nonstationarities in time series (shown in Figure 4.(b)) visually. From Figure 4.(a) it can be seen that the mean and standard deviation of the data remain virtually unchanged during the parameter changes. Traditional SPC methods will be unable to detect these changes. More advanced techniques, which can characterise the dynamic behaviour of the system, is needed State Space Reconstruction of the Autocatalytic System The first step in state space reconstruction is to calculate the embedding parameters, i.e. the time delay (or time lag) and the embedding dimension. Either the autocorrelation function, or the average mutual information (AMI) criterion can be used to determine the time delay. The average mutual information statistic gives the following result (Figure 4.3):

93 CASE STUDIES Average Mutual Information Figure 4.3: The average mutual information, as function of the time delay, for the autocatalytic reaction. Embedding lag The most suitable time delay is where the first minimum value of the average mutual information statistic occurs (Fraser & Swinney, 1986) in this case 17 time steps. Calculating the time delay with the autocorrelation function gives Figure 4.4 as result. Closer inspection of the graph suggests 15 time steps as the embedding delay the first value where the autocorrelation function is zero. The values for both the average mutual information and autocorrelation function are very close. This is not always the case. In principle, the AMI statistic gives better results than the autocorrelation function. The reason being that the autocorrelation function is a linear statistic, while the AMI statistic takes the aspect of chaotic behaviour into consideration (as discussed in Chapter ). The autocorrelation function is also known to sometimes overestimate the time delay, in which case it is better to use the mutual information statistic (Barnard & Aldrich, ).

94 CASE STUDIES Correlation [ACF(Tau)] Signal Time Shift (Tau) Figure 4.4: Autocorrelation function as statistic to determine time delay for the autocatalytic reaction. The embedding dimension is determined by the FNN (False Nearest Neighbours) algorithm, using 17 time steps as the embedding delay. Figure 4.5 shows that the fraction of false nearest neighbours is almost zero at an embedding dimension of Fraction of FNN Embedding dimension Figure 4.5: Fraction of FNN as function of embedding dimension for the autocatalytic reaction.

95 CASE STUDIES 4-8 The system s attractor can now be reconstructed using a time delay of 17 and an embedding dimension of 3. The reconstructed attractor is shown in Figure X X X Figure 4.6: Reconstructed attractor for the autocatalytic reaction system 4.1. Surrogate Data Analysis for the Autocatalytic System When investigating any system, it is important to know whether it exhibits deterministic behaviour, or linear stochastic behaviour. This classification will provide valuable insight into the dynamics of the autocatalytic process, and will confirm whether more advanced change detection techniques need to be used. Fifteen (15) type surrogate data sets are generated using the Amplitude Adjusted Fourier Transform (AAFT) algorithm. These surrogates are compared with the actual data using Judd s correlation dimension ( d c ) estimate as statistic. The data sets (both the actual data and its surrogates) are embedded into 1 dimensions to ensure that the attractor from the actual data set has completely unfolded.

96 CASE STUDIES Surrogates d c (ε ) Actual data logε Figure 4.7: Correlation dimension curves for surrogates and actual data for autocatalytic reaction. There is a definite separation between the surrogate data sets and the actual data (Figure 4.7). This, as well as the low dimensionality of the dc - curve from the actual data, suggests that the system is nonlinear and governed by highly deterministic rules. The outcome is as expected, since the data were simulated from three nonlinear differential equations governing the dynamics of the system resulting in a low (three) dimensional attractor Modelling the Autocatalytic System Knowing that the system under investigation is most likely nonlinear deterministic, the use of MLP-neural networks will probably serve best in modelling the data. The idea is to see how nonstationarities in the data influence the possibility of modelling the

97 CASE STUDIES 4-1 system. A deterioration in the performance of the model will serve as a test for detecting nonstationarities in the data. To start with, a model is fitted to the first 5 data points of the time series. The Levenberg-Marquardt algorithm is applied to estimate the model parameters. The algorithm is instructed to search for an optimum model that has between and hidden nodes. It is important not to overfit the model, as it may lead to the problems described in Section The algorithm attempts to minimise the Schwartz Information Criterion (SIC) for the model. As soon as there are 3 consecutive model estimates, for which the SIC does not improve on the current global minimum, the parameter estimation ends. The number of hidden nodes that corresponds to this minimum SIC, is the amount used in the MPL network. Figure 4.8 illustrates how the number of hidden nodes was optimised. In this particular instance, 17 nodes is the optimum x 14-7 minimum SIC Number of hidden nodes Figure 4.8: History of the moving global minimum of the Schwartz Information Criterion versus the number of hidden nodes for the autocatalytic system. The model, estimated from the first 5 data points of the time series, is now used to predict unseen data from the rest of the time series, while its performance is evaluated. The free-run prediction is used to validate the model, as it is a more rigorous test of the validity of the model than the one-step prediction. Since the simulated data are noise free, the one-step validation will give very low prediction errors ( R 1) in

98 CASE STUDIES 4-11 particular, and more insight will be gained by analysing the free-run predictions. This was verified by fitting the model to any part of the time series, even after the parameters had changed, and it still gave a R value of essentially one. The stationarity of the process is assessed by performing a series of free-run predictions on the different dynamic regions of the time series: 1) Data points 5-5 part of the first dynamical stationary segment of the data. ) Data points part of the segment where slow parameter drift occured. 3) Data points 5-5 part of the segment where the process parameters had stabilized again. Although the model still performs reasonably well in all three instances, it is evident (by examining Figures 4.9 to 4.14), that the performance of the model deteriorates when trying to predict data outside its dynamic range. The model gives an excellent prediction of unseen data within the same dynamic range (Figure 4.9). It predicts the data points without a problem, and this is verified by the low residuals 1 (Figure 4.1). Predictions outside the model s dynamic range become increasingly more difficult. The model cannot predict the data with the slow parameter drift as well as before (Figure 4.11) and the residuals are almost 1 times larger than on data with the same dynamic behaviour (Figure 4.1). The free-run prediction almost breaks down when predicting data from the segment with the new process parameters. It is only capable of predicting about 3 data points into the future (Figure 4.13) and the residuals are again almost 1 times larger than in the previous segment (Figure 4.14). This deteriorating performance of the initial model indicates that there might have been some changes in the process parameters, had it been unknown beforehand. Although this is not a refined method for detecting nonstationarity, it can serve as an indication of changing parameters. 1 The residual is the difference between the actual and predicted value.

99 CASE STUDIES Observed data Predicted data.4 Y,x record index Figure 4.9: Free-run prediction of data points Ys Y x 1-5 residual Figure 4.1: Plot of predicted values versus actual values (5-5), and the residuals.

100 CASE STUDIES Observed data Predicted data.4.3 Y,x record index Figure 4.11: Free-run prediction of data points Ys x 1-4 Y residual Figure 4.1: Plot of predicted values versus actual values (15-15), and the residuals.

101 CASE STUDIES Observed data Predicted data.4 Y,x record index Figure 4.13: Free-run prediction of data points Ys x 1-3 Y residual Figure 4.14: Plot of predicted values versus actual values (5-5), and the residuals.

102 CASE STUDIES Detecting Dynamic Change in the Autocatalytic Reactor using Nonlinear Cross Prediction Schreiber s Approach In Schreiber s (1997) approach the nonlinear cross-prediction error between different segments of the time series is calculated and viewed in the form of a mutual map. The time series is divided into 3 segments, each containing 1 data points. The mutual cross-prediction errors between these segments are calculated using the previously determined embedding parameters of l = 17 and d = 3. The results can be viewed in the form of a 3-D surface plot (Figure 4.15) and a -D colour-coded surface (Figure 4.16). The plots of the mutual prediction errors indicate that there is a change in process parameters round about segment 15-17, which corresponds to data points Segments 17-3 are useless for predicting segments 1-17, which suggests that the process dynamics from these two parts are not the same. Something interesting is revealed by a closer inspection of the mutual prediction maps. While segments 17-3 cannot predict segments 1-17 with great success, segments 1-17 are still very useful for predicting segments This may seem contradictory, as one usually expects that a change in parameters would result in an equally bad result both ways. In the previous chapter it was stated that, while the prediction error is mostly symmetric, in general γ ( X, Y) γ ( Y, X). The asymmetric result point towards the possibility that the attractor for the data points 17-3 (more or less the time where the process parameters stabilized at their new values) is somehow embedded within the attractor for the data points 1-15 (more or less the time of the original fixed process parameters). This is confirmed by plotting the two parts of the reconstructed attractor on the same axis (Figure 4.17).

103 CASE STUDIES 4-16 Figure 4.15: 3-D surface plot of mutual cross-prediction errors for the autocatalytic reaction. Figure 4.16: -D colour-coded mutual prediction map for autocatalytic reaction.

104 CASE STUDIES X3.4.. X X1. Figure 4.17: Attractors form by the two different parameters sets are embedded into each other. This explains why the neural net model was able to predict data points in the latter part of the time series sequence (data points -3) reasonably accurately. The situation underlines the inherent problems when using models to detect dynamic change. Has the parameter change been the other way around, the model performance would have deteriorated much quicker. The manner in which the time series sequence is structured now, however, favours the model fitness, which is why the model-based technique did not detect nonstationarity with as much confidence.

105 CASE STUDIES Using the Correlation Dimension to Detect Dynamic Change in the Autocatalytic Reactor The correlation dimension ( d c ) is a statistic that characterizes the topology of an attractor. When process parameters change, the geometrical shape of the system attractor also changes. This makes the correlation dimension ( d c ) potentially an ideal statistic for detecting nonstationarity. Especially when using Judd s algorithm, as it provides details about the geometry of the attractor on different length scales. The same embedding parameters ( d = 3 and l = 17 ) are used to reconstruct the attractor from which the d ( ) c ε - curves will be calculated. The first step in detecting nonstationarity is to calculate the d ( ) c ε - curves for the first and second halve of the time series. If the d ( ) c ε -curves differ from each other, with respect to either the region they occupy on the d ( ) c ε -ε chart and/or their shape, the process is presumed nonstationary. The d ( ) c ε -curves for the two halves of the time series is shown in Figure d c (ε ) logε Figure 4.18: d ( ) c ε -curves from the two halves of the autocatalytic time series.

106 CASE STUDIES 4-19 The dissimilarity of the d ( ) c ε -curves in Figure 4.18 confirms nonstationarity. Although it has now been determined that the time series is nonstationary, the extent of the parameter change is still unknown. Dividing the data set into smaller segments (6 segments with 5 points each) proves to be more useful (Figure 4.19) d c (ε ) logε Figure 4.19: d ( ) c ε -curves from six segments, each containing 5 data points, of the autocatalytic time series. The similarity of the first two d ( ) c ε -curves (segments containing data points 1-5 and 5-1) point towards the fact that no, or very slight, parameter change took place. The part of the time series containing the first 1 data points is thus considered stationary. The parameter change is revealed by the d ( ) c ε -curve from the third segment (1-15). The fact that this d ( ) c ε -curve, as well as the curve from segment four (15-), moves out of the stationary region, indicates a change in process parameters. The curves from the last two segments (-5 and 5-3) are again very similar and suggest that the drifting parameters have stabilized at new values.

107 CASE STUDIES 4- When analysing this particular graph (Figure 4.19) it is clear that nonstationarity can only be reliably detected after observing 15 data points from the process; although the parameter drift started just after the 1 th point. This can be improved by increasing the number of segments, and thus pinpointing the time at which the parameter change took place with greater accuracy. The only problem with this advance is that Judd s algorithm is very data hungry and gets less accurate with a decreasing number of data points. Furthermore, by decreasing the number of data points in a segment, one risks working with a data set that is nonstationary as a result of too few data points. One way to avoid this problem is to use a moving window. Figure 4. illustrates this approach using a fixed-size moving window of 6 data points, which moves forward by data points with each step. d c (ε ) logε Figure 4.: Calculating d ( ) c ε -curves from a moving window for the autocatalytic reaction time series. In this plot the blue curves represents the part of the time series were the process parameters are constant, the red curves where the parameters are in the process of changing, and the green curves the part where the parameters stay fixed at their new value. No parameter change can be detected from the first four moving window

108 CASE STUDIES 4-1 segments (up to data point 1). The fifth moving window segment (8-14) shows the first signs of a possible parameter change. The moving window approach is an improvement on the previous 6-segment approach in that only 14 data points are necessary to detect the parameter drift, compared to 15. The fourth moving window segment (6-1) is not capable of picking up the change, as 4 of the 6 data points are still part of the stationary part of the time series and only slight parameter change takes place in the next points. The moving window approach can be especially useful for online monitoring of changes in process parameters, but also where the available amount of data is limited The Effect of Noise in the Autocatalytic Process Data One might argue that data measured from real-life processes are not, as is the case for the previous simulated data, noise free. To test the effectiveness of the change detection techniques on noisy data, Gaussian measurement noise was added to the autocatalytic process data used in the previous case study. The noise level was set to 1% of the standard deviation of the data. The effect of adding measurement noise to the data is evident from Figure X Time index x 1 4 Figure 4.1: Data points 975 to 15 from the nonstationary time series generated by the autocatalytic reaction noisy data.

109 CASE STUDIES 4- As in the case for the noise free data, the optimal embedding parameters are determined by the AMI and FNN statistics. An embedding delay of l = 15 (Figure 4.) and an embedding dimension of d = 6 (Figure 4.3) are chosen. While the embedding delay for the noisy data is close to that of the noise free data (15 time steps vs. 17 time steps), the embedding dimension for the noisy data are considerably higher compared to that of the noise free data (6 dimensions vs. 3 dimensions). This demonstrates that noise increases the dimensionality of the data. The resulting reconstructed attractor is shown in Figure 4.4. The presence of noise in the reconstructed attractor is clearly visible; as well as the dissimilarity in shape when compared to that of the noise free data.. Average Mutual Information Embedding lag Figure 4.: The average mutual information, as function of the time delay, for the autocatalytic reaction noisy data.

110 CASE STUDIES Fraction of FNN Embedding dimension Figure 4.3: Fraction of FNN as function of embedding dimension for the autocatalytic reaction noisy data. PC PC 1 PC1 3 3 Figure 4.4: Reconstructed attractor for the autocatalytic reaction system, projected onto the first three principal components noisy data.

111 CASE STUDIES 4-4 Surrogate analysis of the noisy data also suggests that the process which generated the data could be nonlinear deterministic. The noise had no effect on the analysis, apart from slightly increasing the d ( ) c ε -values for the actual data Surrogates 4.5 d c (ε ) Actual data logε Figure 4.5: Correlation dimension curves for surrogates and actual data for autocatalytic reaction noisy data. The same approach, used for the noise free data, is used to detect dynamic changes in the noisy data. The model parameters are estimated from the first 5 data points of the noisy time series, and then used to predict unseen data from the rest of the time series. Figure.6 shows a free-run prediction of the model on the first stationary part of the time series. The performance of the model deteriorates when predicting data from the last part of the time series, which suggest that there was a change in the process parameters (Figure.7). This is consistent with the results obtained from the noise free data.

112 CASE STUDIES Observed data Predicted data.1.5 Y record index Figure 4.6: Free run prediction on the first part of the time series where the process parameters were unchanged noisy data Y1 -.5 Observed data Predicted data record index Figure 4.7: Free run prediction on the last part of the time series where the process parameters had already changed noisy data.

113 CASE STUDIES 4-6 The method of nonlinear cross predictions (Schreiber, 1997) gives essentially the same results (a parameter change roundabout data point 15) on the noisy data, as it did on the noise free data (Figure 4.8 and Figure 4.9). The only real difference is the magnitude of the prediction errors. As one would expect, the prediction errors are larger for the noisy data than for the noise free data. Figure 4.8: 3-D surface plot of mutual cross-prediction errors for the autocatalytic reaction noisy data. Figure 4.9: -D colour-coded mutual prediction map for autocatalytic reaction noisy data.

114 CASE STUDIES 4-7 An assessment of the correlation dimension statistics from the noisy data gives similar change detection results to that obtained from the noise free data. The moving window approach (Figure 4.3) confirms a parameter change from the 5 th moving window segment (data points 8-14) onwards. d c (ε ) Figure 4.3: Calculating d ( ) c ε -curves from a moving window for the autocatalytic reaction time series noisy data. logε The amount of measurement noise added to the data in this case study, did not have a significant impact on the ability to reliably detect changes.

115 CASE STUDIES The Baker s Map Although this system does not represent a chemical process, it is a well-known chaotic system and is a popular choice for benchmarking stationarity tests (Schreiber, 1997; Yu et al., 1998; Yu et al., 1999). This is the main motivation for using the baker s map as a case study, as it will give a good idea of how the change detection techniques discussed in this work perform in comparison to other proposed methods. A generalization of the baker s map is given by the equations, un+ 1 = βun ( vn α) v = v α n+ 1 un+ 1 =.5 + βun ( vn > α) v = ( v α) (1 α) n+ 1 n n. (4.14) The system is a two dimensional piecewise linear mapping defined for v [,1] and β ],1[. A nonstationary time series is created by varying β slowly with time, n, α n β =, N while the parameter α is kept fixed at α =.4. Forty thousand data points are measured ( N = 4 ) by recording w= u+ v. The running mean is subtracted from this value ( w ), and the time series is normalized to unit running variance. The n resulting time series becomes: x n = w n w k ( w w ) k k (4.15) Here k represents the average over the indices ' n n k n k = ( ),...,( + ), with k = 5 in this instance. Since most of the observables in this sequence remain unchanged, the nonstationarity is very hard to detect. The running mean and variance are constant up to finite sample fluctuations, and autocorrelations show only very small variation (Screiber, 1997). When inspecting the entire time series visually (Figure 4.31(a)), one would only suspect that the process parameters have changed after observing at least

116 CASE STUDIES data points (indicated by an increase in the variance of the observations). Since the system is subjected to a slow parameter drift, chances of observing parameter changes when viewing only a small number of data points are even more remote (Figure.31(b)). x W Record index x1 4 Figure 4.31(a): Time series data for the nonstationary baker s map all 4 data points x w Record index x 1 4 Figure 4.31(b): Time series data for the nonstationary baker s map data points 45 to 55.

117 CASE STUDIES State Space Reconstruction of the Baker s Map Researchers that used the baker s map for benchmarking their stationarity tests, reconstructed the attractor by embedding the time series into two dimensions ( d = ) using a time delay of one ( l = 1). To be consistent and ensure comparable results, the same embedding parameters are used in this case study. The reconstructed attractor is shown in Figure X X1 Figure 4.3: Reconstructed attractor for the baker s map. 4.. Surrogate Data Analysis for the Baker s Map To determine if the baker s map exhibit nonlinear deterministic behaviour, 15 type surrogate data sets are generated, again using the AAFT algorithm. The d ( ) c ε -curves from the actual data and surrogate data (embedded in 1 dimensions) are compared in Figure The result from the surrogate data analysis cannot confirm nonlinear determinism with confidence. While the d ( ) c ε -curve for the actual data is lower dimensional and separated from the surrogates for logε < 1.6, the opposite is true for the larger length scales. This inconclusive result is most probably due to instabilities in the correlation dimension algorithm.

118 CASE STUDIES d c (ε ) Surrogates Data logε Figure 4.33: d ( ) c ε -curves for surrogate and actual data when embedding into 1 dimensions Modelling the Baker s Map Modelling the baker s map proved to be more challenging than that of the autocatalytic system. In the case of the autocatalytic system, a one-step prediction R -value of essentially one were obtained on unseen data, but the best model for the baker s map could only achieve a R -value of.91. The optimal model, as determined by the SIC, has 11 hidden nodes (Figure 4.34).

119 CASE STUDIES x minimum SIC Number of hidden nodes Figure 4.34: History of the moving global minimum of the Schwartz Information Criterion versus the number of hidden nodes for the baker s map. Since the estimated model does not provide a very good fit to the data, the one step prediction, rather than the free-run prediction, will be used to validate the model. It will be useless to compare free-run predictions for detecting nonstationarity, as even the training ( seen ) data gave bad free-run results (Figure 4.35). 1 x 1-4 Observed data Predicted data 5 Y,w record index Figure 4.35: Free-run prediction of model training data for baker s map.

120 CASE STUDIES x Y,W Observed data Predicted data Record index Figure 4.36: One step prediction for data points 77 to 775 to illustrate the model fitness. The R validation results, as calculated from different segments of the time series, are given in Table 4.1. Figure 4.36 shows the one step prediction model fit for the segment 5-1. Table 4.1. R - statistic results for one-step prediction of different segments from the baker s map time series. Data Segment R -statistic Data Segment R -statistic

121 CASE STUDIES 4-34 It is hard to make meaningful conclusions, with respect to detecting parameter change, when validating the model on consecutive segments of the time series. At first, it seems as if the model gets worse when predicting observations further away from the training data set. This is evident when comparing the decreasing R -values for segments 5-1 to -5. For segments 5-3 to 35-4 the R -values increase again, which could lead one to believe that the parameters must be returning to their previous states. This is not the case. The fact that the R -values from different segments of the time series is not statistically the same, suggests that the process generating the data might be nonstationary. Although this approach at least suggests that there might be parameter change, it gives no real indication of the extent of the nonstationarity Detecting Dynamic Change in the Baker s Map Using Nonlinear Cross Prediction For this approach, the baker s map time series is divided into 4 segments of 1 data points each, and embedded into two dimensions with a time delay of one. The mutual nonlinear cross predictions between the segments are again illustrated by means of a 3- D surface plot (Figure 4.37) and a -D colour-coded surface (Figure 4.38). Both figures indicate that predictability degrades rapidly with the temporal distance of the segments. Nonstationarity can only be confirmed with confidence after taking into account at least 15 to segments. In other words, about 15 to data points must be observed before a change in the process parameters can be detected. The fairly constant rate (evident when analysing Figure 4.37) at which the predictability degrades indicates that the change in dynamics may be as result of a slow parameter drift, rather than to an instantaneous parameter change.

122 CASE STUDIES 4-35 Figure 4.37: 3-D surface plot of mutual cross-prediction errors for the baker s map Figure 4.38: -D colour-coded mutual prediction map for the baker s map

123 CASE STUDIES Using the Correlation Dimension to Detect Dynamic Change in the Baker s Map Using the same embedding parameters than before ( d = and l = 1), a d ( ) c ε -curve for each of the two halves of the baker s map time series is calculated as a starting point. The dissimilarity of the two curves (Figure 4.39) indicates that there must be differences in the geometrical structure of the attractors reconstructed from the two halves of the time series suggesting nonstationarity in the data d c (ε ) logε Figure 4.39: d ( ) c ε -curves from the two halves of the baker s map time series. To get a better idea of when the parameter change took place, and the extent of the change, the time series is divided into 8 segments containing 5 data points each. The d ( ) c ε -curves calculated from these segments are given in Figure 4.4. Dissimilarity in the curves calculated from the consecutive segments of the data again confirms that the process parameters have not stayed constant throughout the observation period. No two curves look similar, which implies that there is a continuous parameter drift

124 CASE STUDIES 4-37 throughout the observation period. Using this particular segmentation, it is not possible to identify any stationary part in the time series. d c (ε ) logε Figure 4.4: d ( ) c ε -curves calculated from eight consecutive segments, each containing 5 data points, of the baker s map time series It would be interesting to see how quickly nonstationary can be confirmed. In other words, how many data points are necessary before a change in the parameters can be detected with confidence? The moving window approach is better suited for such an analysis. Using a window size of 5 data points with a step size of 1 data points generates Figure A comparison of the first two curves from the moving window segments already suggests that nonstationarity may be present. Comparing curves from consecutive moving window segments confirms the suspicion of nonstationarity in the data. Only about 7-8 observations are necessary to reliably detect nonstationarity, using correlation dimension as test statistic, in the moving window approach.

125 CASE STUDIES 4-38 d c (ε ) logε Figure 4.41: Calculating d ( ) c ε -curves from a 5 point moving window for the baker s map d c (ε ) logε Figure 4.4: Calculating d ( ) c ε -curves from a point moving window for the baker s map.

126 CASE STUDIES 4-39 The size of the moving window can even be reduced to data points, to reduce the number of data points needed to detect the parameter change (Figure 4.4). One should just remember that decreasing the window size, increases the risk of working with segments that are nonstationary as result of too few data points. In this instance, only 4 data points is necessary to reliably detect nonstationarity. This is an improvement on Schreiber s mutual nonlinear cross prediction method where at least 15 observations are needed. It also outperforms the technique used by Yu et al. (1998 & 1999) where 5 to 1 observations are needed to reliably detect such hidden nonstationarity. They make use of space time-index plots to probe dynamical nonstationarities in the data. They believe that in a space time-index plot, the density distributions as a function of normalized time-index are V-shaped when the data are nonstationarity.

127 CASE STUDIES Real Data from a Metal Leaching Plant The problem with using actual plant data to evaluate dynamical change detection techniques, is to obtain time series data that are believed (or known) to be nonstationary. One way to do this is by visual inspection, although it can sometimes be misleading. In this case study, data from a metal leaching plant will be used (Figure 4.43). The data set consists of 1747 observations, where the concentration of the valuable metal is the measured variable. The time interval between successive observations is one day. Although the time span of the data set under consideration may seem long (about 5 years), a dynamic change in this process can typically be as result of a change in the composition of the ore. Such a change may take years to manifest. So even if the change detection techniques are only able to detect changes in the process after say a few years, it is still fast compared to the dynamics of the overall process Metal concentration Record index (Time) Figure 4.43: Data from a metal leaching plant. Visual inspection of the data plot suggests that there might be a change in the process parameters roundabout observation 8. The data prior to this point look more or less stationary, but the dynamic behaviour of the process from that point onwards seems to have changed. To remove the obvious linear nonstationarity (changing mean) from the data, a 1-point moving average is subtracted from the observations. The resulting time

128 CASE STUDIES 4-41 series is shown in Figure The potential nonstationarity in the data is now less obvious. 4 3 Metal concentration Record index (Time) Figure 4.44: Linearly adjusted data from the metal leaching plant The fact that the data set is very short might prove to be problematic. When evaluating measures for detecting nonstationarity, large data sets are preferred. Results become increasingly more unreliable with shorter data sets State Space Reconstruction of Metal Leaching Data Being a real-life process, one must assume that there is some degree of measurement noise present in the data. State space reconstruction by means of singular spectral analysis (SSA) is more robust against noisy data when compared to traditional embedding methods. One of the reasons is that the PCA step of SSA filters out some of the unwanted noise, which normally results in a cleaner reconstructed attractor (Kugiumtzis & Christopherson, 1997). It is for this reason that SSA will be used to reconstruct the attractor.

129 CASE STUDIES 4-4 The plot of the autocorrelation function suggests an initial embedding dimension of 5 (Figure 4.45). The eigenvalues of the covariance matrix are given in Figure Each of the eigenvalues explains a certain variance in the data. When the initial embedding dimension for a given data set is large (e.g. > 1), the embedding dimension can be reduced by selecting only the most prominent eigenvalues, i.e. the eigenvalues that explain most of the variance. The embedding dimension is then reduced to the number of eigenvalues chosen and PCA is used to reduce correlations between the reconstructed vectors Correlation [ACF(Tau)] Signal Time Shift (Tau) Figure 4.45: Autocorrelation function of the metal leaching data It is important however to make sure that at least 95% (rule of thumb) of the variance in the data can still be explained by the retained eigenvalues. In this case, there is no need to reduce the dimension as the initial embedding dimension is already at a low value of 5. The reconstructed attractor (Figure 4.47) does not seem to have any geometrical structure, as is expected from a deterministic system. This does not automatically mean that the data are stochastic. Remember that the attractor can only be viewed in three dimensions (two dimensions actually, as it is viewed on a flat surface), although the actual reconstructed attractor is 5-dimensional. In three dimensions, the attractor is

130 CASE STUDIES 4-43 projected onto its three principal axes as determined by the three largest eigenvalues. The first three principal components explain only about 91.6 % of the variance in the data. It is therefore still possible that the attractor has some geometrical structure in five dimensions. The best way to see if the data exhibit some deterministic behaviour is by doing surrogate data analysis. x Eigenvalues Figure 4.46: Eigenvalues of the covariance matrix of the metal leaching data. 1 PC3 (1.7 %) PC (9.7 %) PC1 (51. %) 6 Figure 4.47: Reconstructed attractor for the metal leaching data, projected onto the first three principal components. The variance explained by each principal component is given in brackets.

Delay Coordinate Embedding

Delay Coordinate Embedding Chapter 7 Delay Coordinate Embedding Up to this point, we have known our state space explicitly. But what if we do not know it? How can we then study the dynamics is phase space? A typical case is when

More information

Reconstruction Deconstruction:

Reconstruction Deconstruction: Reconstruction Deconstruction: A Brief History of Building Models of Nonlinear Dynamical Systems Jim Crutchfield Center for Computational Science & Engineering Physics Department University of California,

More information

Modeling and Predicting Chaotic Time Series

Modeling and Predicting Chaotic Time Series Chapter 14 Modeling and Predicting Chaotic Time Series To understand the behavior of a dynamical system in terms of some meaningful parameters we seek the appropriate mathematical model that captures the

More information

Sjoerd Verduyn Lunel (Utrecht University)

Sjoerd Verduyn Lunel (Utrecht University) Universiteit Utrecht, February 9, 016 Wasserstein distances in the analysis of time series and dynamical systems Sjoerd Verduyn Lunel (Utrecht University) S.M.VerduynLunel@uu.nl Abstract A new approach

More information

One dimensional Maps

One dimensional Maps Chapter 4 One dimensional Maps The ordinary differential equation studied in chapters 1-3 provide a close link to actual physical systems it is easy to believe these equations provide at least an approximate

More information

What is Chaos? Implications of Chaos 4/12/2010

What is Chaos? Implications of Chaos 4/12/2010 Joseph Engler Adaptive Systems Rockwell Collins, Inc & Intelligent Systems Laboratory The University of Iowa When we see irregularity we cling to randomness and disorder for explanations. Why should this

More information

1 Random walks and data

1 Random walks and data Inference, Models and Simulation for Complex Systems CSCI 7-1 Lecture 7 15 September 11 Prof. Aaron Clauset 1 Random walks and data Supposeyou have some time-series data x 1,x,x 3,...,x T and you want

More information

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption ANDRÉ NUNES DE SOUZA, JOSÉ ALFREDO C. ULSON, IVAN NUNES

More information

CHAPTER 5 ROBUSTNESS ANALYSIS OF THE CONTROLLER

CHAPTER 5 ROBUSTNESS ANALYSIS OF THE CONTROLLER 114 CHAPTER 5 ROBUSTNESS ANALYSIS OF THE CONTROLLER 5.1 INTRODUCTION Robust control is a branch of control theory that explicitly deals with uncertainty in its approach to controller design. It also refers

More information

6.2 Brief review of fundamental concepts about chaotic systems

6.2 Brief review of fundamental concepts about chaotic systems 6.2 Brief review of fundamental concepts about chaotic systems Lorenz (1963) introduced a 3-variable model that is a prototypical example of chaos theory. These equations were derived as a simplification

More information

Chapter 1: Introduction

Chapter 1: Introduction 1 Chapter 1: Introduction Stochastic Processes The concept of a random variable gives a mathematical formulation to the notion of a quantifiable event whose current value is not exactly known or whose

More information

Chaos, Complexity, and Inference (36-462)

Chaos, Complexity, and Inference (36-462) Chaos, Complexity, and Inference (36-462) Lecture 4 Cosma Shalizi 22 January 2009 Reconstruction Inferring the attractor from a time series; powerful in a weird way Using the reconstructed attractor to

More information

Eksterne eksaminator / External examiner: Dr. P Ntumba Interne eksaminatore / Internal examiners: Prof. I Broere, Prof. JE vd Berg, Dr.

Eksterne eksaminator / External examiner: Dr. P Ntumba Interne eksaminatore / Internal examiners: Prof. I Broere, Prof. JE vd Berg, Dr. VAN / SURNAME: UNIVERSITEIT VAN PRETORIA / UNIVERSITY OF PRETORIA FAKULTEIT NATUUR- EN LANDBOUWETENSKAPPE / FACULTY OF NATURAL AND AGRICULTURAL SCIENCES DEPARTEMENT WISKUNDE EN TOEGEPASTE WISKUNDE / DEPARTMENT

More information

Persistent Chaos in High-Dimensional Neural Networks

Persistent Chaos in High-Dimensional Neural Networks Persistent Chaos in High-Dimensional Neural Networks D. J. Albers with J. C. Sprott and James P. Crutchfield February 20, 2005 1 Outline: Introduction and motivation Mathematical versus computational dynamics

More information

Adaptive Series-Parallel Identification of Dynamical Systems with Uncertain Bifurcations and Chaos

Adaptive Series-Parallel Identification of Dynamical Systems with Uncertain Bifurcations and Chaos Adaptive Series-Parallel Identification of Dynamical Systems with Uncertain Bifurcations and Chaos James T Lo Department of Math and Stat University of Maryland Baltimore County Baltimore, MD, 225 Email:

More information

Generalised density function estimation using moments and the characteristic function

Generalised density function estimation using moments and the characteristic function Generalised density function estimation using moments and the characteristic function Gerhard Esterhuizen Thesis presented in partial fulfilment of the requirements for the degree of Master of Science

More information

Phase-Space Reconstruction. Gerrit Ansmann

Phase-Space Reconstruction. Gerrit Ansmann Phase-Space Reconstruction Gerrit Ansmann Reprise: The Need for Non-Linear Methods. Lorenz oscillator x = 1(y x), y = x(28 z) y, z = xy 8z 3 Autoregressive process measured with non-linearity: y t =.8y

More information

Removing the Noise from Chaos Plus Noise

Removing the Noise from Chaos Plus Noise Removing the Noise from Chaos Plus Noise Steven P. Lalley Department of Statistics University of Chicago November 5, 2 Abstract The problem of extracting a signal x n generated by a dynamical system from

More information

Analysis of Electroencephologram Data Using Time-Delay Embeddings to Reconstruct Phase Space

Analysis of Electroencephologram Data Using Time-Delay Embeddings to Reconstruct Phase Space Dynamics at the Horsetooth Volume 1, 2009. Analysis of Electroencephologram Data Using Time-Delay Embeddings to Reconstruct Phase Space Department of Mathematics Colorado State University Report submitted

More information

NONLINEAR TIME SERIES ANALYSIS, WITH APPLICATIONS TO MEDICINE

NONLINEAR TIME SERIES ANALYSIS, WITH APPLICATIONS TO MEDICINE NONLINEAR TIME SERIES ANALYSIS, WITH APPLICATIONS TO MEDICINE José María Amigó Centro de Investigación Operativa, Universidad Miguel Hernández, Elche (Spain) J.M. Amigó (CIO) Nonlinear time series analysis

More information

Testing for Chaos in Type-I ELM Dynamics on JET with the ILW. Fabio Pisano

Testing for Chaos in Type-I ELM Dynamics on JET with the ILW. Fabio Pisano Testing for Chaos in Type-I ELM Dynamics on JET with the ILW Fabio Pisano ACKNOWLEDGMENTS B. Cannas 1, A. Fanni 1, A. Murari 2, F. Pisano 1 and JET Contributors* EUROfusion Consortium, JET, Culham Science

More information

Bearing fault diagnosis based on EMD-KPCA and ELM

Bearing fault diagnosis based on EMD-KPCA and ELM Bearing fault diagnosis based on EMD-KPCA and ELM Zihan Chen, Hang Yuan 2 School of Reliability and Systems Engineering, Beihang University, Beijing 9, China Science and Technology on Reliability & Environmental

More information

The Behaviour of a Mobile Robot Is Chaotic

The Behaviour of a Mobile Robot Is Chaotic AISB Journal 1(4), c SSAISB, 2003 The Behaviour of a Mobile Robot Is Chaotic Ulrich Nehmzow and Keith Walker Department of Computer Science, University of Essex, Colchester CO4 3SQ Department of Physics

More information

More Details Fixed point of mapping is point that maps into itself, i.e., x n+1 = x n.

More Details Fixed point of mapping is point that maps into itself, i.e., x n+1 = x n. More Details Fixed point of mapping is point that maps into itself, i.e., x n+1 = x n. If there are points which, after many iterations of map then fixed point called an attractor. fixed point, If λ

More information

5.1 2D example 59 Figure 5.1: Parabolic velocity field in a straight two-dimensional pipe. Figure 5.2: Concentration on the input boundary of the pipe. The vertical axis corresponds to r 2 -coordinate,

More information

Chapter 1 Direct Modeling for Computational Fluid Dynamics

Chapter 1 Direct Modeling for Computational Fluid Dynamics Chapter 1 Direct Modeling for Computational Fluid Dynamics Computational fluid dynamics (CFD) is a scientific discipline, which aims to capture fluid motion in a discretized space. The description of the

More information

Introduction Knot Theory Nonlinear Dynamics Topology in Chaos Open Questions Summary. Topology in Chaos

Introduction Knot Theory Nonlinear Dynamics Topology in Chaos Open Questions Summary. Topology in Chaos Introduction Knot Theory Nonlinear Dynamics Open Questions Summary A tangled tale about knot, link, template, and strange attractor Centre for Chaos & Complex Networks City University of Hong Kong Email:

More information

UNIVERSITEIT VAN PRETORIA / UNIVERSITY OF PRETORIA DEPT WISKUNDE EN TOEGEPASTE WISKUNDE DEPT OF MATHEMATICS AND APPLIED MATHEMATICS

UNIVERSITEIT VAN PRETORIA / UNIVERSITY OF PRETORIA DEPT WISKUNDE EN TOEGEPASTE WISKUNDE DEPT OF MATHEMATICS AND APPLIED MATHEMATICS UNIVERSITEIT VAN PRETORIA / UNIVERSITY OF PRETORIA DEPT WISKUNDE EN TOEGEPASTE WISKUNDE DEPT OF MATHEMATICS AND APPLIED MATHEMATICS VAN/SURNAME: VOORNAME/FIRST NAMES: WTW 218 - CALCULUS SEMESTERTOETS /

More information

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore Lecture - 27 Multilayer Feedforward Neural networks with Sigmoidal

More information

Lecture 2: Linear regression

Lecture 2: Linear regression Lecture 2: Linear regression Roger Grosse 1 Introduction Let s ump right in and look at our first machine learning algorithm, linear regression. In regression, we are interested in predicting a scalar-valued

More information

THREE DIMENSIONAL SYSTEMS. Lecture 6: The Lorenz Equations

THREE DIMENSIONAL SYSTEMS. Lecture 6: The Lorenz Equations THREE DIMENSIONAL SYSTEMS Lecture 6: The Lorenz Equations 6. The Lorenz (1963) Equations The Lorenz equations were originally derived by Saltzman (1962) as a minimalist model of thermal convection in a

More information

Comparison of Modern Stochastic Optimization Algorithms

Comparison of Modern Stochastic Optimization Algorithms Comparison of Modern Stochastic Optimization Algorithms George Papamakarios December 214 Abstract Gradient-based optimization methods are popular in machine learning applications. In large-scale problems,

More information

Investigation of Chaotic Nature of Sunspot Data by Nonlinear Analysis Techniques

Investigation of Chaotic Nature of Sunspot Data by Nonlinear Analysis Techniques American-Eurasian Journal of Scientific Research 10 (5): 272-277, 2015 ISSN 1818-6785 IDOSI Publications, 2015 DOI: 10.5829/idosi.aejsr.2015.10.5.1150 Investigation of Chaotic Nature of Sunspot Data by

More information

FUZZY LOGIC CONTROLLER AS MODELING TOOL FOR THE BURNING PROCESS OF A CEMENT PRODUCTION PLANT. P. B. Osofisan and J. Esara

FUZZY LOGIC CONTROLLER AS MODELING TOOL FOR THE BURNING PROCESS OF A CEMENT PRODUCTION PLANT. P. B. Osofisan and J. Esara FUZZY LOGIC CONTROLLER AS MODELING TOOL FOR THE BURNING PROCESS OF A CEMENT PRODUCTION PLANT P. B. Osofisan and J. Esara Department of Electrical and Electronics Engineering University of Lagos, Nigeria

More information

Available online at AASRI Procedia 1 (2012 ) AASRI Conference on Computational Intelligence and Bioinformatics

Available online at  AASRI Procedia 1 (2012 ) AASRI Conference on Computational Intelligence and Bioinformatics Available online at www.sciencedirect.com AASRI Procedia ( ) 377 383 AASRI Procedia www.elsevier.com/locate/procedia AASRI Conference on Computational Intelligence and Bioinformatics Chaotic Time Series

More information

Detection of Nonlinearity and Stochastic Nature in Time Series by Delay Vector Variance Method

Detection of Nonlinearity and Stochastic Nature in Time Series by Delay Vector Variance Method International Journal of Engineering & Technology IJET-IJENS Vol:10 No:02 11 Detection of Nonlinearity and Stochastic Nature in Time Series by Delay Vector Variance Method Imtiaz Ahmed Abstract-- This

More information

MATHEMATICS GRADE 10 TASK 1 INVESTIGATION Marks: 55

MATHEMATICS GRADE 10 TASK 1 INVESTIGATION Marks: 55 WISKUNDE GRAAD 10 TAAK 1 ONDERSOEK Punte: 55 MATHEMATICS GRADE 10 TASK 1 INVESTIGATION Marks: 55 INSTRUKSIES: 1. Die taak moet ingehandig word op 2 Maart 2015. 2. Beantwoord al die vrae. 3. Slegs vrae

More information

Detecting Change in Complex Process Systems with Phase Space Methods

Detecting Change in Complex Process Systems with Phase Space Methods Detecting Change in Complex Process Systems with Phase Space Methods by Paul Jacobus Botha Thesis submitted in partial fulfilment of the requirements for the Degree of MASTER OF SCIENCE IN CHEMICAL ENGINEERING

More information

A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling

A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling G. B. Kingston, H. R. Maier and M. F. Lambert Centre for Applied Modelling in Water Engineering, School

More information

DESIGN OF AN ON-LINE TITRATOR FOR NONLINEAR ph CONTROL

DESIGN OF AN ON-LINE TITRATOR FOR NONLINEAR ph CONTROL DESIGN OF AN ON-LINE TITRATOR FOR NONLINEAR CONTROL Alex D. Kalafatis Liuping Wang William R. Cluett AspenTech, Toronto, Canada School of Electrical & Computer Engineering, RMIT University, Melbourne,

More information

Timeseries of Determinisic Dynamic Systems

Timeseries of Determinisic Dynamic Systems Timeseries of Determinisic Dynamic Systems Sven Laur November 1, 2004 1 Deterministic dynamic systems Starting from Renaissance rationality has dominated in the science. W. G. Leibnitz and I. Newton based

More information

AFAULT diagnosis procedure is typically divided into three

AFAULT diagnosis procedure is typically divided into three 576 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 47, NO. 4, APRIL 2002 A Robust Detection and Isolation Scheme for Abrupt and Incipient Faults in Nonlinear Systems Xiaodong Zhang, Marios M. Polycarpou,

More information

MODULE -4 BAYEIAN LEARNING

MODULE -4 BAYEIAN LEARNING MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities

More information

Revisiting linear and non-linear methodologies for time series prediction - application to ESTSP 08 competition data

Revisiting linear and non-linear methodologies for time series prediction - application to ESTSP 08 competition data Revisiting linear and non-linear methodologies for time series - application to ESTSP 08 competition data Madalina Olteanu Universite Paris 1 - SAMOS CES 90 Rue de Tolbiac, 75013 Paris - France Abstract.

More information

22/04/2014. Economic Research

22/04/2014. Economic Research 22/04/2014 Economic Research Forecasting Models for Exchange Rate Tuesday, April 22, 2014 The science of prognostics has been going through a rapid and fruitful development in the past decades, with various

More information

Recursive Neural Filters and Dynamical Range Transformers

Recursive Neural Filters and Dynamical Range Transformers Recursive Neural Filters and Dynamical Range Transformers JAMES T. LO AND LEI YU Invited Paper A recursive neural filter employs a recursive neural network to process a measurement process to estimate

More information

Lecture 1: A Preliminary to Nonlinear Dynamics and Chaos

Lecture 1: A Preliminary to Nonlinear Dynamics and Chaos Lecture 1: A Preliminary to Nonlinear Dynamics and Chaos Autonomous Systems A set of coupled autonomous 1st-order ODEs. Here "autonomous" means that the right hand side of the equations does not explicitly

More information

Detecting chaos in pseudoperiodic time series without embedding

Detecting chaos in pseudoperiodic time series without embedding Detecting chaos in pseudoperiodic time series without embedding J. Zhang,* X. Luo, and M. Small Department of Electronic and Information Engineering, Hong Kong Polytechnic University, Hung Hom, Kowloon,

More information

No. 6 Determining the input dimension of a To model a nonlinear time series with the widely used feed-forward neural network means to fit the a

No. 6 Determining the input dimension of a To model a nonlinear time series with the widely used feed-forward neural network means to fit the a Vol 12 No 6, June 2003 cfl 2003 Chin. Phys. Soc. 1009-1963/2003/12(06)/0594-05 Chinese Physics and IOP Publishing Ltd Determining the input dimension of a neural network for nonlinear time series prediction

More information

Vulnerability of economic systems

Vulnerability of economic systems Vulnerability of economic systems Quantitative description of U.S. business cycles using multivariate singular spectrum analysis Andreas Groth* Michael Ghil, Stéphane Hallegatte, Patrice Dumas * Laboratoire

More information

Stabilization of Hyperbolic Chaos by the Pyragas Method

Stabilization of Hyperbolic Chaos by the Pyragas Method Journal of Mathematics and System Science 4 (014) 755-76 D DAVID PUBLISHING Stabilization of Hyperbolic Chaos by the Pyragas Method Sergey Belyakin, Arsen Dzanoev, Sergey Kuznetsov Physics Faculty, Moscow

More information

Advanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland

Advanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland EnviroInfo 2004 (Geneva) Sh@ring EnviroInfo 2004 Advanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland Mikhail Kanevski 1, Michel Maignan 1

More information

PHONEME CLASSIFICATION OVER THE RECONSTRUCTED PHASE SPACE USING PRINCIPAL COMPONENT ANALYSIS

PHONEME CLASSIFICATION OVER THE RECONSTRUCTED PHASE SPACE USING PRINCIPAL COMPONENT ANALYSIS PHONEME CLASSIFICATION OVER THE RECONSTRUCTED PHASE SPACE USING PRINCIPAL COMPONENT ANALYSIS Jinjin Ye jinjin.ye@mu.edu Michael T. Johnson mike.johnson@mu.edu Richard J. Povinelli richard.povinelli@mu.edu

More information

Deep Learning Architecture for Univariate Time Series Forecasting

Deep Learning Architecture for Univariate Time Series Forecasting CS229,Technical Report, 2014 Deep Learning Architecture for Univariate Time Series Forecasting Dmitry Vengertsev 1 Abstract This paper studies the problem of applying machine learning with deep architecture

More information

Information Dynamics Foundations and Applications

Information Dynamics Foundations and Applications Gustavo Deco Bernd Schürmann Information Dynamics Foundations and Applications With 89 Illustrations Springer PREFACE vii CHAPTER 1 Introduction 1 CHAPTER 2 Dynamical Systems: An Overview 7 2.1 Deterministic

More information

1.1 OBJECTIVE AND CONTENTS OF THE BOOK

1.1 OBJECTIVE AND CONTENTS OF THE BOOK 1 Introduction 1.1 OBJECTIVE AND CONTENTS OF THE BOOK Hysteresis is a nonlinear phenomenon exhibited by systems stemming from various science and engineering areas: under a low-frequency periodic excitation,

More information

ESTIMATING THE ATTRACTOR DIMENSION OF THE EQUATORIAL WEATHER SYSTEM M. Leok B.T.

ESTIMATING THE ATTRACTOR DIMENSION OF THE EQUATORIAL WEATHER SYSTEM M. Leok B.T. This paper was awarded in the I International Competition (99/9) First Step to Nobel Prize in Physics and published in the competition proceedings (Acta Phys. Pol. A 8 Supplement, S- (99)). The paper is

More information

Andrea Zanchettin Automatic Control AUTOMATIC CONTROL. Andrea M. Zanchettin, PhD Spring Semester, Linear systems (frequency domain)

Andrea Zanchettin Automatic Control AUTOMATIC CONTROL. Andrea M. Zanchettin, PhD Spring Semester, Linear systems (frequency domain) 1 AUTOMATIC CONTROL Andrea M. Zanchettin, PhD Spring Semester, 2018 Linear systems (frequency domain) 2 Motivations Consider an LTI system Thanks to the Lagrange s formula we can compute the motion of

More information

Neural Networks. Nethra Sambamoorthi, Ph.D. Jan CRMportals Inc., Nethra Sambamoorthi, Ph.D. Phone:

Neural Networks. Nethra Sambamoorthi, Ph.D. Jan CRMportals Inc., Nethra Sambamoorthi, Ph.D. Phone: Neural Networks Nethra Sambamoorthi, Ph.D Jan 2003 CRMportals Inc., Nethra Sambamoorthi, Ph.D Phone: 732-972-8969 Nethra@crmportals.com What? Saying it Again in Different ways Artificial neural network

More information

Lecture 9. Time series prediction

Lecture 9. Time series prediction Lecture 9 Time series prediction Prediction is about function fitting To predict we need to model There are a bewildering number of models for data we look at some of the major approaches in this lecture

More information

CSC 411 Lecture 10: Neural Networks

CSC 411 Lecture 10: Neural Networks CSC 411 Lecture 10: Neural Networks Roger Grosse, Amir-massoud Farahmand, and Juan Carrasquilla University of Toronto UofT CSC 411: 10-Neural Networks 1 / 35 Inspiration: The Brain Our brain has 10 11

More information

Session-Based Queueing Systems

Session-Based Queueing Systems Session-Based Queueing Systems Modelling, Simulation, and Approximation Jeroen Horters Supervisor VU: Sandjai Bhulai Executive Summary Companies often offer services that require multiple steps on the

More information

Target Tracking Using Double Pendulum

Target Tracking Using Double Pendulum Target Tracking Using Double Pendulum Brian Spackman 1, Anusna Chakraborty 1 Department of Electrical and Computer Engineering Utah State University Abstract: This paper deals with the design, implementation

More information

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage

More information

DETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION. Alexandre Iline, Harri Valpola and Erkki Oja

DETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION. Alexandre Iline, Harri Valpola and Erkki Oja DETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION Alexandre Iline, Harri Valpola and Erkki Oja Laboratory of Computer and Information Science Helsinki University of Technology P.O.Box

More information

Introduction to Dynamical Systems Basic Concepts of Dynamics

Introduction to Dynamical Systems Basic Concepts of Dynamics Introduction to Dynamical Systems Basic Concepts of Dynamics A dynamical system: Has a notion of state, which contains all the information upon which the dynamical system acts. A simple set of deterministic

More information

Kwadratiese rye - Graad 11

Kwadratiese rye - Graad 11 OpenStax-CNX module: m38240 1 Kwadratiese rye - Graad 11 Wiehan Agenbag Free High School Science Texts Project Based on Quadratic Sequences - Grade 11 by Rory Adams Free High School Science Texts Project

More information

Chapter 2 Chaos theory and its relationship to complexity

Chapter 2 Chaos theory and its relationship to complexity Chapter 2 Chaos theory and its relationship to complexity David Kernick This chapter introduces chaos theory and the concept of non-linearity. It highlights the importance of reiteration and the system

More information

dynamical zeta functions: what, why and what are the good for?

dynamical zeta functions: what, why and what are the good for? dynamical zeta functions: what, why and what are the good for? Predrag Cvitanović Georgia Institute of Technology November 2 2011 life is intractable in physics, no problem is tractable I accept chaos

More information

Discriminative Direction for Kernel Classifiers

Discriminative Direction for Kernel Classifiers Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering

More information

Characterization of normality of chaotic systems including prediction and detection of anomalies

Characterization of normality of chaotic systems including prediction and detection of anomalies University of Iowa Iowa Research Online Theses and Dissertations Spring 2011 Characterization of normality of chaotic systems including prediction and detection of anomalies Joseph John Engler University

More information

Chapter 6 Nonlinear Systems and Phenomena. Friday, November 2, 12

Chapter 6 Nonlinear Systems and Phenomena. Friday, November 2, 12 Chapter 6 Nonlinear Systems and Phenomena 6.1 Stability and the Phase Plane We now move to nonlinear systems Begin with the first-order system for x(t) d dt x = f(x,t), x(0) = x 0 In particular, consider

More information

Balancing and Control of a Freely-Swinging Pendulum Using a Model-Free Reinforcement Learning Algorithm

Balancing and Control of a Freely-Swinging Pendulum Using a Model-Free Reinforcement Learning Algorithm Balancing and Control of a Freely-Swinging Pendulum Using a Model-Free Reinforcement Learning Algorithm Michail G. Lagoudakis Department of Computer Science Duke University Durham, NC 2778 mgl@cs.duke.edu

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction

More information

Stability theory is a fundamental topic in mathematics and engineering, that include every

Stability theory is a fundamental topic in mathematics and engineering, that include every Stability Theory Stability theory is a fundamental topic in mathematics and engineering, that include every branches of control theory. For a control system, the least requirement is that the system is

More information

Sjoerd Verduyn Lunel (Utrecht University)

Sjoerd Verduyn Lunel (Utrecht University) Lorentz Center Leiden, April 16 2013 Wasserstein distances in the analysis of time series and dynamical systems Sjoerd Verduyn Lunel (Utrecht University) Workshop Mathematics and Biology: a Roundtrip in

More information

Khinchin s approach to statistical mechanics

Khinchin s approach to statistical mechanics Chapter 7 Khinchin s approach to statistical mechanics 7.1 Introduction In his Mathematical Foundations of Statistical Mechanics Khinchin presents an ergodic theorem which is valid also for systems that

More information

The Effects of Dynamical Noises on the Identification of Chaotic Systems: with Application to Streamflow Processes

The Effects of Dynamical Noises on the Identification of Chaotic Systems: with Application to Streamflow Processes Fourth International Conference on Natural Computation The Effects of Dynamical Noises on the Identification of Chaotic Systems: with Application to Streamflow Processes Wen Wang, Pieter H.A.J.M. Van Gelder,

More information

Applications of Hurst Coefficient Analysis to Chaotic Response of ODE Systems: Part 1a, The Original Lorenz System of 1963

Applications of Hurst Coefficient Analysis to Chaotic Response of ODE Systems: Part 1a, The Original Lorenz System of 1963 Applications of Hurst Coefficient Analysis to Chaotic Response of ODE Systems: Part 1a, The Original Lorenz System of 1963 Dan Hughes December 2008 Abstract I have coded the process for calculating Hurst

More information

The application of neural networks to the paper-making industry

The application of neural networks to the paper-making industry The application of neural networks to the paper-making industry P. J. Edwards y, A.F. Murray y, G. Papadopoulos y, A.R. Wallace y and J. Barnard x y Dept. of Electronics and Electrical Eng., Edinburgh

More information

Stochastic Gradient Descent. CS 584: Big Data Analytics

Stochastic Gradient Descent. CS 584: Big Data Analytics Stochastic Gradient Descent CS 584: Big Data Analytics Gradient Descent Recap Simplest and extremely popular Main Idea: take a step proportional to the negative of the gradient Easy to implement Each iteration

More information

CONTROLLING IN BETWEEN THE LORENZ AND THE CHEN SYSTEMS

CONTROLLING IN BETWEEN THE LORENZ AND THE CHEN SYSTEMS International Journal of Bifurcation and Chaos, Vol. 12, No. 6 (22) 1417 1422 c World Scientific Publishing Company CONTROLLING IN BETWEEN THE LORENZ AND THE CHEN SYSTEMS JINHU LÜ Institute of Systems

More information

A plane autonomous system is a pair of simultaneous first-order differential equations,

A plane autonomous system is a pair of simultaneous first-order differential equations, Chapter 11 Phase-Plane Techniques 11.1 Plane Autonomous Systems A plane autonomous system is a pair of simultaneous first-order differential equations, ẋ = f(x, y), ẏ = g(x, y). This system has an equilibrium

More information

MATH 415, WEEK 12 & 13: Higher-Dimensional Systems, Lorenz Equations, Chaotic Behavior

MATH 415, WEEK 12 & 13: Higher-Dimensional Systems, Lorenz Equations, Chaotic Behavior MATH 415, WEEK 1 & 13: Higher-Dimensional Systems, Lorenz Equations, Chaotic Behavior 1 Higher-Dimensional Systems Consider the following system of differential equations: dx = x y dt dy dt = xy y dz dt

More information

A benchmark test A benchmark test of accuracy and precision in estimating dynamical systems characteristics from a time series

A benchmark test A benchmark test of accuracy and precision in estimating dynamical systems characteristics from a time series 2 A benchmark test A benchmark test of accuracy and precision in estimating dynamical systems characteristics from a time series, SM Rispens, M Pijnappels, JH van Dieën, KS van Schooten, PJ Beek, A Daffertshofer

More information

Chaos and Liapunov exponents

Chaos and Liapunov exponents PHYS347 INTRODUCTION TO NONLINEAR PHYSICS - 2/22 Chaos and Liapunov exponents Definition of chaos In the lectures we followed Strogatz and defined chaos as aperiodic long-term behaviour in a deterministic

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Stable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems

Stable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems Stable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems Thore Graepel and Nicol N. Schraudolph Institute of Computational Science ETH Zürich, Switzerland {graepel,schraudo}@inf.ethz.ch

More information

Control-oriented model learning with a recurrent neural network

Control-oriented model learning with a recurrent neural network Control-oriented model learning with a recurrent neural network M. A. Bucci O. Semeraro A. Allauzen L. Cordier G. Wisniewski L. Mathelin 20 November 2018, APS Atlanta Kuramoto-Sivashinsky (KS) u t = 4

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Fundamentals of Dynamical Systems / Discrete-Time Models. Dr. Dylan McNamara people.uncw.edu/ mcnamarad

Fundamentals of Dynamical Systems / Discrete-Time Models. Dr. Dylan McNamara people.uncw.edu/ mcnamarad Fundamentals of Dynamical Systems / Discrete-Time Models Dr. Dylan McNamara people.uncw.edu/ mcnamarad Dynamical systems theory Considers how systems autonomously change along time Ranges from Newtonian

More information

November 2005 TYD/TIME: 90 min PUNTE / MARKS: 35 VAN/SURNAME: VOORNAME/FIRST NAMES: STUDENTENOMMER/STUDENT NUMBER: HANDTEKENING/SIGNATURE:

November 2005 TYD/TIME: 90 min PUNTE / MARKS: 35 VAN/SURNAME: VOORNAME/FIRST NAMES: STUDENTENOMMER/STUDENT NUMBER: HANDTEKENING/SIGNATURE: UNIVERSITEIT VAN PRETORIA / UNIVERSITY OF PRETORIA DEPT WISKUNDE EN TOEGEPASTE WISKUNDE DEPT OF MATHEMATICS AND APPLIED MATHEMATICS WTW 263 - NUMERIESE METODES / NUMERICAL METHODS EKSAMEN / EXAMINATION

More information

Research Article Hidden Periodicity and Chaos in the Sequence of Prime Numbers

Research Article Hidden Periodicity and Chaos in the Sequence of Prime Numbers Advances in Mathematical Physics Volume 2, Article ID 5978, 8 pages doi:.55/2/5978 Research Article Hidden Periodicity and Chaos in the Sequence of Prime Numbers A. Bershadskii Physics Department, ICAR,

More information

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M. TIME SERIES ANALYSIS Forecasting and Control Fifth Edition GEORGE E. P. BOX GWILYM M. JENKINS GREGORY C. REINSEL GRETA M. LJUNG Wiley CONTENTS PREFACE TO THE FIFTH EDITION PREFACE TO THE FOURTH EDITION

More information

Terminology and Concepts

Terminology and Concepts Terminology and Concepts Prof. Naga Kandasamy 1 Goals of Fault Tolerance Dependability is an umbrella term encompassing the concepts of reliability, availability, performability, safety, and testability.

More information

Introduction to. Process Control. Ahmet Palazoglu. Second Edition. Jose A. Romagnoli. CRC Press. Taylor & Francis Group. Taylor & Francis Group,

Introduction to. Process Control. Ahmet Palazoglu. Second Edition. Jose A. Romagnoli. CRC Press. Taylor & Francis Group. Taylor & Francis Group, Introduction to Process Control Second Edition Jose A. Romagnoli Ahmet Palazoglu CRC Press Taylor & Francis Group Boca Raton London NewYork CRC Press is an imprint of the Taylor & Francis Group, an informa

More information

THEODORE VORONOV DIFFERENTIABLE MANIFOLDS. Fall Last updated: November 26, (Under construction.)

THEODORE VORONOV DIFFERENTIABLE MANIFOLDS. Fall Last updated: November 26, (Under construction.) 4 Vector fields Last updated: November 26, 2009. (Under construction.) 4.1 Tangent vectors as derivations After we have introduced topological notions, we can come back to analysis on manifolds. Let M

More information

Neural Networks and the Back-propagation Algorithm

Neural Networks and the Back-propagation Algorithm Neural Networks and the Back-propagation Algorithm Francisco S. Melo In these notes, we provide a brief overview of the main concepts concerning neural networks and the back-propagation algorithm. We closely

More information

Synergy between Data Reconciliation and Principal Component Analysis.

Synergy between Data Reconciliation and Principal Component Analysis. Plant Monitoring and Fault Detection Synergy between Data Reconciliation and Principal Component Analysis. Th. Amand a, G. Heyen a, B. Kalitventzeff b Thierry.Amand@ulg.ac.be, G.Heyen@ulg.ac.be, B.Kalitventzeff@ulg.ac.be

More information

In this chapter, we provide an introduction to covariate shift adaptation toward machine learning in a non-stationary environment.

In this chapter, we provide an introduction to covariate shift adaptation toward machine learning in a non-stationary environment. 1 Introduction and Problem Formulation In this chapter, we provide an introduction to covariate shift adaptation toward machine learning in a non-stationary environment. 1.1 Machine Learning under Covariate

More information