Non-linear Measure Based Process Monitoring and Fault Diagnosis

Similar documents
Principles of Pattern Recognition. C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata

EEE 241: Linear Systems

Lecture 4: Feed Forward Neural Networks

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Computational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification

Data Mining Part 5. Prediction

Feature selection and extraction Spectral domain quality estimation Alternatives

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Artificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence

Neural Networks Introduction

Artificial Neural Networks. Edward Gatt

An artificial neural networks (ANNs) model is a functional abstraction of the

Bearing fault diagnosis based on EMD-KPCA and ELM

Part 8: Neural Networks

Iterative Laplacian Score for Feature Selection

Kyle Reing University of Southern California April 18, 2018

Artificial Neural Networks Examination, March 2004

Artificial Neural Networks

Statistical Pattern Recognition

When Dictionary Learning Meets Classification

Introduction to Biomedical Engineering

In the Name of God. Lectures 15&16: Radial Basis Function Networks

8. Lecture Neural Networks

Neural Networks Lecture 2:Single Layer Classifiers

How to do backpropagation in a brain

Unsupervised Learning Methods

6.036 midterm review. Wednesday, March 18, 15

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

ESANN'2001 proceedings - European Symposium on Artificial Neural Networks Bruges (Belgium), April 2001, D-Facto public., ISBN ,

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

LEC 3: Fisher Discriminant Analysis (FDA)

Brief Introduction of Machine Learning Techniques for Content Analysis

Advanced Acoustic Emission Data Analysis Pattern Recognition & Neural Networks Software

Detecting Changes in Global Dynamics with Principal Curves and Information Theory

Unsupervised Feature Extraction by Time-Contrastive Learning and Nonlinear ICA

8. Classification and Pattern Recognition

An introduction to clustering techniques

Introduction to representational similarity analysis

Representational similarity analysis. Nikolaus Kriegeskorte MRC Cognition and Brain Sciences Unit Cambridge, UK

Neural Network to Control Output of Hidden Node According to Input Patterns

Computational statistics

Multilayer Perceptron

Supervised locally linear embedding

Density Estimation: ML, MAP, Bayesian estimation

University of Florida CISE department Gator Engineering. Clustering Part 1

Intelligent Modular Neural Network for Dynamic System Parameter Estimation

Artificial Neural Networks Examination, June 2005

Neural Networks Lecture 4: Radial Bases Function Networks

Deep Learning. Convolutional Neural Network (CNNs) Ali Ghodsi. October 30, Slides are partially based on Book in preparation, Deep Learning

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

Hypothesis Evaluation

Lecture 4: Perceptrons and Multilayer Perceptrons

Measuring time irreversibility using symbolization

Artificial Neural Networks. Historical description

Artificial Neural Network

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015

Does the Wake-sleep Algorithm Produce Good Density Estimators?

Mathematical Tools for Neuroscience (NEU 314) Princeton University, Spring 2016 Jonathan Pillow. Homework 8: Logistic Regression & Information Theory

Simulating Neural Networks. Lawrence Ward P465A

CS 4700: Foundations of Artificial Intelligence

Pattern recognition. "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher

Artificial Neural Networks Examination, June 2004

Introduction to Neural Networks

Statistical Independence and Novelty Detection with Information Preserving Nonlinear Maps

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

Markov Modeling of Time-Series Data. via Spectral Analysis

Pattern Recognition and Machine Learning

Neural Networks. Nethra Sambamoorthi, Ph.D. Jan CRMportals Inc., Nethra Sambamoorthi, Ph.D. Phone:

ECE662: Pattern Recognition and Decision Making Processes: HW TWO

Radial Basis Function Networks. Ravi Kaushik Project 1 CSC Neural Networks and Pattern Recognition

COMS 4771 Introduction to Machine Learning. Nakul Verma

Design of Basic Logic Gates using CMOS and Artificial Neural Networks (ANN)

Effect of number of hidden neurons on learning in large-scale layered neural networks

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi

WP4: Neural Inspired Information Processing. C. Masoller, A. Pons, M.C. Torrent, J. García Ojalvo UPC, Terrassa, Barcelona, Spain

A. Pelliccioni (*), R. Cotroneo (*), F. Pungì (*) (*)ISPESL-DIPIA, Via Fontana Candida 1, 00040, Monteporzio Catone (RM), Italy.

GLR-Entropy Model for ECG Arrhythmia Detection

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine

Lecture 7 Artificial neural networks: Supervised learning

CS:4420 Artificial Intelligence

Artificial Neural Networks. MGS Lecture 2

Aruna Bhat Research Scholar, Department of Electrical Engineering, IIT Delhi, India

Reservoir Computing and Echo State Networks

Artificial Neural Networks

Improved Kalman Filter Initialisation using Neurofuzzy Estimation

Artificial Intelligence

Information Dynamics Foundations and Applications

COMPARING PERFORMANCE OF NEURAL NETWORKS RECOGNIZING MACHINE GENERATED CHARACTERS

Tutorial on Methods for Interpreting and Understanding Deep Neural Networks. Part 3: Applications & Discussion

Artificial neural networks

Unit III. A Survey of Neural Network Model

Neural Networks and the Back-propagation Algorithm

Knowledge Extraction from DBNs for Images

Deep Learning and Information Theory

Introduction to Neural Networks

Active Sonar Target Classification Using Classifier Ensembles

Weight Quantization for Multi-layer Perceptrons Using Soft Weight Sharing

PATTERN CLASSIFICATION

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

Neural Nets in PR. Pattern Recognition XII. Michal Haindl. Outline. Neural Nets in PR 2

Transcription:

Non-linear Measure Based Process Monitoring and Fault Diagnosis 12 th Annual AIChE Meeting, Reno, NV [275] Data Driven Approaches to Process Control 4:40 PM, Nov. 6, 2001 Sandeep Rajput Duane D. Bruns C. Stuart Daw Department of Chemical Engineering, The University of Tennessee, Knoxville Engineering Technology Division [!?], Oak Ridge National Laboratory

Outline 1. Introduction and motivation 2. Review of popular approaches 3. Nonlinear measures and symbolization 4. Methodology 5. Results 6. Discussion 7. Future Directions

Introduction Complex processes Need to know the process state for Improved control and monitoring Fault Diagnosis Event Detection/Predicting failures GOAL Design Expert System to predict process state online

Motivation Most popular methods assume linear structure in data Need to use nonlinear measures for nonlinear and chaotic processes Difference choices for Feature vector formation Measures of similarity This is a study to investigate the potential of nonlinear measures in helping distinguish different nonlinear processes, with obvious extensions to monitoring and fault diagnosis

Popular Approaches Form a feature vector that sufficiently characterizes the process state Choices for feature vectors Power spectral density Wavelet coefficients Vector made by linear or nonlinear model coefficients Classification Schemes Clustering Artificial Intelligence Multivariate Statistical Methods

Nonlinear measures Correlation sum-based measures: characterize the overall geometry The vectors x i and x j are produced by juxtaposing m equally spaced measurements Statistics based on correlation sum or their variants Very Tedious Inaccurate in presence of noise

Nonlinear measures (contd.) Measures for characterize the dynamics Shannon Entropy as a measure of redundancy of information Nonlinear autocorrelation or mutual information Measures computed on code series generated by symbolization. Using fluidized bed data for illustrating the method

Symbolization Look for patterns of values (measurements) Symbolization Analysis Symbolization tries to assign a symbol value to every measurement Every measurement can be called a 0, 1, 2, Helps reduce the noise Preserves the necessary information in less storage space Data compression techniques can be used from information theory Symbol sequences can be made by juxtaposing symbols

Symbolization Symbolization Parameters Set Size Sequence Length Symbolization Interval Number of distinct symbols Number of symbols in the sequence Time difference between symbols in the sequence The optimum values of parameters have to be found by trial and error Symbolization parameters used in this study Set Size = 3 Sequence Length = 5 Symbolization Interval = 1

Illustrating Symbolization Set Size=3 Seq. Length =4 Sym. Interval=1 2 Multiplier 27 9 3 1. Code 2 1 2 1 70 1 Partition #2 1 2 1 0 48 Partition #1 Symbol Sequences 2 1 0 2 1 0 2 1 65 35 0 0 2 1 1 22 2 1 1 2 68 1 1 2 0 42 Code Series 2 1 2 1 0 2 1 1 2 0 1 0 2 2 2 2 0 1 2 0 0 0 1 1 70 48 65

General Methodology Data Preliminary Analysis/ Insight Fourier/Wavelets/Chaos Feature Vector Formation Dimension Reduction Pattern Vectors Clustering Training Neural Net Classification Scheme Validation EXPERT SYSTEM

Feature dimension reduction Fischer s information criterion Discriminatory Power of each feature J 2 µ i, f µ k j, f k f ( i, j) = J j i for k k f (, ) = = k 2 2 σ + σ i, f k j, f k 1,2,..L Choose features with highest discriminatory power

Expert systems Neural Networks Mimic connections between human neurons in the brain Learn by back propagating the error Can learn non-linear or curved separation boundary A priori information is required in most cases Clustering Methods Utilizing the similarity between two vectors to group them Supervised or Unsupervised training Can learn non-linear or curved separation boundary Clustering can be arbitrary A priori information not necessary

Fluidized Bed Out Zone 3 Zone 2 Gas Zone 1 Solid Gas

Process states available 19 flow rates or states available from an experimental fluidized bed (courtesy Dr. Charles Finney) Increasing flow rates (1 through 19) cover fluidization behavior from bubbling to extremely turbulent A sample of time series and power spectral density Note the varying power in different time series segments

Expert System used in this study Clustering Methods K-means Algorithm seeks to minimize the sum of squared variance within clusters Need to know the number of clusters required Euclidean distance used as the measure of [dis]similarity Compare clustering efficiency using Mutual Information (MIF) Power spectral density (PSD)

Features Comparing similar dynamic states 3 8 14 Classification Error (%) MIF 0% 0% 0% PSD 22% 34% 33% Using a biased estimate of classification error. However it should give general idea about clustering efficiency

Comparing very similar dynamic states Features 3 8 14 Classification Error MIF 14% 10% 9% PSD 29% 17% 34% 9 Flow #10 MIF is a better indicator than PSD, but the fact remains that it is hard to distinguish very similar states

Comparing very similar chaotic states 3 8 15 Features Classification Error MIF 42% 38% 29% PSD 39% 48% 54% To distinguish between very similar chaotic states is very difficult

Comparing very similar chaotic states : In Pairs Features 2 5 10 Class 1 Class 2 Classification Error 17 18 17% 13% 7% 18 19 13% 16% 15% Input to the clustering algorithm was the reduced Mutual Information Vector Comparing classes in pairs results in less classification errors. However, the clustering algorithm still can not perfectly distinguish two very close dynamic states

Flow Diagram DATA Feature Dimension Reduction (FIC) Holdout Clustering (K-Means) Distance Metric (Euclidean) VALIDATION FINAL WEIGHTS

Discussion Mutual Information vector can classify similar dynamic states We tested the limits of the scheme by comparing nearly identical dynamic states We demonstrated that mutual information is superior to spectral density as a feature vector in state recognition Can use reduced pattern vector for process monitoring given the state-clusters Advantages Computing Mutual Information at a specified number of lags is quick Less storage space required to store symbolized data Compression techniques from Information theory can be used

Future Directions Neural networks can be used Use different clustering algorithms Use other nonlinear measures A way to deal with slowly changing dynamics Fuzzy clustering

Thank you!