Advanced Methods for Fault Detection

Similar documents
Modern Reliability and Maintenance Engineering for Modern Industry

Fault Detection in Nuclear Power Plants Components by a Combination of Statistical Methods

Condition Monitoring of Electrical Power. Plant Components During Operational

Using Principal Component Analysis Modeling to Monitor Temperature Sensors in a Nuclear Research Reactor

Principal Components Analysis (PCA)

Artificial Neural Networks Francesco DI MAIO, Ph.D., Politecnico di Milano Department of Energy - Nuclear Division IEEE - Italian Reliability Chapter

Introduction to Machine Learning

MACHINE LEARNING ADVANCED MACHINE LEARNING

Lecture 3: Latent Variables Models and Learning with the EM Algorithm. Sam Roweis. Tuesday July25, 2006 Machine Learning Summer School, Taiwan

1 Principal Components Analysis

A Data-driven Approach for Remaining Useful Life Prediction of Critical Components

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Robust signal reconstruction for condition monitoring of industrial components via a modified Auto Associative Kernel Regression method

MACHINE LEARNING ADVANCED MACHINE LEARNING

Advances in Military Technology Vol. 8, No. 2, December Fault Detection and Diagnosis Based on Extensions of PCA

Condition Monitoring for Maintenance Support

Covariance and Correlation Matrix

Dimensionality Reduction

Issues and Techniques in Pattern Classification

Face Recognition and Biometric Systems

Principal Component Analysis

Principal Component Analysis

Independent Component Analysis for Redundant Sensor Validation

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

Machine Learning Lecture 2

Karhunen-Loéve (PCA) based detection of multiple oscillations in multiple measurement signals from large-scale process plants

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

Synergy between Data Reconciliation and Principal Component Analysis.

Principal Component Analysis

Process-Quality Monitoring Using Semi-supervised Probability Latent Variable Regression Models

Metric-based classifiers. Nuno Vasconcelos UCSD

ON MODEL SELECTION FOR STATE ESTIMATION FOR NONLINEAR SYSTEMS. Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof

Likelihood Bounds for Constrained Estimation with Uncertainty

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

FAULT DETECTION AND ISOLATION WITH ROBUST PRINCIPAL COMPONENT ANALYSIS

Kernel-Based Principal Component Analysis (KPCA) and Its Applications. Nonlinear PCA

6.867 Machine learning

Machine Learning Lecture 2

Principal Component Analysis

Principal Component Analysis and Singular Value Decomposition. Volker Tresp, Clemens Otte Summer 2014

Experimental Design and Data Analysis for Biologists

EECS 275 Matrix Computation

7. Variable extraction and dimensionality reduction

Semi-Supervised Laplacian Regularization of Kernel Canonical Correlation Analysis

APPENDIX 1 NEYMAN PEARSON CRITERIA

Classification. The goal: map from input X to a label Y. Y has a discrete set of possible values. We focused on binary Y (values 0 or 1).

Statistical Machine Learning

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Linear Algebra Methods for Data Mining

Advanced Introduction to Machine Learning CMU-10715

Principal component analysis

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

A methodology for fault detection in rolling element bearings using singular spectrum analysis

Keywords: Multimode process monitoring, Joint probability, Weighted probabilistic PCA, Coefficient of variation.

Reservoir Computing Methods for Prognostics and Health Management (PHM) Piero Baraldi Energy Department Politecnico di Milano Italy

Strategy for signal classification to improve data quality for Advanced Detectors gravitational-wave searches

A Data-Driven Model for Software Reliability Prediction

STA 414/2104: Lecture 8

MRC: The Maximum Rejection Classifier for Pattern Detection. With Michael Elad, Renato Keshet

Basics of Multivariate Modelling and Data Analysis

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Preprocessing & dimensionality reduction

Introduction to Signal Detection and Classification. Phani Chavali

Independent Component Analysis. Slides from Paris Smaragdis UIUC

PDEEC Machine Learning 2016/17

CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION

Failure Prognostics with Missing Data Using Extended Kalman Filter

Machine Learning. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Time Series Classification

CHAPTER 3 FEATURE EXTRACTION USING GENETIC ALGORITHM BASED PRINCIPAL COMPONENT ANALYSIS

PCA and admixture models

Plant-wide Root Cause Identification of Transient Disturbances with Application to a Board Machine

Problem Set 3 Due Oct, 5

Lecture 7: Con3nuous Latent Variable Models

Building knowledge from plant operating data for process improvement. applications

STA 414/2104: Lecture 8

Bayes Decision Theory - I

Machine Learning for Software Engineering

Dynamic Data-Driven Adaptive Sampling and Monitoring of Big Spatial-Temporal Data Streams for Real-Time Solar Flare Detection

CS281 Section 4: Factor Analysis and PCA

PCA, Kernel PCA, ICA

Impeller Fault Detection for a Centrifugal Pump Using Principal Component Analysis of Time Domain Vibration Features

Machine learning for pervasive systems Classification in high-dimensional spaces

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Interacting multiple-models, state augmented Particle Filtering for Fault Diagnostics

Discriminant analysis and supervised classification

Info-Greedy Sequential Adaptive Compressed Sensing

Machine Learning 4771

Data Preprocessing Tasks

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier

Neuroscience Introduction

Data reduction for multivariate analysis

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

Least Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e.

COS 429: COMPUTER VISON Face Recognition

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Bearing fault diagnosis based on EMD-KPCA and ELM

PRINCIPAL COMPONENT ANALYSIS

Applied Machine Learning for Biomedical Engineering. Enrico Grisan

Transcription:

Advanced Methods for Fault Detection Piero Baraldi Agip KCO Introduction Piping and long to eploration distance pipelines activities Piero Baraldi

Maintenance Intervention Approaches & PHM Maintenance Intervention Unplanned Planned Corrective Replacement or repair of failed units Scheduled Replacement or Repair following a predefined schedule Conditionbased Monitor the health of the component and then decide on repair actions based on the degradation level assessed Predictive Predict the Remaining Useful Life (RUL) of the component and then decide on repair actions based on the predicted RUL PHM

3 Fault Detection Piero Baraldi 3 Agip KCO Introduction Piping and long to eploration distance pipelines activities 3 Piero Baraldi

Fault Detection: objective 4 Equipment Measured signals Piero Baraldi

Fault Detection: objective 5 5 f Automatic algorithm Normal condition Equipment f Forcing functions Measured signals Piero Baraldi 5

Fault Detection: analogy 6 6 f Automatic algorithm Normal condition Equipment f Forcing functions Measured signals f f Piero Baraldi 6

7 Methods for Fault Detection: Limit-based Data-driven Piero Baraldi 7

8 Methods for Fault Detection: Limit-based Data-driven Piero Baraldi 8

Data & Information for fault detection (I) 9 Normal operation ranges of key signals Eample: Pressurizer of a PWR nuclear reactor Water level. m 3.8 m Abnormal condition Abnormal condition time Normal operation range

Methods for fault detection (I) Normal operation ranges of key signals Limit Value-Based Fault Detection Eample: Pressurizer of a PWR nuclear reactor Water level. m 3.8 m Abnormal condition Abnormal condition time Normal operation range

Methods for fault detection (I) Normal operation ranges of key signals Limit Value-Based Fault Detection Eample: Pressurizer of a PWR nuclear reactor Water level Drawbacks: comple plant too many signals to monitor no early detection control systems operations may hide small anomalies (the signal remains in the normal range although there is a process anomaly) not applicable to fault detection during operational transients. m 3.8 m Abnormal condition Abnormal condition time Normal operation range

Methods for Fault Detection: Limit-based Data-driven

Data & Information for fault detection (III) 3 Normal operation ranges of key signals Historical signal measurements in normal condition Eample: Liquid Steam Pressuretemperattemperat ure ure Spray flow Surge line flow Heaters power Level 5. 3 36 539 44 7. Pressure 5.4 3 363 68 34 7.5 5.3 33 364 69 335 44 7.7 Water level

s The fault detection approach Real measurements MODEL OF COMPONENT BEHAVIOR IN NORMAL CONDITION ŝ ŝ t 4 Signal reconstructions s t Pb. t COMPARISON t s ŝ s ŝ Residuals Pb. t DECISION NORMAL CONDITION: No maintenance ABNORMAL CONDITION: maintenance required Piero Baraldi 4

s The fault detection approach Real measurements MODEL OF COMPONENT BEHAVIOR IN NORMAL CONDITION ŝ ŝ t 5 Signal reconstructions s t Pb. t COMPARISON t s ŝ s ŝ Residuals Pb. t DECISION NORMAL CONDITION: No maintenance ABNORMAL CONDITION: maintenance required Piero Baraldi 5

6 Data-driven fault detection The approach Modeling the component behaviour in normal condition Model requirements Modeling techniques Principal Component Analysis (PCA) AutoAssociative Kernel Regression (AAKR) Eample of application Analysis of the residuals: Sequential Probability Ratio Test (SPRT) Some Conclusions

7 Data-driven fault detection The approach Modeling the component behaviour in normal condition Model requirements Modeling techniques Principal Component Analysis (PCA) AutoAssociative Kernel Regression (AAKR) Eample of application Analysis of the residuals: Sequential Probability Ratio Test (SPRT) Some Conclusions

Model Requirements () 8 NORMAL CONDITION Real measurements Signal reconstructions t t M O D E L nc ˆ nc ˆ t t ACCURACY: ˆ nc 8

Model Requirements () 9 ABNORMAL CONDITION Real measurements Signal reconstructions t M O D E nc ˆ nc ˆ t OK t L t ROBUSTNESS ˆ nc nc 9

Model Requirements (3) ABNORMAL CONDITION Real measurements Signal reconstructions t t M O D E L nc ˆ nc ˆ t t NO OK NO SPILL-OVER

Data-driven fault detection The approach Modeling the component behaviour in normal condition Model requirements Modeling techniques Principal Component Analysis (PCA) AutoAssociative Kernel Regression (AAKR) Eample of application Analysis of the residuals: Sequential Probability Ratio Test (SPRT) Some Conclusions

PRINCIPAL COMPONENT ANALYSIS (PCA) ˆ t t MODEL OF COMPONENT BEHAVIOR IN NORMAL CONDITIONS ˆ t t Piero Baraldi

PCA for fault detection Auto-associative model n Auto- Associative Model nc ˆ nc ˆ nc ˆn nc ˆ i f,,, i,, n n Piero Baraldi

PCA for fault detection Auto-associative model n Auto- Associative Model nc ˆ nc ˆ nc ˆn nc ˆ i f,,, i,, n n Empirical model built using training patterns = historical signal measurements in normal plant condition Signal X nc k N nc nc j kj Nj nc n kn nc Nn Observation Piero Baraldi

PCA for fault detection 5 Training pattern Test pattern: input Test pattern: output X nc k N nc nc nc ˆ ( ˆ j kj Nj (,, ) nc,, ˆ X nc n n ) nc nc n kn nc Nn = historical signal measurements in normal plant condition = measured signals at current time = signal reconstructions (epected values of the signals in normal condition) n PCA nc ˆ nc ˆ nc ˆn Piero Baraldi

What is PCA? 6 PCA: space transformation from an n-dimensional space to a λ-dimensional space (λ<n) Retaining most of the information (loosing the least information). 3 Piero Baraldi

PCA Eample 7 B (,) A (,) X C (, ) Piero Baraldi

PCA Eample 8 B (,) ' A (,) X C (, ) ' U Piero Baraldi

PCA Eample 9 B (,) ' A (,) X C (, ) ' U Piero Baraldi

Basic Concepts 3 Two signals are highly correlated or dependent One is enough! -. -.4 -.6 u s - -. -.8 -.4 -.6 -.8 -..4.6.8..4.6.8 s Piero Baraldi

Basic Concepts 3 Two signals are highly correlated or dependent key underlying phenomena One is enough! Areas of variance in data Focus on directions along which the ervations have high variance -. -.4 -.6 u 8 6 s - -. s -.8 4 -.4 -.6 - -.8-4 -..4.6.8..4.6.8 s -6-3 4 5 s Piero Baraldi

What does PCA do? 3 ) Find Principal Components: PC is the direction of maimum variance PC,, PC n are orthogonal to PC and describe maimum residual variance s 8 6 4 PC PC -4-6 - 3 4 5 s - Piero Baraldi

What does PCA do? 33 ) Find Principal Components: PC is the direction of maimum variance PC,, PC n are orthogonal to PC and describe maimum residual variance 3 Piero Baraldi

What does PCA do? 34 Step ) Find Principal Components: PC is the direction of maimum variance PC,, PC n are orthogonal to PC and describe maimum residual variance s 8 6 4 PC PC -6-3 4 5 Step ) PCA approimation: ignore the PCs of lower significance. - -4 s Lost small information number of dimensions from n= to = 4 Piero Baraldi

How does PCA works? (step ) 35 ) Find principal components: V Compute the covariance matri of X nc 8 V nc nc T nc nc X X X X Empirical mean matri 6 4 - -4-6 -8 - - -8-6 -4-4 6 8 N i i i V.79.5538.5538 8.7539 Piero Baraldi

How does PCA works? (step ) 36 ) Find principal components: V Compute the covariance matri of X nc V nc nc T nc nc X X X X 8 6 4 N i i i - -4-6 -8 V.79.5538.5538 8.7539 - - -8-6 -4-4 6 8 Piero Baraldi

Piero Baraldi PCA Eample : Variance Matri 37 X X X X X X V T T A (,) ), ( C (,) B X X

How does PCA works? (step ) 38 8 ) Find principal components: V Compute the covariance matri of nc X, p,, p n n Find the n eigenvectors of and the corresponding eigenvalues : p V Vp p 6 4 - -4-6 -8 - - -8-6 -4-4 6 8 Piero Baraldi

Piero Baraldi PCA Eample : Eigenvalue 39 V Eigenvalues from: 4 ) det( 4 4 det det ) det( I V I V p Vp

PCA Eample: Eigenvectors 4 Vp p V 4 First eigenvector: p p p Vp p p p p 4 p p p p p 4 p 4 p p p p p p Piero Baraldi

Piero Baraldi PCA Eample: Eigenvectors 4 Second eigenvector: V p Vp p p p p p p p p p p p p p p Vp p p p 4

How does PCA works? (step ) 4 8 ) Find principal components: V Compute the covariance matri of nc X, p,, p n n Find the n eigenvectors of and the corresponding eigenvalues : p V Vp p 6 4 - p p.8.96.96 T.8 T 9.5 highest.4 smallest -4-6 -8 - - -8-6 -4-4 6 8 Piero Baraldi

Eample: the trasformed basis 43 8 6 4 p p p p.8.96.96 T.8 T - -4-6 -8 - - -8-6 -4-4 6 8 Piero Baraldi

How does PCA works? (step ) 44 8 ) Find principal components: V Compute the covariance matri of nc X, p,, p n n Find the n eigenvectors of and the corresponding eigenvalues : p V Vp p 6 4 - p p.8.96.96 T.8 T 9.5 highest.4 smallest -4-6 -8 - - -8-6 -4-4 6 8 P.8.96.96.8 p p, Piero Baraldi

How does PCA works? (step ) 45 ) Find principal components: V Compute the covariance matri of Find the n eigenvalues of and the corresponding eigenvectors p, p,, p n P represents an orthonormal basis: pi p i p PP T I j i j X n nc V Piero Baraldi

Piero Baraldi PCA Eample: P 46 V p Vp 4 p p P

Piero Baraldi PCA eample: P is an orthonormal basis 47 i i i p p p p p I PP T p p p p p

How does PCA works? (step ) 48 ) Find principal components: V Compute the covariance matri of V nc nc nc T nc nc X X X X P ( n, n) Find the n eigenvectors of and the corresponding eigenvalues : X V n P represents an orthonormal basis U the projection of X -nc on the new basis is given by: U X nc P The inverse relation is: nc T X UP Piero Baraldi

PCA Eample : projection of X 49 B (,) A (,) C (, ) X Piero Baraldi

PCA Eample : projection of X 5 C B (,) (, ) p p PC A (,) PC Piero Baraldi

PCA Eample : projection of X 5 PC PC C B (,) (, ) p A (,) p C ( ) A ( ) B (,) PC PC U XP Piero Baraldi

X Eample: projection of on the new basis 5 nc X nc U X nc P PC 8 8 6 6 4 4 p p - - -4-4 -6-6 -8-8 - - -8-6 -4-4 6 8 - - -8-6 -4-4 6 8 PC Piero Baraldi

X Eample: projection of on the new basis 53 nc U X nc P PC 8 8 6 6 4 4 p p - - -4-4 -6-6 -8-8 - - -8-6 -4-4 6 8 - - -8-6 -4-4 6 8 nc T X UP PC Piero Baraldi

Properties of the PCs 54 PCs are uncorrelated p i p j Data can be transformed from the original to the trasformed bases and viceversa without any loss of information (multiplication for P and P T ) PCs are ordered the first few retains the most of the variation present in the data. % Var PC i i i,, n i Piero Baraldi

Why does PCA work? () 55 PC arg w ma Var X nc w Var w ma T X w X w X w T X nc nc T X T T Var X w L( w, ) w Vw w w nc nc nc w w T Vw nc Lagrange multiplier To be maimized! L( w, ) Vw w Eigenvectors T w are solutions of Which eigenvector is maimizing the variance w T Vw? Piero Baraldi L( w, )

Why does PCA work? () 56 Eigenvectors w T are solutions of Vw w Which eigenvector is maimizing the variance w T Vw? T T w Vw w w The eigenvector associated to the largest eigenvalue Piero Baraldi

How does PCA work? (Step ) 57 ) PCA approimation = ignore the PCs of lower significance map the ervations in a subspace identified by the first eigenvectors p n,, p XP 8 nc X nc X P PC 8.8 X nc P 6 6.6 4 4.4. - -4-6 -8 - -4-6 -. -.4 -.6 PC - - -8-6 -4-4 6 8-8 - - -8-6 -4-4 6 8 PC P -.8 - -4-3 - - 3 4 5 p p p Piero Baraldi

How does PCA work? (Step 3) 58 ) PCA approimation = map the ervations in a subspace n identified by the first n eigenvectors p,, p XP.6.4..8 -. -.4 -.6 3) Antitrasform: XP P T -.8 - -4-3 - - 3 4 5 PC 8 6 4 - -4-6 -8 - -3 - - 3 Piero Baraldi

Piero Baraldi 59 PCA for fault detection: Historical data nc Nn Nj nc N kn kj k nc n j nc nc X P nc X Find from

PCA for fault detection: 6 Historical data nc k N nc nc Measured signals at present time: X j kj Nj nc n kn nc Nn Find (,, n ) P from X nc Trasform and project P I m looking at the measurements considering only the directions that are most meaningful in normal condition (directions of maimum variance) Antitrasform ˆ nc P P T Signal reconstructions ˆ ˆ nc normal condition nc abnormal condition Piero Baraldi I loose only the irrelevant noise The process is changed

Eercise Measured signals at present time: Signal reconstructions? Normal or abnormal condition? (, ) 8 6 4 - -4-6 -8 - - -8-6 -4-4 6 8 available historical signal measurements in normal plant condition Piero Baraldi

Eercise : Solution Measured signals at present time: p Step : find principal components:, (, p ) 8 8 6 6 4 4 p p - - -4-4 -6-6 -8-8 - - -8-6 -4-4 6 8 - - -8-6 -4-4 6 8 available historical signal measurements in normal plant condition Piero Baraldi

Eercise : Solution Measured signals at present time: p Step : find principal components, p (, ) 8 8 6 6 4 4.8 p p.6.4. - - -. -.4-4 -4 -.6-6 -6 -.8 - -4-3 - - 3 4 5-8 - - -8-6 -4-4 6 8-8 - - -8-6 -4-4 6 8 PC Piero Baraldi

Eercise : Solution Measured signals at present time: p Step : find principal components, (, Step (PCA approimation): keep only PC of p ),i.e. p 8 8 6 6 4 4.8 p p.6.4. - - -. -.4-4 -4 -.6-6 -6 -.8 - -4-3 - - 3 4 5-8 - - -8-6 -4-4 6 8-8 - - -8-6 -4-4 6 8 PC Piero Baraldi

Eercise : Solution Measured signals at present time: Step : find principal components (, Step (PCA approimation): keep only PC of Step 3 (antitrasform): 8 6 measuraments reconstruction ˆ nc P P T ),i.e. p 4 - -4-6 -8 - - -8-6 -4-4 6 8 ˆ Piero Baraldi nc normal condition

Eercise Measured signals at present time: Signal reconstructions? Normal or abnormal condition? (, ) 8 6 4 - -4-6 -8 - - -8-6 -4-4 6 8 available historical signal measurements in normal plant condition Piero Baraldi

Eercise : Solution Measured signals at present time: Step : find principal components p (, p Step (PCA approimation): keep only PC of ),i.e. p 8 8 6 6.8 4 4.6.4 p p. - - -. -.4-4 -4 -.6-6 -6 -.8-8 - - -8-6 -4-4 6 8-8 - - -8-6 -4-4 6 8 - -4-3 - - 3 4 5 PC Piero Baraldi

Eercise : Solution Measured signals at present time: Step : find principal components p (, p Step (PCA approimation): keep only PC of ),i.e. p 8 8 6 6.8 4 4.6.4 p p. - - -. -.4-4 -4 -.6-6 -6 -.8-8 - - -8-6 -4-4 6 8-8 - - -8-6 -4-4 6 8 - -4-3 - - 3 4 5 PC Piero Baraldi

Eercise : Solution Measured signals at present time: Step : find principal components (, Step (PCA approimation): keep only PC of Step 3 (antitrasform): 8 ˆ nc P P T ),i.e. p 6 4 measuraments reconstruction - -4-6 -8 ˆ Piero Baraldi nc - - -8-6 -4-4 6 8 Abnormal condition

PCA remarks 7 Computational time: Training time = computational time necessary to find the Principal Components is proportional to the number of measured signals (n) Eecution time: very short (only matri multiplications) OK for online applications Piero Baraldi

PCA remarks 7 Performance: Accuracy = satisfactory t t M O D E L ˆ t t Piero Baraldi

PCA remarks 7 Performance: Accuracy = satisfactory Low robustness and spillover effects on highly correlated signals t t M O D E L ˆ t t Unsatisfactory for dataset characterized by highly non linear relationships. Piero Baraldi

73 Data-driven fault detection The approach Modeling the component behaviour in normal condition Model requirements Modeling techniques Principal Component Analysis (PCA) AutoAssociative Kernel Regression (AAKR) Eample of application Analysis of the residuals: Sequential Probability Ratio Test (SPRT) Some Conclusions

Auto Associative Kernel Regression (AAKR)

What is AAKR? Auto-associative model n Auto- Associative Model ˆ ˆ ˆn ˆ i f,,, i,, n n Empirical model built using training patterns = historical signal measurements in normal plant condition Signal nc nc j n nc X k kj kn nc nc N Nj Nn Observation 75

76 How does AAKR work? Training pattern Test pattern: input Output X nc k N nc nc j kj Nj (,, n ) nc nc nc ˆ ( ˆ,, ˆ ) X n nc nc n kn nc Nn = historical signal measurements in normal plant condition = measured signals at current time = signal reconstructions (epected values of the signals in normal condition) nc ˆ n AAKR nc ˆ nc ˆn 76

77 How does AAKR work? Training pattern X Test pattern: input Test pattern: output nc k N nc nc j kj Nj (,, n ) nc nc nc ˆ ( ˆ,, ˆ ) n nc n kn nc Nn = historical signal measurements in normal plant condition = measured signals at current time = weighted sum of the training patterns: 77

78 How does AAKR work? Training pattern X Test pattern: input Test pattern: output nc k N nc nc j kj Nj (,, n ) nc nc nc ˆ ( ˆ,, ˆ ) n nc n kn nc Nn = historical signal measurements in normal plant condition = measured signals at current time = weighted sum of the training patterns: On all the training pattern ˆ nc j N k w( k) N k w( k) nc kj 78

79 How does AAKR work? Output nc ˆ ( ˆ nc,, ˆ nc n ) = weighted sum of the training patterns: On all the training pattern ˆ nc j N k w( k) k nc kj weights w(k) = similarity measures between (the test and the k-th training pattern): n w N w( k) d ( k ) h ( k) e h and nc k nc with d ( k) ( j ) Euclidean distance between and j kj low weight nc k high weight h = bandwidth parameter 79

Bandwidth parameter d= w=.4/h d=h w=.4/h d=h w=.5/h d=3h w=.4/h w d h w( d 3h).4.4 6 w 4 w 8 6 h=. h= 4-6 -4-4 6 d 8

Eample Signal values at current time: Signal reconstructions? Normal or abnormal condition? (,, n ) available historical signal measurements in normal plant condition 8

Eample : Solution Signal values at current time: ( ˆ nc nc ˆ,, (,, n nc ˆ n ) Signal reconstructions: based on the available historical signal measurements in normal plant condition ) ˆ nc normal condition 8

Eample Signal values at current time: Signal reconstructions? Normal or abnormal condition? (,, n ) available historical signal measurements in normal plant condition 83

84 Eample : Solution Signal values at current time: ( ˆ nc nc ˆ,, (,, n nc ˆ n ) Signal reconstructions: based on the available historical signal measurements in normal plant condition ) ˆ nc abnormal condition available historical signal measurements in normal plant condition 84

85 AAKR: Computational Time Computational time: No training of the model Test: computational time depends on the number of training patterns (N) and on the number of signals (n) d ( k) n j ( j nc kj ) 85

86 AAKR Performance: Accuracy Accuracy: depends on the training set: N Accuracy 86

87 AAKR Performance: Accuracy () Accuracy: depends on the training set: N Accuracy Few patterns and not well distributed in the training space Inaccurate reconstruction 87

AAKR Robustness and resistance to the spill-over effect 88

Case study: Monitoring a Turbine for Energy Production 89 6 Temperature Sensors in different position 3 566 58 564 575 56 57 56 565 558 56 556 558 556 554 55 55 548 546 56 56 558 556 554 55 55 554 555 544 548 55 55 3 4 5 6 7 564 54 3 4 5 6 7 56 546 3 4 5 6 7 568 56 566 56 558 555 564 56 556 554 55 56 558 4 5 6 55 55 548 545 556 554 55 546 3 4 5 6 7 54 3 4 5 6 7 55 3 4 5 6 7 Highly Correlated Signals Abnormal conditions occur 3 4 5 6.97.98.98.99.98.97 on the first.95 signal.99.98.96 3.98.95.96.99.99 4.98.99.96.98.97 5.99.98.99.98.99 6.98.96.99.97.99

Residual Residual Case study: Traditional AAKR 55 4 6 8 9 545 4 6 8 5 5 Traditional AAKR Simulated abnormal residuals conditions -5 4 6 8 Time - - -3 Traditional AAKR Simulated abnormal residuals conditions -4 4 6 8 Time DELAY IN THE DETECTION IMPOSSIBILITY TO IDENTIFY THE SIGNALS TRIGGERING THE ABNORMAL BEHAVIOR

Our Contribution: A modified AAKR method 9 Ԧ - Δ Ԧ Fault Detection Fault No Fault Modified AAKR Ԧ nc Ԧ Pre-processing ψ Ԧ Kernel-Based Reconstruction Ԧ nc Projecting the measurement and the historical data in a new space Ԧ, Ԧ hist k ψ Ԧ, ψ Ԧ hist k Ԧ hist Ԧhist Ԧ Malfunctions causing variations of a small number of signals are more frequent than those causing variations of a large number of signals ψ Ԧ ψ Ԧ hist ψ Ԧ hist Ԧ Ԧ hist Ԧ hist Dilatation matri ψ Ԧ hist = D Ԧp P perm Ԧ hist Permutation matri Modification of the loci of equisimilarity points

Residual Residual Results 9 6 4 Modified AAKR Traditional AAKR Simulated abnormal conditions - 8-6 -3 4 3 4 5 6 Time -4-5 -6 Modified AAKR Traditional AAKR Simulated abnormal conditions 3 4 5 6 Time Early Detection 3 Correct Diagnosis of the signal that triggers the alarms More Accurate 4 5 6

Application to a real case study 93

Application to a real case study 94

Application to a real case study 95

Visual Comparison with historical normal condition values 96

Visual Comparison with historical normal condition values 97

Correlation coefficients 98 Historical Values Modified AAKR reconstruction Traditional AAKR reconstruction

Modified AAKR 99 Traditional AAKR Modify AAKR Loci of equisimilarity points Ԧ Ԧ hist Ԧ hist Ԧ Ԧ hist Ԧ hist Accuracy OK! OK! Robustness NO! Especially with correlated signals. Robust reconstruction of the values epected in normal conditions. Correct identification of signals affected by abnormal condition 3. Good performance with correlated signals

Data-driven fault detection The approach Modeling the component behaviour in normal condition Model requirements Modeling techniques Principal Component Analysis (PCA) AutoAssociative Kernel Regression (AAKR) Eample of application Analysis of the residuals: Sequential Probability Ratio Test (SPRT) Some Conclusions

Problem Statement (EdF): hundreds of signals Grouping AUTO-ASSOCIATIVE MODEL OF PLANT BEHAVIOR IN NORMAL CONDITION MEASURED SIGNALS COMPARISON MODEL MODEL MODEL m

Challenge: Optimal Grouping MEASURED SIGNALS COMPARISON HOW TO GROUP SIGNALS? MODEL MODEL MODEL m

Case Study (EdF) 3 COMPONENT TO BE MONITORED Reactor Coolant Pumps of a PWR Nuclear Power Plant 4 MEASURED 94 signals, 48 of which considered SIGNALS as most important by EdF eperts RECONSTRUCTION Auto-Associative Kernel Regression MODEL (AAKR) historical measurements test measurements reconstruction

Approaches to signal grouping 4 HOW TO SPLIT THE SIGNALS INTO SUBGROUPS? A-PRIORI CRITERIA (signals divided on the basis of ) location in the power plant correlation physical homogeneity others

Results 5 BEST A-PRIORI GROUPING: correlation signals WHITE: high correlation (.7 ] 33 GREY: medium correlation (.4.7] BLACK: low correlation [.4] signal belonging to group signal belonging to group signal belonging to group 3 signal belonging to group 4 signal belonging to group 5 4 signals CRITICALITY: how to group the signals which have a low degree of correlation with all the others?

Approaches to Signal Grouping 6 HOW TO SPLIT THE SIGNALS INTO SUBGROUPS? A-PRIORI CRITERIA (signals divided on the basis of ) location in the power plant correlation physical homogeneity others WRAPPER APPROACH n SIGNALS SEARCH ENGINE SEARCH ALGORITHM CANDIDATE GROUPS AAKR PERFORMANCE EVALUATION OPTIMAL GROUPING

The Wrapper Approach: Genetic Algorithms 7 SEARCH ENGINE Genetic Algorithms CHROMOSOME n genes = n signals GENE signal signal signal 3 signal i signal n GROUP LABEL (integer number) 3 3 3 PERFORMANCE EVALUATION fitness = Accuracy Robustness

Results 8 OBTAINED GROUPINGS correlation GA MAIN DIFFERENCE: signals which have a low degree of correlation are divided in groups and/or mied with signals with a high degree of correlation

Results 9 correlation GA Accuracy = = Robustness NO Spill-Over = =

Results: Wrapper Approach AUTO-ASSOCIATIVE MODEL OF PLANT BEHAVIOR IN NORMAL CONDITIONS ŝ t COMPARISON MEASURED SIGNALS s s ŝ (SEAL OUTCOMING FLOW) t DECISION t ABNORMAL CONDITION: seal deterioration NORMAL CONDITION ABNORMAL CONDITION

Conclusions groups of correlated signals accurate AAKR reconstructions; big groups robust AAKR reconstructions; the wrapper GA search finds the optimal groups also for those signals with low degree of correlation with all the others.

Data-driven fault detection The approach Modeling the component behaviour in normal condition Model requirements Modeling techniques Principal Component Analysis (PCA) AutoAssociative Kernel Regression (AAKR) Eample of application Analysis of the residuals: Sequential Probability Ratio Test (SPRT) Some Conclusions

Model based fault detection 3 SIGNAL RECONSTRUCTIONS ŝ MEASURED SIGNALS s MODEL OF COMPONENT BEHAVIOR IN NORMAL CONDITIONS ŝ t t t s COMPARISON t RESIDUALS Pb. s ŝ s ŝ t DECISION t NORMAL CONDITION: No maintenance ABNORMAL CONDITION: maintenance required 3

Abnormal condition detection: decision Basics of the decision: residual analysis r ˆ r r Normal condition Abnormal condition Methods Threshold-based approaches Stochastic approaches 4

Abnormal condition detection: decision Basics of the decision: residual analysis r ˆ r r Normal condition Abnormal condition Methods Threshold-based approaches Stochastic approaches 5

Threshold-based approaches r Abnormal condition detection t r Too small thresholds high false alarm rates (α) Normal conditions t Too large thresholds high missing alarm rates (β) r Abnormal conditions t 6

Stochastic approaches Residual (r)= random variable described by a probability law The probability law is different in case of normal/abnormal condition.5 Normal condition.5 Abnormal condition.4.4.3.3.... -6-4 - 4 6 r r ~ N(, ) -6-4 - 4 6 r r~ N(, ) 7

Stochastic approach: Sequential Probability Ratio Test (SPRT) R r r,, sequence of residuals at time t,t,,t N n, r N Binary hypothesis test: Null hypothesis (H ) Normal condition ~ i i, r N(, ).5.4.3 f r, N.. -6-4 - 4 6 r Alternative hypothesis (H ) Abnormal condition ~ i i, r N(, ).5.4.3 f r, N.. -6-4 - 4 6 r 8

SPRT: the decision L n P P R R n n H H is true is true f f r fr frn r fr frn ln L n H is true B Not enough information A H is true 9

SPRT Theorem False alarm B A ln ln P P D H is true D H is true Missing alarm

SPRT for the positive mean test Null hypothesis (H ) Normal condition Alternative hypothesis (H ) Abnormal condition ), N( r ~ ), ( N r ~ Sequential Formula! = ) ln( ln n n n n k k n k k n r L r r r L. ),,, ( ),,, ( n k k n k r k r n n n e e H r r r P H r r r P L

Eample: SPRT B ln ln L n H is true L ln ln ln( L L lnl L lnl ) r r 3 Not enough information Not enough information t A ln H is true

SPRT: parameters to be set the residual variance in normal condition (σ ) the epected offset amplitude (μ ) the maimum acceptable false alarm rate (α) the maimum acceptable missing alarm rate (β) 3

Signal value Eample: Reactor Coolant Pumps in a PWR Time interval Offset [-] No [-4] Yes (amplitude =.) [4-6] Yes (amplitude =.3) [6-8] Yes (amplitude =.34) [8-] Yes (amplitude =.46) 53 5 5 5 49 48 47 erved True signal value Reconstruction 46 3 4 5 6 7 8 9 Time 4

Residuals Eample: residuals.8.6.4. -. -.4 -.6 -.8 3 4 5 6 7 8 9 Time No offset Maimum offset 5

L ln ln Eample: SPRT Parameter Value α. Β. μ μ.46 σ. L L L L r r H B ln is true SPRT inde 5 4 3 - - -3 Upper Threshold Lower Threshold SPRT value A ln H is true -4-5 3 4 5 6 7 8 9 Time No offset Maimum offset 6

SPRT: performance Average Sample Number (ASN) necessary to deliver a decision Time interval Offset Estimated ASN Number of times in which a normal condition has been detected Number of times in which an abnormal condition has been detected [-] No. 5 [-4] amplitude =..9 7 5 [4-6] amplitude =.3.4 5 7 [6-8] amplitude =.34. 94 [8-] amplitude =.46. 4 7

8 Data-driven fault detection The approach Modeling the component behaviour in normal condition Model requirements Modeling techniques Principal Component Analysis (PCA) AutoAssociative Kernel Regression (AAKR) Eample of application Analysis of the residuals: Sequential Probability Ratio Test (SPRT) Some Conclusions

9 Fault detection challenges Fault detection systems are currently used to monitor component of comple industrial plants (e.g. electric power plants, transportation systems, process plants) Commercial software for fault detection is available A Challenge: Adaptability 9

Fault Detection challenge: adaptability of the methods 3 Eample: monitoring the turbine of an electric power plant t Identify historical signal measurements in normal plant operation P P Develop the model False alarm rate Plant operations slowly change with time T T t t time A New Model is necessary 3

Fault Detection challenge: adaptability of the methods 3 Eample: monitoring the turbine of an electric power plant t Identify historical signal measurements in normal plant operation P P Develop the model False alarm rate Plant operations slowly change with time T T t t time t Identify recent historical signal measurements in normal plant operation P Develop the new model P A New Model is necessary False alarm rates T T t t time Periodic Human Interventions for developing new models! high costs! 3

Fault detection challenge: adaptability of the methods 3 The detection model should be able to follow the process changes: Incremental learning of the new data that gradually becomes available No necessity of human intervention for: selecting recent normal operation data building the new model New data are coming P P T T Automatic updating of the model P T 3