Elman Recurrent Neural Network in Thermal Modeling of Power Transformers

Similar documents
Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

Transformer Top-Oil Temperature Modeling and Simulation

A STATE-SPACE NEURAL NETWORK FOR MODELING DYNAMICAL NONLINEAR SYSTEMS

Neural Networks and the Back-propagation Algorithm

POWER SYSTEM DYNAMIC SECURITY ASSESSMENT CLASSICAL TO MODERN APPROACH

A New Look at Nonlinear Time Series Prediction with NARX Recurrent Neural Network. José Maria P. Menezes Jr. and Guilherme A.

Neural Network Identification of Non Linear Systems Using State Space Techniques.

Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions

ECE Introduction to Artificial Neural Network and Fuzzy Systems

Chapter 4 Neural Networks in System Identification

Load Forecasting Using Artificial Neural Networks and Support Vector Regression

Neural Networks Introduction

Accurate Location of Transformer Hottest Spot by FEM and Thermal Models

ANN Control of Non-Linear and Unstable System and its Implementation on Inverted Pendulum

Temporal Backpropagation for FIR Neural Networks

An Inverse Vibration Problem Solved by an Artificial Neural Network

T Machine Learning and Neural Networks

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Supervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir

Artificial Intelligence

Estimation of Current Transformer Insulation Life with an Effect of Environmental Variables

Intelligent Modular Neural Network for Dynamic System Parameter Estimation

Experimental study of dynamic thermal behaviour of an 11 kv distribution transformer

4. Multilayer Perceptrons

Complexity Bounds of Radial Basis Functions and Multi-Objective Learning

Adaptive Inverse Control

Artificial Neural Network

Introduction to Neural Networks

Artificial Neural Network Simulation of Battery Performance

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

Online Identification And Control of A PV-Supplied DC Motor Using Universal Learning Networks

Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters

Designing Dynamic Neural Network for Non-Linear System Identification

Estimation of the Pre-Consolidation Pressure in Soils Using ANN method

Deep Feedforward Networks

DETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION. Alexandre Iline, Harri Valpola and Erkki Oja

Using Neural Networks for Identification and Control of Systems

Mr. Harshit K. Dave 1, Dr. Keyur P. Desai 2, Dr. Harit K. Raval 3

A recursive algorithm based on the extended Kalman filter for the training of feedforward neural models. Isabelle Rivals and Léon Personnaz

Analysis of Multilayer Neural Network Modeling and Long Short-Term Memory

Recurrent neural networks with trainable amplitude of activation functions

1. Introduction. 2. Artificial Neural Networks and Fuzzy Time Series

Data Mining Part 5. Prediction

New Recursive-Least-Squares Algorithms for Nonlinear Active Control of Sound and Vibration Using Neural Networks

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

STRUCTURED NEURAL NETWORK FOR NONLINEAR DYNAMIC SYSTEMS MODELING

Extensions to the Finite Element Technique for the Magneto-Thermal Analysis of Aged Oil Cooled-Insulated Power Transformers

Deep Feedforward Networks

Lazy learning for control design

Christian Mohr

Identification of two-mass system parameters using neural networks

Dynamic Behavior of Three phase Inductions Motors as Loads in an Electric Power System with Distributed Generation, a Case of Study.

Parallel layer perceptron

P. M. FONTE GONÇALO XUFRE SILVA J. C. QUADRADO DEEA Centro de Matemática DEEA ISEL Rua Conselheiro Emídio Navarro, LISBOA PORTUGAL

Neural Networks Lecture 4: Radial Bases Function Networks

NONLINEAR PLANT IDENTIFICATION BY WAVELETS

A Method for Estimating Transformer Temperatures and Elapsed Lives Considering Operation Loads

HOTTEST SPOT AND LIFE EVALUATION OF POWER TRANSFORMER DESIGN USING FINITE ELEMENT METHOD

A NEURAL FUZZY APPROACH TO MODELING THE THERMAL BEHAVIOR OF POWER TRANSFORMERS

Deep Feedforward Networks

Lecture 7 Artificial neural networks: Supervised learning

IEEE TRANSACTIONS ON POWER DELIVERY, VOL. 18, NO. 4, OCTOBER

y(x n, w) t n 2. (1)

Study of a Transformer Thermal System

Deep Feedforward Networks. Sargur N. Srihari

Solar Activity Forecasting on by Means of Artificial Neural Networks

Comparative analysis of neural network and regression based condition monitoring approaches for wind turbine fault detection

Learning and Memory in Neural Networks

A New Model Reference Adaptive Formulation to Estimate Stator Resistance in Field Oriented Induction Motor Drive

System Identification for the Hodgkin-Huxley Model using Artificial Neural Networks

Different Techniques for Calculating Apparent and Incremental Inductances using Finite Element Method

Neural Network Based Methodology for Cavitation Detection in Pressure Dropping Devices of PFBR

Research Article Dynamical Models for Computer Viruses Propagation

inear Adaptive Inverse Control

Artificial Neural Networks

CSTR CONTROL USING MULTIPLE MODELS

Unit 8: Introduction to neural networks. Perceptrons

An Improved Method for Determining Voltage Collapse Proximity of Radial Distribution Networks

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Lecture 4: Perceptrons and Multilayer Perceptrons

Reinforcement Learning, Neural Networks and PI Control Applied to a Heating Coil

Introduction to feedforward neural networks

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

THERMAL BEHAVIOR OF DISTRIBUTION TRANSFORMERS IN SUMMERTIME AND SEVERE LOADING CONDITIONS

Accuracy improvement program for VMap1 to Multinational Geospatial Co-production Program (MGCP) using artificial neural networks

Multilayer Perceptron

A Comparison Between Multilayer Perceptron and Fuzzy ARTMAP Neural Network in Power System Dynamic Stability

( t) Identification and Control of a Nonlinear Bioreactor Plant Using Classical and Dynamical Neural Networks

Neural networks. Chapter 19, Sections 1 5 1

Non-linear Predictive Control with Multi Design Variables for PEM-FC

Short Term Load Forecasting Based Artificial Neural Network

Modeling Economic Time Series Using a Focused Time Lagged FeedForward Neural Network

Neural Network Based Response Surface Methods a Comparative Study

MODELING USING NEURAL NETWORKS: APPLICATION TO A LINEAR INCREMENTAL MACHINE

Nonlinear system modeling with deep neural networks and autoencoders algorithm

Modelling and Control of Madeira Wine Aging process Using Directive Inverse Control

Artificial Neural Network and Fuzzy Logic

EE-588 ADVANCED TOPICS IN NEURAL NETWORK

Introduction to Neural Networks: Structure and Training

Transcription:

Elman Recurrent Neural Network in Thermal Modeling of Power Transformers M. HELL, F. GOMIDE Department of Computer Engineering and Industrial Automation - DCA School of Electrical and Computer Engineering - FEEC State University of Campinas - UNICAMP Av. Albert Einstein, 0, Cidade Universitria Zeferino Vaz, 183-9, Campinas, SP BRAZIL L. SECCO, and P. COSTA Jr. Graduate Program in Electrical Engineering Pontifical Catholic University of Minas Gerais Av. Dom José Gaspar, 0, Prédio 3, 535-610, Belo Horizonte, MG BRAZIL Abstract: This work suggests a Elman Recurrent Networks (ELRN) as a means to model thermal condition of power transformers. Experimental results with actual data reported in the literature show that ELRN modeling requires less computational effort, and is more robust and efficient than multilayer feedforward networks, radial basis function network, and classic deterministic modeling approaches. Key-Words: Power transformers; recurrent networks; Elman networks; thermal modeling. 1 Introduction Power transformers are key pieces in transmission and distribution of electric energy. Transformer failures can profoundly affect power systems, causing potentially widespread power outages and, consequently, economic and social losses. Condition monitoring is very important to guarantee safe operation and to increase power availability and service life of power transformers. One main factor in power transformer operation concerns its working temperature, that is, its hot-spot temperature, the temperature at the top or in the center of the high or low voltage windings. The hot-spot temperature has a key role in insulation aging and in service life of power transformers. Monitoring of hotspot temperature is vital to develop overload protection systems. High hot-spot temperature means acceleration in transformer aging. Dynamic modeling of the thermal behavior is one of the most important aspects to monitor power transformer conditions. Comparisons between measured and predicted working temperature provide information about transformer conditions and indicate potential anomalies. Load capability of power transformers can be found using transient heat equations and their specific thermal characteristics [4]. Currently, the use of conventional modeling techniques to determine the internal temperature of power transformers is a challenging task. Security induces adoption of conservative safety factors in computation and often this means underutilized transformers once hot-spot portions of the coils are kept from overheating and failing prematurely. Consequently, the maximum power transfer computed may be 20-% less than nominal power transformer capacity [5]. To increase system operational margins during overload periods, new loading approaches are needed to bring power transformers beyond the corresponding nominal specifications. The approaches should be capable to timely change load capability rates to use full transformer capacity, meaning less revenue losses and lower maintenance and upgrade costs. Several modeling techniques are available today,

artificial neural networks-based modeling (ANN) being among the most relevant and efficient. ANN models, especially multilayer feedforward structures, have been used extensively in system modeling due to its ability to learn complex non-linear relationships that are difficult for conventional approaches [8]. Feedforward neural models do not have internal dynamics. They are static, may not comply with actual system dynamics and fail to provide efficient models. Despite the efficiency of supervised learning schemes, the ones that use input/output system data, imprecision and inadequacy of input/output data [8] may potentially harm feedforward ANN performance. An alternative that has received considerable attention and shown significant progress during the last years concerns recurrent network modeling approaches. Similar to static feedforward neural networks, dynamic recurrent neural neural networks consist of global recursive network and partial recursive network [1]. In the global recursive network, there are full connections between each two nodes, and all weights are trainable. Because complex recurrent connections of the nodes in fully recurrent neural networks result in poor convergence speed. In contranst, in the partial recursive network, feedforward loop is the basic structure in which the weights are variable, and the back-forward loop consists of Context in which the weights are fixed [2]. So it is investigated widely and deeply. Elman neural network is a partial recurrent network model introduced by Elman [3] that lies somewhere between a classic feddforward multilayer perceptron and a pure recurrent network. Because of the existence of contest nodes and local recurrent connections between the context layer and the hidden layer, it has certain dynamical advantages over static neural networks, such a Multi- Perceptron (MLP) and Radial Basis Function Networks (RBFN). This highlight makes Elman neural network very suitable to be utilized in the dynamic system modeling problem. This paper suggests a Elman Recurrent Network (ELRN) as an alternative to model the thermal condition of power transformers. Experimental results reported here show that the Elman model is more effective than multilayer feedforward backpropagation network (MLP), radial basis function network (RBF), and a classic deterministic model. Using actual data reported in the literature, the ELRN network provides an effective and robust model, learns very quickly, and efficiently manages learning data imprecision. 2 Deterministic Modeling Nowadays, as mentioned above, load capability is found using transient heating equations and specific thermal characteristics and parameters of power transformers. Load capability calculation requires knowledge of load curves and operating conditions (especially the operating temperature), whose values are fixed and usually conservative [5]. As discussed in [4], load capability calculations require the following variables and parameters: Variables (functions of time, t): Θ A = environment temperature, C. Θ TO = top oil temperature, C. Θ H = hot-spot temperature, C. Θ H = hot-spot rise above top oil temperature, C. Θ H,U = ultimate hot-spot temperature rise over top oil (for a given load current), C Θ TO,U = ultimate top oil temperature rise over environment (for a given load current), C K = load current, per unit. Parameters (constants): Θ TO,R = rated top oil temperature rise over environment, C Θ H,R = rated hot-spot temperature rise over top oil, C τ TO = top oil rise time constant, hours τ H = hot-spot rise time constant, hours R = ratio of load loss at rated-load to no-load loss at applicable tap position, dimensionless m = empirically derived value, depends on the cooling method, dimensionless. n = empirically derived value, depends on the cooling method, dimensionless. Using these variables and parameters, the heat transfer equations and the step-by-step load capability calculation process are as follows: At each time step, compute the ultimate top oil rise( Θ TO,U ) using the load current value at that instant and: [ K 2 ] n R + 1 Θ TO,U = Θ TO,R (1) R + 1

From (1), and the environment temperature at each time step, compute the increment in the top oil temperature, using the differential equation: τ TO dθ TO dt = [ Θ TO,U + Θ A ] Θ TO (2) in its finite difference equation form: DΘ TO = Dt τ TO ([ Θ TO,U + Θ A ] Θ TO ) (3) 3 Elman Network Structure The basic structure of Elman neural network is composed by four layers: input layer, hidden layer, context layer, and output layer, as depicted in the figure 1. There are adjustable weights connecting each two neighboring layers. Generally, it is considered as a special kind of feedforward neural network with additional memory neurons and local feedback [6]. where the prefix D implies a small finite step. Next compute the ultimate hot-spot rise using: PSfrag replacements Θ H,U = Θ H,R K 2m (4) The increment in the hot-spot temperature rise is found using the differential equation: τ H d Θ H dt = Θ H,U Θ H (5) in its finite difference equation form: D Θ H = Dt τ H [ Θ H,U Θ H ] (6) Finally, add the top oil temperature to the hotspot rise to get the hot-spot temperature, that is, set: Θ H = Θ TO + Θ H (7) The model described by (1) - (7), presumes some simplifying assumptions such as: the oil temperature profile inside winding increases linearly from bottom to top; the difference between the winding temperature and the oil temperature is constant along the winding; the hot-spot temperature rise is higher than the temperature rise of the wire at the top of the winding, introducing a conservative factor; the environment temperature drives the oil temperature up and down with the same time constant as does the winding temperature; the incidence of solar flux is neglected. Therefore, the use of the deterministic thermal model (1) - (7) may produce substantial error when computing load capability rate in real-time. Thus, there is a need to adopt new, more precise and robust methods to compute load capability and use it as a means to increase system operation margin in presence of overload conditions. In what follows, we suggest an alternative based on a Elman recurrent network. y 1 (k) Hidden x 1 (k) u 1 (k) Output Input u m (k) y n (k) x r (k) c 1 (k) Context c r (k) Figure 1. Structure of the Elman network Z 1 Z 1 The self connections of the context nodes in the Elman network make it also sensitive to the history of input data which is very useful in dynamic modeling [7]. The notation used in this section is given below: W1 i j = The weight that connects node i in the input layer to node j in the hidden layer W2 jq = The weight that connects node j in the hidden layer to node q in the output layer W3 l j = The weight that connects context node l to node j in the hidden layer m,n,r = The number of nodes in the input, output and hidden layers respectively. u i (k),y j (k) = Inputs and outputs of the Elman network, where i = 1,,m and j = 1,,n x i (k) = Output of the hidden node i, i = 1,,r c i (k) = Output of the context node i, i.e., the output of the hidden node i of last time. Z 1 = A unit delay. For each unit in the hidden layer an additional unit called context unit is added. The context unit is fully connected with all the hidden units in a forward manner. This means that there is a weight from every context unit to every hidden unit. Furthermore, there are

recurrent connections from the hidden units back to the context units. But each hidden unit is only connected to its associated context unit. The weights of the recurrent connections are fixed and the forward weights get trained by using backpropagation. In the forward phase the context units behave like input units. The values of the hidden units and of the output units get calculated in the same way as it is done for feedforward networks. After calculating the outputs of the hidden units, the current values get copied into the correspondentpsfrag context replacements unit via recurrent connections (through a unit delay). These values are used in the next time step. At the first time step they have to be set to some initial values. During the backward phase of the training, target values for the outputs are used and the forward weights are adjusted by backpropagation. The inputs of network are: u(k) R m, y(k) R n, x(k) R r, the outputs in each layer can be given by: x j (k) = f ( m i=1 w1 i j u i (k) + r l=1 w3 l j c l (k) c l (k) = x j (k 1) ) (8) r y q (k) = g( w2 jq x j (k) j=1 where, f ( ) and g( ) are the linear or nonlinear output function of hidden layer and output layer respectively. Because the dynamic characteristics of Elman network are provided only by internal connection, so it needn t use the state as input pr training signal. This is the advantage of the Elman network in contrast with static feedforward network. 4 Simulation Results In this section the recurrent Elman model presented in the previous section is used to estimate the hot-spot temperature of an actual transformer. The data set adopted during the experiments reported in this paper is the same of [5]. The data were collected from measurements performed in an experimental power transformer whose characteristics are summarized in Table 1. The results are compared with three alternative models: the deterministic model (DM) summarized in Section 2, a backpropagation feedforward multilayer neural network (MLP), and a radial basis function neural network (RBFN) as used in [5]. ) All four models (deterministic, backpropagation, radial basis and Elman) were trained using data describing the behavior of transformer hot-spot temperature during a horizon of 24 hours with 5 minutes sampling period. The learning data is depicted in Figure 2. Measured Data (p.u.) 1.2 1 0.8 0.6 0.4 Load Current(p.u.) Top Oil Temp.(p.u.) hot-spot Temp.(p.u.) 0.2 0 1 200 2 Figure 2. Learning data After learning, the approximation, generalization and robustness capabilities of the networks were tested using two data sets describing different load capability conditions than those in learning data [5]. Test date, shown in Figure 3, were collected in the same way as training data. Figures 4 and 5 depict the results provided by the deterministic (DM), Elman recurrent network (ELRN), backpropagation multilayer network (MLP) and the radial basis function network (RBFN) models. Figure 4 shows the actual transformer output and the corresponding models outputs under no overload condition whereas 5 shows the same outputs, but under overload conditions. Table 2 summarizes the performance index values achieved by the models during the simulation experiments. In the table, MSE Dt1 is the mean square error (MSE) obtained when running the models using the test data with no overload condition (Figure 3(a)), MSE Dt2 is the MSE obtained when using Table 1. Main Characteristics of the Power Transformer Nameplate rating 25 kva Vprimary / Vsecundary 10 kv / 3 V Iron losses 195 W Copper losses (full load) 776 W Top oil temp. rise at full load 73.1 C Length x width x height 64 x 16 x cm of the tank Type of cooling ONAN* Factory / year MACE/87

ments Measured Data (p.u.) Measured Data (p.u.) 1.2 1 0.8 0.6 0.4 0.2 0 1 200 2 1.2 1 0.8 0.6 0.4 0.2 0 1 200 2 (a) (b) Determ. Model 0 200 0 MLP 0 200 0 ELRN 0 200 0 RBFN 0 200 0 Figure 3. Test data: (a) no overload (b) with overload the models with test data with overload condition (Figure 3(b)), and MSE Dts is the MSE obtained when running the models using both data sets. Table 2. Simulation Results Model Learning Time (seg.) MSE-Dt1 MSE-Dt2 MSE-Dts MD - 17, 3489 6, 7822 12, 0655 ELRN 59,39 0, 47 0, 2, 1073 1, 2926 MLP 92,76 0, 71 2, 4885 1, 6393 RBFN 82,83 0, 2565 0, 9917 0, 6241 As Figures 4 and 5 show, ELRN, MLP and RBFN models provide good results when modeling the hotspot temperature. However, MLP and RBFN require complex learning processes whereas the recurrent model (ELRN) uses a faster and simpler learning procedure. On the other hand, when training data do not assemble an ideal learning data set, a data set that it is not representative of the target behavior, the ELRN outperforms MLP and RBFN because of its robustness when dealing with imprecision in data. To verify this important issue, we use a data set that does not fully represent the transformer dynamics. Next, we train the MLP, ELRN and RBFN neural networks using this same data set. In our experiment, as an extreme case, we took the data shown in Figure 3(a) to train the models and data of Figure 3(b) to verify the approximation, robustness and generalization capabilities. Figure 6 shows the results provided by the models in this circumstance. Table 3 summarizes Figure 4. Transformer and models outputs with no overload condition the mean square errors in this case. As we can easily see, the behavior of both, the MLP and ELRN models have shown substantial changes. Contrary, the recurrent model was able to keep the same hot-spot temperature prediction performance as before. Clearly, the recurrent model approximates the true transformer behavior closely, is able to generalize properly, and is more robust to data imprecision than its counterparts. Table 3. Results using non ideal training data MD ELRN MLP RBFN MSE-Dt3 6,7822 1.8896 1,2487 10 2 4,6579 10 4 5 Conclusion This paper has introduced an alternative Elman recurrent network to model the thermal condition of power transformers. Simulation experiments have shown that the Elman model is more effective than multilayer feedforward backpropagation network (MLP), radial basis function network (RBF), and a deterministic model because it approximates and generalizes transformer dynamics properly and is able to manage imprecise data. The recurrent network provides an effective and robust model, learns quickly, and requires modest computational power. Therefore, it is a candidate to safely contribute to increase power transformer real-time load capability during overload periods. In

Determ. Model ELRN Determ. Model ELRN 0 200 0 0 200 0 0 200 0 0 200 0 MLP RBFN MLP RBFN 0 200 0 0 200 0 0 200 0 0 200 0 Figure 5. Transformer and models outputs with overload condition Figure 6. Models robustness evaluation the future we expect to use a similar approach to estimate the loss in service lifes of transformers when submitted to overload conditions. Acknowledgment The authors acknowledge FAPESP, the Research Foundation of the State of São Paulo, for fellowship 03/042-1, CNPq, the Brazilian National Research Council for grant 0729/86-3, the Pontifical Catholic University of Minas Gerais (IPUC, PUC-MG) for its support, and Álvaro Martins and Ronald Moura, engineers of the Energetic Company of Minas Gerais (CEMIG), for their comments during the development of this work. References: [1] C.Lee Giles, Gary M.Kuhn, Ronald J.Willians, Dynamic Recurrent Neural Networks: Theory and Applications, IEEE Transactions on Neural Networks, Vol. 5, No. 2, pp. 153-155, 1994. [2] Y. Cheng, W. Qi, W. Cai, Dynamic Properties of Elman and Modified Elman Neural Networks, Proceedings of the First International Conference on Machine Learning and Cybernetics, Bejing,pp. 637-6, Nov. 2002. [4] G. W. Swift et. al., Adaptative Transformer Thermal Overload Protection, IEEE Transactions on Power Delivery, Vol. 16, No. 4, pp. 516-521, 2001. [5] V. Galdi, L. Ippolito, A. Piccolo and A. Vaccaro, Neural Diagnostic System for Transformer Thermal Overload Protection, IEE Proceedings of Electric Power Applications, Vol. 147, No. 5, pp. 415-421, 2000. [6] P.S. Sastry, G. Santharam, K.P. Unnikrishnan, Memory Neuron Networks for Identification and Control of Dynamic Systems, IEEE Transactions on Neural Networks, Vol. 5, pp. 6-319, 1994. [7] K.S. Narendra, K. Parthasarathy, Identification and Control of Dynamic Systems using Neural Networks, IEEE Transactions on Neural Networks, Vol. 1, pp. 4-27, 19. [8] G. Goodwin and R. Payne, Dynamic System Identification: Experiment Design and Data Analysis, Academic Press, Inc., New York-USA, 1977. [9] S. Haykin, Neural Networks: A Comprehensive Foundation, Prentice Hall, NJ-USA, ed. 2, 1998. [3] J. Elman, Finding Structure in Time, Cognitive Science, Vol. 14, pp. 179-211, 19.