Neural Network Based Response Surface Methods a Comparative Study

Similar documents
Neural Networks and the Back-propagation Algorithm

Artificial Neural Network

Data Mining Part 5. Prediction

4. Multilayer Perceptrons

Global Sensitivity Analysis in Structural Optimization

Lecture 7 Artificial neural networks: Supervised learning

Artificial Neural Networks Examination, June 2005

Global Sensitivity Analysis in Industrial Application with LS-OPT

Artificial Neural Network and Fuzzy Logic

Artificial Neural Networks. Edward Gatt

Artificial Intelligence

New Developments in LS-OPT - Robustness Studies

Artificial Neural Networks

CMSC 421: Neural Computation. Applications of Neural Networks

COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

COMP-4360 Machine Learning Neural Networks

Machine Learning

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

COMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines. COMP9444 c Alan Blair, 2017

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

How to do backpropagation in a brain

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Introduction to Neural Networks

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

EEE 241: Linear Systems

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

Simple Neural Nets For Pattern Classification

Lecture 4: Perceptrons and Multilayer Perceptrons

Two-Stage Stochastic and Deterministic Optimization

Artifical Neural Networks

Learning and Memory in Neural Networks

Neural Networks. Nethra Sambamoorthi, Ph.D. Jan CRMportals Inc., Nethra Sambamoorthi, Ph.D. Phone:

Multilayer Perceptron

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

Neural Networks Introduction

Artificial Neural Networks Examination, June 2004

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Artificial Neural Networks Examination, March 2004

Instituto Tecnológico y de Estudios Superiores de Occidente Departamento de Electrónica, Sistemas e Informática. Introductory Notes on Neural Networks

Reading Group on Deep Learning Session 1

Backpropagation Neural Net

Artificial Neural Network Based Approach for Design of RCC Columns

Revision: Neural Network

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Lecture 5: Logistic Regression. Neural Networks

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

Introduction Biologically Motivated Crude Model Backpropagation

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

Lecture 6. Notes on Linear Algebra. Perceptron

Unit III. A Survey of Neural Network Model

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

ECE521 Lectures 9 Fully Connected Neural Networks

Neural Networks. Textbook. Other Textbooks and Books. Course Info. (click on

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

C4 Phenomenological Modeling - Regression & Neural Networks : Computational Modeling and Simulation Instructor: Linwei Wang

Internal Covariate Shift Batch Normalization Implementation Experiments. Batch Normalization. Devin Willmott. University of Kentucky.

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Simple neuron model Components of simple neuron

Estimation of Inelastic Response Spectra Using Artificial Neural Networks

Neural networks. Chapter 19, Sections 1 5 1

Course 395: Machine Learning - Lectures

Lecture 10. Neural networks and optimization. Machine Learning and Data Mining November Nando de Freitas UBC. Nonlinear Supervised Learning

Notes on Back Propagation in 4 Lines

Neural Networks and Ensemble Methods for Classification

Neural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Neural Networks and Deep Learning

y(x n, w) t n 2. (1)

Introduction to feedforward neural networks

Computational statistics

Blind Equalization Formulated as a Self-organized Learning Process

Artificial Neural Networks (ANN) Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso

Recurrent Neural Networks

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

Introduction to Neural Networks

The error-backpropagation algorithm is one of the most important and widely used (and some would say wildly used) learning techniques for neural

Feedforward Neural Nets and Backpropagation

Application of an Artificial Neural Network Based Tool for Prediction of Pavement Performance

An artificial neural networks (ANNs) model is a functional abstraction of the

S. Freitag, B. T. Cao, J. Ninić & G. Meschke

Neural network modelling of reinforced concrete beam shear capacity

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Neural networks. Chapter 20, Section 5 1

Neural Networks biological neuron artificial neuron 1

Artificial Neural Networks The Introduction

Financial Risk and Returns Prediction with Modular Networked Learning

BACKPROPAGATION. Neural network training optimization problem. Deriving backpropagation

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Linear & nonlinear classifiers

EPL442: Computational

Artificial Neural Networks

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Transcription:

. LS-DYNA Anwenderforum, Ulm Robustheit / Optimierung II Neural Network Based Response Surface Methods a Comparative Study Wolfram Beyer, Martin Liebscher, Michael Beer, Wolfgang Graf TU Dresden, Germany Abstract: This paper deals with the application of the response surface method based on neural networks in the field of engineering. A multi-layer feedforward neural network is employed to replace the underlying structural analysis. It is trained on a data set from initial computations. The application of the backpropagation algorithm for training a neural network requires a decision for a particular training mode. For the examined examples it is shown that the incremental mode possesses different advantages compared to the batch mode. Furthermore, an approach to improve the approximation quality given by a neural network is introduced and demonstrated by means of a numerical example. The authors would like to thank Dynamore GmbH, Stuttgart, Germany for supporting this study. Keywords: response surface method, neural networks, backpropagation Copyright by DYNAmore GmbH K - II - 9

Robustheit / Optimierung II. LS-DYNA Anwenderforum, Ulm Introduction The structural analysis of complex structures is usually solved by means of a computational model. The computational model is an abstraction of the structure and its environment including their interactions. Such a model is used to map structural parameters (e.g. material parameters, geometrical parameters, loads) onto structural responses (e.g. stresses, displacements, deformations). The complexity of the model increases the more information it has to contain. In many fields of engineering the Finite Element method is employed to perform tasks of arbitrary complexity. Structural analysis represents the basis for the structural design in a variety of approaches. The values of the structural parameters are generally not predetermined but vary within specific ranges. This phenomenon is referred to as uncertainty and is accounted for with the aid of an uncertain structural analysis. For processing uncertain quantities of different characteristics associated numerical procedures are available. These primarily comprise stochastic analysis, fuy analysis, and fuy stochastic analysis. In any case, an uncertain structural analysis requires the repeated application of the underlying deterministic computational model. The more complex the model is the higher is the computational effort needed to perform the computation. This effort may be reduced by simplifying the computational model and reducing the number of calls of the deterministic computational model. In this context the use of an adequate approximation of the structural response represents the most effective measure to achieve a reasonable reduction of the numerical effort. Subsequently, an approximation by means of the response surface method based on neural networks is elucidated. Neural Network Concept. Basic idea The idea of artificial neural networks is based on the design of the human brain. The human brain is constituted by information-processing units (so called neurons) that are connected by synapses, and it forms the kernel of the human nervous system. It is capable of processing input signals that are derived from the environment and of providing appropriate output signals (e.g. certain actions). The advantages of the human information processing system are complexity, nonlinearity, and parallelism. An artificial neural network resembles the human brain in many respects. It is constituted by neurons which are connected by synapses, it has the ability of mapping input signals onto output signals and to adapt to certain tasks during a training phase. The output produced by a neural network is called response surface. The idea of such a response surface in engineering is then to replace the deterministic computational model for structural analysis by a neural network. That is, the input signals comprise structural parameters such as loads, material parameters, and geometrical parameters and the network output provides the associated response surface in the form of stresses, displacements, or deformations. For this purpose, the network first needs to learn the features of the underlying deterministic computational model. This learning is based on initially performed structural analyses.. Network Architecture There exist a variety of alternatives to design a neural network, e.g., feedforward neural networks, recurrent neural networks, radial basis function networks and committee machines. The focus of this study is set on feedforward neural networks which are already applied successfully in many fields of engineering. A specific network architecture is built by combining the two main components of a neural network, the neurons and the synapses. Each synapse connects two neurons with each other and possesses a synaptic weight w. It enables the signal flow from one neuron to the next one. A neuron represents an information processing unit that maps an input signal onto an output signal as illustrated in Fig. (a). It contains a summing junction ν which lumps together the incoming signals x, each weighted by a specific synaptic weight w. The summation of input signals for neuron k (see Eq. ()) also involves an external term called bias b. ν k = m j= w k j x j + b k () K - II - Copyright by DYNAmore GmbH

. LS-DYNA Anwenderforum, Ulm Robustheit / Optimierung II The weights w and the bias b allow the neuron to be adjusted to particular conditions. The summing junction ν is used as input argument for the subsequently called activation function ϕ(.), which produces the output y of the neuron. Different types of activation functions are available, such as threshold function, piecewise linear function and sigmoidal functions. It is often desired to have a differentiable activation function (see Sec. ). In this study further consideration is only given to sigmoidal functions (see Fig. ) as the most popular type of activation functions. input layer first hidden layer second hidden layer output layer x input signals x x x m w k w k w km synaptic weights bias b k Ó summing junction activation function í k ö(.) y k (a) Constitution of a neuron output signal x x m b b b input signals, bias (b) Constitution of a feedforward neural network y= output signal Figure : Neural network A multi-layer feedforward neural network (see Fig. (b)) permits signal flow exclusively in forward direction through the network. The neurons are organied in different layers: one input layer, one or more hidden layers, and one output layer. The produced output y of each neuron of a layer is transmitted to one or more neurons of the following layer by synaptic connections. Within a fully connected network every neuron is linked to all neurons of the following layer. Otherwise, if several connections are missing, the network is referred to as partially connected. Furthermore, it is possible to introduce shortcut connections between neurons of non-adjacent layers. ö() í ö() í - - - - í - - - - í - (a) Logistic sigmoid function (b) Sigmoid symmetric function Figure : Sigmoidal activation functions The output signals y of the neurons of the output layer constitute the response of the whole neural network. The number of input and output signals determines the number of the neurons of the input and the output layer, respectively. The number of hidden layers and the dimensionality of these should be chosen problem-specific. neural network x y = Synaptic weights w are adjustable free parameters of a neural network and have the task to strengthen or weaken the signals transferred by the synaptic connections. These values are problem-specific adjusted during the training phase (see Sec. ) of the neural network. The knowledge represented by a neural network after the training is stored in its synaptic weights. Usually the training phase is followed by a validation or testing phase in which the synaptic weights of the network are not changed anymore. The approximation-quality may be evaluated by calculating an error between the prediction of the neural network and the desired output. () Copyright by DYNAmore GmbH K - II -

Robustheit / Optimierung II. LS-DYNA Anwenderforum, Ulm Comparison of Training Algorithms. General Aspects of Backpropagation The iterative process of adjustment of weights of a neural network is called training. The aim is to derive a neural network response that approximates the underlying data set with a maximum quality. There exist a variety of different methods to train a neural network. In this study the focus is on the backpropagation algorithm as one form of supervised error-correction learning. This algorithm requires training data in form of input-output pairs that are obtained by repeatedly applying the computational model for structural analysis. The backpropagation algorithm may be coarsely described with the following three steps, which have to be applied several times in an iteration.. Forward computation of input signal of training sample and determination of neural network response. Computation of an error between desired response and neural network response. Backward computation of the error and calculation of corrections to synaptic weights and biases The initial values of the synaptic weights are obtained using random numbers. After the computation of the neural network response to an input signal the response error is determined and used to compute adequate changes of the synaptic weights with the aim of improving the quality of the neural network response in the next step. By applying these corrections to the weights it is attempted to minimie the error surface. Backpropagation is based on a standard gradient method. For the computation of the gradient the activation function needs to be differentiable. The correction of the synaptic weights may be expressed by the following Eq. () and is discussed in detail in [, ]. Within the backpropagation algorithm two different modes can be applied, which are called incremental and batch. ( weight correction ) ( learning-rate = parameter ) ( local gradient ) ( input signal ) () Incremental Mode The incremental mode, also known as single, on-line or sequential mode, applies a weight correction after each presentation of one sample of training data. The training sample is chosen randomly out of the training data, leading to a stochastic nature of the search for the minimum of the error surface. This maximies the probability of finding the global minimum. Furthermore, the incremental mode is resistant against redundant training data. As the weight correction is applied after each sample, the exact error of the specific sample is used. Batch Mode The batch mode applies a weight correction only once after each epoch. During one epoch every sample of the training data is presented to the neural network. The weight correction according to each training sample needs to be stored and the weights are updated after the presentation of the whole training data using all stored weight changes. The training samples are not chosen randomly which assures a convergence to a (local) minimum. Further, redundant training data is disadvantageous, because it yields longer computation time.. Numerical Comparison by Examples For the comparison of the incremental mode and the batch mode calculations are performed under identical conditions. That means, the sie of the used network, the training and testing data as well as the number of trained epochs are the same for the application of both modes. The training behavior varies in dependence on the randomly selected initial values for the synaptic weights and biases and on the training mode incremental or batch. The evaluation of training results is performed by means of the root mean square error e rms (see Eq. ()) of the approximation. In Eq. () N is the number of K - II - Copyright by DYNAmore GmbH

. LS-DYNA Anwenderforum, Ulm Robustheit / Optimierung II training samples, d i is the desired exact output of one training sample i, and y i is the neural network output generated from training sample i. e rms = N N i= (d i y i ) During a testing phase the trained neural network is shown a data set containing samples not used during the training phase. The synaptic weights are not adjusted during this phase. The testing phase and the computation of a testing error gives information about generaliation qualities of the neural network. () Example The first example is taken from automobile industry (see Fig. (a)). x,...,x x sheet thickness strain rate x x rigid wall x x courtesy of Daimler Chrysler (a) Position of chassis beam in vehicle (b) Parameters of chassis beam Figure : Example chassis beam [7] The displayed chassis beam is an important component for the crash behavior of the vehicle. With the aid of a numerical structural analysis it is aimed at finding an appropriate design for this beam. The input parameters x i, i =,..., which need to be adjusted in the design process are depicted in Fig. (b). The response which is approximated by means of a neural network is the stonewall force affecting the rigid wall. The development of the error, which is plotted over the number of epochs, is shown in Fig. (a). Example The second example deals with the planned bridge of Messina in Italy (see Fig. ) which carries two railways as well as two highways. Figure : Example bridge of Messina [, ] The region around the bridge is subject to seismic activity. Therefore, the bridge behavior under seismic loads needs to be examined. The input parameters are seismic loads at different points. The structural response comprises displacements at 8 points of the bridge. As an example only one of these displacements is chosen to be approximated by means of a neural network. Training and testing results are shown in Fig. (b). Copyright by DYNAmore GmbH K - II -

Robustheit / Optimierung II. LS-DYNA Anwenderforum, Ulm training incremental training batch test incremental (.88) test batch (.78).8 training incremental training batch test incremental (.) test batch (.7).. e rms e rms.. epochs epochs (a) Example (b) Example Figure : Comparison of the incremental mode and the batch mode Interpretation For both examined examples it is obvious that both modes incremental and batch yield a similar remaining error of the approximation of the neural network with respect to the exact response. The incremental mode of backpropagation training, which updates the synaptic weights multiple times during one epoch, achieves a convergence within fewer epochs than the batch mode. For this reason the incremental mode needs less computation time for training and provides a better approximation quality after few epochs. Improvement of Approximation Accuracy In many cases the quality of a response surface based on a neural network even after a fair amount of training epochs still raises the desire for further improvement. Different approaches to improve the approximation quality are, e.g., an increase of complexity of the neural network by adding more hidden layers and / or hidden neurons and the use of more than one neural network for the response approximation. Multiple neural networks are used by committee machines (as implemented in LS-Opt), by section-wise approximations (e.g. using a clustering algorithm for subdividing), and by neural network composites. Subsequently, the specific use of a neural network composite is explained and examined in detail.. Selected Strategy The term composite in connection with neural networks used in this study is associated with a perspective different from the definition given in []. The proposed approach is visualied in Fig.. Herein, a first neural network is trained to approximate a certain response. The underlying task is very complex. Thus the neural network response is only capable of approximating certain features of the desired response. The idea is now to use the computed remaining error surface to train a second neural network. The overall response is then constituted by adding the responses of both networks. For further improvement of the response surface more neural networks may be added to approximate the remaining errors in this manner, respectively. K - II - Copyright by DYNAmore GmbH

. LS-DYNA Anwenderforum, Ulm Robustheit / Optimierung II n- x,-( Ó i ) i= training neural network n n x neural network n n x,-( + ) training neural network x x neural network n Ó i= i x,- training neural network x neural network x, training neural network (a) Training x neural network (b) Evaluation Figure : Neural network composite. Numerical Example The use of a neural network composite is demonstrated by means of a numerical example. The response shown in Fig. 7(a) needs to be approximated. 8 - - x - - (a) Desired response x 8 8 - - x - - x - - x - - x (b) Response of network (c) Response of network + Figure 7: Responses The response of the first trained network is plotted in Fig. 7(b). It is visible that the first neural network is only capable of approximating the global trend of the desired response. Local features such as extreme values are not reflected correctly. The second neural network is trained to approximate the error surface. The sum of the responses from both networks achieves a better approximation quality by giving respect to local features of the response (see Fig. 7(c)). The described procedure may be repeated to add more neural networks to eliminate the remaining errors. The quality improvement by adding the second network becomes particularly obvious in the two-dimensional plot at cross section x = in Fig. 8. Copyright by DYNAmore GmbH K - II -

Robustheit / Optimierung II. LS-DYNA Anwenderforum, Ulm 8 cross section x = - - x - - (a) Position of cross section x = x desired response 8 response of network + network response of network - - x (b) Cross section x = Figure 8: Comparison of responses Conclusions This study discusses two different issues of response surface approximation based on neural networks. First, the differences between the incremental mode and the batch mode of the backpropagation algorithm applied for training neural networks are investigated. For the examined examples it is found that the incremental mode possesses advantages such as faster convergence and a less computational effort compared to the batch mode. Consequently, an application of the incremental mode is proposed in approaches on this basis such as structural design and optimiation. For example, an improvement of the successive response surface method (see []) used in LS-Opt may be achieved. The second issue is the quality improvement of the response approximation by using more than one neural network. Different from the committee machine used by LS-Opt in this study a neural network composite is proposed. Such a network composite applies a stepwise reduction of the approximation error by using the error surface for network training. In the chosen example it is shown that a network composite is able to approximate several features of the desired response with a remarkable quality. K - II - Copyright by DYNAmore GmbH

. LS-DYNA Anwenderforum, Ulm Robustheit / Optimierung II References [] C. M. Bishop. Neural Networks for Pattern Recognition. Oxford University Press, Inc., 997. [] G. Deco and D. Obradovic. An Information Theoretic Approach to Neural Computing. Springer, 99. [] S. Haykin. Neural Networks. Tom Robbins / Prentice-Hall International, Inc., 999. [] L. Scambi and F. Bontempi. A genetic algorithm approach for performance evaluation of long span suspension bridges. In Proceedings of the Tenth International Conference on Civil, Structural and Environmental Engineering Computing. Civil-Comp Press, Stirling, Scotland,. [] L. Scambi and F. Bontempi. A fuy approach in the seismic analysis of long span suspension bridges. In th World Conference on Earthquake Engineering,. [] N. Stander, W. Roux, T. Eggleston, and K. Craig. LS-Opt Manual. Livermore Software Technology Corporation, March. [7] M. Thiele, M. Liebscher, and W. Graf. Fuy analysis as alternative to stochastic methods - a comparison by means of a crash analysis. In th German LS-DYNA Forum,. Copyright by DYNAmore GmbH K - II - 7

Robustheit / Optimierung II. LS-DYNA Anwenderforum, Ulm K - II - 8 Copyright by DYNAmore GmbH