Back Propagation Neural Controller for a Two-Drive Robot Vehicle

Similar documents
Effect of number of hidden neurons on learning in large-scale layered neural networks

Robust Controller Design for Speed Control of an Indirect Field Oriented Induction Machine Drive

A Novel Activity Detection Method

An ANN based Rotor Flux Estimator for Vector Controlled Induction Motor Drive

POWER SYSTEM DYNAMIC SECURITY ASSESSMENT CLASSICAL TO MODERN APPROACH

Chapter 7 Control. Part Classical Control. Mobile Robotics - Prof Alonzo Kelly, CMU RI

Artificial Neural Network

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

Computational statistics

EPL442: Computational

Application of Neuro Fuzzy Reduced Order Observer in Magnetic Bearing Systems

Lecture 6: Control Problems and Solutions. CS 344R: Robotics Benjamin Kuipers

Neural Network Control of Robot Manipulators and Nonlinear Systems

Introduction to Machine Learning

CONTROL OF ROBOT CAMERA SYSTEM WITH ACTUATOR S DYNAMICS TO TRACK MOVING OBJECT

Artificial Neural Networks Examination, June 2004

Decentralized Stabilization of Heterogeneous Linear Multi-Agent Systems

ECE521 Lectures 9 Fully Connected Neural Networks

International Journal of Advanced Research in Computer Science and Software Engineering

Enhancing a Model-Free Adaptive Controller through Evolutionary Computation

DESIGN AND IMPLEMENTATION OF SENSORLESS SPEED CONTROL FOR INDUCTION MOTOR DRIVE USING AN OPTIMIZED EXTENDED KALMAN FILTER

A Sliding Mode Controller Using Neural Networks for Robot Manipulator

Design Collocation Neural Network to Solve Singular Perturbed Problems with Initial Conditions

Video 5.1 Vijay Kumar and Ani Hsieh

Analysis and Design of Hybrid AI/Control Systems

Nonlinear PD Controllers with Gravity Compensation for Robot Manipulators

Laboratory Exercise 1 DC servo

General procedure for formulation of robot dynamics STEP 1 STEP 3. Module 9 : Robot Dynamics & controls

QUICK AND PRECISE POSITION CONTROL OF ULTRASONIC MOTORS USING ADAPTIVE CONTROLLER WITH DEAD ZONE COMPENSATION

Bidirectional Representation and Backpropagation Learning

Journal of Chemical and Pharmaceutical Research, 2014, 6(3): Research Article

Overview of the Seminar Topic

(Refer Slide Time: 00:01:30 min)

Real Time wave forecasting using artificial neural network with varying input parameter

N. Sarikaya Department of Aircraft Electrical and Electronics Civil Aviation School Erciyes University Kayseri 38039, Turkey

A Sliding Mode Control based on Nonlinear Disturbance Observer for the Mobile Manipulator

Unifying Behavior-Based Control Design and Hybrid Stability Theory

Reinforcement Learning, Neural Networks and PI Control Applied to a Heating Coil

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Process Control and Instrumentation Prof. D. Sarkar Department of Chemical Engineering Indian Institute of Technology, Kharagpur

Serious limitations of (single-layer) perceptrons: Cannot learn non-linearly separable tasks. Cannot approximate (learn) non-linear functions

ECE Introduction to Artificial Neural Network and Fuzzy Systems

Artificial Neural Networks

A new large projection operator for the redundancy framework

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

CDS 101/110a: Lecture 1.1 Introduction to Feedback & Control. CDS 101/110 Course Sequence

Analysis of Multilayer Neural Network Modeling and Long Short-Term Memory

SRV02-Series Rotary Experiment # 1. Position Control. Student Handout

Adaptive Inverse Control

Neural networks. Chapter 20. Chapter 20 1

Armin Rasch * Abstract

FAULT DETECTION AND ISOLATION FOR ROBOT MANIPULATORS USING ANFIS AND WAVELET

Adaptive Robust Tracking Control of Robot Manipulators in the Task-space under Uncertainties

The error-backpropagation algorithm is one of the most important and widely used (and some would say wildly used) learning techniques for neural

Control of Mobile Robots

Dreem Challenge report (team Bussanati)

Modeling and Compensation for Capacitive Pressure Sensor by RBF Neural Networks

B1-1. Closed-loop control. Chapter 1. Fundamentals of closed-loop control technology. Festo Didactic Process Control System

Thunderstorm Forecasting by using Artificial Neural Network

Modelling Multivariate Data by Neuro-Fuzzy Systems

EE C128 / ME C134 Feedback Control Systems

Modeling of Hysteresis Effect of SMA using Neuro Fuzzy Inference System

A Neuro-Fuzzy Scheme for Integrated Input Fuzzy Set Selection and Optimal Fuzzy Rule Generation for Classification

AN APPROACH TO FIND THE TRANSITION PROBABILITIES IN MARKOV CHAIN FOR EARLY PREDICTION OF SOFTWARE RELIABILITY

Unit 8: Introduction to neural networks. Perceptrons

COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

4. Multilayer Perceptrons

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Speed Control of PMSM Drives by Using Neural Network Controller

Online Identification And Control of A PV-Supplied DC Motor Using Universal Learning Networks

Intelligent Systems and Control Prof. Laxmidhar Behera Indian Institute of Technology, Kanpur

Week 5: The Backpropagation Algorithm

Lab 6d: Self-Erecting Inverted Pendulum (SEIP)

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

SRV02-Series Rotary Experiment # 7. Rotary Inverted Pendulum. Student Handout

Multi-Robotic Systems

Nonlinear Identification of Backlash in Robot Transmissions

Output Regulation of the Arneodo Chaotic System

Identification and Control of Nonlinear Systems using Soft Computing Techniques

HIGH PERFORMANCE ADAPTIVE INTELLIGENT DIRECT TORQUE CONTROL SCHEMES FOR INDUCTION MOTOR DRIVES

A Fractal-ANN approach for quality control

Research Article NEURAL NETWORK TECHNIQUE IN DATA MINING FOR PREDICTION OF EARTH QUAKE K.Karthikeyan, Sayantani Basu

Where we ve been, where we are, and where we are going next.

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

Predictive Control of a Single Link Flexible Joint Robot Based on Neural Network and Feedback Linearization

Data Mining Part 5. Prediction

Lab 5: 16 th April Exercises on Neural Networks

An Adaptive LQG Combined With the MRAS Based LFFC for Motion Control Systems

Alexander Scheinker Miroslav Krstić. Model-Free Stabilization by Extremum Seeking

Deep Learning and Information Theory

How to do backpropagation in a brain

Switched Reluctance Machine Modeling Applied on a MRAC Scheme

Lecture 4: Feed Forward Neural Networks

We provide two sections from the book (in preparation) Intelligent and Autonomous Road Vehicles, by Ozguner, Acarman and Redmill.

Artificial Neural Network Based Approach for Design of RCC Columns

Neural Networks and Ensemble Methods for Classification

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

ARTIFICIAL NEURAL NETWORKS APPROACH IN MICROWAVE FILTER TUNING

Simulation Study on Pressure Control using Nonlinear Input/Output Linearization Method and Classical PID Approach

Transcription:

Proceedings of the World Congress on Engineering and Computer Science 2 Vol I WCECS 2, October 2-22, 2, San Francisco, USA Back Propagation Neural Controller for a Two-Drive Robot Vehicle Riyadh Kenaya, Member, IEEE, IAENG and Ka C Cheok, Member, IEEE Abstract Controlling robot vehicles has become one of the most important applications in our automated world. Many control algorithms have been applied and used to accomplish the control task. et some of these robust algorithms still face stability and nonlinearity problems. In this paper we build a back propagation-based neural controller to do the control job. The said neural controller is trained according to the data collected from a controller that was pre-designed according to Lyapunov second theorem on stability. The neural controller is used to control a two-drive robot vehicle in such away to track a moving target in the plane and keep a fixed distance between the two. The trained neural controller is tested for shapes of target motion that were not seen during training. that each neuron receives a signal from the neurons in the previous layer, and each of those signals is multiplied by a separate weight value. The weighted inputs are summed, and passed through a limiting function which scales the output to a fixed range of values. The output of the limiter is then broadcast to all of the neurons in the next layer. So, to use the network to solve a problem, we apply the input values to the inputs of the first layer, allow the signals to propagate through the network, and read the output values. Index Terms Neural Controllers, Back Propagation Neural Networks, Robust Control Systems, Nonlinear Control Systems. I. INTRODUCTION Back propagation neural networks [] [2] [3] are known for their ability to recognize patterns. The concept of pattern recognition could be further expanded beyond the idea of identifying faces and images. It could be expanded to cover other important fields in our life. Control applications are examples of these important fields [4]. In this paper we build a neural system capable of imitating another system in such a way that both of the systems will do approximately the same job in response to the same input. The original system we are talking about is a controller designed according to Lyapunov second method [5] [6] of stability and will be referred to as Lyapunov controller. The plant to be controlled is a robot vehicle that tries to track a moving target in the plane. The control action is divided into two sub-actions. The first is to control the vehicle speed in such away to keep a fixed distance between the vehicle and the moving target. The second is to govern the vehicle steering in order to keep the target tracking. These two actions could be thought of as independent or related actions. Consequently, we can have either one neural controller to control both steering and distance simultaneously, or we have two different controllers one for distance and the other for steering. Fig.. General structure of a back propagation neural network. Stimulation signal is applied to the input layer neurons and the response of the said layer propagates through the middle layers. Middle layers are referred to as hidden layers. Each connection between neurons has a unique weighting value. Fig. 2 shows the neuron structure. It simply receives input from the previous layer neurons. These inputs are weighted and summed to come up with a value to be applied to a nonlinear threshold function. II. BACK PROPAGATION NEURAL NETWORKS The basic structure of a back propagation neural network is shown in fig.. It simply consists of a group of mathematically identified modules called neurons shown in fig. 2. These neurons are distributed in layer forms and interconnected with each others via weights, which means Fig. 2. Typical neuron structure in back propagation neural network. ISBN: 978-988-72--6 ISSN: 278-958 (Print); ISSN: 278-966 (Online) WCECS 2

Proceedings of the World Congress on Engineering and Computer Science 2 Vol I WCECS 2, October 2-22, 2, San Francisco, USA The said nonlinear function is usually an exponential function that ranges between two levels: and + and the output value is passed on to the neurons in the next layer [7]. The connection weights in the neural network need to be tuned (adjusted) to obtain an overall response to certain stimulation in such a way to solve a particular problem. What makes each network different from another is the value of its weights. The process of adjusting the weights to come up with a response that is perfect or close to perfect is referred to as training. The training algorithm used in this paper is called back propagation (BP). It is an iterative process and based on collecting input/output data from a certain system we aim to imitate. The idea here is to adjust the weights to bring the response of the neural network close to the response of the original system, whenever the same stimulation is applied to the both. At the beginning of the process of training, the weights are initialized to random values. The corresponding output is compared to the known-good output, and a mean-squared error signal is calculated. The error value is then propagated backwards through the network, and small changes are made to the weights in each layer. The weight changes are calculated to reduce the error signal for the case in question. The whole process is repeated for each of the example cases, then back to the first case again, and so on. The cycle is repeated until the overall error value drops below some pre-determined threshold. At this point we say that the network has learned the problem "well enough", yet the network will never exactly learn the ideal function, but rather it will asymptotically approach the ideal function. The mathematical derivation for the BP training algorithm is beyond the scope of this paper. The original controller is the above mentioned Lyapunov controller. The design details of the said controller are beyond the scope of this paper. The data used to train the BP neural network are all collected from Lyapunov controller. The error signal vector and control signal vector are both the input/output data to be used in training process. Fig. 4 shows the overall system that contains the robot vehicle as a plant and Lyapunov controller as a plant governor. target target Constant speed target target Dist Error dg Error eading sensors Dst Err Dst Ctrl dg Err eading dg Ctrl Lyapunov MRAC subsystem distance & heading controllers Manual swi tch Uf Us Scope forward & steeri ng to left & right conversion Vr Vl Volt l Volt r L & R speed controll ers Fig. 4. Lyapunov controller for the robot vehicle. In In2 Out Out2 power ampl ifi ers Va Left Va Right Left speed Vl Forward Vel Vf Steering Vel Vs Road inclinatn Right speed Vr mobile robot motors & motion Vf Vs mobile robot kinematics The target is simulated by a cross in the plane. It takes its motion from two independent signal generators for x and y axes. Fig. 5 shows the Lyapunov controlled vehicle tracking the moving target. Road incli natn III. TWO WEEL ROBOT VEICLE The block diagram of the robot vehicle to be controlled is shown in fig. 3. The said vehicle is equipped with a necessary set of sensors that provides the controller with the necessary information about the vehicle location (x and y coordinates) and the vehicle heading. It is also equipped with radar that gives information about the target location. This information is used to calculate both distance and heading error signals. Vr v olt l In Out Left Speed Vl Va Lef t Fig. 5. The controlled robot vehicle tracks the moving target. Uf Us Forward & Steer to Left & Right conversion Vl Volt r L & R speed controllers In2 Fig. 3. Robot vehicle block diagram. Out2 Power amplifiers Road Inclinatn Va Right Forward Vel Vf Steer Vel Vs Road inclinatn Right Speed Vr Motors & Motion Vf Vs Kinematics IV. LAPUNOV CONTROLLER DATA Fig. 6 shows Lyapunov controller and sensor block diagrams. Both the distance and steering error signals are applied to the controller. The controller generates two output control signals, one for distance control and the other for steering control. It is important to mention that the design of Lyapunov controller is based on a second order reference model [8]. The main purpose of Lyapunov controller here is to build a control system that generates control signals in the same way a linear second order reference system generates. Accordingly, the cost function we try to minimize is the error difference between the second order controller and its ISBN: 978-988-72--6 ISSN: 278-958 (Print); ISSN: 278-966 (Online) WCECS 2

Proceedings of the World Congress on Engineering and Computer Science 2 Vol I WCECS 2, October 2-22, 2, San Francisco, USA corresponding Lyapunov one. As a result, this method makes the nonlinear system behave like a linear system whose control parameters (like overshoot and steady state error) are design values to be set by the designer. Figs. and 2 show the overall vehicle response to the target motion. One can easily tell that Lyapunov controller just does the right control job. One can also see that the distance between the two is almost kept constants and close to 3 feet. target Dist Error Dst Err target dg Error Dst Ctrl dg Err eading eading target dg Ctrl Fig.. Robot vehicle tracking the moving target (x-axis tracking). target Sensors Lyapunov Controller Fig. 6. Robot vehicle Lyapunov controller Figs. 7- show Lyapunov controller data. These data are collected in response to a sinusoidal target motion in the plane. In this experiment we desire to make the vehicle track the target and keep three feet of distance between the two. Fig. 2. Robot vehicle tracking the moving target (y-axis tracking). Fig. 7. Lyapunov controller distance error signal. Fig. 8. Lyapunov controller distance control signal. V. MIED SIGNAL-BASED NEURAL CONTROLLER RESULTS The data given in figs. 7-2 are used as training data for the BP neural controller. Distance and heading error signals are considered as stimulation data, and the corresponding distance and steering control signals are considered as response data. In this section, both of the distance and heading error signals would be treated as members of one stimulation vector. The said vector is refereed to as error vector. In the same sense, the distance and steering control signals are considered as members of the same response vector and referred to as control vector. The network structure is shown in fig. 3. It is a two layer network and this is one of many possible structures. In our design, we choose 5 TANSIG neurons for the first layer and one PURLIN neuron for the second layer. The first layer weight matrix is W 52. W2 25 is the second layer matrix. s y s2 y2 In u W* u Weight W netsum tansig W2* u Weight W2 netsum purelin Out y2 Fig. 9. Lyapunov controller heading error signal. Constant b* u Constant b2* u Bias B2 Bias B Fig. 3. Back propagation neural controller for the robot vehicle. Fig.. Lyapunov controller heading control signal. W matrix dimensions give indication to the number of neurons and input variables applied to the network. 5 is the number of neurons, whereas 2 is the number of variables applied. B 5 is the bias matrix for the first layer. ISBN: 978-988-72--6 ISSN: 278-958 (Print); ISSN: 278-966 (Online) WCECS 2

Proceedings of the World Congress on Engineering and Computer Science 2 Vol I WCECS 2, October 2-22, 2, San Francisco, USA B2 is the second layer bias. The training epoch number is chosen to be 5. Increasing the epoch number may result in reducing the training mean square error. The trained network replaces the original Lyapunov controller and the experiment is rerun. Fig. 4 shows the overall control system with the BP neural network as the system controller. Notice how the distance error and heading error signals are combined in one vector. Also, the control signals for distance and heading are both combined in one control vector. Fig. 8. BP neural controller heading control signal. target Constant5 Speed target Dist Error target dg Error eading In u Out y 2 BP Controller Manual Switch Uf Us Vr v olt l In Out Lef t Speed Vl Va Lef t Forward Vel Vf Vf x-target motion x-vehicle motion - 5 5 2 target Sensors Volt r Vl Forward & Steer to Left & Right conversion L & R speed controllers In2 Out2 Power amplifiers Va Right Steer Vel Vs Road inclinatn Right Speed Vr Vs Kinematics Fig. 9. BP neural controller overall tracking signals (x-axis motion). BP_Controller_Data To File Road Inclinatn Motors & Motion Fig. 4. Back propagation neural controller for the robot vehicle. Figs. 5-2 show the neural system performance when exposed to the same operational conditions that the original controller has witnessed. Fig. 5. BP neural controller distance error signal. Fig. 6. BP neural controller distance control signal. Fig. 7. BP neural controller heading error signal. 5 y-target motion y-vehicle motion -5 5 5 2 Fig. 2. BP neural controller overall tracking signals (y-axis motion). VI. SEPARATE SIGNAL-BASED NEURAL CONTROLLER RESULTS In this section, we break the neural controller introduced in the previous section into two different controllers. One for distance, and the other, for heading control. Each one of these two controllers is trained independently from the other. This gives us some flexibility in choosing different training parameters for the two controllers. The importance of this measure is represented by the ability of selecting certain range of training data for one controller and selecting another range for the other. Even the epoch number and number of neurons per layer of one of the controllers could be completely different from those of the other. We have trained the distance controller based on the data collected from the original controller while tracking the target that moves in a sinusoidal shape. On the other hand, the heading controller is trained based on the data collected from the original controller while tracking the target whose motion is of step shape. The reason we went with two different types of training data, is to ensure the maximum stability of the neural controller. Our experiments have shown that choosing any other set of training data did always cause an unstable behavior of the robot vehicle. Figs. 2-26 show the performance of the separate signal controller. One can easily notice that the three systems (Lyapunov, separate, and mixed signal controllers) have almost the same response to the same excitation. It is obvious that the three systems show stable behavior for a sinusoidal target motion. The three controllers succeed in keeping a fixed distance of 3 feet between the vehicle and the target. ISBN: 978-988-72--6 ISSN: 278-958 (Print); ISSN: 278-966 (Online) WCECS 2

Proceedings of the World Congress on Engineering and Computer Science 2 Vol I WCECS 2, October 2-22, 2, San Francisco, USA Fig. 2. Separate signal BP neural controller distance error signal. training [6] [9]. In this section we test the BP mixed signal controller for special types of target motion. A mixture of step and sinusoidal Target motion is an example of these excitations that were not seen during training. The neural controller behavior is compared with that of Lyapunov controller (Figs. 27 and 28). It is obvious that Lyapunov controller shows a very poor performance. It completely diverges away from the target after few seconds of relative stability. On the other hand, we can see that the mixed signal neural controller retains its stability and tries to reduce the sudden error caused by the target step motion (figs. 29 and 3). This error reduction is affected by the robot vehicle moment of inertia. Fig. 22. Separate signal BP neural controller distance control signal. x-target motion x-vehicle motion - 5 5 2 Fig. 27. Lyapunov controller overall tracking signals (x-axis target step motion). Fig. 23. Separate signal BP neural controller heading error signal. y-target motion y-vehicle motion - 5 5 2 Fig. 28. Lyapunov controller overall tracking signals (y-axis target sinusoidal motion). Fig. 24. Separate signal BP neural controller heading control signal. x-target motion x-vehicle motion - 5 5 2 Fig. 25. Separate signal BP neural controller x-axis tracking signals. 5 y-target motion y-vehicle motion -5 5 5 2 Fig. 26. Separate signal BP neural controller y-axis tracking signals. x-target motion x-vehicle motion - 5 5 2 Fig. 29. BP mixed signal neural controller overall tracking signals (x-axis target step motion). y-target motion y-vehicle motion - 5 5 2 Fig. 3. BP mixed signal neural controller overall tracking signals (y-axis target sinusoidal motion). VII. GENERALIT OF BP NEURAL CONTROLLERS Generality of any neural system is represented by its ability to respond to the excitations that were not seen during ISBN: 978-988-72--6 ISSN: 278-958 (Print); ISSN: 278-966 (Online) VIII. CONCLUSIONS AND FUTURE WORK Lyapunov controllers could be substituted with back propagation neural controllers. The neural controllers show stable behavior and accurate tracking. Mixed signal WCECS 2

Proceedings of the World Congress on Engineering and Computer Science 2 Vol I WCECS 2, October 2-22, 2, San Francisco, USA controllers could be also replaced with their separate signal counterparts. Signal separation enables us to use different sets of training parameters for the independent controllers. It also makes it easy for us to use different sets and combinations of training vectors. We still get the system to work and perform well. The future work would highly concentrate on optimizing the length of training data chains and studying the effect of selecting certain parts of data other than using the whole data chain. This is to be done by eliminating the redundant parts of data and retain the part that conveys core and non-repetitive information. This would result in reducing the training time and the hardware size needed to realize the network. Neural network optimization is considered as one of the most challenging research topics. REFERENCES [] J. S. R. Jang, C. T. Sun, and E. Mizutani, Neuro Fuzzy and Soft Computing. Prentice all, 997. [2] Z. Effendi, R. Ramli and J.A. Ghani, Back Propagation Neural Networks for Grading Jatropha curcas Fruits Maturity, American Journal of Applied Sciences, 7 (3): 39-394, 2. [3] Michael C.O'Neill, Training back-propagation neural networks to define and detect DNA-binding sites, Nucleic Acids Research, Vol. 9, No. 2, 33. [4] Panos J. Antsaklis, Neural Networks in Control Systems, IEEE Control Systems Magazine, April 99. [5] M. Kanat Camlibel, Jong-Shi Pang, and Jinglai Shen, Lyapunov Stability of Complementarily and Extended Systems, 26 Society for Industrial and Applied Mathematics, Vol. 7, No. 4, pp. 56-. [6] R. Kenaya, Euclidean Adaptive Resonance Theory with Application to Nonlinear and Adaptive Control Systems, PhD Dissertation, Oakland University, Rochester, Michigan, 29. [7] T. Catfolis, Mapping a Complex Temporal Problem into a Combination of Static and Dynamic Neural Networks, SIGART Bulletin, Vol. 5, No. 3. [8] P. Vempaty, K. Cheok, and R. Loh, Model Reference Adaptive Control for Actuators of a Biped Robot Locomotion, Proceedings of the World Congress on Engineering and Computer Science, Vol. II,WCECS October 2-22, 29, San Francisco, USA. [9] C. Cox, K. Mathia, J. Edwards, and R. Akita, Modern Adaptive Control With Neural Networks, International Conference on Neural Networks Information Processing, ICONIP 96, ong Kong. ISBN: 978-988-72--6 ISSN: 278-958 (Print); ISSN: 278-966 (Online) WCECS 2