Optimal Tuning of PID Parameters Using Iterative Learning Approach

Similar documents
Iterative Controller Tuning Using Bode s Integrals

IMC based automatic tuning method for PID controllers in a Smith predictor configuration

CHBE320 LECTURE XI CONTROLLER DESIGN AND PID CONTOLLER TUNING. Professor Dae Ryook Yang

Feedback Control of Linear SISO systems. Process Dynamics and Control

Model-based PID tuning for high-order processes: when to approximate

Improved Identification and Control of 2-by-2 MIMO System using Relay Feedback

Iterative Feedback Tuning

Simulation Study on Pressure Control using Nonlinear Input/Output Linearization Method and Classical PID Approach

Improved Autotuning Using the Shape Factor from Relay Feedback

PID control of FOPDT plants with dominant dead time based on the modulus optimum criterion

Design and Tuning of Fractional-order PID Controllers for Time-delayed Processes

6.1 Sketch the z-domain root locus and find the critical gain for the following systems K., the closed-loop characteristic equation is K + z 0.

An Improved Relay Auto Tuning of PID Controllers for SOPTD Systems

Design of Decentralised PI Controller using Model Reference Adaptive Control for Quadruple Tank Process

Robust Loop Shaping Controller Design for Spectral Models by Quadratic Programming

Limit Cycles in High-Resolution Quantized Feedback Systems

Iterative Feedback Tuning for robust controller design and optimization

MULTILOOP PI CONTROLLER FOR ACHIEVING SIMULTANEOUS TIME AND FREQUENCY DOMAIN SPECIFICATIONS

ADAPTIVE PID CONTROLLER WITH ON LINE IDENTIFICATION

DESIGN OF AN ON-LINE TITRATOR FOR NONLINEAR ph CONTROL

Process Identification for an SOPDT Model Using Rectangular Pulse Input

Performance assessment of MIMO systems under partial information

Quantitative Feedback Theory based Controller Design of an Unstable System

A NEW APPROACH TO MIXED H 2 /H OPTIMAL PI/PID CONTROLLER DESIGN

ThM06-2. Coprime Factor Based Closed-Loop Model Validation Applied to a Flexible Structure

Humanoid Based Intelligence Control Strategy of Plastic Cement Die Press Work-Piece Forming Process for Polymer Plastics

EE3CL4: Introduction to Linear Control Systems

IMPROVED TECHNIQUE OF MULTI-STAGE COMPENSATION. K. M. Yanev A. Obok Opok

Quanser NI-ELVIS Trainer (QNET) Series: QNET Experiment #02: DC Motor Position Control. DC Motor Control Trainer (DCMCT) Student Manual

Analysis and Design of Control Systems in the Time Domain

THE ANNALS OF "DUNAREA DE JOS" UNIVERSITY OF GALATI FASCICLE III, 2000 ISSN X ELECTROTECHNICS, ELECTRONICS, AUTOMATIC CONTROL, INFORMATICS

Automatic Control 2. Loop shaping. Prof. Alberto Bemporad. University of Trento. Academic year

Research Article Convex Polyhedron Method to Stability of Continuous Systems with Two Additive Time-Varying Delay Components

CHAPTER 3 TUNING METHODS OF CONTROLLER

Lecture 5 Classical Control Overview III. Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore

MAS107 Control Theory Exam Solutions 2008

A unified double-loop multi-scale control strategy for NMP integrating-unstable systems

YTÜ Mechanical Engineering Department

Compensator Design to Improve Transient Performance Using Root Locus

Inertia Identification and Auto-Tuning. of Induction Motor Using MRAS

Tuning Rules for Proportional Resonant Controllers

Dr Ian R. Manchester

EEE 550 ADVANCED CONTROL SYSTEMS

Robust Tuning of Power System Stabilizers Using Coefficient Diagram Method

A FEEDBACK STRUCTURE WITH HIGHER ORDER DERIVATIVES IN REGULATOR. Ryszard Gessing

A 2D Systems Approach to Iterative Learning Control with Experimental Validation

Digital Control Engineering Analysis and Design

Experiment 81 - Design of a Feedback Control System

YTÜ Mechanical Engineering Department

Part II. Advanced PID Design Methods

Tuning Method of PI Controller with Desired Damping Coefficient for a First-order Lag Plus Deadtime System

CompensatorTuning for Didturbance Rejection Associated with Delayed Double Integrating Processes, Part II: Feedback Lag-lead First-order Compensator

Department of Electrical and Computer Engineering. EE461: Digital Control - Lab Manual

CHAPTER 7 STEADY-STATE RESPONSE ANALYSES

A Comparative Study on Automatic Flight Control for small UAV

A Simple PID Control Design for Systems with Time Delay

Transient Response of a Second-Order System

Professional Portfolio Selection Techniques: From Markowitz to Innovative Engineering

Nonlinear PD Controllers with Gravity Compensation for Robot Manipulators

Self-tuning Control Based on Discrete Sliding Mode

Root Locus Design Example #3

Contents. PART I METHODS AND CONCEPTS 2. Transfer Function Approach Frequency Domain Representations... 42

Ian G. Horn, Jeffery R. Arulandu, Christopher J. Gombas, Jeremy G. VanAntwerp, and Richard D. Braatz*

Autonomous Mobile Robot Design

Research Article. World Journal of Engineering Research and Technology WJERT.

Dynamic Operability for the Calculation of Transient Output Constraints for Non-Square Linear Model Predictive Controllers

DECENTRALIZED PI CONTROLLER DESIGN FOR NON LINEAR MULTIVARIABLE SYSTEMS BASED ON IDEAL DECOUPLER

Design of Decentralized Fuzzy Controllers for Quadruple tank Process

STABILITY ANALYSIS OF DAMPED SDOF SYSTEMS WITH TWO TIME DELAYS IN STATE FEEDBACK

Improving the Control System for Pumped Storage Hydro Plant

Robust fixed-order H Controller Design for Spectral Models by Convex Optimization

Robust Internal Model Control for Impulse Elimination of Singular Systems

Dr Ian R. Manchester Dr Ian R. Manchester AMME 3500 : Review

GAIN SCHEDULING CONTROL WITH MULTI-LOOP PID FOR 2- DOF ARM ROBOT TRAJECTORY CONTROL

Virtual Reference Feedback Tuning for non-linear systems

Video 5.1 Vijay Kumar and Ani Hsieh

Experimental Study of Fractional Order Proportional Integral (FOPI) Controller for Water Level Control

AN EXTENSION OF GENERALIZED BILINEAR TRANSFORMATION FOR DIGITAL REDESIGN. Received October 2010; revised March 2011

AN INTELLIGENT HYBRID FUZZY PID CONTROLLER

Unified Power Flow Controller (UPFC) Based Damping Controllers for Damping Low Frequency Oscillations in a Power System

PID Tuning of Plants With Time Delay Using Root Locus

Modelling the Dynamics of Flight Control Surfaces Under Actuation Compliances and Losses

Control of Electromechanical Systems

NonlinearControlofpHSystemforChangeOverTitrationCurve

Attitude Regulation About a Fixed Rotation Axis

Laboratory Exercise 1 DC servo

VALLIAMMAI ENGINEERING COLLEGE

(Continued on next page)

arxiv: v1 [cs.sy] 30 Nov 2017

FINITE HORIZON ROBUST MODEL PREDICTIVE CONTROL USING LINEAR MATRIX INEQUALITIES. Danlei Chu, Tongwen Chen, Horacio J. Marquez

Chapter 7. Digital Control Systems

Iterative Learning Control Analysis and Design I

A Novel Integral-Based Event Triggering Control for Linear Time-Invariant Systems

Additional Closed-Loop Frequency Response Material (Second edition, Chapter 14)

Design of Classical and State Space Controller Using a Moving Coil Machine Folayan Gbenga, Adeoye O. S.

Algorithm for Multiple Model Adaptive Control Based on Input-Output Plant Model

Actuator Fault Tolerant PID Controllers

Performance Comparison of PSO Based State Feedback Gain (K) Controller with LQR-PI and Integral Controller for Automatic Frequency Regulation

Index Accumulation, 53 Accuracy: numerical integration, sensor, 383, Adaptive tuning: expert system, 528 gain scheduling, 518, 529, 709,

A Method for PID Controller Tuning Using Nonlinear Control Techniques*

Transcription:

SICE Journal of Control, Measurement, and System Integration, Vol., No. 2, pp. 43 54, March 2008 Optimal Tuning of PID Parameters Using Iterative Learning Approach Jian-Xin XU, Deqing HUANG, and Srinivas PINDI Abstract : Proportional-integral-derivative (PID) controller is the most predominant industrial controller that constitutes more than 90% feedback loops. Time domain performance of PID, including peak overshoot, settling time and rise time, is directly relevant to PID parameters. In this work we propose an iterative learning tuning method (ILT) an optimal tuning method for PID parameters by means of iterative learning. PID parameters are updated whenever the same control task is repeated. The first novel property of the new tuning method is that the time domain performance or requirements can be incorporated directly into the objective function to be minimized. The second novel property is that the optimal tuning does not require as much the plant model knowledge as other PID tuning methods. The new tuning method is essentially applicable to any plants that are stabilizable by PID controllers. The third novel property is that any existing PID auto-tuning methods can be used to provide the initial setting of PID parameters, and the iterative learning process guarantees that a better PID controller can be achieved. The fourth novel property is that the iterative learning of PID parameters can be applied straightforward to discrete-time or sampled-data systems, in contrast to existing PID auto-tuning methods which are dedicated to continuous-time plants. In this paper, we further exploit efficient searching methods for the optimal tuning of PID parameters. Through theoretical analysis, comprehensive investigations on benchmarking examples, and real-time experiments on the level control of a coupled-tank system, the effectiveness of the proposed method is validated. Key Words : PID tuning, iterative learning, objective functions, searching methods, iterative learning tuning (ILT).. Introduction Among all the known controllers, the proportional-integralderivative (PID) controllers are always the first choice for industrial control processes owing to the simple structure, robust performance, and balanced control functionality under a wide range of operating conditions. However, the exact workings and mathematics behind PID methods vary with different industrial users. Tuning PID parameters (gains) remains a challenging issue and directly determines the effectiveness of PID controllers [] [3]. To address the PID design issue, much effort has been invested in developing systematic auto-tuning methods. These methods can be divided into three categories, where the classification is based on the availability of a plant model and model type, (i) non-parametric model methods; (ii) parametric model methods; (iii) model free methods. The non-parametric model methods use partial modelling information, usually including the steady state model and critical frequency points. These methods are more suitable for closedloop tuning and applied without the need for extensive priori plant information []. Relay feedback tuning method [4] [7] is a representative method of the first category. The parametric model methods require a linear model of the plant either transfer function matrix or state space model. To obtain such a model, standard off-line or on-line identification methods are often employed to acquire the model data. Thus parametric model methods are more suitable for off-line PID Department of Electrical and Computer Engineering, National University of Singapore, Singapore 7576. Department of Electrical Engineering, Indian Institute of Technology Guwahati, Guwahati-78039, India. E-mail: elexujx@nus.edu.sg (Received November 27, 2007) tuning [8]. When the plant model is known with a parametric structure, optimal design methods can be applied [9],[0]. As for PID parameter tuning, it can be formulated as the minimization of an objective function with possible design specifications such as the nominal performance, minimum input energy, robust stability, operational constraints, etc. In model free methods, no model or any particular points of the plant are identified. Three typical tuning methods are unfalsified control [], iterative feedback tuning [2] and extreme seeking [3]. In [], input-output data is used to determine whether a set of PID parameters meets performance specifications and these PID parameters are updated by an adaptive law based on whether or not the controller falsifies a given criterion. In [2], the PID controller is updated through minimizing an objective function that evaluate the closed-loop performance and estimating the system gradient. In [3], adaptive updating is conducted to tune PID parameters such that the output of the cost function reaches a local minimum or local maximum. In practice, when the plant model is partially unknown, it would be difficult to compute PID parameters even if the relationship between transient specifications and PID parameters can be derived. In many of existing PID tuning methods, whether model free or model based, test signals will have to be injected into the plant in order to find certain relevant information for setting controller parameters. This testing process may however be unacceptable in many real-time control tasks. On the other hand, many control tasks are carried out repeatedly, such as in batch processors. The first objective of this work is to explore the possibility of fully utilizing the task repetitiveness property, consequently provide a learning approach to improve PID controllers through iteratively tuning parameters when the transient behavior is of the main concern. JCMSI 0002/08/002 043 c 2007 SICE

44 SICE JCMSI, Vol., No. 2, March 2008 In most learning methods including neural learning and iterative learning [4], the process Jacobian or gradient plays the key role by providing the greatest descending direction for the learning mechanism to update inputs. The convergence property of these learning methods is solely dependent on the availability of the current information on the gradient. The gradient between the transient control specifications and PID parameters, however, may not be available if the plant model is unknown or partially unknown. Further, the gradient is a function of PID parameters, thus the magnitude and even the sign may vary. The most difficult scenario is when we do not know the sign changes apriori. In such circumstances, traditional learning methods cannot achieve learning convergence. The second objective of this work is to extend the iterative learning approach to deal with the unknown gradient problem for PID parameter tuning. It should be noted that in many industrial control problems such as in process industry, the plant is stable in a wide operation range under closed-loop PID, and the major concern for a PID tuning is the transient behaviors either in the time domain, such as peak overshoot, rise time, settling time, or in the frequency domain such as bandwidth, damping ratio and undamped natural frequency. From the control engineering point of view, it is one of the most challenges to directly address the transient performance, in comparison with the stability issues, by means of tuning control parameters. Even for a lower order linear time invariant plant under PID, the transient performance indices such as overshoot could be highly nonlinear in PID parameters and an analytical inverse mapping from overshoot to PID parameters may not exist. In other words, from the control specification on overshoot we are unable to decide the PID parameters analytically. The third objective of this work is to link these transient specifications with PID parameters and give a systematic tuning method. Another issue is concerned with the redundancy in PID parameter tuning when only one or two transient specifications are required, for instance when only overshoot is specified, or only the integrated absolute error is to be minimized. In order to fully utilize the extra degrees of freedom of the controller, the most common approach is to introduce an objective function and optimize the PID parameters accordingly. This traditional approach is however not directly applicable because of the unknown plant model, and in particular the unknown varying gradient. A solution to this problem is still iterative learning. An objective function, which is accessible, is chosen as the first step for PID parameter optimization. Since the goal is to minimize the objective function, the control inputs will be updated along the greatest descending direction, namely the gradient, of the objective function. In other words, the PID parameters are chosen to directly reduce the objective function, and the objective function is treated as the plant output and used to update the PID parameters. When the gradient is varying and unknown, extra learning trials will be conducted to search the best descending direction. The paper is organized as follows. Section 2 gives the formulation of PID auto-tuning problem. Section 3 introduces the iterative learning tuning approach. Section 4 shows the comparative studies on benchmark examples. Furthermore, Section 5 addresses the real-time implementation on a laboratory pilot plant. Section 6 concludes the work. 2. Formulation of PID Auto-Tuning Problem 2. PID Auto-Tuning In this work, we consider a fundamental PID controller in continuous or discrete-time C(s) = k p + k i s + k ds T s z C(z) = k p + k i z + k z d T s z where k p is the proportional gain, k i the integral gain, k d is the derivative gain, s is the Laplace operator, z is the Z operator, T s is the sampling period. Denote k = [k p, k i, k d ] T. The aim of PID tuning is to find appropriate values for PID parameters such that the closed-loop response can be significantly improved when comparing with the open-loop response. Since the PID control performance is determined by PID parameters k, by choosing a set of performance indices x, for instance overshoot and settling time, there exists a unique relationship or mapping f between x and k x = f(k). The PID auto-tuning problem can be mathematically formulated as to look for a set of k such that x meet the control requirements specified by x d. If the inverse mapping is available, we have k = f (x d ). The mapping f, however, is a vector valued function of x, k and the plant, and is in general a highly nonlinear mapping. Thus its inverse mapping in general is either not analytic solvable or not uniquely existing. Above all, the most difficult problem in PID auto-tuning is the lack of plant model, hence the mapping f is unknown or partially unknown. The importance of PID and challenge in PID auto-tuning attracted numerous researchers and come up with various autotuning or detuning methods, each has unique advantages and limitations. In this work we propose a new auto-tuning method using iterative learning and optimization, which complements existing PID tuning methods. 2.2 Performance Requirements and Objective Functions An objective function, or cost function, quantifies the effectiveness of a given controller in terms of the closed-loop response, either in time domain or frequency domain. A widely used objective function in PID auto-tuning is the integrated square error (ISE) function T J(k) = e 2 (t, k)dt, () T t 0 t 0 where the error e(t, k) = r(t) y(t, k) is the difference between the reference, r(t), and the output signal of the closed-loop system, y(t). T and t 0, with 0 t 0 < T <, are two design parameters. In several auto-tuning methods such as IFT and ES, t 0 is set approximately at the time when the step response of the closed-loop system reaches the first peak. Hence the cost function effectively places zero weighting on the initial transient portion of the response and the controller is tuned to minimize the error beyond the peak time. Similar objective functions, such as integrated absolute error, integrated time weighted absolute error, integrated time weighted square error, T t 0 T t 0 e dt, T t 0 T t 0 t e dt, T t 0 T t 0 te 2 dt,

SICE JCMSI, Vol., No. 2, March 2008 45 Fig. The nonlinear mapping between the peak overshoot 00M p and PD gains (k p, k d ) in continuous-time. Fig. 2 The nonlinear mapping between the settling time t s and PD gains (k p, k d ) in continuous-time. have also been widely used in the process of PID parameter tuning. In many control applications, however, the transient performance, such as overshoot M p, settling time t s, rise time t r, could be of the main concern. An objective function that can capture the transient response directly, is highly desirable. For this purpose, we propose a quadratic function J = (x d x) T Q(x d x) + k T Rk, (2) where x = [00M p, t s, t r ] T, x d is the desired x, Q and R are two constant non-negative weighting matrices. The selection of Q and R matrices will yield effects on the optimization results, that is, produce different closed-loop responses. A larger Q highlights more on the transient performance, whereas a larger R gives more control penalty. 2.3 A Second Order Example Consider a second order plant under a unity feedback with a PD controller. The PD controller and plant are respectively k C(s) = k p + k d s, G(s) = s 2 + as + b, (3) where k p, k d, k, a, b are all non-negative constants. The closedloop system is stable if a + d > 0andb + p > 0, where p = kk p and d = kk d. The nonlinear mapping f between the closed-loop transient response x = [M p, t s ] and the PD parameters k = [k p, k d ]is derived in Appendix, and shown in Figs. and 2, where a = 0., b = 0, and k =. It can be seen that the mapping f is nonlinear or even discontinuous. The nonlinear mappingf for discrete-time control system can also be derived but omitted here due to the complexity. Discretizing the plant in (3) with a sampling time T s = 0.s, the nonlinear mapping between M p and (k p, k d ) is shown in Fig. 3. It can be seen that there exist local minima in the surface, and the gradient may vary and take either positive or negative values. On the other hand, it can also been seen from those figures that the transient responses may vary drastically while the control parameters only vary slightly. This indicates the impor- Fig. 3 The nonlinear mapping between the peak overshoot 00M p and PD gains (k p, k d ) in discrete-time. An upper bound of 00 is applied to crop the vertical values. tance for PID parameter auto-tuning and the necessity for finding an effective tuning method. 3. Iterative Learning Approach Iterative learning is adopted to provide a solution to the PID auto-tuning. The iterative learning offers a desirable feature that it can guarantee the learning convergence even if the plant model is partially unknown. 3. Principal Idea of Iterative Learning The concept of iterative learning was first introduced in control to deal with a repeated control task without requiring the perfect knowledge such as the plant model or parameters [5]. It learns to generate a control action directly instead of doing a model identification. The iterative learning mechanism updates the present control action using information obtained from previous control actions and previous error signals. Let us first give the basic formulation of iterative learning in PID auto-tuning. From preceding discussions, the PID autotuning problem can be described by the mapping x = f(k)

46 SICE JCMSI, Vol., No. 2, March 2008 where k Ω k R n and x Ω x R m,wheren and m are integer numbers. The learning objective is to find a suitable set k such that the transient response x can reach a given region around the control specifications x d. The principal idea of iterative learning is to construct a contractive mapping A x d x i+ = A(x d x i ), where the norm of A is strictly less than, and the subscript i indicates that the quantity is in the ith iteration or learning trial. To achieve this contractive mapping, a simple iterative learning law is k i+ = k i +Γ i (x d x i ) (4) where Γ i R n m is a learning gain matrix. It can be seen that the learning law (4) generates a set of updated parameters from the previously tuned parameters, k i, and previous performance deviations x d x i. The schematic of the iterative learning mechanism for PID auto-tuning is shown in Fig. 4. When n = m, define the process gradient F(k) = f(k) k, we can derive the condition for the contractive mapping A x d x i+ = x d x i (x i+ x i ) = x d x i f(k i ) k (k i+ k i ) = [I F(k i )Γ i](x d x i ) (5) where k i [min{k i, k i+ }, max{k i, k i+ }] Ω k. Therefore we have a contractive mapping A x d x i+ = A(x d x i ) as far as the magnitude A = I F(k i )Γ i ρ<. (6) When n > m, there exists an infinite number of solutions because of redundancy in control parameters. With the extra degrees of freedom in PID, optimality can be exploited, for instance the shortest settling time, minimum peak overshoot, minimum values of control parameters, etc. A suitable objective function to be minimized could be a non-negative function J(x, k) 0, where J is accessible, such as the quadratic one (2). The minimization is in fact a searching task min J(x, k) k Ω k where x = f(k). The optimal parameters k opt can be obtained by differentiating J and computing dj dk = J dx x dk + J k = J x F + J = 0. (7) k In most cases k opt can only be found numerically due to the highly nonlinear relationship between k and quantities J, x. A major limitation of optimization methods is the demand for the complete plant model knowledge or the mapping f. On the contrary, iterative learning only requires the bounding knowledge of the process gradient. The principal idea of iterative learning can be extended to solving the optimization problem under the assumption that all gradient components with respect to k have known limiting bounds. Since the learning objective now is to directly reduce the value of J, the objective function J can be regarded as the process output to be minimized. The new iterative learning tuning law is k i+ = k i γ i J(k i ), (8) where γ i = [γ,i,,γ n,i ] T and J(k i ) denotes J(k i, f(k i )). To show the contractive mapping, note that J(k i+ ) = J(k i ) + [J(k i+ ) J(k i )] ( dj(k ) = J(k i ) + i ) T (k i+ k i ) dk where k i Ω k is in a region specified by k i and k i+. By substitution of the ILT law (8), we have ( dj(k ) J(k i+ ) = i ) T γ dk i J(k i ). (9) Fig. 4 The schematic block diagram of the iterative learning mechanism and PID control loop. The parameter correction is generated by the performance deviations x d x i multiplied by a learning gain Γ i. The operator z denotes one iteration delay. The new PID parameters k i+ consists of the previous k i and the correction term, analogous to a discrete-time integrator. The iterative learning tuning mechanism is shown by the block enclosed by the dashed line. r is the desired output and the block M is a feature extraction mechanism that records the required transient quantities such as overshoot from the output response y i+. The convergence property is determined by the learning gains γ i and the gradient dj/dk. 3.2 Learning Gain Design Based on Gradient Information To guarantee the contractive mapping (9), the magnitude relationship must satisfy D T i γ i ρ<, where ( dj(k ) D i = i ) dk is the gradient of the objective function. The selection of learning gain γ i is highly related to the prior knowledge on the gradient D i. Consider three scenarios and assume D i = (D,i, D 2,i, D 3,i ), that is, a PID controller is used. When D i is known apriori, we can choose γ j,i = D j,i /3. Such a selection produces the fastest learning convergence speed, that is, convergence in one iteration because D T i γ i = 0. When the bounding knowledge and the sign information of D j,i are available, the learning convergence can also be guaranteed. For instance assume 0 < α j D j,i β j < for

SICE JCMSI, Vol., No. 2, March 2008 47 k i Ω k,whereα j and β j are respectively lower and upper bounds of the gradient components D j,i. In such circumstances, choosing γ j,i = /3β j, the upper bound of the magnitude is ρ = α /3β α 2 /3β 2 α 3 /3β 3 <. The most difficult scenario is when bounding functions or signs of D j,i are unknown. In order to derive the iterative learning convergence, it can be seen from (9) that we do not need the exact knowledge about the mapping f. It is adequate to know the bounding knowledge (amplitude and sign) of the gradient D(k i )orf(k i ). Although an invariant gradient is assumed for most iterative learning problems, the elements in D(k i )may change sign and take either positive or negative values. Without knowing the exact knowledge of the mapping f, wemay not be able to predict the varying signs in the gradient. Now, in the iterative learning law (8) the learning gains will have to change signs according to the gradient. Here the question is how to change the signs of the learning gains when we do not know the signs of the gradient components, i.e. how to search the direction of the gradient, which obviously can only be done in a model free manner if f is unknown. A solution to the problem with unknown gradient is to conduct extra learning trials to determine the direction of gradient or the signs of the learning gains directly γ i = [±γ,, ±γ n ] T, (0) where γ i are positive constants. From the derivation (5), when learning gains are chosen appropriately, A < and the learning error reduces. On the other hand, if learning gains are chosen inappropriately, then A > and the error increases after this learning trial. Therefore, several learning trials are adequate for ILT mechanism to determine the correct signs of the learning gains. In general, when there are two gradient components (D,i, D 2,i ), there are 4 sets of signs {, }, {, }, {, }, and {, }, corresponding to all possibles signs of the gradient (D,i, D 2,i ). In such circumstances, at most 4 learning trials are sufficient to find the greatest descending among the four control directions, as shown in Fig. 5. Similarly, if there are three free tuning parameters, there will be 8 sets of signs in the gradient (D,i, D 2,i, D 3,i ), as shown in Fig. 6. In general, if there are n control tuning parameters, there will be n gradient components D j,i. Since each γ j,i takes either positive or negative sign, there will be 2 n combinations and at most 2 n learning trials are required. In addition to the estimation of the gradient direction or learning direction, the magnitudes of learning gains γ i should also be adjusted to satisfy the learning convergence condition D T i γ i ρ<. Since the gradient D i is a function of PID parameters k i, the magnitude of D i varies at different iterations. When the magnitudes of the gradient is unknown, extra learning trials will be needed to search for suitable magnitudes of learning gains. In this work, we adopt a self-adaption rule [6] which scales learning gains up and down by a factor ζ =.839, that is, each component γ j,i will be adjusted to γ j,i ζ and γ j,i /ζ. Extra learning trials will be performed with the scaled learning gains. It is reported that the choice of such a scaling factor ζ will lead to a linear convergence rate [6],[7]. To facilitate searching and reduce the number of trials, we can also estimate gradient components D j,i numerically using Fig. 5 Fig. 6 There are four pairs of signs for the gradient (D, D 2 ) as indicated by the arrows. Hence there are four possible updating directions, in which one pair gives the fastest descending direction. There are three gradient components D, D 2 and D 3 with respect to three control parameters. Consequently there are 8 possible tuning directions and at most 8 learning trials are required to find the correct updating direction. the values of the objective function and PID parameters obtained from previous two iterations ˆD j,i = J(k i ) J(k i 2 ). () k j,i k j,i 2 The learning gain can be revised accordingly as γ j,i = λ j ˆD j,i where λ j is a constant gain in the interval of (0, ]. This method is in essence the Secant method along the iteration axis, which can effectively expedite the learning speed [8]. Since gradient direction is a critical issue in searching, and the approximation () may not always guarantee a correct sign, we can use the estimation result in () partially by retaining the magnitude estimation, while still searching the correct control direction k j,i k j,i 2 γ j,i = ±λ j J(k i ) J(k i 2 ). (2) 3.3 Iterative Searching Methods Three iterative searching methods are considered for ILT in this work with verifications and comparisons. They are M 0 an exhaustive searching method in directions and magnitude, M an exhaustive searching method in directions, M 2 a lazy searching method. M 0 does exhaustive searching in all 2 n directions and exhaustive searching in all 2 n magnitudes using self-adaptation with the factor ζ. Then PID parameters will be updated using the set that generates the best closed-loop response or yields the

48 SICE JCMSI, Vol., No. 2, March 2008 biggest drop of J in that trial. With the best tuned PID parameters as the initial setting, the ILT mechanism enters another run of exhaustive searching for the best response and best PID parameters. The searching process repeats until the stopping criterion is met. This is the worst case searching where neither the gradient directions nor the gradient magnitudes are available. For each run of searching the greatest descending direction, 4 n trials are performed. For PID tuning where n = 3, 64 trials are needed for one run of searching. Clearly, this searching method is not efficient. In M the entire searching and updating process is similar to M 0 except that the magnitudes of learning gains are determined using the formula (2). Hence it reduces the total number of trials from 4 n in M 0 to 2 n for each run of searching. For PID tuning where n = 3, only 8 trials are needed. M 2 does exhaustive searching in directions in the first run of searching, and the magnitudes of learning gains are determined using the formula (2). Then the greatest descending direction will be used for subsequent runs of searching, with the assumption that the mapping f is in general smooth and drastic variations in gradient rarely occur. The exhaustive searching in directions will be activated again when the stopping criterion is met. The searching process will permanently stop if the stop criterion is still met after exhaustive searching in all directions. The initial magnitudes of learning gains can be set as [γ,0,γ 2,0,γ 3,0 ] = γ 0 [k p,0, k i,0, k d,0 ], (3) J 0 where γ 0 is a positive constant, and chosen to be 0. inthis work. k p,0, k i,0, k d,0 are initial values of PID parameters determined using any existing PID auto-tuning methods. J 0 is calculated with k p,0, k i,0, k d,0 and the corresponding closed-loop response. 4. Comparative Studies on Benchmark Examples In this section we conduct comprehensive tests on 8 benchmark plant models. Four plant models G to G 4 were used in [3] to compare and validate several iterative tuning methods G (s) = + 20s e 5s, G 2 (s) = + 20s e 20s, G 3 (s) = ( + 0s), 8 5s G 4 (s) = ( + 0s)( + 20s), where G and G 2 have relatively large normalized time delay (NTD), G 3 has high-order repeated poles, and G 4 is nonminimum phase. The other four plant models G 5 to G 8 were used in [9],[20] to validate their auto-tuning methods G 5 (s) = G 6 (s) = G 7 (s) = G 8 (s) = (s + )(s + 5) 2 e 0.5s, (25s 2 + 5s + )(5s + ) e s, (s 2 + 2s + 3)(s + 3) e 0.3s, (s 2 + s + )(s + 2) 2 e 0.s, where G 5 is a high-order plant with medium NTD, G 6 is a highorder and moderately oscillatory plant with short NTD, G 7 is a high-order and heavily oscillatory plant with short NTD, and G 8 has both oscillatory and repeated poles. To make fair comparisons, we choose initial learning gains with form (3) for PID parameters in all case studies. In all searching results, we use N to denote the total number of trials. 4. Comparisons between Objective Functions Objective function ISE () has been widely investigated and adopted in PID tuning. The quadratic objective function (2), on the other hand, has more weights to prioritize transient performance requirements. When control requirements are directly concerned with transient response such as M p or t s, we can only use the quadratic objective function (2). To show the effects of different objective functions and weight selections, we use plants G G 4. To make fair comparisons, the learning process starts from the same initial setting for PID gains generated by Ziegler Nicholes (ZN) tuning method [8]. Exhaustive searching method M 0 is employed. The tuning results through iterative learning are summarized in Table. For simplicity only M p and t s are taken into consideration in (2). The learning process stops when the drop of an objective function between two consecutive iterations is lower than ε = 0 6 for () and ε = 0.0 for (2). In ISE, the parameters are T = 00, 300, 500, 200 s and t 0 = 0, 50, 40, 30 s respectively [3]. Usually settling time is much greater than overshoot, thus 00M p is used. When plants have much bigger settling time, we can choose t s instead of (t s ) 2 in the objective function, as shown in cases of G 2 and G 3. Byscalingdownt s in objective functions, overshoot decreases as it is weighted more. Meanwhile t s increases as it is weighted less. Comparing ISE and quadratic objective function, we can see that latter offers more choices. 4.2 Comparisons between ILT and Existing Iterative Tuning Methods Now we compare iterative learning based tuning with other iterative tuning methods such as extremum seeking (ES) tuning and iterative feedback tuning (IFT). ZN tuning method is also included for comparison. G G 4 and the same ISE [3] for each model are used. The PID controller parameters given by ZN are used as a starting point for ILT tuning. Exhaustive searching method M 0 is employed in ILT. The results verify that ILT, ES and IFT give very similar responses which are much superior than that of ZN tuning method. Figure 7 shows the ILT results for G,(a)showsthe decreasing objective function J, (b) shows decreasing performance indices M p and t s, (c) shows the variations of the PID parameters, and (d) compares step responses with four tuning methods. Figure 8 shows the searching results of the gradient directions and the variations of the learning gains through selfadaptation. It can be seen that the gradients undergo changes in signs, hence it is in general a difficult and challenging optimization problem. Table 2 summarizes the comparative results with all four plants G G 4 when 4 tuning methods were applied. The iteration numbers when applying ILT for G G 4 are 786, 52, 024 and 26 respectively. Although ILT shares similar performance as ES and IFT, it

Table SICE JCMSI, Vol., No. 2, March 2008 49 Control performances of G G 4 using the proposed ILT method. Plant J PID Controller 00M p t s N ISE 3.59 + 0.3s + 7.54s 4.48 2.35 768 G (00M p ) 2 + 0.5t 2 s 3.38 + 0.3s + 7.05s.99 2.7 024 (00M p ) 2 + 0.t 2 s 3.25 + 0.2s + 6.30s 0.63 2.83 832 (00M p ) 2 + 0.0t 2 s 3.7 + 0.s + 9.s 0.53 22.24 52 ISE 0.93 + 0.03s + 5.67s 0.7 50.67 52 G 2 (00M p ) 2 + t s 0.99 + 0.032s + 6.87s.06 47.99 600 (00M p ) 2 + 0.2t s.05 + 0.028s + 9.79s 0.29 82.74 52 (00M p ) 2 + 0.t s.03 + 0.029s + 9.8s 0.20 83.6 640 ISE 0.64 + 0.02s +.3s 0.49 37.3 024 G 3 (00M p ) 2 + t s 0.76 + 0.03s + 6.65s.93 20.56 576 (00M p ) 2 + 0.2t s 0.85 + 0.04s + 25.77s 0.66 22.04 92 (00M p ) 2 + 0.t s 0.83 + 0.04s + 24.9s 0.62 22.76 92 ISE 5.0 + 0.092s + 25.59s 3.05 25.2 26 G 4 (00M p ) 2 + 0.25t 2 s 4.3 + 0.075s + 22.9s.8 8.63 52 (00M p ) 2 + 0.t 2 s 3.89 + 0.07s + 22.28s.70 20.56 384 (00M p ) 2 + 0.0t 2 s 4.5 + 0.075s + 23.96s 0.06 9.27 26 Fig. 7 ILT performance for G. (a) The evolution of the objective function; (b) The evolution of overshoot and settling time; (c) The evolution of PID parameters; (d) The comparisons of step responses among ZN, IFT, ES and ILT, where IFT, ES and ILT show almost the same responses. is worth to highlight some important factors in the tuning process. In ES tuning, there are more than 0 design parameters to be set properly. From [3], design parameters take rather specific values. In IFT, the initial values of the PID parameters must be chosen in such a way as to give an initial response that is very slow and with no overshoot. Further, in IFT the transient performance is purposely excluded from the objective function by choosing a sufficiently large t 0. On the contrary, in ILT only initial learning gains need to be preset, and we choose all three learning gains with a uniform value γ 0 = 0. in(3) for the ease of ILT design. In fact, by choosing initial learning gains with different values, we can achieve much better responses than those in Table 2. Needless to mention that ILT can handle both types of objective functions () and (2) with the flexibility to highlight transient behaviors. 4.3 Comparisons between ILT and Existing Auto-Tuning Methods PID auto-tuning methods [8],[9],[20] provided several effective ways to determine PID parameters. In this subsection we compare ILT with the auto-tuning method [8] based on internal model control (IMC), and the auto-tuning method [20] based on pole-placement (PPT) which shows superior performance than [9]. Comparisons are made based on plants G 5 G 8 which were used as benchmarks. Using ISE as the objective function and searching method M 0 in ILT, the tuning results are summarized in Table 3, where the start points of PID parameters are adopted from the tuning results in [9]. Comparing with the results auto-tuned by the IMC method and PPT method, ILT achieves better performance after learning.

50 SICE JCMSI, Vol., No. 2, March 2008 Fig. 8 ILT searching results for G. (a) The evolution of the gradient directions; (b) The evolution of the magnitudes of learning gains with self-adaptation. Table 2 Control performances of G G 4 using methods ZN, IFT, ES and ILT. Plant Method PID Controller 00M p t s ZN 4.06 + 0.44s + 9.38s 46.50 47.90 G IFT 3.67 + 0.3s + 7.74s 5.38 2.38 ES 3.58 + 0.3s + 7.68s 3.3 2.35 ILT 3.59 + 0.3s + 7.54s 4.48 2.35 ZN.33 + 0.043s + 0.30s 2.75 09.4 G 2 IFT 0.93 + 0.03s + 5.64s 0.80 50.33 ES.0 + 0.032s + 7.23s.37 76.6 ILT 0.93 + 0.03s + 5.67s 0.7 50.67 ZN.0 + 0.05s + 20.9s 3.95 336.90 G 3 IFT 0.66 + 0.02s + 2.08s 0.98 32.05 ES 0.68 + 0.03s + 3.30s 0.96 30.4 ILT 0.64 + 0.02s +.30s 0.49 37.3 ZN 3.53 + 0.2s + 4.80s 53.70 86.2 G 4 IFT 3.03 + 0.065s + 8.42s 0.55 28.74 ES 3.35 + 0.068s + 2.40s 0.8 29.80 ILT 5.0 + 0.092s + 25.59s 3.05 25.2 Table 3 Control performances of G 5 G 8 usingimc,pptandiltmethods. Plant Method PID Controller 00M p t s IMC 33.46 + 6.67s + 8.54s 2.09 5.7 G 5 PPT 27.02 + 2.4s + 6.45s 2.2 5.0 ILT 3.96 + 8.8s +.0s.3 2.09 IMC 0.88 + 0.066s +.20s 3.46 74.6 G 6 PPT 0.44 + 0.074s + 2.54s 5.0 49.8 ILT 0.57 + 0.075s + 2.78s.7 23.9 IMC 7.06 + 5.29s +.64s 5.5 8.37 G 7 PPT 3.89 + 5.39s + 2.5s 3.2 5.32 ILT 4.5 + 5.40s + 2.53s 0.44 2.93 IMC 2.79 +.33s +.9s 3.0 4.74 G 8 PPT.50 +.37s +.72s 3.04 9.44 ILT 2.8 +.48s + 2.34s.8 4.6 4.4 Comparisons between Searching Methods Now we investigate the effects of searching methods M 0, M and M 2.PlantsG G 8 are used. The objective function to be used is J = (00M p ) 2 + t2 s 00. (4) Set λ j = 0.2, j =, 2, 3 in (2) for M and M 2. The stopping criterion for ILT is ε = 0.0. The control performance after learning is summarized in Tables 4 and 5. It can be seen that M and M 2 achieve similar performance, which is slightly inferior than M 0. However, comparing with M 0 the learning trial numbers in M and M 2 have been significantly reduced. 4.5 ILT for Sampled-Data Systems A promising feature of ILT is the applicability to sampleddata or discrete-time systems. To illustrate how ILT works for digital systems, consider plant G 4 which can be discretized using sampler and zero order hold ( 2.5e.05T s +.5e.T s + ) z + e.5t s 2.5e 0.T s +.5e.05T s G 4 (z) =, z 2 (e 0.T s + e 0.05T s)z + e 0.5T s where the sampling period T s = 0. s. The digital PID controller is used. Choose again (4) as the objective function, and use ZN to generate the initial values for PID parameters. The closed-loop responses using ILT are summarized in Table 6. For comparison all three searching methods M 0, M and M 2 are used. It can be seen that in all cases the control responses have been improved drastically, especially the reduction in overshoot. 5. Real-Time Implementation In order to show the applicability and effectiveness of the proposed ILT, real-time experiment has been carried out on a couple tank. 5. Experimental Setup and Plant Modelling The couple tank equipment consists of two small perspex tower-type tanks and interconnected through an hole which yields a hydraulic resistance (Fig. 9). The experimental setup in this work was configured such that the level is measured from the tank-2 while the water is pumped into the tank- as the control input. The outlet of tank-2 is used to discharge the water into the reservoir. The measurement data is collected from couple tank using NI data acquisition card USB-6008. The control method is programmed using Labview. A window-type smoothing filter using 00 samples is implemented to mitigate measurement noise. The sampling period is 0.25 second.

SICE JCMSI, Vol., No. 2, March 2008 5 Table 4 Control performance of G G 8 using searching methods M 0, M, M 2. Plant M 0 M M 2 00M p t s N 00M p t s N 00M p t s N G 0.53 22.24 52 0.24 22.3 60 0.23 22.23 29 G 2.80 78.64 256 0.73 79.93 72.47 79.9 43 G 3.8 2.29 92.36 209.77 48.8 205.35 4 G 4 0.06 9.27 26 0 5.92 20 0 22.93 4 G 5 0.282 2.6 640 0.2 2.44 48 0 3.34 70 G 6 0.605 25.04 320 0.03 48.04 40 0.03 48.04 9 G 7 0 6.05 92 0 6.3 32 0 6.5 28 G 8 0.9 8.25 768 0 0.53 32 0 0.74 8 Table 5 Final controllers for G G 8 by using searching methods M 0, M, M 2. Plant M 0 M M 2 G 3.59 + 0.3s + 7.54s 3.42 + 0.s + 8.77s 3.38 + 0.2s + 7.80s G 2 0.93 + 0.029s + 9.43s 0.94 + 0.030s + 7.28s.05 + 0.03s + 8.98s G 3 0.75 + 0.03s + 6.26s 0.87 + 0.02s + 24.73s 0.72 + 0.02s + 2.42s G 4 4.5 + 0.075s + 23.96s 5.99 + 0.097s + 30.70s 6.92 + 0.0s + 39.8s G 5 3.52 + 8.3s + 0.76s 29.63 + 7.65s + 0.9s 7.00 + 3.68s + 0.54s G 6 0.5 + 0.070s + 2.4s 0.60 + 0.069s + 2.7s 0.60 + 0.069s + 2.7s G 7 5.34 + 4.86s +.28s 5.02 + 4.8s +.3s 4. + 4.8s +.3s G 8 2.03 +.27s + 2.7s 2.3 +.2s + 0.94s.74 +.2s + 0.94s Table 6 Digital Control Results. Initial performance is achieved by ZN tuned PID. Final performance is achieved by ILT. T s Initial Performance Method Final Final Performance 00M p t s k p, k i, k d 00M p t s N M 0 3.56, 0.0, 2.62 0.3 4.77 896 0.0 39.6 64.4 M 4.04, 0.2, 25.26 0.43.25 80 M 2 3.24, 0., 2.2.2 7.64 43 M 0 3.45, 0.0, 20.89 0.057 5.45 768 0.05 39.52 63.90 M 4.04, 0.2, 25.26 0.55 0.85 80 M 2 3.25, 0., 2.2.0 7.35 43 M 0 3.53, 0., 2.76 0.060 3.60 024 0.2 40.92 63.00 M 4.08, 0.5, 28.2.02.80 96 M 2 2.80, 0.090, 6.4 0.70 20.20 56 M 0 3.39, 0., 2.60 0.00.50 832 0.5 44.26 77.00 M 3.48, 0.3, 24.05 0.97 5.50 72 M 2 2.69, 0.086, 5.6 0.74 20.00 48 M 0 2.22, 0.082, 7.60.08 34.00 576 2 76.96 88.00 M 2.9, 0.078, 5.22 0.32 34.00 28 M 2 2.6, 0.075, 3.94 0.0020 8.00 8 A step response is conducted to approximate the couple tank dynamics with a first order plus time delay model [8] k G(s) = τs + e sl (5) where k is the plant DC gain, τ is the time constant, L is the transportation lag. As shown in Fig. 0, after conducting a step response the obtained plant is G(s) = 68 + 65s e 3.6s. (6) 5.2 Application of ILT Method In the experiments, we conducts a series of tests to investigate ILT. The quadratic objective function is J = (00M p ) 2 + qt 2 s where two values q = 0.andq = 0.0 are used. The ISE is Fig. 9 Diagram of couple tank apparatus. T t 0 T t 0 e 2 dt,

52 SICE JCMSI, Vol., No. 2, March 2008 Fig. 0 Step response based modelling. Table 7 Experimental results. Method Constraint J PID Controller 00M p t s N M 0 C (00M p ) 2 + 0.0t 2 s 0. + 0.0029s + 0.26s 0.089 6.65 384 C ISE 0.6 + 0.0045s + 0.32s 4.8 34.04 56 C 2 ISE 0.4 + 0.0039s + 0.20s 4.29 30.22 32 M C (00M p ) 2 ++0.0t 2 s 0.2 + 0.0036s + 0.43s 0. 6.58 56 C 2 (00M p ) 2 + 0.0t 2 s 0.6 + 0.004s + 0.34s 0.78 0.45 24 C (00M p ) 2 + 0.t 2 s 0.26 + 0.0037s + 0.45s.40 9.35 48 C 2 (00M p ) 2 + 0.t 2 s 0.22 + 0.0036s + 0.44s.44 0.0 32 C ISE 0.26 + 0.0040s + 0.62s.30 0.50 5 C 2 ISE 0.23 + 0.0042s + 0.56s 2.9 2.98 3 M 2 C (00M p ) 2 + 0.0t 2 s 0.32 + 0.0034s + 0.76s 0.27 7.30 7 C 2 (00M p ) 2 + 0.0t 2 s 0.30 + 0.0034s + 0.72s 0.50 6.63 4 C (00M p ) 2 + 0.t 2 s 0.30 + 0.0038s + 0.72s.30 6.33 4 C 2 (00M p ) 2 + 0.t 2 s 0.23 + 0.0037s + 0.56s.7 5.99 0 where T = 24 s and t 0 = 27.4 s. All three search methods M 0, M and M 2 are applied. It is observed in the experiment that the calculated gradient components () or (2) may be singular some times. This is due to the presence of measurement noise and the closeness of the values of J at two adjacent iterations. On the other hand, the learning process may become sluggish when the PID parameters at two adjacent iterations are too close, yielding a very lower learning gains γ j,i. To solve these two problems, a constraint is applied when updating the learning gains c k j,i γ j,i c 2 k j,i, (7) where 0 c < c 2. The upper and lower learning gain bounds are made in proportion to PID parameters. The rationale is clear. When a controller parameter is bigger, we can update it with a larger bound without incurring drastic changes. If setting absolute bounds for the learning gain, an overly small bound would limit the parameter updating speed, and an overly large bound would make the constraint ineffective. In the realtime application we consider 2 sets of boundaries C : c = 0.02 c 2 = 0.2, C 2 : c = 0.05 c 2 = 0.4. The initial PID controller is found using ZN tuning method 0. + 0.0 + 0.26s. s The transient performance with ZN tuned PID is 00M p = 30.02 and t s = 58.73 s. ILT is applied to improve the performance. 5.3 Experimental Results The results are summarized in Table 7. It can be seen that searching methods M and M 2 can significantly reduce the number of learning iterations while M p and t s can be maintained at almost the same level. By changing the weight for t s, the final overshoot M p and settling time t s can be adjusted. The constraint C 2 can also reduce the trial number, because higher limits in learning gains will expedite the learning progress. In experiments, it is observed that M 2 can be further simplified by removing the last run of exhaustive searching, so long as the variation of J reaches the preset threshold ε. 6. Conclusion A new PID auto-tuning method is developed and compared with several well established auto-tuning methods including ZN, IFT, ES, PPT, IMC. Iterative learning tuning provides a sustained improvement in closed-loop control performance, and offers extra degrees of freedom in specifying the transient control requirements through a new objective function. The iterative learning tuning method proposed in this work can be further extended to various optimal designs for controllers, owing to its model free nature. For instance, by iterative searching and task repeatability, we can tune the parameters of lead and lag compensators, filters, observers, and use both time domain and frequency performance indices in objective functions. The future issues associated with the iterative learning tuning approach we are looking into include the robustness of ILT when the control process is not completely repeatable, the incorporation of various practical constraints, improvement of searching efficiency for high dimension multi-loop control systems, as well as ILT for closed-loop systems with lower sta-

SICE JCMSI, Vol., No. 2, March 2008 53 bility margins such as the sampled-data system with large time delays. References [] M. A. Johnson and M. H. Moradi: PID Controller Design, Spring-Verlag, London, 2003. [2] A. O Dwyer: Handbook of PI and PID Controller Tuning Rules, Imperial College Press, London, 2003. [3] K. J. Astrom, T. Hagglund, C. C. Hang, and W. K. Ho: Automatic tuning and adaptation for PID controllers A survery, Control Engineering Practice, Vol., No. 4, pp. 699 74, 993. [4] K. J. Astrom and T. Hagglund: Automatic tuning of simple regulators with specifications on phase and amplitude margins, Automatica, Vol. 20, No. 5, pp. 645 65, 984. [5] A. A. Voda and I. D. Landau: A method for the auto-calibration of PID controllers, Automatica, Vol. 3, No., pp. 4 53, 995. [6] A. Leva: PID autotuning method based on relay feedback, IEEE Proceeding D Control Theory and Applications, Vol. 40, No. 5, pp. 328 338, 993. [7] G. K. I. Mann, B.-G. Hu, and R. G. Gosine: Time-domain based design and analysis of new PID tuning rules, IEE Proceedings on Control Theory and Applications, Vol. 48, No. 3, pp. 25 26, 200. [8] K. J. Astrom and T. Hagglund: PID Controllers: Theory, Design and Tuning, 2nd ed., Research Triangle Park, NC: Instrum. Soc. Amer., 995. [9] M. Ikeda and D. D. Siljak: Optimality and robustness of linearquadratic control for nonlinear systems, Automatica, Vol. 26, No. 3, pp. 499 5, 990. [0] C. Y. Kuo, C. L. Yang, and C. Margolin: Optimal controller design for nonlinear systems, IEE Proceedings on Control Theory and Applications, Vol. 45, No., pp. 97 05, 998. [] M. Saeki: Unfalsified control approach to parameter space design of PID controllers, Proceedings of 42nd IEEE Conference on Decision and Control, pp. 786 79, 2003. [2] O. Lequin, E. Bosmans, and T. Triest: Iterative feedback tuning of PID parameters: Comparison with classical tuning rules, Control Engineering Practice, Vol., No. 9, pp. 023 033, 2003. [3] N. J. Killingsworth and M. Krstic: PID tuning using extremum seeking: online, model-free performance optimization, IEEE Control Systems Magazine, Vol. 26, No., pp. 70 79, 2006. [4] Z. Bien and J.-X. Xu: Iterative Learning Control - Analysis, Design, Integration and Applications, Kluwer Academic Press, Boston, USA, 998. [5] A. Arimoto, S. Kawamura, and F. Miyazaki: Bettering operation of robots by learning, Journal of Robotic Systems, Vol., pp. 23 40, 984. [6] R. Salomon: Evolutionary methods and Gradient Search: Similarities and Differences, IEEE Transactions On Evolutionary Computation, Vol. 2, No. 2, 998. [7] R. Salomon and J. L. van Hemmen: Accelerating backpropagation through dynamic self-adaptation, Neural Networks,Vol. 9, No. 4, pp. 589 60, 996. [8] J.-X. Xu and Y. Tan: Linear and Nonlinear Iterative Learning Control, Springer-Verlag, Berlin, 2003. [9] W. K. Ho, O. P. Gan, E. B. Tay, and E. L. Ang: Performance and gain and phase margins of well-known PID tuning formulas, IEEE Transactions on Control Systems Technology,Vol.4, No. 3, pp. 473 477, 996. [20] Q. Wang, T. Lee, H. Fung, Q. Bi, and Y. Zhang: PID tuning for improved performance, IEEE Transactions on Control Systems Technology, Vol. 7, No. 4, pp. 457 465, 999. Appendix The nonlinear mapping f will be explored for three cases with underdamped, overdamped and critical damped characteristics. Underdamped Case When the system parameters satisfy (a+kk d ) 2 4(b+kk p ) < 0, the closed-loop system shows underdamped response with two complex poles. The system response in the time domain is where y(t) = A A + λ 2 e αt cos(βt tan λ), (A. ) A = β = 2 λ = p b + p, α = a + d 2, 4(b + p) (a + d) 2, ap 2bd pd p 4(b + p) (a + d) 2 with p = kk p and d = kk d. It is easy to see that the steady-state output is A and M p = + λ 2 e αt p cos ( βt p tan λ ), (A. 2) t s = α ln ( 50 + λ 2), (A. 3) where the peak time is t p = β tan βλ α β+αλ β [ π + tan βλ α β+αλ ] if βλ α β+αλ 0, if βλ α β+αλ < 0. Note that the settling time t s, corresponding to 2% of the final steady state value, is calculated from the worst case condition + λ2 e αt cos ( βt tan λ ) < 0.02 by ignoring the factor cos ( βt tan λ ). Overdamped Case When the system parameters satisfy (a + kk d ) 2 4(b + kk p ) > 0, the closed-loop shows overdamped response and the closedloop system has two distinct real poles. The system response in the time domain is y(t) = A [ ] d 2 β p e (α+β)t β(α + β) + [ ] d 2 β p e (α β)t. (A. 4) β(α β) where α = a + d 2, β = 2 (a + d) 2 4(b + p). The relationship α>βalways holds since b + p > 0. Therefore the steady state is still A. First derive the peak overshoot. From dy dt = 2β [(dα + dβ p)e (α+β)t + ( dα + dβ + p)e (α β)t ], and let dy/dt = 0 we can obtain the peak time t p = ( ) dα dβ p 2β ln. dα + dβ p The output has overshoot if and only if a, b, p, d lie in the domain

54 SICE JCMSI, Vol., No. 2, March 2008 D = {(a, b, p, d) dα dβ p > 0}. When the system parameters lie in the domain D, M p = y(t p) A A = ( ) dα dβ p α β 2β (α + β)(dα dβ p) 2βp dα + dβ p ) dα dβ p α+β 2β (α β)(dα + dβ p) dα + dβ p. Now derive the settling time. Let y(t) A A < 0.02 for all t t s. If the overshoot exceeds 2%, then t s is the bigger real root of the following equation 2 [ d β p β(α + β) = 0.02 p b + p ; ] e (α+β)t + [ d 2 β ] p e (α β)t β(α β) otherwise, t s is the unique real root of the following equation [ ] d 2 β p e (α+β)t + [ ] d β(α + β) 2 β p e (α β)t β(α β) = 0.02 p b + p. Note that in neither cases can we obtain the analytic expression of t s. Critical Damped Case When the system parameters satisfy (a+kk d ) 2 4(b+kk p ) = 0, the closed-loop system shows critical damped response with two identical real poles. The system response in the time domain is p + (p dα)αt y(t) = A e αt (A. 5) b + p with the steady-state A. First derive the peak overshoot. Letting Jian-Xin XU Received the Bachelor degree from Zhejiang University, China in 982. He attended the University of Tokyo, Japan, where he received the Master s and PhD degrees in 986 and 989 respectively. All degrees are in Electrical Engineering. He worked for one year in the Hitachi research Laboratory, Japan; for more than one year in Ohio State University, U.S.A. as a visiting scholar; and for 6 months in Yale University as a visiting research Fellow. In 99 he joined the National University of Singapore, and is currently a professor in the Department of Electrical and Computer Engineering. He is a senior member of IEEE. Deqing HUANG Received the Bachelor and PhD degrees from Mathematical College Sichuan University, China in 2002 and 2007 respectively. Currently he is a PhD student at the Department of Electrical and Computer Engineering, National University of Singapore. His research interests lie in the area of iterative learning control. Srinivas PINDI Has done his BE in Electrical and Electronics from Osmania University, Hyderabad, India. after he worked as Scientific Officer in Bhabha Atomic Research Centre (BARC), India for 2 years. Currently pursuing his Master of Technology from Indian Institute of Technology Guwahati. dy dt = α2 (d + (p dα)t) e αt = 0 b + p yields a unique solution t p = d p dα = 2d 2p ad d, 2 implying that the overshoot exists if and only if a 2 < 4b + d 2. The overshoot is M p = d2 (a + d) 2 p( a 2 + 4b + 3d 2 + 2ad) e 2d(a+d) p( a 2 + 4b + d 2 a ) 2 +4b+d 2. Next derive the settling time. If the overshoot exceeds 2%, then t s is the bigger real root of the following equation p + (p dα)αt e αt = 0.02 p b + p b + p ; otherwise, t s is the unique real root of the following equation p + (p dα)αt e αt = 0.02 p b + p b + p. Note that in neither cases can we obtain the analytic expression of t s.