# Minimum Fuel Optimal Control Example For A Scalar System

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 Minimum Fuel Optimal Control Example For A Scalar System A. Problem Statement This example illustrates the minimum fuel optimal control problem for a particular first-order (scalar) system. The derivation of the general solution to this problem is found in the course textbook. The expressions for the switching curve, the switching time, and the minimum time will be developed, and several examples with different initial states and different target states will be presented. The general form for the system model, control constraint, and minimum fuel performance index are Z T =f [,t]+b[,t] u(t), u i (t), i [,m], J = c T u(t) dt () where x < n and u < m. The specific expressions for the first-order system for this example, with n = m =, are given by Z T = +u(t), u(t), J = u(t) dt () For this system the terms in the general state space model and performance index have the following values: f [,t]=, B [,t]=,andc T =. The Hamiltonian for the system is H(t) = u(t) λ(t)+λ(t)u(t) (3) where λ(t) is the scalar costate variable (Lagrange multiplier) for this system. The solution to the minimum fuel problem is the control signal u(t) that satisfies the state and costate equations and that minimizes the Hamiltonian, such that which for this problem becomes H [x,λ,u,t] H [x,λ,u,t] admissible u(t) (4) Minimize u(t) + λ(t)u(t), u(t) (5) The solution to this minimization problem follows the form for the general case, and is b u T (t) = dez i λ(t) = dez [λ(t)] (6) where dez is the deadzone function defined by OUT =, when IN > dez = OUT = when IN < OUT = when IN c i B. Costate Trajectories and Admissible Controls The costate equation and its solution are H x = λ(t) =λ(t) λ(t) =λ()e t (8) There are five possible solution trajectories for λ(t) that affect the control signal u(t). These different trajectories depend on the initial condition λ() and its relation to and ±. The trajectories are shown in Fig., where Case corresponds to λ() >, Case corresponds to <λ() <, Case 3 corresponds to <λ() <, Case 4 corresponds to λ() <, and Case 5 corresponds to λ() =. The time indicated in the figure as t is when the trajectories for Case and Case 3 cross ±. Since the solution to the costate differential equation is a single growing exponential, λ(t) never decreases and is always increasing if λ() 6=.If < λ(), the initial control will be u() =. As t increases, the costate variable will be F.L. Lewis and V.L. Syrmos, Optimal Control, nd Ed., Wiley-Interscience, 995, Chap. 5. (7)

2 4 3 Case Case Costate λ(t) t = t Case 5 Case 3 3 Case Fig.. The 5 possible costate trajectories that affect the optimal control. λ(t) =at t = t. At that time, the control signal will switch to u (t ) =. It will remain at that value for t <t T. If λ() >, then u (t) =for t T. Therefore, if u () =, it will never switch to u(t) =. Control sequences of the form [+, ], [, ], [+,, ], [,, +], etc. are excluded from being admissible controls. The only admissible control sequences for this system are shown in Table. The Minimum Principle effectively filters out all other control sequences and provides us with the only admissible ones, regardless of the initial condition or the target state. Other control sequences could obviously be applied to the system, but they would not represent minimum fuel solutions. Table : Admissible control sequences for minimum fuel control. Case u (t) Comments [ ] No switching [, ] Switch at t = t 3 [, +] Switch at t = t 4 [+] No switching 5 [] No switching At the final time, t = T, the minimum fuel problem ends. The control that is applied for t>t depends on the performance requirements that are imposed for that time period. If it is desired to maintain the state at the target state, so that =x T for t T, then for the system model used in this example, the equilibrium control is u(t) =x T for t>t.this control will make =. The control used to maintain this equilibrium condition is not included in the minimum fuel performance index since J only includes time in the interval t T. C. State Trajectories The state trajectories can be found by substituting each of the allowed values for the control signal u(t) into the state equation, and then solving the state equation for. This will determine what target states can be reached and whether or

3 3 not the reachable target states depend on the initial condition. For the general linear, time-invariant system described by the solution to the differential equation is =A+Bu(t) (9) Z t x(t )=e A(t t ) x (t )+ e A(t τ) Bu(τ)dτ () t For the system model used in this example (A =,B =), and the allowed control values, the state trajectories are given by the following expressions. For u(t) = = x(t )=x (t ) e (t t ) which indicates that decays toward the origin x =in accordance with its natural dynamics given by the open-loop eigenvalue at s =. For u(t) =+ () Z t = + x(t )=x(t ) e (t t) + t e (t τ) dτ () x (t ) = e (t t ) [x (t ) ] + showing that the state asymptotically approaches =at (t t ), regardless of the initial condition x. For u(t) = Z t = x(t )=x(t ) e (t t ) t e (t τ) dτ (3) x (t ) = e (t t) [x (t )+] showing that with this input signal the state asymptotically approaches = at (t t ). Figure shows typical state trajectories for the three values of u(t). From the trajectories that are shown, the following observations can be made: If the specified final time T is finite, then u(t) =cannot be the only control value used if the target state is x(t )=x T = or if the target state has the opposite sign of the initial state x, that is, if x x T <. The target state has to be in the open interval <x T < in order for the target state to be reached in a finite time from any initial condition. If the target state were outside that interval, x T, there would be some initial conditions that could not reach the target state at all. For example, no state trajectory with initial condition x > can ever reach any target state x T <, and it can only reach x T = in infinite time. On the other hand, an initial condition x > + can reach a target state x >x T > + in a finite time with the optimal control being either u =or u =[, ], depending on the value of T. The control u =+would not be used since it has the same fuel cost as u = and yet takes longer to drive the state toward its target value. However, x T does limit valid initial conditions. For u(t) =+or u(t) =, the state trajectory tends toward =u(t) as t, asymptotically approaching that value. Therefore, the state trajectory cannot cross =+from either direction when u(t) =+. A similar statement holds for = with u(t) =. Any target state x T > can be reached in a finite time from any initial state x >x T with the control u(t) =. Therefore, the possible minimum fuel control sequences for this scenario would be u(t) = or u(t) =[, ] depending on the value of T. Likewise, any target state x T < + can be reached in a finite time from any initial state x <x T with the control u(t) =+. Therefore, the possible minimum fuel control sequences in this case would be u(t) =+or u(t) =[, +], again depending on the value of T. Example : Let the system model be given by (), and let the target state and initial condition be x(t ) = x T =, x() = x =5. If only u(t) = is used, then the target state can be reached, and it will be reached in minimum time. J.G. Reid, Linear System Fundamentals, 983, McGraw Hill, NY, Chap. 7.

4 4 Control Signal is u(t) = State Control Signal is u(t) = + State Control Signal is u(t) = State Fig.. State trajectories for u(t) =, u(t) =+, and u(t) =.

5 5 The performance index is J = R T dt = T, and the value of T necessary to reach x T from x with u = is T = Ln [ + 5] =.79 seconds. If we specify that u(t) =[, ] and specify the switching time to be t =seconds, then Z Z T J = dt + dt = T (4) and T =+Ln +5e =.57 seconds, so J =.57. This is less than the minimum time solution using only u(t) =. If we let the switching time be t =3seconds, then J 3 = Z 3 dt + Z T 3 dt = T 3 (5) and T =3+Ln +5e 3 =3. seconds, so J 3 =.. This is less than either of the two previous solutions. This implies that the longer we allow the system to take in order to reach the target state, the lower the cost in terms of fuel to reach the target state. In this example J 3 <J <J. This is generally the case. However there is a scenario in which letting the system take longer to reach the target state actually increases the cost as well as the time. This scenario will be presented later, in Examples 8 and 9. In each of the cases in the example above, the value of the performance index J for this system is the length of time that a non-zero control is being used. This is a general result for single-input systems. Therefore, Z T Z t Z T J = u(t) dt = dt + ± dt (6) t J = T t (7) If T = T Min, the minimum time solution, then the switching time is t =, and J = T = T Min. For T À T Min, the performance index approaches a limiting value. If the control law needed to drive x to x T includes u(t) =+or u(t) =, the limiting value of J depends only on the absolute value of the target state x T. The limiting value of J depends neither on the sign of x T nor on the value of the control signal used to drive toward x T. In this case the limiting value of the performance index is J Lim =max[ Ln ( + x T ), Ln ( x T )] (8) The validity of (8) is seen in the next section in Eqn. (4) and the discussion following. If x T =, or if u(t) =will allow the state to decay from x to x T in a finite time, then J Lim =;otherwise, J Lim >. In order for J to actually represent a quantity of fuel that is being minimized, the control signal u(t) must have the units of fuel used per unit time. The integration over time then yields a value for J having units of total fuel consumed. In these notes, values for J will be given in each example without any associated units. D. Determining the Switching Time t and the Minimum Time T Min Assume that the target state is <x T and the initial condition is x > so that the control u(t) =will not force the system to reach the target state in a finite time. Therefore, either u(t) = or u(t) =[, ] must be used, depending on the value of T. Assuming that T>T Min so that the control sequence will be u(t) =[, ], let t = t be the time at which the control switches from u(t) =to u(t) =. The initial time will be taken as t =. From () and (3) the state trajectory will be x(t ) = x e t Z T x(t ) = x T = x (t ) e (T t) e (T τ) dτ = x e T e T e T e t () t x T = x e T +e (T t) = x e T +e T e t () e t = e T (x T +) x or e (T t) = x T + x e T () The first expression in () can be solved for the relationship between t,x,andx T. t = Ln e T (x T +) x An alternate equation for t can be developed by solving the second expression in () for T t. T t = J = Ln x T + x e T (4) (9) (3)

6 6 If T À, then the exponential on the right side of (4) will be negligible, and T t Ln [x T +]. If the target state is the origin, then T t for large T. This means that u(t) =is used for most of the time, so the system coasts for as long as possible. Since J = T t, for the control sequence u(t) =[, ], the cost can be made arbitrarily small (for x T =)bylettingthefinal time be arbitrarily large. However, there is no solution to the optimal control problem if the target state is the origin and T is free. If x T 6=, then the limiting value of the performance index will be positive, with its value depending only on x T. If the final time is fixed, then (4) can be solved for the switching time t t = T + Ln x T + x e T (5) The final time has to be fixed at a value greater than or equal to the minimum time T Min in order for there to be any solution to the minimum fuel control problem, and T>T Min is necessary for t >. The time optimal solution can be obtained from (9) and () by setting t =, and it is given by x T = x e T e T e T = e T [x +] (6) e T [x T +] = [x +] (7) T x + = T Min = Ln (8) x T + The expressions developed above for the performance index and minimum time are also valid for <x T <x and when <x T <x. The governing rule for this condition is that (x +)(x T +) > and (x +) > (x T +) and x T >. When the third condition, based on requiring the target state to be in the open interval <x T <, is included, this simplifies to x >x T. Likewise, the expressions for the switching time are valid as long as the relationship between x, x T, and T are such that switching occurs. Otherwise, t =. The expressions are summarized in Table. Table : Summary of equations for u = or u =[, ]. Characteristic Equation Eqn. No. Switching Time t = Ln e T (x T +) x 3 Performance Index T t = J = Ln x T + x e T 4 Switching Time (alternate) t = T + Ln x T + x e T 5 Minimum Time T x + = T Min = Ln 8 x T + Similar expressions can be developed for the case where x < and x T <, which requires u =+. The general rule for this case is (x ) (x T ) > and (x ) > (x T ) and x T <, which simplifies to x <x T when the restriction on the target state is imposed. The expressions are developed in the same way and are given in Table 3. Table 3: Summary of equations for u =+or u =[, +]. Characteristic Equation Switching Time t = Ln e T (x T ) + x Performance Index T t = J = Ln x T ++x e T Switching Time (alternate) t = T + Ln x T ++x e T Minimum Time T x = T Min = Ln x T Example : Using the same initial condition x =5and target state x T =as in Example, Table 4 shows the switching time and total cost for several different values of T, including the minimum time T =.79. Table 4: Switching time and total cost as a function of final time T. T t J

7 7 Relationship Between Final Time and Interval of Non Zero Control; x = 5, x T =, T min =.79 sec J = T t.8 J =.749 for T =.5 sec.6.4. Limit on J = Final Time T (sec) Fig. 3. Plot of performance index J vs. final time T. The entries in the table clearly show the reduction in J as T increases. For large T, t T and J. Figure 3 graphically shows the relationship between the specified final time T and the performance index J = T t. The vertical dashed line in the figure is at T = T Min =.79 seconds. The graph clearly shows that when x T =, the limiting value of the performance index is J Lim =, and that J decreases fairly rapidly with increasing T above T Min. For a specific value of T, the corresponding value of J can be obtained from the graph, so the cost of operating the system from x can be determined in advance. The value of J =.749 for T =.5 seconds is shown for illustration. E. The Switching Curve and Closed-Loop Control Once the target state x T, the initial condition x, and the final time T are known, the value of the switching time t can be computed from the appropriate entry in Table or Table 3. The system can then be operated with u(t) =for t t and u(t) = (or +) for t <t T. The only requirement is to keep track of time. This is open-loop control since the current value of the state is not used to determine the control signal. It would be preferable to use closed-loop control for the usual reasons of reducing the effects of uncertainty and disturbances. This can be done by comparing the current state with the value of a switching curve. Assume that the initial condition and target state satisfy the condition x >x T, requiring the non-zero portion of the minimum fuel control to be u(t) =. Then the switching curve is the state trajectory that reaches the target state x(t )=x T at the final time T using only the control u(t) =. Call this switching curve z(t). The expression for z(t) has the same form as (3), and is given by where the initial condition is z(t) =e t (z +) (9) z = e T (x T +) (3)

8 8 9 State Response and Switching Curve with T =.5 sec, Switching Time: t =.5 sec, J = Switching Curve z(t) with u(t) = t with u(t) = Target Reached & Held T Fig. 4. State trajectory and switching curve for x =5,x T =, and T =.5 sec.; t =.5 sec. and J =.749 sec. Substituting (3) into (9) and letting t = T yields the result that z(t )=x T. Note that the switching curve depends only on the target state x T and the final time T, not on the initial condition x. The corresponding equations for the condition x <x T, with u(t) =+, are z(t) = e t (z ) + (3) z = e T (x T ) + (3) The closed-loop control law is shown in Table 5, which can be simplified as shown in (33) and (34). Table 5: Closed-loop, minimum fuel control law. x >x T x <x T u(t) =, if <z(t) u(t) =, if >z(t) u(t) =, if z(t) u(t) =+, if z(t) u(t) =, if (x x T )[ z(t)] < (33) u(t) = SGN (x x T ), if (x x T )[ z(t)] (34) The minimum fuel control law allows this system which is open-loop stable to respond naturally for as long as possible with zero control. A non-zero control is applied only at the time necessary to force the state so that it arrives at the target state at the specified final time. Figure 4 shows the switching curve for T =.5 seconds and x T =. The initial condition for the state is x =5. The heavy line in the figure is the actual trajectory that the state variable would follow, reaching the origin at the specified value of T. The switching takes place at t =.5 seconds.

9 9 F. Illustrative Examples ) Overview of the Examples: Several examples will be presented in this section to illustrate the use of the various expressions developed in the previous sections and to point out one scenario that has a peculiar twist to it. These examples are certainly not exhaustive, but they do show the major aspects of minimum fuel control for this first-order stable system. Changing the signs of the initial condition x and the target state x T only changes the sign of the control signal u(t); there is no change in the fundamental properties of the control or state trajectories. The examples to be presented are listed below and summarized in Table 6. Ex. 3. x >, x T (, ),T>T Min Ex. 4. x >, x T (, ),T= T Min Ex. 5. x <, x T (, +),T>T Min Ex. 6. x >, x T (,x ),T Min <T <T u Ex. 7. x >, x T (,x ),T Min <T u T Ex. 8. x >, x T (x, +),T>T Min,T fixed in value Ex. 9. x >, x T (x, +),T>T Min,T set equal to T Min The parameter T u is the time required for the state to decay from x to x T only using the control u(t) =. This can only occur for the situation defined by x T 6=and x x T > and x > x T. If T u T, then u(t) =is the minimum fuel control, and T will be reduced to T u if the specified T is greater than T u. Examples 8 and 9 illustrate the scenario that has the peculiar twist that was previously mentioned. In Example, shown on page 6, it was seen that increasing the value of the final time T reduces the amount of fuel used to drive the state from x to x T. This characteristic will also be seen in Examples 3 7 in this section. For the conditions in Examples 8 and 9, however, increasing the final time actually increases the amount of fuel used. The true minimum fuel solution for these conditions is also the minimum time solution. Those two examples will illustrate the results when T is required to be fixed at its specified value and when it can be decreased to equal the optimal minimum time value. Table 6: Summary of examples. Example x x T T (sec) t (sec) J T Min (sec) J Lim u (t) [, ] [ ] [, +] [, ] N/A.336 [] [, +] [+] ) Description of the Examples: The examples will be described on the following pages. The important points of each example will be highlighted to illustrate how that example is similar to or different from the other examples. For each of the examples, three plots are provided. The upper left graph shows the state trajectory (blue line) and switching curve z(t) (red line). The upper right graph shows the control signal u(t). In each case, it is assumed that the goal is to maintain the state at the target value for t T. Therefore, the control signal takes on its equilibrium value u(t) =x T for t>t.the bottom graph shows the relationship between the value of the performance index J = T t and the final time T. In each of the examples, the relationships between the initial state, the final state, and the final time should be noted, as well as the relationship between the final time and the minimum time T Min, and the relationship between the state and the switching curve z(t).

10 Example 3: x =3,x T =.5, T=.5 sec. The graphs for this example are shown in Fig. 5. Since x > and x T <, the control signal must have the value u(t) = for at least part of the time since u(t) =cannot force to change sign. From (8), T Min = Ln [(x +)/ (x T +)]=.79 sec., so it is clear that T>T Min. This allows the control signal to be u(t) = for part of the time. From (3) the time at which the control switches from u(t) = to u(t) = is t = Ln e T (x T +) x =.9 sec. The state response is seen to decay naturally from t =to t = t with the applied control being u(t) =. At time t = t, the state trajectory intersects the switching curve trajectory, and the control switches from u(t) =to u(t) =. The switching curve and state reach the desired value x T =.5 at t = T =.5 sec. It is assumed that the goal is for the state to be held at that final value for t>t,so the control signal is switched to u(t) =x T =.5 at t = T. This makes = x T + x T =for t T. Since x T <, the control signal can take on this equilibrium value under the specified control constraints. The control signal is seen to be piecewise constant, switching values once at t = t and again at t = T in order to establish equilibrium at x = x T. The value of the performance index for this example is J = T t =.37. The bottom plot in Fig. 5 shows that relationship between the final time T and the value of J. The maximum cost would be incurred if T = T Min =.79 sec. In that case, the switching time would be t =(the control starts at u(t) = ), and the value of the performance index would be J = T t =.79 = T Min. The plot shows that increasing T decreases the value of J, but the performance index does reach a limiting value. Since x T > and u(t) 6= is required, the limiting value of J is non-zero. In accordance with (8), J Lim = Ln ( + x T )= State: T =.5 sec, t =.9 sec, J =.37.5 Control Signal u(t) x = 3, x T =.5, T min =.79 sec J = T t.5 J =.37 for T =.5 sec Limit on J = Final Time T (sec) Fig. 5. Results from Example 3: x =3,x T =.5, T=.5 sec.

11 Example 4: x =3,x T =.5, T=.79 sec. The graphs for this example are shown in Fig. 6. As in Example 3, the control signal must have the value u(t) = for at least part of the time since x > and x T < so that u(t) =cannot drive the system state to the target state. The minimum time depends only on the initial and target states, so it has the same value as before: T Min =.79 sec. The final time in this example is chosen to be equal to the minimum time, so the control signal cannot be u(t) = any time during the trajectory. Thus, the time at which the control switches from u(t) =to u(t) = is t = sec. (the control starts at u(t) = ). There is no natural decaying of the state in this situation. The switching curve has the initial condition z = x, so it and the state reach the desired value x T =.5 at t = T =.79 sec. As before, it is assumed that the goal is for the state to be held at that final value for t>t,so the control signal is switched to u(t) =x T =.5 at t = T. This makes =for t T. The value of the performance index for this example is J = T t =.79, which is obviously higher than in Example 3. Since T is the independent variable in the bottom plot of the figure cost vs. final time and the cost depends on T, x, and x T, the curve showing that relationship for this example is the same as in the previous example since x and x T are unchanged. As seen in the bottom plot, the maximum cost has been incurred since T = T Min =.79 sec. Since the limiting value of J depends only on x T (except when u(t) =is a valid control by itself), that limiting value is the same as in Example 3, namely, J Lim = Ln ( + x T )= State: T =.79 sec, t =. sec, J =.79.5 Control Signal.5 u(t) x = 3, x T =.5, T min =.79 sec J =.79 for T =.79 sec J = T t.5.5 Limit on J = Final Time T (sec) Fig. 6. Results from Example 4: x =3,x T =.5, T= T Min =.79 sec.

12 Example 5: x = 3, x T =.5, T=.5 sec. The graphs for this example are shown in Fig. 7. This example is just the mirror image of Example 3. The signs of the initial condition and target state are both reversed from the previous example. The values of switching time, minimum time, performance index, and limiting value for the performance index are unchanged, although the times are computed from different expressions because of the changes in signs of x and x T. The correct expressions for t,t Min, and J are obtained from Table 3, and the expression for J Lim is J Lim = Ln ( x T ). The control sequence is u(t) =[, +] rather than u(t) =[, ] as it was in Example 3. Reversing the signs on both x and x T has no effect except to change the sign on the non-zero portion of u(t). For any of these examples, the bottom plot can be used to determine the value of the final time T needed in order to achieve a particular value of the performance index J = J in the interval T Min J >J Lim. This can be done by substituting the desired value J into (4), or the corresponding expression in Table 3, and solving for the value of T. The resulting value of T is +xt e J T = Ln, x 6= (35) If x =, then J = J Lim =max[ Ln ( + x T ), Ln ( x T )] regardless of the value of T. x State: T =.5 sec, t =.9 sec, J =.37.5 Control Signal u(t) x = 3, x T =.5, T min =.79 sec J = T t.5 J =.37 for T =.5 sec Limit on J = Final Time T (sec) Fig. 7. Results from Example 5: x = 3, x T =.5, T=.5 sec.

13 3 Example 6: x =.75, x T =.5, T=.5 sec. The graphs for this example are shown in Fig. 8. This example and the following one illustrate the situation when the target state can be reached from the initial state using u(t) =only if T is sufficiently long. Since the open-loop system is stable, x will decay naturally to x T in a finite time if x > x T >. The time needed for this decay is T u = Ln (x /x T ).Ifthespecified final time T is greater than or equal to T u, then the minimum fuel control would be u(t) =, and the cost would be J =. If T<T u, then u(t) =+or u(t) = is required, depending on the sign of x. For the values given in this example, T u = Ln(3) =.99 sec., so with T =.5 sec. the control signal cannot be u(t) =by itself. The switching time is computed in the normal fashion, and the control signal switches from u(t) =to u(t) = at t = t. The top two graphs in Fig. 8 show the state trajectory, switching curve, and control signal. As with the previous examples, the cost decreases with increasing T, as seen in the bottom graph, with a limiting value of J =. Unlike the examples shown in Fig. 3 and Figs. 5 7, however, the limiting value is not reached asymptotically in this example. The graph of J vs. T in Fig. 8 appears to be approaching a negative value as T increases. The limiting value J Lim =is reached at T = T u instead of as T. This can be seen from J = Ln x T + x e T h = Ln J = Ln h x T + x e Ln(xT /x)i = Ln i x T + x e Ln(x /x T ) (36) x T + x xt (37) J = Ln [] = when T = T u (38) x For T>T u, the switching time t is greater than the final time. This would make the performance index negative, which is not a valid situation. Therefore, for T T u,t T u, no switching occurs, u(t) =, and J =. This is illustrated in the next example. State: T =.5 sec, t =.7 sec, J =.9.5 Control Signal u(t) x =.75, x T =.5, T min =.336 sec.4 J = T t.3. J =.9 for T =.5 sec. Limit on J = Final Time T (sec) Fig. 8. Results from Example 6: x =.75, x T =.5, T=.5 sec.

14 4 Example 7: x =.75, x T =.5, T=.99 sec. The graphs for this example are shown in Fig. 9. The initial state and target state are the same as in Example 6, but now the final time has been increased to T = T u =.99 sec. The system state can decay naturally from the initial state to the final state, so no control is necessary. The optimal minimum fuel control is u(t) =, and the corresponding cost is J =. The state and control trajectories are shown in the top two graphs. There is no switching curve shown in the figure since z(t) would correspond to the control signal u(t) =, and the only control used in this scenario is u(t) =. The state will not coincide with z(t) anywhere except at t = T, where they both have the value x(t )=z(t)=x T. If the final time is specified at a larger value, T>T u, then the state will still decay from x to x T in T u seconds. There is no way to increase that length of time and still implement a minimum fuel control law. Since u(t) =[+, ] and u(t) =[+,, ] are not valid control sequences for this system with the minimum fuel performance index, there is no way to move the state away from the target state in order for the trajectory to take a longer period of time. When the initial and target states are such that x > x T >, then T u is the maximum time that is acceptable for minimum fuel control. Therefore, T Min T T u is required in this situation. State: T = T u =.99 sec, J =..5 Control Signal u(t) x =.75, x T =.5, T min =.336 sec.4 J = T t.3. J =. for T = T u =.99 sec. Limit on J = Final Time T (sec) Fig. 9. Results from Example 7: x =.75, x T =.5, T= T u =.99 sec.

15 5 Example 8: x =.35, x T =.75, T=.75 sec. The graphs for this example are shown in Fig.. Examples 8 and 9 illustrate the scenario mentioned previously that has a peculiar twist to it, relative to the other scenarios studied. This scenario is characterized by the fact that the control signal u(t) =allows the system state to move in the direction away from the target state. The target state can still be reached using a control sequence of u(t) =[, +] or u(t) =[, ], depending on the sign of x. However, the true minimum fuel control for this scenario is also the minimum time control, u(t) =+or u(t) =. It actually takes less fuel to force the system to move from x to x T as quickly as possible rather than allowing the system to coast. The reason for this is that when the system is coasting with u(t) =, it is coasting in the wrong direction. When the control signal u(t) =is finally applied, the state is farther away from x T than it initially was, and the non-zero control has to be applied for a longer period of time. Thus, the cost goes up with increasing T. The relationship between x and x T to have this scenario is one of the following: > x T >x >, x > (39) < x T <x <, x < (4) In this example, the final time is set at T =.75 sec., which is longer than the minimum time T Min =.956 sec. However, this example forces the specified final time to be used for the simulation. Therefore, the initial control is u(t) =. The switching time t is computed in the normal fashion, and the control switches to u(t) =+at t = t. The state will reach the target state at the specified final time, as seen in the upper left graph of Fig.. The bottom graph in the figure shows the increase in cost with increasing T. This is opposite of what was seen in all the other examples, and this characteristic of the J vs. T curve is specific to this scenario. For the specified final time of T =.75 sec., the cost is J =.69. State: T =.75 sec, t =.58 sec, J =.69.5 Control Signal u(t) J = T t.5.5 J =.69 for T =.75 sec x =.35, x T =.75, T min =.956 sec Limit on J = Final Time T (sec) Fig.. Results from Example 8: x =.35, x T =.75, T=.75 sec.

16 6 Example 9: x =.35, x T =.75, T=.956 sec. The graphs for this example are shown in Fig.. The same values for initial and target states are used here as were used in Example 8. In this example, the final time T is allowed to be the minimum time value of T Min =.956 sec. The only control that is used is u(t) =+. The switching time is t =sec., since this is equivalent to a minimum time problem, so the cost is J = T t = T Min =.956. This cost is lower than the value achieved in the previous example, and the target state was reached more rapidly as well. This illustrates the paradox of this particular scenario a faster response time requires less fuel. With this particular scenario, > x T > x >, the benefits of the shortest possible response time and the least amount of fuel consumed are both achieved with the same control law, u(t) =SGN [x ]. The bottom graph in the figure illustrates that the minimum time solution is also the minimum fuel solution for this set of conditions. State: T =.956 sec, t =. sec, J = Control Signal u(t) x =.35, x T =.75, T min =.956 sec Limit on J =.386 J = T t.5 J =.956 for T =.956 sec Final Time T (sec) Fig.. Results from Example 9: x =.35, x T =.75, T= T Min =.956 sec.

### Section 14.1 Vector Functions and Space Curves

Section 14.1 Vector Functions and Space Curves Functions whose range does not consists of numbers A bulk of elementary mathematics involves the study of functions - rules that assign to a given input a

### Lecture 4 Continuous time linear quadratic regulator

EE363 Winter 2008-09 Lecture 4 Continuous time linear quadratic regulator continuous-time LQR problem dynamic programming solution Hamiltonian system and two point boundary value problem infinite horizon

### Final Exam Solutions

EE55: Linear Systems Final Exam SIST, ShanghaiTech Final Exam Solutions Course: Linear Systems Teacher: Prof. Boris Houska Duration: 85min YOUR NAME: (type in English letters) I Introduction This exam

### Solution via Laplace transform and matrix exponential

EE263 Autumn 2015 S. Boyd and S. Lall Solution via Laplace transform and matrix exponential Laplace transform solving ẋ = Ax via Laplace transform state transition matrix matrix exponential qualitative

### Pontryagin s Minimum Principle 1

ECE 680 Fall 2013 Pontryagin s Minimum Principle 1 In this handout, we provide a derivation of the minimum principle of Pontryagin, which is a generalization of the Euler-Lagrange equations that also includes

### Math 312 Lecture Notes Linear Two-dimensional Systems of Differential Equations

Math 2 Lecture Notes Linear Two-dimensional Systems of Differential Equations Warren Weckesser Department of Mathematics Colgate University February 2005 In these notes, we consider the linear system of

### Network Flows that Solve Linear Equations

Network Flows that Solve Linear Equations Guodong Shi, Brian D. O. Anderson and Uwe Helmke Abstract We study distributed network flows as solvers in continuous time for the linear algebraic equation arxiv:1510.05176v3

### 1. The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ = Ax + Bu is

ECE 55, Fall 2007 Problem Set #4 Solution The Transition Matrix (Hint: Recall that the solution to the linear equation ẋ Ax + Bu is x(t) e A(t ) x( ) + e A(t τ) Bu(τ)dτ () This formula is extremely important

### 1 Steady State Error (30 pts)

Professor Fearing EECS C28/ME C34 Problem Set Fall 2 Steady State Error (3 pts) Given the following continuous time (CT) system ] ẋ = A x + B u = x + 2 7 ] u(t), y = ] x () a) Given error e(t) = r(t) y(t)

### Optimal Control Theory - Module 3 - Maximum Principle

Optimal Control Theory - Module 3 - Maximum Principle Fall, 215 - University of Notre Dame 7.1 - Statement of Maximum Principle Consider the problem of minimizing J(u, t f ) = f t L(x, u)dt subject to

### 9 Controller Discretization

9 Controller Discretization In most applications, a control system is implemented in a digital fashion on a computer. This implies that the measurements that are supplied to the control system must be

### Intersecting Two Lines, Part One

Module 1.4 Page 118 of 1124. Module 1.4: Intersecting Two Lines, Part One This module will explain to you several common methods used for intersecting two lines. By this, we mean finding the point x, y)

### 0 t < 0 1 t 1. u(t) =

A. M. Niknejad University of California, Berkeley EE 100 / 42 Lecture 13 p. 22/33 Step Response A unit step function is described by u(t) = ( 0 t < 0 1 t 1 While the waveform has an artificial jump (difficult

### 10 Transfer Matrix Models

MIT EECS 6.241 (FALL 26) LECTURE NOTES BY A. MEGRETSKI 1 Transfer Matrix Models So far, transfer matrices were introduced for finite order state space LTI models, in which case they serve as an important

### LTI Systems (Continuous & Discrete) - Basics

LTI Systems (Continuous & Discrete) - Basics 1. A system with an input x(t) and output y(t) is described by the relation: y(t) = t. x(t). This system is (a) linear and time-invariant (b) linear and time-varying

### OPTIMAL CONTROL SYSTEMS

SYSTEMS MIN-MAX Harry G. Kwatny Department of Mechanical Engineering & Mechanics Drexel University OUTLINE MIN-MAX CONTROL Problem Definition HJB Equation Example GAME THEORY Differential Games Isaacs

### Section 6.1: Composite Functions

Section 6.1: Composite Functions Def: Given two function f and g, the composite function, which we denote by f g and read as f composed with g, is defined by (f g)(x) = f(g(x)). In other words, the function

### PID Control. Objectives

PID Control Objectives The objective of this lab is to study basic design issues for proportional-integral-derivative control laws. Emphasis is placed on transient responses and steady-state errors. The

### Exponential stability of families of linear delay systems

Exponential stability of families of linear delay systems F. Wirth Zentrum für Technomathematik Universität Bremen 28334 Bremen, Germany fabian@math.uni-bremen.de Keywords: Abstract Stability, delay systems,

### Aim: How do we prepare for AP Problems on limits, continuity and differentiability? f (x)

Name AP Calculus Date Supplemental Review 1 Aim: How do we prepare for AP Problems on limits, continuity and differentiability? Do Now: Use the graph of f(x) to evaluate each of the following: 1. lim x

### Deterministic Dynamic Programming

Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

### Tracking Control of a Class of Differential Inclusion Systems via Sliding Mode Technique

International Journal of Automation and Computing (3), June 24, 38-32 DOI: 7/s633-4-793-6 Tracking Control of a Class of Differential Inclusion Systems via Sliding Mode Technique Lei-Po Liu Zhu-Mu Fu Xiao-Na

### Automatic Control 2. Nonlinear systems. Prof. Alberto Bemporad. University of Trento. Academic year

Automatic Control 2 Nonlinear systems Prof. Alberto Bemporad University of Trento Academic year 2010-2011 Prof. Alberto Bemporad (University of Trento) Automatic Control 2 Academic year 2010-2011 1 / 18

### 3 Rigid Spacecraft Attitude Control

1 3 Rigid Spacecraft Attitude Control Consider a rigid spacecraft with body-fixed frame F b with origin O at the mass centre. Let ω denote the angular velocity of F b with respect to an inertial frame

### 2D-Volterra-Lotka Modeling For 2 Species

Majalat Al-Ulum Al-Insaniya wat - Tatbiqiya 2D-Volterra-Lotka Modeling For 2 Species Alhashmi Darah 1 University of Almergeb Department of Mathematics Faculty of Science Zliten Libya. Abstract The purpose

### Generalized Cost-Benefit-Analysis and Social Discounting with Intertemporally Dependent Preferences

Generalized Cost-Benefit-Analysis and Social Discounting with Intertemporally Dependent Preferences February 29 Christian Traeger, Larry Karp Department of Agricultural & Resource Economics, UC Berkeley

### 20D - Homework Assignment 5

Brian Bowers TA for Hui Sun MATH D Homework Assignment 5 November 8, 3 D - Homework Assignment 5 First, I present the list of all matrix row operations. We use combinations of these steps to row reduce

### Averaging II: Adiabatic Invariance for Integrable Systems (argued via the Averaging Principle)

Averaging II: Adiabatic Invariance for Integrable Systems (argued via the Averaging Principle In classical mechanics an adiabatic invariant is defined as follows[1]. Consider the Hamiltonian system with

### Math Fall Linear Filters

Math 658-6 Fall 212 Linear Filters 1. Convolutions and filters. A filter is a black box that takes an input signal, processes it, and then returns an output signal that in some way modifies the input.

### 12. Lines in R 3. through P 0 and is parallel to v is written parametrically as a function of t: Using vector notation, the same line is written

12. Lines in R 3 Given a point P 0 = (x 0, y 0, z 0 ) and a direction vector v 1 = a, b, c in R 3, a line L that passes through P 0 and is parallel to v is written parametrically as a function of t: x(t)

### Mat104 Fall 2002, Improper Integrals From Old Exams

Mat4 Fall 22, Improper Integrals From Old Eams For the following integrals, state whether they are convergent or divergent, and give your reasons. () (2) (3) (4) (5) converges. Break it up as 3 + 2 3 +

### Control Systems I. Lecture 4: Diagonalization, Modal Analysis, Intro to Feedback. Readings: Emilio Frazzoli

Control Systems I Lecture 4: Diagonalization, Modal Analysis, Intro to Feedback Readings: Emilio Frazzoli Institute for Dynamic Systems and Control D-MAVT ETH Zürich October 13, 2017 E. Frazzoli (ETH)

### Chapter 5 - Differentiating Functions

Chapter 5 - Differentiating Functions Section 5.1 - Differentiating Functions Differentiation is the process of finding the rate of change of a function. We have proven that if f is a variable dependent

### Limits, Continuity, and Differentiability Solutions

Limits, Continuity, and Differentiability Solutions We have intentionally included more material than can be covered in most Student Study Sessions to account for groups that are able to answer the questions

### Control Systems. Time response. L. Lanari

Control Systems Time response L. Lanari outline zero-state solution matrix exponential total response (sum of zero-state and zero-input responses) Dirac impulse impulse response change of coordinates (state)

### ECE 3620: Laplace Transforms: Chapter 3:

ECE 3620: Laplace Transforms: Chapter 3: 3.1-3.4 Prof. K. Chandra ECE, UMASS Lowell September 21, 2016 1 Analysis of LTI Systems in the Frequency Domain Thus far we have understood the relationship between

### Basic matrix operations

Roberto s Notes on Linear Algebra Chapter 4: Matrix algebra Section 3 Basic matrix operations What you need to know already: What a matrix is. he basic special types of matrices What you can learn here:

### FUNNEL CONTROL WITH ONLINEFORESIGHT

FUNNEL CONTROL WITH ONLINEFORESIGHT C.M. Hackl and Dierk Schröder Institute for Electrical Drive Systems Technical University of Munich Arcisstr., D-8333 Munich, Germany email: christoph.hackl@tum.de ABSTRACT

### Solve by factoring and applying the Zero Product Property. Review Solving Quadratic Equations. Three methods to solve equations of the

Topic 0: Review Solving Quadratic Equations Three methods to solve equations of the form ax 2 bx c 0. 1. Factoring the expression and applying the Zero Product Property 2. Completing the square and applying

### EE451/551: Digital Control. Chapter 8: Properties of State Space Models

EE451/551: Digital Control Chapter 8: Properties of State Space Models Equilibrium State Definition 8.1: An equilibrium point or state is an initial state from which the system nevers departs unless perturbed

### Systems Analysis and Control

Systems Analysis and Control Matthew M. Peet Arizona State University Lecture 8: Response Characteristics Overview In this Lecture, you will learn: Characteristics of the Response Stability Real Poles

### Practice Problems For Test 3

Practice Problems For Test 3 Power Series Preliminary Material. Find the interval of convergence of the following. Be sure to determine the convergence at the endpoints. (a) ( ) k (x ) k (x 3) k= k (b)

### Linear Filters. L[e iωt ] = 2π ĥ(ω)eiωt. Proof: Let L[e iωt ] = ẽ ω (t). Because L is time-invariant, we have that. L[e iω(t a) ] = ẽ ω (t a).

Linear Filters 1. Convolutions and filters. A filter is a black box that takes an input signal, processes it, and then returns an output signal that in some way modifies the input. For example, if the

### Written Homework 7 Solutions

Written Homework 7 Solutions Section 4.3 20. Find the local maxima and minima using the First and Second Derivative tests: Solution: First start by finding the first derivative. f (x) = x2 x 1 f (x) =

### Nonlinear and robust MPC with applications in robotics

Nonlinear and robust MPC with applications in robotics Boris Houska, Mario Villanueva, Benoît Chachuat ShanghaiTech, Texas A&M, Imperial College London 1 Overview Introduction to Robust MPC Min-Max Differential

### Chemical Process Dynamics and Control. Aisha Osman Mohamed Ahmed Department of Chemical Engineering Faculty of Engineering, Red Sea University

Chemical Process Dynamics and Control Aisha Osman Mohamed Ahmed Department of Chemical Engineering Faculty of Engineering, Red Sea University 1 Chapter 4 System Stability 2 Chapter Objectives End of this

### Pulsed Lasers Revised: 2/12/14 15: , Henry Zmuda Set 5a Pulsed Lasers

Pulsed Lasers Revised: 2/12/14 15:27 2014, Henry Zmuda Set 5a Pulsed Lasers 1 Laser Dynamics Puled Lasers More efficient pulsing schemes are based on turning the laser itself on and off by means of an

### Aspects of Continuous- and Discrete-Time Signals and Systems

Aspects of Continuous- and Discrete-Time Signals and Systems C.S. Ramalingam Department of Electrical Engineering IIT Madras C.S. Ramalingam (EE Dept., IIT Madras) Networks and Systems 1 / 45 Scaling the

### GRAPHIC REALIZATIONS OF SEQUENCES. Under the direction of Dr. John S. Caughman

GRAPHIC REALIZATIONS OF SEQUENCES JOSEPH RICHARDS Under the direction of Dr. John S. Caughman A Math 501 Project Submitted in partial fulfillment of the requirements for the degree of Master of Science

### Time optimal trajectories for bounded velocity differential drive vehicles

Time optimal trajectories for bounded velocity differential drive vehicles Devin J. Balkcom Matthew T. Mason June 3, 00 Abstract This paper presents the time optimal trajectories for differential drive

### Conservation laws and some applications to traffic flows

Conservation laws and some applications to traffic flows Khai T. Nguyen Department of Mathematics, Penn State University ktn2@psu.edu 46th Annual John H. Barrett Memorial Lectures May 16 18, 2016 Khai

### 1 Functions, Graphs and Limits

1 Functions, Graphs and Limits 1.1 The Cartesian Plane In this course we will be dealing a lot with the Cartesian plane (also called the xy-plane), so this section should serve as a review of it and its

### Infinite series, improper integrals, and Taylor series

Chapter 2 Infinite series, improper integrals, and Taylor series 2. Introduction to series In studying calculus, we have explored a variety of functions. Among the most basic are polynomials, i.e. functions

### ( ) and then eliminate t to get y(x) or x(y). Not that the

2. First-order linear equations We want the solution (x,y) of () ax, ( y) x + bx, ( y) y = 0 ( ) and b( x,y) are given functions. If a and b are constants, then =F(bx-ay) where ax, y where F is an arbitrary

### Sommerfeld-Drude model. Ground state of ideal electron gas

Sommerfeld-Drude model Recap of Drude model: 1. Treated electrons as free particles moving in a constant potential background. 2. Treated electrons as identical and distinguishable. 3. Applied classical

### Damped harmonic motion

Damped harmonic motion March 3, 016 Harmonic motion is studied in the presence of a damping force proportional to the velocity. The complex method is introduced, and the different cases of under-damping,

### Discrete-time linear systems

Automatic Control Discrete-time linear systems Prof. Alberto Bemporad University of Trento Academic year 2-2 Prof. Alberto Bemporad (University of Trento) Automatic Control Academic year 2-2 / 34 Introduction

### Review of Linear Time-Invariant Network Analysis

D1 APPENDIX D Review of Linear Time-Invariant Network Analysis Consider a network with input x(t) and output y(t) as shown in Figure D-1. If an input x 1 (t) produces an output y 1 (t), and an input x

### Lecture 9 - Increasing and Decreasing Functions, Extrema, and the First Derivative Test

Lecture 9 - Increasing and Decreasing Functions, Extrema, and the First Derivative Test 9.1 Increasing and Decreasing Functions One of our goals is to be able to solve max/min problems, especially economics

### Lagrange Multipliers

Optimization with Constraints As long as algebra and geometry have been separated, their progress have been slow and their uses limited; but when these two sciences have been united, they have lent each

### EEE 303 Notes: System properties

EEE 303 Notes: System properties Kostas Tsakalis January 27, 2000 1 Introduction The purpose of this note is to provide a brief background and some examples on the fundamental system properties. In particular,

### Analysis of Discrete-Time Systems

TU Berlin Discrete-Time Control Systems 1 Analysis of Discrete-Time Systems Overview Stability Sensitivity and Robustness Controllability, Reachability, Observability, and Detectabiliy TU Berlin Discrete-Time

### Order Preserving Properties of Vehicle Dynamics with Respect to the Driver s Input

Order Preserving Properties of Vehicle Dynamics with Respect to the Driver s Input Mojtaba Forghani and Domitilla Del Vecchio Massachusetts Institute of Technology September 19, 214 1 Introduction In this

### Chapter 6. Exercise Solutions. 1. From Eq. 6.1, density = mass / volume. So, mass = density volume. styrofoam: density = 75 kg/m 3 volume = 1 cm 3

Chapter 6 Exercise Solutions 1. From Eq. 6.1, density mass / volume So, mass density volume styrofoam: density 75 kg/m 3 volume 1 cm 3 A unit conversion factor will be needed in order to cancel m and cm.

### EECE251. Circuit Analysis I. Set 4: Capacitors, Inductors, and First-Order Linear Circuits

EECE25 Circuit Analysis I Set 4: Capacitors, Inductors, and First-Order Linear Circuits Shahriar Mirabbasi Department of Electrical and Computer Engineering University of British Columbia shahriar@ece.ubc.ca

### Analog Signals and Systems and their properties

Analog Signals and Systems and their properties Main Course Objective: Recall course objectives Understand the fundamentals of systems/signals interaction (know how systems can transform or filter signals)

### Calculus at Rutgers. Course descriptions

Calculus at Rutgers This edition of Jon Rogawski s text, Calculus Early Transcendentals, is intended for students to use in the three-semester calculus sequence Math 151/152/251 beginning with Math 151

### 3.9 Derivatives of Exponential and Logarithmic Functions

322 Chapter 3 Derivatives 3.9 Derivatives of Exponential and Logarithmic Functions Learning Objectives 3.9.1 Find the derivative of exponential functions. 3.9.2 Find the derivative of logarithmic functions.

### 2.004 Dynamics and Control II Spring 2008

MIT OpenCourseWare http://ocw.mit.edu 2.004 Dynamics and Control II Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 8 I * * Massachusetts

### Solve by factoring and applying the Zero Product Property. Review Solving Quadratic Equations. Three methods to solve equations of the

Hartfield College Algebra (Version 2015b - Thomas Hartfield) Unit ONE Page - 1 - of 26 Topic 0: Review Solving Quadratic Equations Three methods to solve equations of the form ax 2 bx c 0. 1. Factoring

### Chapter 2. Review of Mathematics. 2.1 Exponents

Chapter 2 Review of Mathematics In this chapter, we will briefly review some of the mathematical concepts used in this textbook. Knowing these concepts will make it much easier to understand the mathematical

### Part I Stochastic variables and Markov chains

Part I Stochastic variables and Markov chains Random variables describe the behaviour of a phenomenon independent of any specific sample space Distribution function (cdf, cumulative distribution function)

### Math 124A October 11, 2011

Math 14A October 11, 11 Viktor Grigoryan 6 Wave equation: solution In this lecture we will solve the wave equation on the entire real line x R. This corresponds to a string of infinite length. Although

### In Chapters 3 and 4 we introduced linear programming

SUPPLEMENT The Simplex Method CD3 In Chapters 3 and 4 we introduced linear programming and showed how models with two variables can be solved graphically. We relied on computer programs (WINQSB, Excel,

### ORDINARY DIFFERENTIAL EQUATIONS

Page 1 of 20 ORDINARY DIFFERENTIAL EQUATIONS Lecture 11 Fundamental Solution Sets for Linear Homogeneous Second-Order ODEs (Revised 22 March 2009 @ 09:16) Professor Stephen H Saperstone Department of Mathematical

### AP Calculus AB Worksheet - Differentiability

Name AP Calculus AB Worksheet - Differentiability MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. The figure shows the graph of a function. At the

Null Controllability and Stabilization of Linear Systems Subject to Asymmetric Actuator Saturation Tingshu Hu, Achilleas N. Pitsillides & Zongli Lin Department of Electrical Engineering, University of

### CONTROL OF DIGITAL SYSTEMS

AUTOMATIC CONTROL AND SYSTEM THEORY CONTROL OF DIGITAL SYSTEMS Gianluca Palli Dipartimento di Ingegneria dell Energia Elettrica e dell Informazione (DEI) Università di Bologna Email: gianluca.palli@unibo.it

### The Hurewicz Theorem

The Hurewicz Theorem April 5, 011 1 Introduction The fundamental group and homology groups both give extremely useful information, particularly about path-connected spaces. Both can be considered as functors,

### Fair Exponential Smoothing with Small Alpha

Fair Exponential Smoothing with Small Alpha (fessa) Juha Reunanen juha.reunanen@tomaattinen.com October 12, 2015 Abstract This short paper shows that standard exponential smoothing with a small α may not

### Chapter 2. Motion along a straight line

Chapter 2 Motion along a straight line Introduction: Study of the motion of objects Physics studies: Properties of matter and energy: solid state physics, thermal physics/ thermodynamics, atomic physics,

### Learn2Control Laboratory

Learn2Control Laboratory Version 3.2 Summer Term 2014 1 This Script is for use in the scope of the Process Control lab. It is in no way claimed to be in any scientific way complete or unique. Errors should

### AP Calculus Testbank (Chapter 6) (Mr. Surowski)

AP Calculus Testbank (Chapter 6) (Mr. Surowski) Part I. Multiple-Choice Questions 1. Suppose that f is an odd differentiable function. Then (A) f(1); (B) f (1) (C) f(1) f( 1) (D) 0 (E). 1 1 xf (x) =. The

### Fluctuations of Time Averaged Quantum Fields

Fluctuations of Time Averaged Quantum Fields IOP Academia Sinica Larry Ford Tufts University January 17, 2015 Time averages of linear quantum fields Let E (x, t) be one component of the electric field

### Order of Operations. Real numbers

Order of Operations When simplifying algebraic expressions we use the following order: 1. Perform operations within a parenthesis. 2. Evaluate exponents. 3. Multiply and divide from left to right. 4. Add

### Consensus Protocols for Networks of Dynamic Agents

Consensus Protocols for Networks of Dynamic Agents Reza Olfati Saber Richard M. Murray Control and Dynamical Systems California Institute of Technology Pasadena, CA 91125 e-mail: {olfati,murray}@cds.caltech.edu

### The Relation Between the 3-D Bode Diagram and the Root Locus. Insights into the connection between these classical methods. By Panagiotis Tsiotras

F E A T U R E The Relation Between the -D Bode Diagram and the Root Locus Insights into the connection between these classical methods Bode diagrams and root locus plots have been the cornerstone of control

### e st f (t) dt = e st tf(t) dt = L {t f(t)} s

Additional operational properties How to find the Laplace transform of a function f (t) that is multiplied by a monomial t n, the transform of a special type of integral, and the transform of a periodic

### Balancing of Lossless and Passive Systems

Balancing of Lossless and Passive Systems Arjan van der Schaft Abstract Different balancing techniques are applied to lossless nonlinear systems, with open-loop balancing applied to their scattering representation.

### Systems of Linear Equations and Inequalities

Systems of Linear Equations and Inequalities Alex Moore February 4, 017 1 What is a system? Now that we have studied linear equations and linear inequalities, it is time to consider the question, What

### 9709 MATHEMATICS. 9709/31 Paper 3, maximum raw mark 75

CAMBRIDGE INTERNATIONAL EXAMINATIONS GCE Advanced Level MARK SCHEME for the May/June 0 series 9709 MATHEMATICS 9709/ Paper, maximum raw mark 75 This mark scheme is published as an aid to teachers and candidates,

### Relativity. An explanation of Brownian motion in terms of atoms. An explanation of the photoelectric effect ==> Quantum Theory

Relativity Relativity In 1905 Albert Einstein published five articles in Annalen Der Physik that had a major effect upon our understanding of physics. They included:- An explanation of Brownian motion

### Invariant Manifolds of Dynamical Systems and an application to Space Exploration

Invariant Manifolds of Dynamical Systems and an application to Space Exploration Mateo Wirth January 13, 2014 1 Abstract In this paper we go over the basics of stable and unstable manifolds associated

### Direct-Current Circuits. Physics 231 Lecture 6-1

Direct-Current Circuits Physics 231 Lecture 6-1 esistors in Series and Parallel As with capacitors, resistors are often in series and parallel configurations in circuits Series Parallel The question then

### Übersetzungshilfe / Translation aid (English) To be returned at the end of the exam!

Prüfung Regelungstechnik I (Control Systems I) Prof. Dr. Lino Guzzella 3. 8. 24 Übersetzungshilfe / Translation aid (English) To be returned at the end of the exam! Do not mark up this translation aid

### Integral action in state feedback control

Automatic Control 1 in state feedback control Prof. Alberto Bemporad University of Trento Academic year 21-211 Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 21-211 1 /

### 18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

806 Problem Set 8 - Solutions Due Wednesday, 4 November 2007 at 4 pm in 2-06 08 03 Problem : 205+5+5+5 Consider the matrix A 02 07 a Check that A is a positive Markov matrix, and find its steady state

### A sub-optimal second order sliding mode controller for systems with saturating actuators

28 American Control Conference Westin Seattle Hotel, Seattle, Washington, USA June -3, 28 FrB2.5 A sub-optimal second order sliding mode for systems with saturating actuators Antonella Ferrara and Matteo