Professional Documents
Culture Documents
x(t)
= Ax(t) B[Cx(t), t]
ABSOLUTE STABILITY
is exponentially stable.
Analysis of dynamical systems and circuits is mostly done under the assumption that the system is linear and time-invariant. Powerful mathematical techniques are then available for
analysis of stability and performance of the system, for example, superposition and frequency domain analysis. In fact,
even if the system is nonlinear and time-varying, such assumptions can often be used to get a first estimate of the system properties.
The purpose of absolute stability theory is to carry the
analysis one step further and get a bound on the possible influence of the nonlinear or time-varying components. The approach was suggested by the Russian mathematician Lure in
the 1940s and has, since then, developed into a cornerstone
of nonlinear systems theory.
The basic setup is illustrated in Fig. 1, where the linear
time-invariant part is represented by the transfer function
G(s) and the nonlinear parts of the system are represented by
a feedback loop w (v, t). The analysis of the system is
based on conic bounds on the nonlinearity (Fig. 2):
(v, t)/v
for all v = 0
The name of this result comes from its graphical interpretation. The Nyquist plot, that is, the plot of G(i) in the complex plane as R, must not cross or circumscribe the circle
centered on the real axis and passing through 1/ and 1/
(Fig. 3).
An important aspect of the circle criterion is that it demonstrates how frequency domain properties can be used in a
nonlinear setting. It is instructive to compare with the Nyquist criterion, which states that the closed-loop system with
linear feedback w(t) kv(t) is stable for all k [, ], provided that G(i) does not intersect the real axis outside the
interval [1/ , 1/]. The circle criterion replaces the interval condition with a circular disk. As a consequence, the stability assertion is extended from constant feedback to nonlinear and time-varying feedback.
The proof of the circle criterion is based on a quadratic
Lyapunov function of the form
V (x) = x Px
(1)
x = Ax Bw
x(t)
= Ax(t) B[Cx(t)]
(6)
G(i) + 1
0 < Re
G(i) + 1
(5)
(2)
v = Cx
for all v = 0
(4)
(7)
is exponentially stable.
Note that the circle criterion is recovered with 0. Also the
Popov criterion can be illustrated graphically. Introduce the
Popov plot, where ImG(i) is plotted versus ReG(i). Then
(3)
1
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
ABSOLUTE STABILITY
15
(, t)
10
v
5
Im
G(s)
1/
1/
G(i)
10
15
10
0
Re
10
(a)
15
10
x2 = x1 2x2 + 21(x1 )
s + 82
G(s) = C(i A)1 B = 2
s + 7s + 6
Im
let
1/
1/
The plot in Fig. 4 then shows that the Popov criterion is satisfied for 1.
The theory of absolute stability had a strong development
in the 1960s, and various improvements to the circle and Popov criteria were generated, for example, by Yakubovich.
Many types of nonlinearities were considered and stronger
criteria were obtained in several special cases (24). Important aspects of the theory were summarized by Willems (5),
using the notions of dissipativity and storage function.
G(i)
10
15
10
0
Re
10
(b)
Figure 3. The circle criterion proves stability as long as the Nyquist
plot does not cross or circumscribe the circle corresponding to the
conic bounds on the nonlinearity. (a) 0 . (b) 0 .
2
0
1/
w
w= (v)
Im G(i)
2
4
6
8
10
12
5
v
Figure 2. The nonlinearity is bounded by linear functions.
5
Re G(i)
10
15
ABSOLUTE STABILITY
T
0
w(t)[v(t) w(t)] dt
0
|u|2 dt
0
(8)
(w + h w)(v w) dt
(9)
Using this inequality, the Popov criterion [Eq. (6)] can be replaced by the condition
Re[(1 + i + H(i))(G(i) + 1)]] > 0
v = Gw + f
wR
(10)
w = Fv + e
define a bounded causal map from the inputs (e, f) to the outputs (v, w).
It is worthwhile to make a comparison with the circle criterion. Consider the case when . Then the condition [Eq.
(3)] becomes
[0, ]
|G(i)| < 1
|Fu|2 dt C2
u(t)y(t) dt
Ax + B
x = Ax Bsat(Cx) = Ax + BCx
Ax B
if Cx 1
if |Cx| < 1
if Cx 1
BIBLIOGRAPHY
1. V. M. Popov, Absolute stability of nonlinear systems of automatic
control, Autom. Remote Control, 22: 857875, 1962. (Original in
Russian, August, 1961.)
2. V. A. Yakubovich, Absolute stability of nonlinear control systems
in critical cases, parts 13, Avtomaika i Telemechanika, 24 (3):
293302; 24 (6): 717731, 1963; 25 (25): 601612, 1964. (English
translation in Autom. Remote Control.)
3. V. A. Yakubovich, On an abstract theory of absolute stability of
nonlinear systems, Vestnik Leningrad Univ. Math., 10: 341361,
1982. (Original in Russian, 1977.)
4. S. Lefschetz, Stability of Nonlinear Control Systems, New York:
Academic Press, 1963.
5. J. C. Willems, Dissipative dynamical systems, part 1, General theory; part 2, Linear systems with quadratic supply rates, Arch. Rational Mech. Anal., 45 (5): 321393, 1972.
6. G. Zames, On the inputoutput stability of nonlinear time-varying
feedback systemspart 1, Conditions derived using concepts of
loop gain; part 2, Conditions involving circles in the frequency
plane and sector nonlinearities, IEEE Trans. Autom. Control, 11:
228238, 1966.
ANDERS RANTZER
Lund Institute of Technology
ADAPTIVE CONTROL
According to Websters dictionary, to adapt means to change (oneself) so that ones behavior will conform to
new or changed circumstances. The words adaptive systems and adaptive control were used as early as
1950 (1). The design of autopilots for high-performance aircraft was one of the primary motivations for active
research on adaptive control in the early 1950s. Aircraft operate over a wide range of speeds and altitudes,
and their dynamics are nonlinear and conceptually time-varying. For a given operating point, specified by the
aircraft speed (Mach number) and altitude, the complex aircraft dynamics can be approximated by a linear
model of the form (2)
where Ai , Bi , Ci , and Di are functions of the operating point i. As the aircraft goes through different flight
conditions, the operating point changes, leading to different values for Ai , Bi , Ci , and Di . Because the output
response y(t) carries information about the state x as well as the parameters, one may argue that in principle,
a sophisticated feedback controller should be able to learn about parameter changes by processing y(t) and
use the appropriate gains to accommodate them. This argument led to a feedback control structure on which
adaptive control is based. The controller structure consists of a feedback loop and a controller with adjustable
gains as shown in Fig. 1. The way the controller parameters are adjusted on line leads to different adaptive
control schemes.
Gain Scheduling
The approach of gain scheduling is illustrated in Fig. 2. The gain scheduler consists of a lookup table
and the appropriate logic for detecting the operating point and choosing the corresponding value of the control
parameter vector . For example, let us consider the aircraft model in Eq. (1) where for each operating point i,
i = 1, 2, . . ., N, the parameters Ai , Bi , Ci , and Di are known. For each operating point i, a feedback controller
with constant gains, say i , is designed to meet the performance requirements for the corresponding linear
model. This leads to a controller, say C(), with a set of gains {1 , 2 , . . ., i , . . ., N } covering N operating points.
Once the operating point, say i, is detected, the controller gains can be changed to the appropriate value of i
obtained from the precomputed gain set. Transitions between different operating points that lead to significant
parameter changes may be handled by interpolation or by increasing the number of operating points. The two
elements that are essential in implementing this approach are a lookup table to store the values of i and the
plant auxiliary measurements that correlate well with changes in the operating points.
Direct and Indirect Adaptive Control
A wide class of adaptive controllers is formed by combining an on-line parameter estimator, which provides
estimates of unknown parameters at each time instant, with a control law that is motivated from the knownparameter case. The way the parameter estimator, also referred to as the adaptive law, is combined with the
control law gives rise to two different approaches. In the first approach, referred to as indirect adaptive control
1
ADAPTIVE CONTROL
(shown in Fig. 3), the plant parameters are estimated online and used to calculate the controller parameters.
This approach has also been referred to as explicit adaptive control, because the design is based on an explicit
plant model.
In the second approach, referred to as direct adaptive control (shown in Fig. 4), the plant model is
parametrized in terms of the controller parameters, which are estimated directly without intermediate calculations involving plant parameter estimates. This approach has also been referred to as implicit adaptive
control, because the design is based on the estimation of an implicit plant model.
The principle behind the design of direct and indirect adaptive control shown in Figs. 3 and 4 is conceptually simple. The design of C(c ) treats the estimates c (t) (in the case of direct adaptive control) or the estimates
(t) (in the case of indirect adaptive control) as if they were the true parameters. This design approach is
called certainty equivalence and can be used to generate a wide class of adaptive control schemes by combining
different on-line parameter estimators with different control laws.
The idea behind the certainty equivalence approach is that as the parameter estimates c (t) and (t)
converge to the true ones c and , respectively, the performance of the adaptive controller C(c ) tends to that
achieved by C( c ) in the case of known parameters.
Model Reference Adaptive Control
Model reference adaptive control (MRAC) is derived from the model-following problem or model reference
control (MRC) problem. The structure of an MRC scheme for a LTI, single-input single-output (SISO) plant is
shown in Fig. 5. The transfer function W m (s) of the reference model is designed so that for a given reference
input signal r(t) the output ym (t) of the reference model represents the desired response the plant output y(t)
should follow. The feedback controller, denoted by C( c ), is designed so that all signals are bounded and the
closed-loop plant transfer function from r to y is equal to W m (s). This transfer function matching guarantees
that for any given reference input r(t), the tracking error e1 y ym , which represents the deviation of the
ADAPTIVE CONTROL
plant output from the desired trajectory ym , converges to zero with time. The transfer function matching is
achieved by canceling the zeros of the plant transfer function G(s) and replacing them with those of W m (s)
through the use of the feedback controller C( c ). The cancellation of the plant zeros puts a restriction on the
plant to be minimum-phase, that is, have stable zeros. If any plant zero is unstable, its cancellation may easily
lead to unbounded signals.
The design of C( c ) requires the knowledge of the coefficients of the plant transfer G(s). When is
unknown we use the certainty equivalence approach to replace the unknown c in the control law with its
estimate c (t) obtained using the direct or the indirect approach. The resulting control schemes are known
as MRAC and can be classified as indirect MRAC of the structure shown in Fig. 3 and direct MRAC of the
structure shown in Fig. 4.
ADAPTIVE CONTROL
where is a vector with the coefficients of the plant transfer function G(s).
As in the case of MRC, we can deal with the unknown-parameter case by using the certainty equivalence
approach to replace the unknown vector c with its estimate c (t). The resulting scheme is referred to as
adaptive pole placement control (APPC). If c (t) is updated directly using an on-line parameter estimator, the
scheme is referred to as direct APPC. If c (t) is calculated using the equation
where (t) is the estimate of generated by an on-line estimator, the scheme is referred to as indirect APPC.
The structure of direct and indirect APPC is the same as that shown in Figs. 3 and 4, respectively, for the
general case. The design of APPC schemes is very flexible with respect to the choice of the form of the controller
C(c ) and of the on-line parameter estimator.
Design of On-Line Parameter Estimators
As we mentioned in the previous sections, an adaptive controller may be considered as a combination of
an on-line parameter estimator with a control law that is derived from the known-parameter case. The way this
combination occurs and the type of estimator and control law used give rise to a wide class of different adaptive
controllers with different properties. In the literature of adaptive control the on-line parameter estimator has
often been referred to as the adaptive law, update law, or adjustment mechanism.
Some of the basic methods used to design adaptive laws are
(1) Sensitivity methods
(2) Positivity and Lyapunov design
(3) Gradient method and least-squares methods based on estimation error cost criteria
The sensitivity method is one of the oldest methods used in the design of adaptive laws and is briefly
explained below.
ADAPTIVE CONTROL
Sensitivity Method. This method became very popular in the 1960s (3), and it is still used in many
industrial applications for controlling plants with uncertainties. In adaptive control, the sensitivity method is
used to design the adaptive law so that the estimated parameters are adjusted in a direction that minimizes a
certain performance function. The adaptive law is driven by the partial derivative of the performance function
with respect to the estimated parameters multiplied by an error signal that characterizes the mismatch between
the actual and desired behavior. This derivative is called the sensitivity function, and if it can be generated
online, then the adaptive law is implementable. In most earlier formulations of adaptive control, the sensitivity
function cannot be generated online, and this constitutes one of the main drawbacks of the method. The use
of approximate sensitivity functions that are implementable leads to adaptive control schemes whose stability
properties are either weak or cannot be established.
Positivity and Lyapunov Design. This method of developing adaptive laws is based on the direct
method of Lyapunov and its relationship to positive real functions. In this approach, the problem of designing
an adaptive law is formulated as a stability problem where the differential equation of the adaptive law is
chosen so that certain stability conditions based on Lyapunov theory are satisfied. The adaptive law developed
is very similar to that based on the sensitivity method. The only difference is that the sensitivity functions in
the approach are replaced with ones that can be generated online. In addition, the Lyapunov-based adaptive
control schemes have none of the drawbacks of the MIT rule-based schemes.
The design of adaptive laws using Lyapunovs direct method was suggested by Grayson (4) and Parks
(5) in the early 1960s. The method was subsequently advanced and generalized to a wider class of plants by
Phillipson (6), Monopoli (7), and others (8,9,10,11).
Gradient and Least-Squares Methods Based on Estimation Error Cost Criteria. The main drawback of the sensitivity methods used in the 1960s is that the minimization of the performance cost function led
to sensitivity functions that are not implementable. One way to avoid this drawback is to choose a cost function
criterion that leads to sensitivity functions that are available for measurement. A class of such cost criteria is
based on an error, referred to as the estimation error (12), that provides a measure of the discrepancy between
the estimated and actual parameters. The relationship of the estimation error with the estimated parameters
is chosen so that the cost function is convex, and its gradient with respect to the estimated parameters is implementable. Several different cost criteria may be used, and methods such as the gradient and least-squares
may be adopted to generate the appropriate sensitivity functions.
ADAPTIVE CONTROL
where x Rn and only y, u are available for measurement. Note that the plant equation can also be written as
an nth-order differential equation
where
We can avoid the use of differentiators by filtering with our nth-order stable filter 1/(s) to obtain
where
where W(s) is an appropriate proper stable transfer function and , are defined similarly to , .
In Eqs. (5), (6) the unknown vectors , appear linearly in equations where all the other terms are
known a priori or can be measured online. We use these parametric models to estimate the unknown vectors
or by using the following approaches.
SPRLyapunov Design Approach. We start by rewriting Eq. (6) as follows (for simplicity we drop
the subscript ):
where = L 1 (s)and L(s) is chosen so that L 1 (s) is a proper transfer function and W(s)L(s) is a proper strictly
positive real (SPR) transfer function. Let (t) denote the estimate of at time t. Then the estimated value of z
ADAPTIVE CONTROL
Let
denote the normalized estimation error, where n2 s is the normalizing signal, which we design to satisfy
Typical choices for ns that satisfy this condition are n2 s = T P for any P = PT > 0, and the like. When L ,
the condition is satisfied with m = 1, that is, ns = 0, in which case = 1 . We express in terms of the parameter
error = :
For simplicity, let us assume that L(s) is chosen so that WL is strictly proper and consider the following state
space representation of Eq. (8):
where Ac , Bc , and Cc are the matrices associated with a state-space representation that has a transfer function
W(s)L(s) = CT c (sI Ac ) 1 Bc .
Let us now consider the following Lyapunov-like function for the differential equation (9):
where = T > 0 is a constant matrix and Pc = PT c > 0 satisfies the algebraic equations
for some vector q, matrix Lc = LT c > 0, and small constant > 0. The existence of Pc = PT c > 0 satisfying the
above equations is guaranteed by the SPR property (12) of W(s)L(s) = CT c (sI Ac ) 1 Bc . The time derivative of
ADAPTIVE CONTROL
V is given by
are
We now need to choose = as a function of signals that can be measured so that the indefinite terms in V
canceled out. Because e is not available for measurement, cannot depend on e explicitly.
We know that Pc Bc = Cc , which implies that eT Pc Bc = eT Cc = . Therefore
we have
Gradient Method. In this method, we consider the parametric model in Eq. (5). Similar to the previous
subsection, we define (t) to be the on line estimate of and the normalized estimation error as
where z = T (t) and m2 = 1 + n2 s and n2 s is chosen so that /m L . The adaptive law is designed to minimize
the performance index J(), i.e.,
ADAPTIVE CONTROL
which gives
Different choices for the performance index lead to different adaptive laws.
Let us consider the simple quadratic cost function (instantaneous cost function)
Applying the gradient method, the minimizing trajectory (t) is generated by the differential equation
where = T > 0 is a scaling matrix that we refer to as the adaptive gain. We have
Least Squares. Let (t), , z be defined as above, and let m2 = 1 + n2 s , (t) be the estimate of at time
t, and m satisfy /m L . We consider the following cost function:
where Q0 = QT 0 > 0, 0, 0 = (0), which includes discounting of past data and a penalty on the initial estimate
0 of . Because z/m, /m L , we have that J() is a convex function of over Rn at each time t. Hence, any
local minimum is also global and satisfies
10
ADAPTIVE CONTROL
where
We refer to Eqs. (22) and (23) as the continuous-time recursive least-squares algorithm with forgetting factor.
The stability properties of the least-squares algorithm depend on the value of the forgetting factor .
In the identification literature, Eqs. (22) and (23) with = 0 are referred to as the pure least-squares
algorithm and have a very similar form to the Kalman filter. For this reason, the matrix P is usually called
the covariance matrix. The pure least-squares algorithm has the unique property of guaranteeing parameter
convergence to constant values as described by the following theorem:
Theorem 3. The pure least-squares algorithm guarantees that:
(i)
(ii)
(iii)
(iv)
P L .
, ns , , ,
, ns , L 2 .
where is a constant vector.
limt (t) = ,
If ns , L and is PE, then (t) converges to as t .
The proof of the theorem can be found in 12.
Bilinear Parametric Model. As will be shown in the next sections, a certain class of plants can be
parametrized in terms of their desired controller parameters, which are related to the plant parameters via
a Diophantine equation. Such parametrizations and their related estimation problem arise in direct adaptive
control and, in particular, direct MRAC, which is discussed in the next section.
In these cases appears in the form
where is an unknown constant, z, , z0 are signals that can be measured, and W(s) is a known proper transfer
function with stable poles. Because the unknown parameters , appear in a special bilinear form, we refer
to Eq. (24) as the bilinear parametric model.
Known Sign of . The SPRLyapunov design approach and the gradient method with an instantaneous
cost function discussed in the linear parametric case extend to the bilinear one in a rather straightforward
manner.
Let us start with the SPRLyapunov design approach. We rewrite Eq. (24) in the form
ADAPTIVE CONTROL
11
where z1 = L 1 (s)z0 , = L 1 (s), and L(s) is chosen so that L 1 (s) is proper and stable and WL is proper and
SPR. The estimate z of z and the normalized estimation error are generated as
and (t), (t) are the estimates of , at time t, respectively. Letting , , it follows from
Eqs. (25) to (27) that
Now T T = T T + T T = T
T , and therefore,
By choosing
, , L .
L 2 .
, ns , ,
If , L , is PE, and L 2 , then (t) converges to as t .
If L 2 , the estimate converges to a constant independent of the properties of .
The proof of the theorem can be found in Ref. 12. The case where the sign of is unknown is also given
in Ref. 12.
12
ADAPTIVE CONTROL
where Zp , Rp are monic polynomials and kp is a constant referred to as the high-frequency gain.
The reference model, selected by the designer to describe the desired characteristics of the plant, is
described by
Plant Assumptions.
P1. Zp (s) is a monic Hurwitz polynomial of degree mp .
P2. An upper bound n on the degree np of Rp (s) is known.
P3. The relative degree n = np mp of Gp (s) is known.
P4. The sign of the high-frequency gain kp is known.
MRC Schemes: Known Plant Parameters. In addition to assumptions P1 to P4 and M1, M2, let us
also assume that the plant parameters, that is, the coefficients of Gp (s), are known exactly. Because the plant
is LTI and known, the design of the MRC scheme is achieved using linear system theory.
We consider the feedback control law
ADAPTIVE CONTROL
13
where
where
We can now meet the control objective if we select the controller parameters 1 , 2 , 3 , c 0 so that the
closed-loop poles are stable and the closed-loop transfer function Gc (s) = W m (s) is satisfied for all s C. Choosing
and using (s) = 0 (s)Zm (s), the matching equation Gc (s) = W m (s) becomes
Equating the coefficients of the powers of s on both sides of Eq. (38), we can express it in terms of the algebraic
equation
14
ADAPTIVE CONTROL
The proof of the lemma can be found in Ref. 12.
MRAC for SISO Plants. The design of MRAC schemes for the plant in Eq. (31) with unknown parameters is based on the certainty equivalence approach and is conceptually simple. With this approach, we develop
a wide class of MRAC schemes by combining the MRC law, where is replaced by its estimate (t), with
different adaptive laws for generating (t) online. We design the adaptive laws by first developing appropriate
parametric models for , which we then use to pick up the adaptive law of our choice from the preceding
section.
Let us start with the control law
where = [T 1 , T 2 , 3 , c0 ]T and = [T 1 , T 2 , yp , r]T , and search for an adaptive law to generate (t), the estimate
of the desired parameter vector .
It can be seen that under the above control law, the tracking error satisfies
where = 1/c 0 , = [T 1 , T 2 , 3 , c 0 ]T . The above parametric model may be developed by using the matching
Eq. (38) to substitute for the unknown plant polynomial Rp (s) in the plant equation and by canceling the Hurwitz
polynomial Zp (s). The parametric model in Eq. (42) holds for any relative degree of the plant transfer function.
A linear parametric model for may also be developed from Eq. (42). Such a model takes the form
where
The main equations of several MRAC schemes formed by combining Eq. (41) with an adaptive law based
on Eq. (42) or (43). The following theorem gives the stability properties of the MRAC scheme.
Theorem 5. The closed-loop MRAC scheme in Eq. (41), with (t) adjusted with any adaptive law with normalization based on the model in Eq. (42) or (43) as described in the preceding section, has the following
properties:
(i) All signals are uniformly bounded.
(ii) The tracking error e1 = yp ym converges to zero as t .
ADAPTIVE CONTROL
15
(iii) If the reference input signal r is sufficiently rich of order 2n, r L , and Rp , Zp are coprime, then the
tracking error e1 and parameter error = converge to zero for the adaptive law with known sgn(kp ).
The proof of the theorem can be found in Ref. 12.
where Gp (s) is proper and Rp (s) is a monic polynomial. The control objective is to choose the plant input up so
that the closed-loop poles are assigned to those of a given monic Hurwitz polynomial A(s). The polynomial
A(s), referred to as the desired closed-loop characteristic polynomial, is chosen according to the closed-loop
performance requirements. To meet the control objective, we make the following assumptions about the plant:
P1. Rp (s) is a monic polynomial whose degreen n is known.
P2. Zp (s), Rp (s) are coprime, and degree(Zp ) < n.
Assumptions P1 and P2 allow Zp , Rp to be non-Hurwitz, in contrast to the MRC case, where Zp is required
to be Hurwitz. If, however, Zp is Hurwitz, the MRC problem is a special case of the general pole placement
problem defined above with A(s) restricted to have Zp as a factor.
In general, by assigning the closed-loop poles to those of A(s), we can guarantee closed-loop stability and
convergence of the plant output yp to zero provided there is no external input. We can also extend the PPC
objective to include tracking, where yp is required to follow a certain class of reference signals ym , by using the
internal model principle as fol8lows: The reference signal ym L is assumed to satisfy
where Qm (s), the internal model of ym , is a known monic polynomial of degree q with nonrepeated roots on the
j axis and satisfies
P3. Qm (s), Zp (s) are coprime.
For example, if yp is required to track the reference signal ym = 2 + sin 2t, then Qm (s) = s(s2 + 4) and
therefore, according to assumption P3, Zp (s) should not have s or s2 + 4 as a factor.
16
ADAPTIVE CONTROL
In addition to assumptions P1 to P3, let us also assume that the coefficients of Zp (s), Rp (s), i.e., the plant
parameters are known exactly.
We consider the control law
where P(s), L(s), M(s) are polynomials [with L(s) monic] of degree q + n 1, n 1, q + n 1, respectively, to
be found, and Qm (s) satisfies Eq. (45) and assumption P3.
Applying Eq. (46) to the plant in Eq. (44), we obtain the closed-loop plant equation
is satisfied for a given monic Hurwitz polynomial A(s) of degree 2n + q 1. It can be seen that assumptions
P2 and P3 guarantee that L, P satisfying Eq. (49) exist and are unique. The solution for the coefficients of L(s),
P(s) of Eq. (49) may be obtained by solving the algebraic equation
and l, p, a are the vectors whose entries are the coefficients of L(s), P(s) and A(s), respectively. The coprimeness
of Qm Rp , Zp guarantees that Sl is nonsingular; therefore, the coefficients of L(s), P(s) may be computed from
the equation
ADAPTIVE CONTROL
17
For zero tracking error, Eq. (51) suggests the choice of M(s) = P(s) to null the first term. The second term in Eq.
(51) is nulled by using Qm ym = 0. Therefore, the pole placement and tracking objective are achieved by using
the control law
which is implemented using n + q 1 integrators to realize C(s) = P(s)/Qm (s)L(s). Because L(s) is not necessarily
Hurwitz, the realization of Eq. (52) with n + q 1 integrators may have a transfer function, namely C(s), with
poles outside C . An alternative realization of Eq. (52) is obtained by rewriting it as
where is any monic Hurwitz polynomial of degree n + q 1. The control law (53) is implemented using 2(n
+ q 1) integrators to realize the proper stable transfer functions ( LQm )/, P/.
APPC. The APPC scheme that meets the control objective for the unknown plant is formed by combining the control law in Eq. (53) with an adaptive law based on the parametric model in Eq. (5). The adaptive law generates on-line estimates a , b of the coefficient vectors a of Rp (s) = sn + T a n 1 (s) and b
p (s, t) = sn + T a n 1 (s), Z p (s,
of Zp (s) = T b n 1 (s), respectively, to form the estimated plant polynomials R
T
t) = b n 1 (s). The estimated plant polynomials are used to compute the estimated controller polynomials L(s,
t), P(s,
t) by solving the Diophantine equation
Because different adaptive laws may be picked up from the section On-Line Parameter Estimation above, a
wide class of APPC schemes may be developed.
P or
The implementation of the APPC scheme requires that the solution of the polynomial Eq. (54) for L,
of the algebraic Eq. (55) for l exists at each time. The existence of this solution is guaranteed provided that
l (t) is nonsingular at each time t.
p (s, t)Qm (s), Z p (s, t) are coprime at each time t, that is, the Sylvester matrix S
R
p Qm , Z p are strongly coprime at each time t. Then
Theorem 6. Assume that the estimated plant polynomials R
all the signals in the closed-loop APPC scheme are u.b., and the tracking error converges to zero asymptotically
with time.
The proof of the theorem can be found in Ref. 12. The assumption that the estimated polynomials are
strongly coprime at each time t is restrictive and cannot be guaranteed by the adaptive law. Methods that relax
this assumption are given in Ref. 12.
18
ADAPTIVE CONTROL
where y, u R are the scalar system output and input, respectively, f , g are smooth vector fields, and x
[x1 , x2 , . . ., xn ]T is the state vector of the system. In order for the system in Eq. (57) to be controllable and
feedback-linearizable we assume that
A1. A lower bound for g(x) [i.e., |g(x)| > > 0 x Rn ] and the sign of g(x) are known.
The control objective is to find the control input u that guarantees signal boundedness and forces y to
follow the output ym of the reference model
= sn + k1 sn 1 + + kn be a Hurwitz polynomial (here s denotes the d/dt operator). Also let [e,
Let h(s)
e , . . ., e(n 1) ]T . Under assumption A1, the system Eq. (57) is a feedback-linearizable system. Therefore, if we
know the vector fields f and g, we can apply the static feedback
which implies that e, L and therefore all closed-loop signals are bounded, and limt e(t) = 0.
ADAPTIVE CONTROL
19
In many cases, the vector fields f and g are not completely known and thus adaptive versions of the
feedback law (60) have to be applied. For instance, using the usual assumption of linear parametrization, if the
vector fields f and g are of the form
where i , i = 1, 2, are vectors with unknown constant parameters, one may replace the feedback law in Eq. (60)
with the certainty-equivalent one [the certainty-equivalent feedback-linearizing (CEFL) controller]
where i , i = 1, 2, are the estimates of the unknown parameter vectors i , i = 1, 2. These estimates are generated
by an on-line adaptive law. We next propose the following adaptive laws for updating i :
where i , i = 1, 2, are symmetric positive definite matrices and 1 = bc f , 2 = ubc g , bc = [0, . . ., 0, 1] . The
next theorem summarizes the properties of the proposed control law.
Theorem 7. Consider the system in Eq. (57) and the feedback control law in Eqs. (62) and (63). Let assumption
A1 hold. Then, if T 2 (t)g (x(t)) = 0 for all t, all the closed-loop signals are bounded and the tracking error
converges to asymptotically to zero.
Parametric-Pure-Feedback Systems. Let us now try to extend the results of the previous section to
nonlinear systems that take the form
where u, zi R, f ij , gnj are smooth functions, and Rp is the vector of constant but unknown system parameters.
Let us rewrite Eq. (64) as
where f i1 () = f i0 () zi+1 . Systems of the form in Eq. (65) are called parametric-pure-feedback (PPF) systems
(13,14). Note that the above class of systems includes as a special case the system in Eq. (57) of the previous
section.
20
ADAPTIVE CONTROL
The control objective is to force the system output y to asymptotically track a reference signal ym . We
assume that the first n 1 time derivatives of ym are known. Also it will be assumed that ym and its first n
1 time derivatives are bounded and smooth signals.
Let us now assume that the parameter vector is known and construct a control law that meets the
control objectives. Before we design the feedback law, we will transform the system in Eq. (64) into a suitable
form. The procedure we will follow is based on the backstepping integrator principle (13).
Step 0. Let 1 z1 ym . Let also c1 , . . ., cn be positive constants to be chosen later.
Step 1. Using the chain-of-integrators method, we see that, if z2 were the control input in the z1 part of Eq.
(65) and were known, then the control law
would result in a globally asymptotically stable tracking, since such a control law would transform the z1
part of Eq. (65) as follows:
However, the state z2 is not the control. We therefore define 2 to be the difference between the actual z2
and its desired expression in Eq. (66):
Step 2. Using the above definition of 2 , the definition of 1 , and the z1 part of Eq. (65), we find that
where (2) is a (p + p2 )-dimensional vector that consists of all elements that are either of the form 2,i or
of the form 2,i j , where by 2,i we denote the ith entry of the vector . In the system (69) we will think of
z3 as our control input. Therefore, as in step 1, we define the new state 3 as
ADAPTIVE CONTROL
21
Step i (2 < i n 1). Using the definitions of 1 , . . ., i and working as in the previous steps, we may express
the derivative of i as
where the vector (i) contains all the terms of the form i1 i2 i with 1 j i. Defining now i+1 as
we obtain that
Step n.Using the definitions of 1 , . . ., n 1 and working as in the previous steps, we may express the derivative
of n as follows:
where the vector contains all the terms of the form i1 i2 ij with 1 j n, Y m [ym , y m , y m , . . .,
y(n 1) m ]T , and [ 0 (z1 , . . ., zn ) + T 1 (z1 , . . ., zn )] is given by
Using the definitions of 1 , . . ., n , and rearranging terms, we may rewrite Eq. (75) as follows:
22
ADAPTIVE CONTROL
Using the above methodology, we have therefore transformed the system in Eq. (65) into the following
one:
It can be shown that the matrix A0 is a stability matrix, provided that ci > 2.
Theorem 8. The control law in Eq. (79) guarantees that all the closed-loop signals are bounded and that the
tracking error converges to zero exponentially fast, provided that the design constants ci satisfy ci > 2.
In the case where the vector is not known, the certainty-equivalence principle can be employed in order
to design an adaptive controller for the system. However, the problem of designing parameter estimators for
the unknown parameters is not as easy as it was in the linear case. This can be seen from the fact that
the states i , i > 1, are not available for measurement, since they depend on the unknown parameters. To
overcome this problem a recursive design approach similar to the approach above can be constructed. The
difference between this approach [called adaptive backstepping (13)] and the approach presented above is the
following: in the approach presented above the states i , i > 1, depend on the unknown vector of parameters
; in the new approach they are appropriately redefined so they depend on the parameter estimate vector .
ADAPTIVE CONTROL
23
In order to overcome
Then the derivatives of i , i > 1, depend on the derivatives of the parameter estimates .
this problem, the adaptive backstepping approach makes use of the so-called tuning functions (13).
Next we present the adaptive controller that results from applying the adaptive backstepping procedure
to the system in Eq. (65) for the case where
A2. T f i2 () are independent of zi+1 and T gn2 = 0.
Also for simplicity, and without loss of generality, we will assume that f i1 () = 0. The case where assumption
A2 is not valid will be treated in the next subsection. The adaptive controller that results from applying the
adaptive backstepping procedure is recursively defined as follows:
Control law:
Tuning functions:
Regressor vectors:
Here ci > 2, i are positive design constants, and = T > 0 is a positive definite design matrix. The next
theorem summarizes the properties of the above control law.
Theorem 9. Suppose that assumptions A1 , A2 hold. Then the above adaptive control law guarantees that all
the closed-loop signals are bounded and that the tracking error converges asymptotically to zero.
The proof of the theorem can be found in Ref. 13.
24
ADAPTIVE CONTROL
A Control Law That Overcomes the Loss-Of-Controllability Problem. A significant problem that
arises in adaptive control of linear-in-the-parameters feedback-linearizable systems is the computation of
the feedback control law when the identification model becomes uncontrollable although the actual system
is controllable; so far, there is no known solution to this problem. For instance, for the case of the system
in Eq. (57) the parameter estimation techniques used in adaptive control cannot guarantee, in general, that
| 2 (t)T g (x(t))| > 0 for each time t, that is, they cannot guarantee that the identification model is controllable.
Also, for the case of PPF systems presented in the previous subsection, the adaptive backstepping techniques
guarantee global stability only in the case where assumption A2 is valid. Such restrictions are made because
the computation of the adaptive control law depends on the existence of the inverse of the matrix that consists
of the estimated input vector fields (or the Lie derivatives of the output functions along those vector fields).
Even in the case of known parameters where the inverse of the corresponding matrix exists (this is trivially
satisfied for feedback-linearizable systems), the inverse of the estimate of this matrix might not exist at each
time due to insufficiently rich regressor signals, large initial parameter estimation errors, and so on.
We next show how one can overcome the problem where the estimation model becomes uncontrollable, by
appropriately using switching adaptive control. We will apply the switching adaptive control methodology to
the PPF system of the previous subsection, by removing assumption A2.
Consider the Lyapunov function for the PPF system of the previous subsection.
Let us define
Note now that, using the definition of i , we can rewrite the i s as follows:
where i and w
i are appropriately defined known functions. Therefore, we have that
where is defined to be the vector whose entries are the elements i j , and 0 , 1 areappropriately defined
known functions.
We are now ready to present the proposed controller. The control input is chosen as follows:
ADAPTIVE CONTROL
where:
One has
is a continuous-switching signal that is used to switch from control u1 to control u2 and vice versa:
where is a symmetric positive definite design matrix and PC is defined as follows (15):
25
26
ADAPTIVE CONTROL
where
C, where
C2. j are sufficiently small. Moreover, (0)
Then for any compact X 0 Rn and for any positive constant c the following holds: there exist a positive
constant K such that, for any initial state x0 X 0 , the control law in Eqs. (83, 84, 85, 86, 87, 88, 89) with K > K
guarantees that all the closed-loop signals are bounded and, moreover, that the tracking error 1 converges in
finite time to the residual set
The idea of using switching adaptive controllers of the form presented above was first introduced in Ref.
16, where the proposed methodology was applied to systems of the form in Eq. (57). The controller of Ref. 16
was extended in Ref. 17 for PPF systems of the form in Eq. (65).
Acknowledgment
This article was supported by NASA grant NAGW-4103.
BIBLIOGRAPHY
1. J. A. Aseltine, A. R. Manchini, and C. W. Sartune, A survey of adaptive control systems, IRE Trans. Automat. Control,
6 (3): 1958.
2. D. McRuer, I. Ashkenas, D. Graham, Aircraft Dynamics and Automatic Control, Princeton, NJ: Princeton Univ. Press,
1973.
3. J. B. Cruz, Jr., System Sensitivity Analysis, Stroudsburg, PA: Dowden, Hutchinson & Ross, 1973.
4. L. P. Grayson, Design via Lyapunov second method, Proc. 4th JACC, 1963.
5. P. C. Parks, Lyapunov redesign of model reference control systems, IEEE Trans. Autom. Control, 11: 1966.
6. P. H. Phillipson, Design methods for model reference adaptive systems, Proc. Inst. Mech. Eng., 183 (35):, 695706, 1968.
7. R. V. Monopoli, Lyapunovs method for adaptive control design, IEEE Trans. Autom. Control, 3: 1967.
8. I. D. Landau, Adaptive Control: The Model Reference Approach, New York: Marcel Dekker, 1979.
ADAPTIVE CONTROL
27
9. K. S. Narendra, A. M. Annaswamy, Stable Adaptive Systems, Englewood Cliffs, NJ: Prentice-Hall, 1989.
10. S. Sastry, M. Bodson, Adaptive Control: Stability, Convergence and Robustness, Englewood Cliffs, NJ: Prentice-Hall,
1989.
11. K. J. Astrom, B. Wittenmark, Adaptive Control, Reading, MA: Addison-Wesley, 1989.
12. P. A. Ioannou, J. Sun, Robust Adaptive Control, Upper Saddle River, NJ: Prentice-Hall, 1996.
13. M. Krstic, M. I. Kanellakopoulos, and R. Kokotovic, Nonlinear and Adaptive Control Design, New York: Wiley, 1995.
14. D. Seto, A. M. Annaswamy, J. Baillieul, Adaptive control of nonlinear systems with a triangular structure, IEEE Trans.
Autom. Control, 7: 1994.
15. J.-B. Pomet, L. Praly, Adaptive nonlinear stabilization: Estimation from the Lyapunov equation, IEEE Trans. Autom.
Control, 37: 729740, 1992.
16. E. B. Kosmatopoulos and P. A. Ioannou, A switching adaptive controller for feedback linearizable systems, IEEE Trans.
Autom. Control, to be published.
17. E. B. Kosmatopoulos and P. A. Ioannou, Robust switching adaptive control using control Lyapunov functions, Proc.
IEEE Conf. Decision and Control, 1997.
PETROS A. IOANNOU
ELIAS B. KOSMATOPOULOS
University of Southern California
BILINEAR SYSTEMS
An electrical circuit or other engineering system often communicates with its external environment by input signals
that control its behavior and output signals; it is then called
a control system. If the components of a control system all
obey Ohms Law or one of its analogs, such as Hookes Law,
the system is called linear. In linear control systems, the
effect of the controls is additive and the output measurement is linear. They are discussed in the article in this
encyclopedia on MULTIVARIABLE SYSTEMS.
What distinguishes a bilinear system (BLS) is that although it is linear in its state variables, some control signal
u(t) exerts its effect multiplicatively. BLS may be given as
mathematical models of circuits or plants or may be chosen by a designer to obtain better system response than is
possible with a linear system. Their study is a rst step toward nonlinear control theory. Industrial process control,
economics, and biology provide examples of BLS with multiplicative controls such as valve settings, interest rates,
and neural signals, respectively. This topic of research began in the early 1960s with independent work in the USSR
and in the USA; see the surveys of Bruni et al. (1) and
Mohler (2, 3) for historical development and reviews of the
early literature.
dxi
(Fi j x j ) + vgi , i = 1, . . . , n; y =
h i xi
=
dt
n
j=1
i=1
(1)
(It is customary to suppress in the notation the timedependence of x, y and often the control v.)
As written, equation 1 has constant coefcients, and
such a control system is called time-invariant, which
means that its behavior does not depend on where we
choose the origin of the independent variable t; the system
can be initialized and control v exerted at any time. When
the coefcients are not constant, that is, made explicit in
the notation, e.g.,
x = F (t)x + vg(t), y = h (t)x
which is called a time-variant linear system.
For both linear and bilinear systems, we will need the
solution of x = Fx, which for a given initial condition x(0)
is x(t) = e Ft x(0). The matrix exponential function is dened
by
eFt = I + Ft +
(1)
F k tk
F 2t2
+ ... +
+ ...
2
k!
(2)
x(t) = e x(0) +
e(tr)F v(r)gdr
Ft
The polynomial
def
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright 2007 John Wiley & Sons, Inc.
Bilinear Systems
x = Ax +
Precisely what are BLS and why should one use bilinear
systems in control engineering? This section will give some
answers to those questions, starting with a formal denition, and give some examples.
Denition. A bilinear system is a rst-order vector differential equation x = p(x, u) in which x is a state vector;
u is a control signal (scalar or vector); and the components
of p(x, u) are polynomials in (x, u) that are linear in x, u
separately, but jointly quadratic, with constant real coefcients. Restating that, for any real numbers ,
p(x, u) p(x, u) + p(x, 0) + p(0, u)
The (optional) output is linear, y = h x.
To begin with, BLS are simpler and better understood
than most other nonlinear systems. Their study involves a
constant interplay between two protable viewpoints: looking at BLS as time-invariant nonlinear control systems and
as time-variant linear systems.
Another answer is that BLS are useful in designing control systems that use a very small control signal to modulate a large current of electricity of uid, apply brakes, or
change rates of growth.
A third answer is that the usual linearization of a nonlinear control system near an equilibrium point can be improved by using a BLS approximation; thus,
=
x = Ax + u(Bx + b), y = h x
(3)
k
(5)
B jx
j=1
(6)
(5)
u jB jx
j=1
rank(g, Fg, . . . , F n1 g) = n
If it is satised, the matrix-vector pair {F, g} is called a
controllable pair. If the controls are not bounded, this condition is sufcient for controllability, but if they are in some
U , the control may be small compared with Fx for large x
and have an insufcient effect.
k
(4)
type. Celikovsk
y and Vanec ek (7) have studied the thirdorder Lorenz system as a BLS with output feedback:
x = Ax + uBx, u(x) = x1 , with > 0, > 0, > 0, and
0
0 0 0
A=
, B = 0 0 1
1 0
0
0
0 1 0
For small all eigenvalues of A are negative, but for > 1
one becomes positive and B generates a rotation. This description seems to be characteristic of several such examples of strange attractors and bounded chaos, including the
Rossler attractor.
Bilinear Systems
1
x 2 = x1 ,
L
u=
1/R, closed
0,
open
(7)
(t0 , t0 ) = I
u(t),
t [0, 1 )
v1 (t), t [1 < t 2 ]
j=1
Bilinear Systems
(v ; , )
(u; , 0) =
(u v; , 0)
(8a)
(u ; 2, )
(u; , 0) = I
(8b)
(u; , 0) = e
At+
t k
i=1
ui (s)Bi ds
(0) = I
x(t) =
(u; t, 0)x(0) +
(u; t, s)u(s)bds
0
(u; T, 0) =
m
e(A+u(k1 )B)k
(9)
k=1
This idea can be carried much further with more analysis: More general (measurable) inputs can be approximated by PWC controls, and in analogy to the denition
of an integral as a limit of sums, the solution to equation 4
for measurable inputs can be written as (in an appropriate sense) the limit of products like equation 9, called a
product-integral.
The representation given by equation 9 generalizes to
the multi-input BLS equation 5 in the obvious way. With
equation 9, one can also easily verify the composition and
(for A = 0) inverse properties. To emphasize that exponential formulas for noncommuting matrices have surprising
behavior, here is a standard example in which you should
notice that A 2 = 0 and B 2 = 0.
0 1
0 0
1 0
A=
and B =
; AB BA =
;
0 0
1 0
0 1
eAt eBt =
1
0
e(A+B)t = exp
t
1
0
t
1
t
t
0
0
1
=
1 + t2
t
cosh(t)
sinh(t)
1
, but
1
sinh(t)
cosh(t)
(10)
Iu = {x|x u 0} = {x|C
(u; t, 0)x = 0, 0 t T }
which is a linear subspace called the u-unobservable subspace of the BLS; the u-observable subspace is the quotient
Bilinear Systems
State Observers
Given a u-observable system with A, B known, it is possible to estimate the initial state (or current state) from
the history HT . The theory of time-variant linear systems
(see Reference 10 or vol. 1 of Reference 3) shows that Ou is
the range of the time-variant version of the observability
Gramian,
I = {x|C
(u; t, 0)x = 0 for all u U}
(11)
def
(u; t, 0)C C
(u; t, 0)
WT =
x(0) =
WT1 (T )
(12a)
(12b)
e = (A + uB uKC)e
(12c)
Observer design for linear systems is concerned with nding K for which the observer is convergent, meaning that
e(t) 0, under some assumptions about u. (Observability is more than what is needed for the error to die out; a
weaker concept, detectability, will do. Roughly put, a system is detectable if what cannot be observed is asymptotically stable.) At least three design problems can be posed
for equation 12.
Bilinear Systems
N
y(tk ) C
(ud ; tk , 0)z2
k=1
CONSEQUENCES OF NONCOMMUTIVITY
The noncommutativity of the coefcient matrices A, B of
a BLS is crucial to its controllability, raising questions
that suggest for their answers some interesting mathematical tools: Lie algebras and Lie groups, named for Norwegian mathematician Sophus Lie, pronounced lee. See the
Reading List at the end of this article for books about them.
Lie Brackets
If A and B do not commute, then solving equation 4 is
more interesting and difcult. In addition to A, B, we will
need AB BA, which is written [A, B] and is called the
Lie bracket of A and B. To see how this matrix might come
into the picture, and obtain a geometric interpretation of
the Lie bracket, consider the two-input BLS with piecewise
constant controls
x = uAx + vBx; u, v {1, 0, 1}
(13a)
Starting at any x(0), use a control with four control segments (u, v), each of small duration > 0:
{1, 0}, {0, 1}, {1, 0}, {0, 1} on intervals
(13b)
(13c)
The nal state is (using the Taylor series of Eq. (7) for the
exponential and keeping only terms up to degree 2)
x(4) = 7eB eA eB eA x(0)
= (I + 2 [A, B] + 3 (higher order terms) + )x(0)
(13d)
The Lie Algebra of a BLS
In much the same way that controllability of the linear system equation 1 is related to the linear span of
{g, Fg, , F n1 g}, controllability of BLS is related to a linear space of matrices generated from A and B by repeated
bracketing, called the Lie algebra of the BLS. A survey of
Lie algebra facts is given in Belinfante and Kolman (16).
The primary concern of this subsection is homogeneous
BLS of the types given in equations 4 and 5. To make clear
what is being discussed, we need a denition.
Denition. A linear space g over a eld K (usually the
real or complex numbers) with a multiplication g g g:
{X, Y } [X, Y ] g will be called a Lie algebra if it satises
the properties
1. [X, Y ] = [X, Y ] = [X, Y ], K.
2. [X, Y ] + [X, X] = 0.
3. [X, [Y, Z]] + [Y, [Z, X]] + [Z, [X, Y ]] =
0 (Jacobi identity).
In mathematical writing, Lie algebras are usually given abstractly, by relations among their elements, and only then
does one represent the elements by matrices acting on some
vector space. In contrast, the Lie algebras of control theory
have specied generators. (In this article, the Lie algebras
are all matrix Lie algebras; for nonlinear vector eld Lie
algebras, see CONTROLLABILITY AND OBSERVABILITY.)
The Lie algebra generated by a BLS is constructed as
follows. Let Mn denote the linear space of all n n real
matrices; its dimension is r 2 . A real linear subspace g Mn
that is closed under the bracket [X, Y ] = XY YX is called a
(real) matrix Lie algebra. The space Mn itself can be identied with a Lie algebra of dimension n 2 over R called the
general linear algebra gl(n, R).
Two matrix Lie algebras g, g are said to be equivalent if
their elements are related by a common similarity transfor = P 1 XP. This denition is justied by the idenmation X
tity
P 1 [A, B]P = [P 1 AP, P 1 BP]
For homogeneous BLS with system matrices A and B as
in equation 4, we generate a Lie algebra in the following
way, based on the theory of free Lie algebras. {A, B}LA is
the subspace of Mn containing A and B and closed under
the Lie bracket and linear span operations. It is not hard to
compute {A, B}LA , using the properties 13 of Lie brackets.
Start with two generators A and B, we write down a tree
Bilinear Systems
The trees indicated structure depends only on the denition of a Lie algebra; T (A, B) is built level by level, bracketing each terminal leaf by A and by B; as shown it has already been pruned of obviously linearly dependent leaves
by the identities [X, Y ] = [Y, X] and [X, X] = 0. By using the Jacobi identity, more leaves can be pruned; e.g.,
note that [B, [A, [A, B]]] = [A, [B, [B, A]]] at level 4 before
building level 5, and so forth. It is a known feature of Lie
algebras that all higher order Lie brackets generated by A
and B can be obtained from those in this tree using the Jacobi identity. The linear span of the members of T (A, B) is
the desired Lie algebra g = {A, B}LA . As our matrices are in
Mn there can be no more than n 2 of them that are linearly
independent; working with the specic matrices A, B more
dependencies can be found (the Cayley-Hamilton Theorem
provides them) so the construction of the tree stops when
the matrices at some level are linearly dependent on their
ancestors.
In this process, we nd enough matrices to obtain the
dimension l and a basis B for the linear space g; we shall
write it as an array B = {C1 , . . . , Cl }. As we assume A and B
are not linearly dependent, it is convenient to take C1 = A,
C2 = B, C3 = [A, B]. If the entries of A and B are rational
numbers, symbolic algebra computer programs can generate T (A, B) and produce B.
It is known that generic pairs A, B will generate a tree
of matrices whose span is Mn . In other words, if you ll
out the entries with randomly chosen real numbers, then
for almost all sample pairs, {A, B}LA = gl(n, R). However,
control systems have structural relationships among their
components that may lead to smaller Lie algebras.
For example, here are some properties of generators that
are preserved under bracketing and linear combination,
and the name of the Lie subalgebra g gl(n, R) with the
given property and largest dimension d.
1
0
0
, J=
1
0
1
1
0
Bilinear Systems
at the set of states that can be reached at , any xed positive time, it is a circle with radius r() = e r0 , which is not
an open set, so the BLS does not have the strong accessibility property. No state with r() < r0 can be reached, which
establishes that the system is not controllable. The set
{r(t), (t)|0 < t < } is the open annulus {(r, )|r0 < r < r0 e },
so the control system does have the accessibility property.
Reverse the roles of the two generators. The BLS becomes x = (J + uI)x; in polar coordinates r = ur, = 1. At
later time T, the radius is arbitrary but (T ) = 2T + 0 .
This system has the accessibility property (but not strong
accessibility) and is controllable. If the trajectory misses
a target state, it takes 2 seconds before it has a second
chance. This peculiar state of affairs reects the fact that
our state space is punctured at the origin, and as we remarked before, it is topologically a cylinder.
Sometimes it is easy to see from the structure or known
symmetries of generators that a given system cannot have
the accessibility property; the most obvious of these, for
equation 5, are 1) the dimension of {A, B}LA is less than
n2 1 or 2) A and B are simultaneously triangularizable.
How can we guarantee accessibility on Rn0 ? For
a given homogeneous BLS, e.g., equation 4 whose
Lie algebra is g = {A, B}LA , construct the n l matrix
def
Bx = [C1 x, C2 x, . . . , Cl x], where l is the dimension of g. Dene the Lie rank (x) as the dimension of the linear span
of {Xx |X g}, or more constructively, (x) = rank(Bx).
For homogeneous BLS, a necessary and sufcient condition for the accessibility property is the Lie rank condition
introduced by J. Kucera (17):
(x) = n, for all x = 0
(14)
x1
x2
x2
x1
X M n+1 |X =
x
, X Mn , x Rn
0
X
0
x1
z = ...
, A = A 0 , B = 0
0
0
b
0
zn+1
with z n+1 = 0 and zn+1 (0) = 1. On Rn+1 , the bilinear system
z = Ax + u(t)Bx has aff (n, R) for its Lie algebra but on P
and is equivalent to the linear system we started with. Note
that
adA (B) =
0
0
Ab
2
, adA
(B) =
0
0 A2 b
, etc.
0 0
Brackets containing the factor B twice will vanish. The hyperplane zn+1 = 1 is invariant, and the BLS is controllable
on that hyperplane under the usual Kalman rank condition
that rank{b, Ab, . . .} = n.
The controllability properties of inhomogeneous BLS
like equation 3 x = Ax + u(Bx + b) can be studied using
the idea and notation of Example II. This BLS system can,
using n + 1 coordinates, be given as a homogeneous BLS
with an invariant hyperplane,
z = Az + u(t)Bz, where A =
A
0
0
, B=
0
B
0
b
0
Bilinear Systems
(15)
or = if all cadk1
A B = 0, k > 0. The BLS is invertible if
< and cadk1
Bx(0)
= 0. Invertibility fails if and only if
A
every control results in the same output.
The conditions of Hirschorn (Eq. 15) and Williamson
(Eq. 11) are interestingly related when one takes into account that in the system-inverse problem x(0) is known.
Assume that the Williamson condition holds, i.e., the rank
of W is n and cAk B = 0, 0 k n 2. They do not force
B = 0, because we can have cAn1 B = 0. Now evaluating
the Hirschorn conditions, in turn
cB = 0; c[A, B] = cAB = 0; c[A, [A, B]] = cA2 B = 0; . . .
n1
B = 0; so the relative order is = n.
but cadn1
A B = cA
Reference 21 points out, in this case the inverse system
becomes an observer, tracking x(t) when y() is supplied as
its input.
CONTROLLABILITY PROBLEMS
Criteria for Controllability. The necessary conditions for
controllability of homogeneous BLS begin with the Lie
rank condition (Eq. 14). Using it is made easier by the list
of transitive Lie algebras in References 1819. A necessary
condition for controllability given by Elliott (22) is that controllable homogeneous BLS on Rn0 , n > 2 have the strong
accessibility property. (The peculiarity of Example I occurs
because the punctured plane is not simply connected.)
For inhomogeneous BLS like equation 3, at x = 0, the
Kalman rank condition on {b, Ab, . . .} is a sufcient condition for local controllability with unbounded controls; as
that rank condition is an open one in the space of coefcients, for sufciently small x, the Bx term does no harm.
There is usually a family of possible equilibria corresponding to solutions of (A + B)x = b and at each of these
one can make such a test. In the rst paper on BLS, Rink
and Mohler (23) assumed such local controllability conditions to show controllability with bounded control u of
x = Ax +
m
(Bk x + bk )uk
k=1
(16)
(17)
10
Bilinear Systems
Lie rank condition is satised. Controllability and noncontrollability results are established for several families of A,
B pairs, especially for n = 2.
Bacciotti (29) completed the study for n = 2. He assumes
the same conditions: A > 0; B is diagonal and nonsingular,
with no repeated eigenvalues; and 2 > 0 (if not, reverse the
sign of the control). Then for the BLS x = Ax + uBx:
1. If 1 > 0, the BLS is completely controllable on Rn+ .
2. If 1 < 0 but = (2 a11 1 a22 )2 + 41 2 a12 a21 > 0 and
1 a22 2 a11 > 0, then the BLS is completely controllable on Rn+ . In any other case, controllability fails.
Sachkov (30) gives answers for m-input problems
(Eq. 5) with m = n 1 or m = n 2, using an idea that
has been successful in other
m controllability problems: if the
symmetric system x = i=1 ui Bi x is controllable on hypersurfaces V (x) = that ll up the (simply connected) state
space Rn+ , and if tne zero-control trajectories of equation
5 can cross all the hypersurfaces in both directions, then
equation 5 is globally controllable on Rn+ .
Stabilization II. At this point it is appropriate to look at
stabilization by state feedback controls that are not constant. For BLS the usual way of attacking this has been
by quadratic Lyapunov functions. Given a vector differential equation x = f(x), f (0) = 0, the basic idea behind
Lyapunovs Direct Method is to nd a family of nested
smooth hypersurfaces around 0, such as concentric ellipsoids, which the trajectories enter and never leave. For example, let us start with a test for stability of the differential equation x = Ax. Choose a symmetric matrix Q that is
positive denite, which means that all its eigenvalues are
positive and is easy to check by the criterion that each of
the n leading minor determinants of Q is positive:
Q11 Q12
> 0, . . . , det(Q) > 0
Q11 > 0,
Q12
Q22
Returning to equation 4, suppose A has all its eigenvalues on the imaginary axis. Then there is a positive denite Q for which QA + AQ = 0; by a change of variables,
Q = I. That is the assumption in Reference 25, which therefore uses the Lyapunov function V (x) = xx; the the adcondition on A and B (Eq. 16) is assumed to hold. The
proposed feedback control is u(x) = xBx; so along the
trajectories of the closed-loop system x = Ax (xBx)Bx,
we have V = (xBx)2 0 x = 0. From the ad-condition,
trajectories cannot remain in the set {xBx = 0|x = 0},
and V (x(t)) 0 as t . However, looking at the onedimensional case, one sees that for the differential equation = 3 , the approach to the origin is of order t 1/2 ,
and that this is also true for the n-dimensional problem
using quadratic feedback. Using the same hypotheses and
nearly the same short proof as in Reference 25, the reader
can verify that the bounded feedback u = xBx/xx provides exponential stability.
In linear system work, the choice of feedback control
is linear, u = c x. Applied to BLS x = Ax + uBx, this results in a quadratic system, such as the Lorenz system,
and is the topic of References 6 and 7. Reference 7 is concerned with linear feedback for a class of BLS in which B
generates a rotation and, for some constant , A + B has
real eigenvalues of mixed sign. For n = 2, after a similarity
transformation
A + B = diag(1 , 2 ); 1 > 0, 2 < 0, B = J
Using a constant control > , the BLS can be globally
practically stabilized; i.e., the trajectory eventually enters a ball of radius of order 1/. For n = 3, the same type of
system [now with 3 < 0, B so (3)] can be globally asymptotically stabilized given certain polynomial inequalities
in the i and B. Given some simpler inequalities, such as
tr(A) < 0 but allowing 1 > 0, the system is practically stabilized by a family of linear feedbacks with parameter .
Stabilization of homogeneous BLS in the plane has been
fully analyzed; it is not necessary that A have imaginary
eigenvalues. Bacciotti and Boieri (32) using constant, linear, and quadratic feedbacks, and Reference 8 using feedbacks differentiable except perhaps at 0, have given complete classications of the possibilities for stabilizability of
equation 4 on R20 . The methods of analysis in these papers
include Lyapunov functions, center manifolds, and properties of plane curves, depending on the various cases. These
cases are specied by the determinants and traces of A and
B, diagonalizability, and a few other structural features. In
Reference 8, a feedback control, useful even when A is not
stable, is u = xRx/xT x where matrices R, T are found
on a case by case basis. These controls are homogeneous
of degree zero, and if not constant are discontinuous at 0
but differentiate for x = 0. The stabilized trajectories typically approach the origin in a spiraling fashion, often with
large excursions; over each revolution the distance to 0 decreases in a constant ratio. In that sense they generalize
the constant controls; the signum controls are essentially
homogeneous of degree zero, and there may be other possibilities.
Recent work on more general problems of nonlinear system stabilization suggests that a time-periodic feedback
Bilinear Systems
k
ui Bi )
,
(0, 0
=I
(18)
As {
} contains I and is closed under composition [eq. 8a]
and inverse [eq. 8b], we see that it constitutes a group, in
fact a subgroup of GL(n,R). As all matrices in {
} have
11
positive determinants, {
} actually lies in GL+ (n, R).
On the other hand, corresponding to any matrix Lie algebra g, a Lie group we shall call G(g) can be constructed,
consisting of all products of exponentials of matrices in g;
see Reference 16 or Rossmanns book in the Reading List. If
a basis of g is {C1 , . . . , Cl }, in a neighborhood of the identity
G(g) has coordinates {exp(C1 t1 ), . . . , exp(C1 tl )} (t1 , . . . , tl ).
(Thus, g is the tangent space to G at I.) Using elements
of G to translate this coordinate patch anywhere on the
group, it can be observed that G(g) has an atlas of coordinate charts that are related by differentiable transformations where they overlap, like a geographic atlas, and on
which the group multiplication and inverse are differentiable functions. Lie groups occur in many applications of
mathematics to classical mechanics, quantum mechanics,
chemistry, and the control of robots and aerospace vehicles.
At this point we can note that the mathematics of controllability for symmetric BLS is rather simple. If a matrix
Lie group G has the property that given any two states x
and z in Rn0 , X G exists such that Xx = z, then G is called
transitive on Rn0 and g is a transitive Lie algebra.
In Reference 20, it was shown that G({B1 , . . . , Bk }LA ) =
{
}; that is, all matrices in the Lie group can be obtained as
transition matrices. This is a simple version of the Chow
Theorem of nonlinear control. Its meanings for BLS are
that, once the Lie algebra has been identied, the structure
of the group of transition matrices is completely known;
and that any matrix M G({B1 , . . . , Bk }LA ) can be written
as a product
M = et1 B1 es1 B2 . . . etk B1 esk B2
for some nite sequence of reals {t1 , s1 , . . . , tk , sk }. Thus, a
few generators are as good as l of them.
Some examples of Lie groups are useful; in some of those
listed below, the identity det(eA ) = etr(A) is relevant:
Lie algebra g
gl(n, R) = Mn
sl(n, R) = {X|tr(X) = 0}
so(n) = {X|X = X}
G GL(n, R)
GL+ (n, R) = {Q|det(Q) > 0}
SL(n, R) = {Q|det(Q) = 1}
SO(n) = {Q|Q Q = I}
12
Bilinear Systems
x = Ax + uBx; A =
AB =
1 + 1
1
1
2+ 0
0
2
1 1
;
1
1
2
A+B =
2
, B=
xs = Ax + bc x
(19)
(20)
v
c x
cos() sin()
, B=I
sin() cos()
Bilinear Systems
matrix
cb
c Ab
cA b
..
.
..
.
..
.
c Ab c A2 b c A3 b
..
.
= z1 + (z2 + z21 )u
= z2 + 3z21 + (z1 2z1 z2 2z31 )u
This was cooked by using the one-to-one coordinate transformation x1 = z1 , x2 = z2 + z21 , from the BLS x = Ax + uBx
where
A=
1
0
0
, B=
1
0
1
1
0
t
W0 (t) + n=1 0 0 n1
Wn (t, 1 , . . . , n )u(1 ) u(n )d1 dn
Treat our usual system x = Ax + u(t)Bx, as if it were inhomogeneous to get the integral equation
x(t) = eAt x(0) +
The Lie algebra of the BLS is sl(2, R). The matrices A,B are
the Jacobians at 0 of the nonlinear vector functions in the
rst system, so the BLS is also the classic linearization
of the nonlinear system, but what we have here is not an
approximation but a system equivalence. The Lie algebra
of nonlinear vector elds (see CONTROLLABILITY AND OBSERVABILITY, Section Nonlinear Finite-Dimensional Systems)
generated by our example is also sl(2, R).
It has been shown by Sedwick and Elliott (42) that given
a family F of real-analytic nonlinear vector elds on Rn that
vanish at 0, if F and the family F1 of their linear terms isomorphically generate a transitive Lie algebra, then a realanalytic coordinate transformation exists that transforms
F system to its equivalent BLS F1 , and that can be found
by solving linear partial differential equations. Of Lie algebras not on that list, the compact and the semisimple ones
also permit this linearization, but all others either fail to
have differentiable transformations or do not have any, as
far as is known; and nonlinear systems axe prone to have
innite-dimensional Lie algebras.
13
= A
+
k
Bi
(20)
i=1
14
Bilinear Systems
7. Celikovsk
y,
S.; Vanec ek, A., Bilinear Systems and Chaos. Kybernetika (Prague) 1994, 30, pp 403424.
8. Chabour, R.; Sallet, G.; Vivalda, J. C., Stabilization of Nonlinear Systems: A Bilinear Approach. Math. Contr. Signals Syst.
1993, 6, pp 224246.
9. Luesink, R.; Nijmeijer, H., On the Stabilization of Bilinear
Systems via Constant Feedback. Linear Algebra Appl. 1989,
122124, pp 457474.
10. Kailath, T., Linear Systems; Prentice-Hall: Englewood Cliffs,
NJ, 1980.
11. Grasselli, O. M.; Isidori, A., Deterministic State Reconstruction and Reachability of Bilinear Control Processes;
Proc. Joint Automatic Control Conf. San Francisco, June
2225, 1977; IEEE: New York, pp 14231427.
12. Williamson, D., Observation of Bilinear Systems with Application to Biological Control. Automatica: J. IFAC 1977, 13, pp
243254.
13. Frelek, B. A.; Elliott, D. L., Optimal Observation for VariableStructure Systems; Proc. VI IFAC Congress, Boston, Mass.;
vol I, Paper 29.5, 1975.
14. Grasselli, O. M.; Isidori, A., An Existence Theorem for Observers of Bilinear Systems. IEEE Trans. Autom. Contr. 1981,
AC-26, pp 12991300;Erratum 1982, AC-27, 284.
15. Sen, P., On the Choice of Input for Observability in Bilinear
Systems. IEEE Trans. Autom. Contr. 1981, AC-26, pp 451
454.
16. Belinfante, J. G. F.; Kolman, B., A Survey of Lie Groups and
Lie Algebras with Applications and Computational Methods;
Philadelphia: SIAM, 1972.
17. Kucera, J., Solution in Large of Control Problem x = (A(1
u) + Bu)x. Czechoslovak Math. J. 1966, 16, pp 600623.
18. Boothby, W. M.; Wilson, E. N., Determination on the Transitivity of Bilinera System. SIAM J. Contr. Optim. 1979, 17, pp
212221.
19. Kramer, L., Twotransitive Lie groups. J. Reine Angew Math.
2003, 563, pp 83113.
20. Brockett, R. W., System Theory on Group Manifolds and Coset
Spaces. SIAM J. Contr. 1972, 10, pp 265284.
21. Hirschorn, R. M., Invertibility of Control Systems on Lie
Groups. SIAM J. Contr. Optim. 1977, 15, pp 10341049.
22. Elliott, D. L., A Consequence of Controllability. J. Diff. Equations 1971, 10, pp 364370.
23. Rink, R. E.; Mohler, R. R., Completely Controllable Bilinear
Systems. SIAM J. Contr. 1968, 6, pp 477486.
24. Cheng, G.-S. J.; Tarn, T.-J.; Elliott, D. L., Controllability of Bilinear Systems. In Variable Structure Systems with Applications to Economics and Biology;Roberti, A.;Mohler, R. R., Eds.;
Springer-Verlag: Berlin, 1975; pp 83100.
25. Jurdjevic, V.; Quinn, J., Controllability and Stability. J. Diff.
Equation 1978, 28, pp 381389.
Bilinear Systems
Further Reading
The following books and articles are listed more or less in the order
of the topics of this article.
DAVID L. ELLIOTT
University of Maryland
15
281
as performance indexes in terms of state variables, then modern control approaches must be used.
The systems that may be designed by a conventional or
classical approach are usually limited to single-inputsingleoutput, linear time-invariant systems. The designer seeks to
satisfy all performance specifications by means of educated
trial-and-error repetition. After a system is designed, the designer checks to see if the designed system satisfies all the
performance specifications. If it does not, then he or she repeats the design process by adjusting parameter settings or
by changing the system configuration until the given specifications are met. Although the design is based on a trial-anderror procedure, the ingenuity and know-how of the designer
will play an important role in a successful design. An experienced designer may be able to design an acceptable system
without using many trials.
The primary objective of this article is to present procedures for the design and compensation of single-inputsingleoutput linear time-invariant control systems. Compensation
is the modification of the system dynamics to satisfy the given
specifications. The methods to the control system design and
compensation used in this article are the root-locus method
and frequency-response method. These methods are commonly called the classical or conventional methods of control
systems design. Note that in designing control systems by the
root-locus or frequency-response methods the final result is
not unique, because the best or optimal solution may not be
precisely defined if the time-domain specifications or frequency-domain specifications are given.
SYSTEM COMPENSATION
Setting the gain is the first step in adjusting the system for
satisfactory performance. In many practical cases, however,
the adjustment of the gain alone may not provide sufficient
alteration of the system behavior to meet the given specifications. As is frequently the case, increasing the gain value will
improve the steady-state behavior but will result in poor stability or even instability. It is then necessary to redesign the
system by modifying the structure or by incorporating additional devices or components to alter the overall behavior so
that the system will behave as desired. A device inserted into
the system for the purpose of satisfying the specifications is
called a compensator. The compensator compensates for deficit performance of the original system.
In discussing compensators, we frequently use such terminologies as lead network, lag network, and lag-lead network.
If a sinusoidal input ei is applied to the input of a network
and the steady-state output eo (which is also sinusoidal) has
a phase lead, then the network is called a lead network. (The
amount of phase lead angle is a function of the input frequency.) If the steady-state output eo has a phase lag, then
the network is called a lag network. In a lag-lead network,
both phase lag and phase lead occur in the output but in different frequency regions; phase lag occurs in the low-frequency region and phase lead occurs in the high-frequency
region. A compensator having a characteristic of a lead network, lag network, or lag-lead network is called a lead compensator, lag compensator, or lag-lead compensator.
In this article we specifically consider the design of lead
compensators, lag compensators, and lag-lead compensators.
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
282
R(s)
C(s)
G(s)
H(s)
Figure 1. Control system.
(1)
ROOT-LOCUS METHOD
The basic characteristic of the transient response of a closedloop system is closely related to the location of the closed-loop
poles. If the system has a variable loop gain, then the location
of the closed-loop poles depends on the value of the loop gain
chosen. It is important, therefore, that the designer know how
the closed-loop poles move in the s-plane as the loop gain is
varied.
From the design viewpoint, in some systems simple gain
adjustment may move the closed-loop poles to desired locations. Then the design problem may become the selection of
an appropriate gain value. If the gain adjustment alone does
not yield a desired result, addition of a compensator to the
system will become necessary.
A simple method for finding the roots of the characteristic
equation has been developed by W. R. Evans and used extensively in control engineering. This method, called the rootlocus method, is one in which the roots of the characteristic
equation are plotted for all values of a system parameter. The
roots corresponding to a particular value of this parameter
can then be located on the resulting graph. Note that the pa-
Angle condition:
&
G(s)H(s) = 180 (2k + 1)
(k = 0, 1, 2, . . . )
Magnitude condition:
|G(s)H(s)| = 1
The values of s that fulfill both the angle and magnitude conditions are the roots of the characteristic equation, or the
closed-loop poles. A plot of the points of the complex plane
satisfying the angle condition alone is the root locus. The
roots of the characteristic equation (the closed-loop poles) corresponding to a given value of the gain can be determined
from the magnitude condition.
FREQUENCY-RESPONSE METHOD
By the term frequency response, we mean the steady-state response of a system to a sinusoidal input. In frequencyresponse methods, we vary the frequency of the input signal
over a certain range and study the resulting response.
283
fied, then a suitable compensator that will reshape the openloop transfer function is determined. Finally, if there are any
other requirements to be met, we try to satisfy them, unless
some of them are contradictory to the other.
ROOT-LOCUS APPROACH TO THE
DESIGN OF CONTROL SYSTEMS
The root-locus approach to design is very powerful when the
specifications are given in terms of time-domain quantities,
such as the damping ratio and undamped natural frequency
of the desired dominant closed-loop poles, maximum overshoot, rise time, and settling time.
Consider a design problem in which the original system
either is unstable for all values of gain or is stable but has
undesirable transient-response characteristics. In such a
case, the reshaping of the root locus is necessary in the broad
neighborhood of the j axis and the origin in order that the
dominant closed-loop poles be at desired locations in the complex plane. This problem may be solved by inserting an appropriate lead compensator in cascade with the feedforward
transfer function.
If it is desired to improve steady-state performance (such
as to reduce the error in following the ramp input), insertion
of a lag compensator in the feedforward path will do the job.
If it is desired to improve both the transient-response and
steady-state performance, insertion of a lag-lead compensator
will accomplish the job. In what follows we discuss the lead,
lag, and lag-lead compensation techniques.
Lead Compensation
The procedure for designing a lead compensator for the system shown in Fig. 2 by the root-locus method may be stated
as follows:
1. From the performance specifications, determine the desired location for the dominant closed-loop poles.
2. By drawing the root-locus plot, ascertain whether or not
the gain adjustment alone can yield the desired closedloop poles. If not, calculate the angle deficiency . This
angle must be contributed by the lead compensator if
the new root locus is to pass through the desired locations for the dominant closed-loop poles.
3. Assume the lead compensator Gc(s) to be
1
s+
Ts + 1
T
, (0 < < 1)
= Kc
Gc (s) = Kc
1
Ts + 1
s+
T
(2)
where and T are determined from the angle deficiency. Kc is determined from the requirement of the
open-loop gain.
Gc(s)
G(s)
284
j
j3
j2
Closed-loop poles
j1
5 4
R(s)
4
s(s + 2)
C(s)
1
j1
j2
j3
(a)
4
s(s + 2)
The root-locus plot for this system is shown in Fig. 3(b). The
closed-loop poles are located at
s = 1 j 3
The damping ratio of the closed-loop poles is 0.5. The undamped natural frequency of the closed-loop poles is 2 rad/s.
The static velocity error constant is 2 s1.
It is desired to modify the closed-loop poles so that an undamped natural frequency n 4 rad/s is obtained, without
changing the value of the damping ratio, 0.5. In the present example, the desired locations of the closed-loop poles are
s = 2 j2 3
In some cases, after the root loci of the original system have
been obtained, the dominant closed-loop poles may be moved
(b)
Gc (s)G(s) =
1
T
Kc
1
s+
T
s+
G(s)
where
1
s+
Ts + 1
T
= Kc
Gc (s) = Kc
,
1
Ts + 1
s + T
(0 < < 1)
Notice that there are many possible values for T and that
will yield the necessary angle contribution at the desired
closed-loop poles.
The next step is to determine the locations of the zero and
pole of the lead compensator. There are many possibilities for
the choice of such locations. (See the comments at the end of
this example problem.) In what follows, we shall introduce a
procedure to obtain the largest possible value for . (Note that
a larger value of will produce a larger value of Kv. In most
cases, the larger the Kv is, the better the system performance.) First, draw a horizontal line passing through point
P, the desired location for one of the dominant closed-loop
poles. This is shown as line PA in Fig. 4. Draw also a line
connecting point P and the origin. Bisect the angle between
the lines PA and PO, as shown in Fig. 4. Draw two lines PC
and PD that make angles /2 with the bisector PB. The
intersections of PC and PD with the negative real axis give
the necessary location for the pole and zero of the lead net-
j
P
A
1
T
1
T
285
where K 4Kc. The root-locus plot for the compensated system is shown in Fig. 5. The gain K is evaluated from the magnitude condition as follows: Referring to the root-locus plot for
the compensated system shown in Fig. 5, the gain K is evaluated from the magnitude condition as
"
"
" K(s + 2.9) "
"
"
=1
" s(s + 2)(s + 5.4) "
s=2+ j2 3
or
K = 18.7
It follows that
Gc (s)G(s) =
"
18.7
= 4.68
4
s=2+ j2 3
Gc (s) = 2.51
Thus, if we need to force the root locus to go through the desired closed-loop pole, the lead compensator must contribute
30 at this point. By following the foregoing design procedure, we determine the zero and pole of the lead compensator,
as shown in Fig. 5, to be
18.7(s + 2.9)
s(s + 2)(s + 5.4)
s + 2.9
0.345s + 1
= 4.68
0.185s + 1
s + 5.4
Zero at s = 2.9,
Pole at s = 5.4
s18.7(s + 2.9)
s(s + 2)(s + 5.4)
= 5.02 s1
= lim
s0
or
T=
1
= 0.345,
2.9
T =
1
= 0.185
5.4
Thus 0.537. The open-loop transfer function of the compensated system becomes
K(s + 2.9)
s + 2.9
4
=
Gc (s)G(s) = Kc
s + 5.4 s(s + 2)
s(s + 2)(s + 5.4)
j4
j2
15
6
2
4
5.4
2.9
15
j2
j4
Figure 5. Root-locus plot of the compensated system.
286
Lag Compensation
Consider the case where the system exhibits satisfactory
transient-response characteristics but unsatisfactory steadystate characteristics. Compensation in this case essentially
consists of increasing the open-loop gain without appreciably
changing the transient-response characteristics. This means
that the root locus in the neighborhood of the dominant
closed-loop poles should not be changed appreciably, but the
open-loop gain should be increased as much as needed. This
can be accomplished if a lag compensator is put in cascade
with the given feedforward transfer function.
The procedure for designing lag compensator for the system shown in Fig. 2 by the root-locus method may be stated
as follows:
1. Draw the root-locus plot for the uncompensated system
whose open-loop transfer function is G(s). Based on the
transient-response specifications, locate the dominant
closed-loop poles on the root locus.
2. Assume the transfer function of the lag compensator to
be
1
s+
Ts + 1
T
= Kc
Gc (s) = Kc
1
Ts + 1
s+
T
1.06
C(s)
=
R(s)
s(s + 1)(s + 2) + 1.06
=
1.06
(s + 0.3307 j0.5864)(s + 0.3307 + j0.5864)(s + 2.3386)
( > 1)
(3)
Then the open-loop transfer function of the compensated system becomes Gc(s)G(s).
3. Evaluate the particular static error constant specified
in the problem.
4. Determine the amount of increase in the static error
constant necessary to satisfy the specifications.
5. Determine the pole and zero of the lag compensator that
produce the necessary increase in the particular static
j2
Closed-loop poles
j1
3
+
1.06
s(s + 1) (s + 2)
j1
j2
(a)
(b)
s + 0.05
s + 0.005
s + 0.05
1.06
s + 0.005 s(s + 1)(s + 2)
K(s + 0.05)
=
s(s + 0.005)(s + 1)(s + 2)
1.5
Uncompensated system
1
Imag axis
K = 1.06K c
1.06
s(s + 1) (s + 2)
(a)
Gc (s)G(s) = K c
where
s +0.05
s + 0.005
Kc = 0.966
Kc
287
0.5
New
closed-loop
pole
0
0.5
1
1.5
2
3 2.5 2 1.5 1 0.5
Real axis
0.5
(b)
s2 = 0.31 j0.55
0.5
0.4
= 1.0235
0.3
c is determined as
Then the lag compensator gain K
0.2
Imag axis
1.0235
K
=
= 0.9656
K c =
1.06
1.06
0.1
0
0.1
0.2
0.3
0.4
0.5
0.6
1.0235(s + 0.05)
G1 (s) =
s(s + 0.005)(s + 1)(s + 2)
5.12(20s + 1)
=
s(200s + 1)(s + 1)(0.5s + 1)
The static velocity error constant Kv is
Kv = lim sG1 (s) = 5.12 s1
s0
0.4
0.2
0
0.2
Real axis
0.4
0.6
(c)
Figure 7. (a) Compensated system; (b) root-locus plots of the compensated system and the uncompensated system; (c) root-locus plot of
compensated system near the origin.
s3 = 2.326,
1.4
1.2
Outputs c1 and c2
nal value. (The steady-state error with ramp inputs has decreased to about 10% of that of the original system.) We have
essentially accomplished the design objective of increasing
the static velocity error constant to about 5 s1.
Note that, since the pole and zero of the lag compensator
are placed close together and are located very near the origin,
their effect on the shape of the original root loci has been
small. Except for the presence of a small closed root locus
near the origin, the root loci of the compensated and the uncompensated systems are very similar to each other. However, the value of the static velocity error constant of the compensated system is 9.66 times greater than that of the
uncompensated system.
The two other closed-loop poles for the compensated system are found as follows:
Compensated system
1
Uncompensated system
0.8
0.6
0.4
0.2
0
10
15
20
s4 = 0.0549
25
30
35
40
t (s)
(a)
50
45
40
Outputs c1 and c2
288
35
30
Compensated system
25
Uncompensated system
20
15
10
5
0
10
15
20
25
30
35
40
45
50
t (s)
(b)
Figure 8. (a) Unit-step response curves for the compensated and uncompensated systems; (b) unit-ramp response curves for both
systems.
Lag-Lead Compensation
Lead compensation basically speeds up the response and increases the stability of the system. Lag compensation improves the steady-state accuracy of the system but reduces
the speed of the response.
If improvements in both transient response and steadystate response are desired, then both a lead compensator and
a lag compensator may be used simultaneously. Rather than
introducing both a lead compensator and a lag compensator
as separate elements, however, it is economical to use a single
lag-lead compensator.
Consider the system shown in Fig. 2. Assume that we use
the following lag-lead compensator:
1
1
s+
s+
(T1 s + 1)(T2 s + 1)
T1
T2
Gc (s) = Kc
= Kc
T1
1
s + 1 (T2 s + 1)
s+
s+
T1
T2
(4)
4
s(s + 0.5)
289
1
1
s+
4
T1
T2
Gc (s)G(s) = Kc
1
s(s + 0.5)
s+
s+
T1
T2
s+
"
1
"
" s+
"
T
2
"
"
"s + 1
" 1 T
s0
1
s1 +
T1
s1 +
T1
Kc = 10
Noting that
"
"
"
"
G(s1 )"" = 1
"
"
1
T1
=
s1 +
T1
s1 +
"
1
"
" s1 +
"
T
2
"
"
"s + 1
" 1 T
5 <
s0
Thus
"
"
"
"
"
"
"
"
"
"
"
"
"
4
"
s(s + 0.5) ""
"
"
is approximately unity, where s s1 is one of the dominant closed-loop poles. Determine the values of T1 and
from the magnitude and angle conditions:
"
"
"
"
"Kc
"
"
"
4
= 8Kc = 80
0.5
"
"
"
"
"+1
"
"
"
= 235
s=2.50+ j4.33
1
s1 +
T2
< 0
1
s1 +
T2
APB = 55 ,
PA
PB
4.77
8
j
The value of T2, the largest time constant of the laglead compensator, should not be too large to be physically realized.
j5
j4
j3
55
j2
j1
B
10 9 8
A
7 6 5 4 3 2 1 0
1
j1
j2
j3
j4
290
1.8
1.6
1.4
BO = 8.34
Uncompensated system
Outputs
1.2
or
1
= 0.420,
T1 =
2.38
Compensated
system
0.8
0.6
= 8.34T1 = 3.503
0.4
0.2
10
s + 2.38
0.5
1.5
s + 8.34
2.5
t (s)
(a)
3.5
4.5
T2 = 10
3.5
3
Outputs
Then
1
1
= 0.0285
=
T2
3.503 10
Gc (s) = (10)
s + 8.34
s + 0.1
s + 0.0285
Compensated
system
Uncompensated system
Gc (s)G(s) =
2
1.5
s + 2.38
2.5
0.5
0
0.5
1.5
2
t (s)
(b)
2.5
3.5
Figure 11. (a) Unit-step response curves for the compensated and
uncompensated systems; (b) unit-ramp response curves for both
systems.
No cancellation occurs in this case, and the compensated system is of fourth order. Because the angle contribution of the
phase lag portion of the lag-lead network is quite small, the
dominant closed-loop poles are located very near the desired
location. In fact, the dominant closed-loop poles are located at
s 2.4539 j4.3099. The two other closed-loop poles are
located at
uncompensated systems are shown in Fig. 11(a). The unitramp response curves for both systems are depicted in Fig.
11(b).
FREQUENCY-RESPONSE APPROACH
TO THE DESIGN OF CONTROL SYSTEMS
Lead Compensation
s = 0.1003,
s = 3.8604
jT + 1
jT + 1
(0 < < 1)
Im
Gc(s)
291
G(s)
m
m
1
(1 )
2
=0
0
Re
1 (1 + )
2
Figure 12. Polar plot of a lead compensator ( jT 1)/( jT 1),
where 0 1.
1
1
sin m = 2 =
1+
1+
2
(5)
Equation (5) relates the maximum phase lead angle and the
value of .
Figure 13 shows the Bode diagram of a lead compensator
when Kc 1 and 0.1. The corner frequencies for the lead
compensator are 1/T and 1/(T) 10/T. By examining Fig. 13, we see that m is the geometric mean of the two
corner frequencies, or
log m =
1
log
1
1
+ log
T
T
1
s+
Ts + 1
T
= Kc
Gc (s) = Kc
1
Ts + 1
s+
T
(0 < < 1)
(7)
Define
Kc = K
Then
Gc (s) = K
Ts + 1
Ts + 1
Hence,
1
m =
aT
Lead Compensation Techniques Based on the Frequency-Response Approach. The primary function of the lead compensator is to reshape the frequency-response curve to provide sufficient phase-lead angle to offset the excessive phase lag
associated with the components of the fixed system.
Consider the system shown in Fig. 14. Assume that the
performance specifications are given in terms of phase margin, gain margin, static velocity error constants, and so on.
The procedure for designing a lead compensator by the frequency-response approach may be stated as follows:
(6)
As seen from Fig. 13, the lead compensator is basically a highpass filter. (The high frequencies are passed, but low frequencies are attenuated.)
Ts + 1
Ts + 1
G(s) =
KG(s)
Ts + 1
Ts + 1
Ts + 1
=
G (s)
Ts + 1 1
Gc (s)G(s) = K
where
G1 (s) = KG(s)
10
dB
0
10
20
90
m
0
0.1
T
1
T
10
T
in rad/s
10
T
100
T
292
1
T
1
=
T
4K
s(s + 2)
where K Kc.
The first step in the design is to adjust the gain K to meet
the steady-state performance specification or to provide the
required static velocity error constant. Since this constant is
given as 20 s1, we obtain
s0
dB
0
20
40
0
90
G(s) =
20
40
Ts + 1
G (s)
Ts + 1 1
s4K
= 2K = 20
= lim
s0 s(s + 2)
17
180
1
810
in rad/s
With K 10, the compensated system will satisfy the steadystate requirement.
We shall next plot the Bode diagram of
20
40
=
j( j + 2)
j(0.5 j + 1)
G1 ( j) =
4
s(s + 2)
40 60 100
or
K = 10
20
1
1+
=1/( aT )
"
"
1 "
"
" 1 + j "
"
1
""
= ""
" =
1
" 1 + j "
"
"
Note that
+
1
1
1
= 6.2 dB
=
=
0.49
0.24
41.7(s + 4.41)
s + 18.4
and
Gc (s)G(s) = 41.7
1
c
= = 18.4
T
10
K
=
= 41.7
0.24
0.227s + 1
s + 4.41
= 10
s + 18.4
0.054s + 1
Note that
Gc (s)
Gc (s)
G1 (s) =
10G(s) = Gc (s)G(s)
K
10
20
Gc
10
6.2 dB
4
s + 4.41
s + 18.4 s(s + 2)
40
dB
4
s(s + 2)
The magnitude curve and phase-angle curve for Gc( j)/10 are
shown in Fig. 17. The compensated system has the following
open-loop transfer function:
1
= c = 4.41
T
Gc (s) = Kc
293
Kv
GcG
20
1.4
Compensated system
1.2
G1 = 10G
1
Outputs
40
0
Gc
10
90
G1 = 10G
0.6
0.4
0.2
GcG
50
180
1
Uncompensated system
0.8
10
20
40 60 100
in rad/s
Figure 17. Bode diagram for the compensated system.
3
t (s)
Figure 19. Unit-step response curves of the compensated and uncompensated systems.
294
30
5
4.5
dB
10
3.5
Compensated system
3
Outputs
20
2.5
90
0.01
T
2
1.5
1
T
10
T
in rad/s
Uncompensated system
0.5
0
0.1
T
0.5
1.5
2.5
3.5
4.5
Define
t (s)
Figure 20. Unit-ramp response curves of the compensated and uncompensated systems.
Kc = K
Then
1
s+
Ts + 1
T
= Kc
Gc (s) = Kc
1
Ts + 1
s+
T
( > 1)
Im
Kc
Kc
0
=0
Re
Gc (s) = K
Ts + 1
Ts + 1
Ts + 1
Ts + 1
G(s) =
KG(s)
Ts + 1
Ts + 1
Ts + 1
=
G (s)
Ts + 1 1
Gc (s)G(s) = K
where
G1 (s) = KG(s)
Determine gain K to satisfy the requirement on the
given static error constant.
2. If the uncompensated system G1( j) KG( j) does not
satisfy the specifications on the phase and gain margins, then find the frequency point where the phase
angle of the open-loop transfer function is equal to
180 plus the required phase margin. The required
phase margin is the specified phase margin plus 5 to
12. (The addition of 5 to 12 compensates for the phase
lag of the lag compensator.) Choose this frequency as
the new gain crossover frequency.
3. To prevent detrimental effects of phase lag due to the
lag compensator, the pole and zero of the lag compensator must be located substantially lower than the new
gain crossover frequency. Therefore, choose the corner
frequency 1/T (corresponding to the zero of the lag
compensator) 1 octave to 1 decade below the new gain
crossover frequency. (If the time constants of the lag
compensator do not become too large, the corner frequency 1/T may be chosen 1 decade below the new
gain crossover frequency.)
4. Determine the attenuation necessary to bring the magnitude curve down to 0 dB at the new gain crossover
frequency. Noting that this attenuation is 20 log ,
determine the value of . Then the other corner frequency (corresponding to the pole of the lag compensator) is determined from 1/(T).
295
40
G1
20
dB
K
Kc =
GcG
0
20
40
0
1
G(s) =
s(s + 1)(0.5s + 1)
90
11 dB
Gc
Gc
G1
40
180
GcG
It is desired to compensate the system so that the static velocity error constant Kv is 5 s1, the phase margin is at least 40,
and the gain margin is at least 10 dB.
We shall use a lag compensator of the form
1
s+
Ts + 1
T
= Kc
Gc (s) = Kc
1
Ts + 1
s+
T
Kc = K
Define also
K
s(s 1)(0.5s + 1)
The first step in the design is to adjust the gain K to meet the
required static velocity error constant. Thus,
= lim
s0
s0
0.1
in rad/s
Figure 24. Bode diagrams for the uncompensated system, the compensator, and the compensated system. (G1: uncompensated system,
Gc: compensator, GcG: compensated system.)
( > 1)
Define
G1 (s) = KG(s) =
270
0.004 0.01 0.02 0.04
Ts + 1
G (s) = lim sG1 (s)
s0
Ts + 1 1
sK
=K=5
s(s + 1)(0.5s + 1)
5
G1 ( j) =
j( j + 1)(0.5 j + 1)
1
= 20
or
= 10
+
s(s + 1) (0.5s + 1)
296
20
18
1
s+
10s + 1
10
= Kc
Gc (s) = Kc (10)
1
100s + 1
s+
100
16
Since the gain K was determined to be 5 and was determined to be 10, we have
5
K
=
= 0.5
Kc =
10
12
10
8
6
Outputs
14
5(10s + 1)
s(100s + 1)(s + 1)(0.5s + 1)
Uncompensated system
Compensated system
1.4
10
t (s)
12
14
16
18
20
1
1
s+
(T1 s + 1)(T2 s + 1)
T1
T2
Gc (s) = Kc
= Kc
T1
1
s + 1 (T2 s + 1)
s+
s+
T1
T2
(8)
s+
where 1. The phase lead portion of the lag-lead compensator (the portion involving T1) alters the frequency-response
curve by adding phase lead angle and increasing the phase
margin at the gain crossover frequency. The phase lag portion
(the portion involving T2) provides attenuation near and
above the gain crossover frequency and thereby allows an increase of gain at the low-frequency range to improve the
steady-state performance.
Figure 27 shows a Bode diagram of a lag-lead compensator
when Kc 1, 10, and T2 10T1. Notice that the magnitude curve has the value 0 dB at both low-frequency and highfrequency regions.
10
0
dB
Outputs
Figure 26. Unit-ramp response curves for the compensated and uncompensated systems.
Compensated system
1.2
Uncompensated system
0.8
10
20
30
0.6
90
0.4
0
0.2
0
10
15
20
25
30
35
40
t (s)
Figure 25. Unit-step response curves for the compensated and uncompensated systems.
90
0.001
T1
0.01
T1
0.1
T1
1
T1
10
T1
100
T1
in rad/s
Figure 27. Bode diagram of a lag-lead compensator given by Eq. (8)
with Kc 1, 10, and T2 10T1.
60
40
G
20
dB
20
16 dB
GcG
0
Gc
40
90
1
1
=
sin m =
1
+1
1+
0
Gc
90
50
180
270
0.01
297
GcG
32
4 6 10
in rad/s
Figure 28. Bode diagrams for the uncompensated system, the compensator, and the compensated system. (G: uncompensated system,
Gc: compensator, GcG: compensated system.)
We shall illustrate the details of the procedure for designing a lag-lead compensator by an example.
= 10
Then the corner frequency 1/T2 (which corresponds to
the pole of the phase lag portion of the compensator) becomes
0.015 rad/s. The transfer function of the phase lag portion of the lag-lead compensator then becomes
s + 0.15
6.67s + 1
= 10
s + 0.015
66.7s + 1
K
s(s + 1)(s + 2)
K
K
=
= 10
Kv = lim sGc (s)G(s) = lim sGc (s)
s0
s0
s(s + 1)(s + 2)
2
Hence,
K = 20
We shall next draw the Bode diagram of the uncompensated
system with K 20, as shown in Fig. 28. The phase margin
of the uncompensated system is found to be 32, which indicates that the incompensated system is unstable.
The next step in the design of a lag-lead compensator is to
choose a new gain crossover frequency. From the phase angle
curve for G( j), we notice that G( j) 180 at 1.5
rad/s. It is convenient to choose the new gain crossover frequency to be 1.5 rad/s so that the phase-lead angle required
at 1.5 rad/s is about 50, which is quite possible by use
of a single lag-lead compensator.
1.43s + 1
0.143s + 1
s + 0.7
s+7
s + 0.15
s + 0.015
=
1.43s + 1
0.143s + 1
6.67s + 1
66.7s + 1
The magnitude and phase-angle curves of the lag-lead compensator just designed are shown in Fig. 28. The open-loop
transfer function of the compensated system is
(s + 0.7)(s + 0.15)20
(s + 7)(s + 0.015)s(s + 1)(s + 2)
10(1.43s + 1)(6.67s + 1)
=
s(0.143s + 1)(66.7s + 1)(s + 1)(0.5s + 1)
Gc (s)G(s) =
(9)
298
1.6
1.4
3.
1.2
Output
1
0.8
0.6
4.
0.4
0.2
0
0
10 12
t (s)
14 16 18 20
20
5.
6.
MULTI-DEGREES-OF-FREEDOM CONTROL
In the classical design approaches presented in this article,
we design control systems such that the response to the reference input is satisfactory. If the control system is subjected to
other inputs, such as disturbance input and noise input, it is
not possible to design the system such that the responses to
the disturbance input and noise input are also satisfactory, in
addition to the primary requirement that the response to the
reference input is satisfactory. This is because the systems we
considered so far simply do not have the freedom to satisfy
requirements on the responses to disturbances and noises.
If we wish to design high-performance control systems in
the presence of disturbances and sensor noises, we must
change the configuration of the control system. This means
that we must provide additional degrees of freedom to the
control system to handle additional requirements.
18
16
D(s)
Output
14
12
R(s)
10
Gc(s)
U(s) +
+
Y(s)
Gp(s)
6
4
B(s)
2
0
0
10 12 14 16 18 20
t (s)
N(s)
+
299
D(s)
D(s)
Gc2(s)
R(s)
+
Gc1(s)
U(s) +
+
Y(s)
Gp(s)
R(s)
Gc1(s)
+ U(s) +
+
Y(s)
Gp(s)
Gc2(s)
B(s)
B(s)
B(s)
N(s)
N(s)
In what follows we first discuss the single-degree-of-freedom control systems and then discuss the two-degrees-offreedom control systems. Finally, we present an example of
three-degrees-of-freedom control systems that can satisfy the
requirements on the responses to the reference input, disturbance input, and noise input.
Single-Degree-of-Freedom Control
Two-Degrees-of-Freedom Control
Next consider the system shown in Fig. 32, where Gp(s) is the
transfer function of the plant and is assumed to be fixed and
unalterable. For this system, closed-loop transfer functions
Gyr, Gyn, and Gyd are given, respectively, by
Gyr
Gyd
Gyn
Gc Gp
Y (s)
=
=
R(s)
1 + Gc Gp
Gp
Y (s)
=
=
D(s)
1 + Gc Gp
Gc Gp
Y (s)
=
=
N(s)
1 + Gc Gp
Gyn =
Hence, we have
Gp Gyd
Gp
Gyd Gp
Gp
R(s)
Gc1
Gyn =
Gyd Gp
Gp
In this case, if Gyd is given, then Gyn is fixed, but Gyr is not
fixed, because Gc1 is independent of Gyd. Thus, two closed-loop
transfer functions among three closed-loop transfer functions
Gyr, Gyd, and Gyn are independent. Hence, this system is a twodegrees-of-freedom control system.
Similarly, the system shown in Fig. 33 is also a twodegrees-of-freedom control system, because for this system
D(s)
Gc2
Gyd
Gyn
Gyr =
Gc1 Gp
Y (s)
=
R(s)
1 + (Gc1 + Gc2 )Gp
Gp
Y (s)
=
=
D(s)
1 + (Gc1 + Gc2 )Gp
Gc1 + Gc2 )Gp
Y (s)
=
=
N(s)
1 + (Gc1 + Gc2 )Gp
Gyr =
Gc3
Y(s)
G1
G2
N(s)
+
Figure 34. Three-degrees-of-freedom system.
300
CONVEX OPTIMIZATION
Gc1 Gp
Y (s)
Gc2 Gp
=
+
R(s)
1 + Gc1 Gp
1 + Gc1 Gp
Gp
Y (s)
=
=
D(s)
1 + Gc1 Gp
Gc1 Gp
Y (s)
=
=
N(s)
1 + Gc1 Gp
Gyr =
Gyd
Gyn
Hence,
KATSUHIKO OGATA
Gp Gyd
University of Minnesota
Gp
Gyd Gp
Gp
Clearly, if Gyd is given, then Gyn is fixed, but Gyr is not fixed
because Gc2 is independent of Gyd.
Three-Degrees-of-Freedom Control
In the control system shown in Fig. 34, the transfer functions
Gc1, Gc2, and Gc3 are controllers, and transfer functions G1 and
G2 are plant transfer functions that are unalterable. It can be
shown that this control system is a three-degrees-of-freedom
system. If a system has this configuration, then it is possible
to design three controllers by use of the root-locus method
and/or frequency-response method (or other methods) such
that the responses to all three inputs are acceptable.
CONCLUDING COMMENTS
This article has presented easy-to-understand procedures for
designing lead compensators, lag compensators, and lag-lead
compensators by use of the root-locus method or frequencyresponse method. The systems are limited to single-input
single-output, linear time-invariant control systems. For such
systems various design methods are available in addition to
the root-locus method and frequency-response method. Interested readers are referred to specialized books on control systems, as listed in the Reading List.
Toward the end of this article we included discussions on
multi-degrees-of-freedom control systems for the informed
specialist.
Most of the materials presented in this article were taken,
with permission, from Katsuhiko Ogata, Modern Control Engineering 3/e, 1997. Prentice Hall, Upper Saddle River,
New Jersey.
Reading List
H. W. Bode, Network Analysis and Feedback Design, New York: Van
Nostrand Reinhold, 1945.
R. C. Dorf, Modern Control Systems, 6th ed., Reading, MA: AddisonWesley, 1992.
W. R. Evans, Graphical analysis of control systems, AIEE Trans. Part
II, 67: 547551, 1948.
W. R. Evans, Control system synthesis by root locus method, AIEE
Trans. Part II, 69: 6669, 1950.
W. R. Evans, The use of zeros and poles for frequency response or
transient response, ASME Trans., 76: 11351144, 1954.
G. F. Franklin, J. D. Powell, and A. Emami-Naeini, Feedback Control
of Dynamic Systems, Reading, MA: Addison-Wesley, 1986.
B. C. Kuo, Automatic Control Systems, 6th ed., Upper Saddle River,
NJ: Prentice Hall, 1991.
643
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
644
R(z)
Compensator
Gc(z)
Plant
Gp(z)
Y(z)
TE (z) =
(2)
Az
z1
R(z) =
ATz
(z 1)2
R(z) =
z1
(1)
z1
1
A(kT )2 u(kT )
2
or
KG(z)H(z) =
(1 z1 )R(z)
1 + KG(z)H(z)
or
r(kT ) = AkTu(kT )
Gc (z)Gp (z)
Y (z)
=
R(z)
1 + Gc (z)Gp (z)
K(z + 1 )(z + 2 ) . . . (z + l )
(z 1)n (z + 1 )(z + 2 ) . . . (z + m )
KN(z)
=
(z 1)n D(z)
1
E(z)
=
R(z)
1 + KG(z)H(z)
Assuming that all the closed-loop poles of the system are inside the unit circle on the z plane, the steady-state error to a
power-of-time input is given by the final value theorem:
R(z)
then
Y(z)
T 2 Az(z + 1)
2 (z 1)3
H(z)
645
System
Type
Steady-State Error
to Ramp Input
ATz
R(z)
(z 1)2
Steady-State Error to
Parabolic Input
ATz
R(z)
(z 1)2
A
0
1K
N(1)
D(1)
AT
N(1)
K
D(1)
AT 2
N(1)
K
D(1)
is superimposed the locus of the poles of the closed-loop transfer function as some parameter is varied. For the configuration shown in Fig. 2 where the constant gain K is the parameter of interest, the overall transfer function of this system is
T (z) =
KG(z)
1 + KG(z)H(z)
R(z)
Tape motion
dynamics
Digital controller
Desired
tape
position
E(z)
Gc(z)
G(s) =
D/A
with
S/H
(1/120)s
40e
s + 40
Positionvelocity
relation
Tape
position
1
s
Position
sensor
A/D
(a)
Gp(z)
R(z) +
E(z)
Gc(z)
D/A
with
S/H
Gp(z)
(b)
1
s
P(s)
P(z)
A/D
Figure 3. Videotape-positioning system. (a) Block diagram. (b) Relation between discrete-time signals.
646
1 e(1/120)s 40e(1/120)s 1
G(z) = Z
s
s + 40
s
1
1/40
(1/120)s (1/120)s 1/40
+ 2+
= Z [1 e
]e
s
s
s + 40
z/40
z/120
z/40
= (1 z1 )z1
+
+
z1
(z 1)2
z 0.72
0.00133(z + 0.75)
=
z(z 1)(z 0.72)
R(z)
E(z)
Gp(z) =
0.00133 (z + 0.75)
z (z 1) (z 0.72)
Gc(z) = K
P(z)
(a)
Im
Unit
circle
K = 1000
The position error signal, in terms of the compensators ztransfer function Gc(z), is given by
Gc (z)Gp (z)
R(z)
1 + Gc (z)Gp (z)
E(z) = R(z) Y (z) = 1
95
K=
1000 K=100
1
=
R(z)
1 + Gc (z)Gp (z)
0.00133
Re
0.75
10
0.72
1
1 + Gc (z)Gp (z)
z
z+1
Circle of
radius
0.56
z1
z1
(b)
1
1 + Gc (z)Gp (z)
Im
Unit
circle
0.00133
0.75
0.72
Re
c = 0.56
K(z 0.72)
Gc (z) =
z
T (z) =
(c)
(3)
0.2(z + 0.75)
Gc (z)Gp (z)
= 3
1 + Gc (z)Gp (z)
z 0.6z2 0.2z + 0.15
z3 0.6z2 0.4z
=0
z3 0.6z2 0.2z + 0.15
1
(z2 + 0.4z)
1
120
=
limit 3
z1 z 0.6z2 0.2z + 0.15
30
R(z)
+
Gp(z) =
0.00133 (z + 0.75)
z (z 1) (z 0.72)
Gc(z) =
K(z 0.72)
z
E(z)
647
Im
P(z)
Circle of
radius 0.56
Unit circle
(a)
a = 0.2
Im
Circle of
radius
0.56
K=1000
a = 0.4
K
K
0.75
K
a = 0.6
370
a = 0.6
Unit circle
90
0.00133
Re
K 90
370
(b)
150(z 0.72)
z+a
1
Re
K = 1000
0.2(z + 0.75)
Gc (z)Gp (z)
= 3
1 + Gc (z)Gp (z)
z z2 + 0.2z + 0.15 + a(z2 z)
R(z)
E(z)
Gp(z) =
0.00133 (z + 0.75)
z (z 1) (z 0.72)
Gc(z) =
K(z 0.72)
z + 0.4
P(z)
(a)
Gc (z)Gp (z) =
Im
Circle of
radius 0.56
K = 1000
Unit circle
K = 1000
K = 150
0.75 0.4
K = 150
0.00133
Re
(b)
Figure 6. Compensator with zero at z 0.72 and pole at z 0.4.
(a) Block diagram. (b) Root locus for positive K.
0.1995(e jT + 0.75)
+ 0.4)(e jT 1)
jT (e jT
(4)
648
Gain (dB)
50
50 1
10
100
101
102
103
102
103
Frequency (rad/s)
Phase (deg)
0
90
Phase margin
180
270
360
101
w=
Gain margin
100
101
Frequency (rad/s)
z1
,
z+1
z=
w+1
1w
z1
z+1
w+1
1w
to obtain G(w)H(w).
2. Substitute w j into G(w)H(w) and generate frequency response plots in terms of the real frequency ,
such as Nyquist, Bode, Nichols, and so on. The w plane
can be thought of as if it were the s plane.
3. Determine the stability margins, crossover frequencies,
bandwidth, closed-loop frequency response, or any other
desired frequency response characteristics.
4. If it is necessary, design a compensator Gc(w) to satisfy
the frequency domain performance requirements.
Plant
(k)
u(k)
+
2
= tan1
T
designer can freely choose the character of the overall systems transient performance. When the plant state vector is
not available for feedback, as is usually the case, an observer
can be designed to estimate the state vector. As we shall see
later, the observer state estimate can be used for feedback in
place of the state itself.
649
Letting
Ev i = i
we obtain
v i = (iI A )1 B i
(5)
...
v 1v 2 . . . v n ] = EV = = [ 1 2 n ]
E [v
and the desired feedback gain matrix is
V 1
E = V
1 x1 (k)
1
+
u(k)
0 x2 (k)
2
x1 (k + 1)
0.5
0
1
x1 (k)
0.5 2 x2 (k)
x2 (k + 1) = 0
x3 (k + 1)
x3 (k)
1
1
0
1
0
u1 (k)
+ 0 2
u2 (k)
1
1
we let
u1 (k) = 3(k)
= [1.5
1]
1
2
1
2
1
3
3
1
1
+ 0.5
3
3
1
0
and
u2 (k) = (k)
0]
0]
e = [0
(6)
1]
x1 (k + 1)
1
=
x2 (k + 1)
3
Thus
x1 (k + 1)
0.5
x2 (k + 1) = 0
x3 (k + 1)
1
0
0.5
1
1
3
x1 (k)
2 x2 (k) + 2 (k)
x3 (k)
0
2
650
which is a controllable single input system. If the desired eigenvalues are located at z1 0.1, z2 0.15, and z3 0.1,
Ackermanns formula gives
e = [0.152
0.0223
0.2807]
and hence the feedback gain matrix for the multiple input
system is
E=
0.4559
0.1520
0.0669
0.0223
0.8420
0.2807
6.5871
v 1 = 4.5506 ,
0.3652
6.9304
v 2 = 4.8442 ,
0.5744
5.5076
v 3 = 3.4772
0.3046
and therefore
3
E=
1
3
1
3
0.4559
1
V =
1
0.1520
0.0669
0.0223
0.8420
0.2807
will place the feedback system eigenvalues arbitrarily, provided that the matrix I EC1D is nonsingular. If it is singular, a small change in the feedback gain matrix E which corresponds to small changes in the desired feedback system
eigenvalue locations eliminates the singularity. If the nth-order system has more than n outputs, only n of these can be
linearly independent, so excess linearly dependent output
equations can simply be ignored when recovering a systems
state from its output. To improve a feedback systems reliability and its performance in the presence of noise, one may wish
instead to combine linearly dependent outputs with other outputs rather than ignore them.
When the nth-order plant does not have n linearly independent measurement outputs, it still might be possible to
select a feedback matrix E in
u (k) = E y (k) Du
Du(k) + (k) = ECx
ECx(k) + (k)
to place all of the feedback system eigenvalues, those of (A
BEC), acceptably. Generally, however, output feedback alone
does not allow arbitrary feedback system eigenvalue placement.
Pole Placement with Feedback Compensation
We now present another viewpoint for placing the feedback
system poles using a transfer function approach. Although
our discussion is limited to single-input and single-output
plants, the results can be generalized to the case of plants
with multiple inputs and multiple outputs. Similar to output
feedback, pole placement with feedback compensation assumes that the measurement outputs of a plant, not the state
vector, are available for feedback.
For an nth-order, linear, step-invariant, discrete-time system described by the transfer function G(z), arbitrary pole
placement of the feedback system can be accomplished with
an mth-order feedback compensator as shown in Fig. 10.
Let the numerator and denominator polynomials on G(z)
be Np(z) and Dp(z), respectively. Also, let the numerator and
denominator of the compensator transfer function H(z) be
Nc(z) and Dc(z), respectively. Then, the overall transfer function of the system is
T (z) =
x (k + 1) = Ax
Ax(k) + Bu
Bu(k)
Np (z)Dc (z)
P(z)
G(z)
=
=
1 + G(z)H(z)
Dp (z)Dc (z) + Np (z)Nc (z)
Q(z)
y (k) = Cx
Cx(k) + Du
Du(k)
has n linearly independent outputs, that is, if the output coupling matrix C has n linearly independent rows, then the
plant state can be recovered from the plant inputs and the
measurement outputs:
x (k) = C 1 {yy (k) Du
Du(k)}
The output feedback
u (k) = Ex
Ex(k) + (k) = EC 1 {yy (k) Du
Du(k)} + (k)
G(z) =
Np(z)
Y(z)
Dp(z)
Feedback
compensator
Nc(z)
H(z) =
Dc(z)
or
u (k) = (II + EC 1D )1EC 1y (k) + (II + EC 1D )1 (k)
which has closed-loop zeros in P(z) that are those of the plant,
in Np(z), together with zeros that are the poles of the feedback
compensator, in Dc(z).
For a desired set of poles of T(z), given with an unknown
multiplicative constant by the polynomial Q(z), we obtain
Dp (z)Dc (z) + Np (z)Nc (z) = Q(z)
(7)
m n1
(8)
(9)
1 z + 2
Nc (z)
=
z + 3
Dc (z)
(10)
1 = 0.325,
2 = 0.00265,
0.325z 0.00265
z + 0.1185
or
G(z) =
651
3 = 0.1185
P (N)xx (N) +
J = x (N)P
N1
k=0
652
[x (i)Q(i)x(i) u (i)R(i)u(i)]
N1
i0
min
J x(0)P(0)x(0)
u(0), . . ., u(N 1)
nique developed by Richard Bellman in the 1950s in connection with his invention of dynamic programming. To apply
the principle of optimality, one begins at the next-to-last step
N 1 and finds the last input u(N 1) that minimizes the
cost of control from step N 1 to step N, J(N 1, N), as a
function of the beginning state for that step, x(N 1). Then
the input u(N 2) is found that minimizes J(N 2, N) when
u(N 1) is as previously determined. One proceeds in this
manner finding one control vector at a time, from the last to
the first, as a function of the systems state. This results in a
recursive calculation of the optimal feedback gains for the linear-quadratic regulator as given in Table 2. Beginning with
known N, P(N), Q, and R, the last feedback gain matrix
E(N 1) is calculated. Using E(N 1), the matrix P(N 1)
is computed. Then all of the indices are stepped backward one
step and, with P(N 1), the feedback gain matrix E(N 2)
is calculated. Using E(N 2), we can calculate P(N 2). The
cycle is continued until E(0) is found. A formidable amount of
algebraic computation is required; the user should therefore
have digital computer aid for all but the lowest-order
problems.
ing at N and proceeding backward is always the same independent of the value of N. The procedure summarized in Table 3 for calculating the optimal regulator gain can be easily
adapted to the steady-state regulator gain by replacing steps
5, 6, and 7 with the following single step 5:
5. Form
1
1
B W 11W 1
E = [B
21 B + R ] B W 11W 21 A
When the plant state vector is not entirely available for feedback, as is usually the case, the state is estimated with an
observer, and the estimated state can be used in place of the
actual state for feedback (see Refs. 9, 10).
For an nth-order step-invariant discrete-time plant,
x (k + 1) = Ax
Ax(k) + Bu
Bu(k)
y (k) = Cx
Cx(k) + Du
Du(k)
(12)
(13)
or
A GC
x (k) (k) = (A
GC)k [xx (0) (0)] = F k [xx (0) (0)]
[x (i)Qx(i) u (i)Ru(i)]
N1
i0
A QA1BR1B
A1BR1B
QA1
A1
n = 0
and since every matrix satisfies its own characteristic equation, then
0
1
Fn = 0
At the nth step, the error between the plant state and the
observer state is given by
W11
W21
653
W12
W22
so that
5. Form
x (n) = (n)
x (k + 1) (k + 1) = Ax
Ax(k) + Bu
Bu(k) F (k) Gy
Gy(k) Hu
Hu(k)
= Ax
Ax(k) + Bu
Bu(k) F (k) GCx
GCx(k)
and the observer state equals the plant state. Such an observer is termed deadbeat. In subsequent steps, the observer
state continues to equal the plant state.
There are several methods for calculating the observer
gain matrix g for single-output plants. Similar to the situation with state feedback, if a single-output plant is in observable canonical form, finding the elements of the observer gain
vector g for arbitrary eigenvalue placement is simple, because
each element of the observer gain vector determines one coefficient of the observer characteristic equation. Usually, however, the plant is not in observable canonical form. One way
of designing an observer for a completely observable singleoutput plant that is not in observable form is to change the
GDu
GDu(k) Hu
Hu(k)
A GC
= (A
GC)xx (k) F (k)
B GD H )u
u (k)
+ (B
Observer
(14)
H = B GD
(15)
654
(16)
c
...
c A
...
2
c A
M0 =
...
.
...
c A n
Mx
Mx(k + 1) (k + 1)
0
. . .
0
. . .
0
jn =
. . .
.
.
.
. . .
1
Another popular form of a full-order observer which is viewed
as an error feedback system can be obtained by expressing
the observer Eqs. (13), (14), and (15) in the form
A GC
B GD
u (k)
(k + 1) = (A
GC) (k) + Gy
Gy(k) + (B
GD)u
= MAx
MAx(k) + MBu
MBu(k) F (k) Gy
Gy(k) Hu
Hu(k)
MA GC
MB GD H )u
u (k)
= (MA
GC)xx (k) F (k) + (MB
where M is m n. For the observer error system to be autonomous, we require
FM = MA GC
H = MB GD
(17)
C
W (k)
x (k) = Nx
Nx(k)
M
= A (k) + Bu
Bu(k) + G [yy (k) w (k)]
where
w (k) = C (k) + Du
Du(k)
Here, the observer consists of a model of the plant driven by
the input u(k) and the error between the plant output y(k)
and the plant output that is estimated by the model w(k).
This form of a full-order observer is similar to the Kalman
Bucy filter (see Ref. 1).
655
D
+
(k) = 0
+
u(k)
x(k+1)
Unit
delay
y(k)
x(k)
C
+
A
D
+
(k+1)
Unit
delay
(k)
C
(18)
(19)
(20)
can be designed such that the overall feedback system eigenvalues are specified by the designer. The design procedure
proceeds in two steps. First, the state feedback is designed
to place the n-state feedback system eigenvalues at desired
locations as if the state vector were accessible. Second, the
state feedback is replaced by feedback of an observer estimate
of the same linear transformations of the state. As an example of eigenvalue placement with observer feedback, Figure
11 shows eigenvalue placement with full order state observer.
The eigenvalues of the overall system are those of the state
feedback and those of the full-order observer.
A tracking system in which the plant outputs are controlled so that they become and remain nearly equal to
externally applied reference signals r(k) is shown in Fig.
12(a). The outputs y(k) are said to track or follow the reference inputs.
As shown in Fig. 12(b), a linear, step-invariant controller
of a multiple-input/multiple-output plant is described by two
transfer function matrices: one relating the reference inputs
to the plant inputs, and the other relating the output feedback vector to the plant inputs. The feedback compensator is
used for shaping the plants zero-input response by placing
the feedback system eigenvalues at desired locations as was
discussed in the previous subsections. The input compensator,
on the other hand, is designed to achieve good tracking of the
reference inputs by the system outputs.
The output of any linear system can always be decomposed
into two parts: the zero-input component due to the initial
conditions alone, and the zero-state component due to the input alone. That is,
y (k) = y zero input (k) + y zero state (k)
Basically, there are three methods for tracking system design:
1. Ideal tracking system design
2. Response model design
656
Plant
inputs
Controller
Tracking and
feedback outputs
y(k)
Plant
u(k)
(a)
G (z) = I
T (z)G
Input
compensator
(k)
u(k)
Plant
y(k)
Feedback
compensator
(b)
r(k)
Input
(k)
compensator
Gr(z)
u(k)
Plant
y(k)
Feedback
compensator
G (z) = (z)
T (z)G
(c)
Reference
inputs
Reference
r(k)
input filter
Gr(z)
Plant
inputs
r(k)
Plant with
feedback
T(z)
Tracking
outputs
y(k)
(d)
the designer to specify a class of representative reference inputs that are to be tracked exactly, rather than having to
specify acceptable response models for all the possible inputs.
In the reference model tracking system design, additional
external input signals r(k) to the composite system are applied to the original plant inputs and to the observer state
equations so that the feedback system is, instead of being described by Eqs. (18), (19), and (20), described by Eq. (12) and
(k + 1) = F (k) + Gy
Gy(k) + Hu
Hu(k) + Jr
Jr(k)
w (k) = L (k) + N [yy (k) Du
Du(k)]
Observer of the
signal model
Fictitous
autonomous
reference signal
model
r(k)
Plant with
observer
feedback
y(k)
(21)
(22)
with
u (k) = Ew
Ew(k) + Pr
Pr(k)
657
(23)
= A x (k) + B rr(k)
(25)
and for
DEL
DEL]xx (k) + DPr
DPr(k)
y (k) r (k)
= C x (k) + D rr(k)
it is necessary that
where
H = H + GD
(k + 1) = (k)
r (k) =
(k)
(24)
(26)
SIMULATION
One of the most important control system design tools is simulationthat is, computer modeling of the plant and controller to verify the properties of a preliminary design and to test
its performance under conditions (e.g., noise, disturbances,
parameter variations, and nonlinearities) that might be difficult or cumbersome to study analytically. It is usually
through simulation that difficulties with between-sample
plant response are covered and solved.
When a continuous-time plant is simulated on a digital
computer, its response is computed at closely spaced discrete
times. It is plotted by joining the closely spaced calculated
response values with straight line segments in approximation
658
DISPATCHING
of continuous curve. A digital computer simulation of discretetime control of a continuous-time system involves at least two
sets of discrete-time calculations. One runs at a high rate for
simulation of the continuous-time plant. The other runs at a
lower rate (say once every 10 to 50 of the former calculations)
to generate new control signals at each discrete control step.
BIBLIOGRAPHY
1. M. S. Santina, A. R. Stubberud, and G. H. Hostetter, Digital Control System Design, 2nd ed., Philadelphia: Saunders, 1994.
2. J. J. DiStefano III, A. R. Stubberud, and I. J. Williams, Feedback
and Control Systems (Schaums Outline), 2nd ed., New York:
McGraw-Hill, 1990.
3. B. C. Kuo, Digital Control Systems, 2nd ed., Philadelphia: Saunders, 1992.
4. G. F. Franklin, J. D. Powell, and M. L. Workman, Digital Control
of Dynamic Systems, 2nd ed., Reading, MA: Addison-Wesley,
1990.
5. T. Kailath, Linear Systems, Englewood Cliffs, NJ: Prentice-Hall,
1980.
6. C. T. Chen, Linear System Theory and Design, Philadelphia:
Saunders, 1984.
strom and B. Wittenmark, Computer Controlled Systems,
7. K. J. A
Englewood Cliffs, NJ: Prentice-Hall, 1987.
8. K. Ogata, Discrete-Time Control Systems, Englewood Cliffs, NJ:
Prentice-Hall, 1987.
9. B. Friedland, Control System Design, New York: McGraw-Hill,
1986.
10. W .M. Wonham, Linear Multivariable Control: A Geometric Approach, 3rd ed., New York, Springer-Verlag, 1985.
MOHAMMED S. SANTINA
Boeing Defense and Space Group
ALLEN R. STUBBERUD
University of California, Irvine
264
(3)
u
w=
y
d
x(t), t R
x = f (x, u) where ( x)(t) := dt
(1)
x(t + 1), t Z
y = h(x, u)
(2)
or
u
w = x
y
(4)
where as in Eq. (1), denotes the derivative operator for continuous-time systems and the (backwards) shift operator for
discrete-time systems. The input u(t) and state x(t) of the system at time t belong, respectively, to the input space U m
and state space X n. Moreover,
G : U X,
F:X X
are linear maps; the first is called the input map, while the
second describes the dynamics or internal evolution of the
system.
The output Eq. (2), for both discrete- and continuous-time
linear systems, is composed of a set of linear algebraic equations
y = Hx + Ju
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
(5)
:=
F
H
G
J
, F Rnn , G Rnm , H R pn , J R pm
(6)
(u; x0 ; t) = eF (tt 0 ) x0 +
t0
eF (t ) Gu( ) d , t t0
(8)
t1
F t1 j Gu( j), t t0
(9)
=
F n1 G]
(11)
j=t 0
1
y = HT
x + Ju
(u; x0 ; t) = F tt 0 x0 +
then Eqs. (4) and (5) when expressed in terms of the new
state x, will become
1
x = TFT
TG u,
x +
J :U Y
265
P (t) :=
eF GG eF d , t > 0
(12)
t1
F k GG (F )k , t > 0
(13)
k=0
If we transform the state under a linear change of coordinates, the corresponding matrices describing the system will
change. In particular, if the new state is
x := Tx, det T = 0
(T t )
, Rn , t, T R, 0 t T
(14)
266
(15)
Proposition 1. The reachability Grammians have the following properties: (a) P (t) P *(t) 0, and (b) their columns
span the reachability subspace, that is
im P (t) = im Rn (F, G)
This relationship holds for continuous-time systems for all
t 0, and for discrete-time systems (at least for t n).
Corollary 2
Definition 4. Given
=
=
(16)
reach
In general, reachability is an analytic concept. The previous theorem, however, shows that for linear, finite-dimensional, time-invariant systems, reachability reduces to an algebraic concept depending only on properties of F, G and in
particular on the rank of the reachability matrix R n(F, G),
but independent of time and the input function. It is also
worthwhile to note that Eq. (16) is valid for both continuousand discrete-time systems. This, together with a similar result on observability [see Eq. (23)], has as a consequence the
fact that many tools for studying linear systems are algebraic.
It should be noticed however that the physical significance of
F and G is different for the discrete- and continuous-time
cases; if for instance we discretize the continuous-time system
d
x(t) = Fcontx(t) + Gcontu(t), to
dt
x(t + 1) = Fdiscrx(t) + Gdiscru(t)
then Fdiscr eFcont.
I f ( ) f ( ), t Z
2
f := f, f :=
f ( ) f ( ) d , t R
I
The input function w,T defined by Eqs. (14) and (15) has the
following property:
Proposition 2. Given the reachable state x Xreach, let
u be any input function which reaches x at time T, that is,
(u; 0; T) x. There exists n satisfying:
x = P (T )
for both the continuous- and discrete-time case, Xreach is a linear subspace of X, given by the formula
X reach = im Rn (F, G)
(17)
(18)
=
the nonlinearities of the system are. We conclude this subsection by stating a result on various equivalent conditions for
reachability.
Theorem 2 Reachability Conditions. The following are
equivalent:
1.
=
F nxn, G nm is reachable.
2. The rank of the reachability matrix is full: rank R n(F,
G) n.
3. The reachability Grammian is positive definite, that is,
P (t) 0, for some t 0.
4. No left eigenvector v* of F is in the left kernel of G:
v F = v v G = 0
5. rank(In F G) n, for all
6. The polynomial matrices sI F and G are left coprime.
The fourth and fifth conditions in this theorem are known
as the PHB or PopovHautusBelevich tests for reachability.
The equivalence of the fifth and sixth conditions is a straightforward consequence of the theory of polynomial matrices; it
will not be discussed in this article.
Remark. Reachability is a generic property. This means intuitively that almost every n n, n m pair of matrices F,
G satisfies
rank Rn (F, G) = n
Put in a different way, in the space of all n n, n m pairs
of matrices, the unreachable pairs form a hypersurface of
measure zero.
The next theorem shows that for continuous-time systems,
the concepts of reachability and controllability are equivalent
while for discrete-time systems the latter is weaker. This is
easily seen by considering the system with state equation:
x(t 1) 0. Clearly, for this system all states are controllable, while none is reachable. Often for this reason, only the
notion of reachability is used.
Theorem 3. Given is
=
(a) For continuous-time systems Xcontr X reach. (b) For discretetime systems X reach. X contr; in particular X contr X reach
ker F n.
for some T,
H
HF
HF 2
On (H, F ) :=
(20)
.
.
.
HF n1
Definition 6. Let
=
F
H
eF H HeF d , t > 0
(21)
(22)
Q(t) :=
Remark. It follows from the previous results that for any
two states x1, x2 X reach there exist u12, T12 such that x1
(u12; x2; T12). Since x2 is reachable it is also controllable; thus
there exist u2, T2 such that (u2; x2; T2) 0. Finally, the reach-
267
268
F
H
G
J
Theorem 4. Given
=
F
H
F : X X , G : X U , H : Y X , J : Y U
unobs
is a lin-
(23)
reach
unobs
(X
) = X
On (H, F )x(0) = Y0
(24)
It can also be shown that controllability and reconstructibility are dual concepts. We conclude this subsection by stating the dual to theorem 2.
Theorem 6. Observability conditions. The following are
equivalent:
1.
F
H
F
H
H pn, F nn is observable.
2. The rank of the observability matrix is full: rank
O n(H, F) n.
3. The observability Grammian is positive definite Q (t)
0, for some t 0.
4. No right eigenvector v of F is in the right kernel of H:
Fv = v Hv = 0
5.
In F
rank
H
Proposition 4. Given is
=
for all
=n
Fo Foo
F
= 0 Fo
H
0 Ho
where
There exists a basis in X such that F, G have the following
matrix representations:
Fr
= 0
Gr
0
Frr
Fr
(26)
:=
Fr
Gr
is reachable.
Thus every system
Fr
Gr
Fo
Ho
is observable.
The reachable and observable canonical decompositions
given in lemmas 1 and 2 can be combined to obtain the following decomposition of the triple (H, F, G):
Lemma 3. Reachable-observable canonical decomposition.
Given
F G
=
H
there exists a basis in X such that F, G, and H have the following matrix representations
269
=
F
H
Fro
0
0
0
0
F12
Fro
0
0
Hro
F13
0
Fro
0
0
Fro
Hro
Gro
F14
F24
F34
Fro
Hro
Gro
Gro
0
0
(27)
r
Fr
ro
:=
F
G
270
c
F
H
Eo2 = x Qx
P :=
eF GG eF d ,
Q :=
eF H HeF d
(28)
F Q + QF + H H = 0
(29)
d
F
H
F t1 GG (F )t1 ,
t>0
Q :=
(F )t1 H HF t1
(30)
t>0
is stable, that is, all eigenvalues of F are inside the unit disk,
the grammian Eqs. (13) as well as (22) are defined for t
P :=
(34)
F QF + H H = Q
(31)
Recall Eqs. (18), (19), and (24), valid for both discrete- and
continuous-time systems. From the definition of the grammians follows that:
P (t2 ) P (t1 ), Q(t2 ) Q(t1 ), t2 t1
(32)
irrespective of whether we are dealing with discrete- or continuous-time systems. Hence from Eq. (19) it follows that the
minimal energy Er, required for the transfer of state 0 to x, is
obtained as the alloted time T tends to infinity. Assuming
reachability, this minimal energy is:
Er2 = x P 1 x
(33)
(35)
w (t) =
w 1 (t) for t < 0
w 2 (t) for t > t
(36)
It relates
w
B
a
to its shifts or derivatives. This equation can be written explicitly in terms of the coefficient matrices of R and M. Let
R(s) :=
271
1
Ri si , Ri R pq , M(s) :=
i=0
2
Mi si , Mi R pr
i=0
Ri ( i w)(t) =
2
Mi ( i a)(t)
i=0
w
B
a
Eq. (35) is referred to as an annihilating behavioral equation.
The special case of SV Eq. (4) is described by w u, a x,
R() G and M(s) I F; while that of input/output equations is described by
u
w=
y
a is nonexistent, and R() [N() D()], where D is a
square, nonsingular polynomial matrix of size p.
A further important aspect at which the behavioral formalism departs from, and generalizes, the classical formalism is
that related to controllability; controllability becomes namely
an attribute of the system (i.e., of a collection of trajectories)
as opposed to an attribute of a system representation (i.e., of
equations generating these trajectories).
Roughly speaking, a system is controllable if its behavior
has the property: whatever the past history (trajectory), it can
always be steered to any desired future trajectory. More precisely, a dynamical system with behavior B is said to be controllable, if for any w1, w2 B , there exists a t 0 and a
w B such that
Notice that the first condition stated here is the same as condition 5 of theorem 2. Thus the behavioral definition of controllability provides a generalization of the classic concept.
Furthermore the second condition provides a way of defining
controllability without the definition of state.
Examples
1. The previous issues are illustrated by means of two examples. First, consider an RLC circuit composed of the
parallel connection of two branches: the first branch is
composed of an inductor L in series with a resistor RL;
the second is composed of a capacitor C in series with a
resistor RC. The driving force u is a voltage source applied to the two branches. Let the state variables x1, x2
be the current through the inductor, the voltage across
the capacitor respectively. The state equations are
RL
1
d
R
L
x =
x + u
0
dt 1
L 1 L
L
F=
1
d
1
1
x =
x +
u
RC C
dt 2
RCC 2 RCC
L
G=
1
RCC
Since this is a continuous-time system, reachability and
controllability are equivalent notions. The reachability
matrix is
L
R2 = [G FG] = 1
RCC
RL
2
L
1
2 2
RCC
1
RCCL
RL
1
L
RCC
= 0
272
g21
(1 e2 L T )
2L
P (T ) =
g1 g2
(1 e( L + C )T )
L + C
Controllability: Continuous-Time
In continuous time, we consider systems of differential equations of the following general form:
g1 g2
(1 e( L + C )T )
L + C
g22
(1 e2 C T )
2C
where
L =
RL
1
1
, C =
, g = , g 2 = C
L
RCC 1
L
g21
g1 g2
L + C
2L
P =
2
g2
g1 g2
L + C
2C
and it can be verified that FP P F* GG* 0.
Assume now that the variable observed y is the sum
of the voltages across the capacitor and across the resistor RL:
y = RL x1 + x2 H = [RL
d
x(t) = f (x(t), u(t))
dt
x1
x1
x x
x = 2 = 2
x 3
x4
1]
1
R
L
O2 =
1 det O2 = RL RC LC det R2
R2L
L
RC C
Thus reachability and observability are lost simultaneously. If this happens, then one can reach any given
state x1 and x2 L/RL x1; while only the linear combination RLx1 x2 can be deduced, but not x1 or x2 individually.
2. Consider the system given in input/output form by the
equation
d
d
y=
u, u(t), y(t) R
dt
dt
We will show that this system is not controllable.
This equation can be rewritten as d/dt v 0, where
v y u. All trajectories composing the behavior of
this system are constants. But a trajectory defined by
v(t) c1, t T, and v(t) c2, t T, c1 c2, does not
belong to the behavior, as its derivative is not zero.
Hence, since the trajectories of this system are not concatenable, the conclusion follows.
(37)
x2
(x1,x2)
We turn attention now to nonlinear systems. Both continuous- and discrete-time systems will be discussed.
0
cos( + )
0
sin( + )
d
x = u1 + u2
0
sin
dt
1
0
(38)
273
2 [ f,g]
+ o(t 2 )( )
(40)
tg t f
tg
tf
Observe that (0) and that the values of (t) are all in the
set of points S attainable by positive and negative time solutions of the differential equations corresponding to f and g;
the above expansion implies that d/dt (0) [f, g](), that is,
there is a curve in S whose tangent is the Lie bracket of the
two vector fields. Thus, Lie brackets provide new directions of
infinitesimal movement in addition to those corresponding to
f and g themselves (and their linear combinations).
Given now a system by Eq. (37), we consider, for each possible control value u m, the following vector field:
f u : Rn Rn : x f (x, u)
{( , u) Rn |u Inp}
Definition. The system Eq. (37) is accessible from the state
if the reachable set R () contains an open subset of n (that
is, it has a nonempty interior).
(39)
274
0
cos( + )
0
sin( + )
g1 = , g2 =
0
sin
1
0
Computing some brackets, we get the two new vector fields
[again, we borrow our terminology from (11)]:
sin( + )
cos( + )
cos
0
and
(41)
for some vector fields gis. It is easy to verify that, for such
systems, L is the Lie algebra generated by taking all possible
iterated Lie brackets starting from the gis.
For example, consider a linear system
d
x = Fx + Gu
dt
Here g0(x) Fx is a linear vector field and the gi(x)s are the
constant vector fields defined by the m columns of the matrix
G. It then follows that, for each state , L () is the span of
the vector F together with the columns of G, FG, . . .,
Fn1G. In particular, for 0, one has that L (0) is the same
as the right-hand side of Eq. (39). Seen in that context, the
following result, which is valid in general, is not surprising:
Theorem. The system Eq. (37) is accessible from if and
only if the accessibility rank condition holds at .
There is a subclass of systems for which far stronger conclusions can be drawn. This subclass includes all purely kinematic mechanical models. It is the class of affine systems
without drift, that is, systems affine in u (as in Eq. 41, but
for which, in addition, g0 0). We say that a system is completely controllable if R () n for every n, that is to
say, every state can be steered to every other state by means
of an appropriate control action.
Theorem. A system affine without drift is completely controllable if and only if the accessibility rank condition holds
at every state.
sin
cos
0
0
(The bracket [wriggle, steer] equals drive, so it is redundant,
in so far as checking the accessibility rank condition is concerned.) It turns out that these four brackets are enough to
satisfy the accessibility test. Indeed, one computes
det(steer, drive, wriggle, slide) 1
so there is accessibility from every state. Moreover, since this
system is affine without drift, it is completely controllable. (Of
course, it is quite obvious from physical reasoning, for this
example, that complete controllability holds.)
Consider in particular the problem of accessibility starting
from the special state 0 (corresponding to the problem of
exiting from a parallel parked spot). For 0, wriggle
is the vector (0, 1, 1, 0), a mix of sliding in the x2 direction
and a rotation, and slide is the vector (0, 1, 0, 0) corresponding to sliding in the x2 direction. This means that one can in
principle implement infinitesimally both of these motions.
The wriggling motion is, based on the characterization of Lie
brackets mentioned earlier, the one that arises, in a limiting
sense, from fast repetitions of the following sequence of four
basic actions:
steer drive reverse steer reverse drive ( )
The Car Example. In the notations for systems affine in controls, we have the vector fields g1 and g2, which we call (follow-
1
0
d
+ u2
x = u1
dt
0
(x1 )
275
Controllability: Discrete-Time
(42)
=0
1
f u f u+
e (x)
i
d
x = x21 x2 + sin x1
dt 1
d
x = x1 ex 2 + u cos x1
dt 1
because, up to first order around 0 one has x12x2 sin
x1 x1 h1, x1ex2 u cos x1 x1 u h1, where h1 and
h2 are higher-order terms in states and controls, which means
that the linearization at the equilibrium 0 is the linear
system with matrices
0 1
0
F=
, G=
1 0
1
which is controllable. This is only a sufficient condition. the
system d/dt x u3, in dimension one, is clearly (even) completely controllable, but its linearization at 0 gives the
noncontrollable linear system d/dt x 0. A necessary and
sufficient condition does exist linking accessibility and linear
controllability, but it is more subtle. It is illustrated next, for
simplicity, only for discrete-time systems. In continuous time,
an analogous result holds, but it is slightly more complicated
to state and prove (12).
Observe that accessibility from corresponds to the requirement that the union of the images of the composed maps
f k ( , ) : (Rm )k Rn k 0
cover an open subset, where we are denoting
f k (x, (u1 , . . ., uk )) := f ( f (. . . f ( f (x, u1 ), u2 ), . . ., uk1 ), uk )
for every state x and sequence of controls u1, . . . uk. A simple
argument, based on a standard result in analysis (Sards the-
276
f [xt , ut ],
x
Gt =
f [xt , ut ]
u
Then, accessibility is equivalent to the existence of some sequence of controls u1, . . . uk for which this linearization is
controllable as a time-varying linear system.
Observability: Continuous-Time
We present a brief outline of a nonlinear observability test,
for the special case of continuous-time systems affine in Eq.
(41), with an output map h : n p added to the system description. Two states and are distinguishable by input/
output experiments if there is at least some input which,
when applied to the system in initial state , gives a different
output than when applied to the system in state . An observable system is one with the property that every pair of distinct states is distinguishable. Thus an observable system is
one for which, at least in principle, it is possible to distinguish
between internal states by means of input/output measurements alone.
Consider the vector space spanned by the set of all functions of the type
Lg i . . . Lg i h j (x)
1
(43)
(44)
= LX LX . . . LX h j ( )
1
where Xl(x) g0(x) uilgi(x). This expression is a multilinear function of the uils, and a further derivation with respect
to these control value coordinates shows that the generators
in Eq. (43) must coincide at x1 and x2. It turns out that (for
analytic vector fields, as considered in this exposition), separability by O is necessary as well as sufficient, because Eq.
(44) can be expressed as a power series in terms of the generators of Eq. (43). Thus, observability is equivalent to separation by the functions in O .
The observability rank condition at a state is the condition that the dimension of the span of
m
i1
(45)
(46)
where the input values u(t), state x(t), and output values y(t)
are elements of, for instance, Hilbert spaces. For example, if
we consider the heat equation
2x
x
(t, ) = 2 (t, ) + G( )u(t)
t
u(t)
v(t h)
v(t)
G0
+
+ +
ehs
1/s
277
y(t)
H0
F0
F1
d
v(t) = F0 v(t) + F1 v(t h) + G0 u(t)
dt
y(t) = H0 v(t h), v(t) R n
(47)
#
$ # $
F1 Eh
G0
v(t)
+
u(t)
0
zt ( )
=: Fx(t) + Gu(t)
#
$
x(t)
y(t) = [0 H0 Eh ]
, zt () (L2 [h, 0])n
zt ()
#
$
F0
d v(t)
=
dt zt ()
0
(48)
(49)
where x(t) : [v(t) zt( )], and zt( ) is the state in the delay
element ehs which is related to v(t) via zt() v(t ). Eh
denotes the point evaluation operator Ehz( ) : z(h). The
domain of F in Eq. (48) is [v z( )] R n (H2(h, 0))n :
z(0) v. Here the input operator G is bounded but the output
operator C [0 C0Eh] is not. (Point evaluation in L2 cannot
be continuous.)
We thus consider the abstract system Eqs. (45), (46) with
x(t) in a Hilbert space X, with input/output variables u(t) and
y(t) being m- and p-valued. The operators G : m X and
H : X p are often assumed to be bounded. However, there
are many systems that do not satisfy this property: G may
not take values in X or H may be only densely defined. The
system described by Eqs. (48), (49) is an example of the latter
case. For brevity of exposition, however, we will not be very
rigorous about this point in what follows. Also, since J plays
no role in controllability/observability, we will assume J 0.
Denote this system by (X, F, G, H).
The solution of Eq. (45) can be written as
x(t) = S(t)x0 +
S(t )Gu( ) d
0
278
S(T )Gu( )d
(50)
range (T )
T >0
range (T )
(51)
T >0
t0
(53)
2
where Lloc
[0, ) is the space of locally square integrable functions. Then is observable if and only if is one-to-one.
For finite-dimensional systems these observability notions
imply the following consequences:
(T )u :=
range (T )
(52)
T >0
0tT
%
T >0
L2 [0, T] X : u
S(t)Gu(t) dt
(54)
t0
(55)
complete with suitable feedback. Since feedback does not alter controllability, this latter condition preserves controllability.
Similar spectral tests are possible for systems with spectral F operators. For example, consider the case where F is
the self-adjoint operator defined by a spectral representation
(32):
S(t) Met
Fx =
n=1
then it is possible to relate controllability/observability to Lyapunov equations. In this case it is possible to extend the domain of and restrict the codomain of to L2[0, ). Define
P := =
Q := =
S(t)GG S (t) dt
Gu :=
x1 , x2 D(F )
(x, n j )n j
j=1
m
gi u i ,
gi X
i=1
rn
S (t)H HS(t) dt
Then is approximately controllable if and only if P is positive definite, and observable if and only if Q is positive definite. P and Q are called controllability and observability
Grammians. Actually, when is observable, Q is a unique
self-adjoint solution to the Lyapunov equation
279
(gm , n 1 )
(g1 , n 1 )
..
..
= rn
rank
.
.
(g1 , n r n ) (gm , n r n )
Remarks
Controllability/observability questions for distributed parameter systems were first addressed by Fattorini (33). Since then
numerous papers have appeared, and the literature is too
vast to be listed or surveyed here. For developments up to
1978, consult Russels survey (34). The recent textbook by
Curtain and Zwart (32) gives an extensive set of further references. For abstract operator settings of linear systems, see
Refs. 35, 36, and 37. There is also a vast amount of literature
concerning controllability of systems described by partial differential equations. For the developments along this line, consult Refs. 28, 29, and 30. We discuss here some limited references that deal with further related subjects.
Controllability/observability for delay systems has received considerable attention: see the work by Datko, Delfour,
Langenhop, Kamen, Manitius, Mitter, OConnor, Pandolfi,
Salamon, Triggiani, Yamamoto, and others [references cited
in (32)]. The spectral condition for approximate controllability
of retarded delay systems cited previously is due to the work
of Manitius and Triggiani (38,39). It is also extended to various situations including neutral systems; see, for example,
Refs. 40, 41, and 42.
Controllability and observability are also important in realization theory. It was shown by Baras, Brockett, and Fuhrmann (43) that an approximately controllable and observable
realization need not be unique. One needs a stronger condition to guarantee uniqueness; the results cited in Ref. 31 are
one example.
As with finite-dimensional systems, controllability and observability are closely related to the question of coprimeness
of factorizations of transfer functions. For example, a factorization is (left) coprime (in an appropriately defined sense) if
a certain system associated with this factorization is approximately controllable (see, for example, Ref. 44). However, in
280
contrast to the finite-dimensional context, there is a large variety of freedom in choosing the algebra over which the factorization is considered. Typical examples are the Callier-Desoer
algebra, the space of H functions, and the algebra of distributions with compact support. Each different choice leads to
a different theory of realization/stabilization/controller parameterization as is the case with the finite-dimensional theory, but with much wider freedom. Each has its advantage in
different contexts, reflecting the variety of distributed parameter systems. The theories can therefore hardly be expected
to be complete.
Another topic that is in close connection with
controllability/observability is the existence and uniqueness
of solutions to Riccati equations. This topic is also related
with H control theory and under active research. The current
trend is in spectral factorization in an abstract functional
equation setting. To this end, various questions concerning
the well-posedness of system equations and transfer functions
arise and are currently under study.
BIBLIOGRAPHY
1. J. W. Polderman and J. C. Willems, Introduction to Mathematical
System Theory: A Behavioral Approach, New York: SpringerVerlag, 1998.
2. R. E. Kalman, On the general theory of control systems, Proc.
1st Int. Cong. IFAC, Moscow 1960, London: Butterworths, 1960,
pp. 481492.
3. R. E. Kalman, Y. C. Ho, and K. Narendra, Controllability of linear dynamical systems, Contributions to Differ. Equations, 1:
189213, 1963.
4. R. E. Kalman, Lectures on controllability and observability,
Centro Internazionale Matematico Estrio, Bologna, 1968.
5. R. E. Kalman, P. L. Falb, and M. A. Arbib, Topics in Mathematical System Theory, New York: McGraw-Hill, 1969.
6. E. D. Sontag, Mathematical Control Theory: Deterministic Finite
Dimensional Systems, 2nd ed., New York: Springer-Verlag, 1998.
7. W. J. Rugh, Linear System Theory, 2nd ed., Upper Saddle River,
NJ: Prentice-Hall, 1996.
8. K. Zhou, J. C. Doyle, and K. Glover, Robust and Optimal Control,
Upper Saddle River, NJ: Prentice-Hall, 1996.
9. W. L. Brogan, Modern Control Theory, Upper Saddle River, NJ:
Prentice-Hall, 1991.
10. J. C. Willems, Paradigms and puzzles in the theory of dynamical
systems, IEEE Trans. Autom. Control, AC-36: 259294, 1991.
11. E. Nelson, Tensor Analysis, Princeton, NJ: Princeton University
Press, 1967.
12. E. D. Sontag, Universal nonsingular controls, Syst. Control Lett.,
19: 221224, 1992.
ber Systeme von linearen partiellen differen13. W. L. Chow, U
tialgleichungen erster ord-nung, Math. Ann., 117: 98105, 1939.
14. R. Hermann, On the accessibility problem in control theory, in
Int. Symp. Differ. Equations and Mech., New York: Academic
Press, 1973.
15. C. Lobry, Controllabilite des syste`mes non lineaires, SIAM J.
Contr., 8: 573605, 1970.
16. H. J. Sussmann and V. Jurdjevic, Controllability of nonlinear
systems, J. Differ. Equations 12: 95116, 1972.
17. A. Krener, A generalization of Chows theorem and the bang-bang
theorem to nonlinear control systems, SIAM J. Control, 12: 43
52, 1974.
A. C. ANTOULAS
Rice University
E. D. SONTAG
Rutgers University
Y. YAMAMOTO
Kyoto University
SERVOMECHANISMS.
281
194
x k+1 = f (xx k , k; p ),
NONLINEAR DYNAMICS
Unlike linear systems, many nonlinear dynamical systems do
not show orderly, regular, and long-term predictable responses to simple inputs. Instead, they display complex, random-like, seemingly irregular, yet well-defined output behaviors. This dynamical phenomenon is known as chaos.
The term chaos, originating from the Greek word ,
was designated as the primeval emptiness of the universe
before things came into being of the abyss of Tartarus, the
underworld. . . . In the later cosmologies Chaos generally
designated the original state of things, however conceived.
The modern meaning of the word is derived from Ovid, who
saw Chaos as the original disordered and formless mass, from
which the maker of the Cosmos produced the ordered universe (1). There also is an interpretation of chaos in ancient
Chinese literature, which refers to the spirit existing in the
center of the universe (2). In modern scientific terminology,
chaos has a fairly precise but rather complicated definition by
means of the dynamics of a generally nonlinear system. For
example, in theoretical physics, chaos is a type of moderated
randomness that, unlike true randomness, contains complex
patterns that are mostly unknown (3).
Bifurcation, as a twin of chaos, is another prominent phenomenon of nonlinear dynamical systems: Quantitative
change of system parameters leads to qualitative change of
system properties such as the number and the stability of system response equilibria. Typical bifurcations include transcritical, saddle-node, pitchfork, hysteresis, and Hopf bifurcations. In particular, period-doubling bifurcation is a route to
chaos. To introduce the concepts of chaos and bifurcations as
well as their control (4,5), some preliminaries on nonlinear
dynamical systems are in order.
Nonlinear Dynamical Systems
A nonlinear system refers to a set of nonlinear equations,
which can be algebraic, difference, differential, integral, functional, and abstract operator equations, or a certain combination of these. A nonlinear system is used to describe a physical device or process that otherwise cannot be well defined by
a set of linear equations of any kind. The term dynamical
system is used as a synonym of mathematical or physical system, in which the output behavior evolves with time and/or
other varying system parameters (6).
In general, a continuous-time dynamical system can be described by either a differential equation
x = f (xx, t; p ),
t [t0 , )
(1)
F : x g (xx, t; p ),
t [t0 , )
(2)
(3)
or a map
F : x k g (xx k , k; p ),
k = 0, 1, . . .
(4)
t [t0 , )
(5)
Classification of Equilibria
For illustration, consider a general two-dimensional autonomous system
x = f (x, y)
(6)
y = g(x, y)
with given initial conditions (x0, y0), where f and g are two
smooth nonlinear functions that together describe the vector
field of the system.
The path traveled by a solution of Eq. (6), starting from
the initial state (x0, y0), is a solution trajectory, or orbit, of the
system and is sometimes denoted by t(x0, y0). For autonomous systems, two different orbits will never cross each other
(i.e., never intersect) in the x-y plane. This x-y coordinate
plane is called the (generalized) phase plane (phase space in
the higher-dimensional case). The orbit family of a general
autonomous system, corresponding to all possible initial conditions, is called solution flow in the phase space.
Equilibria, or fixed points, of Eq. (6), if they exist, are the
solutions of two homogeneous equations:
f (x, y) = 0
or a map
k = 0, 1, . . .
and g(x, y) = 0
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
Stable node
Unstable node
Stable focus
Unstable focus
Saddle
Center
point and center, as summarized in Fig. 1. The type of equilibria is determined by the eigenvalues, 1,2, of the system Jacobian
J :=
fx
gx
fy
gy
195
The most basic problem in studying the general nonlinear dynamical system of Eq. (1) is to understand and/or to solve for
the system solutions. The asymptotic behavior of a system
solution, as t , is called the steady state of the solution,
while the solution trajectory between its initial state and the
steady state is the transient state.
For a given dynamical system, a point x in the state space
is an -limit point of the system state orbit x(t) if, for every
open neighborhood U of x, the trajectory of x(t) will enter U
at a (large enough) value of t. Consequently, x(t) will repeatedly enter U infinitely many times, as t . The set of all
such -limit points of x(t) is called the -limit set of x(t) and
is denoted x. An -limit set of x(t) is attracting if there exists
an open neighborhood V of x such that whenever a system
orbit enters V then it will approach x as t . The basin of
attraction of an attracting set is the union of all such open
neighborhoods. An -limit set is repelling if the nearby system orbits always move away from it. An attractor is an limit set having the property that all orbits nearby have it as
their -limit sets.
For a given map, F, and a given initial state, x0, an -limit
set is obtained from the orbit Fk(x0) as k . This -limit
set, x, is an invariant set of the map, in the sense that
F(x) x. Thus, -limit sets include equilibria and periodic
orbits.
Poincare Maps
Assume that the general n-dimensional nonlinear autonomous system of Eq. (5) has a tp-periodic limit cycle, , and let
x
x
=J
y
y
in a neighborhood of the equilibrium (x, y).
196
(a)
(b)
(c)
(d)
(e)
(f)
x* be a point on the limit cycle and be an (n 1)-dimensional hyperplane transversal to at x*, as shown in Fig. 3.
Here, the transversality of to at x* means that and the
tangent line of at x* together span the entire n-dimensional
space (hence, this tangent line of cannot be tangent to at
x*). Since is tp-periodic, the orbit starting from x* will return to x* in time tp. Any orbit starting from a point, x, in a
small neighborhood U of x* on will return and hit at a
point, denoted P(x), in the vicinity V of x*. Therefore, a map
P : U V can be uniquely defined by , along with the solution flow of the autonomous system. This map is called the
Poincare map associated with the system and the cross section . For different choices of the cross section , Poincare
maps are similarly defined. Note that a Poincare map is only
locally defined and is a diffeomorphismnamely, a differentiable map that has an inverse and the inverse is also differentiable. If a cross section is suitably chosen, the orbit will
repeatedly return and pass through the section. The Poincare
map together with the first return orbit is particularly important, which is called the first return Poincare map. Poincare
maps can also be defined for nonautonomous systems in a
similar way, where, however, each return map depends on
the initial time in a nonuniform fashion.
P (x)
x*
x
x*
(a)
x1
x2
(b)
197
t t0
(8)
xx (t) 0
as t
(9)
x (t0 ) = x 0
(10)
0 t < tp }
x (t0 ) = x 0
(7)
d(t ( x 0 ), ) < ,
t t0
Lyapunov Stability Theorems. Two cornerstones in the Lyapunov stability theory for dynamical systems are the Lyapunov first (or indirect) method and the Lyapunov second (or
direct) method.
198
mathematical definition of chaos does not seem to be available anytime soon, some fundamental features of chaos are
well received and can be used to signify or identify chaos in
most cases.
xkk0
, generated by the iterations of the map starting from
any given initial state x0, is defined to be
i (xx0 ) = lim
i < 0, i = 1, . . ., n
Features of Chaos
1
ln |i (Jk (xx k ) . . . J0 (xx 0 ))|,
k
i = 1, . . ., n
(11)
1
ln xx (t; x 0 )/xx 0
t
199
stable equilibrium
1 = 0, i < 0, i = 2, . . ., n
1 = 2 = 0, i < 0, i = 3, . . ., n
stable two-torus
1 = = m = 0,
i < 0, i = m + 1, . . ., n
stable m-torus
Here, a two-torus is a bagel-shaped surface in three-dimensional space, and an m-torus is its geometrical generalization
in (m 1)-dimensional space.
It is now well known that one and two-dimensional continuous-time autonomous dynamical systems cannot produce
chaos. For a three-dimensional continuous-time autonomous
system, the only possibility for chaos to exist is that the three
Lyapunov exponents are
(+, 0, ) := (1 > 0, 2 = 0, 3 < 0) and 3 < 1
Intuitively, this means that the system orbit in the phase
space expands in one direction but shrinks in another direction, thereby yielding many complex (stretching and folding)
dynamical phenomena within a bounded region. The discretetime case is different, however. A prominent example is the
one-dimensional logistic map, discussed in more detail later,
which is chaotic but has (the only) one positive Lyapunov exponent. For four-dimensional continuous-time autonomous
systems, there are only three possibilities for chaos to emerge:
1. (, 0, , ): 1 0, 2 0, 4 3 0; leading to chaos
2. (, , 0, ): 1 2 0, 3 0, 4 0; leading to
hyperchaos
3. (, 0, 0, ): 1 0, 2 3 0, 4 0; leading to a
chaotic two-torus (this special orbit has not been
experimentally observed).
Simple Zero of the Melnikov Function. The Melnikov theory
of chaotic dynamics deals with the saddle points of Poincare
maps of continuous solution flows in the phase space. The
Melnikov function provides a measure of the distance between the stable and unstable manifolds near a saddle point.
To introduce the Melnikov function, consider a nonlinear
oscillator described by the Hamiltonian system
+ f1
p =
q
H
q =
+ f2
p
where f : [f 1(p, q, t), f 2(p, q, t)] has state variables (p(t),
q(t)), 0 is small, and H H(p, q) EK EP is the Hamilton function for the undamped, unforced (when 0) oscillator, in which EK and EP are the kinetic and potential energy
of the system, respectively. Suppose that the unperturbed
(unforced and undamped) oscillator has a saddle-node equilibrium (e.g., the undamped pendulum) and that f is tp-periodic
with phase frequency . If the forced motion is described in
200
the three-dimensional phase space (p, q, t), then the Melnikov function is defined by
F (d ) =
[H( p, q)] f dt
(12)
consider k1
dk for different coverings, and let infC dk be the
smallest value of the sum over all possible such coverings. In
the limit 0, this value will diverge if h but tends to
zero if h for some constant h 0 (need not be an integer).
This value, h, is called the Hausdorff dimension of the set S.
If such a limit exists for h, then the Hausdorff measure
of the set S is defined to be
h (S) := lim inf
0 C
dk
k=1
There is an interesting conjecture that the Lyapunov exponents k (indicating the dynamics) and the Hausdorff dimension h (indicating the geometry) of a strange attractor have
the relation
h=k+
k
k+1
i=1
i ,
1 2 n
k
i 0. This
where k is the largest integer that satisfies i1
formula has been mathematically proved for large families of
three-dimensional continuous-time autonomous systems and
of two-dimensional discrete-time systems.
A notion that is closely related to the Hausdorff dimension
is fractal. Fractal was first coined and defined by Mandelbrot
in the 1970s to be a set with Hausdorff dimension strictly
greater than its topological dimension (which is always an
integer). Roughly, a fractal is a set that has a fractional
Hausdorff dimension and possesses certain self-similarities.
An illustration of the concept of self-similarity and fractal is
given in Fig. 13. There is a strong connection between fractal
and chaos. Chaotic orbits often possess fractal structures in
the phase space. For conservative systems, the KolmogorovArnold-Moser (KAM) theorem implies that the boundary between the region of regular motion and that of chaos is fractal. However, some chaotic systems have nonfractal limit sets,
and many fractal structures are not chaotic.
Pk ln(Pk )
n1
201
For example, a pendulum controlled by a proportional-derivative controller can behave chaotically when the tracking signal is periodic, with energy dissipation, even for the case of
small controller gains. In addition, chaos has been found in
many engineering applications, such as design of control circuits for switched-mode power conversion equipment, highperformance digital robot controllers, second-order systems
containing a relay with hysteresis, and various biochemical
control systems.
Chaos also occurs frequently in discrete-time feedback control systems due to sampling, quantization, and roundoff effects. Discrete-time linear control systems with dead-zone
nonlinearity have global bifurcations, unstable periodic orbits, scenarios leading to chaotic attractors, and crises of chaotic attractors changing to periodic orbits. Chaos also exists in
digitally controlled systems, feedback types of digital filtering
systems (either with or without control), and even the linear
Kalman filter when numerical truncations are involved.
Many adaptive systems are inherently nonlinear, and thus
bifurcation and chaos in such systems are often inevitable.
The instances of chaos in adaptive control systems usually
come from several possible sources: the nonlinearities of the
plant and the estimation scheme, external excitation or disturbances, and the adaptation mechanism. Chaos can occur
in typical model-referenced adaptive control (MRAC) and selftuning adaptive control (STAC) systems, as well as some
other classes of adaptive feedback control systems of arbitrary
order that contain unmodeled dynamics and disturbances. In
such adaptive control systems, typical failure modes include
convergence to undesirable local minima and nonlinear selfoscillation, such as bursting, limit cycling, and chaos. In indirect adaptive control of linear discrete-time plants, strange
system behaviors can arise due to unmodeled dynamics (or
disturbances), bad combination of parameter estimation and
control law, and lack of persistency of excitation. For example, chaos can be found in set-point tracking control of a linear discrete-time system of unknown order, where the adaptive control scheme is either to estimate the order of the plant
or to track the reference directly.
Chaos also emerges from various types of neural networks.
Similar to biological neural networks, most artificial neural
networks can display complex dynamics, including bifurcations, strange attractors, and chaos. Even a very simple recurrent two-neuron model with only one self-interaction can produce chaos. A simple three-neuron recurrent neural network
can also create period-doubling bifurcations leading to chaos.
A four-neuron network and multineuron networks, of course,
have higher chances of producing complex dynamical patterns
such as bifurcations and chaos. A typical example is cellular
neural networks, which have very rich complex dynamical behaviors.
Chaos has also been experienced in some fuzzy control systems. The fact that fuzzy logic can produce complex dynamics
is more or less intuitive, inspired by the nonlinear nature of
the fuzzy systems. This has been justified not only experimentally but also both mathematically and logically. Chaos has
been observed, for example, from a coupled fuzzy control system. The change in the shapes of the fuzzy membership functions can significantly alter the dynamical behavior of a fuzzy
control system, potentially leading to the occurrence of chaos.
Many specific examples of chaos in control systems can be
given. Therefore, controlling chaos is not only interesting as
202
a subject for scientific research but also relevant to the objectives of traditional control engineering. Simply, it is not an
issue that can be treated with ignorance or neglect.
x2 = p
BIFURCATIONS
Associated with chaos is bifurcation, another typical phenomenon of nonlinear dynamical systems that quantifies the
change of system properties (such as the number and the stabilities of the system equilibria) due to the variation of system
parameters. Chaos and bifurcation have a very strong connection; often they coexist in a complex dynamical system.
Basic Types of Bifurcations
To illustrate various bifurcation phenomena, it is convenient
to consider the two-dimensional, parametrized, nonlinear dynamical system
x = f (x, y; p)
(14)
y = g(x, y; p)
x1 = x1
x2 = p + x2 x32
has equilibria
x1 = 0 and
p x2 + x32 = 0
x2 = p
t
t
Figure 7. The transcritical bifurcation.
x2
p0
p0
203
x (t0 ) = x 0
x
u
x = A(p)x + B(p)u
y = C(p)xx
u = g (yy; p)
(16)
p < p0
p = p0
p > p0
(15)
where x Rn, p is a real variable parameter, and f is differentiable. The most fundamental result on the Hopf bifurcation of this system is the following theorem, which is stated
here for the special two-dimensional case.
d{(p)}
>0
dp
p= p
p
p
p = p0
y
y
x
204
{ ( ;p)}
( ;p)
( ;p)
where
y(t) y +
n
yk e jkt
k=0
where y is the equilibrium solution and the complex coefficients, yk, are determined as follows. For the approximation
with n 2, first define an auxiliary vector
1 ()
=
p)]h
h1
ll T [H( j ;
l Tr
2 g (y; p)
D2 =
y2 y=y
3g (y; p)
D3 =
y3 y=y
1
p)]
1 G(0; p)D
z 02 = [1 + H(0; p)J(
2r r
4
1
p)]
1 H(2 j ;
z 22 = [1 + H(2 j ;
p)D
2r r
4
y0 = z 02 p p0
1/2
y2 = z 22 p p0
{ ( ;p)}
2 1()
I {1 ( j )}
{1 ( j )}
d
d
det
= 0
(;
p)
=
(;
p)
=
I
d
d
(17)
y1 = r | p p0
P = 1 + 2 1( )
( j0 ; p0 ) = 1 + j0,
205
1.00
0.00
(18)
xk+1 = pxk (1 xk )
p2 2p 3)/(2p)
xn
1.0
0.5
0.0
2.8 3.0
1+ 6
1+ 8
3.56994
4.0
1.00
2.00
3.00
2.50
2.88
3.25
p
3.63
4.00
Figure 14 shows the Lyapunov exponent versus the parameter p, in the interval [2.5, 4]. This figure corresponds to
the period-doubling diagram shown in Fig. 13.
The most significant discovery about the phenomenon of
period-doubling bifurcation route to chaos is Feigenbaums
observation in 1978: The convergence of the period-doubling
bifurcating parameters has a geometric rate, p pk k,
where
pk+1 pk
= 4.6692 . . .
pk+2 pk+1
(k )
This is known as a universal number for a large class of chaotic dynamical systems.
Bifurcations in Control Systems
Not only chaos but also bifurcations can exist in feedback and
adaptive control systems. Generally speaking, local instability
and complex dynamical behavior can result from feedback
and adaptive mechanisms when adequate process information is not available for feedback transmission or for parameter estimation. In this situation, one or more poles of the linearized closed-loop transfer function may move to cross over
the stability boundary, thereby causing signal divergence as
the control process continues. However, this sometimes may
not lead to global unboundedness, but rather, to self-excited
oscillations or self-stabilization, creating very complex dynamical phenomena.
Several examples of bifurcations in feedback control systems include the automatic gain control loop system, which
has bifurcations transmitting to Smale horseshoe chaos and
the common route of period-doubling bifurcations to chaos.
Surprisingly enough, in some situations even a single pendulum controlled by a linear proportional-derivative controller
can display rich bifurcations in addition to chaos.
Adaptive control systems are more likely to produce bifurcations than a simple feedback control system due to changes
of stabilities in adaptation. The complex dynamics emerging
from an adaptive control system are often caused by estimation instabilities. It is known that certain prototypes of
MRAC systems can experience various bifurcations.
206
Bifurcation theory has been employed for analzying complex dynamical systems. For instance, in an MRAC system, a
few pathways leading to estimator instability have been identified via bifurcation analysis:
1. A sign change in the adaptation law, leading to a reversal of the gradient direction as well as an infinite linear
drift.
2. The instability caused by high control gains, leading to
global divergence through period-doubling bifurcations.
3. A Hopf bifurcation type of instability, complicated by a
number of nonlocal phenomena, leading to parameter
drift and bursting in a bounded regime through a sequence of global bifurcations.
Both instabilities of types 1 and 2 can be avoided by gain
tuning or simple algorithmic modifications. The third instability, however, is generally due to the unmodeled dynamics and
a poor signal-to-noise ratio, and so cannot be avoided by simple tuning methods. This instability is closely related to the
presence of a degenerate set and a period-two attractor.
Similarly, in the discrete-time case, a simple adaptive control system can have rich bifurcation phenomena, such as period-doubling bifurcation (due to high adaptive control gains)
and Hopf and global bifurcations (due to insufficient excitation).
Like the omnipresent chaos, bifurcations exist in many
physical systems (4). For instance, power systems generally
have various bifurcation phenomena. When consumers demands for power reach peaks, the stability of an electric
power network may move to its margin, leading to serious
oscillations and stability bifurcations, which may quickly result in voltage collapse. As another example, a typical double
pendulum can display bifurcations as well as chaotic motions.
Some rotational mechanical systems also have similar behavior. Even a common road vehicle driven by a pilot with driver
steering control can have Hopf bifurcation when its stability
is lost, which may also develop chaos and even hyperchaos. A
hopping robot, or a simple two-degree-of-freedom flexible robot arm, can response strange vibrations undergoing period
doubling, which eventually lead to chaos. An aircraft stalling
for flight below a critical speed or over a critical angle of attack can cause various bifurcations. Dynamics of a ship can
exhibit stability bifurcation according to wave frequencies
that are close to the natural frequency of the ship, which creates oscillations and chaotic motions leading the ship to capsize. Simple nonlinear circuits are rich sources of different
types of bifurcations as well as chaos. Other systems that
have bifurcation properties include cellular neural networks,
lasers, aeroengine compressors, weather systems, and biological population dynamics, to name just a few.
CONTROLLING CHAOS
Understanding chaos has long been the main focus of research in the field of nonlinear dynamics. The idea that chaos
can be controlled is perhaps counterintuitive. Indeed, the extreme sensitivity of a chaotic system to initial conditions once
led to the impression and argument that chaotic motion is in
general neither predictable nor controllable.
207
208
4.
5.
6.
7.
It is also worth mentioning an additional distinctive feature of a controlled chaotic system that differs from an uncontrolled chaotic system. The controlled chaotic system is generally nonautonomous and cannot be reformulated as an
autonomous system by defining the control input as a new
state variable, since the controller is physically not a system
state variable and, moreover, it has to be determined via design for performance specifications. Hence, a controlled chaotic system is intrinsically much more difficult to design than
it appears (e.g., many invariant properties of autonomous systems are no longer valid). This observation raises the question of extending some existing theories and techniques from
autonomous system dynamics to nonautonomous, controlled,
dynamical systems, including such complex phenomena as degenerate bifurcations and hyperchaos in the system dynamics
when a controller is involved. Unless suppressing complex dynamics in a process is the only purpose for control, understanding and utilizing the rich dynamics of a controlled system are very important for design and applications.
Representative Approaches to Chaos Control
There are various conventional and nonconventional control
methods available for bifurcations and chaos control (4, 11,
12). To introduce a few representative ones, only three categories of methodologies are briefly described in this section.
Parametric Variation Control. This approach for controlling
a chaotic dynamical system, proposed by Ott, Grebogi, and
Yorke (13,20) and known as the OGY method, is to stabilize
one of its unstable periodic orbits embedded in an existing
chaotic attractor, via small time-dependent perturbations of
the key system parameter. This methodology utilizes the
special feature of chaos that a chaotic attractor typically
has embedded within it a dense set of unstable periodic
orbits.
To introduce this control strategy, consider a general continuous-time parametrized nonlinear autonomous system
x (t) = f (xx (t), p)
(19)
(20)
pk pmax
The iteration of the map near the desired orbit are then observed, and the local properties of this chosen periodic orbit
are obtained. To do so, the map is first locally linearized,
yielding a linear approximation of P near *f and p*, as
k+1 f + Lk (k f ) + v k (pk p )
(21)
tabl
em
ani
fold
le
an
ifo
ld
St
Uns
f = P( f , p )
ab
209
Target
k + 1 iterate
without
perturbation
kth iterate
or
k+1 Lk k + v k pk
k + 1 iterate
with parameter
perturbation
(22)
where
k = k f ,
Lk =
P( f ,
pk = pk p ,
p )/k , v k =
P( f ,
p )/ pk
gsT,k es ,k = guT ,k eu ,k = 1,
gsT,k eu ,k = guT ,k es ,k = 0
pk = u ,k
then the Jacobian Lk can be expressed as
Lk = u ,k eu ,k guT ,k + s ,k es ,k gsT,k
(23)
guT ,k k
guT ,kv k
(24)
where it is assumed that gu,kvk 0. This is the control formula for determining the variation of the adjustable system
parameter p at each step, k 1, 2, . The controlled orbit
thus is expected to approach *f at a geometric rate.
Note that this calculated pk is used to adjust the parameter p only if pk pmax. When pk pmax, however, one
should set pk 0. Also, when k1 falls on a local stable manifold of *f , one should set pk 0 because the stable manifold
might lead the orbit directly to the target.
Note also that the preceding derivation is based on the assumption that the Poincare map, P, always possesses a stable
and an unstable direction (saddle-type orbits). This may not
be the case in many systems, particularly those with high periodic orbits. Moreover, it is necessary that the number of accessible parameters for control is at least equal to the number
of unstable eigenvalues of the periodic orbit to be stabilized.
In particular, when some of such key system parameters are
unaccessible, the algorithm is not applicable or has to be modified. Also, if a system has multiattractors the system orbit
may never return to the opened window but move to another
nontarget limit set. In addition, the technique is successful
only if the control is applied after the system orbit moves into
the small window covering the target, over which the local
210
k = 0, 1, . . .
Gk, k 1, 2, . Let C k1
Ck, and denote the basin of
entrainment for the goal by
B = {xx0 Rn : lim xx k g k = 0}
k
or continuous-time system,
x (t) = f (xx (t)),
x (0) = x 0
g k ) = x k+1 f k (xx k ) = 0,
g k+1 f k (g
or
k kb
(25)
x k Rn
is an open-loop controller, which is directly added to the righthand side of the original system.
A meaningful application of the entrainment control is for
multiattractor systems, to which the parametric variation
control method is not applicable. Another important application is for a system with an asymptotic goal gk g, an equilibrium of the given system. In this case, the basin of entrainment is a convex region in the phase space:
Be = {x0 Rn : xx 0 g < r( g )}
where
r( g ) = max{r : xx 0 g < r lim xx k g = 0}
r
The entrainment-migration control method is straightforward, easily implementable, and flexible in design. However,
it requires the dynamics of the system be accurately described
by either a map or a differential equation. Also, in order for
the system state to be entrained to the given equilibrium, the
equilibrium must lie in a particular subset of the convergent
region. This can be a technical issue, particularly for higherdimensional systems. In addition, due to the open-loop na-
211
x (0) = x 0
(26)
(27)
where g is generally a piecewise continuous nonlinear function, such that the feedback-controlled system
x (t) = f (xx, g (xx, t), t)
(28)
(29)
(30)
(31)
(32)
212
ex
f (.)
g(.)
using
u (t) = f (xx (t), t) + y (t) + K(xx (t) y (t))
x = f (xx, 0, t)
x1 (t) = x2 (t)
x (t) = x (t)
2
..
Subtracting Eq. (33) from Eq. (26) then yields the error dynamics:
e x = f e (eex , x , t)
r +
ex
Kc
x
+
(33)
(34)
where
ex (t) = x (t) xx (t), f e (eex , x , t) = f (xx, g (xx, x , t), t) f (xx, 0, t)
Here, it is important to note that in order to perform correct
stability analysis later on, in the error dynamical system of
Eq. (34) the function f e must not explicitly contain x; if so, x
should be replaced by ex x (see Eq. (38) below). This is because Eq. (34) should only contain the dynamics of ex but not
(35)
(36)
(37)
The controller is required to drive the trajectory of the controlled system of Eq. (37) to approach the target orbit x.
The error dynamics of Eq. (34) now takes the form
ex = f e (eex , t) + Kc ex + g (eex , k c , t)
(38)
where
ex = x x ,
213
(39)
where
f e (eex , t)
A(xx, t) =
eex
e x =0
and h(ex, Kc, kc, t) contains the rest of the Taylor expansion.
The design is then to determine both the constant control
gains Kc and kc as well as the nonlinear function g( , , t)
based on the linearized model of Eq. (39), such that ex 0 as
t . When this controller is applied to the original system,
the goal of both chaos suppression and target tracking will be
achieved. For illustration, two controllability conditions established based on the boundedness of the chaotic attractors
as well as the Lyapunov first and second methods, respectively, are summarized next (24).
Suppose that in Eq. (39), h(0, Kc, kc, t) 0 and A(x, t) A
is a constant matrix whose eigenvalues all have negative real
parts, and let P be the positive definite and symmetric solution of the Lyapunov equation
PA + AT P = I
where I is the identity matrix. If Kc is designed to satisfy
h (eex , Kc , k c , t) ceex
h
for a constant c max(P) for t0 t , where max(P) is the
maximum eigenvalue of P, then the controller u(t), defined in
Eq. (36), will drive the trajectory x of the controlled system of
Eq. (37) to the target, x, as t .
For Eq. (39), since x is tp-periodic, associated with the matrix A(x, t) there always exist a tp-periodic nonsingular matrix M(x, t) and a constant matrix Q such that the fundamental matrix (consisting of n independent solution vectors) has
the expression
(xx, t) = M(xx, t)etQ
The eigenvalues of the constant matrix etpQ are called the Floquet multipliers of the system matrix A(x, t).
In Eq. (39), assume h(0, Kc, kc, t) 0 and h(ex, Kc, kc, t)
and h(ex, Kc, kc, t)/ex are both continuous in a bounded
neighborhood of the origin in Rn. Assume also that
lim
e x 0
h (eex , Kc , k c , t)
h
=0
eex
uniformly with respect to t [t0, ). If the nonlinear controller of Eq. (36) is so designed that all Floquet multipliers i
of the system matrix A(x, t) satisfy
i (t) < 1,
i = 1, . . ., n,
t [t0 , )
then the controller will drive the chaotic orbit x of the controlled system of Eq. (37) to the target orbit, x, as t .
Various Feedback Methods for Chaos Control. In addition to
the general nonlinear feedback control approach described pre-
214
k = 0, 1, . . .
(40)
k = 0, 1, . . .
f (.;p)
Kc
g(.;p)
where and represent the convolution and composition operations, respectively, as shown in Fig. 18. First, suppose that
a system S S(f, g) is given as shown in the figure without
the feedback controller, Kc. Assume also that two system parameter values, ph and pc, are specified, which define a Hopf
bifurcation and a supercritical predicted period-doubling bifurcation, respectively. Moreover, assume that the system has
a family of predicted first-order limit cycles, stable in the
range ph p pc.
Under this system setup, the problem for investigation is
to design a feedback controller, Kc, added to the system as
shown in Fig. 18, such that the controlled system, S*
S*(f, g, Kc), has the following properties:
1. S* has a Hopf bifurcation at p*h ph.
2. S* has a supercritical predicted period-doubling bifurcation for p*c pc.
3. S* has a one-parameter family of stable predicted limit
cycles for p*h p p*c .
4. S* has the same set of equilibria as S.
Only the one-dimensional case is discussed here. First, one
can design a washout filter with the transfer function s/(s
a), where a 0, such that it preserves the equilibria of the
given nonlinear system. Then note that any predicted firstorder limit cycle can be well approximated by
y
(1)
(t) = y0 + y1 sin(t)
s(s2 + 2 (ph ))
(s + a)3
where kc is the constant control gain, and (ph) is the frequency of the limit cycle emerged from the Hopf bifurcation
at the point p ph. This controller also preserves the Hopf
bifurcation at the same point. More importantly, since a 0,
the controller is stable, so by continuity in a small neighborhood of kc the Hopf bifurcation of S* not only remains supercritical but also has a supercritical predicted period-doubling
bifurcation (say at pc(kc), close to ph) and a one-parameter
family of stable predicted limit cycles for ph p pc(kc).
The design is then to determine kc such that the predicted
period-doubling bifurcation can be delayed, to a desired parameter value p*c . For this purpose, the harmonic balance approximation method (10) is useful, which leads to a solution
of y(1) by obtaining values of y0, y1, and (they are functions
of p, depending on kc and a, within the range ph p p*c ).
The harmonic balance also yields conditions, in terms of kc, a
215
H( j )
1
( + j)2 H (s) +
2
(41)
m
z 1,2i+1 2i+1 = H( j)
i=0
m
r 1,2i+1 2i+1
i=1
[H( j )
J + I](rr + z 13 3 + z 15 5 + )
= H( j )[h
h1 3 + h 2 5 + ]
(42)
(43)
216
G1 = f 1 (G, p, 0),
G2 = f 1 (G1 , p, 0) = f 2 (G, p, 0),
..
.
Gm = f m (G, p, 0)
As m increases, the width of Gm in the unstable direction(s)
has a general tendency to shrink exponentially. This suggests
the following control scheme (28):
Pick a suitable value of m, denoted m0. Assume that the orbit initially starts outside the region Gm01 Gm0 G1 G. If
the orbit lands in Gm01 at iterate , the control ul is applied to
kick the orbit out of Gm0 at the next iterate. Since Gm0 is thin, this
217
(46)
0k
(44)
where xk Rn, x0 is given, and f k is assumed to be continuously differentiable, at least locally in the region of interest.
The anticontrol problem for this dynamical system is to
design a linear state-feedback control sequence, uk Bkxk,
with uniformly bounded constant control gain matrices,
Bks u , where s is the spectral norm for a matrix,
such that the output states of the controlled system
i = lim
1
ln i(k)
k
i = 1, . . ., n
(k1)i
Clearly, if (k)
is used in the design, then all (k)
i e
i will
not be zero for any finite values of i, for all i 1, , n and
k 0, 1, . Thus, Tk(x0) is always nonsingular. Consequently, a control-gain sequence Bk can be designed such
that the singular values of the matrix Tk(x0) are exactly equal
n
to ekii1
: At the kth step, k 0, 1, 2, , one may simply
choose the control gain matrix to be
Bk = ( f + ec )In ,
for all k = 0, 1, 2, . . .
where the constants c and f are given in Eqs. (45) and (46),
respectively (29). This ensures Eq. (45) to hold.
Finally, in conjunction with the previously designed controllerthat is,
u k = Bkx k = ( f + ec )xx k
anticontrol can be accomplished by imposing the mod operation in the controlled system:
x k+1 = f k (xx k ) + u k
(45)
x k+1 = f k (xx k ) + u k
(mod 1)
This results in the expected chaotic system whose trajectories remain within a bounded region in the phase space
and, moreover, satisfies the aforementioned three basic
properties that together define discrete chaos. This approach yields rigorous anticontrol of chaos for any given
discrete-time systems, including all higher-dimensional, linear time-invariant systems; that is, with f k(xk) Axk in Eq.
(44), where the constant matrix A can be arbitrary (even
asymptotically stable).
Although uk Bkxk is a linear state-feedback controller,
it uses full-order state variables, and the mod operation is
inherently nonlinear. Hence, other types of (simple) feedback controllers are expected to be developed in the near
future for rigorous anticontrol of chaos, particularly for continuous-time dynamical systems [which is apparently much
more difficult (30), especially if small control input is desired].
BIBLIOGRAPHY
1. The New Encyclopaedia Britannica (Ready Reference and Index),
Micropaedia, Vol. 2, Chicago: Encyclopaedia Britannica, Inc.,
1984, p. 743.
2. T. Matsumoto, Chaos in electronic circuits, Proceedings of the
IEEE, 75: 10331057, 1987.
218
GUANRONG CHEN
University of Houston
300
CONVEX OPTIMIZATION
CONVEX OPTIMIZATION
An optimization problem can be stated in the so-called standard form as follows:
minimize
f (xx ) : R R
subject to g (xx ) 0,
g : Rm Rn
(NLP)
CONVEX OPTIMIZATION
301
x1
x2
A
x1
x2
Convex set
Nonconvex set
c Rn ,
bR
c Rn ,
bR
Polyhedra. A (convex) polyhedron is defined as an intersection of half spaces, and is given by the equation
P = {xx|Axx b ),
A Rmn ,
b Rm
DEFINITIONS OF CONVEXITY
Convex Sets
Definition. A set C Rn is said to be a convex set if, for
every x1, x2 C and every real number , 0 1, the
point x1 (1 )x2 C.
This definition can be interpreted geometrically as stating
that a set is convex if, given two points in the set, every point
on the line segment joining the two points is also a member
of the set. Examples of convex and nonconvex sets are shown
in Fig. 1.
x
Figure 2. An example convex polytope in two dimensions as an intersection of five half spaces.
302
CONVEX OPTIMIZATION
x1
x3
x2
x4
x5
c =1 c = 2
c=3
x
X
m
y=
X
m
i x i ,
i 0 i,
i=1
i = 1
i=1
The convex hull is thus the smallest convex set that contains the m points. An example of the convex hull of five
points in the plane is shown by the shaded region in Fig. 5. If
the set of points xi is of finite cardinality (i.e., m is finite),
then the convex hull is a polytope. Hence, a polytope is also
often described as the convex hull of its vertices.
Convex Functions
Definition. A function f defined on a convex set is said to
be a convex function if, for every x1, x2 and every , 0
1,
f [xx 1 + (1 )xx 2 ] f (xx 1 ) + (1 ) f (xx 2 )
f is said to be strictly convex if the above inequality is strict
for 0 1.
Geometrically, a function is convex if the line joining two
points on its graph is always above the graph. Examples of
convex and nonconvex functions on Rn are shown in Fig. 6.
2 f (xx 0 ) 0
f (x)
g(x)
x1
Figure 6. Convex and nonconvex functions.
xx 0 S
This is a very useful property, and many convex optimization algorithms depend on the fact that the constraints are
defined by an intersection of level sets of convex functions.
x2
Convex function
x1
x2
Nonconvex function
CONVEX OPTIMIZATION
CONVEX OPTIMIZATION
Convex Programming
Definition. A convex programming problem is an optimization problem that can be stated as follows:
f (xx )
minimize
minimize c Tx
subject to
where
Axx b ,
AR
mn
x0
bR ,
m
(CP)
such that x S
(LP)
cR ,
n
xR
The feasible region for a linear program corresponds to a polyhedron in Rn. It can be shown that an optimal solution to a
linear program must necessarily occur at one of the vertices
of the constraining polyhedron. The most commonly used
technique for solution of linear programs, the simplex method
(11), is based on this principle and operates by a systematic
search of the vertices of the polyhedron. The computational
complexity of this method can show exponential behavior for
pathological cases, but for most practical problems it has been
observed to grow linearly with the number of variables and
sublinearly with the number of constraints. Algorithms with
polynomial-time worst-case complexity do exist; these include
Karmarkars method (2) and the ShorKhachiyan ellipsoidal
method (12). The computational times of the latter, however,
are often seen to be impractical.
In the remainder of this section, we will describe various
methods used for convex programming and for mapping problems to convex programs.
303
S = {xx| gi (xx ) 0, i = 1, 2, . . ., m}
where each gi(x) is a convex function. The computation required of the method is dependent on the choice of the barrier
function. In this connection, the logarithmic barrier function
(abbreviated as the log barrier function) for the set of inequalities is defined as
(xx ) =
( Pn
x )]
i=1 log[gi (x
for x S
otherwise
Intuitively, the idea of the barrier is that any iterative gradient-based method that tries to minimize the barrier function
will be forced to remain in the feasible region, due to the singularity at the boundary of the feasible region. It can be
shown that (x) is a convex function over S and its value
approaches infinity as x approaches the boundary of S. Intuitively, it can be seen that (x) becomes smallest when x is,
in some sense, farthest away from all of the boundaries of S.
The value of x at which the function (x) is minimized is
called the analytic center of S.
Example. For a linear programming problem of the type described in Eq. (LP), with constraint inequalities described by
aTi x bi, the barrier function in the feasible region is given
by
X
n
(xx ) =
log(bi a Ti x )
i=1
The value of bi aTi x represents the slack in the ith inequality, i.e., the distance between the point x and the corresponding constraint hyperplane. The log barrier function, therefore,
is a measure of the product of the distances from a point x to
each hyperplane, as shown in Fig. 8(a). The value of (x) is
n
minimized when i1
(bi aTi x) is maximized. Coarsely speaking, this occurs when the distance to each constraint hyperplane is sufficiently large.
As a cautionary note, we add that while the analytic center
is a good estimate of the center in the case where all con-
Path-Following Methods
This class of methods proceeds iteratively by solving a sequence of unconstrained problems that lead to the solution of
the original problem. In each iteration, a technique based on
barrier methods is used to find the optimal solution. If we
denote the optimal solution in the kth iteration as x*k , then
the path x*1 , x*2 , . . . in Rn leads to the optimal solution, and
hence techniques of this class are known as path-following
methods.
Analytic
center
p
(a)
Analytic
center
p
(b)
Figure 8. (a) Physical meaning of the barrier function for the feasible region of a linear program; (b) the effect of redundant constraints
on the analytic center.
304
CONVEX OPTIMIZATION
f (xx )
i = 1, 2, . . ., m
>0
Optimum
= 10
=0
(Analytic
center)
= 100
X PAX Xcc
PAX Xcc
CONVEX OPTIMIZATION
the current ellipsoid, y is the variable over which the minimization is being performed, and H is the Hessian of the logn
barrier function i1 log[gi(y)]. The problem now reduces
to
{min b Ty : (yy y c )T H(yy y c ) 2 }
which has a closed-form solution of the form
H 1b
y ( ) = y
b T H 1b
This is used as the center of the ellipsoid in the next iteration.
The procedure continues iteratively until convergence.
Potential-Based Methods. These methods formulate a potential function that provides a coarse estimate of how far the
solution at the current iterate is from the optimal value. At
each step of a potential reduction method, a direction and
step size are prescribed; however, the potential may be minimized further by the use of a line search (large steps). This is
in contrast with a path-following method that must maintain
proximity to a prescribed central path. An example of a potential-based technique is one that utilizes a weighted sum of the
gap between the value of the primal optimization problem
and its dual, and of the log barrier function value as the potential. For a more detailed description of this and other
potential-based methods, the reader is referred to Refs. 6
and 13.
Localization Approaches
Polytope Reduction. This method begins with a polytope
P Rn that contains the optimal solution, xopt. The polytope P may, for example, be initially selected to be an ndimensional box described by the set
{xx|xmin x(i) xmax }
where xmin and xmax are the minimum and maximum values of
each variable, respectively. In each iteration, the volume of
the polytope is shrunk while keeping xopt within the polytope,
until the polytope becomes sufficiently small. The algorithm
proceeds iteratively as follows.
Step 1. A center xc deep in the interior of the polytope P
is found.
Step 2. The feasibility of the center xc is determined by
verifying whether all of the constraints of the optimization problem are satisfied at xc. If the point xc is infeasible, it is possible to find a separating hyperplane passing through xc that divides P into two parts, such that
the feasible region lies entirely in the part satisfying the
constraint
c Tx
where c [gp(x)] is the negative of the gradient of
a violated constraint gp, and cTxc. The separating
hyperplane above corresponds to the tangent plane to
the violated constraint. If the point xc lies within the
feasible region, then there exists a hyperplane cT x
T
305
that divides the polytope into two parts such that xopt is
contained in one of them, with c [f(x)]T being the
negative of the gradient of the objective function, and
being defined as cT xc once again. This hyperplane
is the supporting hyperplane for the set f(x) f(xc) and
thus eliminates from the polytope a set of points whose
value is larger than the value at the current center. In
either case, a new constraint cT x is added to the
current polytope to give a new polytope that has roughly
half the original volume.
Step 3. Go to step 1 and repeat the process until the polytope is sufficiently small.
Note that the center in step 1 is ideally the center of gravity of the current polytope, since a hyperplane passing
through the center of gravity is guaranteed to reduce the volume of the polytope by a factor of 1 1/e in each iteration.
However, since finding the center of gravity is prohibitively
expensive in terms of computation, the analytic center is an
acceptable approximation.
Example. The algorithm is illustrated by using it to solve
the following problem in two dimensions:
minimize
f (x1 , x2 )
such that
(x1 , x2 ) S
f decreasing
xc
xopt
xc
xopt
(b)
(a)
xc
xc
xopt
xopt
(c)
(d)
306
CONVEX OPTIMIZATION
Ek
f(xk )
xk
Ek + 1
Figure 11. The ellipsoidal method.
Other Characterizations of Quasiconvexity. A function f defined on a convex set is quasiconvex if, for every x1, x2 :
smaller polytope. Its center lies inside the feasible region, and
hence the gradient of the objective function is used to generate a hyperplane that further shrinks the size of the polytope,
as shown in Fig. 10(c). The result of another iteration is illustrated in Fig. 10(d). The process continues until the polytope
has been shrunk sufficiently.
Ellipsoidal Method. The ellipsoidal method begins with a
sufficiently large ellipsoid that contains the solution to the
convex optimization problem. In each iteration, the size of the
ellipsoid is shrunk, while maintaining the invariant that the
solution lies within the ellipsoid, until the ellipsoid becomes
sufficiently small. The process is illustrated in Fig. 11.
The kth iteration consists of the following steps, starting
from the fact that the center xk of the ellipsoid E k is known:
Step 1. In case the center is not in the feasible region, the
gradient of the violated constraint is evaluated; if it is
feasible, the gradient of the objective function is found.
In either case, we will refer to the computed gradient
as h(xk).
Step 2. A new ellipsoid containing the half ellipsoid given
by
Ek {xx|h(xx k ) x h(xx k ) x k )
T
Ak+1 =
1
A g
n+1 k k
n2
2
T
A g g A
Ak
n2 1
n+1 k k k k
minimize c Tx
subject to F (xx ) 0
where
x k+1 = x k
where
h(xx k )
g k =
h(xx k )T Ak h(xx k )
Step 3. Repeat the iterations in steps 1 and 2 until the
ellipsoid is sufficiently small.
RELATED TECHNIQUES
Quasiconvex Optimization
Definition. A function f : S R, where S is a convex set, is
quasiconvex if every level set La xf(x) a is a convex set.
A function g is quasiconcave if g is quasiconvex over S. A
function is quasilinear if it is both quasiconvex and quasiconcave.
F (xx ) R
mm
(SDP)
,
c, x R
(cc Tx )2
d Tx
subject to Axx + b 0
minimize
CONVEX OPTIMIZATION
minimize t
subject to
2
diag(Axx + b )
64
0
0
0
t
c Tx
3
0
7
c Tx 5 0
d Tx
cT x
0
d Tx
c Tx
The two tricks shown here, namely, the reformulation of linear inequations and the use of Schur complements, are often
used to formulate optimization problems as semidefinite programs.
Geometric Programming
Definition. A posynomial is a function h of a positive variable x Rm that has the form
X Y
n
h(xx ) =
307
j)
x (i,
i
i=1
f (xx ) =
1
(2 )n det
e0.5(xxc )
T 1 (xx )
c
where 0.
2. The exponential distribution
f (xx ) =
Y
n
!
(i) e((1)x(1)+(2)x(2)++(n)x(n))
i=1
ENGINEERING PROBLEMS AS
CONVEX OPTIMIZATION PROBLEMS
There has been an enormous amount of recent interest in
applying convex optimization to engineering problems, particularly as the optimizers have grown more efficient. The
reader is referred to Boyd and Vandenberghes lecture notes
(19) for a treatment of this subject. In this section, we present
a sampling of engineering problems that can be posed as convex programs to illustrate the power of the technique in
practice.
308
CONVEX OPTIMIZATION
Design Centering
While manufacturing any system, it is inevitable that process
variations will cause design parameters, such as component
values, to waver from their nominal values. As a result, after
manufacture, the system may no longer meet some behavioral
specifications, such as requirements on the delay, gain, and
bandwidth, that it has been designed to satisfy. The procedure of design centering attempts to select the nominal values
of design parameters to ensure that the behavior of the system remains within specifications with the greatest probability and thus maximize the manufacturing yield.
The random variations in the values of the design parameters are modeled by a probability density function (x,xc) :
Rn [0,1], with a mean corresponding to the nominal value
of the design parameters. The yield of the manufacturing process, Y, as a function of the mean xc is given by
Y (xx c ) =
(xx, x c ) dxx
xF
h|h
h = h c + F p , p
p 1}
H = {h
where p (pTp)1/2, h Rk1, F R(k1)q.
The robust version of the optimization problem above must
ensure that the error is minimized over all possible values of
h within the ellipsoid. To consider the worst-case tracking error, the optimization problem may be written as
hH t=1,2,..., M
t = 1, 2, . . ., N
|u(t + 1) u(t)| S,
t = 1, 2, . . ., N 1
t=1,2,..., M
minimize
subject to Ulow u(t) Uhigh,
t = 1, 2, . . ., N
S u(t + 1) u(t) S,
t = 1, 2, . . ., N 1
h Dt u ydes (t)]
F Dt u + [h
T
(xx, x c ) dxx
xP
minimize error =
t=1,2,..., M
t = 1, 2, . . ., N
t = 1, 2, . . ., N 1
h T Dt u ydes (t)]
F T Dt u [h
Optimizing Structural Dynamics
Consider a linear elastic structure consisting of a stack of k
linear elastic bars connecting a set of p nodes. The topology
and lengths of the bars and their material are fixed, and the
appropriate cross-sectional widths of the bars are to be determined. The elastic stored energy of this system is given by
f Td, where f is the vector of load forces and d is the vector of
(small) node displacements. The relation between f and d is
given by f A(x) d, where A(x), called the stiffness matrix,
k
is an affine sum of the variables xi, given by A(x) i1xi Ai
with the matrices A being all symmetric positive semidefinite.
The optimization problem of minimizing the elastic stored energy (24) can then be stated as follows:
minimize
f Td
k
subject to
ljxj v
j=1
d
f = A(xx )d
x j,min x j x j,max
for j = 1, 2, . . ., k
CONVEX OPTIMIZATION
minimize
f T A(xx )1 f
X
k
subject to
ljxj v
j=1
x j,min x j x j,max
for j = 1, 2, . . ., k
Using Schur complements, this leads to the semidefinite programming formulation in x and t given by
minimize t
subject to
t
f
k
X
fT
0
A(x)
309
X
n
ljxj v
minimize
j=1
l i xi
i=1
x j,min x j x j,max
for j = 1, 2, . . ., k
x min x x max
VLSI Transistor and Wire Sizing
Convex Optimization Formulation. Circuit delays in integrated circuits often have to be reduced to obtain faster response times. Given the circuit topology, the delay of a combinational circuit can be controlled by varying the sizes of
transistors, giving rise to an optimization problem of finding
the appropriate areadelay tradeoff. The formal statement of
the problem is as follows:
minimize area
subject to
delay Tspec
(TS)
X
n
area =
xi
i=1
maximize
log det B
subject to B = BT > 0
a i bi a Ti d ,
Ba
i = 1, 2, . . ., L
X
N
B = BT > 0
G( )
i=1
wi gi ( ) exp j
2
(x cos + yi sin )
i
310
CONVEX OPTIMIZATION
minimize |G(i )|
subject to |G(i )| Ui
|wk | W
for j = 1, 2, . . ., M
for k = 1, 2, . . ., N
G(0 ) = 1
BIBLIOGRAPHY
1. W. Stadler, Multicriteria Optimization in Engineering and in the
Sciences, New York: Plenum, 1988.
2. N. Karmarkar, A new polynomial-time algorithm for linear programming, Combinatorica, 4: 373395, 1984.
3. A. V. Fiacco and G. P. McCormick, Nonlinear Programming, New
York: Wiley, 1968.
4. J. Renegar, A polynomial time algorithm, based on Newtons
method, for linear programming, Math. Programming, 40: 59
93, 1988.
5. C. C. Gonzaga, An algorithm for solving linear programming
problems in O(nL) operations, in Progress in Mathematical Programming: Interior Point and Related Methods, N. Meggido (ed.),
New York: Springer-Verlag, 1988, pp. 128.
minimize t
subject to |G(i )| t,
i = 1, 2, . . ., L
|G(i )| U j
|wk | W
for j = 1, 2, . . ., M
for k = 1, 2, . . ., N
G(0 ) = 1
The last constraint here is a normalization constraint and can
be handled by decreasing the number of variables by one. Recalling that each wj is complex, we may choose the vector x
as
6. Y. Nesterov and A. Nemirovskii, Interior-Point Polynomial Algorithms in Convex Programming, SIAM Studies in Applied Mathematics, Philadelphia: Society for Industrial and Applied Mathematics, 1994.
7. P. M. Vaidya, An algorithm for linear programming which requires O(((m n)n2 (m n)1.5n)L) arithmetic operations, Math.
Programming, 47: 175201, 1990.
8. Y. Ye, An O(n3L) potential reduction algorithm for linear programming, Math. Programming, 50: 239258, 1991.
9. R. T. Rockafellar, Convex Analysis, Princeton, NJ: Princeton University Press, 1970.
11. P. E. Gill, W. Murray, and M. H. Wright, Numerical Linear Algebra and Optimization, Reading, MA: Addison-Wesley, 1991.
i = 1, 2, . . ., L + M + N
c i = {[0, 0, . . ., 0, 1]
[0, 0, 0, . . ., 0]
d i = {[0, 0, . . ., 0, 0]
UiL
W
for i = 1, 2, . . ., L
for i = L + 1, . . ., L + M + N
for i = 1, 2, . . . L
for i = L + 1, . . ., M
for i = L + M + 1, . . ., L + M + N
19. S. Boyd and L. Vandenberghe, Introduction to Convex Optimization with Engineering Applications, Lecture Notes, Electrical Engineering Department, Stanford University, 1995. Available from
http://www-isl.stanford.edu/boyd.
CONCLUSION
This overview has presented an outline of convex programming. The use of specialized techniques that exploit the convexity properties of the problem have led to rapid recent advances in efficient solution techniques for convex programs,
which have been outlined here. The applications of convex optimization to real problems of engineering design have been
illustrated. Convex optimization techniques are used widely
in control, for example, in Youla-based design and in design
by linear matrix inequalities (LMIs). For Youla-based design,
the reader is referred to Refs. 30 and 31. A good sourcebook
for design by LMIs is Ref. 32 and a useful practical design
tool is the LMI Control Toolbox (33).
CONVOLUTION
25. S. S. Sapatnekar and S. M. Kang, Design Automation for TimingDriven Layout Synthesis, Boston: Kluwer Academic, 1993.
26. J. Fishburn and A. E. Dunlop, TILOS: A posynomial programming approach to transistor sizing, in Proc. IEEE Int. Conf.
Comput.-Aided Des., 1985, pp. 326328.
27. S. S. Sapatnekar et al., An exact solution to the transistor sizing
problem using convex optimization, IEEE Trans. Comput.-Aided
Des. Integr. Circuits Syst., 12: 16211632, 1993.
28. L. Vandenberghe, S. Boyd, and A. El Gamal, Optimal wire and
transistor sizing for circuits with non-tree topology, in Proc. IEEE
Int. Conf. Comput.-Aided Des., 1997, pp. 252259.
29. H. Lebret, Optimal beamforming via interior point methods, J.
VLSI Signal Process., 14 (1): 2941, 1996.
30. S. P. Boyd and C. H. Baratt, Linear Controller Design: Limits of
Performance, Englewood Cliffs, NJ: Prentice-Hall, 1991.
31. M. A. Dahleh and I. J. Diaz-Bobillo, Control of Uncertain Systems:
A Linear Programming Approach, Englewood Cliffs, NJ: PrenticeHall, 1995.
32. S. P. Boyd et al., Linear Matrix Inequalities in System and Control
Theory, SIAM Studies in Applied Mathematics, Philadelphia, PA:
SIAM, 1994.
33. P. Gahinet et al., LMI Control Toolbox, Natick, MA: The MathWorks, 1995.
SACHIN S. SAPATNEKAR
University of Minnesota
311
DELAY SYSTEMS
139
DELAY SYSTEMS
In most applications of mathematics to engineering it is
tacitly assumed that the systems under consideration are
causal. That is, the future state of the system depends
only on its present state. In reality most electrical systems,
particularly control systems, are subject to transportation
and/or processing delays. Usually these delays are ignored,
either because they are considered small or because they
complicate the mathematical model. Thus a dilemma arises.
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
140
DELAY SYSTEMS
When does the realistic modeling of a physical system require the introduction of time delays into the mathematical
model? One purpose of this article is to introduce the
reader to the fundamental properties of time delay differential equations and to compare these to the corresponding
ones for ordinary differential equations. The other is to sum
up the structure and fundamental properties of delay systems of the type most frequently encountered in electrical
engineering.
Ordinary differential equations (ODE) in which a part of
the past history affects the present state are called delay differential equations or functional differential equations (FDE).
Some examples are
x(t)
= ax(t) + bx(t 1) +
0
1
x(t + ) d
d
(x(t) dx(t 1)) = ax(t) + bx(t 1) + c
dt
(1)
0
1
x(t + ) d
(2)
tems using only the finite-dimensional approach. The monograph by J. K. Hale and S. M. Verduyn-Lunel (5) develops
both the coarse and fine structure of delay systems using the
powerful tools of infinite-dimensional analysis together with
the more mundane methods of linear algebra and complex
analysis.
There are two main categories of FDE considered in the
engineering literature, retarded functional differential equations (RFDE) and neutral functional differential equations
(NFDE). A delay system, written as a first order system, is a
RFDE if the derivative contains no delay terms, Eq. (1) is a
RFDE. If the derivative contains delay terms in a first-order
system, the equation is called a neutral functional differential
equation. Eq. (2) for d 0 is a NFDE. In engineering practice,
only a certain class of NFDE is considered; namely, D-stable
NFDE. A D-stable NFDE is one in which the difference equation associated with the derivative is uniformly exponentially
stable (u.e.s.) Eq. (2) is a D-stable NFDE if d 1 since the
respective difference equation,
y(t) dy(t 1) = 0
(3)
DELAY SYSTEMS
141
142
DELAY SYSTEMS
r
x(,
) = S() (0)
D j (h j )
+ S()
L (x(t))() = x()
r
h j j=1
(A j + D j )e(+hj) ( ) d + S()B
u()
(7)
x(t, , u) = S(t) (0)
+
r
j=1
r
D j (h j )
j=1
0
h j
)) d (8)
S(t h j )(A j ( ) + D j (
S(t )Bu( ) dv
+
0
where
x(t)et dt
S(t) = L 1 (S())(t)
is the Laplace transform of x(t). The inverse Laplace transform of any vector or matrix valued function, provided it exists, is denoted by L 1(S())(t).
The delay systems most commonly encountered in electrical engineering are LTI systems with discrete time delays.
The fundamental properties of these systems serve as a paradigm for most other systems one encounters. These systems
which include RFDE and NFDE often have their dynamics
described by
r
d
x(t)
D j x(t h j )
dt
j=1
= A0 x(t) +
j=1
(9)
The function S(t) formally satisfies Eq. (4) with the initial
matrix S(0) I, S() 0 for 0 and is referred to as the
fundamental matrix solution of Eq. (4).
If the matrices in Eq. (4) are time varying, there is a representation of the solution similar to Eq. (8), but it is not obtained by using the Laplace transform. The matrix S(t) is replaced by a matrix S(t, ), where S(, ) I, and S(t, ) 0 if
t. The matrix function S(t, ) is formally a matrix solution
of Eq. (4) for t (5).
Stability
r
A j x(t h j ) + Bu(t),
t0
(4)
j=1
x(t)
and initial values
r
D j x(t h j ) = 0
(10)
j=1
x(t) = (t), h t 0,
C(h)
(5)
In Eq. (4), the matrices Dj, Aj, and A0 are n n-matrices
with real entries; the matrix B is an n m-matrix with real
entries and 0 hj h, 1 j r. The m-vector u(t) is called
the control.
A solution x(t, , u) of Eq. (4), Eq. (5) is a function which
satisfies Eq. (5), is continuously differentiable for t 0, and
has a right hand derivative at t 0 which satisfies Eq. (4).
We can use the Laplace transform to obtain the existence of
and a specific representation for a solution. Let
!
1
r
r
h j
h j
Dje
A0
Aje
S() = I
j=1
j=1
(6)
det I
r
Dje
h j
=0
(11)
j=1
DELAY SYSTEMS
det I
r
D j eh j
A0
j=1
r
=0
d
[x(t) x(t 1)] = x(t)
dt
(13)
d
d2
[x(t) 2x(t 1) + x(t 2)] + 2 [x(t) x(t 1)] + x(t) = 0
dt 2
dt
(14)
The eigenvalues of both systems are solutions of the equation
(1 e) 1 0 which has all its solutions in Re 0.
Equation (13) has all its solutions tending to zero as t tends
to infinity, but it is not u.e.s. Equation (14) has solutions
which tend to infinity as t tends to infinity.
The stability or instability of LTI FDE can be determined
by LyapunovKrasovskii functions and sometimes by Razumikhin functions. However as was mentioned above, these
are difficult to find for all but the simplest systems and even
then are usually selected on an ad hoc basis. To illustrate
this, consider the scalar system
a > 0, a > |b| > 0, r > 0
(15)
(( )) d
2
= wr, w > 0
= cT x
(17)
0
r
dV
(x(t)) (a |b|)(x(t)) 0
dt
(12)
j=1
x(t)
= ax(t) + bx(t r),
If we consider only those solutions x(t) that satisfy the relation x(t) x(t r), then
A j eh j
143
(18)
(16)
(19)
(20)
dV
(xt ) = (a + c)(x(t)2 + bx(t r)x(t) c(x(t r))2
dt
is u.e.s. and let
This functional will be negative on C(r) if we choose c a/2.
Therefore, by Theorem 2.1, Chapter 5, in Ref. 5, the region of
u.e.s. contains the set of coefficients a, b with a b. Notice
that this simple choice for the LyapunovKrasovskii functional yielded the stability region which is completely independent of the size of the delay.
Now, consider the Razumikhin function for Eq. (15) given
by V(x(t)) x2 /2. Along differentiable trajectories of
Eq. (15),
dV
(x(t)) = x(t)(ax(t) + bx(t r))
dt
1
0
a
(21)
144
DELAY SYSTEMS
V (, ) = Q(, ) +
f (s) ds,
>0
(22)
(23)
but remark that the basic condition holds for any linear FDE
whose coefficient matrices are uniformly bounded.
We assume that the square matrices A0(t) and A1(t) have
all their entries uniformly bounded on R. It is known (5) that
the solutions of Eq. (23) with initial conditions zero in C(h)
may be represented in the form
x(t, t0 , f ) =
S(t, ) f ( ) d
(24)
t0
(25)
Theorem. A necessary and sufficient condition for the homogeneous version of Eq. (23) to be u.e.s. is that for all f which
are uniformly bounded on R, the vector function in Eq. (24)
satisfies an inequality of the form x(t, t0, f) Mf, where Mf
is finite and depends only on f (10).
Although the Perron condition or BIBO condition may be
theoretically difficult to verify, a modified form often is used
in control engineering. The linear system is subjected to periodic forcing functions at a variety of frequencies and with uniformly bounded gains. If the outputs are uniformly bounded
over long time periods, the system is considered u.e.s.
Control and Stabilization
Controllability for an FDE system is function space control;
that is, one seeks to control a given initial point in C(h) to a
(26)
1
x(,
, u) =
a beh
(0) +
be
h
( +h)
( ) d + u()
(27)
Since both xt(, u) and u(t) are identically zero after time T,
the functions x(, , u) and u() in Eq. (27) must be entire
analytic functions (19). This means that u() must be chosen
so that the numerator in Eq. (27) is zero when the denominator is zero. But the zeros of a beh are infinite in number and can at best be approximated.
There are several versions of the Pontryagin Maximum
Principle for FDE control problems, and the theoretical
method used to solve control problems is the Method of Dynamic Programming. From an engineering point of view,
these are only of academic interest. Comprehensive reference
sources for this area of control are Refs. 12 and 20.
For multiple-input-multiple-output LTI ODE, there are
three basic methods of feedback stabilization. These are pole
placement, linear quadratic regulator (LQR)-optimization,
and H-optimization pole placement. The latter has the simplest numerical structure but has less eclat than the other
two methods. Pole placement methods are possible for LTI
FDE. In practice the best one can hope for are constant gain
feedbacks which guarantee a given decay rate. However,
LQR-optimization and H-optimization have in theory been
completely extended from LTI ODE to LTI FDE. There are at
least two ways to look at these extensions. One way relates
to the specific delay structure and minimizes the use of Banach space theory. The other embeds LTI delay systems into
a general class of LTI infinite-dimensional systems known as
Pritchard-Salamon systems (P-S systems) and makes extensive use of the theory of Sobelev spaces (21). Here, we confine
ourselves to the first approach.
A typical LQR-optimization problem is the following. For a
given positive definite matrix W and a given C(h), choose
the control function u(t) to minimize the functional
F (u) =
(28)
(29)
DELAY SYSTEMS
(30)
u(t) = BT Kxt , d(t) =
(31)
BBT
=0
I + AT0 + AT1 eh
q(t)
(32)
xT (t)Wx(t) + uT (t)u(t)
(33)
1 T
d
(t)d(t)
dt
2
(34)
where A0, A1, B, and W are the matrices in Eqs. (28) and
(29), and L in an n r-matrix. It is assumed that the
systems
x(t)
= A0 x(t) + A1 x(t h) + Bu(t)
(36)
x(t)
= A0 x(t) + A1 x(t h) + Ld(t)
(37)
and
min max F (u, d)
u
(38)
(41)
C(u, d) =
0
1
xT (t h)x(t h) + uT (t)u(t) 2 d T (t) d(t) dt
(0)
x(t)
= A0 x(t) + A1 x(t h) + Ld(t) + Bu(t)
(40)
F (u, d) =
(39)
1 T
L Kxt
2
I A0 A1 eh
det
W
145
r
j=1
D j (h
j ) A0 (0)
r
A j (h j ) = 0
(42)
j=1
146
DELAY SYSTEMS
LTI RFDE and LTI parabolic PDE are dissimilar. For example, parabolic PDE generate analytic semigroups, and RFDE
generate C0-semigroups which are compact for t h but are
not analytic. One basic reason for this difference is that the
eigenvalues of LTI RFDE do not belong to a sector in the complex plane. On the other hand, some LTI hyperbolic PDE and
LTI D-stable NFDE have identical mathematical structures.
In particular, the equations for the dynamics of transmission
lines described by the telegraph equation can be transformed
to D-stable NFDE (5). An example which illustrates the similarity between some hyperbolic PDE and D-stable NFDE, are
the PDE system
wtt = wxx 2awt a2 w = 0,
w(0, t) = 0,
(43)
(44)
y = x(t)
(46)
1
2
+ sin 2t x(t 1)
(47)
Let V be a continuous mapping from C(h) into R which satisfies the following two conditions:
(48)
1
[V (xh )() V ()] b(|(0)|)
h
(49)
(50)
DELAY SYSTEMS
I A
W
BBT
I + AT
1
x0
Kx0
(51)
in Z2n has no poles in the right half complex plane. Since the
poles of the matrix function in Eq. (51) are symmetric with
respect to the imaginary axis, the matrix K is uniquely determined once the solutions of
det
I A
W
BBT
I + AT
=0
(52)
147
ation of solutions, continuous dependence on data, and parameters, etc. for these systems are similar to the corresponding ones for delay systems in Rn. The major exception to this
statement occurs for properties which depend on the compactness of closed bounded sets in Rn or Zn. These systems are
often encountered in models in population ecology, genetic repression, control theory, climatology, coupled oscillators, age
dependent populations, etc. (29,30).
TIME DELAYS IN CONTROL SYSTEMS
Time delays are sometimes desired in the design of control
systems. For example in self-tuning control, one encounters
systems of the form
y(t) + a1 y(t 1)a2 y(t 2) + ana y(t na)
= b1 w(t 1) + b2 u(t 2) + + bub u(t nb)
wtt = wxx ,
ut duxx =
(53)
(25), which is a nonlinear diffusion equation with a time delay. Extensions of time delay systems to PDE and abstract
Banach spaces may be found in Refs. 2629.
Time independent versions of these systems are often of
the form
x(t)
= Ax(t) + f (xt )
(57)
(58)
(56)
then a small time delay in the control will not destroy the
stabilization. Of course, the word small depends on the particular system. However, if system [Eq. (55)] is infinite-dimensional, it may be unable to tolerate any delays, particularly if
it is an abstract representation of a boundary stabilized hyperbolic PDE. The simplest example of such a system is given
by
w(0, t) = 0,
EXTENSIONS TO INFINITE DIMENSIONAL PHASE SPACES
(55)
(54)
u(t) = wt (1, t)
(59)
h>0
(60)
(61)
148
DELAY SYSTEMS
has solutions
w(x, t) = e n t sinh nt
(62)
(63)
(64)
(65)
2
n
(66)
This result implies that large-dimensional Galerkin approximations to systems of the type in Eqs. (57) and (58) become unstable for small time delays. Thus, there is a tradeoff
between the dimension of the approximation and the tolerance for delays.
The above example illustrates one of the important differences between LTI evolutionary systems whose initial space
is infinite dimensional and those for which this space is finite
dimensional. It is instructive to make some general remarks
about why such a situation might occur. Suppose that is a
parameter varying in a subset S of a Banach space, and
T(t), t 0 is a C0-semigroup of linear transformations on a
Banach space X which is continuous in ; that is, T(t)x is
continuous in (, t, x). Let r be the radius of the spectrum of
T(1). The asymptotic behavior of the semigroup is determined
by the spectrum (T(1)) of T(1). If (T(1)) is inside the unit
circle in the complex plane, then each orbit T(t), t 0 will
approach zero exponentially and uniformly. If there is a point
in (T(1)) outside the unit circle, then there is an unbounded
orbit, and we have instability. A fundamental role in the
study of stability and the preservation of stability under perturbations in the parameter is the behavior of the essential
spectrum e(T(1)) of T(1). Let re r(e(T(1))) denote the radius of the essential spectrum of T(1). If it is known that
re 1, then the stability or instability of 0 is determined by
eigenvalues of (T(1)). Furthermore, if 0 is unstable, then it
is due to only a finite number of eigenvalues; that is, the instability occurs in a finite dimensional subspace of X. If the
latter situation occurs, then it is natural to expect that the
stabilization of the system could be accomplished by using a
finite dimensional control. However, if it is required that the
stabilization be insensitive to small changes in the parame-
ter, then this may not be the case. If re 1, then the instability of the system is of such a nature that it cannot be controlled by a finite dimensional control.
In general, eigenvalues of T(1) can be chosen to be continuous functions of . On the other hand, the function re may
not be continuous in . The function re will be continuous in
at a point 0 if it is known that T(1) T0(1) is compact.
This condition is not necessary, but it is sufficient. If this difference is only bounded, then there is the possibility of a large
shift in re if we vary . For example, if re0 1, and the perturbation is only bounded, it is possible to have re 1 for a
sequence of j 0, and the semigroup Tj(t) will be unstable
for each j.
Let us interpret these remarks in terms of the examples
that we have been discussing above. For finite dimensional
problems, the semigroup is compact and thus the asymptotic
behavior of orbits is determined by the eigenvalues. Since the
semigroup for a LTI RFDE is compact for t h, the continuous spectrum always is the point zero, and so the asymptotic
behavior is again determined by dominant eigenvalues (finite
in number), and these are generally continuous in parameters.
For D-stable NFDE, the essential spectrum lies in the unit
circle for all values of the delay. Therefore, the introduction
of small delays in the control function does not disturb the
stabilization property of the feedback control.
It can happen that the solutions of the difference equation associated with the difference operator D for an NFDE
has all solutions approaching zero exponentially and uniformly for a particular value of the delays, and a small
change in the delays leads to exponential instability. If this
is the case, then the asymptotic behavior of solutions of
NFDE subjected to small variations in the delays is not
determined by the eigenvalues of the semigroup, but by the
essential spectrum, which in turn is determined by the
eigenvalues of the difference equations associated to the
difference operator D.
In the example [Eqs. (57) and (59)], the natural space of
initial data is HB1 (0, 1) L2(0, 1), where B represents the homogeneous boundary conditions w 0 at x 0, wx 0 at
x 1. In this case, the boundary control wt is bounded but
not compact. If this control is implemented with a delay, then
the radius of the essential spectrum is increased considerably
and, in fact, leads to instability.
It is possible to consider a more physical version of Eqs.
(5759), for which the boundary control problem is insensitive
to small delays in the time at which it is implemented. Consider the equation
wtt wxx cwxxt = 0,
(67)
(68)
DEMODULATORS
BIBLIOGRAPHY
1. N. Minorsky, Self-excited oscillations in dynamical systems possessing retarded action, J. Appl. Mech., 9: 6571, 1942.
2. A. D. Mishkis, General theory of differential equations with a
retarded argument, Amer. Math. Soc. Trans., No. 55, 1951.
3. N. N. Krasovskii, Stability of Motion, Moscow, 1959, Translation:
Palo Alto: Stanford Univ. Press, 1963.
149
5. J. K. Hale and S. Verduyn-Lunel, Introduction to functional differential equations, Appl. Math. Sci., Vol. 99, New York: SpringerVerlag, 1993.
32. R. Datko, Two examples of ill-posedness with respect to time delays revisited, IEEE Trans. Autom. Control, 42: 511, 515, 1997.
6. M. A. Cruz and J. K. Hale, Asymptotic behavior of neutral functional differential equations, Arch. Ration. Mech. Anal., 34: 331
353, 1969.
7. D. Henry, Linear autonomous neutral functional differential
equations, J. Differ. Equ., 15: 106128, 1974.
8. Yu. M. Repin, Quadratic Lyapunov functionals for systems with
delay (in Russian), Prikl. Mat. Mekh., 29: 564566, 1965.
9. B. S. Razumikhin, Applications of Lyapunovs method to problems in the stability of systems with a delay (in Russian), Autom.
Telemekh., 21: 740749, 1960.
10. A. Halanay, Differential Equations, New York: Academic, 1966.
11. E. A. Andreeva, V. B. Kolmanovskii, and P. E. Shaikhet, Control
of Systems with After Effect, Moscow: Nauka, 1992.
12. X. Li and J. Yong, Optimal Control Theory for Infinite Dimensional
Systems, Boston: Birkhauser, 1995.
13. C. Foias, A. Tannenbaum, and G. Zames, Weighted sensitivity
minimization for delay systems, IEEE Trans. Autom. Control, 31:
763766, 1986.
Addison-Wesley, 1989.
R. DATKO
Georgetown University
J. K. HALE
Georgia Institute of Technology
15. R. Datko, An example of an unstable neutral differential equation, Int. J. Control, 38: 263267, 1983.
16. K. Cooke and J. Ferreira, Stability conditions for linear retarded
functional differential equations, J. Math. Anal. Appl., 96: 480
504, 1983.
17. S. Saks and A. Zygmund, Analytic Functions, Warsaw: Polish Scientific Publishers, 1971.
18. V. M. Popov and A. Halanay, On the stability of nonlinear control
systems with time-lag (in Russian), Autom. Telemekh., 23: 31
38, 1962.
19. G. Doetsch, Handbuch der Laplace-Transformation, Birkhauser:
Basel, 1956, Vol. 3.
20. R. Gabasov and F. M. Kirillova, The Maximum Principle in the
Theory of Optimal Control (in Russian), Mrnsk: Minsk Science
and Technology, 1974.
21. B. von Keulen, H-Control for Distributed Parameter Systems: A
State-Space Approach, Birkhauser: Basel, 1993.
22. R. Datko, Two questions concerning the boundary control of certain elastic systems, J. Differ. Equ., 92: 2744, 1991.
23. R. Datko, The point spectrum of some LQR problems. Appl. Math.
Optim., 31: 8599, 1995.
24. J. Louis and D. Wexler, The Hilbert space regulator problem and
operator Riccati equation under stabilization, Ann. Soci. Sci.
Brux., 105: 137165, 1991.
25. K. Yoshida, The Hopf bifurcation and its stability for semilinear
diffusion equation with time delay arising in ecology, Hiroshima
Math. J., 12: 321348, 1982.
26. W. E. Fitzgibbon, Semilinear functional differential equations in
Banach space, J. Differ. Equ., 29: 114, 1978.
DIGITAL CONTROL
445
DIGITAL CONTROL
The revolutionary advances in computer technology today
have made it possible to replace conventional controllers with
digital computers. Digital control thus refers to the control
scheme in which the controller is a digital device, generally a
digital computer. This means that we can make use of a much
more advanced control logic and versatility than those made
possible with conventional analog controllers. On the other
hand, we also need an interface that connects a computer
with real plants. In particular,
Measurement is made at discrete instants in time
Data must be spatially discretized to allow digital data
handling
In other words, digital controllers can handle data that are
discretized both in time and space. The former discretization
is usually referred to as sampling and the latter quantization.
These two features place digital control systems outside the
scope of the usual linear, time-invariant control systems.
(There is also the problem of saturation effect when controllers have a fixed word length. But this problem is much less
studied in the context of digital control.)
To see the situation more concretely, consider the unityfeedback digital control system shown in Fig. 1. Here r is the
reference signal, y the system output, and e the error signal.
These are continuous-time signals. The error e(t) goes
through the sampler (or an A/D converter) S . This sampler
r+
ed
e
S
ud
C(z)
y
H
C(z)
Figure 1. A unity-feedback digital control system consisting of a continuous-time plant P(s), discrete-time controller C(z), sampler S and
a hold device H .
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
446
DIGITAL CONTROL
Rf
;
;
; ;
u[t]
2R
4R
8R
Figure 4. A D/A converter is constructed with an operational amplifier, switching, and resistors.
reads out the values of e(t) at every time step h called the
sampling period, and produces a discrete-time signal ed[k],
k 0, 1, 2, . . .. In this process, a quantization error (due to
round-off) as shown in Fig. 2 occurs. The sampling operator
S acts on a continuous-time signal w(t), t 0 as
S (w)[k] := w(kh),
k = 0, 1, 2, . . .
(The quantization effect is omitted here.) The discretized signal is then processed by a discrete-time controller C(z) and
becomes a control input ud. This signal then goes through another interface H called a hold device or a D/A converter to
become a continuous-time signal. A typical example is the
zero-order hold, where H simply keeps the value of a discretetime signal w[k] as a constant until the next sampling time:
(H (w[k]))(t) := w[k],
for kh t < (k + 1) h
z-TRANSFORM
We start with a fundamental description of systems and se
denote a sequence with values in some
quences. Let w[k]k0
vector space X. Typically, X is the n-dimensional vector space
n, but we will later encounter an example where X is not
w[k]zk
k=0
Figure 3. A simple sample-hold combination maps a continuoustime signal to a piecewise step function.
DIGITAL CONTROL
(w u)[k] :=
k
w[k j]u[ j]
j=0
y(t) = Cx(t)
w[k 1]k0.
The z-transformation plays the role of the Laplace transformation in the continuous-time case. As with Laplace transforms, it is useful to consider the substitution of a complex
number to the variable z. For example, the geometric se
quence kvk0
has the z-transform
k zk v =
k=0
zv
z
x[k + 1] = x((k + 1)h) = e
Ah
0
h
= e Ah x[k] +
0
x(kh)zk
z
z eh
x[k] +
k=0
Z [et ](z) =
(4)
The first objective is to give a description of this plant at sampling instants t kh, k 0, 1, 2, . . .. By the zero-order hold,
the input to the plant for kh t (k 1)h is the constant
ud[k]. Suppose that the state of the plant at t kh is x[k]
x(kh). Then by integrating Eq. (4) from kh to (k 1)h, we
obtain
(1)
Z [w u] = Z [w]Z [u]
447
(3)
zx = Ax + Bu
y = Cy + Du
Solving this, we have
eA Bd
0
y = [1
0] x
y[k] = [1
0] x[k]
448
DIGITAL CONTROL
1
P(s)
ds
2 j 1 ehs z1
P(s)
Res
=
1 ehs z1
poles of P(s)
Z [x](z, m) :=
x(kh + mh)zk
(9)
k=0
P(s)emhs
1
ds
2 j C 1 emhs z1
P(s)emhs
Res
=
1 emhs z1
poles of P(s)
Z [P](z, m) =
Z [P](z) =
(6)
Once the sample-point behavior is specified by the procedure above, it is easy to give a description of the intersample
behavior of the output or the state. Suppose, for example, that
the state of the plant P(s) takes values x[k] at t kh, k 0,
1, 2, . . .. By integrating Eq. (4) from t kh to t kh
(0 h), we get
x(kh + ) = e
x[k] +
e
0
A( )
Bud [k] d
(7)
and
y(kh + ) = Cx(kh + )
(8)
We briefly review the classical design methods and their limitations. The first is a design method based on the continuoustime design.
Continuous-Time Design
A simple, straightforward method is to employ a continuoustime design, obtain a continuous-time controller, and then
convert the controller to a discrete-time system via some kind
of discretization. Let Cc(s) be a continuous-time controller.
Typical discretization methods are the following:
Use the Tustin (bilinear) transformation:
C(z) = Cc
2
z1
h z+1
Employ the backward difference (z 1)/hz for approximating the differential operator s.
Approximate Cc(s) by using the sample/hold equivalent
Eq. (5).
Among these, the first method is most commonly used.
It is well known that the Tustin transformation preserves
stability: if Cc(s) is a stable transfer function (in the sense of
continuous-time systems), then the transformed function
Cc(2(z 1)/h(z 1)) gives a stable (in the sense of discretetime systems) discrete-time transfer function. Although this
is a great advantage in signal processing, care must be exercised in control system design, because this property does not
guarantee the closed-loop stability. In fact, as h becomes
larger, there is even a case in which the closed-loop stability
is violated (see the example in the section entitled H Design). This is because the original continuous-time design
does not usually take account of the sampling period. To take
care of this, one has to pay more attention to various robustness properties, such as gain and phase margins, and so
DIGITAL CONTROL
on. To discuss such properties, frequency domain considerations are highly desirable.
However, the notion of frequency response is not readily
available. To see the situation, let C(z) be a discrete-time
transfer function. Suppose that a sinusoid ejt is applied after
. The response is
exactly the same sampled values ekjhk0
then governed by the same C(ejh). This means that sampling
does not have enough resolution to distinguish all these sinusoids, and the notion of phase, which depends on the response
against sinusoids, is unclear in such a sampled-data control
setup. Another way of seeing this is to note that ejh ejh
and hence C(ej(2/h)h) C(ejh). This means that beyond the
frequency /h, the same gain characteristic repeats periodically and C(ejh) cannot be treated as the same frequency response concept as in the continuous-time case. This is related
to the notion of aliasing, which we examine in more detail in
the section Modern Approach.
It may be still possible to execute an elaborate continuoustime design that also works well in the sampled-data setting
by looking more closely into the nature of the Tustin transformation. However, in such a method, a systematic design
method such as H design theory is difficult to apply. Furthermore, one needs a more concrete understanding of the phenomena above, and this is much better done in the modern
approach treated in the subsequent sections.
Discrete-Time Design
Yet another classical approach is based on the pulse transfer
function Pd(z). As far as stability is concerned, one can deal
only with Pd(z). It was also recognized that sampled-data control can achieve performance that is not possible with linear,
time-invariant, continuous-time controller. For example, the
so-called deadbeat control achieves the property that the output (or state) settles exactly to zero after a finite time period.
This is done by placing all poles of the closed-loop system to
zero; the output, at least at sampled instants, then becomes
zero after a finite timea performance not possible with continuous-time controllers.
It should be, however, noted that such a classical treatment also shares the weakness of the classical transfer function approach. Namely, it did not take account of hidden
polezero configurations. In particular, it was observed that
merely settling the output might sometimes induce very large
intersample ripples.
It was Kalman and Bertram (1) who introduced the state
space approach for sampled-data systems. As we have already
seen, the sample-time inputoutput relation is described by
x[k + 1] = Ad x[k] + Bd ud [k] = e Ah x[k] +
449
Let us consider the stabilization by state feedback. If we employ a sampled state feedback
ud [k] = Kx[k]
then by Theorem 3 the stability of the closed-loop system is
determined by the spectrum of
A d Bd K
(10)
sin(1 + 20 )t
r+
ed
e
S
e2h1
z e2h
ud
H
s2
1
+1
h
0
e A Bd ud [k]
450
DIGITAL CONTROL
0.5
SAMPLING THEOREM
0
0.5
1
43
43.2
43.4
43.6
43.8
44
Time (sec)
Figure 6. The simulation of Figure 5 shows that the input sin(1
20)t does not yield a sinusoid at the same frequency, and large intersample ripples result.
f (kh)zk
k=
We have extended the definition in a natural way to the negative ks. We also recall that the multiplication by z1 is the
right shift operator. Since the kth signal value f(kh) is placed
at t kh, this right shift corresponds to the right shift by
time length h in the continuous-time domain. It is also well
known that in the Laplace transform domain the right shift
operator by h is represented by the multiplication by ehs.
Therefore, it is natural to represent the Laplace transform of
f (kh)e khs
k=
f (kh)(t kh)
(11)
k=
(t kh)
(12)
k=
khs
k f(kh)e
just as well. How much can we recover the
original signal f(t) out of this piece of data?
If we impose no condition on f(t), then the solution is
clearly nonunique. There is infinite freedom in the intersampling periods while passing through f(kh), k . A typical
solution is obtained by assuming that f(t) is band-limited;
that is, its Fourier transform is zero outside a bounded interval. This is the content of the following Shannon sampling
theorem.
Theorem 6. Let f be a continuous function that is Fourier
transformable. Suppose that its Fourier transform is identi-
DIGITAL CONTROL
f (nh)
n=
sin (t/h n)
(t/h n)
(t nh)
n=
It follows that
f (nh)(t nh)
2
h
n=
=F
f (t)
n=
1
f()
2
1
f()
h
2n
h
n=
1
2n
f
h n=
h
f( )
f( 2 /h)
/h
2n/h)
f(
n=
f (nh)(t nh)
n=
f (nh)sinc(t nh)
n=
sinc t =
2n
h
2n
2
t
h n=
h
= h F []
()
(14)
(t nh)
1
2n
=
f t
h n=
h
f (t) = F [f()]
=F
n=
(13)
451
/h
sin t/h
t/h
452
DIGITAL CONTROL
MODERN APPROACH
x[k + 1] = e
Ah
e A(h ) Bu[k]( ) d
x[k] +
(17)
f [k]( ) = f (kh + ),
The difference here is that the lifted input u[k]( ) need not
be a constant on (kh, (k 1)h), and the right-hand side integral gives an operator
L2 [0, h) Rn : u()
e A(h ) Bu( ) d
x[k]( ) = e A x[k] +
e A( ) Bu[k]( ) d
y[k]( ) = Ce A x[k] +
Ce A( ) Bu[k]( ) d
0 <h
(16)
y(t) = Cx(t)
f(t)
A :
Rn Rn : x e Ah x
B:
L2 [0, h) Rn : u
e A(h ) Bu( ) d
C :
Rn L2 [0, h) : x Ce A x
D:
L2 [0, h) L2 [0, h) : u
(19)
Ce A( ) Bu( ) d
Thus the continuous-time plant in Eq. (16) can be described by a time-invariant discrete-time model. Once this is
done, it is entirely routine to connect this expression with a
discrete-time controller, and hence sampled-data systems (for
example, see Fig. 1) can be fully described by time-invariant
discrete-time equations, this time without sacrificing the intersampling information. We will also denote this overall
equation abstractly as
x[k + 1] = A x[k] + Bu[k]
2h
3h
4h
(18)
0 h
0 h
0 h
0 h
(20)
DIGITAL CONTROL
present situation without any change. For example, the transfer function is defined as
G(z) := D + C (zI A )1 B
When an input u(z) is applied to system in Eq. (20), its zeroinitial state output is given by G(z)u(z). Note also that operator A in Eq. (19) is a matrix, and hence A in Eq. (20) is also
a matrix. This means that the stability of Eq. (20) can be
tested by the poles of G(z). It is stable if G(z) is analytic for
z : z 1.
Steady State and Frequency Response
The time-invariance of the lifted model Eq. (20) naturally
yields a definition of stead-state and frequency responses.
Let G(z) be a stable transfer function as defined above.
Take a sinusoid ejt as an input. Its lifted image is
{e jkh e j }
k=0
According to Eq. (1), the z-transform of this sequence is
ze j
z e jh
( ) =
G(z)
ze
zG(e )e
=
z e jh
z e jh
jh
an e j n
(24)
n=
with
453
+ G(z)e
an =
1
h
1
(
jn )
h
e j n ( ) d =
(25)
2 = h
|an |2
(26)
n=
(21)
as k .
Unless ejh 1, the asymptotic response above is really not
in steady state. However, its modulus G(ejh)ej remains invariant at each sampling time. The change at each step is a
phase shift induced by the multiplication by ejh. This explains
why the ripples in Fig. 6 look similar but not really the same
in different sampling periods.
This observation motivates the following definition:
Definition 1 Let G(z) be the transfer function of the lifted
system as above. The frequency response operator is the operator
G(e
jh
) : L [0, h) L [0, h)
2
sup
vL 2 [0, h)
G(e jh )v
v
jkh
G(e
jh
)[e
v
e j
(27)
j
]v
g
n e j n
n=
e jkh G(e jh )[v] = e jkh G(e jh )
(22)
= n=
g
n e j n v
454
DIGITAL CONTROL
= n=
g
n e j n t v
n=
g
n ()v
e j n t
(28)
This means that the response against vejt is again expressible as an infinite sum of all such aliased signals. It
should be intuitively clear that the largest gain among them
again gives the gain of the frequency response, when such
signals are equipped with norm (nv2)1/2. This isometric correspondence is guaranteed by the Parseval identity Eq. (26).
This is the viewpoint adopted in Refs. 14 and 15 to discuss
the frequency response of sampled-data systems; see also Ref.
16. It is also closer to the classical treatment based on the
impulse modulation (3,7).
Gain Computation
The gain function G(ejh) is given as the operator norm at each
frequency, and its computation is primarily an infinite-dimensional problem. However, for most of the practical purposes,
it can be computed as the maximal singular value (17).
Our problem is thus reduced to that of solving the singular
value equation
[ 2 I G G(e jh )]w = 0
y(t) = C2 xc (t)
Here w is the exogenous input, u(t) is the control input, y(t)
is the measured output, and z(t) is the controlled output. The
controller is of the following form:
xd [k + 1] = Ad xd [k] + Bd S y[k]
v[k] = Cd xd [k] + Dd S y[k]
(29)
u[k]( ) = H( )v[k]
where H() is a suitable hold function. This is shown in Fig.
9. The objective here is to design or characterize a controller
that achieves a prescribed performance level > 0 in such
a way that
Tzw <
(31)
jh
w
A
C1
C2
B1 B2
D11 D12
0
0
x = A x + Bw
p = e jh A p + C (C x + Dw)
( 2 D D)w = e jh B p + D C x
H( )
jh
I
0
BR1
B
A + C DR1
B
A + BR1
D C
1
C (I + DR D )C
Ad
Cd
!
0
x
=0
I
p
(30)
Bd
Dd
Figure 9. Generalized plant construction of a sampled-feedback system where z denotes the controlled output, y is the measured output,
w is the exogenous input, and u is the control input.
DIGITAL CONTROL
Amplitude
455
0.5
0
0.5
1
1.5
SOME EXAMPLES
2
To see the power of the modern design methods, let us consider two design examples. We start with the H design.
10
15
20
25
30
Time (s)
Figure 11. Time responses for h 0.1 by sampled-data (solid) and
continuous-time (dash) H designs do not show much difference.
H Design
Consider the unstable second-order plant
P(s) := C p (sI A p )1 B p =
1
s2 0.1s + 1
1
hs + 1
In the continuous-time design, the antialiasing filter is bypassed. On the other hand, it is inserted in the sampled-data
design to make the total design well posed.
1.5
Q1/2
E
zc
wc
R1/2
N
Amplitude
1
0.5
0
0.5
1
1.5
Faa
Cp
(sI Ap)1
Bp
10
15
20
25
30
Time (s)
Figure 10. Generalized plant for sampled-data and continuous-time
H design.
456
DIGITAL CONTROL
Frequencygain response
50
40
30
0
Gain (dB)
20
10
0
10
30
101
100
Frequency (rad/s)
101
Figure 13. Frequency response plots for h 0.55 support the observation in Fig. 11.
5
0
5
Gain (dB)
Time
20
10
15
20
25
30
35
0.01
0.1
Figure 15. Time responses for sampled-data (solid) and discretetime (dash) design show a clear difference.
10
Frequency (rad/s)
Figure 14. Frequency response plots show the difference between
the sampled-data (solid), discrete-time (dashed) difference; the thin
dotted curve shows the frequency response with intersample behavior ignored.
BIBLIOGRAPHICAL NOTES
For classical treatments of sampled-data control, it is instructive to consult Refs. 2, 3, and 7. The textbooks (24,25) cover
both classical and modern aspects of digital control. For dis-
DIGITAL CONTROL
BIBLIOGRAPHY
19.
20.
21.
22.
23.
24.
25.
26.
3. J. R. Ragazzini and G. F. Franklin, Sampled-Data Control Systems, New York: McGraw-Hill, 1958.
27.
4. R. E. Kalman, Y. C. Ho, and K. Narendra, Controllability of linear dynamical systems, Contrib. Differ. Equations, 1: 189213,
1963.
28.
5. T. Chen and B. A. Francis, Optimal Sampled-Data Control Systems, New York: Springer, 1995.
29.
30.
31.
32.
33.
34.
11. Y. Yamamoto, New approach to sampled-data systems: A function space method, Proc. 29th CDC, 1990, pp. 18821887.
35.
36.
13. B. Friedland, Sampled-data control systems containing periodically time varying members, Proc. 1st IFAC Congr., 1961, pp.
361367.
37.
14. M. Araki, Y. Ito, and T. Hagiwara, Frequency response of sampled-data systems, Automatica, 32: 483497, 1996.
38.
39.
40.
457
YUTAKA YAMAMOTO
Kyoto University
458
DIGITAL FILTERS
DIGITAL CONTROL SYSTEMS DESIGN. See DISCRETE TIME SYSTEMS DESIGN METHODS.
602
, n = 0, 1, . . .
server j with probability qi, j and leaves the network with probM
ability qi,0. We have j0 qi, j 1, i 1, 2, , M. The service
time of server i is exponentially distributed with mean si
1/i, i 1, 2, , M.
The system state is n (n1, n2, , nM), where ni is the
number of customers in server i. Let i be the arrival rate of
the customers to server i. Then
i = 0,i +
M
j q j,i
i = 1, 2, . . ., M
j=1
p(n) = p(n1 , n2 , . . ., nM ) =
M
p(nk )
k=1
with
n
p(nk ) = (1 k )k k
k =
k
k
k = 1, 2, . . ., M
A load-independent closed Jackson (Gordon-Newell) network is similar to the open Jackson network described above,
except that there are N customers circulating among servers
M
according to the routing probabilities qi, j, k1 qi,k 1, i 1,
M
2, , M. We have k1 nk N. We consider a more general
case: the service requirement of each customer is exponential
with a mean 1; the service rates, however, depend on the
number of customers in the server. Let i,ni be the service rate
of server i when there are ni customers in the server, 0
i,ni , ni 1, 2, , N, i 1, 2, , M. We call this a
load-dependent network. In a load-independent network,
i,ni i for all ni, i 1, 2, , M.
The state of such a network is n (n1, n2, , nM). We
use ni, j (n1, , ni 1, , nj 1, , nM), ni 0, to
denote a neighboring state of n. Let
(nk ) =
603
1 if nk > 0
0 if nk = 0
and let
(n) =
M
(nk )k,n
k=1
Then the ow balance equation for the steady state probability p(n) is
(n)p(n) =
M
M
i=1 j=1
(n j )i,n
i +1
yi =
M
j=1
q j,i y j
j = 1, 2, . . ., M
604
Ai (k) =
b1
k
i, j
b2
Mach. 2
Mach. 1
i = 1, 2, . . ., M
b3
j=1
Exit
Gm (n) =
m
n 1 ++n m =n i=1
yn i
Ai (ni )
Then we have
n
p(n) =
M
yi i
1
GM (N) i=1 Ai (ni )
Gm (n) =
m
n 1 ++n m =n i=1
xi i
and
p(n) =
M
1
n
x i
GM (N) i=1 i
(1)
where xi yi /i yisi, i 1, 2, , M.
There are a number of numerical methods for calculating
p(n) and the steady state performance, among them are the
convolution algorithm and the mean value analysis (7); in addition, analytical expressions exist for the normalizing constant GM(N) (10). For more about queueing theory, see, for
example, Refs. 11 and 12.
One typical example of the queueing model is the resource
sharing problem. Consider the case where M resources are
shared by N users and each resource can be held by only one
user at any time. Every time a user grasps resource i, it holds
the resource for a random time with si, A user, after the completion of its usage of resource i, requests the hold of resource
j with a probability qi, j. This problem can be modeled exactly
as a closed queueing network with N customers and M
servers. This model can be successfully used in analyzing the
performance of packet switches, where the users are the
head-of-line packets and the resources are the channels, and
the performance of data-base systems, where the users are
programs and the resources are data records.
Petri Nets
Many DEDS consist of components [e.g., central processing
units (CPUs), disks, memories, and peripheral devices in computer systems; and machines, pallets, tools, and control units
in manufacturing systems] that are shared by many users
and exhibit concurrency. This feature makes Petri nets a suitable model for DEDS.
In a graphical representation, the structure of a Petri net
is dened by three sets: a set of places P p1, p2, , pn,
a set of transitions T t1, t2, , tm, and a set of arcs. A
place is represented by a circle and a transition by a bar. An
arc is represented by an arrow from a transition to a place or
a place to a transition. A place is an input (output) to a transition if an arc exists from the place (transition) to the transition (place).
The dynamic feature of a Petri net is represented by tokens, which are assigned to the places. Tokens move from
place to place during the execution of a Petri net. Tokens are
drawn as small dots inside the circle representing the places.
The marking of a Petri net is a vector M (m1, m2, ,
mn), where mi is the number of tokens in place pi, i 1, 2,
, n. A marking corresponds to a state of the system. A
Petri net executes by ring transitions. A transition is enabled if its every input place contains at least one token.
When a transition is enabled, it may re immediately or after
a ring delay, which can be a random number. The ring delay is used to model the service times. When a transition res,
one token is removed from each input place and one token
added to each output place. Thus, the number of tokens in a
place and in the system my change during the execution. In
addition to the arcs described above, another type of arc,
called the inhibitor arc, is often used to model the priority
among services. An inhibitor arc is drawn from a place to a
transition, with a small circle at its end. When an inhibitor
arc is used, if there is at least one token in the place, the
transition cannot re.
As an example, we consider the reentrant line (13) shown
in Fig. 1. The system consists of two machines and three buffers. Work pieces arrive at buffer b1 with rate and get service
from machine 1 with rate 1; after the service in b1, the piece
moves to buffer b2 and gets service at machine 2 with rate
2; after the service in b2, the piece reenters machine 1 at
buffer b3 and receives service with rate 3. Machine 1 can
serve one piece at a time, and pieces in b3 have a nonpreemptive higher priority than those in b1.
The Petri net model for the system is shown in Fig. 2. In
the gure, places bi, i 1, 2, 3, represent the buffers, and
transitions pi, i 1, 2, 3, represent the service processes of
the pieces in the three buffers. If there is a token in places
mi, i 1, 2, then machine i is available; if there is a token in
si, i 1, 3, then the work piece in buffer bi is under service.
It is clear that machine 1 is shared by the pieces in both b1
and b3, and the inhibitor arc from b3 to p4 models the priority.
(For more about Petri net theory and applications, see Refs.
4 and 1416).
The state process of a queueing system or a Petri net can
be modeled as a Markov process, or more generally, as a generalized semi-Markov process. In this sense, both Markov processes and GSMPs are more general than queueing systems
and Petri nets; however, these general models do not enjoy
the structural property that queueing systems and Petri nets
possess. In fact, the GSMP model is a formal description of
the evolution mechanism of a queueing system. Readers are
referred to Refs. 5 and 8 for a discussion of GSMP.
b1
in
p1
b2
p2
K
s1
(2i)
(2i+1)
p4
m2
p5
3
k = 1, 2, . . .
where
a
A=
d
H
d
xk+1 = Ak xk
(2)
5
3
=4
2
1
1
=4K
2
Perturbation analysis of DEDS is a multidisciplinary research area developed since early 1980s, with the initial
work of Ho et al. (19). PA provides the sensitivities of
performance measures with respect to system parameters
by analyzing a single sample path of a DEDS. This area is
promising because of its practical usefulness. First, compared with the standard procedure, which uses the difference of performance measures with two slightly different
values for every parameter, this technique saves a great
amount of computation in simulation, because PA algorithms can provide the derivatives with respect to all the
parameters by using only a single simulation run. In addition, PA estimates are more accurate than those obtained
from nite differences because the latter may encounter
numerical problems caused by dividing two small numbers.
Second, PA can be applied to on-line performance optimization of real-world systems by observing a sample path of
an existing system; for these systems, changing the values
of their parameters may not be feasible.
Cao (20) observed that the simple PA algorithms based on
a single sample path, called innitesimal perturbation analysis (IPA), in fact yield sample derivatives of the performance;
although these sample derivatives are unbiased or strongly
consistent for many systems, this is not the case for many
others. This insight has set up two fundamental research directions: to establish IPA theory, including the proof of convergence of IPA algorithms, and to develop new algorithms
for systems where IPA does not work well. After the hard
work of many researchers in more than one decade, the theory for IPA is relatively mature, and many results have been
obtained for problems where IPA does not provide accurate
estimates.
Infinitesimal Perturbation Analysis
1
M=
3
1
2
PERTURBATION ANALYSIS
p3
ak
xk =
dk1
1
0
b3
Exit
0
=
3
2
=
1
m1
s3
605
606
i,1
i,1 i,2
(3)
t1
and
t2
( ) = lim L (, )
L
w.p.1.
(4)
where E denotes the expectation with respect to the probability measure P . We assume that both the mean and limit in
Eq. (3) and Eq. (4) exist. Thus L(, ) is an unbiased estimate
of L() and an strongly consistent estimate of (), respectively.
The goal of perturbation analysis is to obtain the performance derivative with respect to by analyzing a single sample path (, ). That is, we want to derive a quantity based on
a sample path (, ) and use it as an estimate of L()/ or
()/.
Given a single sample path, the realization of the random
vector is xed. Therefore, we x and consider L(, ) as a
function of . This function is called a sample performance
function. Now, we consider the following question: given a
sample path (, ), can we determine the sample path (
, ) with the same and / 1? If we can, then we can
get the performance for the perturbed system, L( , ),
and furthermore, the derivative of the sample performance
function:
( + , ) L ( ), )
(, ) = lim L
0
L
(5)
[L (, )] =
E[L (, )]
(6)
or
lim
[L (, )] =
lim [L (, )]
All the papers published in the area of IPA deal with these
two basic issues. Roughly speaking, the interchangeability requires that the sample performance function L(, ) be continuous with respect to . General conditions can be found in
Refs. 20 and 8.
The algorithms for obtaining sample derivatives are called
IPA algorithms. Given a nite-length sample path (, ), we
rst, by applying IPA algorithms, ctitiously construct a sample path for the DEDS with a slightly changed parameter and
the same random vector, ( , ), called a perturbed sample
path. The derivative of the performance with respect to can
be obtained by comparing these two sample paths, the original one and the perturbed one.
The principles used in IPA to determine the perturbed
path are very simple. We take closed queueing networks as
an example. The basic idea is that a change in parameter
(say, a mean service time) will induce changes of the service
completion times, and a change of a customers service completion time will affect the other customers service completion times. IPA rules describe how these changes can be determined.
Figure 3 illustrates a busy period of a server, say server i,
in a queueing network. Let Fi(s, ) be its service time distribution. The service time of its kth customer is
si,k = Fi1 (i,k , ) = sup{s : F (s, ) i,k },
where i,k, k 1, 2, , are uniformly distributed random
variables on [0, 1). With the same i,k, in the perturbed sys(i,k, ). Thus, the
tem, the service time changes to F 1
i
service time increases by
=F (s , )
i,k
(7)
(8)
i,k
si
s
si i,k
(9)
t2
t3
Server 1
Server 2
t0
t1
t4
busy period will be delayed by j1 i, j, with i, j being determined by Eq. (8) or Eq. (9). This can be summarized as follows: a perturbation of a customer will be propagated to the
next customer in the same busy period; the perturbation of a
customer equals the perturbation generated in its service period plus the perturbation propagated from the preceding customer.
If a perturbation at the end of a busy period is smaller
than the length of the idle period following the busy period,
the perturbation will not affect (i.e., will not be propagated
to) the next busy period, because the arrival time of the next
busy period depends on another servers service completion
time.
A perturbation at one server may affect other servers
through idle periods. To see how servers may affect each
other, we study the evolution of a single perturbation. In Fig.
4, at t1, server 1 has a perturbation , and before t1, server 2
is idle. At t1, a customer arrives from server 1 to server 2.
Because server 1s service completion time is delayed by ,
server 2s service starting time will also be delayed by ; and
as a result, its service completion time will also be delayed by
the same amount. We say the perturbation is propagated
from server 1 to server 2 through an idle period (server 2 has
the same perturbation as server 1 after t1). At t3, this perturbation is propagated back to server 1.
In summary, if a perturbation is smaller than the lengths
of idle periods (we say that the original and the perturbed
paths are similar), then the evolution of this perturbation on
the sample path can be determined by the following IPA perturbation propagation rules:
k
1. A perturbation of a customer at a server will be propagated to the next customer at the server until it meets
an idle period, and
2. If, after an idle period, a server receives a customer
from another server, then after this idle period the former will have the same perturbation as the latter (the
perturbation is propagated from the latter to the
former).
The perturbation generation rule describes how perturbations are generated because of a change in the value of a parameter; perturbation propagation rules describe how these
perturbations evolve along a sample path after being generated. Combining these rules together, we can determine the
perturbed path.
To apply the propagation rules, the size of the perturbation
at the end of each busy period should not be larger than the
length of the idle period that follows, and the size of the perturbation of a customer that terminates an idle period should
607
not be larger than the length of the idle period; otherwise the
idle period in the original sample path will disappear in the
perturbed one and the simple propagation rules illustrated by
Fig. 4 no longer hold. It is easy to see that for any nitelength sample path (, ), we can always (with probability
one) choose a that is small enough (the size depends on )
such that the perturbations of all customers in the nite sample path are smaller than the shortest length of all the idle
periods in the sample path. This explains the word innitesimal in IPA. Therefore, we can always use IPA propagation
rules to get the perturbed sample path and the sample derivative.
The perturbed sample path is completely determined by
the perturbations of the servers. Given a sample path of a
single-class closed Jackson network of M servers with sm, m
1, 2, , M, being the mean service times, the perturbations of the perturbed system with si changed to si si (with
i xed) can be determined by the following algorithm.
IPA Algorithm for Closed Jackson Networks
0. Create a vector v (v1, v2, , vM); set its initial value
v (0, 0, , 0)
1. At the kth, k 1, 2, , service completion time of
server i, set vi : vi si,k
2. If on the sample path, a customer from server j terminates an idle period of server l, then set vl : vj.
Note that for simplicity in the algorithm we add si,k, instead
of (si /si) si,k, to the perturbation vector. Thus, the perturbation of server m, m 1, 2, , M, is (si /si) vm, with vm being
determined by the algorithm. We shall see that the term
si /si is eventually cancelled in Eq. (11).
The sample derivative can be obtained from these perturbations. Let the sample performance measure be
L( f ) =
1
L
TL
f [N(t)] dt
(10)
w.p.1
608
H
L
and
L( f ) (si , i ) =
si H
si L
L( f ) (si , i )
si
H
=
(f)
si
H
L (si , )
(11)
L( f ) (si , i )
si
si ( f )
=
L ( f ) (s , )
si
( f ) si
i
L
lim
w.p.1
Similar algorithms and convergence results have been obtained for open networks, networks with general service time
distributions, and networks in which the service rates depend
on the system states (6). Glasserman (8) studied various IPA
algorithms and their unbiasedness and strong consistency in
the GSMP framework.
Extensions of IPA
For a sample derivative (i.e., the IPA estimate) to be unbiased
and strongly consistent, usually requires that the sample
function be continuous. This request, however, is not always
satised. A typical example illustrating the failure of IPA is
the two-server multiclass queueing network discussed in Ref.
(21); later, Heidelberger et al. (22) discussed in detail a few
extensions of the example.
In the past decade, many methods have been proposed to
extend IPA to a wider class of problems. Each of these
methods has some success on some problems at the cost of
increasing analytical difculty and computational complexity.
We shall review only briey the basic concepts of these
methods.
Smoothed Perturbation Analysis. The idea of smoothed perturbation analysis (SPA) is to average out the discontinuity
over a set of sample paths before taking the derivative and
expectation. To illustrate the idea, we rst write the expected
value of L(, ) as
L ( ) = E[L (, )] = E{E[L (, ) | Z ]}
where Z represents some random events in (, F , P ). Let
L(, Z ) E[L(, )Z ], then
L ( ) = E[L (, Z )]
and Eq. (6) becomes
[L (, Z )] =
E[L (, Z )]
E
(12)
SPA attempts to nd a suitable Z so that the average performance function L(, Z is smooth enough and the interchangeability of Eq. (12) holds, even if Eq. (6) does not. If this
is the case and the derivative of the conditional mean L(,
Z ) can be calculated, then / L(, Z ) can be used as an
unbiased estimate of / L().
The method was rst proposed in Ref. (23); a recent book
Ref. (5) contains a detailed discussion about it. The main issues associated with this method are that it may not be easy
to calculate / E[L(, Z )] and that the computation effort
required may be large.
Finite Perturbation Analysis. The sample derivative does not
contain any information about the jumps in the performance
function. This is because as goes to zero, the event sequence of any perturbed sample path is the same as that of
the original path (two paths are similar). Thus, with IPA, we
do not study the possibility that because of a parameter
change , two events may change their order. In nite perturbation analysis (FPA), a xed size of is assumed. For
any xed , the event sequence in the perturbed path may
be different from that in the original path. FPA develops some
rules that determine the perturbation when the event order
changes. The FPA algorithm is more complicated than IPA,
and it is usually approximate since only order changes between adjacent events are taken into account (9).
Sample Path Constructability Techniques. Given the nature
of IPA, it cannot be applied to sensitivities with respect to
changes of a xed size or changes in discrete parameters. Motivated by the principles of IPA, we ask the following question: Given a sample path of discrete event system under parameter , it is possible to construct a sample path of the
same system under a different parameter ? This problem is
formulated as the sample path constructability (7). Normally,
such constructability requires that the sets of events and
states of the sample path to be constructed (with parameter
) belong to the sets of events and states of the original sample path. For example, one may construct a sample path for
an M/M/1/K 1 queue (where K 1 denotes the buffer size)
from a sample path of an M/M/1/K queue. Ref. 24 shows that
for some systems with additional computation such sample
path construction can be done even if some states in the sample path to be constructed do not appear in the original sample path.
Techniques in this class include augmented system analysis
(7,25), extended perturbation analysis (26), and the standard
clock approach (27).
Structural Infinitesimal Perturbation Analysis. Structural innitesimal perturbation analysis (SIPA) was developed to address the problem of estimating the performance sensitivity
with respect to a class of parameters such as the transition
probabilities in Markov chains. At each state transition, in
addition to the simulation of the original sample path, an extra simulation is performed to obtain a quantity needed to get
the performance sensitivity. It has been shown that the extra
simulation requires bounded computational effort, and that
in some cases the method can be efcient (28). It is interesting
to note that this approach can be explained by using the concept of realization discussed in the next subsection.
Rare Perturbation Analysis. Bremaud (29) studies the performance sensitivity with respect to the rate of a point process
and proposes the method of rare perturbation analysis (RPA).
The basic idea is that the perturbed Poisson process with
rate with 0 is the superposition of the original
Poisson process with rate and an additional Poisson process
with rate . Thus, in a nite interval, the difference between
the perturbed path and the original one is rare. The performance derivative is then obtained by studying the effect of
these rare but big (meaning nite) perturbations on the system performance. The case 0 is called the positive
RPA.
When 0, the perturbed Poisson process with rate
can be constructed by thinning the original Poisson process with the thinning probability / . That is, some arrival
points in the original process will be taken away. The performance derivative is then obtained by studying the effect of
the removal of these rare arrival points. This is called the
negative RPA. Others in this direction include Refs. 30
and 31.
Estimation of Second Order Derivatives. The single path
based approach can also be used to estimate the second order
derivatives of the performance of a DEDS by calculating the
conditional expectations. See Ref. 32 for GI/G/1 queues and
Ref. 33 for Jackson networks.
Others. In addition to the above direct extensions of IPA,
it also motivated the study of a number of other topics, such
as the Maclaurin series expansion of the performance of some
queueing systems (35), the rational approximation approach
for performance analysis (36), and the analysis of performance discontinuity (37).
Finally, besides the PA method, there is another approach,
called the likelihood ratio (LR) method (3840), that can be
applied to obtain estimates of performance derivatives. The
method is based on the importance sampling technique in
simulation. Compared with IPA, the LR method may be applied to more systems but the variances of the LR estimates
are usually larger than those of IPA.
Perturbation Realization
One important concept regarding the sensitivity of steady
state performance of a DEDS is the perturbation realization. The main quantity related to this concept is called
the realization factor. This concept may provide a uniform
framework for IPA and non-IPA methods. The main idea
is: The realization factor measures the nal effect of a
single perturbation on the performance measure of a DEDS;
the sensitivity of the performance measure with respect to
a parameter can be decomposed into a sum of the nal
effects of all the single perturbations induced by the parameter change.
Perturbation Realization For Closed Jackson Networks. Suppose that at time t 0, the network state is n and server i
obtains a small perturbation , which is the only perturbation
generated on the sample path. This perturbation will be propagated through the sample path according to the IPA propagation rules and will affect system performance. The realiza-
609
c
(f)
(n, i) = lim E
L
1
TL
TL
f [N (t)] dt
f [N(t)] dt
(13)
where TL and N(t) represents the quantities in the perturbed path.
A perturbation is said to be realized if at some time Tl
all the servers have the same perturbation ; it is said to be
lost if at some time Tl no server has any perturbation. It was
proved that in an irreducible closed network a perturbation
will be either realized or lost with probability one. The probability that a perturbation is realized is called the realization
probability.
Suppose that a perturbation is realized or lost at TL*. L*
depends on the sample path, that is, . If the perturbation is
lost, then f[N(t)] f[N(t)], for all t TL*; if it is realized,
then f[N(t)] f[N(t )] for all t TL*. Therefore, from the
Markov property, Eq. (13) becomes
c
(f)
n, i) = E
(n
1
TL
TL
f [N (t)] dt
f [N(t)] dt
(14)
where L* is a random number, which is nite with probability one.
Realization factors can be uniquely determined by a set of
linear equations (6). The steady state performance sensitivity
can be obtained by
si ( f )
n )c( f ) (n
n, i)
=
p(n
(I
)
si
all n
(15)
where I(n) 1 for all n and p(n) is the steady state probability of n.
A close examination reveals that the IPA algorithm provides a simple way for estimating the quantity alln
p(n)c(f)(n, i) on a single sample path. The theory has been extended to more general networks, including open networks,
state-dependent networks, and networks with generally distributed service times (6).
Perturbation Realization for Markov Processes. Consider an
irreducible and aperiodic Markov chain X Xn; n 0 on a
nite state space E 1, 2, , M with transition probabilM
ity matrix P [pij]M
i1j1. Let (1, 2, , M) be the vector
representing its steady state probabilities, and f [f(1), f(2),
, f(M)]T be the performance vector, where T represents
transpose and f is a column vector. The performance measure
is dened as its expected value with respect to :
= E ( f ) =
M
i f (i) = f.
(16)
i=1
610
di j = E
[ f (Xn ) f (Xn )]|X0 = i, X0 = j
n=0
i, j = 1, 2, . . ., M
(17)
n
( )
(22)
where
n
( ) =
n n
n
( ),
( ), ,
( )
1
2
M
n1 n and n1 n2 .
Many results have been obtained in this direction. For example, Ref. 41 studied the optimization of J() T() C()
for a single server queues, where T() is the mean system
time, C() a cost function, and the mean service time. It was
proved that under some mild conditions, the RobbinsMonro
type of algoritm (22) converges even if we update using the
IPA gradient estimate at any random times (e.g., at every
customer arrival time). Other works include Refs. 42, 43, 44,
and 45.
The optimization procedures using perturbation analysis
have been applied to a number of real-world problems. Successful examples include the bandwidth allocation problem in
communications, (46,47), and optimization of manufacturing
systems (4852).
For performance optimization over discrete parameters, for
example, in problems of choosing the best transition matrix,
we may use the approach of realization matrix and potentials
discussed in the last section. It is interesting to note that in
this context, PA is equivalent to the Markov decision process
(MDP) approach. To see this, let be vector of the steady
state probability for the Markov chain with transition matrix
P. From the Poisson equation [Eq. (21)], it is easy to prove
(I P + e )g = f
= Qg
i, j = 1, 2, . . ., M
(18)
= QDT T
Q
(19)
= Qg.
Q
(20)
(21)
(23)
The right-hand side of Eq. (23) is the same as that of Eq. (20)
except is replaced by . In policy iteration of an MDP problem, we choose the P corresponding to the largest Qg
(P P)g (component-wise) as the next policy. This corresponds to choosing the largest /Q in PA, because all the
components of and are positive. Therefore, the policy iteration procedure in MDP in fact chooses the steepest direction
BIBLIOGRAPHY
1. Y. C. Ho (ed.), Dynamics of Discrete Event Systems, Proc. IEEE,
77: 1232, 1989.
2. P. J. Ramadge and W. M. Wonham, Supervisory control of a class
of discrete event processes, SIAM J. Control Optim., 25: 206
230, 1987.
3. W. H. Fleming, (chair), Future directions in control theoryA
mathematical perspective, Report of the Panel on Future Directions in Control Theory, 1988.
4. R. David and H. Alla, Petri nets for modeling of dynamic systemsA survey, Automatica, 30: 175202, 1994.
5. M. C. Fu and J. Q. Hu, Conditional Monte Carlo: Gradient Estimation and Optimization Applications, Norwell, MA: Kluwer Academic Publishers, 1997.
6. X. R. Cao, Realization Probabilities: the Dynamic of Queueing Systems, New York: Springer-Verlag, 1994.
7. C. G. Cassandras, Discrete Event Systems: Modeling and Performance Analysis, Aksen Associates, Inc., 1993.
8. P. Glasserman, Gradient Estimation Via Perturbation Analysis,
Norwell, MA: Kluwer Academic Publishers, 1991.
9. Y. C. Ho and X. R. Cao, Perturbation Analysis of Discrete-Event
Dynamic Systems, Norwell, MA: Kluwer Academic Publishers,
1991.
10. P. G. Harrison, On normalizing constants in queueing networks,
Oper. Res. 33: 464468, 1985.
11. L. Kleinrock, Queueing Systems, vols. I, II, New York: Wiley,
1975.
12. J. Walrand, An Introduction to Queueing Networks, Englewood
Cliffs, NJ: Prentice Hall, 1988.
13. S. H. Lu and P. R. Kumar, Distributed scheduling based on due
dates and buffer priorities, IEEE Trans. Autom. Control, AC-36:
14061416, 1991.
14. J. L. Peterson, Petri Net Theory and the Modeling of Systems,
Englewood Cliffs, NJ: Prentice Hall, 1981.
15. C. Reutenauer, The Mathematics of Petri Nets, London: PrenticeHall, 1990.
16. M. C. Zhou and F. DiCesare, Petri Net Synthesis for Discrete Event
Control of Manufacturing Systems, Norwell, MA: Kluwer Academic Publisher, 1993.
17. G. Cohen et al., A linear system theoretical view of discrete event
processes and its use for performance evaluation in manufacturing, IEEE Trans. Autom. Control, AC-30: 210220, 1985.
18. F. Baccelli et al., Synchronization and Linearity, New York: Wiley, 1992.
19. Y. C. Ho, A. Eyler, and T. T. Chien, A gradient technique for
general buffer storage design in a serial production line, Int. J.
Production Res., 17: 557580, 1979.
20. X. R. Cao, Convergence of parameter sensitivity estimates in a
stochastic experiment, IEEE Trans. Autom. Control, AC-30: 834
843, 1985.
611
21. X. R. Cao, First-order perturbation analysis of a single multiclass finite source queue, Performance Evaluation, 7: 3141,
1987.
22. P. Heidelberger et al., Convergence properties of infinitesimal
perturbation analysis estimates, Management Science, 34: 1281
1302, 1988.
23. W. B. Gong and Y. C. Ho, Smoothed perturbation analysis of discrete event dynamic systems, IEEE Trans. Autom. Control, 32:
858866, 1987.
24. Y. Park and E. K. P. Chong, Distributed inversion in timed discrete event systems, Discrete Event Dynamic Systems: Theory and
Applications, 5: 219241, 1995.
25. A. A. Gaivoronski, L. Y. Shi, and R. S. Sreenivas, Augmented
infinitesimal perturbation analysis: An alternate explanation,
Discrete Event Dynamic Systems: Theory and Applications, 2: 121
138, 1992.
26. Y. C. Ho and S. Li, Extensions of perturbation analysis of discrete
event dynamic systems, IEEE Trans. Autom. Control, 33: 427
438, 1988.
27. P. Vakili, Using a standard clock technique for efficient simulation, Oper. Res. Lett., 11: 445452, 1991.
28. L. Y. Dai and Y. C. Ho, Structural infinitesimal perturbation
analysis (SIPA) for derivative estimation of discrete event dynamic systems, IEEE Trans. Autom. Control, 40: 11541166,
1995.
29. P. Bremaud, Maximal coupling and rare perturbation sensitivity
analysis, Queueing Systems: Theory and Applications, 10: 249
270, 1992.
30. F. Baccelli, M. Klein, and S. Zuyev, Perturbation analysis of functionals of random measures, Adv. in Appl. Prob., 27: 306325,
1995.
31. F. J. Vazquez-Abad and K. Davis, Efficient implementation of the
phantom RPA method, with an application to a priority queueing
system, Adv. in Appl. Prob., submitted, 1995.
32. M. A. Zazanis and R. Suri, Perturbation analysis of the GI/G/1
queue, Queueing Systems: Theory and Applications, 18: 199
248, 1994.
33. B. Gang, C. Cassandras, and M. A. Zazanis, First and second
derivative estimators for closed Jackson-like queueing networks
using perturbation analysis techniques, Discrete Event Dynamic
Systems: Theory and Applications, 7: 2968, 1997.
34. X. R. Cao and H. F. Chen, Perturbation realization, potentials,
and sensitivity analysis of Markov processes, IEEE Trans. Autom.
Control, 42: 13821393, 1997.
35. W. B. Gong and S. Nananukul, Rational interpolation for rare
event probabilities, Stochastic Networks: Stability and Rare Event,
New York: Springer, 1996.
36. W. B. Gong and J. Q. Hu, On the MacLauring Series of GI/G/1
Queue, J. Appl. Prob., 29: 176184, 1991.
37. X. R. Cao, W. G. Gong, and Y. Wardi, Ill-conditioned performance
functions of queueing systems, IEEE Trans. Autom. Control, 40:
10741079, 1995.
38. P. W. Glynn, Likelihood ratio gradient estimation: An overview,
Proc. Winter Simulation Conf., 366375, 1987.
39. M. I. Reiman and A. Weiss, Sensitivity analysis via likelihood
ratio, Oper. Res., 37: 830, 844, 1989.
40. R. Y. Rubinstein, Sensitivity analysis and performance extrapolation for computer simulation models, Oper. Res., 37: 7281,
1989.
41. E. K. P. Chong and P. J. Ramadge, Optimization of queues using
infinitesimal perturbation analysis-based algorithms with general update times, SIAM J. Control Optim., 31: 698732, 1993.
42. Y. C. Ho and X. R. Cao, Perturbation analysis and optimization
of queueing networks, J. Optim. Theory Appl., 40: 559582, 1983.
612
43. C. G. Cassandras and S. G. Strickland, On-line sensitivity analysis of Markov chains, IEEE Trans. Autom. Control, 34: 7686,
1989.
44. R. Suri and Y. T. Leung, Single run optimization of discrete event
simulationsan empirical study using the M/M/1 queue, IIE
Trans., 21: 3549, 1989.
45. Q. Y. Tang and H. F. Chen, Convergence of perturbation analysis
based optimization algorithm with fixed number of customers period, Discrete Event Dynamic Systems: Theory and Applications,
4: 359375, 1994.
46. C. A. Brooks and P. Varaiya, Using perturbation analysis to solve
the capacity and flow assignment problem for general and ATM
networks, IEEE Globcom, 1994.
47. N. Xiao, F. F. Wu, and S. M. Lun, Dynamic bandwidth allocation
using infinitesimal perturbation analysis, IEEE Infocom, 383
389, 1994.
48. M. Caramanis and G. Liberopoulos, Perturbation analysis for the
design of flexible manufacturing system flow controllers, Oper.
Res., 40: 11071125, 1992.
49. A. Haurie, P. LEcuyer, and C. van Delft, Convergence of stochastic approximation coupled with perturbation analysis in a class
of manufacturing flow control models, Discrete Event Dynamic
Systems: Theory and Applications, 4: 87111, 1994.
50. H. M. Yan and X. Y. Zhou, Finding optimal number of Kanbans
in a manufacturing system via perturbation analysis, Lecture
Notes in Control and Information Sciences, 199, Springer-Verlag,
572578, 1994.
51. H. M. Yan, G. Yin, and S. X. C. Lou, Using stochastic optimization to determine threshold values for the control of unreliable
manufacturing systems, J. Optim. Theory Appl., 83: 511539,
1994.
52. N. Miyoshi and T. Hasegawa, On-line derivative estimation for
the multiclass single-server priority queue using perturbation
analysis, IEEE Trans. Autom. Control, 41: 300305, 1996.
53. Y. C. Ho, Soft Optimization for Hard Problems, Computerized
lecture via private communication, 1996.
XI-REN CAO
The Hong Kong University of
Science and Technology
480
on an ongoing basis from the observation process. The filtering problem is to estimate the signal process at the same
time. The smoothing problem has a couple of variants: (1)
Given the observation process in a fixed time interval, estimate the signal process at each element in the time interval,
and (2) estimate the signal process at a time that is a fixed lag
behind the observation process. The approach of Kolmogorov,
Krein, and Wiener to these problems assumed that the stochastic processes are (wide-sense) stationary and that the infinite past of the observation process is available. Both of
these assumptions are not physically reasonable, so there was
a need to relax these assumptions.
In the late 1950s, control and system theory were undergoing a significant change from the frequency-domain approach
to the state-space approach. Transfer function descriptions of
linear systems were replaced by ordinary differential equation descriptions of linear systems. This state-space approach
provided an impetus to reexamine the linear filtering problem. Using this approach the signal process is modeled as the
solution of a linear differential equation with a Gaussian
white noise input so the signal process is GaussMarkov. The
differential of the observations process is a linear transformation of the signal process plus Gaussian white noise. This filtering model does not require the infinite past of the observations. The signal and the observation processes can evolve
from some fixed time with a Gaussian random variable as the
initial condition for the differential equation that describes
the signal process. The processes are not required to be stationary; and, in fact, the coefficients of the differential equation for the signal process and the linear transformation for
the signal in the observation equation can be time-varying.
While it is not required that the ordinary differential equation for the signal process be stable, which is implicit in the
description for stationary processes, it is necessary to be able
to model the signal process as the solution of an ordinary differential equation with a white noise input. In general a stationary Gaussian process may not have such a model.
With the success of these linear filtering results that were
developed particularly by Kalman (9) for discrete-time processes and by Kalman-Bucy (10) for continuous-time processes, an interest developed in trying to solve a filtering
problem where the signal is a solution to a nonlinear differential equation with a white noise input. It is natural to call
such a problem a nonlinear filtering problem. The precise description of such a problem required the introduction of a significant amount of the modern theory of stochastic processes.
The major technique for describing the signal process is the
theory of stochastic differential equations that was initiated
by K. Ito (11).
The Gaussian white noise processes that appear as inputs
in the nonlinear differential equations require more sophisticated mathematical methods than do the inputs to linear differential equations. This occurs because the linear transformations of white noise have one natural interpretation but
the nonlinear transformations of white noise have no single
natural interpretation.
Interestingly, it was Wiener (12) who first constructed the
basic sample path property of the integral of Gaussian white
noise that is called the Wiener process or Brownian motion
and which provided the basis for the interpretations of nonlinear transformations of white noise. Many important properties of Brownian motion were determined by P. Levy (13). The
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
solution of a stochastic differential equation (a nonlinear differential equation with a white noise input) required the theory of stochastic integrals (14), which depends on some martingale theory (15) associated with Brownian motion.
No one definition of stochastic integrals arises naturally
from the Riemann sum approximations to the stochastic integral. This phenomenon occurs because Brownian motion does
not have bounded variation. The definition of K. Ito (14) is
the most satisfying probabilistically because it preserves the
martingale property of Brownian motion and Wiener integrals (stochastic integrals with deterministic integrands).
However, the calculus associated with the Ito definition of stochastic integral is somewhat unusual. The FiskStratonovich
definition of stochastic integral (16,17) preserves the usual
calculus properties, but the family of integrable functions is
significantly smaller. An uncountable family of distinct definitions of stochastic integrals can be easily exhibited (18).
This choice or ambiguity in the definition of a stochastic integral has played an important role in nonlinear filtering because initially some nonlinear filtering solutions were given
without specifying the interpretation or the definition of the
stochastic integrals. This ambiguity often arose by a formal
passage to the limit from discrete time.
In general, to compute conditional statistics of the state
process given the observation process, it is necessary to compute the conditional density of the state process given the observation process. For linear filtering the signal and the observation processes are Gaussian, so the conditional density
is determined by the conditional mean and the conditional
covariance. The conditional covariance is not random, so it
does not depend on the observation process. The conditional
mean can be shown to satisfy a stochastic differential equation that models the signal process and that has the observations as the input. These two conditional statistics (i.e., function of the observations are called sufficient conditional
statistics (19) because the conditional density can be recovered from them. For nonlinear filtering the solution does not
simplify so easily. In general there is no finite family of sufficient conditional statistics for a nonlinear filtering problem.
The conditional density can be shown to satisfy a nonlinear
stochastic partial differential equation (20,21). This equation
is especially difficult to solve because it is a stochastic partial
differential equaton and it is nonlinear. Even approximations
are difficult to obtain. The conditional density can be expressed using Bayes formula (22,23), so that it has the same
form as the Bayes formula in elementary probability though
it requires function space integrals. The numerator in the
Bayes formula expression for the conditional density is called
the unnormalized conditional density. This unnormalized conditional density satisfies a stochastic partial differential equation that is linear. It is called the DuncanMortensenZakai
(DMZ) equation of nonlinear filtering (2426).
In nonlinear filtering, the question of finite dimensional
filters describes the problem of finding finite dimensional solutions to the DMZ equation or to finite families of conditional
statistics. A basic approach to this question on the existence
or the nonexistence of finite-dimensional filters is the estimation algebra (27,28,28a), which is a Lie algebra of differential
operators that is generated by the differential operators in
the DMZ equation. Some families of nonlinear filtering problems have been given that exhibit finite-dimensional filters
(e.g., see Ref. 29).
481
482
spectively. The following assumptions are made on the coefficients of the stochastic differential equations (1) and (2) that
describe the signal or state process and the observation process, respectively.
The stochastic processes are defined on a fixed probability
space (, F , P) with a filtration (F t, t 0). Often the space
can be realized as a family of continuous functions. The
-algebras are assumed to be complete with respect to the
probability measure P.
A transition probability measure or a transition probability function for a Markov process is a function P(s, x; t, ) for
s [0, t), x d, and Bd the Borel -algebra on d that
satisfies the following:
(1)
ci j (t, x)
xi
2 ci j (t, x)
xi x j
i, j {1, . . ., n}
(3)
(4)
(2)
With the notion of transition probability measure (function), a Markov process can be defined.
Definition. Let P(s, x; t, ) be a transition probability measure and be a probability measure on (d, Bd). A probability
P(X (0) ) = ( ),
L=
(5)
(6)
f : lim
t0
d
i=1
ai
Pt f f
exists
t
2
1
+
ci j
xi
2
xi x j
(9)
where c bTb.
An analogous result holds if the Markov process is not homogeneous, so that
dX (t) = a(t, X (t)) dt + b(t, X (t)) dB(t)
where the theory of two-parameter semigroups is used so that
Ps,t f (x) = EX (s)=x [ f (X (t))]
and
dPs,t f
= Lf
dt
ai
1
2
+
ci j
xi
2
xi x j
(10)
1 2
(ai ) +
(c )
xi
2
xi x j i j
(11)
(12)
(13)
or
(s) dM(s)
(14)
(8)
L=
for
f D(L) =
where
483
This integral is defined from the limit of finite sums that are
formed from partitions as in RiemannStieltjes integration
where the function is evaluated at the midpoint of each interval formed from the partition. Recall that the Ito integral
is formed by evaluating the integrand at the left endpoint of
each of the subintervals formed from the partition (40).
For the linear filtering problem of GaussMarkov processes it is elementary to show that the conditional probability density is Gaussian so that only the conditional mean and
the conditional covariance have to be determined. Furthermore, the estimate of the state, given the observations that
minimizes the variance, is the conditional mean. Thus one
approach to the nonlinear filtering problem is to obtain a stochastic equation for the conditional mean or for other conditional statistics. The difficulty with this approach is that typically no finite family of equations for the conditional statistics
is closed; that is, any finite family of equations depends on
other conditional statistics.
The conditional probability density of the state X(t), given
the observations (Y(u), 0 u t), is the density for the conditional probability that represents all of the probabilistic information about X(t) from the observations (Y(u), 0 u t). A
conditional statistic can be computed by integrating the conditional density with a suitable function of the state.
To obtain a useful expression for the conditional probability measure, it is necessary to use a result for the absolute
484
(15)
B(t)
= B(t)
where
M(t) = exp
(s), dB(s)
0
1
2
|(s)|2 ds
Y (t) = B(t)
(s) ds
(17)
(18)
(t) = exp
1
c1 (s, X (s))a(s, X (s)), a(s, X (s))
ds
1
2
(19)
(t) = exp
0
t
P(, t | x0 , Yu , 0 u t) =
E Z [1 t t ]
E Z [t t ]
(21)
0
t
f 1 (s, Y (s))g(s, X (s),Y (s)), g(s, X (s), Y (s))
ds
(20)
To indicate the expectation with respect to one of the function
space measures, E is subscripted by the measurefor example, EX.
where
p(t) = p(X (t), t | x0 , Y (u), 0 u t)
(23)
(24)
=
g(t)
E X [ (t)g(t)]
E X [ (t)]
(25)
485
Theorem. Let (X(t), Y(t), t 0) satisfy Eqs. (1) and (2). Assume that a(t, x) and b(t, x) in Eq. (1) are continuous in t and
globally Lipschitz continuous in x, h(t, x, y) is continuous in t
and globally Lipschitz continuous in x and y, and g(t, y) is
continuous in t and globally Lipschitz continuous in y and
f gTg is strictly positive definite uniformly in (t, y). If
C2(n, ) such that
q(t) = E X [ (t)]
E| (X (t))| dt <
(26)
(27)
(28)
(32)
so that
p(x, t | x0 , Y (u), 0 u t) = r(t)q1 (t)
E
0
where
then the conditional expectation of (X(t)), given the observations (Y(u), u t),
To apply algebro-geometric methods to the nonlinear filtering problem the following form of the DMZ equation is
used
(X (t)) =
E X [ (X (t)) (t)]
E X [ (t)]
(30)
The stochastic equation for the condition probability density is a nonlinear stochastic partial differential equation. The
stochastic equation for a conditional statistic is typically coupled to an infinite famiily of such equations. The conditional
density is more useful because it represents all of the probabilistic information about the state given the observations, but
it is a nonlinear equation. If the so-called unnormalized conditional density is used, then the stochastic partial differential
equation is linear. This unnormalized conditional density was
given by Duncan (24), Mortensen (25), and Zakai (26). The
equation is usually called the DuncanMortensenZakai
(DMZ) equation.
Theorem. Let (X(t), Y(t), t 0) be the processes that are the
solutions to Eqs. (1) and (2). Let r be given by
r(x, t | x0 , Y (u), 0 u t) = E X [ (t) | X (t) = x]pX (0, x0 ; t, x)
(31)
(35)
(36)
= A p(t)
dt
(37)
where A is the intensity matrix or the transpose of the generator of the Markov process (X(t), t 0). The dependence of p
on the initial value X(0) has been suppressed for notational
convenience. By analogy with the DMZ equations (31) and
(36) in this case for the finite-state Markov process (49) it
follows that the unnormalized conditional density (t) satisfies
(t) = (0) +
A(s) ds +
0
B(s) dY (s)
0
(38)
486
or
(t) = (0) +
(A (1/2)B2 )(s) ds +
0
B(s) dY (s)
0
(39)
n
i
i=1 (si ) (t)
n
i
i=1 (t)
(40)
(41)
where x(t) M, a smooth d-dimensional manifold. A controllability property has local significance in analogy to its global
significance for linear control systems.
Definition. The controllability Lie algebra of Eq. (41) is the
Lie algebra L generated by f, g1, . . ., gm.L (x) is the linear
space of vectors in TxM, the tangent space of M at x, spanned
by the vector fields of L at x. The dimension of L (x) has implication for the local reachable set starting at x M.
Another basic notion in system theory is observability.
This condition implies that different statesthat is, different points in Mcan be distinguished using an appropriate
control.
Definition. Consider the control system (41). Let h
C(M, ) give the observation as
y(t) = h(x(t))
(42)
(43)
t () = ((t))
(44)
where (t) n.
The application of the Lie algebraic methods described
above and the use of nonlinear system theory presupposed a
finite-dimensional manifold. For the DMZ equation (36) the
solution evolves in an infinite-dimensional manifold. Thus it
is necessary to be precise when translating these finite-dimensional algebro-geometric results to the DMZ equation. If
this approach can be applied to the DMZ equation, then the
questions of the existence or the nonexistence of finite-dimensional stochastic equations for conditional statistics and the
equivalence of two nonlinear filtering problems can be resolved. Even for finite-state Markov processes it can be determined if some conditional statistics are the solutions of stochastic equations whose dimension is significantly smaller
than the number of states of the Markov process (52).
For the DMZ equation (36) by analogy with the finite-dimensional inputoutput systems (41) and (42), the Lie algebra generated by the operators L* (1/2)h, h and h,
acting on smooth (C) functions is called the estimation algebra associated with Eqs. (41) and (42) (53,54,54a).
To identify equivalent filtering problems it is important to
investigate transformations that induce isomorphic estimation algebras. A simple, important transformation is a change
of scale of the unnormalized conditional probability density
r( ). Let :n be a strictly positive, smooth function and
let r(t) r(t). This transformation acts on the generators of
the estimation algebra as
L 1
1
h, h
2
g1 ,
Thus the estimation algebras are formally isomorphic. Furthermore, a smooth homeomorphism of the state space induces an estimation algebra that is formally isomorphic to the
initial estimation algebra. The above two operations on the
estimation algebra have been called the estimation formal
equivalence group (55).
If for some distribution of X(0) a conditional statistic,
t(), can be described by a minimal finite-dimensional (recursive) filter of the form (43) and (44), then the Lie algebra
of this system should be a homomorphism of the estimation
algebra for this filtering problem. This property has been
called the homomorphism principle for the filtering problem
(56). This homomorphism principle can be a guide in the investigation of the existence of finite-dimensional filters.
A specific example of this homomorphism property occurs
when the estimation algebra is one of the Weyl algebras. The
Weyl algebra Wn is the algebra of polynomial differential operators over with operators x1, . . ., xn, /x1, . . ., /xn.
The Lie bracket is the usual commutator for differential operators. This Lie algebra has a one-dimensional center and the
quotient Wn / is simple; that is, it contains no nontrivial ideals. For the estimation algebra these two properties imply
that if Wn is the estimation algebra for a filtering problem,
then either the unnormalized conditional density can be computed by a finite-dimensional filter or no conditional statistic
can be computed by a finite-dimensional filter of the form (43)
and (44). More specifically, for n 0 there are no nonconstant
homomorphisms from Wn or Wn / to the Lie algebra of smooth
vector fields on a smooth manifold (57).
As an example of a Weyl algebra occurring as an estimation algebra, consider
dX (t) = dB(t)
(45)
(46)
(47)
(48)
1 2
x2 , x, , 1
2 x2
2
x
This algebra is called the oscillator algebra in physics (27,28).
The oscillator algebra is the semidirect product of 1 and
the Heisenberg algebra that is generated by
1
1 2
x2 , x, and
2 x2
2
x
487
(49)
(50)
1 2
f x2 , x
2
2 x
x
2
and is four-dimensional and solvable. The estimation algebra
L for (47)(48) arises for the algebra L by letting f 0. To
x
associate L with the estimation algebra L let F(x) f,
(x) exp(F(x)) and r(t, x) (x)r(t, x). Then the DMZ equation for (47) and (48) is transformed by the gauge transformation as
dr =
1 2
1
2
[(a + 1)x + bx + c] r + xr dY
2 x2
2
(51)
This has the same form as the DMZ equation for (49) and
(50). Thus the nonlinear filtering problem (49) and (50) is obtained from the linear filtering problem (47) and (48) by a
gauge transformation of the conditional density. Various
other examples of finite-dimensional filters are available (e.g.,
see Refs. 5961).
Ocone (62) showed that for a scalar filtering problem with
the observation equation of the form (50) the two examples
(47)(48) and (49)(50) are the only ones that give a finitedimensional estimation algebra. This result is given in the
following theorem.
Theorem. Let n m 1 in (49)(50) and let g 1 in the
observation equation (2). Then the dimension of the estimation algebra is finite only if:
1. h(x) ax and
df
+ f 2 = ax2 + bx + c
dx
or
2. h(x) ax2 x, a 0 and
df
+ f 2 (x) = h2 (x) + a(2ax + )2 + b + c(2ax + )1
dx
488
or
and
df
+ f 2 (x) = h2 (x) + ax2 + bx + c
dx
dR(t)
= B(t,Y )B(t, Y ) + A(t,Y )R(t)
+ R(t)A(t,Y
)
d
[G j (t, Y )X (t) + g j (t, Y )] dY j (t)
G j (t, Y )R(t)G
(t, Y )
(56)
j
[G j (t, Y )R(t)
+ R(t)G
(t, Y ) ] dY j (t)
j
(52)
(0) R0.
where m
(0) m0 and R
(53)
j=1
where X(0) is a Gaussian random variable, Y(0) 0. The random variable X(t), given Y (t) (Y(u), u t), is conditionally
Gaussian. More precisely, it is assumed that (X(t), t 0) is
an n-valued (F t)-adapted process, (Y(t), t 0) is an m-val (t), t 0) are
ued (F t)-adapted process, (B(t), t 0) and (B
n
m
independent standard - and -valued Brownian motions,
respectively, in the filtered probability space (, F , (F t), P),
and X(0) is N(m0, R0). The functions A, a, B, Gj, gj, H, and h
are defined on C(, m) with values in a suitable Euclidean space, and they are progressively measurable. The
1
functions A2, a, B2, Gj2, H, and h are in Lloc
() for each
m
1
y C(, ). For each T 0, E (T) 1 where
T
(T ) = exp
0
1
2
(54)
dm(t)
= A(t,Y )m(t)
+ a(t,Y )
R(t)H(t,
Y ) [H(t, Y )m(t)
+ h(t, Y )]
j
+
G j (t, Y )R(t)H
(t,Y ) dt
(57)
where X(0), X(t) H, a separable, infinite-dimensional Hilbert space, and (W(t), t 0) is a standard cylindrical Wiener
process. A standard cylindrical Wiener process means that if
1, 2 H H*, 1, 2 0, and 1, 1 2, 2 1 where
, is the inner product in H, then (1, W(t), t 0) and
(2, W(t), t 0) are independent standard Wiener processes.
If A is the generator of an analytic semigroup (S(t), t 0)
and S(r)Q1/2 is HilbertSchmidt for each r 0 and
t
0
|S(r)Q1/2 |2L
2 (H )
dr <
(58)
(55)
Z(t) =
(59)
j
+
[G j (t,Y )m(t)
+ g j (t,Y ) + R(t)H
(t, Y ) dY j (t)]
j
(61)
(62)
489
490
For an example of this method of logarithmic transformation consider the following linear parabolic partial differential
equation:
1
tr a(x)pxx + g(x, t), px
+ V (x, t)p
2
p(x, 0) = p0 (x)
pt =
(63)
1
tr a(x)Sxx + H(x, t, Sx )
2
(64)
(65)
1
a(x), Sx , Sx
V (x, t)
2
(66)
J(x, t, u) = Ex
(67)
(68)
where
L(x, t, u) =
1 1
a (x)u, u
V (x, t)
2
(69)
With suitable assumptions on the family of admissible controls and conditions on the terms in the model it can be shown
from the Verification Theorem (87) that Eq. (63) is the dynamic programming equation for this stochastic control problem. This approach can provide a rigorous basis for the formal
maximization of a likelihood function in function space. See
Ref. 87a.
An approach to the robustness of the nonlinear filter
(87b,87c) is to obtain a so-called pathwise solution to the DuncanMortensenZakai (DMZ) equation by expressing the solution as an (observation) path dependent semigroup. The infinitesimal generator of this semigroup is the conjugation of
the generator of the signal process by the observation path
multiplied by the drift in the observation where the Stratonovich form of the DMZ equation is used. The fact that the
observation path appears explicitly rather than its differential implies the robustness of the solution of the DMZ
equation.
It is important to obtain estimation methods that are applicable to both stochastic disturbances (noise) and deterministic disturbances. For Brownian motion, a Hilbert (or Sobolev
space) of functions that are functions that are absolutely continuous and whose derivatives are square integrable having
1
(t, x) (y)
!
(70)
491
14. K. Ito, Stochastic integral, Proc. Imp. Acad. Tokyo, 20: 519524,
1944.
15. P. A. Meyer, Probability and Potentials, Waltham, MA: Blaisdell, 1966.
16. D. L. Fisk, Quasi-martingales and stochastic integrals, Technical
Report 1, Dept. of Mathematics, Michigan State University, 1963.
17. R. L. Stratonovich, Conditional Markov process theory, Theory
Probability Appl. (USSR), 5: 156178, 1960.
18. E. J. McShane, Stochastic differential equations and models of
random processes, In Proceedings of the 6th Symposium on Mathematical, Statistics, and Probability, Berkeley, CA: University of
California Press, 1973.
19. S. S. Wilks, Mathematical Statistics, New York: Wiley, 1962.
20. H. Kushner, On the differential equations satisfied by conditional
densities of Markov processes, with applications, SIAM J. Control, 2: 106119, 1964.
21. R. L. Stratonovich, Application of the theory of Markov processes
for optimum filtration of signals, Radio Eng. Electron. Phys.
(URRS), 1: 119, 1960.
22. R. S. Bucy, Nonlinear filtering theory, IEEE Trans. Autom. Control, AC-10: 198, 1965.
23. G. Kallianpur and C. Striebel, Estimation of stochastic systems:
Arbitrary system process with additive white observation errors,
Ann. Math. Statist., 39: 785801, 1969.
24. T. E. Duncan, Probability Densities for Diffusion Processes with
Applications to Nonlinear Filtering Theory and Detection Theory,
Ph.D. Dissertation, Stanford University, Stanford, CA, June 1967.
25. R. E. Mortensen, Optimal Control of Continuous-Time Stochastic
Systems, Ph.D. dissertation, University of California, Berkeley,
CA, 1966.
26. M. Zakai, On the optimal filtering of diffusion processes, Z.
Wahrsch, Verw. Geb. 11: 230243, 1969.
27. R. Brockett, Remarks on finite dimensional nonlinear estimation,
Asterisque, 764: 4755, 1980.
28. S. K. Mitter, On the analogy between mathematical problems on
nonlinear filtering and quantum physics, Ric. Autom., 10 (2):
163216, 1980.
28a. S. K. Mitter, Filtering theory and quantum fields, Asterisque,
7576: 199205, 1980.
29. V. E. Benes, Exact finite dimensional filters for certain diffusions
with nonlinear drift, Stochastics, 5: 6592, 1981.
30. R. H. Cameron and W. T. Martin, The orthogonal development of
non-linear functionals in a series of FourierHermite functions,
Ann. Math., 48: 385392, 1947.
31. K. Ito, Multiple Wiener integral, J. Math. Soc. Japan, 3: 157
169, 1951.
32. Y. Hu and P. A. Meyer, Sur les integrales multiples de Stratonovich, Lecture Notes in Mathematics, No 1321, Berlin: SpringerVerlag, pp. 7281, 1988.
492
39. A. Pazy, Semigroups of Linear Operators and Applications to Partial Differential Equations, New York: Springer-Verlag, 1983.
40. I. Karatzas and S. E. Shreve, Brownian Motion and Stochastic
Calculus, 2nd ed., New York: Springer-Verlag, 1991.
41. I. E. Segal, Tensor algebras over Hilbert spaces. I, Trans. Am.
Math. Soc., 81: 106134, 1956.
42. A. V. Skorokhod, On the differentiability of measures which correspond to stochastic processes. I: Processes with independent increments, Theory Probability Appl., 2: 407432, 1957.
43. A. V. Skorokhod, Studies in the Theory of Random Processes, Reading, MA: Addison-Wesley, 1965.
44. I. V. Girsanov, On transforming a certain class of stochastic processes by absolutely continuous substitution of measures, Theory
Probability Appl., 5: 285301, 1960.
45. T. E. Duncan and P. Varaiya, On the solutions of a stochastic
control system. II, SIAM J. Control, 13: 10771092, 1975.
46. A. N. Shiryayev, On stochastic equations in the theory of conditional Markov processes, Theor. Probability Appl., 11: 179184,
1966.
47. M. Fujisaki, G. Kallianpur, and H. Kunita, Stochastic differential
equations for non-linear filtering, Osaka J. Math., 9: 1940, 1972.
48. G. Kallianpur, Stochastic Filtering Theory, New York: SpringerVerlag, 1980.
49. W. M. Wonham, Some applications of stochastic differential equations to optimal nonlinear filtering, SIAM J. Control, 2: 347
369, 1965.
50. H. Sussmann, Existence and uniqueness of minimal realizations
of nonlinear systems, Math. Syst. Theory, 10: 263284, 1977.
51. R. Brockett and J. M. C. Clark, The geometry of the conditional
density equation, In O. Jacobs et al. (eds.), Analysis and Optimization of Stochastic Systems, New York: Academic Press, 1980,
pp. 299310.
52. S. I. Marcus, Low dimensional filters for a class of finite state
estimation problems with Poisson observations, Syst. Control
Lett., 1: 237241, 1982.
53. M. Davis and S. Marcus, An introduction to nonlinear filtering,
In M. Hazewinkel and J. Willems (eds.), Stochastic Systems: The
Mathematics of Filtering and Identification and Applications, Dordrecht, The Netherlands: D. Riedel, 1981, pp. 5375.
54. S. I. Marcus, Algebraic and geometric methods in nonlinear filtering, SIAM J. Control, 22: 817844, 1984.
54a. S. K. Mitter, Geometric theory of nonlinear filtering, Outils et
Mode`les Mathematiques pour lAutomatiques, 3: CNRS, 3760,
1983.
55. R. W. Brockett, Classification and equivalence in estimation theory, In Proceedings of the 18th IEEE Conference on Decision and
Control, Ft. Lauderdale, FL, 1979, pp. 172175.
56. R. Brockett, Nonlinear systems and nonlinear estimation theory,
In M. Hazewinkel and J. Willems (eds.), Stochastic Systems: The
Mathematics of Filtering and Identification and Applications, Dordrecht, The Netherlands: D. Riedel, pp. 442477, 1981.
57. M. Hazewinkel and S. Marcus, Some results and speculations on
the role of Lie algebras in filtering, In M. Hazewinkel and J. Willems (eds.), Stochastic Systems: The Mathematics of Filtering and
Identification and Applications, Dordrecht, The Netherlands: D.
Riedel, pp. 591604, 1981.
58. M. Hazewinkel, S. Marcus, and H. Sussmann, Nonexistence of
exact finite dimensional filters for conditional statistics of the cubic sensor problem, Systems Control Lett., 3: 331340, 1983.
59. D. Ocone, J. S. Baras, and S. I. Marcus, Explicit filters for diffusions with certain nonlinear drifts, Stochastics, 8: 116, 1982.
60. W. Wong, Theorems on the structure of finite dimensional estimation algebras, Systems Control Lett., 9: 117124, 1987.
FILTERING THEORY
84. A. E. Bryson and M. Frazier, Smoothing for linear and nonlinear
dynamic systems, in Proceedings of the Optimum System Synthesis
Conference, Wright-Patterson Air Force Base, Ohio, September,
1962, AST-TDR-63-119.
85. R. E. Mortensen, Maximum likelihood recursive nonlinear filtering, J. Optim. Theory Appl., 2: 386394), 1968.
86. W. H. Fleming and S. K. Mitter, Optimal control and nonlinear
filtering for nondegenerate diffusion processes, Stochastics, 8:
6377, 1982.
87. W. H. Fleming and R. W. Rishel, Deterministic and Stochastic
Optimal Control, New York: Springer-Verlag, 1975.
87a. S. K. Mitter, Lectures on nonlinear filtering and stochastic control, Lecture Notes in Math, 972: 170207, 1983.
87b. M. H. A. Davis, On a multiplicative functional transformation
arising in nonlinear filtering, Z. Wahr. Verw. Geb., 54: 125139,
1980.
87c. M. H. A. Davis, A pathwise solution of the equations of nonlinear filtering, Theory Prob. Appl., 27: 167175, 1982.
87d. O. Hijab, Minimum energy estimation, Ph.D. dissertation, Univ.
California, Berkeley, CA, 1980.
87e. W. H. Fleming and W. M. McEneaney, Risk sensitive and robust
nonlinear filtering, Proc. 36th IEEE Conf. Decision Cont., San
Diego, 10881093, 1997.
87f. B. Z. Bobrovsky and M. Zakai, Asymptotic bounds on the minimal
error of nonlinear filtering, Stochastic Systems: the mathematics of
filtering and identification and applications, M. Hazeainkel and J.
Willems (eds.), Dordrecht: Reidel, 1981, 573582.
87g. D. Ocone and E. Pardoux, Asymptotic stability of the optimal
filter with respect to the initial conditions, SIAM J. Con. Optim.,
34: 226243, 1996.
88. A. Bensoussan, R. Glowinski, and R. Rascanu, Approximations of
the Zakai equation by splitting up method, SIAM J. Control Optim., 28: 14201431, 1990.
89. R. J. Elliott and R. Glowinski, Approximations to solutions of the
Zakai filtering equation, Stochastic Anal. Appl., 7: 145168, 1989.
90. K. Ito, Approximation of the Zakai equation for nonlinear filtering, SIAM J. Control Optim., 34: 620634, 1996.
91. H. P. McKean, Geometry of differential space, Ann. Prob., 1: 197
206, 1973.
92. S. Lototsky, R. Mikulevicius, and B. L. Rozovskii, Nonlinear filtering revisited: A spectral approach, SIAM J. Control Optim.,
35: 435461, 1997.
93. A. Budhiraja and G. Kallianpur, Approximation to the solution
of the Zakai equation using multiple Wiener and Stratonovich
integral expansions, Stochastics and Stoch. Repts., 56: 271315,
1996.
94. W.-L. Chiou and S.-T. Yau, Finite-dimensional filters with nonlinear drift. II: Brocketts problem on classification of finite-dimensional estimation algebras, SIAM J. Control Optim., 32: 297
310, 1994.
TYRONE E. DUNCAN
University of Kansas
493
GAIN SCHEDULING
MOTIVATION
A classical tradeoff in control design is model accuracy versus
model simplicity. While sophisticated models better represent
a physical systems behavior, the resulting analysis and control design are more involved.
One manifestation of this tradeoff is the use of nonlinear
versus linear models. Most physical systems exhibit nonlinear
behavior. Some common examples are saturations, rate limiters, hysteresis, and backlash. Predominantly nonlinear behavior may be found in robotic manipulator dynamics, aircraft or missile ight dynamics, undersea vehicle dynamics,
jet engine combustion dynamics, and satellite attitude dynamics. Although analysis and control design for nonlinear
systems remains an active topic of research, analysis of linear
systems is signicantly less complicated, and there is an
abundance of control design methodologies, ranging from
classical control to multivariable robust control.
One compromise is to linearize the system behavior, that
is, approximate the behavior of a nonlinear system near a
particular operating condition by a linear system. This simplication allows one to draw upon analysis and design methods
for linear systems. However, this simplication comes at the
cost of certain limitations:
The nonlinear system must be conned to operating near
the specied operating condition of the linearization.
The linearization analysis may give misleading or inconclusive results.
The linearization may ignore important nonlinear phenomena which dominate the system behavior.
Despite these limitations, linearization remains a widely used
method for control system analysis and design.
In many cases, conning a nonlinear system to operating
near a specied operating condition is too restrictive. One example is ight control. An aircraft typically experiences several ight conditions, including take-off, cruising at various
altitudes, specialized maneuvers, and landing. No single
linearization can adequately describe the aircraft dynamics at
all of these conditions. Another example is boiler-turbine control in power generation. Typical operating conditions include
power ramp up, steady power delivery at various levels, and
power ramp down. Again, no linearization can adequately describe the dynamics at all operating conditions.
Scheduling
variables
Environmental
conditions
Reference
commands
Gain scheduling is an approach to overcome the local limitations associated with linearizations. The idea is simple and
intuitively appealing. Given a nonlinear plant with a wide
range of operating conditions, one can select several representative operating conditions within the operating regime, perform several linearizations of the nonlinear dynamics, and de-
Operating
conditions
Scheduled
controller
Nonlinear
plant
Regulated
outputs
Measured
outputs
205
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
206
GAIN SCHEDULING
xx o
r(x)
=0
|(x x0 )|
(5)
(4)
(6)
x1 |x1 | + x2
x21 x2
(7)
Then
2|x1 |
D f (x) =
2x1 x2
1
x21
(8)
x0 =
1
2
leads to
LINEARIZATION
Linearization of Functions
We begin by recalling some concepts from multivariable calculus. Let f: R n R p denote a multivariable function which
maps vectors in R n to vectors in R p. In terms of the individual components,
f 1 (x1 , . . ., xn )
..
f (x) =
.
f p (x1 , . . ., xn )
(1)
x1
.
.
D f (x) =
.
fp
x1
(x)
..
(x)
f1
(x)
xn
..
.
fp
(x)
xn
(2)
f 1 (x1 , . . ., xn , u1 , . . ., um )
..
f (x, u) =
.
f p (x1 , . . ., xn , u1 , . . ., um )
(3)
(10)
In case f is differentiable, D1f(x, u) denotes the p n Jacobian matrix of partial derivatives with respect to the rst
variable,
For f continuously differentiable, i.e., if Df(x) has continuous elements, we may approximate f(x) by the truncated Taylors series
f (x) f (x0 ) + D f (x0 )(x x0 )
(9)
(x, u)
x1
..
D1 f (x, u) =
.
fp
(x, u)
x1
..
f1
(x, u)
xn
..
fp
(x, u)
xn
(11)
GAIN SCHEDULING
1
(x, u)
u1
..
D2 f (x, u) =
.
fp
(x, u)
u1
..
f1
(x, u)
um
..
fp
(x, u)
um
(12)
(14)
r(x, u)
lim
=0
|x x0 |2 + |u u0 |2
xx 0
uu 0
(15)
(16)
Otherwise, x0 is unstable.
x0 is asymptotically stable if in addition to being stable,
(25)
(17)
The reasoning for this terminology is that the initial condition x(0) x0 leads to the solution x(t) x0 for all time. So if
the solution starts at x0, it remains at x0, hence the term equilibrium.
In case f is continuously differentiable, we may rewrite
(16) as
(18)
(23)
A = D f (x0 )
In the last section, we saw how to approximate the static behavior of a nonlinear function by using a truncated Taylors
series. We now show how similar tools can be used to approximate the dynamic behavior of a nonlinear system.
Consider an autonomous nonlinear system
(21)
Substituting these formulas into (18) and neglecting the residual term, r(x), leads to the approximate dynamics
.
x = Ax
(22)
f (x0 ) = 0
f (x0 ) = 0
x = f (x)
and
where
r(x, u) = f (x, u) f (x0 , u0 ) + D1 f (x0 , u0 )(x x0 )
+ D2 (x0 , u0 )(u u0 )
207
(20)
(26)
x2
d x1
(27)
= g
sin(x1 ) cx2
dt x2
where x1 , and x2 .
208
GAIN SCHEDULING
x0 =
0
leads to
x =
0
g/
1
x
c
(29)
with the constant input u(t) u0 leads to the constant solution x(t) x0.
Proceeding as in the autonomous case, whenever f is continuously differentiable, we may rewrite (33) as
x = f (x0 , u0 ) + D1 f (x0 , u0 )(x x0 )
+ D2 f (x0 , u0 )(u u0 ) + r(x, u)]
(35)
(36)
Dropping this residual term and using that f(x0, u0) 0 leads
to the approximate linear dynamics
.
x = Ax + Bu
(37)
where
which is asymptotically stable. Therefore, the hanging equilibrium of the nonlinear pendulum is asymptotically stable,
as expected.
The above example demonstrates that different equilibrium points of the same nonlinear system can have different
stability conditions.
In some cases, using the linearization to assess stability
may be inconclusive.
A = D1 f (x0 , u0 ),
B = D2 f (x0 , u0 )
(38)
(30)
x = x3
(31)
x = f (x, G(x))
(39)
and
The following is a direct consequence of Theorem 4.
(note that x represents different quantities in the two linearizations). In this case, stability analysis of linearization is inconclusive; it does not indicate either the stability or instability of the equilibrium x0 0.
u = u0 K(x x0 )
(33)
(34)
u = K x
(40)
Theorem 9 states that we can construct stabilizing feedback for a nonlinear system by designing stabilizing feedback
for its linearization.
Example 10. Recall the simple pendulum example, but now
with a control torque input,
x2
d x1
(42)
= g
sin(x1 ) cx2 + u
dt x2
Linearizing about the upright equilibrium leads to
.
0
1
0
x =
x +
u
g/ c
1
(43)
GAIN SCHEDULING
The feedback
209
k2 )x
k2 )(x x0 )
k2 )x
(45)
.
x = Ax + Bx
(44)
where
A = D1 f (x0 , u0 ),
(46)
(47)
C = Dg(x0 )
(54)
Theorem 13. Let f in Eq. (33) and g in Eq. (49) be continuously differentiable, and let Eq. (53) be the linearization of
Eq. (33) about the equilibrium (x0, u0). Suppose the linear
feedback,
z = Az + By
(55)
u = Cz
B = D2 f (x0 , u0 ),
approximates the input-output behavior of Eq. (33) with measurement of Eq. (49). Here, x approximates x x0, y approximates y y0, and u exactly represents u u0.
(53)
y = Cx
z = Az + B(y g(x0 ))
(56)
u = u0 + Cz
(48)
Example 14. Suppose we wish to control the simple pendulum under the output feedback
which is stable.
Now suppose that the state is not available for feedback.
Rather, the control is restricted to measurements
y = g(x)
(49)
By using a similar analysis, we can construct stabilizing output feedback for the nonlinear system based on stabilizing
output feedback for the linearization.
Definition 12. The equilibrium (x0, u0) of Eq. (33) is stabilized by the dynamic output feedback
z = F (z, y)
u = G(z)
y = (1 0)x
(57)
.
x =
0
g/
1
0
def
x +
u = Ax + Bu
c
1
y = (1 0)x = Cx
The observer based controller
z = (A BK HC)z + H y
(50)
(58)
def
u = Kz
(59)
(51)
(52)
K = (k1
k2 ),
h1
H=
h2
(60)
210
GAIN SCHEDULING
(61)
By linearizing about the equilibrium (x0, u0) (0, 0), we obtain the linear control u x r. For r 0, this control law
stabilizes the equilibrium (0, 0). The resulting closed loop system is
x = x|x| x
(62)
For x(0) 1, the solution asymptotically approaches 0. However, for x(0) 1, the solution diverges to innity.
Step 1: Construction of Linearization Family. Gain scheduling
attempts to overcome local limitations by considering a family
of linearizations, rather than a single linearization.
Definition 16. The functions (xeq( ), ueq( )) dene an equilibrium family for the nonlinear system Eq. (33) over the set
S if
f (xeq (s), ueq (s)) = 0
(63)
for all s S.
x2
d x1
= g
sin(x1 ) cx2 + u
dt x2
An equilibrium family over the set S [, ] is
s
g
xeq (s) =
, ueq (s) = sin s
0
The associated linearization family is
0
1
0
x +
u
x = g
1
cos(s) c
(67)
(68)
(69)
Step 2: Fixed Operating Condition Designs. Let us select several operating conditions,
{s1 , s2 , . . ., sN } S
(70)
z = Ai z + Bi y + Li r
u = Ci z
(71)
(64)
(65)
where
(66)
(72)
This step constitutes the core of gain scheduling, and accordingly, accounts for the bulk of the effort in a gain scheduled control design. Designing xed operating point controllers is especially tedious in the case of several design
operating conditions.
Step 3: Scheduling. The remaining step is to piece together
a global controller from the individual local controllers. As
the scheduling variable varies in time, the control gains are
updated to reect the current operating condition of the plant.
The resulting overall controller is
r = r yeq (si )
(73)
The matrices A(s), B(s), C(s), and L(s) are functions of the
scheduling variable, as are the vectors yeq(s) and ueq(s). It is
important to note that the scheduling variable, s, which was
held constant during the design phase, is now time varying.
These matrix and vector functions are used to update the control parameters according to the variations in the scheduling
variable.
There are different options in how to schedule the controller parameters in Eq. (73).
GAIN SCHEDULING
(74)
A1 , . s R 1 ;
..
, . . ., ueq (s)
A(s) =
A ,
.
s
R
N
N
ueq (s1 ),
s R1 ;
..
=
(75)
.
R
u (s ),
N
eq N
Continuous Scheduling. Any interpolation algorithm is
used to construct continuous matrices which interpolate
the design conditions so that
A(si ) = Ai , . . ., ueq (si ) = ueq (si )
(76)
x = f (x, u)
(77)
(78)
(79)
(80)
r = r yeq (s)
In general, the scheduling variable, s, can be written as
u = x|x| + (r x)
This feedback leads to the closed loop dynamics
x = x + r
LPV Systems. It is convenient to consider slow variation restriction in the context of linear parameter varying (LPV) systems. LPV systems are dened to be linear systems whose
dynamics depend on exogenous time varying parameters
which are unknown a priori, but can be measured upon operation of the control system.
An LPV system can be represented in state space form as
y = C( )x
(81)
(82)
(84)
x = A( )x + B( )u
(83)
s = (x, r)
Example 18. Recall the system of Example 15. The equilibrium family
xeq (s) = s,
211
(85)
(86)
212
GAIN SCHEDULING
d
dt
s
0
= M(s)
+ B(u ueq (s))
v
v veq (s)
(92)
If veq(s) is differentiable,
d
veq (s) = Dveq (s)s
dt
= Dveq (s)M12 (s)(v veq (s)) + Dveq (s)B1 (u ueq (s))
(93)
where the matrices M12(s) and B1 are appropriate sub-matrices of M(s) and B.
Combining these equations leads to the alternate form of
Eq. (89),
d
dt
s
v veq (s)
0
M12 (s)
s
=
0 M22 (s) Dveq (s)M12 (s)
v veq (s)
B1
(u ueq (s))
+
B2 Dveq (s)B1
(94)
(95)
d
dt
s
s
= Anew (s)
+ Bnew (s)u
v
v
These equations can be viewed as a mass-spring-damper system with time-varying spring stiffness. For xed parameter
values, (t) o, the equilibrium xo 0 is asymptotically stable. However, for the parameter trajectory (t) cos(2t), it
becomes unstable. An intuitive explanation is that the stiffness variations are timed to pump energy into the oscillations.
where
These equations represent the extreme case where the nonlinearities are entirely captured in the scheduling variable, s.
Let (xeq(s), ueq(s)) be an equilibrium family, with
s
(90)
xeq (s) =
veq (s)
so that
s
0 = (s) + M(s)
+ Bueq (s)
veq (s)
v(t)
= v(t) veq (s(t)),
u(t)
(96)
(97)
x1 |x1 | + x2
x =
x21 x2 + u
(98)
(91)
s
xeq (s) =
,
s|s|
(99)
GAIN SCHEDULING
d
dt
s
0
1
s
=
x2 (s|s|)
0 s2 2|s|
x2 (s|s|)
0
+
(u (s3 |s|)
1
(100)
Yi = Ki P1
(101)
(102)
(103)
(104)
(1 0
(107)
(106)
Given Q 0 and Yi, one can solve for the original variables P
and Ki. With these variables, condition Eq. (105) is equivalent to
Ai Q BYi + QATi YiT BT < 0
213
...
0)xeq (s) = s
(108)
(109)
(110)
where
(105)
(111)
is
(112)
214
GAIN SCHEDULING
(113)
M. Vidyasagar, Nonlinear Systems Analysis, Nonlinear Systems Analysis, Englewood Cliffs, NJ: Prentice-Hall, Inc., 1993.
One can state appropriate conditions involving partial differential constraint equations under which there exists any G
which achieves this objective.
The intention is that improved closed loop performance is
possible if the closed loop linearizations have some sort of invariance property. Such improved performance, if any, is difcult to quantify, but simulation studies indicate the potential benets.
A. Bacciotti, Local Stabilizability of Nonlinear Control Systems, Singapore: World Scientic Publishing Co., 1992.
(114)
(115)
LPV System Analysis and Control
J. S. Shamma and M. Athans, Guaranteed Properties of Gain Scheduled Control of Linear Parameter Varying Plants, Automatica, 27:
559564, 1991.
S. M. Shahruz and S. Behtash, Design of controllers for linear parameter-varying systems by the gain scheduling technique, J. Math.
Anal. Appl., 168: 125217, 1992.
A. Packard, Gain scheduling via linear fractional transformations,
Syst. Control Lett., 22: 7992, 1994.
(117)
P. Apkarian and P. Gahinet, A convex characterization of gain-scheduled H controllers, IEEE Trans. Autom. Control, 40: 853864,
1995.
P. Apkarian and R. J. Adams, Advanced gain-scheduling techniques
for uncertain systems, IEEE Trans. Control Syst. Technol., 60: 21
32, 1998.
Extended and Pseudo-Linearization
C. Reboulet and C. Champetier, A new method for linearizing nonlinear systems: the pseudolinearization, Int. J. Control, 40: 631
638, 1984.
W. T. Baumann and W. J. Rugh, Feedback control of nonlinear systems by extended linearization, IEEE Trans. Autom. Control, 31:
4046, 1986.
J. Wang and W. J. Rugh, Linearized model matching for single-input
nonlinear systems, IEEE Trans. Autom. Control, 33: 793796,
1988.
J. Huang and W. J. Rugh, On a nonlinear servomechanism problem,
Automatica, 26: 963972, 1990.
J. Huang and W. J. Rugh, Approximate noninteracting control with
stability for nonlinear systems, IEEE Trans. Autom. Control, 36:
295304, 1991.
D. A. Lawrence and W. J. Rugh, Input-output pseudolinearization for
nonlinear systems, IEEE Trans. Autom. Control, 39: 22072218,
1994.
I. Kaminer, A. M. Pascoal, P. P. Khargonekar, and E. Coleman, A
velocity algorithm for the implementation of gain-scheduled controllers, Automatica, 31: 11851192, 1995.
D. A. Lawrence and W. J. Rugh, Gain scheduling dynamic linear controllers for a nonlinear plant, Automatica, 31: 381390, 1995.
JEFF S. SHAMMA
The University of Texas at Austin
215
H INFINITY CONTROL
This article describes an optimal multivariable control system design technique for achieving robustness under
external disturbances and model uncertainties. We show that many robust control problems can be formulated
as H norm optimization problems, and we describe analytically their solutions. We also give guidelines to the
choice of design parameters and insights to this optimal and robust control theory.
One of the motivations for the original introduction of H methods by Zames (1) was to bring plant
uncertainty, specified in the frequency domain, back to the center stage as it had been in classical control, in
contrast to analytic methods such as linear quadratic Gaussian (LQG) control. The H norm was found to be
appropriate for specifying both the level of plant uncertainty and the signal gain from disturbance inputs to
error outputs in the controlled system.
The standard H optimal control problem is concerned with the feedback system shown in Fig. 1(a)
where w represents an external disturbance, y is the measurement available to the controller, u is the output
from the controller, and z is an error signal that should be kept small. The transfer function matrix G represents
not only the conventional plant to be controlled but also any weighting functions included to specify the desired
performance, which will be discussed in more detail later. Suppose that G is partitioned consistent with the
inputs w and u and outputs z and y as
The closed loop transfer function from w to z, denoted by T zw , is defined to be the linear fractional transformation
(LFT) of G on K:
The H optimal control problem is then to design a stabilizing controller K, so as to minimize the H norm of
T zw , which is defined in the next section and is denoted by T zw . The H norm gives the maximum energy
gain, or sinusoidal gain of the system. This is in contrast to the H 2 norm T zw 2 , which for example gives the
variance of the output given white noise disturbances. The important property of the H norm comes from
the application of the small gain theorem, which states that if T zw , then the system in Fig. 1(b) will be
stable for all stable with < 1/. It is probably the case that this robust stability consequence was one
of the main motivations for the development of H methods. The synthesis of controllers that achieve an H
norm specification hence gives a well-defined practical and mathematical problem.
1
H INFINITY CONTROL
Fig. 1. Most control systems can be put in this unified framework where w represents an external disturbance, y is the
measurement available to the controller, u is the output from the controller, and z is an error signal that it is desired to
keep small.
H 2 and H Norms
We consider a q-input and p-output dynamical system with the matrix transfer function G(s). Let G(s) have
the following stabilizable and detectable state space realization:
Many control design problems can be regarded as finding a suitable controller so that the undesirable responses
of the system are made small in some sense. There are obviously many ways to define the smallness for a given
control problem. Here we are mainly interested in one way of defining the smallness: the H norm. For
comparison we shall also mention another more classical way of defining the smallness in terms of the H 2
norm.
Let R H 2 denote the set of strictly proper and real rational stable transfer matrices. In terms of state-space
realizations, R H 2 is simply the set of finite dimensional systems with D = 0 and stable A. The H 2 norm of a
G(s) R H 2 is defined as
H INFINITY CONTROL
where trace(M), M, and i (M) denote, respectively, the trace, the complex conjugate transpose, and the ith
singular value of the matrix M. This norm can also be computed in the time domain as
where g(t) is the inverse Laplace transform of G(s). Thus the H 2 norm of a system is a measure of the total
energy of the system impulse response. It can be computed using state-space realization as
where Q and P are observability Gramian and controllability Gramian, which can be obtained from the following
Lyapunov equations
Let R H denote the set of proper (but not necessarily strictly proper) and real rational stable transfer
matrices. In terms of state-space realizations, R H includes all finite dimensional systems with stable A
matrices. The H norm of a transfer matrix G(s) R H is defined as
where (M)
denotes the largest singular value of a matrix M. When G(s) is a single input and single output
system, the H norm of the G(s) is simply the peak value on the Bode magnitude plot of the frequency response
G(j). It can also be regarded as the largest possible amplification factor of the systems steady state response
to sinusoidal excitations. For example, the steady state response of the system with respect to a sinusoidal
input u(t) = U sin(0 t + ) is
which is precisely the H norm of the transfer function. As an example, consider a standard second-order
system
Then max = n
1 22 and G = |G(jmax )| = 1/(2
1 2 ). If G(s) is the description of
a structure vibration, then max would be the most dangerous exciting frequency.
In the multiple input and multiple output (MIMO) case, the H norm of a transfer matrix G R H is
the peak value on the largest singular value Bode plot of the frequency response G(j). Analogous to the scalar
H INFINITY CONTROL
case, the H norm of G(s) can also be regarded as the largest possible amplification factor of the systems
steady state response to sinusoidal excitations in the following sense: Let the sinusoidal inputs be
where denotes the time domain convolution and L 2 [0, ) denotes the space of all square integrable functions
with the norm defined as f 2 : =
. Thus it is important to make the H norm of all undesirable
transfer functions small in a feedback control system. That is one of the motivations for the development of
H control theory.
This discussion shows that the H norm of a transfer function can, in principle, be obtained either
graphically or experimentally. To get an estimate, set up a fine grid of frequency points, {1 , . . ., N }. Then an
estimate for G is
H INFINITY CONTROL
This value is usually read directly from a singular value Bode plot. The H norm can also be computed directly
using state-space representations. Let
Then
where
Hence, the H norm of a matrix transfer function can be calculated to the specified accuracy by a bisection
search.
Example. Consider a mass/spring/damper system as shown in Fig. 2. The dynamical system can be
described by the following differential equations:
Suppose that G(s) is the transfer matrix from (F 1 , F 2 ) to (x1 , x2 ) and suppose k1 = 1, k2 = 4, b1 = 0.2, b2 =
0.1, m1 = 1, and m2 = 2 with appropriate units. Then the H 2 norm of this transfer matrix is G(s)2 = 2.56,
whereas the H norm of this transfer matrix is G(s) = 11.47, which is shown as the peak of the largest
singular value Bode plot in Fig. 3. Since the peak is achieved at max = 0.8483, exciting the system using the
H INFINITY CONTROL
Fig. 2. A two mass/spring/damper system with the external forces F 1 and F 2 as inputs and the positions of the two masses
as outputs.
Fig. 3. The singular value Bode plot of the two mass/spring/damper system. The H norm G is the peak of the largest
singular value plot of G(j).
This shows that the system response will be amplified 11.47 times for an input signal at the frequency max ,
which could be undesirable if F 1 and F 2 are disturbance force and x1 and x2 are the positions to be kept steady.
We will see later how to design an H feedback controller to suppress this kind of vibration.
We note that the state-space computational method is usually much more accurate than the graphical
method. Consider, for example, the standard second-order system again with n = 1 and = 0.01. By the
H INFINITY CONTROL
analytic formula or the state-space computational method, we get that the H norm is 50.0025. To estimate
the H norm graphically, we set up a frequency grid to compute the frequency response of G over a suitable
range of frequency. Take, for example, 200 points in the frequency range of [0.1, 10] uniformly on the log scale,
then we get an estimate of the norm 33.0743. This shows clearly that the graphical method may lead to a
wrong answer for a lightly damped system if the frequency grid is not sufficiently dense. Indeed, we would
get the H norm 43.5567, 48.1834, and 49.5608 from the graphical method if 400, 800, and 1600 frequency
points are used, respectively.
Weighted H Performance
We now consider how to formulate some performance objectives into mathematically tractable problems. It
is well known that the performance objectives of a feedback system can usually be specified in terms of
requirements on the sensitivity functions and/or complementary sensitivity functions or in terms of some other
closed-loop transfer functions. For instance, the performance criteria for a scalar system may be specified as
requiring
with S(j) = 1/(1 + P(j)K(j)) where P is the plant and K is the controller. However, it is much more convenient
to reflect the system performance objectives by choosing appropriate weighting functions. For example, this
performance objective can be written as
with
In order to use W e in control design, a rational transfer function W e (s) is usually used to approximate this
frequency response.
The advantage of using weighted performance specifications is obvious in multivariable system design.
First, some components of a vector signal are usually more important than others. Second, each component of
the signal may not be measured in the same units; for example, some components of the output error signal
may be measured in terms of length, and the others may be measured in terms of voltage. Weighting functions
are essential to make these components comparable. Also, we might be primarily interested in rejecting errors
in a certain frequency range (e.g., low frequencies); hence, some frequency-dependent weights must be chosen.
Consider a standard feedback diagram in Fig. 4. The weighting functions in Fig. 4 are chosen to reflect
the design objectives and knowledge on the disturbances and sensor noise. For example, W d and W i may be
chosen to reflect the frequency contents of the disturbances d and di . The weighting matrix W n is used to model
the frequency contents of the sensor noise, whereas W e may be used to reflect the requirements on the shape
of certain closed-loop transfer functions (e.g., the shape of the output sensitivity function). Similarly, W u may
be used to reflect some restrictions on the control or actuator signals, and the dashed precompensator W r is an
H INFINITY CONTROL
Fig. 4. Standard feedback configuration with disturbance weights and performance weights. W i , W d , and W n represent
the frequency contents of the input disturbance, output disturbance, and the sensor noise. W e represents the disturbance
rejection requirement, and W u puts the limit on the control effort. W r shapes the input signal.
optional element used to achieve deliberate command shaping or to represent a nonunity feedback system in
equivalent unity feedback form.
A typical control design may involve making the sensitivity function small over a suitable frequency
range while keeping the control effort within a reasonable limit. This may be mathematically formulated as
minimizing
Note that can be absorbed into W u , so there is no loss of generality in assuming = 1. Finding a controller
so that the H norm of a certain closed-loop transfer function, such as the preceding one, is minimized is the
H control problem.
Similar H 2 norm minimization problems can be formulated if the disturbance is modeled as white noise
and the performance is measured in terms of output power.
H INFINITY CONTROL
Robust Stabilization
Another way that a weighted H norm minimization problem can arise naturally is when we consider robust
stability and robust performance of a closed-loop system with model uncertainties. For example, consider a
unity feedback system with a family of additively perturbed uncertain dynamical systems:
and assume that K stabilizes the nominal plant P. Then by the small gain theorem, the uncertain system is
stable for all admissible with < 1/ if and only if
Therefore, a related synthesis problem is to find a controller K so that this inequality holds.
be a normalized coprime factorization, that is,
1N
As another example, let P = M
A controller K stabilizing the nominal system P will robustly stabilize the family Fp if and only if
Similarly, many other robust stability problems can be formulated. It should be noted that the H 2 norm cannot
be used in the robust stability test because it does not satisfy a key multiplicative property (i.e., G1 G2 2
G1 2 G2 2 in general).
10
H INFINITY CONTROL
Consider the SISO feedback system shown in Fig. 4. Then the tracking error is
where S = (I + PK) 1 is the output sensitivity function and T = I S = PK(I + PK) 1 is the output
complementary sensitivity function. Note that tracking error is closely related to the low-frequency gain of S.
In particular, we must keep |S| small over the range of frequencies, typically low frequencies where r and d
are significant. For example, if we need the steady state error with respect to a step input to be no greater
than , then we need |S(0)| . Hence, the steady state tracking requirement can be fulfilled by constraining
the low-frequency gain of S. From classical control theory, we know that the dynamical quality of the system
time response can be quantified by rise time, settling time, and percent overshoot. Furthermore, the speed
of the system response is inversely proportional to the closed-loop bandwidth and the overshoot of the system
response increases with the resonant peak sensitivity defined as M s := S . Let b = min {:|S(j)| 1}.
Then we can regard b as the closed-loop bandwidth because, beyond this frequency, the closed-loop system
will not be able to track the reference and the disturbance will actually be amplified.
Now suppose that we are given the time domain performance specifications. Then we can determine the
corresponding requirements in frequency domain in terms of the low-frequency gain, the bandwidth b , and
the peak sensitivity M s . Hence, a good control design should result in a sensitivity function S satisfying all
these requirements as shown in Fig. 5. These requirements can be approximately represented as
If a steeper transition between low frequency and high frequency is desired, the weight W e can be modified as
follows:
H INFINITY CONTROL
11
The selection of control weighting function W u follows similarly from the preceding discussion by considering the control signal equation
The magnitude of |KS| in the low-frequency range is essentially limited by the allowable cost of control effort
and saturation limit of the actuators; hence, in general the maximum gain M u of KS can be fairly large,
whereas the high-frequency gain is essentially limited by the controller bandwidth bc and the (sensor) noise
frequencies. Ideally, we would like to roll off as fast as possible beyond the desired control bandwidth so that
the high-frequency noises are attenuated as much as possible. Hence, a candidate weight W u would be
However, the standard H control design techniques cannot be applied directly to a problem with an improper
control weighting function. Hence, we shall introduce a faraway pole to make W u proper:
for a small 1 > 0 as shown in Fig. 5. Similarly, if a faster rolloff is desired, we may choose
12
H INFINITY CONTROL
and
For the following discussion, let us assume that state-space models of G and K are available and that their
realizations are assumed to be stabilizable and detectable. We say that a controller is admissible if it internally
stabilizes the system.
Optimal H 2 Control:
Find an admissible controller K(s) such that T zw 2 is minimized.
Optimal H Control:
Find an admissible controller K(s) such that T zw is minimized.
It should be noted that the optimal H controllers as defined here are generally not unique for MIMO systems. Furthermore, finding an optimal H controller is often both numerically and theoretically complicated.
This is certainly in contrast to the standard LQG or H 2 theory, in which the optimal controller is unique and can
be obtained by solving two Riccati equations without iterations. Knowing the achievable optimal (minimum)
H norm may be useful theoretically because it sets a limit on what we can achieve. In practice, however, it
is often not necessary and sometimes even undesirable to design an optimal controller, and it is usually much
cheaper to obtain controllers that are very close in the norm sense to the optimal ones, which will be called
suboptimal controllers. A suboptimal controller may also have other nice properties (e.g., lower bandwidth)
over the optimal ones.
Suboptimal H Control:
Given > 0, find an admissible controller K(s), if there are any, such that T zw < .
For these reasons mentioned, we shall focus our attention on the suboptimal H control. We shall assume
that the realization of the transfer matrix G takes the following form:
which is compatible with the dimensions of z(t) Rp 1 , y(t) Rp 2 , w(t) Rm1 , and u(t) Rm2 and the state x(t)
Rn . We make the following assumptions:
H INFINITY CONTROL
13
Assumption 1 is necessary for the existence of stabilizing controllers. Assumption 2 means that the penalty
on z = C1 x + D12 u includes a nonsingular penalty on the control u, that the exogenous signal w includes
both plant disturbance and sensor noise, and that the sensor noise weighting is nonsingular. Relaxation of
assumption 2 leads to singular control problems.
Assumptions 3 and 4 are made for a technical reason: together with assumption 1 they guarantee that the
two algebraic Riccati equations in the corresponding LQG or H 2 problem have the desired solutions. Dropping
assumptions 3 and 4 would make the solution very complicated.
Define
Theorem 2. Suppose G satisfies assumptions 14. Then there exists an admissible controller K such that
F
(G, K ) < (i.e., T zw < ) if and only if
(1) there exists an X 0 such that
14
H INFINITY CONTROL
Furthermore, if these conditions are satisfied, all internally stabilizing controllers K (s) satisfying F
(G,
K ) < can be parameterized as
is called the central controller. Comparing the H central controller to the H 2 optimal controller, we can see
that the H central controller will approach the optimal H 2 controller as .
Example. Consider again the two mass/spring/damper system shown in Fig. 2. Assume that F1 is the
control force, F 2 is the disturbance force, and the measurements of x1 and x2 are corrupted by measurement
noise:
Our objective is to design a control law so that the effect of the disturbance force F 2 on the positions of the two
masses x1 and x2 are reduced in a frequency range 0 2. The problem can be set up as shown in Fig. 6,
where
H INFINITY CONTROL
15
Fig. 6. Rejecting the disturbance force F 2 of the two mass/spring/damper system by a feedback control of F 1 .
is the performance weight and W u is the control weight. In order to limit the control force, we shall choose
Now let u = F 1 ,
respectively. Let
that is, we want to reject only the effect of the disturbance force F 2 on the position x1 . Then the optimal H 2
performance is F
(G, K 2 )2 = 2.6584, and the H performance with the optimal H 2 controller is F
(G, K 2 )
= 2.6079, whereas the optimal H performance with an H controller is F
(G, K ) = 1.6101. This means
16
H INFINITY CONTROL
Fig. 7. The largest singular value Bode plot of the closed-loop two mass/spring/damper system with an H 2 controller and
an H controller.
that the effect of the disturbance force F 2 in the desired frequency range 0 2 will be effectively reduced
with the H controller K by 5/1.6101 = 3.1054 times at x1 . On the other hand, let
that is, we want to reject only the effect of the disturbance force F 2 on the position x2 . Then the optimal H 2
performance is F
(G, K 2 )2 = 0.1659, and the H performance with the optimal H 2 controller is F
(G, K 2 )
= 0.5202, whereas the optimal H performance with an H controller is F
(G, K ) = 0.5189. This means
that the effect of the disturbance force F 2 in the desired frequency range 0 2 will be effectively reduced
with the H controller K by 5/0.5189 = 9.6358 times at x2 . Finally, set
that is, we want to reject the effect of the disturbance force F 2 on both x1 and x2 . Then the optimal H 2
performance is F
(G, K 2 )2 = 4.087, and the H performance with the optimal H 2 controller is F
(G, K 2 )
= 6.0921, whereas the optimal H performance with an H controller is F
(G, K ) = 4.3611. This means
that the effect of the disturbance force F 2 in the desired frequency range 0 2 will be effectively reduced
only with the H controller K by 5/4.3611 = 1.1465 times at both x1 and x2 . This result shows clearly that it
is very hard to reject the disturbance effect on both positions at the same time. The largest singular value Bode
plots of the closed-loop system are shown in Fig. 7. We note that the H controller typically gives a relatively
flat frequency response because it tries to minimize the peak of the frequency response. On the other hand, the
H 2 controller typically produces a frequency response that rolls off fast in the high-frequency range but with
a large peak in the low-frequency range.
H INFINITY CONTROL
17
H Filtering
In this section we shall illustrate how an H filtering problem can be converted to a special H control
problem. Suppose that a dynamic system is described by the following equations:
The filtering problem is to find an estimate z of z in some sense using the measurement of y. The restriction on
the filtering problem is that the filter has to be causal so that it can be realized (i.e., z must be generated by a
causal system acting on the measurements). We will further restrict our filter to be unbiased, that is, given T
> 0, the estimate z (t) = 0 t [0, T] if y(t) = 0, t [0, T]. Now we can state our H filtering problem.
H Filtering:
Given a > 0, find a causal filter F(s) R H if it exists such that
with z = F(s)y.
This H filtering problem can also be formulated in an LFT framework because
Hence, the filtering problem can be regarded as a special H problem. However, comparing this filtering
problem to the control problems, we can see that there is no internal stability requirement in the filtering
problem. Hence, the solution to this filtering problem can be obtained from the H solution in the last section
by setting B2 = 0 and dropping the internal stability requirement. Thus, a rational causal filter F(s) is given
by
Understanding H Control
Most existing derivations and proofs of the H control results given in Theorem 2 are mathematically quite
complex. Some algebraic derivations are simple but they provide no insight to the theory for control engineers.
In this section, we shall present an intuitive but nonrigorous derivation of the H results by using only
some basic system theoretic concept such as state feedback and state estimation. In fact, we shall construct
intuitively the output feedback H central controller by combining an H state feedback and an observer.
18
H INFINITY CONTROL
A key fact we shall use is the so-called bounded real lemma, which states that for a system z = G(s)w with
state space realization G(s) = C(sI A) 1 B H , G < , which is essentially equivalent to
and A + BB
X/ 2 is stable. Dually, there is a Y = Y
0 such that
and A + YC
C/ 2 is stable.
Note that the system has the following state space realization:
To keep the presentation simple, we shall make some additional assumptions: D 12 C1 = 0, B1 D 21 = 0, D 12 D12
= I, and D21 D 21 = I.
We shall first consider state feedback u = Fx. Then the closed-loop system becomes
which gives
This is exactly the X Riccati equation under the preceding simplified conditions. Hence, we can take F = F
and X = X .
H INFINITY CONTROL
19
Next, suppose that there is an output feedback stabilizing controller such that T zw < . Then x() =
0 because the closed-loop system is stable. Consequently, we have
Substituting x = Ax + B1 w + B2 u and z = C1 x + D12 u into the above integral and using the X equation, and
finally completing the squares with respect to u and w, we get
Hence the original H control problem is equivalent to finding a controller so that T vr < or
Obviously, this also suggests intuitively that the state feedback control can be u = F x and a worst state
feedback disturbance would be w = 2 B 1 X x. Since full state is not available for feedback, we have to
implement the control law using estimated state:
where x is the estimate of x. A standard observer can be constructed from the new system equations as
20
H INFINITY CONTROL
Since it is assumed that T vr < , it follows from the dual version of the bounded real lemma that there
exists a Y 0 such that
which gives
which is exactly the H central controller given in Theorem 2 under the simplified conditions.
We can see that the H central controller can be obtained by connecting a state feedback with a state
estimate under the worst state feedback disturbance.
H Loop Shaping
Consider the family of uncertain systems Fp again. It is now clear that finding a controller K such that
it robustly stabilizes the family Fp is a standard H norm minimization problem. Now suppose P has a
H INFINITY CONTROL
21
N)
given by
Then the left coprime factorization (M,
Define
Then bP,K > 0 implies that K also stabilizes robustly the following family of uncertain systems:
(1)
(2)
(3)
P = P + a such that P and P have the same number of unstable poles and a < bP,K .
P = (I + m )P such that P and P have the same number of unstable poles and m < bP,K .
P = (I + f ) 1 P such that P and P have the same number of unstable poles and f < bP,K .
These conclusions also hold when the roles of plant and controller are interchanged. The number bP,K can
also be related to the classical gain and phase margins of a SISO system:
22
H INFINITY CONTROL
Obviously, bopt is the largest admissible size of perturbation so that the system is stable. It follows from the
standard H solution that
Moreover, for any > 1/bopt (P), a controller achieving bP,K > 1/ is given by
where
This stabilization solution can be used to devise an H loop-sharing design method. The objective of this
approach is to incorporate the simple performance/robustness trade-off obtained in the loop shaping with the
guaranteed stability properties of H design methods. Recall that good performance controller design requires
that
particularly in some low-frequency range where (PK) denotes the smallest singular value. And good robustness requires that
H INFINITY CONTROL
23
A typical design works as follows: the designer inspects the open-loop singular values of the nominal plant,
and shapes these by pre- and/or postcompensation until nominal performance (and possibly robust stability)
specifications are met. (Recall that the open-loop shape is related to closed-loop objectives.) A feedback controller
K with associated stability margin (for the shaped plant) bopt (Ps ) is then synthesized. If bopt (Ps ) is small,
then the specified loop shape is incompatible with robust stability requirements, and should be adjusted
accordingly; then K is reevaluated.
Note that in contrast to the classical loop-sharing approach, the loop shaping here is done without explicit
regard for the nominal plant phase information. That is, closed-loop stability requirements are disregarded at
this stage. Also in contrast with conventional H design, the robust stabilization is done without frequency
weighting.
In fact, the preceding robust stabilization objective can also be interpreted as the more standard H
problem formulation of minimizing the H norm of the frequency weighted gain from disturbances on the
plant input and output to the controller input and output as follows:
This shows that the H loop-shaping design is equivalent to a standard H design with the shaping functions
as weighting functions.
Synthesis
As we discussed at the beginning of this article, F
(G, K) guarantees the robust stability of the uncertain
system shown in Fig. 1(b) for any (s) R H with < 1/. However, if a system is built from components,
which are themselves uncertain, then, in general, the uncertainty in the system level is structured, and this
robust stability guarantee may be overly conservative. Because the interconnection model G can always be
chosen so that (s) is block diagonal, and, by absorbing any weights, < 1. Thus we can assume that (s)
takes the form of
with i < 1 and j < 1. The robust stability analysis for systems with such structured uncertainty is
not as simple but can be formally characterized by using the structured singular value, see Ref. 5.
Define Cnn as
24
H INFINITY CONTROL
The itself is not easy to compute. But good bounds can be obtained efficiently. Let
and the equality holds if 2S + F 3. This bound can be used frequency by frequency to determine the system
robust stability and performance with structured uncertainties. For example, the system in Fig. 1(b) is well
posed, internally stable for all stable (s) with (s0 ) , Re(s0 ) 0, and < 1/ if and only if
This synthesis problem is not yet fully solved in the general case. A reasonable approach is to obtain a solution
to an upper bound:
by iteratively solving for K and D. This is the so-called D-K iteration. The stable and minimum phase scaling
matrix D(s) is chosen such that D(s)(s) = (s)D(s). For a fixed scaling transfer matrix D, minK DF
(G,
K)D 1 is a standard H optimization problem because
Then a D(s) is found to approximate the magnitude frequency response D uniformly (usually by curve fitting).
D-K iterations proceed by performing this two-parameter minimization in sequential fashion: minimizing over
H INFINITY CONTROL
25
K with D(s) fixed, minimizing pointwise over D with K fixed, minimizing again over K, and then again over D,
and so on. With either K or D fixed, the global optimum in the other variable may be found using the and
H solutions. Although the joint optimization of D and K is generally not convex and the global convergence
is not guaranteed, many designs have shown that this approach works very well. In fact, this is probably the
most effective design methodology available today for dealing with such complicated problems.
Additional Applications
There are many additional extensions and development in the H control theory. Here are some of them:
H loop-shaping techniques using -gap metric, see Ref. 6.
Robust control design in the gap metric, see Robust control and Ref. 6.
Linear matrix inequality (LMI) approach to H control, see Convex optimization.
Time-varying and finite horizon H control and game theoretical approach to H control, see Refs. 7 and 8.
Operator theoretic approach to H control, see Ref. 9.
Chain-scattering approach to H control, see Ref. 10.
H control with pole placement, see Ref. 11.
H controller reduction, see Refs. 12 and 13.
Linear parameter varying H control, see Ref. 14.
Sampled-Data H control, see Ref. 15.
H control for infinite dimensional systems, see Refs. 16 and 17.
H control for nonlinear systems, see Ref. 18.
Software and applications, see Ref. 19.
A comprehensive treatment of H control theory can be found in Refs. 13 and 20.
BIBLIOGRAPHY
1. G. Zames, Feedback and optimal sensitivity: model reference transformations, multiplicative semi-norms, and approximate inverses, IEEE Trans. Autom. Control, 26: 301320, 1981.
2. K. Glover and J. Doyle, State-space formulae for all stabilizing controllers that satisfy an H norm bound and relations
to risk sensitivity, Systems and Control Letters, 11: 167172, 1988.
3. J. C. Doyle et al., State-space solutions to standard H 2 and H control problems, IEEE Trans. Autom. Control, 34(8):
831847, 1989.
4. D.C. McFarlane and K. Glover, A loop shaping design procedure using H synthesis, IEEE Trans. Autom. Control,
37(6): 759769, 1992.
5. A. Packard and J. C. Doyle, The complex structured singular value, Automatica, 29: 71109, 1993.
6. G. Vinnicombe, Frequency domain uncertainty and the graph topology, IEEE Trans. Autom. Control, 38(9): 13711383,
1993.
7. M. Green and D. J. N. Limebeer, Linear Robust Control, Englewood Cliffs, NJ: Prentice-Hall, 1995.
8. T. Basaer and P. Bernhard, H -Optimal Control and Related Minimax Design Problems: A Dynamic Game Approach,
1997.
10. H. Kimura, Chain-Scattering Approach to H -Control, Boston: Birkhauser,
26
H INFINITY CONTROL
11. M. Chilali and P. Gahinet, H design with pole placement constraints: An LMI approach, IEEE Trans. Autom. Control,
41(3): 358367, 1996.
12. P. J. Goddard and K. Glover, Performance preserving frequency weighted controller approximation: A coprime factorization approach, Proceedings of the 33nd Conference on Decision and Control, Orlando, FL, pp. 27202725, December
1994.
13. K. Zhou, J. C. Doyle, and K. Glover, Robust and Optimal Control, Upper Saddle River, NJ: Prentice-Hall, 1996.
14. A. Packard, Gain scheduling via linear fractional transformations, Systems and Control Lett., 22: 7992, 1994.
15. T. Chen and B. A. Francis, Optimal Sampled-Data Control Systems, London: Springer-Verlag, 1995.
16. R. F. Curtain and H. J. Zwart, An Introduction to Infinite-Dimensional Linear Systems Theory, New York: SpringerVerlag, 1995.
KEMIN ZHOU
Louisiana State University
POSITION CONTROL
Position control has many applications, such as control of the elevator angle of a fighter, control of the
antenna angle in a satellite tracking system, and control of robot manipulators. Often, the essence of position
control is that the tracking error between the desired system output and the actual system output is used
to generate a suitable control input to drive the tracking error to zero. In other words, tracking control is an
important part of motion control, as it solves the problem of determining the control inputs necessary for a
system to track a desired trajectory and provides a way to achieve accurate performance.
In this article, we present adaptive position control of robot manipulators and teleoperation systems.
Robot manipulators are composed of links connected by joints. The joints may be electrically, hydraulically,
or pneumatically actuated. The number of joints determines the number of degrees of freedom (DOF) of the
manipulator. Position control of a robot manipulator involves control of the positions of the joints. Once given
a set of desired trajectories for all the joints, the controller is designed to track these trajectories so that the
end effector of the manipulator sweeps the desired positions in the workspace. The primary method of sensing
the positions is with position encoders located on the joints, either on the shaft of the motor that actuates the
joint or on the joint itself. At times, direct sensing of the end-effector position with the help of a camera is used
to improve the accuracy of the manipulator in tracking a desired trajectory.
A teleoperation system involves two distant yet coupled robots: a local master robot and a remote slave
robot. In teleoperation, the human operator controls the master robot. Motion commands are measured on the
master robot and transmitted to the slave robot, which executes these commands and is expected to track the
motion of the master robot. In addition, the contact force information sensed by the slave robot is reflected
to the master robot for force perception. Thus, the master acts as an position input device that generates a
desired trajectory. The goal of position or tracking control is to design the necessary control input that makes
the slave track the motion of the master. This control problem is an example of masterslave control.
The article is organized as follows: In the following section, we present robust adaptive control schemes
of robot manipulators. First, we present dynamic models of robot manipulators with time-varying parameters
or unmodeled dynamics. Second, we present the controller structure and adaptive law for the time-varying
parameter case and show the signal boundedness and the tracking performance of the robot system. Third, we
present and analyze stable adaptive control schemes for robot manipulators with unmodeled dynamics. Some
common topics of position control relevant to robot manipulators such as PD control, inverse dynamics, and
path or trajectory interpolation are discussed in the fourth subsection. In the third section, we present adaptive
control of teleoperation systems. Adaptive control schemes for teleoperation systems with unknown jumping
parameters and with parametrizable and unparametrizable smooth time-varying parameters are presented.
We also present some control issues relevant to teleoperation systems with communication time delays.
POSITION CONTROL
where q = (q1 ,. . ., qn )T is a set of position variables of n (n > 0) joints of the robot manipulator, u = (u1 ,. . ., un )T
is the applied joint torque, and L is the Lagrangian defined as L = K P, the difference between the kinetic
energy K and the potential energy P, in the form
POSITION CONTROL
Fig. 1.
Robot manipulator.
with D (q, t) Rnn being the symmetric and positive definite manipulator inertia matrix. For Eqs. (1), (2) to
represent the manipulator dynamics with time-varying parameters, the mass and the moment of inertia of
each link of the manipulator should not explicitly depend on q .
Letting dij be the ijth element of D (q, t) and (q, t) = P (q, t)/q, and substituting Eq. (2) in Eq. (1), we
obtain the manipulator dynamic equation:
where the kjth element of C(q, q , t Rnn is ckj = n i = 1 (dkj /qi + dki /qj dij /qk ) q i.
A key feature of the manipulator model (3) is that the inertia matrix D(q, t) and the potential energy P(q,
t) are explicitly time-dependent, which takes into account the effect of changes in environment of the robot
system or changes of the manipulator dynamics with time. Moreover, an important property of the manipulator
model (3) is
where M(q, q , t) = dD(q, t)/dt D(q, t)/t, whose ijth element is (dij /q)T q . When D(q, t) = D(q) does not
explicitly depend on t, that is, D(q, t)/t = 0, Eq. (4) becomes xT [dD(q)/dt 2C(q, q )]x = 0, which is well known
in the robotics literature.
A manipulator with unmodeled dynamics may be modeled as
POSITION CONTROL
where H i , i = 1, 2, 3, with appropriate dimensions, are linear operators with rational transfer matrices,
representing the unmodeled part of the robot dynamics, and g1 (q), g2 (q), g3 (u) are certain vector functions of
q , q, u. The functions D(q), C(q, q ), (q) have been defined above, and, for simplicity, they are assumed not
explicitly time-dependent in the unmodeled dynamics problem. The manipulator model (5) is generalized from
some practical robot systems [Reed and Ioannou (3); Ortega and Spong (9)].
Control Objective and Parametrization. Our control objective is, for a given reference signal qd (t), to
generate the applied torque u for the manipulator (3) or (5) with unknown parameters so that all signals in the
robot system are bounded and the joint position q tracks qd as closely as possible. To achieve such an objective
we first use the transformation technique developed in Slotine and Li (1,2) to parametrize the manipulator
model (3) or (5).
Let be any n n constant matrix whose eigenvalues have positive real parts; define
and s, v, v depend only on q, qd , q , q d , q d and not on the joint acceleration vector q (t).
Using Eq. (7), we express the manipulator model (3) as
where Y (q, qd , q , q d , q d , t) is an n r matrix of known functions for some r > 0, and (t) Rr contains
parameters, which may be time-varying. In Eq. (8), the regressor Y (q, qd , q , q d , q d , t) is bounded for bounded
q, qd , q , q d , q d
Let xt be the truncated x at time t. Denote by the Euclidean vector norm or the induced matrix norm,
and by (1 , 2 ) the L (L1 , L2 ) vector norm or the induced operator norm [Desoer and Vidyasagar (10)],
as the case may be.
We make the following assumptions about the manipulator model (8):
(t) for some constants 0 > 0, > 0;
(1) (t) 0 ,
(2) D(q, t)/t f (q) for some constant > 0 and known f (q) bounded for bounded q.
Similarly, for the manipulator model (5), we obtain
POSITION CONTROL
Fig. 2.
(1) (g1 (q))t f 1 (q(t)), (g2 (q))t f 2 (q(t)) for some known f 1 (q), f 2 (q) that are bounded for bounded q , q,
and H 1 1 , H 2 2 for some constants 1 > 0, 2 > 0;
(2) (g3 (u))t ut , and H 3 i i , where H 3 i is the ith row of H 3 , i = 1,. . ., n.
We also make an assumption on the desired joint position vector qd (t):
(1) qd (t), q d (t), q d (t) are bounded.
Assumption (A1) requires only the boundedness of the manipulator parameters and their derivatives, not
the smallness of the time variations of the parameters. Smallness of the parameter variations is usually an
assumption for the design of adaptive control schemes for time-varying plants, but it is not needed here because
of the special structure of the robot manipulator dynamics. Assumption (A2) requires that D (q, t)/t satisfy
a certain relative boundedness condition. Assumption (A3) requires that the L gains of H 1 , H 2 be finite and
g1 (q), g2 (q) satisfy certain relative boundedness conditions. Assumption (A4) is similar to (A3), but i 0, i =
1,. . ., n, are to be specified for the robust stability of the adaptive robot system. We note that the bounds , 1 ,
2 are not needed for the adaptive controller design.
An Illustrative Example. In this sub-subsection, we consider a two-link planar manipulator [Spong and
Vidyasagar (8)], shown in Fig. 2 as an illustrative example for the robot system modeling and parametrization.
The manipulator configuration may be described as follows: there are two revolute joints with joint angles
q1 , q2 , and two links with masses M 1 , M 2 , lengths l1 , l2 , distances lc1 , lc2 from the joints to the mass centers,
and rotational inertias I1 , I2 . The inertia matrix D(q, t) has four elements: d11 = M 1 lc1 2 + M 2 (l1 2 lc2 2 + 2l1 lc2
cos q2 )+ I1 +l2 , d12 = d21 = M 2 (lc2 2 + l1 lc2 cos q2 ) + I2 , d22 = M 2 lc2 2 + I2 ; and the potential energy is P(q, t) =
(M 1 lc1 + M 2 l1 ) g sin q1 + M 2 lc2 g sin (q1 + q2 ), where g is the gravitational acceleration. The matrix C(q, q , t)
in Eq. (3) has four elements: c11 = h q 2 , c12 = (q1 + q 2 )h, c21 = q 1 h, c22 = 0, where h = M 2 l1 lc2 sin q2 .
POSITION CONTROL
The manipulator model without parameter variations and unmodeled dynamics is D(q) s + C(q, q )s = u
Y(q, qd , q , q d , q d ), where
When = (t) is time-varying, the manipulator model is Eq. (8) with Y(q, qd , q , q d , q d , t) and
guarantees that the closed-loop system is globally stable and convergent in the sense that q(t), q (t), (t) are
bounded, and limt e(t) = 0, as the positive definite function
(t) = sT (t)K D s(t) 0 [also see Spong et al. (11) for further analysis].
POSITION CONTROL
Fig. 3.
When D(q, t), P(q, t) are both explicitly time-dependent, we have obtained the manipulator model as Eq.
(8) in which (t) is time-varying and the term D(q, t)/t q appears. If the parameters in D(q, t), P(q, t) were
known, then (t) and D(q, t)/t q could be calculated so that the control law u(t) = Y(q, qd , q , q d , q d , t)(t)
+ [D(q, t)/t] q K D s could be implemented, which guarantees global stability and asymptotic tracking. For
unknown D(q, t), P(q, t), next we present an adaptive control scheme that is robust with respect to the time
variation of (t) and [D(q, t)/t] q .
With Y(q, qd , q , q d , q d , t), s, K D , defined before, we propose the following feedback controller structure
for the manipulator (8):
where (t) as shown in Fig. 3 is the switching signal [Ioannou and Tsakalis (12)] using a priori knowledge of
the upper bound M on supt0 (t):
This adaptive control scheme has the following stability and tracking properties.
POSITION CONTROL
Theorem 1. All closed-loop signals are bounded, and the tracking error e(t) = q(t) qd (t) satisfies
for some constants 0 > 0, 0 > 0, and any t2 > t1 0. Moreover, e(t) L2 and limt e(t) = 0 in the absence of
From Eqs. (4, 7, 8), (16, 17, 18) and from (A1), (A2), we have that
it follows from the second inequality of (2.22) that (t) 0 for (t), s(t) outside a certain bounded set. Therefore
s(t) and (t) are bounded, which, in view of Eqs. (7), (16), implies that q(t), q (t), u(t) are also bounded.
Using the fact that (t)
POSITION CONTROL
for some constant k1 > 0. Since V(t) is bounded, from Eq. (24) we have
Since the operator H is exponentially stable and e(t) is bounded, both H2 and t1 t2 H (t t1 )e(t1 )2 dt are
finite for any t2 > t1 0. Hence from Eqs. (25) and (26) we prove Eq. (20).
When (t)
= 0 and D(q, t)/t = 0, that is, = = 0, it follows from Eq. (20) that e(t) L2 . This, together
with the boundedness of e (t) = s(t) e(t), proves limt e(t) = 0.
To implement the controller (16), we need the knowledge of f (q) to generate the bounding signal m(t) in Eq.
(17). A more sophisticated choice of f (q) admits a wider class of D(q, t)/t, but may make the implementation of
m(t) more complicated. We also note that the above design does not need the knowledge of the bounds , . For
a chosen f (q), different choices of k0 in generating m(t) may have different effects on the tracking performance,
while increasing k0 may reduce the effect of in the mean error (20). For the signal boundedness and the
mean tracking error (20), parameter variations characterized by and are not required to be small. This is
an important feature of the robot system. With q , q available for measurement, the manipulator mode (3) is
equivalent to the first-order model (8), for which the adaptive controller allows large parameter variations to
exist.
Solution to the Unmodeled Dynamics Problem. Consider the manipulator (9) with unmodeled
dynamics. If the terms H 1 [g1 (q)](t), H 2 [g2 (q)](t), H 3 [g3 (u)](t) were available for measurement and were
known, then the control law u(t) = Y(q, qd , q , q d , q d ) H 1 [g1 (q)](t) H 2 [g2 (q)](t) H 3 [g3 (u)](t) K D s(t) could
be implemented so that d/dt[sT (t)D(q)s(t)] = 2sT (t)K D s(t), showing the boundedness of s(t) and exponentially
fast tracking. However, to ensure the boundedness of u(t), one needs H 3 = maxi = 1,...,n i < 1.
To solve the adaptive control problem in which H 1 [g1 (q)](t), H 2 [g2 (q)](t), H 3 [g3 (u)(t), and are unknown,
with Y(q, qd , q , q d , q d ), K D , s(t) = (s1 (t),. . ., sn (t))T , (t), defined before, we propose the following feedback
10
POSITION CONTROL
The stability and tracking properties of this adaptive control scheme are:
Theorem 2. All closed-loop signals are bounded for any i [0, i ] i = 1,. . .,n, and the tracking error e(t)
satisfies
for some constants 0 > 0, 0 > 0. Moreover, e(t) L2 and limt e(t) = 0 in the absence of the unmodeled
dynamics, that is, when H 1 = 0, H 2 = 0, H 3 = 0.
POSITION CONTROL
11
Proof:. Considering V(t) defined in (15), from Eqs. (9), (27), (32) and (A3), (A4) we obtain
Hence it follows from Eqs. (34), (35) that, for 0 i i , i = 1,. . .,n, we have (t) 0 whenever (t) and s(t)
are outside a certain bounded set, that is, s(t), (t) are bounded, and so are q(t), q (t).
From Eqs. (27, 28, 29, 30), (A3), (A4), and the boundedness of s(t), q(t), q (t), (t), we have
for some constant k2 > 0, which, together with Eq. (31), implies that u(t) is bounded.
Using Eqs. (34), (35) and the fact that (t)
(t)(t) 0, we obtain
for some constants 1 > 0, 1 > 0, and any t2 > t1 0, which implies Eq. (33).
12
POSITION CONTROL
When H 1 = 0, H 2 = 0, H 3 = 0, the expression (34) for
(t) becomes
Since sT (t)3 (t) 0, Eq. (39) shows that s(t) L2 . Hence, from Eq. (7), it follows that e(t) L2 and e (t) is bounded.
Therefore we have limt e(t) = 0.
We have thus proved the signal boundedness of the closed-loop system in the presence of H 1 [g1 (q)](t),
H 2 [g2 (q)](t), H 3 [g3 (q)](t). The gains of the linear operators H 1 , H 2 are assumed to be finite but not small. The
gain of H 3 is required to be small to ensure the boundedness of u(t).
The modifying term (t)(t) in Eq. (32) can be replaced by 0 (t), 0 > 0. The signal boundedness follows,
but Eq. (33) is changed to
This scheme cannot guarantee asymptotic tracking in the absence of the unmodeled dynamics, though the
scheme does not need the knowledge of the upper bound on .
The use of the bounding signals mi (t) defined in Eq. (28) is the key to ensuring signal boundedness in
the presence of the unmodeled dynamics satisfying (A3), (A4). To generate these signals, the knowledge of the
stability margin of the unmodeled dynamics is not needed. Alternative bounding signals may be used under
other assumptions for the unmodeled dynamics. For example, if 3 i i , where 3 i is the ith row of H 3 (s)(s
variable), i = 1,. . ., n, for some known constant a0 > 0, and (s+a0 ) 1 [g3 (u)])t
+ a0 ) (this s is the Laplace
1
(s+a0 ) [u])t , then we can choose m3 (t) = (s+a0 ) 1 [u])t and set 0 < i a0 . Another choice of m3 (t)
is the bounding signal [ /(s+0 )][u](t) when g3 (u) u and 3 i 1 i , where 3 i is the ith row of
H 3 (s 0 )s with the Laplace variable s, for some known constant 0 > 0. For this m3 (t), the condition on i is
< 0 and i > 0 for i = 1,. . .,n. We note that a similar L1 -norm condition can be established for
(1 ,. . .,m )T
H 1 and H 2 to design a robust adaptive controller.
When is known and H 1 [g1 (q)](t), H 2 [g2 (q)](t), H 3 [g3 (u)](t) are present, Eq. (27) with (t) = becomes a
robust nonadaptive controller, which results in e(t) converging exponentially to a residual set whose size is of
the order of 1 2 /k10 2 + 2 2 /k20 2 + .
Next, we present the robust adaptive control design assuming that the upper bounds on the gains of the
unmodeled dynamics H 1 , H 2 are known:
(1) (A3a)(g1 q ), g2 (q) are the same as in (A3), and H j i : ji for some known constants ji > 0,where H j i is
the ith row of H j , j = 1, 2, i = 1,. . ., n.
We propose to choose j (t) = (j1 (t),. . .,jn (t))T , j = 1,2,3, in Eq. (27) as
POSITION CONTROL
13
This scheme guarantees that all signals in the closed-loop system are bounded for 0 i 3i , i = 1,. . .,
n, and the tracking error e(t) satisfies
for some constants 0 > 0, 0 > 0, and any t2 > t1 0. Moreover, e(t) L2 and limt e(t) = 0 in the absence of
the unmodeled dynamics H 1 , H 2 , and H 3 .
We see from Eq. (43) that the mean tracking error explicitly depends only on the design parameter , not
on the bounds ji defined in (A3a). A smaller may result in a smaller mean tracking error (2.43). Hence, with
the knowledge of the unmodeled dynamics bounds, improvements of the tracking performance may be achieved
by using the control signals defined by Eq. (41).
Another interesting result is the adaptive controller with a so-called variable structure [Utkin (13)]:
letting ji 0 in (2.41), j = 1,2,3, i = 1,. . .,n, we obtain
It can be shown that for 0 i 3i , i = 1,. . .,n, all closed-loop signals are bounded and the tracking error
asymptotically converges to e(t) = 0 with possible chatterings. For a variable structure controller, (t) = 0 can
be used in the update law (32).
As a final remark, we note that the proposed designs can be combined to solve the problem in which both
the parameter variation and unmodeled dynamics are present.
Proportional Derivative Control, Inverse Dynamics, and Path Interpolation. In this subsection,
we discuss some of the general concepts related to the position control of robot manipulators.
Proportional Derivative Control. We first derive a PD control law for each joint of a manipulator based
on a single-input single-output (SISO) model. Coupling effects among the joints are regarded as disturbances.
Permanent-magnet dc motors along with gear reduction are commonly used in practice to actuate the joints of
the manipulator. For such dc-motor-actuated robotic manipulator, a simplified version of the dynamics of the
kth joint can be given as in Spong and Vidyasagar (8),
where J eff k = J mk + rk 2 dkk (q) is the effective joint inertia of the kth actuator (motor plus gear, J mk ) and the
manipulator link [dkk (q) is the kth diagonal element of D(q) in Eq. (3)], Beff k = Bmk + (K bk K mk /Rk ) is the
effective damping of the kth actuator (motor plus gear, Bmk ) with K bk the back emf constant, K mk the torque
constant, and Rk the armature resistance; mk is the kth motor (rotor) angular position; V ak is the armature
voltage of the kth motor; rk is the kth gear ratio; k = K mk /Rk ; and k is the actuator dynamics (3) specified
for the kth joint and is treated as a disturbance to simplify the problem, since in that case, we maintain the
14
POSITION CONTROL
where cijk = (dkj /qi + dki /qj dij /qk ) and k (q) = P(q)/qk with P(q) the potential energy.
The setpoint tracking problem is defined as the problem of tracking a constant step reference command
d = [d1 ,. . .,dn ] for n joints. This type of control is adequate for applications not involving very fast motion,
especially in robots with large gear reduction between the actuators and the links. A PD compensator for each
of the n joints can now be used to achieve setpoint tracking:
where K pk , K dk are the proportional and the derivative gains, respectively. The characteristic polynomial of the
closed-loop system is
indicating that the closed-loop system will be stable for all positive values of K pk and K dk and bounded
disturbances. The tracking error is given by
For a step reference input dk (s) = dk /s and a constant disturbance k (s) = k /s, the steady-state error is ekss
= rk k / k K pk [see Spong and Vidyasagar (8)]. Thus, the steady-state error due to a constant disturbance
is smaller for larger gear reduction and can be made arbitrarily small by making the position gain K pk large.
By using integral control as well (PID), we can achieve zero steady-state error while keeping gains small and
rejecting step disturbances. However, the PD or PID compensators perform poorly for position control when
there are large uncertainties in system parameters, or when varying disturbances and unmodeled dynamics
are present, as is common in applications. In such situations, the adaptive control designs presented in the
preceding two subsections perform much better than the PD or PID controller.
In the PD compensator presented above, the coupling effects among the joints were regarded as disturbances. In reality, the dynamic equations of a robot manipulator form a complex, nonlinear, and multivariable
system. Such a dc-motor-driven n-joint actuator may be represented in a matrix equation as
where D(q) is the time-invariant n n inertia matrix, C(q, q ) and (q) are the time-invariant versions of C(q,
q , t) and (q, t) in Eq. (3) respectively,
has components uk = (K mk /rk Rk ) V ak , and B has elements k = Bmk + K bk Lmk /Rk (with Lmk the inductance),
for k = 1, . . ., n. An independent joint PD control scheme can be written for the system (50) as in Spong and
POSITION CONTROL
15
Vidyasagar (8):
where = qd q is the error between the desired and the actual joint displacements, and P , D are diagonal
matrices of positive proportional and derivative gains, respectively. In the absence of gravity [(q) = 0], the
PD control law (51) achieves asymptotic tracking of the desired joint positions. In presence of gravity, Eq. (51)
alone cannot guarantee asymptotic tracking and has to be modified as
to cancel the steady-state error due to the effect of the gravitational terms [see Spong and Vidyasagar (8)]. For
detailed analysis and performance study of PD controller for robot manipulators, please refer to Spong and
Vidyasagar (8).
Inverse Dynamics. Using inverse dynamics, a more complex nonlinear control technique can be implemented for trajectory tracking of rigid manipulators [Spong and Vidyasagar (8)]. Consider the system given by
Eq. (50) in a more simplified form,
reduces the system q = V a with V a as the new input to the system, the armature voltages to be applied to to the
actuator motors. Thus, we have a double integrator system with n uncoupled double integrators. The nonlinear
control law (54) is called the inverse dynamics control and achieves a new linear and decoupled system, making
it possible to design V ak to control a simple linear second-order system and can be designed as
where 0 , 1 are diagonal matrices of position and velocity gains, respectively, and is the reference. The gains
could be chosen to get a joint response that is equal to the response of a critically damped linear second-order
system with desired natural frequencies for each of the desired speeds of the responses of the joints. The inverse
dynamics can be viewed as an input transformation that transforms the problem from one of choosing torque
input commands, which is difficult, to one of choosing acceleration input commands, which is easy. There are
many crucial issues of implementation and robustness that must be addressed to implement Eq. (54), and the
reader is referred to Spong and Vidyasagar (8).
Path Interpolation. The simplest type of robot motion is point-to-point motion. In this approach the robot
is commanded to go from an initial configuration to a final configuration without regard to the intermediate
path followed by the end effector. To understand the concept of configuration, it is helpful to review some
terminology used in Spong and Vidyasagar (8). Suppose a robot has n + 1 links numbered from 0 to n starting
16
POSITION CONTROL
Fig. 4.
the base of the robot, which is taken as link 0. The joints are numbered 1 to n, and the ith joint is the point
in space where links i 1 and i are connected. The ith joint variable is denoted by qi . A coordinate frame is
attached rigidly to each link. We attach an inertial frame to the base and call it frame 0. Frames 1 to n are
chosen such that frame i is rigidly attached to link i. Now the configuration is given by the transformation
matrix that transforms the coordinates of a point from frame j to frame i and is denoted by T i j . For example, for
a seven-link robot manipulator, the initial and final configurations that are of interest in point-to-point motion
are the transformation matrices that transform the coordinates of frame 6 to frame 0; let them be denoted
by T 0 6 init and T 0 6 final. This type of motion is suitable for materials transfer jobs where the workspace is
clear of obstacles. Given the desired initial and final positions and orientation of the end effector, the inverse
kinematic solution is evaluated to find the required initial and final joint variables. Suppose, di j denotes the
position of frame j with respect to frame i, and Ri j denotes the orientation of frame j relative to frame i. For the
manipulator with seven links, the motion of the first three, joints is calculated by computing the joint variables
q1 , q2 , and q3 corresponding to d0 3 init and d0 3 final. The motion of the final three joint variables is found by
computing a set of Euler angles corresponding to R3 6 init and R3 6 final [Spong and Vidyasagar (8)]. For some
purposes, such as obstacle avoidance, the path of the end effector can be further constrained by the addition of
via points intermediate to the initial and the final configurations as shown in Fig. 4. Different techniques of
generating smooth trajectories in joint space, given the initial and final joint variables, are presented in Spong
and Vidyasagar (8).
POSITION CONTROL
Fig. 5.
17
In this article, we present new transparency concepts suitable for adaptive control of teleoperation systems
with time-varying parameters [Shi et al. (17)]. Adaptive control schemes for teleoperation systems with jumping
or rapidly time-varying parameters are developed [Shi et al. (17)]. The developed adaptive control schemes
lead to stable and transparent teleoperations in the presence of unknown constant or jumping or fast-varying
parameters. The teleoperation systems to be controlled are assumed to have no communication time delay. In
the first subsection, we present the new concepts of weak transparency, asymptotic weak transparency, and
approximate weak transparency, and formulate the transparency control problem for four types of teleoperation
systems with no communication time delay [Shi et al. (17)]. In the next subsection, we present adaptive
control schemes for teleoperation systems with unknown jumping parameters and with parametrizable and
unparametrizable smoothly time-varying parameters [Shi et al. (17)]. In the last subsection we present some
control issues relevant to teleoperation systems with communication time delays.
Teleoperation Systems. In this section, we present the general structure of a teleoperation system
and its dynamic description, introduce several new concepts for transparency of teleoperation systems, and
state the adaptive control objective with which the new transparency concepts are to be verified.
Dynamics of a Teleoperation System. A teleoperation system consists of five subsystems: the human
operator, the master robot, the communication channels, the slave robot, and the slave environment, as shown
in Fig. 5. The term teleoperator refers to the master and slave manipulators connected by the communication
channels. Bilateral teleoperation involves velocity and force information transfer between the master and the
slave. Communication time delays commonly exist in teleoperation systems due to the large distance and
restrictive data transfer. These delays are assumed to be absent in the following analysis, for confinement to
fundamentals and for simplicity of analysis. In Fig. 5, vh is the velocity of the human operators hand, vm is the
velocity of the master end effector, vs is the velocity of slave end effector during contact, F h is the force applied
by the human operator to the master robot, F e is the force exerted by the slave robot on its environment, and
F s is the coordinating torque. In the absence of communication time delay, vsd (t) = vm (t) and F md = F s (t). In the
presence of communication time delay T, vsd (t) = vm (t T) and F md = F s (t T). In the following analysis, no
communication time delay is assumed.
For analysis, a network representation of a teleoperation system is useful and Fig. 6 shows one commonly
used in the literature, in which the human operator and the slave environment are represented by one-port
networks, and the teleoperator by a two-port network. The blocks Zh , Zm , Zs , and Ze represent respectively
the dynamics of a human operator, a master robot, a slave robot, and the slave environment; signals m and s
denote control torques for master and slave robots; signals vh and ve refer to velocities of the human operators
hand and the slave environment. Note that vh equals vm , the velocity of master end effector, and ve equals
vs , the velocity of slave end effector during contact. The signals F h , F h , F e represent respectively the force
generated by the human operator, the force applied by the human operator to the master robot, and the force
exerted by the slave robot on its environment.
As in Hashtrudi-Zaad and Salcudean (15), Lee and Chung (16), and Raju et al. (18), we consider the
dynamics of the master and slave robots as
18
POSITION CONTROL
Fig. 6.
where M, B, and K are inertia, damping, and stiffness parameters; the signal x is the position of end effector;
and the signal denotes the control torque with subscript m for the master and s for the slave. From Eq. (56),
we see that Zm (s) = M m s + Bm + K m /s.
Let Cm and Cs denote the master and slave feedback control, and Ci , i = 1, . . ., 4, represent the data
communication control for signals vm , F e , F h , and vs , respectively. Then the torques m and s have the following
descriptions:
where the minus sign indicates the feedback signals. We also assume the human operator and the slave
environment are passive, and as in Raju et al. (18), we use a generalized massdampingspring model to
describe the human operator and the slave environment,
where M, B, and K are the inertia, damping, and stiffness parameters with subscript h for the human operator
and e for the slave environment. Substituting Eq. (61) into Eq. (57), we get the slave system
where M = M s + M e , B = Bs + Be , and K = K s + K e .
Four types of teleoperation systems are usually met in applications: teleoperation systems (i) with known
time-invariant dynamics, (ii) with unknown constant environment, (iii) with jumping environment parameters,
and (iv) with smooth time-varying environment parameters. The transparency of adaptive teleoperation control
POSITION CONTROL
19
systems is of main interest in this article, for which we will introduce new concepts suitable for adaptive control
when the system parameters are unknown, for different cases of parameter uncertainties.
We first consider teleoperation systems with no communication time delay. The stability analysis is easy
and simple when communication delay is not involved. A closed-loop transfer function can be obtained for the
bilateral system, and the traditional tools such as the root locus technique and the RouthHurwitz stability
criterion can be used for stability analysis. It is reasonable to assume no communication delay to develop the
basic adaptive control techniques along with the stability analysis. In the last subsection below, we present
some of the control issues relevant to teleoperation with communication time delays.
Transparency of a Teleoperation System. The impedance transmitted to or felt by human operator,
Zt (see Fig. 6), is defined by F h = Zt vh , in the frequency domain.
Definition 1. [Lawrence (14)]. A teleoperation system is transparent if
This means that in a transparent teleoperation, the human operator feels as if he were manipulating the
slave environment directly. Note that when the slave robot is in contact with its environment, its velocity, vs
and the environment force F e are related by the impedance Ze as F e = Zs vs in the frequency domain. Since
vh = vm , if the slave exactly reproduces the motion of the master (i.e., vs = vm , and Zt = Ze , then F h = F e ,
that is, the master accurately feels the slave contact force. That is, for a transparent teleoperation, the velocity
tracking from the slave to the master leads to force tracking from the master to the slave.
Definition 2. A teleoperation system is weakly transparent if
The property (64) is called weak transparency because it only needs Zt = Ze for some specific operation
frequencies at which vs = vm .
Definition 3. A teleoperation system is asymptotic weakly transparent if
This weak transparency is ensured in adaptive teleoperation control systems with parametric uncertainties.
Definition 4. A teleoperation system is approximate weakly transparent if
for some constant c1 > 0, c2 > 0, > 0 and some design parameter > 0.
In this case, it is expected that the design parameter > 0 in the control system can be chosen to be large
so that the tracking error vs (t) vm (t) can be made small.
20
POSITION CONTROL
Control Designs. In this subsection, we will first review two existing control designs: one for teleoperation systems with known constant parameters, and one for teleoperation systems with unknown constant
parameters. We will then present new adaptive control schemes for time-varying teleoperation systems with
jumping parameters or with smoothly time-varying parameters. The teleoperation systems in consideration
are assumed to have no communication time delay. For the new proposed control schemes, we will analyze the
system performance in terms of stability and transparency. The teleoperation system is said to be stable if the
state variables of the system are bounded at any time.
Design for System with Known Constant Parameters. As in Lawrence (14), the forces and velocities of
the teleoperator two-port as shown in Figure 6 are related by a hybrid matrix H:
where H ij , i, j = 1, 2, representing the inputoutput relation of the teleoperator two-port, are H 11 = (Zm + Cm )
D(Zs + Cs C3 C4 ) + C4 , H 12 = (Zm + Cm ) D(I C3 C2 ) C2 , H 21 = D(Zs + Cs C3 C4 ), H 22 = (I C3 C2 )
with D = (C1 + C3 Zm + C3 Cm ) 1 . Solving for F h , vh from Eq. (68), we get
POSITION CONTROL
21
useful for adaptive teleoperation systems. To design an adaptive control law, we assume zero communication
time delay and
m [Lee and Chung (16)].
(1) The master position signal xm is bounded with bounded derivatives x m , 022
Design for System with Unknown Constant Parameters. In this section, the slave environment is assumed to have unknown constant parameters. Our control objective, as specified in the sub-subsection Transparency of a Teleoperation System above, is to achieve (i) closed-loop signal boundedness; (ii) limt [vs (t)
vm (t)] = 0, and (iii) F h = F e . The last two properties imply asymptotic weak transparency of the teleoperation
system. A control scheme based on Slotine and Lis (1) design is first applied to achieve signal boundedness
and velocity tracking. Force matching is then designed by using the relationship of (3.13) with H 11 = 0, H 12 =
1.
Adaptive Control Design. The slave system as defined in Eq. (62) is
(Ms2 +
), where
is
. Then it follows that = (K D + B)s 0, which implies that all the closed-loop signals are bounded,
and that e, e converge to zero asymptotically as the time t goes to [Slotine and Li (1)].
Transparency and Stability. With velocity tracking from the slave to the master, the force tracking from
the master to the slave will lead to a weak transparent teleoperation. Because F h is related to F e as F h = H 11
ve + H 12 (F e ), the condition for F h = F e is
2
Thus far asymptotic velocity tracking and force tracking are ensured, which lead to asymptotic weak
transparency for a teleoperation system with unknown constant parameters. The developed teleoperation
system is stable because the master, the slave, and their controllers are passive, and the human operator and
the slave environment are also passive by assumption.
22
POSITION CONTROL
Design for System with Unknown Jumping Parameters. Parameter variations in this case are characterized by piecewise constant behavior. An example is the slave robot intermittently contacting with different
environments. Assume that parameters B and K in the slave system (69) are unknown and piecewise constant,
as modeled by
where Bi and K i , i = 1,. . ., l, are constants, which last for certain durations of time, and f i (t), i = 1,. . ., l, are
functions indicating which values of Bi and K i are taken by B(t) and K(t) at a given time. The indicator functions,
f i (t), are known functions that reveal the durations and the time instants of the parameter discontinuities, and
are defined as
The function f i (t) = 1 indicates the value Bi or K i is taken by the system parameter B(t) or K(t). In addition,
f i (t) indicates that only one value can be active at any time, that is, f i (t)f j (t) = 0 for i = j. With (t) = [M B(t)
= (t), so we cannot use (t) in Eq. (70) to ensure (t) 0 for any t
K(t)]T , it follows that (t) = (t) (t)
> 0.
In this sub-subsection, we first present an adaptive control scheme for the slave system (69) with unknown
jumping parameters described by (73, 74, 75).
Adaptive Control Design. We propose the following controller structure:
where v(t), s(t), Y( v , v, xs ), and K D are defined in the preceding sub-subsection, and
=[
T
estimate of = [M B(t) K(t)] , with (t) = 1 f 1 (t) + 2 f 2 (t) + . . . + l f l (t) and (t) = 1 f 1 (t) +
. . . + l f l (t). Substituting (3.21) into (3.14) reveals
M,
i
Bi , and
]T is the
2 f 2 (t) +
POSITION CONTROL
where m , b i, k i > 0, i = 1,. . ., l. We choose the adaptive laws for
i,
23
and K i as
(t), becomes
where m(t) > 0, B(t) > 0, K(t) > 0 represent the time-varying mass, damping, and spring parameters. This
model follows from the EulerLagrange equation [Spong and Vidyasagar (8)] with kinetic energy K = x s M(t)
x s . Transparent teleoperation designs for known and unknown time-varying parameters are considered in this
section. To achieve weak transparency the key is velocity tracking and force tracking between slave and master
robots.
24
POSITION CONTROL
Control Design for Known Time-Varying Parameters. A control scheme that ensures asymptotically weak
transparency is proposed first for the teleoperation system with known time-varying slave system. This scheme
is then extended to the time-varying slave system with bounded parameter disturbances.
Design I for Known Time-Varying Parameters.Design I for Known Time-Varying Parameters
slave system (83) with known time-varying parameters, we propose the control scheme
For the
where (t) = [M(t) B(t) K(t)]T , Y( v , v, xs ) = [ v , v, xs ], v = x m (t) e(t), s = x s (t) v, and e(t) = xs (t) xm (t), as
in the sub-subsection Design for System with Unknown Constant Parameters above, and K D > 0 is a design
gain to be specified later. Substituting the controller (84) into the slave system (83) reveals
To ensure
(t) of V(t) is
The result that (t) 0 implies that s L2 L . Since s = e + e, we conclude that e and e L2 L . Hence
xs , x s L . From Eqs. (84) and (85) we have s , s L , and therefore e L . Applying Lemma 1, we conclude
that the tracking errors e(t) and e (t) = vs (t) vm (t) go to zero as t goes to .
In summary, we have the following results:
Theorem 4. All signals in the closed-loop system with the time-varying model (83) and the controller (84)
where K D satisfies (3.32) are bounded, and the tracking errors e(t) and e (t) go to zero as t goes to .
Design II for Time-Varying Parameters with Unknown Disturbances.Design II for Time-Varying Parameters with Unknown Disturbances In this case, the system parameters (t) and M(t) satisfy the assumptions:
(1) The time-varying parameter vector (t) satisfies
POSITION CONTROL
25
for some known parameter 0 (t) R3 and some unknown but bounded disturbance (t) R3 such that
(t) < 1 for some constant 1 > 0.
(2) The time-varying parameter (t) satisfies
0 (t),
and some unknown but bounded disturbance M (t) such that |M (t)| < 2
where K D > 0 is the design gain. We choose the positive function V(t) as defined in Eq. (86). Choose K D >
+
M(t)s2 , we have
which implies
and
where k1 > 0, k2 > 0, and 1 > 0 are constants, k0 is a design parameter that can be chosen to be large, and so
is K D > 0. Since s(t) = e (t) + e(t) where > 0 is a constant, we have
26
POSITION CONTROL
where c1 > 0, c2 > 0, d1 > 0, and d2 > 0 are some constants, and 1 =
be chosen to large so that the errors in Eqs. (93) and (94) are small.
In summary, we have the following results.
Theorem 5. All signals in the time-varying system (83) with parameter disturbances (A2), (A3) and controller
(90) are bounded, and the tracking errors e(t), e (t) satisfy Eqs. (93) and (94), respectively. Moreover, e(t) L2 ,
e (t) L2 , and limt e(t) = 0, limt e (t) = 0 in the absence of parameter disturbances, that is, when = 0,
M = 0.
Adaptive Control for Unknown Time-Varying Parameters. Transparent teleoperations are designed for
two types of slave systems: those with unknown smooth time-varying (parametrizable) parameters, and those
with unknown and disturbed time-varying (unparametrizable) parameters. An adaptive control scheme is
proposed for the first type of system to achieve asymptotic weak transparency. With modification, this scheme
ensures approximate weak transparency in the mean for the second type of system.
Design I for Parametrizable Parameter Variations.Design I for Parametrizable Parameter Variations
present an adaptive control design for systems satisfying the following assumptions:
We
for some known function Y 0 (t) R3r and some unknown but constant parameter 0 Rr , for some r
1 [under this assumption, Y( v , v, xs )(t) = Y( v , v, xs )Y 0 (t)0 , so that Slotine and Lis (1) design in the
sub-subsection Design for System with Unknown Jumping Parameters above can be applied].
(2) The time-varying term
for some known function Z(xs , x s , xm , x m , t) R1r and some unknown but constant parameter 0 Rr, for
some r 1.
We propose the adaptive controller structure
POSITION CONTROL
Define the parameter errors
0 ,
where = T > 0 and = T > 0 are constant matrices of the appropriate dimensions. To ensure that
0, we choose the adaptive laws for
and
27
as
With this choice of 0 and 0 , we have = K D s2 0, which implies that s L2 L and 0 , 0 L . Since
s = e + e, we conclude that e, e L2 L . Hence xs , x s L . From Eq. (97) it follows that s , s L ; therefore,
e L . Applying Lemma 3.1, we conclude that the tracking errors e(t) and e (t) = vs (t) vm (t) go to zero as t
goes to .
In summary, we have the following results.
Theorem 6. The adaptive controller (97) with the adaptation law (100) and (101) applied to the time-varying
system (83) guarantees that all closed-loop signals are bounded and the tracking error e(t) and e (t) go to zero
as t goes to .
Design II for Unparametrizable Parameter Variations.Design II for Unparametrizable Parameter Variations We assume the unparametric parameters having a parametric part and bounded disturbance part.
They satisfy the modified assumptions:
(1) The parameter (t) satisfies
where Y 0 (t) and 0 are the same as that defined in (A4), such that 0 < M 1 for some constant M 1 0, and
(t) < 1 for some constants 1 > 0.
(2) The term
(t)( x s + v) satisfies
where Z(xs , x s , xm , x m , t) and 0 are the same as that defined in (A5) such that |0 | < M 2 for some constant
M 2 > 0, and |(t)| < Y 1 (t)2 for some constant 2 > 0 and some known function Y 1 (t). Remark: One choice
of Y 1 (t) is Y 1 (t) = x s (t) + v(t).
28
POSITION CONTROL
We propose the controller structure as
0 and
T
T
and 0 T 0 go unbounded if 0 (t) and 0 (t) go unbounded [see Ioannou and
0
0 0, that
0
Sun (20)], and that and 2 are finite [see assumptions (A4 ) and (A5 )], we have that V(t) is bounded, and
for some constants 0 , 0 > 0, and any t2 > t1 0. Because of the relation s(t) = e (t) + e(t) and > 0 is
constant, we can obtain
POSITION CONTROL
29
and
where 1 , 1 , 2 , and 2 are some positive constants [see Tao (7)]. In this case, the design parameter K D > 0
can be chosen to be large so that the mean errors are small in Eqs. (111) and (112).
In summary, we have the following results.
Theorem 7. All signals in the closed-loop system (83) with unparametric time-varying parameters, and adaptive control law (104) and adaptive law (105, 106, 107, 108, 109, 110), are bounded, and the tracking errors
e(t), e (t) satisfy Eqs. (111) and (112), respectively. Moreover, e(t) L2 , e (t) L2 and limt e(t) = 0 limt e (t)
= 0 in the absence of parameter disturbances, that is, when = 0, = 0.
Transparency and Stability. For teleoperation systems with known time-varying parameters or parametric time-varying parameters, the adaptive control schemes (84) and (97) ensure velocity tracking from the
slave to the master. Therefore the force tracking design as in Eq. (72) will lead to asymptotic weak transparency
(Definition 3). For time-varying systems with bounded disturbances, an arbitrary small tracking error can be
obtained by increasing the design gain K D . By using the force tracking design in Eq. (72), approximate weak
transparency (Definition 4) or approximate weak transparency in the mean (Definition 5) is achieved. Stability
of the resulting releoperation system is ensured by the boundedness of all the closed-loop signals.
Teleoperation with Communication Time Delay. Communication time delay in a bilateral teleoperation system reduces system stability and performance. Delay on the order of a tenth of a second were shown
to destabilize the teleoperator. The stability problem becomes difficult when a communication time delay T is
present, because a time delay introduce a factor e sT into the system and hence makes the system infinitedimensional. In bilateral teleoperation, the force reflection from the slave, introduced for providing the feeling
of the remote task, has effects on the masters motion, which generates disturbances on the desired motion. The
communication delay may worsen the situation as well. With a time delay T, the conventional communication
law results in tracking of both position and force in the steady state. However, with the delay, the system is not
passive, and will probably never reach a steady state. A preliminary result on a modified control scheme that
provides improved tracking performance for a noncontact task in the presence of time delay and for arbitrary
master trajectories has been developed in Shi et al. (21).
In the previous research, the passivity formalism and network expression were used to investigate the
stability of a bilateral teleoperation system. In these methods, the human operator input is assumed to be
bounded, and the human operator and the environment are assumed to be passive. In the presence of communication time delay, and with the passivity assumptions about the operator and the environment, passivity of
the system depends on passivity of the communication block.
Two approaches can be used to produce a passive communication block. The first was developed by
Anderson and Spong (22) using scattering transformation theory. This solution uses the transmission line
equations as a basis for deriving a passive communication control law. By applying scattering theory, it is
shown how conventional approaches lead to infinite gain of a scattering operator at finite frequencies, and how
by implementing a set of time delay equations this instability can be overcome. The resulting system is then
passive for all time delays. The proposed control law maintains steady-state force and velocity tracking.
The second approach to produce a passive communication block is developed by Niemeyer and Slotine
(23). This approach uses an energy formulation to construct a teleoperation system that imitates physical
systems and obeys an energy conservation law. A wave variable is utilized to characterize time delay systems
and leads to a new configuration for force-reflecting teleoperation.
30
POSITION CONTROL
Since the dynamic control of the remote slave by an operator is severely restricted by the time delays in
the transmission, it is important to provide consistent dynamic performance locally at the remote site in the
face of uncertainties and varying operating conditions. With the development of high-speed and high-capacity
computer networks, it is possible to deliver teleoperation over a public computer network. The problem of
varying communication time delays arises in such a teleoperation system. Adaptivity of a teleoperation system
to uncertain time delay in also desirable. The related stability, tracking, and transparency problems of a
bilateral teleoperation system under uncertain environment but now with communication time delay are
important issues to be addressed. Adaptive control solutions proposed in the subsection Control Designs
need to be modified to provide adaptation mechanisms for adjusting the controller parameters to achieve
desired system performance, despite system uncertainties due to the unknown slave environment, and now in
the presence of communication time delays.
The passivity-based solution to the bilateral teleoperator time delay problem developed by Anderson and
Spong (22) is based on the result in circuit theory that a circuit consisting of passive elements only is passive
and therefore stable. However, if some elements in a circuit representing a teleoperation system are not passive,
one cannot use passive network theory to conclude the stability of the teleoperation system. On the other hand,
if the transfer function of a teleoperation system is positive real, then the system is passive. In Shi et al. (24),
a notion of positive realness has been used to investigate the passivity of the teleoperation system proposed
by Anderson and Spong (22). Shi et al. (24) have also proposed a modified control scheme that use the master
accelaration information (with delayed operation, which can be obtained from the velocity information) for
slave control and ensures that in the absence of slave environment torque the slave position tracks that of the
master asymptotically, that is, achieves improved tracking performance for the teleoperation system.
Summary
Position control for robot manipulators and teleoperation systems involves many dimensions of control theory,
such as controller design, robustness analysis, and adaptive designs, along with many practical applications.
Robust adaptive control schemes have been presented to handle situations in which the robot system has
bounded parameter variations or/and unmodeled dynamics of bounded gains. The distinct feature of the manipulator dynamics were used to define bounding signals in the controller structure, whose parameters are
updated from a robust adaptive law to ensure signal boundedness and tracking errors of the order of parameter
variations and unmodeled dynamics, which may not be small. Some common topics relevant to position control
of robot manipulators, such as PD control, inverse dynamics control, and path or trajectory interpolation, were
also discussed.
Adaptive motion control of teleoperation systems was addressed. Several new concepts of transparency
were defined for teleoperation control systems with unknown parameters. These new transparency concepts
are useful for developing adaptive control schemes for control of a teleoperation system with unknown constant
parameters, or with unknown jumping parameters, or with unknown smooth but large time-varying parameters. Weak transparency properties have been established for such adaptive teleoperation control systems.
Some important control issues for teleoperation systems with communication time delay were also discussed.
BIBLIOGRAPHY
1. J. Slotine W. Li On the adaptive control of robot manipulators, Int. J. Robotics Res., 6 (3): 4959, 1987.
2. J. Slotine W. Li Adaptive manipulator control: A case study, IEEE Trans. Automat. Control, AC-33: 9951003, 1988.
3. J. S. Reed P. A. Ioannou Instability analysis and robust adaptive control of robotic manipulators, IEEE Trans. Robotics
Automation, 5: 381386, 1989.
POSITION CONTROL
31
4. M. W. Spong Adaptive control of flexible joint manipulators, Systems Control Lett., 13: 1521, 1989.
5. F. L. Lewis S. Jagannathan A. Yeildirek Neural Network Control of Robot Manipulators and Nonlinear Systems, London:
Taylor & Francis, 1999.
6. S. S. Ge T. H. Lee C. J. Harris Adaptive Neural Network Control of Robotic Manipulators, River Edge, NJ: World
Scientific, 1998.
7. G. Tao On robust adaptive control of robot manipulators, Automatica, 28 (4): 803807, 1992.
8. M. W. Spong M. Vidyasagar Robot Dynamics and Control, New York: Wiley, 1989.
9. R. Ortega M. Spong Adaptive motion control of rigid robots: A tutorial, Automatica, 25 (6): 877888, 1989.
10. C. A. Desoer M. Vidyasagar Feedback Systems: InputOutput Properties, New York: Academic Press, 1975.
11. M. W. Spong R. Ortega R. Kelly Comments on Adaptive manipulator control: A case study, IEEE Trans. Automat.
Control, AC-35: 761762, 1990.
12. P. A. Ioannou K. Tsakalis A robust direct adaptive controller, IEEE Trans. Automat. Control, AC-31: 10331043, 1986.
13. V. Utkin Variable structure systems with sliding modes, IEEE Trans. Automat. Control, AC-22: 212222, 1977.
14. D. A. Lawrence Stability and transparency in bilateral teleoperation, IEEE Trans. Robotics Automation, 9 (5): 624637,
1993.
15. K. Hashtrudi-Zaad S. E. Salcudean Adaptive transparent impedance reflecting teleoperation, Proc. IEEE Int. Conf. on
Robotics and Automation, Minneapolis, 1996, pp. 13691374.
16. H. K. Lee M. J. Chung Adaptive controller of a masterslave system for transparent teleoperation, J. Robotic Systems,
15 (8): 465475, 1998.
17. M.-Y. Shi et al. Adaptive control of teleoperation systems, Int. J. Robotics Res. Preliminary version, Proc. 38th IEEE
Conf. on Decision and Control, Phoenix, AZ, 1999, pp. 791796.
18. G. J. Raju G. C. Verghese T. B. Sheridan Design issues in 2-port network models of bilateral remote manipulation, Proc.
1989 IEEE Int. Conf. on Robotics and Control, 1989, pp. 13161321.
GANG TAO
AVINASH TAWARE
MINYAN SHI
University of Virginia
14
tion and deceleration schedules to provide transient limit protection. More advanced controls schedule variable engine geometry and augmentor fuel, provide fan and booster stall
protection, control variable parasitic engine flows, improve integrated engine-airframe performance, and provide engine
health monitoring and diagnostics.
It should be noted that only recently have electronic computers been used to implement engine controls. This is primarily due to the inherent need for safe operation and the
harsh temperature and vibration environment in which the
computer operates. Many engines in use today are controlled
by a hydromechanical controller commonly referred to as an
HMU. These are ingenious mechanical computers, which implement the desired control strategy in terms of cams and
mechanical integrators. Of necessity, the implemented control
strategies must be fairly simple. A drawing of a typical HMU
is shown in Fig. 1. More detailed discussions of their operation can be found in (4).
The changeover to electronic controllers began in the 1980s
as rugged integrated circuits became available and as the
need for improved performance led to increased functionality
and tighter control. Pratt and Whitney calls its controller a
Digital Engine Control (DEC), while General Electric calls it
a Full Authority Digital Electronic Control (FADEC). These
are highly customized computers, whose complexity depends
mainly on the number of sensor inputs and actuator outputs.
Such electronic controllers result in higher engine operating
efficiencies, by allowing tighter engine control through the
use of higher loop gains and improved strategies to reduce
transient overshoot or undershoot. It also allows implementation of control algorithms, which would be difficult to implement mechanically.
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
15
Forward
thrust
stop
Power
lever
stop
Reverse
thrust
stop
Forward
signal
Shut off
stop
Atmos
Ratio
lever
Servo press.
Bypass to pump
interstage
Regulated press.
Water
drain
Feed-back lever
Accel.
limiting
cam
Press
ratio
reg.
Burnerpress sensor
position adj.
Atmospheric press.
Increasing R.P.M.
Burner press.
Speedsensing
servo
Speed
servo
stop
Flapper valve
Feedback
lever
stop
Droop
lever
Speed
servo
pos. adj.
Speed cam
follower
Dec.
inc
Burnerpressure
output
Mult. lever
Balancing
nozzle
Burner-press.
limiting
valve
Water
drain
Throttle valve
position adj.
Sequencing
valve
Sequencing
valve cam
Throttle-valve
metered press.
Return to engine
pump inlet
(boost press.)
Droop
cam
Atmos
Fixed
orifice
Droop adj.
Min.
ratio
adj.
Droop cam
followers
Filter
90 stop
Max
trim
Idle
trim
Shut off
lever
Thrust
reverser
inlet
signal
Shut off
lever stop
Detent
roller
Power
lever
Atmos
Reverse
signal
Thrustreversing
actuator
Burner-press.limiting-valve adj.
Fail-safe
orifice
Burner
press
Evacuated
bellows
N2 drive
Throttle valve
Min. flow adj.
Fixed orifice
Bypass to pump
interstage
Press-reg valve
Fixed orifice
Fail-safe orifice
Fuel-temp
compensator
Flapper valve
Press-reg
servo sensor
Fuel inlet
from pump
Overboard drain
Speed-sensing governor
Coarse filter
Fine filter
To engine nozzles
16
HPC
inlet
2.5
LPT
discharge
Combustor
discharge
Afterburner
inlet
Exhaust
Exhaust
nozzle
nozzle
inlet
discharge
Exhaust
nozzle
throat
Compression System
The turbofan compression system consists of a fan, a booster,
and one or more compressors. Operating characteristics of
each are a function of their inlet total temperature and pressure, rotational speed, and discharge pressure. It is convenient to normalize flow and speed with respect to inlet temperatures and pressures. The following corrected parameters
can be used to represent operating characteristics independent of the actual inlet temperature and pressure levels:
Tb = Ta
Inlet Compression
Air flow is supplied to the engine by the inlet which compresses the inlet air. Assuming that the ambient air pressure, P0, and temperature, T0, is not moving and the inlet is
moving at the flight mach number, M, the total pressure and
temperature at the inlet is:
1/k
1
M2
P2 = r P0 1 +
2
1
M2
T2 = T0 1 +
2
(1)
(2)
where P is pressure, T is temperature, M is flight Mach number, is the specific heat ratio of air (constant pressure specific heat/constant volume specific heat), k is the ratio (
1)/, and r is the ram recovery (actual total pressure/ideal
total pressure) which is approximately 1 for subsonic flight.
Combustion
Compressor
Nozzle
2
Inlet
(3)
(4)
4
Turbine
3
Turbine
(P /Pa )k 1
1+ b
c
HP = Wa c p (Tb Ta )
Combustion
4
Pressure
Pressure
17
Compressor
2
Inlet
Nozzle
Volume
Entropy
18
Surge line
@ Sea level
@ 35,000 ft
@ 50,000 ft
Acceleration line
Deceleration line
Pt3
Pt2
Low
Low
T2
c pa + f cpf
1+ f
(6)
2000
Effect of varying T3
1800
1600
1400
1200
ou
(5)
e
tr
gio
T3
ow
Bl
Combustion System
Turbine Expansion
W4 c p (T4 T3 )
Q f f cpf T4
Corrected shaft
speed
High
Wa T2
Wf =
Zone of poor
compressor
efficiency
N/ T2 Low
Steady-state
operating line
@ Sea level
@ 35,000 ft
@ 50,000 ft
N/ T2 High
Pressure ratio
Surge zone
1000
800
Operating area
600
400
200
Blowout region
0
10
20 30 40 50 60 70
Combustion total pressure
inches (HG ABS.)
80
A8 =
8 N8 P8
8 = g/R
3
1.2
N8 = M8
1 + .2M82
(9)
(10)
(11)
The throat Mach number must be 1 if the nozzle total-tostatic pressure ratio is greater than the critical value of about
1.8. If the pressure ratio is less than critical, the throat Mach
number will be less than 1, and can be obtained from the
relationship:
M=
(8)
Wg8 T8
Bypass Duct
A fraction of the fan discharge air is bypassed around the
engine core, and exhausted through either a separate bypass
nozzle, or mixed with the core stream and exhausted through
the core nozzle. In either case, the bypassed air improves the
propulsive efficiency of the engine, and makes it the preferred
approach for the worlds large commercial fleet. The bypass
ratio is defined as:
19
2
[(PT /PO )k 1]
1
(12)
The exit area will effect the thrust output of the engine. Similar equations can be used for the bypass or mixed flow
streams.
Augmentor
Exhaust System
The core and bypass streams are expanded through core and
bypass nozzles to the pressure of ambient air. A converging
.28
5.0
Nozzles choke
.26
4.0
.24
.85
.22
T = .84
.20
T4 T5
T4
.18
122
3.0
120
.16
T4
W
.14
P4
= 115
.12
110
.10
60
70
80
90
100
P4/P5 = 2.0
110
120
N/ T4
130
140
Engine Trade-Offs
Engine operating characteristics are set predominantly by
four interacting design variables: (1) bypass ratio, (2) turbine
inlet temperature, (3) overall pressure ratio, and (4) fan pressure ratio. The best design choice will be dependent on the
20
intended application or mission, the level of technology available, the degree of subsequent growth capability required,
and expected competition from other engines.
Bypass ratio will have the most dominant effect on engine
performance. High bypass ratios of 4 to 8 are used for most
large commercial engines. Increased bypass ratio will improve
(decrease) specific fuel consumption (SFC) at cruise, and improve specific thrust (thrust per kilogram/second of air flow)
at takeoff. Ultra-high bypass ratios of 10 to 20 have been considered for improved cruise performance, but would require
an unducted fan (8) or reduction gears between the fan and
low-pressure turbine. Lower bypass ratios of 1 to 3 provide
improved thrust for flight Mach numbers of 1 to 2, and are
used for military fighters and bombers. A pure turbojet has a
zero bypass ratio, and would be used for supersonic transport.
High turbine inlet temperature leads to improved specific
thrust and a lighter engine, but requires more expensive turbine materials and a more complex turbine cooling system,
which reduces cruise performance. The proper balance will
depend on the relative importance of specific thrust, which
sets engine size, and weight and cruise performance, which
sets fuel requirements. Military applications will tend to demand higher temperatures, to achieve a lighter engine
weight, while commercial applications will place a stronger
emphasis on cruise performance.
Overall pressure ratio, which is the compressor discharge
pressure divided by fan inlet pressure, will affect both takeoff
and cruise performance. Optimum pressure ratio tends to increase will increased turbine inlet temperature, but decreases
as flight Mach number increases. Pressure ratios of 40 : 1 and
50 : 1 could be effective with modern temperatures at takeoff,
but should be limited to the 10 : 1 to 15 : 1 range at supersonic
speeds. Extremely high-pressure ratios would require the use
of high alloy steel or titanium at the rear of the compressor,
and cooled-cooling air for use in the high-pressure turbine.
High fan pressure ratio improves takeoff performance, but
increases exhaust noise. The achievable pressure ratio will be
dependent on the work available from the low-pressure turbine. Both will be dependent on turbine inlet temperature
and overall pressure ratio. Fan pressure ratios of 2 to 3 can
be achieved on low bypass military engines, but would be limited to the 1.6 to 2.0 regime for high bypass commercial engines. Mixed-flow turbofans would also require the use of fan
pressure ratios that produce duct pressure levels roughly
equal to the turbine discharge pressure, in order for mixing
to occur.
CONTROL REQUIREMENTS
The overall function of an engine controller is to provide
thrust in response to throttle position. It must achieve the
requested thrust with the lowest specific fuel consumption. It
must also insure that the following limits are not exceeded:
Maximum fan speed
Maximum compressor speed
Maximum turbine temperature
Fan stall
Compressor stall
Maximum compressor discharge pressure
Various structural
and overspeed
limitations
Primary
and augentor
A ignition and
blow out
l
t
i
Low
t
pressure
u
limit
d
e
Constant
T2 limit
T4 and T7
limits
separate
hard
augmentor
light-off
Minimum
burner
pressure
at low
power
Maximum
turbine
inlet
temperature
High pressure
limit
Blade loading
and burner case
limits
Mach number
Figure 8. Engine limits (19).
Fan speed
Core speed
Engine pressure ratio P5 /P2
Fuel flow divided by compressor exit pressure
Low pressure turbine inlet temperature
(13)
Wf
W4
T4
(14)
21
22
Operating
schedule
P3
Desired fan
speed
Gain
Engine
Fan speed
Wf /Ps3
Surge
area
6
Max. ratio line
Rich blowout 5
(8-9)
area
(5-6)
Speed
(78)
(2-3)
3D cam
governor
3D cam
3D
(4-5)
cam
Speed
(6-7)
(9-10)
governor Governor
3D cam
droop
lines
Speed
3
(infinite number)
governor
2
Steady state
lines
(3-4) 4 12
Speed
governor
Speed reset
(11-12)
(10-11)
11 Min. flow stop
line
10
(1-2)
Min. ratio line
Speed
1 governor
0
controller
Idle
N
Percent rotor speed
Nmax
2 psia
[13.79 kPa]
14.7 psia
9 [101.36 kPa]
100%
23
Fan speed
T2
Core speed
T2
Fan speed
controller
PS3
Deceleration
controller
Core speed
T2
Core speed
T2
Core speed
T2
Acceleration
controller
Idle
controller
Min
Max
Min
Engine
Max and min
fuel limits
Max
Operating
schedule
controller
control of acceleration, rather than speed, allows tighter control of engine acceleration, thereby improving transient response and reducing mechanical stress. While rotor acceleration cannot be easily measured directly, a second-order filter
applied to speed can be used to give a good approximation.
The algorithm shown in Fig. 12 replaces that shown in Fig. 9.
The previous acceleration schedule is replaced with one that
directly specifies allowable engine acceleration. A lead lag
compensator is necessary to improve transient response. The
Ndot control will result in a more consistent acceleration between a cold engine that has not been run for at least 30
minutes, and a warm engine that is being reaccelerated in a
shorter time.
Fan Speed Control
Throttle position is converted to a specific fan speed requirement as a function of engine inlet temperature, T2, and inlet
pressure, P2. Fan speed is held to the desired speed by replacing the proportional control with an isochronous integral control. The bandwidth of the loop will determine the speed of
response and, hence, how tightly the fan speed is maintained.
The core speed floats, within limits, at a speed necessary to
provide the sufficient power to maintain fan speed.
The agility of a helicopter (11) depends on maintaining rotor speed and, hence, power turbine speed during maneuvers
such as a waveoff from an autorotational descent. The load
on the rotor is changed by the pilot, changing either the col-
Fan speed
controller
Wf /P3 demand
Integrator
T2
Ndot
schedule
lective or cyclic settings. System responsiveness is determined by the dynamics of the power turbine isochronous governor. Unfortunately, the helicopter rotor first torsional
resonance mode limits the bandwidth of the control. Notch
filters centered at the rotor resonance are used to allow high
crossover and, therefore, a more responsive system.
A further refinement of fan speed control is called Power
Management Control. This control integrates aircraft and engine requirements, to compute the necessary thrust levels to
maintain uniform aircraft speed. The controller uses information on the specific aircraft configuration, inlet pressure, temperature, Mach number, and engine bleeds, to calculate the
desired thrust and a reference fan speed or pressure ratio. In
order to compensate for engine-to-engine quality variation
and deterioration, the pilot then moves the throttle to match
the reference fan speed or pressure ratio. Aircraft speed is
maintained, during variations in inlet pressure and temperature, by closed-loop adjustment of these reference values.
Multivariable Control
As indicated previously, commercial engines are generally
controlled by a single, closed-loop main fuel control. The additional actuators have minimal interactions between each of
the open-loop controls and with the main fuel control. Consequently, single-input, single-output design techniques are adequate.
P3
Compensator
Engine
Min
select
2nd order
rate filter
Compressor speed
24
This is, however, not the case for military engines, which
generally have an afterburner augmentor, to increase engine
thrust for selected segments of the flight, such as takeoff,
climb, acceleration to supersonic speed, or combat. The augmentor fuel flow will determine augmentor temperature, T6,
which affects engine pressure ratio and turbine temperature.
However, the primary combustor fuel control also controls
these variables. Thus the augmentor fuel control loop and the
combustor fuel control strongly interact. Early design approaches used the concept of spectral separation, which made
the combustor fuel control loop an order of magnitude faster
than the augmentor control loop. More recent designs have
used a multivariable approach, to achieve two control loops of
approximately the same response time.
Another application of multivariable control is on a variable cycle engine (VCE), which has a variable area bypass
injector, or VABI. The VABI allows the bypass ratio to be
changed during flight. When closed, the engine behaves more
like a turbojet, providing more thrust during supersonic
flight. Opening the VABI makes the engine more like a turbofan, improving specific fuel consumption during cruise.
Much of the published literature on multivariable engine
control is focused on the regulator problem of maintaining the
engine near the desired operating trajectory. It is based on
linear models valid for small signal analysis, and avoids the
important problem of limit protection. The early 19741982
applications are summarized by Zeller, Lehtinen, and Merrill
(11). The earliest work was that of McMorran and MacFarlane (12,13), which was tested on a two-spool afterburning
engine. They used fuel flow to control compressor speed and
A8 to control fan speed due to the engine bypass duct.
The International Forum on Alternatives for Linear Multivariable Control (14), held in 1977, is of particular interest,
because it showed the use of several multivariable techniques
using a linear model of the F100 engine as a theme problem.
Four authors developed control strategies based on three different multivariable techniques: multivariate transfer functions (15,16), Inverse Nyquist Array (17), and Characteristic
Root Locus (18). Each strategy used three or four of the measured variables, fan speed, N1, compressor speed, N2, compressor exit speed, P3, exhaust pressure, P7, and turbine inlet
temperature, FTIT and controlled fuel flow, Wf , exhaust nozzle areas, A8, and compressor guide vanes, CIVV, or fan guide
vanes, RCVV. In all cases, decoupled control of each of the
measured variables was achieved.
One of the most complete early multivariable designs extensively tested was that by De Hoff and Hall (1922) for
the F100 engine, using extended linear quadratic regulator
techniques. This work went beyond the previous studies, with
ability to handle large power excursions without exceeding
engine or actuator limits and to operate over the entire engine
operating envelope. A block diagram of the control system is
shown in Fig. 13. The feedback law itself is an optimal regulator structure, with integral trims for steady-state accuracy
and engine limit tracking. Linear controllers were designed
at five operating points: sea level static, subsonic and supersonic. Dominant gain elements were determined by assessing
the closed-loop eigenvalue sensitivity to each gain element.
Over 50 percent of the controller gains were eliminated in
the final implementation, with little or no effect on system
performance. Important gain elements were fitted with univariate functions of fan inlet pressure and temperature and
core speed. Unique to this control design is the transition trajectory generator, whose purpose is to smooth rapid throttle
changes by providing a piecewise linear transition from the
current engine state to the requested state. The rate of transition depended on whether the engine is at low, medium, or
high power.
Takeoff thrust for the F100 is defined as the thrust obtained at maximum allowable turbine temperature. At altitude conditions, the minimum burner pressure defines engine
idle. Thus various physical limits must be held exactly at various flight conditions, as shown in Fig. 10. In theory, an optimal controller could have been designed for each limit at appropriate operating conditions. This would have required an
exponentially large number of gain matrices to cover all combinations. De Hoff and Hall (19) used an ad hoc approach to
solve this problem, by designing single-loop spectrally separated integral trims for each input, corresponding to each desired set point and number of unsaturated controls. The control was switched whenever an actuator saturated or an
engine limit was reached.
A very similar multivariable design was developed for a
GE23 variable cycle engine (23,24). A much larger number of
linear design points were needed, due to the greater changes
in engine configuration. Instead of using integral trim to handle engine limits, a model of the engine was incorporated into
the transition logic. This allowed the generated trajectories
always to satisfy engine limits. The problem with this approach is that it is an open-loop feedforward approach. How
well the engine limits are held depends on the accuracy of the
model and how well it can handle engine-to-engine variations.
One possible solution to handling engine limits is demonstrated by Adibhatla (25,26) for a short takeoff and vertical
landing airplane (STOVL) using an F110 engine. This is a
full-range multivariable design, which was pilot evaluated in
the NASA Ames fixed-base and vertical-motion simulators.
The primary objective of this engine control is to manage
thrust, through the aft nozzle during cruise, and through ejectors at the aircraft wing roots and ventral ports on the underside of the aircraft during transition and hover. An estimate
of thrust is determined based on fan speed, fan operating line,
fuel flow, and the three thrust port areas. During cruise a
2 2 controller regulates fan speed and aft thrust, using fuel
flow and aft nozzle area. During transition and hover, the
controller is expanded to a 4 4 controller, in order to additionally regulate the ejector and ventral thrust using the ejector and ventral area actuators. To prevent engine stall and
overtemperature, and to ensure sufficient pressure for customer bleed, three single-input, single-output regulators were
designed as limit regulators.
Hanus technique (27) for integrator antiwindup is used to
transition between these five regulators and to handle actuator limits. In this technique, an observer-based structure is
used, such that the normal controller is modified to assume
the form:
xc = Ac xc + Bc (ys p y) + L(sat(u) u)
u = Cc xc + Dc (ys p y)
(15)
where Ac Bc Cc Dc the state space description of the five controller dynamics with the desired set point ysp and the engine
output y, sat(u) is the actual bounded engine input and L is
the anti-windup gain. When a given regulator is controlling
25
Nominal input
Reference
point
schedules
Transition
control
+
LQR
State
Engine limit
protection
Actuators
Engine
CP
Integral
control
CI
Gain
schedules
Limit flags
CP
Sensors
and FITT
estimator
Outputs
Fuel flow
Inlet guide vanes
A8 nozzle areas
Fan speed
Compressor speed
Bypass duct Mach number
The 2 2 controller used fuel flow and nozzle area to control fan speed and bypass duct Mach number. Both the 2 2
and the 3 3 controllers were run simultaneously, with authority switched between them as the guide vanes either approached or moved off their limit.
During sea level static tests, it was found that this fast
response multivariable controller produced a 60 percent re-
26
HAROLD BROWN
Gray Fox Technologies
INPUT-OUTPUT STABILITY
The earliest mathematical studies of control systems focused solely on the input-output stability of systems
as described in the works of Black (1), Bode (2), and Nyquist (3). In fact, most of the classical control work
was limited to the input-output study of single-inputsingle-output linear and mostly time-invariant systems.
Notions of input-output stability for nonlinear systems were later advanced by Sandberg (4), Zames, (5,6)
Safonov (7), and others. The first book dedicated completely to the subject was by Willems (8) in 1971, followed
shortly by that of Desoer and Vidyasagar in 1975 (9). With the popularity of state-space methods, Lyapunovstability concepts became the preferred analysis and design tools for nonlinear systems until the 1980s when
researchers became interested again in the input-output behavior of systems. The relationships between the
input-output and Lyapunov-stability concepts were developed in Refs. 10 and 11 and culminated in the various
versions of the Kalman-Yakubovich-Popov (KYP) lemma (12). The current studies in input-output systems are
highly dynamic with the introduction of new concepts such as input-to-state stability (13,14,15), the interaction
with geometric nonlinear control (16), applications to robust control (16,17,18), research in the adaptive control
for linear (19) and nonlinear systems (17), the interest in various mechanical and electric systems (20), and
the publication of various theoretical and applied books (18,20,21). It is now clear that the two points of view
(state-space and input-output) are complementary and that many deep relationships between the various
stability concepts are yet to be explored.
In this article we concentrate our discussion on continuous-time systems and survey the classical as well
as the more recent results in the input-output approach. The major results included in this article are taken
mainly from Refs. 9,18,22,23,24, and the reader is referred to those books for most proofs. An excellent 1995
chapter on input-output stability from a distance-separation point of view appeared in Ref. 25. Related concepts
for discrete-time systems were presented in Ref. 9 and recently revived by Byrnes and co-workers (26,27), but
will not be discussed here. In addition, while we mention absolute stability as an important application area of
input-output concepts, we refer the reader to the article on Absolute Stability in this encyclopedia for details.
The article starts with a collection of basic definitions followed by the general results on the basic concepts
of input-output stability and results for testing input-output stability and its relationship with Lyapunov stability. Next, we discuss the stability of interconnected systems and present the small-gain and passivity results.
Related concepts such as absolute stability, dissipativity, and input-to-state and input-to-output stability are
then reviewed, followed by our conclusions. Various technical definitions are presented in the appendices. We
have attempted to include the main references on the subject of input-output stability, striving to be current
and relevant rather than encyclopedic.
Basic Concepts
The general ideas of input-output stability involve the relative size of signals as they are processed by
dynamical systems. We will thus begin by providing mathematical measures of the size of signals and of the
effect that a particular system has on that size.
1
INPUT-OUTPUT STABILITY
In order to introduce the mathematical notions of input-output stability, we need some preliminary
definitions of signal spaces. (A detailed treatment of measure, signal spaces, and signal norms is beyond the
scope of this article. The reader is referred, for example, to Ref. 28.) The Lm p set, with p [1, +), consists of
all functions u: [0, +) m that are Lebesgue measurable [i.e., functions that are the limits (except for a set
of measure zero) of a sequence of piecewise constant functions], such that
Remark 1. For the finite-dimensional set m all q-norms are equivalent, in the sense that for all k, h [1,
+] there exist positive constants c1 and c2 such that
for all x
[0, ).
. This is why, when defining the Lp m set, we do not need to specify q, as u(t)
The Lm set consists of all the functions u: [0, +) m that are measurable and essentially bounded
on [0, +) (where essentially bounded means that the function is bounded except on a set of measure zero).
For a function belonging to Lm p , p [1, +), we introduce the norm
Note that at the left-hand side of definitions (2) and (3), we should specify the dependence of the norm on q,
but we omit it to avoid cumbersome notation; the choice of q will be clear from the context.
For the Lm 2 set, the norm is usually defined by setting q = 2 in Eq. (2), obtaining
The Lm 2 set with the norm (4) is of particular interest, since in some contexts the norm (4) is proportional
to the energy of the signal u, as is the case, for instance, for which u(t) is the voltage across a resistance R.
Then the total energy delivered to the resistance would be found by integrating the instantaneous power
u(t)2 /R. As usual, in the following we shall denote the L1 p set, p [1, +), by Lp .
INPUT-OUTPUT STABILITY
Remark 2. In contrast with finite-dimensional spaces (see Remark 1), for the infinite-dimensional set Lm (the
set of all measurable functions u: [0, ) m ) the p-norms may be nonequivalent. For instance,
u(t) = 1/(1+t) with 0 < < 1 belongs to L and to Lp with p < 1/, but does not belong to Lp for p 1/.
u(t) = sin(t) belongs to L but does not belong to Lp for any p [1, +).
In this article, we are interested in studying the stability properties of a generic operator G that maps a
signal space into another signal space ,
We shall define the operator G in Eq. (5) as being Lp -stable if for any u() Lm p the output y() belongs to
L p . To give a formal definition of Lp stability and to provide a broad definition of unstable systems, we first
need to define precisely the domain and the image of the operator G, which in general are not limited to Lm p
and Lr p , respectively (for example, we may be interested in considering unbounded inputs and/or outputs). For
this reason, we need to introduce some extended Ln p spaces. First, let us define the truncation uT (t) of a signal
u(t) in the following way:
r
The extended set Lm pe , with p [1, +], consists of all the functions u: [0, +)
such that
Introducing this extended space, we can treat unbounded signals. For instance, u(t) = t does not belong to Lp
for any p [1, +] but its truncation
belongs to Lp , with p [1, +], for every finite T. Therefore u(t) = t belongs to Lpe for every p
[1, + ].
Finally we end the section with the definition of causality of an operator G. When we later specify an
operator in terms of state-space differential equations, causality is an intrinsic property of the operator: the
output y at time t depends on the initial conditions and on the values of the input u up to time t. On the other
hand, when dealing with a generic operator as in Eq. (5), we need to enforce causality. An operator G: Lm pe
Lr pe is said to be causal if the value of the output at time t depends only on the values of the input up to time t
as defined next.
Definition 1 (causality). An operator G: Lm pe Lr pe is said to be causal if (G(u))T = (G(uT ))T T 0, u
Lm pe .
We are now ready to present the first input-output stability concept.
INPUT-OUTPUT STABILITY
Lp Stability
Now we give the definition of Lp -stable systems that transform Lp input signals into Lp output signals.
Definition 2. An operator G: Lm pe Lr pe is
Lp -stable if
Finite-gain Lp -stable with zero bias if there exists a finite constant p such that
For finite-gain Lp -stable systems, the smallest scalar p for which there is a p such that relation (7) is
satisfied (when such p exist) is called the gain of the system. Similarly, for finite-gain Lp -stable systems with
zero bias, the smallest scalar p such that relation (8) is satisfied (when such p exist), is called the gain with
zero bias of the system.
Regarding finite-gain Lp stability for a causal operator G, we have the following result.
Lemma 1. (Ref. 29)
Let G: Lm pe Lr pe be a causal finite-gain Lp -stable operator with constants p and p . Then
Proof. Since G is causal (G(u))T = (G(uT ))T . Moreover if u Lm pe , then uT Lm p for all T 0; hence G(uT )
Lm p . Finally, for a generic function x, xT Lp xLp . Therefore the following inequalities hold:
For a causal, finite-gain Lp -stable operator with zero bias, it can be shown in the same way that
Remark 3. Recalling the definition of the Lm set with the norm (3), L stability is in fact what is normally
termed bounded-input, bounded-output (BIBO) or external stability. It is usually defined in reference to a system
specified through a state-space representation, and guarantees that if the input to the system is essentially
bounded, then its output is also essentially bounded.
Example: This example illustrates an L -stable and a finite-gain L -stable operator.
Let us consider the function g: defined by g(u) = uk with k (1, +). Correspondingly, define the
operator G: Le Le that assigns to every input u(t), t 0 the output g(u(t)). We will show that G is L
stable. Indeed set uL = c < +; then we have
INPUT-OUTPUT STABILITY
Therefore G is a BIBO operator. It is not a finite-gain L -stable operator; however, since we cannot find
fixed scalars and such that ck c + holds for all c [0, +).
By similar arguments as above, it is easy to check that G associated with the function g(u) = uk with k
(0, 1) is a finite-gain L -stable operator.
Example: There are some special cases when the gain of an operator can be numerically or explicitly found;
one of these cases is the L2 gain for linear systems. Let us consider a linear time-varying (LTV) system in the
form (for the sake of simplicity, we assume no feedthrough term)
where A(t), B(t), and C(t) are piecewise continuous and bounded.
= A(t)x is exponentially stable (see Appendix 1 for the definition of
We assume that the unforced system
exponential stability and Appendix 2 for a necessary and sufficient condition for exponential stability of LTV
systems). Let G be the input-output operator mapping u to y; then the L2 gain of G (which is also called energy
gain and induced-operator norm of G) is defined by
For a given > 0, we find (see Ref. 30) that 2 if and only if there exists an > 0 such that the Riccati
differential equation
admits a positive definite solution P(t) for which a symmetric matrix-valued function P(t) is said to be positive
definite if there exists a positive such that xT P(t)x x2 for all x n and t 0. Therefore by conducting a
binary search over , the L2 gain of G can be computed up to the desired precision.
The time-invariant case is even simpler, since we only have to deal with the algebraic version of Eq. (11). In
in fact
G(s) = C(sI A) B. It is possible to show that the L2 gain (10) is related to the transfer matrix G(s);
it is given by
INPUT-OUTPUT STABILITY
the G is not an L -stable operator according to Definition 1. However, let |u| c < 1; then
and so
which implies that G is an L -stable operator in a new sense made clear in the next definition.
The following definition is an extension of one originally presented in Ref. 31.
Definition 3. An operator G: Lm pe Lr pe is small-signal Lp -stable if there exists a positive constant r such
that Eq. (6) is satisfied for all u Lm p with supt u(t) < r. Similarly, G is a small-signal finite-gain Lp -stable
[operator small-signal finite-gain Lp -stable operator with zero bias] if there exists a positive constant r such that
inequality (7) [inequality (8)] is satisfied for all u Lm p with supt u(t) < r.
Remark 4. In Definition 3, we do not need to specify a particular norm to evaluate supt u(t), since we are
dealing with a norm in m and norms are equivalent in finite-dimensional spaces (see Remark 1). From Eq. (3)
it is clear that if u(t) is uniformly bounded then so is the signal norm u()L . This is not true in general as
it is easy to construct examples in which the signal norm u()Lp , with p [1, +), is arbitrarily large even if
u(t) is uniformly bounded.
Sufcient conditions for Lp stability. In the previous section we have provided the definition of Lp
stability for a given input-output operator. An important question then arises: How do we check whether an
input-output operator is Lp -stable or not?
To answer this question we should not focus, as done so far, on generic input-output operators; in this
section we assume that the operators under consideration are specified in terms of a state-space representation
of a dynamical system in the form
where x
,u
,y
, and
with D := {x n : x < r}, Du := {u m : u < ru }, where r and ru are positive numbers.
It is important to note that Eqs. (13a) define an input-output operator for any given initial state x0 . In other
words, system (13) defines an entire class of input-output operators, each obtained in relation to a particular x0 .
We should always distinguish between the concept of a system and that of an operator. In many textbooks, when
looking at the input-output behavior of a system, it is assumed that the initial condition is zero, and therefore
the system and the corresponding operator are the same object. As opposed to our treatment, the discussion of
Ref. 25 maintains the distinction between the input-output concepts and the state-space description.
In the remainder of this article, we assume that a state-space description of the dynamical system is given
unless otherwise specified. This will allow us to provide in this section a sufficient condition for Lp stability
INPUT-OUTPUT STABILITY
and, at the same time, to establish the first connection between Lyapunov (internal) stability and Lp stability.
Assume that x = 0 is an equilibrium point for system (13a) with u = 0, that is,
We shall see that, if x = 0 is an exponentially stable equilibrium point (see Appendix A for the definition)
and some other additional technical assumptions hold, the corresponding input-output operator is Lp -stable
for any x0 as described in the following theorem.
Theorem 1. (Corollary 6.1 of Ref. 23)
Assume that
(1) x = 0 is an exponentially stable equilibrium point for system (13a) under the input u = 0;
(2) f is continuously differentiable and the Jacobian matrices f /x and f /u are bounded, uniformly in t
(3) h is of Lipschitz form with respect to x and u, that is, there exist positive constants 1 , 2 such that
always satisfies assumption ii and iii of Theorem 1 globally if A() and B() are continuously differentiable
(actually this hypothesis can be relaxed to piecewise continuity) and uniformly bounded and C() and D() are
uniformly bounded; moreover, the exponential stability of x = 0 of system (14a) with u = 0 is always global.
Therefore we can state the following corollary of Theorem 1.
Corollary 1. Consider the linear system
where A() and B() are continuously differentiable and uniformly bounded and C() and D() are uniformly
bounded. Assume that the equilibrium point x = 0 under u = 0 of Eq. (15a) is exponentially stable; then for each
x0 n the operator defined by system (15) with initial condition x0 is finite gain Lp -stable for each p [1, +].
Recall that the L2 gain of the operator associated with the linear system (14) for x0 = 0 can be computed
according to the procedure detailed in Remark 4. Finally, a sufficient condition for exponential stability is given
in Appendix 2.
INPUT-OUTPUT STABILITY
Relations between Lyapunov stability and Lp stability. So far in this section we have shown the
following (Theorem 1):
is globally reachable if and only if it is reachable and is uniformly observable if and only if it is observable.
Moreover small-signal L2 stability implies L2 stability, and attractivity implies exponential stability. Therefore
we can derive the following corollary of Theorem 2.
Corollary 2. Assume that
(1) system (16) is reachable and observable and
(2) The input-output operator defined by system (16) with initial condition x(0) = 0 is L2 -stable.
Then x = 0 is an exponentially stable equilibrium point for system (16).
Interconnected Systems
One of the main applications of the formalism of input-output stability is the study of the stability of interconnected systems, without explicit knowledge of the internal dynamics of the composite subsystems. Let us
consider the feedback interconnection of Fig. 1, where G1 : Lm pe Lr pe and G2 : Lr pe Lm pe . Input-output
stability allows us to investigate how the signals propagate through this scheme. Before presenting the main
INPUT-OUTPUT STABILITY
results, we need to introduce the concept of well-posedness of the feedback interconnection. Well-posedness
guarantees that for each choice of u1 Lm pe and u2 Lr pe there exist unique solutions e1 ,y2 Lm pe and e2 ,y1
Lr pe that satisfy the loop equations
Small-Gain Theorems. Theorem 3. Consider the feedback system of Fig. 1. Suppose that G1 : Lm pe
L pe and G2 : Lr pe Lm pe are causal finite-gain Lp -stable operators with constants p1 ,p1 and p2 ,p2 , respectively. Moreover, suppose that the feedback interconnection is well-posed. Then the feedback system of
Fig. 1 is a finite-gain Lp -stable system if
r
and
Proof. Consider inputs u1 Lm p and u2 Lr p . Since the closed-loop system is well-posed, there exist unique
solutions e1 ,e2 ,y1 ,y2 . With respect to e1 and e2 we have
10
INPUT-OUTPUT STABILITY
Since G1 and G2 are causal finite-gain Lp -stable operators, we find that (see Lemma 1)
After some trivial manipulations, recalling that p1 p2 < 1 by assumption, Eqs. (18a) become
Now, the right-hand sides of inequalities (20) are independent of T. Therefore it can be easily shown that e1
and e2 belong to Lm p and Lr p , respectively, and that
In a similar way it can be shown that y1 and y2 belong to Lr p and to Lm p , respectively, and that inequalities
(17c) and (17d) hold.
The work of Safonov (7) exploited the general input-output concepts in order to study the robustness
of closed-loop systems. His results and variations were later exploited in H robust control analysis and
design. The small-gain theorem is thus extremely useful in studying the robustness of a closed-loop system,
when a nominal system is subject to a perturbation as shown in Fig. 2. The next example shows one of these
applications.
INPUT-OUTPUT STABILITY
11
Example: Let us consider the feedback scheme of Fig. 2 where G(s) and (s) are asymptotically stable
transfer matrices. The transfer matrix G(s) represents the nominal system, whereas (s) is a model of the
uncertainty. Let u1 and u2 belong to L2 . In Example 2 we have seen how to compute the L2 gains of G(s) and
(s); let
then the closed-loop system is a finite-gain L2 -stable system. In other words, it gives us an estimate of how
large the perturbation (s) can be, in terms of its H norm, preserving the closed-loop L2 stability.
There are various versions of the small-gain theorem, a sample of which is the incremental small-gain
theorem below, which needs a preliminary definition.
Definition 4. An operator G: Lm pe Lr pe is said to be an incrementally finite-gain stable operator if
(1) G(u) Lr p when u 0 and
(2) there exists a constant such that
Passivity Theorems. One of the main related concepts to the input-output stability concepts discussed
so far is the concept of passive systems. In a way, while Lp stability deals with the effect the system has on the
size of signals, passivity results deal with the effect the system has on the energy of signals. We start with
few definitions and follow with the main results for interconnected systems.
12
INPUT-OUTPUT STABILITY
Theorem 5. Consider the interconnected system of Fig. 1. Assume that the systems G1 and G2 satisfy
Note that this theorem does not require both systems G1 and G2 to be passive, as long as one of the two
systems is passive enough. If on the other hand, one of the two systems is passive and the other is strictly
passive, the previous theorem simplifies to the following.
Theorem 6. Consider the interconnected system of Fig. 1, and let u2 = 0. Then, the closed-loop system is
finite-gain L2 stable if one of the following conditions holds:
In the special case of affine in the control systems, the passivity of one system is equivalent to the
L2 -stability of a related system. This is detailed in the following theorem:
Theorem 7. (Proposition 3.2.12 of Ref. 18)
Let
and
where u = 12 (v z) and y = 12 (v + z). Then, system (22) is passive system (23) has L2 gain 1.
INPUT-OUTPUT STABILITY
13
Note that passivity can actually be defined as a special case of dissipativity by letting w(u, y) = uT y
(therefore, the system is square and m = r). We can also define other types of passivity as follows: the system is
an input-strictly-passive (ISP) system if it is dissipative with supply rate w(u, y) = uT y u u2 , u > 0, and it
is an output-strictly-passive (OSP) system if it is dissipative with supply rate w(u, y) = uT y y y2 , y > 0. The
system is a state-strictly-passive (SSP) system if w(u, y) = uT y x (x), x > 0 and (x) is a positive-definite
function of x. Note that an OSP system is necessarily L2 stable (18,23). In addition, one can guarantee the L2
stability of a system by making sure it is dissipative with the particular supply rate w(u, y) = 12 2 u2 y2
for some positive , which then becomes an upper bound on the L2 gain of the system (18,23).
The Linear Case and the KYP Lemma. One of the more important applications of the input-output
approach is in the solution of the so-called Lure problem shown in Fig. 3. The details of this approach are
detailed in another chapter of this encyclopedia and only a few comments are included here for completeness.
The basic question asked by Lure is to find conditions in the single-inputsingle-output case on the linear
system G(s) such that when the nonlinear block (t, y) is static (i.e., a non-time-varying function of y only),
the closed-loop system is stable (32). Popov provided graphical, frequency-domain criterion for the absolute
stability problem when the nonlinear block (y) is time invariant (33,34). Yakubovich (35) and Kalman (36)
introduced different versions of the so-called positive-real or Kalman-Yakubovich-Popov (KYP) lemma to relate
Popovs criterion to the existence of a special Lyapunov function. This then provides another connection between
input-output stability concepts and Lyapunov concepts. Anderson then (12,22) extended the KYP lemma to the
multi-inputmulti-output case. The KYP lemma has found various applications in adaptive control (19) and
14
INPUT-OUTPUT STABILITY
has recently been generalized to the case in which the linear block G(s) is replaced by a nonlinear but affine
nonlinear system = f (x) + g(x)u, y = h(x) (16,18).
In the linear case, passivity concepts may be related to the concept of positive-realness, already introduced
in the study of electrical networks (22). In fact, consider a stable, square, LTI system with minimal state-space
realization
where u, y
,x
Since the state-space realization is minimal, then (A, B) is controllable and (C, A) is observable. Recall that (A,
B) is controllable if and only if rank ( ) = n and (C, A) is observable if and only if rank ( ) = n where = [B
AB An 1 B] and T = [CT AT CT (An 1 )T CT ].
Definition 7. Let H(s) be a proper m m rational transfer matrix. Then
For variations on this definition, the reader should consult Ref. 37, where various PR concepts are
discussed.
Lemma 2. Let H(s) be an m m transfer matrix (25) where
(1) A is stable
(2) (A, B) is controllable and (C, A) is observable
Then, the transfer function H(s) is SPR if and only if there exist a positive definite symmetric matrix P,
matrices W and L, and a real > 0 such that
INPUT-OUTPUT STABILITY
15
Note that if D = 0, that is, H(s) is strictly proper, then Eqs. (26a) simplify to the more familiar
Lemma 3. H(s) is PR if and only if it is dissipative with storage function V(x) = xT Px.
Hill and Moylan (38) and others expanded the dissipativity notions in order to explain the KYP lemmas.
The KYP lemmas have applications to adaptive control for linear systems (19) and a generalization to nonlinear
systems (17). Connections between passivity and stability are provided in the next lemma.
Lemma 4. (Lemma 10.6 in Ref. 23)
Given the autonomous system
Passication via Feedback. In recent years, the input-output approach has gained new footing as a
design tool for nonlinear control systems. One of the main applications of such an approach has been to use
feedback in order to render a closed-loop system passive or strictly passive (or the passification of an open-loop
system). The main motivation for such designs is of course that a passive system will tolerate large-magnitude
uncertainties as long as the uncertainties are passive (see Theorem 5). References 16 and 17 contain a large
number of results on the passification of nonlinear systems. Roughly speaking, all designs require that the
open-loop system be of minimum phase and of a relative degree one in order for it to be made passive using
static output feedback. Such concepts have been generalized to a large class of nonlinear systems. As mentioned
previously, and following the early concepts (9), there has been much recent work on the discrete-time versions
of the input-output stability concepts including the passification designs in Refs. 26 and 27.
Input-to-State and Input-to-Output Stability. In a series of papers (13,14,15), Sontag and co-workers
have advanced the notion of input-to-state stability to study the behavior of state-space systems when the
input is bounded. Roughly, the input-to-state stability concepts guarantee that the state x(t) is bounded for
any bounded input u(t), which may be an external disturbance or a tracking signal. This idea is in some
ways a more restrictive version of the input-output concepts unless y = x and is more tightly coupled to the
Lyapunov-stability concepts. In what follows, we deal with system (13), or with its autonomous version:
16
INPUT-OUTPUT STABILITY
Definition 8. The system (13a) is said to be locally input-to-state stable (ISS) if there exists a class L function
, a class function (see Appendix 1 for the definitions of such functions), and positive constants k1 and k2
such that for any initial state x(t0 ) with x(t0 ) < k1 and any input u(t) with supt t0 u(t) < k2 , the solution
x(t) exists and
The system is said to be ISS stable if the preceding requirement holds globally (i.e., if D =
any bounded input u(t) and any initial condition x(t0 ).
and Du =
) for
, Du =
Note that while this is similar to the Lp -stability concepts presented previously, it is actually more general
as the function need not be linear and the function need not be a constant.
Theorem 9. (Theorem 6.3 of Ref. 23)
Let f (t, x, u) be piecewise continuous in t and of local Lipschitz form in (x, u), and let h be piecewise
continuous in t and continuous in (x, u). Assume that the system (13) is ISS, and that there exists class
functions 1 and 2 such that
Then the system (13) is locally IOS. If all assumptions hold globally and Eq. (13a) is ISS, then it is IOS.
Conclusions
In this article we have attempted to summarize various concepts of input-output stability for nonlinear dynamical systems, focusing on the continuous-time case. We have presented the basic input-output concepts but also
INPUT-OUTPUT STABILITY
17
some extensions and their applications to stability robustness analysis and design and to the adaptive control
of linear and nonlinear systems, as mentioned previously.
It is now clear that the connections between Lyapunov stability and input-output stability are strong and
may be exploited for further design. On the other hand, it is clear that the input-output approach remains a
versatile tool. This approach allows us to be able to determine the stability of the closed-loop system although
we have have little knowledge of the internal dynamics of the open-loop system and its uncertainties. This is
clearly an advantage when dealing with uncertain systems as the dynamics of the systems and its uncertainties
may be unknown. One of the limitations of the input-output approach, however, is that it remains limited to the
study of stability while other objectives such as the optimization of some performance indices remain beyond
the reach of these techniques. We end this article by mentioning that prior to its introduction to feedback
systems and control, the input-output approach was part of operator theory and functional analysis. Finally,
the input-output approach has been applied to various areas such as communications (39,40) and to the study
of neural network stability (41).
Appendix 1: Denitions
Definition 1.1 (Lipschitz functions). A function f (x) is said to be of local Lipschitz form on a domain D
if for each point x0 D there exist a neighborhood D0 of the point x0 and a constant L0 such that
If equation (32) holds for all x D with the same constant L, then f is said to be of Lipschitz form on D. If f is
of Lipschitz form on n than it is said to be of global Lipschitz form.
Definition 1.1 can be extended to the case of f (t, x) provided that the Lipschitz condition holds uniformly
on t for a given time interval.
Definition 1.2. (Function of class .
A continuous function : [0, +) [0, +) is said to be of class
Definition 1.3. (Function of class ).
A function : [0, +) [0, +) is said to be of class
if it is of class
and (s) as s .
The function (s): [0, +) s/(s+1) is of class since
(s) = 1/(s+1)2 > 0, but it is not of class since
lims (s) = 1 < +.
The function (s): [0, +) s is of class since
(s) = 1 > 0 and lims (s) = +.
The function (s, t): [0, +) [0, +) se t is of class L. Indeed it is strictly increasing in s, since
18
INPUT-OUTPUT STABILITY
strictly decreasing in t, since
and (s, t) 0 as t .
Definition 1.5 (State transition matrix). Given the LTV system
with x n and A(t) piecewise continuous, the state transition matrix (c,c):
is defined as the unique solution of the matrix differential equation
nn
, (t, t0 ) (t, t0 ),
where f (): D
Definition 1.6 (Attractivity). Consider the zero-input system (1.2) and denote by s(t, x0 ) the solution starting
from x0 at time t = 0. Assume that x = 0 is an equilibrium point of system (1.2); then x = 0 is attractive if there
exists a domain Da D, 0 Da such that
Definition 1.7 (Stability and Asymptotic Stability). Consider the zero-input system (1.2) and denote by
s(t, x0 ) the solution starting from x0 at time t = 0. Assume that x = 0 is an equilibrium point of system (1.2);
then x = 0 is
stable if, for each > 0, there exists a = () such that
, D = {x
INPUT-OUTPUT STABILITY
19
Definition 1.8 (Exponential stability). Consider the zero-input system (1.3) and assume that x = 0 is an
equilibrium point. Then x = 0 is exponentially stable if there exist positive numbers K, , c such that
The equilibrium x = 0 is globally exponentially stable if the above condition is verified for any initial state.
Definition 1.9 (Zero-State Observability). The system (27) is said to be zero-state observable from the output
y if for all initial conditions, y(t) 0 x(t) 0. The system is zero-state detectable if for all initial conditions
y(t) 0 limt x(t) = 0.
Definition 1.10 (Reachability). Consider the system
and denote by s(t, x0 , u) the solution starting from x0 at time t = 0 under the input u. Then system (1.5) is said
to be reachable if there exists a class function and a set D := {x n : x < r}, such that for all x D there
exists a time t and an input u such that uL (x) and s(t, 0, u) = x. The system is said to be globally
reachable if all the assumptions hold for all x n .
Definition 1.11 (Uniform observability). Consider the system
and denote by s(t, x0 , u) the solution starting from x0 at time t = 0 under the input u. Then system (1.6) is said
to be uniformly observable if there exists a class function such that for all x,
for some positive constants k1 , k2 , k3 , and . Then the equilibrium x = 0 is exponentially stable. Moreover, if all
the assumptions hold globally, x = 0 is globally exponentially stable.
20
INPUT-OUTPUT STABILITY
Theorem 2.2 (Exponential stability of linear systems). The equilibrium point x = 0 of the LTV system
is exponentially stable if and only if there exist positive constants and k such that
BIBLIOGRAPHY
1. H. S. Black Wave Translation System, US Patent No. 2,102,671, December 21, 1937.
2. H. W. Bode FeedbackThe History of an Idea, in R. Bellman and R. Kalaba (eds.), Selected Papers on Mathematical
Trends in Control Theory, pp. 106123. New York: Dover, 1964.
3. H. Nyquist Regeneration theory. Bell System Tech. J., 11: 126147, 1932.
4. I. W. Sandberg On the l2 -boundedness of solutions of nonlinear functional equations, Bell Sys. Tech. J., 43: 15811599,
1964.
5. G. Zames On the input-output stability of nonlinear time-varying feedback systems: Part i, IEEE Trans. Autom. Control,
11: 228238, 1966.
6. G. Zames On the input-output stability of nonlinear time-varying feedback systems: Part ii, IEEE Trans. Autom.
Control, 11: 465477, 1966.
7. M. Safonov Stability and Robustness of Multivariable Feedback Systems, Cambridge: MIT Press, 1980.
8. J. C. Willems The Analysis of Feedback Systems, Cambridge: MIT Press, 1971.
9. C. A. Desoer M. Vidyasagar Feedback Systems: Input Output Properties, New York: Academic Press, 1975.
10. J. C. Willems Dissipative dynamical systems: Part i: General theory, Arch. Ration. Mech. Anal., 45: 325351, 1972.
11. J. C. Willems Dissipative dynamical systems part ii: Linear systems with quadratic supply rate, Arch. Ration. Mech.
Anal., 45: 352393, 1972.
12. B. D. O. Anderson The small-gain theorem, the passivity theorem and their equivalence, J. Franklin Institute, 293(2):
105115, 1972.
13. E. Sontag On the input-to-state stability property, Eur. J. Control, 1: 2436, 1995.
14. E. Sontag Y. Wang Notions of input to output stability, Syst. Control Lett., 38: 235248, 1999.
15. E. Sontag Y. Wang Lyapunov characterizations of input to output stability, SIAM J. Control Optimization, 39, 226249,
2000.
16. C. I. Byrnes A. Isidori J. C. Willems Passivity, feedback equivalence, and the global stabilization of minimum phase
nonlinear systems, IEEE Trans. Autom. Control, 36: 12281240, 1991.
17. M. Krstc Kanellakopoulos P. V. Kokotivc Nonlinear and Adaptive Control Design, New York: Wiley-Interscience, 1995.
18. A. J. van der Schaft L2 -Gain and Passivity Techniques in Nonlinear Control, vol. 218, Lecture Notes in Control and
Information Sciences, Heidelberg: Springer-Verlag, 1996.
19. K. S. Narendra A. M. Annaswamy Stable Adaptive Systems, Englewood Cliffs, NJ: Prentice-Hall, 1985.
20. R. Ortega et al. Passivity-Based Control of Euler-Lagrange Systems: Mechanical, Electrical and Electromechanical
Applications, London: Springer-Verlag, 1998.
INPUT-OUTPUT STABILITY
21
21. R. Lozano O. Egeland B. Maschke Dissipative Systems Analysis and Control: Theory and Applications, London: Springer,
2000.
22. B. D. O. Anderson S. Vongpanitlerd Network Analysis and Synthesis: A Modern Systems Theory Approach, Englewood
Ciffs, NJ: Prentice Hall, 1973.
23. H. K. Khalil Nonlinear Systems, 2d ed., Upper-Saddle River, NJ: Prentice Hall, 1996.
24. S. Sastry Nonlinear Systems: Analysis, Stability, and Control, New York: Springer, 1999.
25. A. R. Teel T. T. Georgiou L. Praly E. Sontag Input-output stability, in W. S. Levine (ed.), The Control Handbook, pp.
895908, Boca Raton FL: CRC Press, 1995.
26. C. I. Byrnes W. Lin Losslessness, feedback equivalence, and the global stabilization of discrete-time nonlinear systems,
IEEE Trans. Autom. Control, 39(1): 8398, 1994.
27. W. Lin C. I. Byrnes Passivity and absolute stabilization of a class of discrete-time nonlinear systems, Automatica, 31(2):
263267, 1995.
28. H. L. Royden Real Analysis, London: Macmillan (Collier-MacMillan Limited), 2d. ed., 1968.
29. M. Vidyasagar Nonlinear System Analysis, 2d ed., Upper-Saddle River, NJ: Prentice Hall, 1993.
30. G. Tadmor Input/output norms in general linear systems, Int. J. Control, 51(4): 911921, April 1990.
31. M. Vidyasagar A. Vannelli New relationships between input-output and Lyapunov stability, IEEE Trans. Autom.
Control, 27(2): 481483, 1982.
32. A. I. Lure Some Nonlinear Problems in the Theory of Automatic Control, London: H.M. Stationery Office, 1957.
33. V. M. Popov Absolute stability of nonlinear systems of automatic control, Autom. Remote Control, 22: 857875, 1962.
34. V. M. Popov Hyperstability of Control Systems, New York: Springer, 1973.
35. V. A. Yakubovich Solution of certain matrix inequalities in the stability theory of nonlinear control systems, Engl.
Translation Sov. Math. Dokl., 3: 620623, 1962.
36. R. E. Kalman Lyapunov functions for the problem of Lure in automatic control, Proc. Nat. Acad. Sci., 49: 201205,
1963.
37. J. T. Wen Time domain and frequency domain conditions for strict positive realness, IEEE Trans. Autom. Control, 33:
988992, 1988.
38. D. J. Hill P. Moylan Dissipative dynamical systems: Basic input-output and state properties, J. Franklin Inst., 5:
327357, 1980.
39. I. W. Sandberg On the properties of systems that distort signals, Bell Syst. Tech. J., 42: 20332047, 1963.
40. I. W. Sandberg Notes on PQ theorems, IEEE Trans. Circuits Syst. I: Fundam Theory Appl., 41: 303307, 1994.
41. A. Mayer-Base F. Ohl H. Scheich Asymptotic hyperstability for the Hopfield model of neural networks, in Proc. IEEE
Joint Int. Conf. Neural Networks, pp. 24362441, Orlando, FL, 1994.
CHAOUKI T. ABDALLAH
University of Tennessee
FRANCESCO AMATO
MARCO ARIOLA
Universita` degli Studi di Napoli Federico II
INTELLIGENT CONTROL
AUTONOMOUS SYSTEMS, CONTROL, ROBOTS,
VEHICLES
The term intelligent control has come to mean, particularly to those outside the control area, some form of control
using methodologies such as intelligent agents, genetic algorithms, or fuzzy and/or neural networks. Intelligent control, however, does not restrict itself only to those methodologies. The fact is that there are problems of control today
that cannot be formulated and studied in the conventional
differential or difference equation mathematical framework using conventional (or traditional) control methodologies; these methodologies were developed in the past
decades to control dynamical systems. To address these
problems in a systematic way, a number of methods have
been developed in recent years that are collectively known
as intelligent control methodologies. It is worth remembering at this point that intelligent control uses conventional
control methods to solve lower level control problems and
that conventional control is included in the area of intelligent control. In summary, intelligent control attempts to
build upon and enhance the conventional control methodologies to solve new challenging control problems.
Conventional and Intelligent Control
The word control in intelligent control has a different,
more general meaning than the word control in conventional control. First, the processes of interest are more general and may be described, for example, by discrete event
system models, differential or difference equation models,
or both. This has led to the development of theories for hybrid control systems, which study the control of continuousstate dynamic processes by discrete-state controllers. In
addition to the more general processes considered in intelligent control, the control objectives can also be more
general. For example, replace part A in satellite can be
the general task for the controller of a space robot arm;
this is then decomposed into a number of subtasks, several of which may include, for instance, follow a particular
trajectory, which may be a problem that can be solved by
conventional control methodologies. To attain such control
goals for complex systems over a period of time, the controller has to cope with signicant uncertainty that xed
feedback robust controllers or adaptive controllers cannot
deal with. Since the goals are to be attained under large
uncertainty, fault diagnosis, and control reconguration,
adaptation and learning are important considerations in
intelligent controllers. It is also clear that task planning is
an important area in intelligent control design. So the control problem in intelligent control is an enhanced version of
the problem in conventional control. It is much more ambitious and general. The area of intelligent control is interdisciplinary, and it attempts to combine and extend theories
and methods from areas such as control, computer science,
and operations research to attain demanding control goals
in complex systems.
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright 2007 John Wiley & Sons, Inc.
Note that the theories and methodologies from the areas of operations research and computer science cannot,
in general, be used directly to solve control problems, as
they were developed to address different needs. The techniques from computer science and operations research are
primarily analysis tools developed for non-real-time systems, while in control, synthesis techniques to design realtime feedback control laws for dynamic systems are mainly
of interest. Recent work on real-time systems in computer
science has moved the research along lines that are of great
interest to intelligent control. In view of this discussion, it
should be clear that intelligent control research, which is
mainly driven by applications, has a very important and
challenging theoretical component. Signicant theoretical
strides must be made to address the open questions. The
problems are nontrivial, but the payoff is very high indeed.
As it was mentioned above, the word control is intelligent control has a more general meaning than in conventional control; in fact it is closer to the way the term control
is used in everyday language. Because intelligent control
addresses more general control problems that also include
the problems addressed by conventional control, it is rather
difcult to come up with meaningful benchmark examples.
Intelligent control can address control problems that cannot be formulated in the language of conventional control.
To illustrate, in a rolling steel mill, for example, conventional controllers may include the speed (revolutions per
minute) regulators of the steel rollers. In the intelligent
control framework one may include, in addition, fault diagnosis, alarm systems, and perhaps the problem of deciding on the set points of the regulators, which are based
on the sequence of orders processed, economic decisions,
maintenance schedules, availability of machines, etc. All
these factors have to be considered as they play a role in
controlling the whole production process, which is really
the overall goal.
Another difference between intelligent and conventional control is the separation between controller and the
system to be controlled. In conventional control the system
to be controlled, called the plant, typically is separate and
distinct from the controller. The controller is designed by
the control designer, while the plant is, in general, given
and cannot be changed. In intelligent control problems,
which are most often complex and challenging, there may
not be a clear separation of the plant and the controller; the
control laws may be embedded in and part of the system to
be controlled. This opens new opportunities and challenges
as it may be possible to affect the design of processes in a
more systematic way.
Areas relevant to intelligent control, in addition to
conventional control, include hybrid systems, planning
and knowledge-based systems, intelligent agents, machine
learning, search algorithms, fault diagnosis and control reconguration, and real-time and embedded systems, predicate logic, automata, Petri nets, neural networks, and fuzzy
logic genetic algorithms. In addition, in order to control
complex systems, one has to deal effectively with the computational complexity issue; this has been in the periphery
of the interests of the researchers in conventional control,
but it is clear that computational complexity is a central
issue whenever one attempts to control complex systems.
or intelligent machines. The term intelligent control system simply stresses the control aspect of the intelligent
system.
Characteristics or Dimensions of Intelligent Systems
There are several essential properties present in different degrees in intelligent systems. One can perceive them
as intelligent system characteristics or dimensions along
which different degrees or levels of intelligence can be measured. In the following we discuss three such characteristics that appear to be rather fundamental in intelligent
control systems.
Adaptation and Learning. The ability to adapt to changing conditions is necessary in an intelligent system. Although adaptation does not necessarily require the ability
to learn, for systems to be able to adapt to a wide variety of
unexpected changes, learning is essential. So the ability to
learn is an important characteristic of (highly) intelligent
systems.
Autonomy and Intelligence. Autonomy in setting and
achieving goals is an important characteristic of intelligent
control systems. When a system has the ability to act appropriately in an uncertain environment for extended periods of time without external intervention it is considered
to be highly autonomous. There are degrees of autonomy;
an adaptive control system can be considered as a system
of higher autonomy than a control system with xed controllers, as it can cope with greater uncertainty than a xed
feedback controller. Although for low autonomy no intelligence (or low intelligence) is necessary, for high degrees
of autonomy, intelligence in the system (or high degrees
of intelligence) is essential.
Structures and Hierarchies. In order to cope with complexity, an intelligent system must have an appropriate
functional architecture or structure for efcient analysis
and evaluation of control strategies. This structure should
provide a mechanism to build levels of abstraction (resolution, granularity) or at least some form of partial ordering so to reduce complexity. An approach to study intelligent machines involving entropy (of Saridis) emphasizes such efcient computational structures. Hierarchies
(that may be approximate, localized, or combined in heterarchies) that are able to adapt may serve as primary
vehicles for such structures to cope with complexity. The
term hierarchies refers to functional hierarchies or hierarchies of range and resolution along spatial or temporal
dimensions, and it does not necessarily imply hierarchical
hardware. Some of these structures may be hardwired in
part. To cope with changing circumstances the ability to
learn is essential so these structures can adapt to signicant, unanticipated changes.
In view of the preceding points, a working characterization of intelligent systems [or of (highly) intelligent (control) systems or machines] that captures the essential characteristics present in any such system follows: An intelligent system must be highly adaptable to signicant unanticipated changes, and so learning is essential. It must ex-
Figure 1. Intelligent autonomous controller functional architecture. The three levels of a hierarchical intelligent control architecture are the execution level, the coordination level, and the management or organization level.
the use of frequency domain methods and Laplace transforms in the 1930s and 1940s and the development of optimal control methods and state-space analysis in the 1950s
and 1960s. Optimal control in the 1950s and 1960s, followed by progress in stochastic, robust, adaptive, nonlinear hybrid, and networked control methods in the 1960s
to today, has made it possible to control signicantly more
complex dynamical systems than the original yball governor more accurately.
Conventional control systems are designed today using
mathematical models of physical systems. A mathematical model that captures the dynamical behavior of interest
is chosen and then control design techniques are applied,
aided by computer-aided design (CAD) packages, to design
the mathematical model of an appropriate controller. The
controller is then realized via hardware or software and it
is used to control the physical system. The procedure may
take several iterations. The mathematical model of the system must be simple enough so that it can be analyzed with
available mathematical techniques and accurate enough
to describe the important aspects of the relevant dynamical behavior. It approximates the behavior of a plant in the
neighborhood of an operating point.
The control method and the underlying mathematical
theory were developed to meet the ever-increasing control
needs of our technology. The need to achieve the demanding
control specications for increasingly complex dynamical
systems has been addressed by using more complex mathematical models such as nonlinear and stochastic ones,
and by developing more sophisticated design algorithms
for, say, optimal control. The use of highly complex mathematical models, however, can seriously inhibit our ability
to develop control algorithms. Fortunately, simpler plant
models, for example, linear models, can be used in the control design; this is possible because of the feedback used
in control that can tolerate signicant model uncertainties. When the xed feedback controllers are not adequate,
then adaptive controllers are used. Controllers can then
be designed to meet the specications around an operating
point where the linear model is valid, and then via a scheduler a controller emerges that can accomplish the control
objectives over the whole operating range. This is, for example, the method typically used for aircraft ight control
and it is a method used to design xed controllers for certain classes of nonlinear systems. Adaptive control in conventional control theory has a specic and rather narrow
meaning. In particular, it typically refers to adapting to
variations in the constant coefcients in the equations describing the linear plant; these new coefcient values are
identied and then used, directly or indirectly, to reassign
the values of the constant coefcients in the equations describing the linear controller. Adaptive controllers provide
for wider operating ranges than xed controllers and so
conventional adaptive control systems can be considered to
have higher degrees of autonomy than control systems employing xed feedback controllers.
Intelligent Control for High-Autonomy Systems
There are cases in which we need to increase the operating range of the system signicantly. We must be able
to deal effectively with signicant uncertainties in models of increasingly complex dynamical systems in addition
to increase the validity range of our control methods. We
need to cope with signicant unmodeled and unanticipated
changes in the plant, in the environment, and in the control
objectives. This will involve the use of intelligent decisionmaking processes to generate control actions so that a certain performance level is maintained even though there are
drastic changes in the operating conditions. I have found
useful to keep in mind an example that helps set goals for
the future and also teaches humility, as it shows how difcult, demanding, and complex autonomous systems can
be. Currently, if there is a problem on the space shuttle,
the problem is addressed by the large number of engineers
working in Houston Control, the ground station. When the
problem is solved the specic detailed instructions about
how to deal with the problem are sent to the shuttle. Imagine the time when we will need the tools and expertise of all
Houston Control engineers aboard the space shuttle or another space vehicle for extended space travel. What needs
to be achieved to accomplish this goal is certainly highly
challenging!
In view of the above it is quite clear that in the control
of systems there are requirements today that cannot be
successfully addressed with the existing conventional control theory. It should be pointed out that several functions
proposed in later sections, to be part of the high-autonomy
control system, have been performed in the past by separate systems; examples include fault trees in chemical process control for failure diagnosis and hazard analysis, and
control system design via expert systems.
An Intelligent Control Architecture for High-Autonomy
Systems
To illustrate the concepts and ideas involved and to provide
a more concrete framework to discuss the issues, a hierarchical functional architecture of an intelligent controller
that is used to attain high degrees of autonomy in future
space vehicles is briey outlined as an example; full details
can be found in Ref. 21. This hierarchical architecture has
three levels, the execution level, the coordination level, and
the management or organization level. The architecture
exhibits certain characteristics, which have been shown in
the literature to be necessary and desirable in autonomous
systems. Based on this architecture we identify the important fundamental issues and concepts that are needed for
an autonomous control theory.
Architecture Overview: Structure and Characteristics. The
overall functional architecture for an autonomous controller is given by the architectural schematic of the Fig. 1.
This is a functional architecture rather than a hardware
processing one; therefore it does not specify the arrangement and duties of the hardware used to implement the
functions described. Note that the processing architecture
also depends on the characteristics of the current processing technology; centralized or distributed processing may
be chosen for function implementation depending on available computer technology.
Research Directions. One can roughly categorize research in the area of intelligent autonomous control into
two areas: conventional control theoretic research, which
addresses the control functions at the execution and coordination levels, and the modeling, analysis, and design
of higher-level decision-making systems found in the management and coordination levels.
HYBRID SYSTEMS
Hybrid control systems contain two distinct types of systems, systems with continous dynamics and systems with
discrete dynamics, that interact with each other. Their
study is central in designing intelligent control systems
with a high degree of autonomy and it is essential in designing discrete event supervisory controllers for continuous systems; see Refs. 126.
Hybrid control systems typically arise when continuous
processes interact with, or are supervised by, sequential
machines. Examples of hybrid control systems are common in practice and are found in such applications as exible manufacturing, chemical process control, electric power
distribution, and computer communication networks. A
simple example of a hybrid control system is the heating
and cooling system of a typical home. The furnace and air
conditioner, along with the heat-ow characteristics of the
home, form a continuous-time system that is to be controlled. The thermostat is a simple discrete-event system
that basically handles the symbols {too hot, too cold} and
{normal}. The temperature of the room is translated into
these representations in the thermostat and the thermostats response is translated back to electrical currents that
control the furnace, air conditioner, blower, etc.
Since the continuous and discrete dynamics coexist and
interact with each other it is important to develop models that accurately describe the dynamic behavior of such
hybrid systems. In this way it is possible to develop control strategies that fully take into consideration the relation and interaction of the continuous and discrete parts
of the system. In the past, models for the continuous- and
discrete-event subsystems were developed separately; the
control law was then derived in a rather empirical fashion,
except in special cases such as the case of digital controllers
for linear time-invariant systems. The study of hybrid systems provides the backbone for the formulation and implementation of learning control policies. In such policies,
the control acquires knowledge (discrete data) to improve
the behavior of the system as it evolves in time. Hybrid
systems have become a distinctive area of study due to
opportunities to improve on traditional control and estimation technologies by providing computationally effective
methodologies for the implementation of digital programs
that design or modify the control law in response to sensordetected events, or as a result of adaptation and learning.
The interested reader should consult Refs. 25, 26 and especially the references therein.
Certain important issues in hybrid systems are now
briey discussed using a paradigm of a continuous system supervised by a DES controller from Refs. 23 and 24.
The hybrid control system of interest here consists of a
continuous-state system to be controlled, also called the
plant, and a discrete-state controller connected to the plant
via an interface; see Fig. 2.
The plant contains all continuous-state subsystems
of the hybrid control system, such as any conventional
continuous-state controllers that may have been developed
and a clock if time and synchronous operations are to be
modeled. The controller is an event-driven, asynchronous
DES, described by a nite state automaton or an ordinary
Petri net. The hybrid control system also contains an interface that provides the means for communication between
the continuous-state plant and the DES controller. The interface receives information from the plant in the form of
a measurement of a continuous variable z(t), such as the
continuous state, and issues a sequence of symbols {z(i)}
10
CONCLUDING REMARKS
There may be the temptation to classify the area of intelligent autonomous systems as simply a collection of methods
and ideas already addressed elsewhere, the need only being some kind of intelligent assembly and integration of
known techniques. This is not true. The theory of control
systems is not covered by, say, the area of applied mathematics, because control has different needs and therefore
asks different questions. The problems of interest in intelligent systems require development of novel concepts, approaches, and methods in computer science, operations research control systems, to mention but a few. The marriage
of all these elds can only be benecial to all. Computer science and operation research methods are increasingly used
in control problems, while control system concepts such
as feedback and methods are providing the base for new
theories and methods in those areas. Intelligent control
for high degree of autonomy systems is a vibrant research
and applications area where developments are followed by
interdisciplinary research and advances in computational
networking, sensing, and artworking technologies.
BIBLIOGRAPHY
1. P. J. Antsaklis, Dening intelligent control, Report of the IEEE
Control Systems Society Task Force on Intelligent Control (P.
J. Antsaklis, Chair; J. Albus, K. Baheti, J. D. Birdwell, M.
Lemmon, M. Mataric, A. Meystel, K. Narendra, K. Passino, H.
Rauch, G. Saridis, H. Stephanou, P. Werbos, Members), IEEE
Control Syst. Mag., 14 (3): 45, 5866, 1994.
2. P. J. Antsaklis, K. M. Passino (eds.), An Introduction to Intelligent and Autonomous Control, Norwell, MA: Kluwer, 1993.
3. A. Meystel, Autonomous Mobile Robots, River Edge, NJ: World
Scientic, 1991.
4. K. P. Valavanis, G. N. Saridis, Intelligent Robotic System Theory: Design and Applications, Norwell, MA: Kluwer, 1992.
5. J. S. Albus and A. M. Meystel, Intelligent Systems:Architecture,
Design, Control, Wiley-Interscience, 2001.
6. R. A. Brooks Cambrian Intelligence: The Early History of the
New AI, The MIT Press, 1999.
7. J. S. Albus, Outline for a theory of intelligence, IEEE Trans.
Syst. Man Cybern., 21 (3): 432509, 1991.
8. G. N. Saridis, K. P. Valavanis, Analytical design of intelligent
machines, Automatica, 24 (2): 123133, 1988.
9. J. S. Sinha and A. M. Meystel, Engineering of Mind: An
Introduction to the Science of Intelligent Systems, WileyInterscience, 2001.
10. D. A. White, D. A. Sofge (eds.), Handbook of Intelligent Control Neural, Fuzzy, and Adaptive Approaches, New York: Van
Nostrand Reinhold, 1992.
11. M. M. Gupta, N. K. Sinha (eds.), Intelligent Control: Theory
and Practice, Piscataway, NJ: IEEE Press, 1994.
12. N. K. Sinha andM. M. Gupta (Eds), Soft Computing & Intelligent Systems: Theory and Applications, Academic Press, 2000.
13. R. S. Michalski, J. G. Carbonell, T. M. Mitchell, Machine
LearningAn Articial Intelligence Approach, Palo Alto, CA:
Tioga, 1983.See also R. S. Michalski andG. Tecuci (eds.), Ma-
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
PANOS J. ANTSAKLIS
University of Notre Dame
Notre Dame, IN, 46556
11
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright 2007 John Wiley & Sons, Inc.
Nonlinear Filters
where E is the expected value operator and T is the transpose of the vector.
The random process x(t) is called uncorrelated if
where (t) is the Dirac delta function (actually, a generalized function), dened by
WHITE NOISE
Q1 (t) and Q2 (k) are the intensity matrices of the white noise
process and sequence. If Q1 (t) and Q2 (t) are constant, the
processes and sequences are stationary. If the probability
distribution of a white noise process at each instant of time
is Gaussian, then it is completely dened by its rst two
moments, mean and variance. If Ex(t) = 0, the process is
called zero mean.
A white noise process or sequence is an example of an
uncorrelated process or sequence. Generally, a white noise
process has no time structure. In other words, knowledge
of the white process value at one instant of time provides
no knowledge of what its value will be (or was) at any other
time point.
Nonlinear Filters
LINEAR ESTIMATION
Linear estimation addresses the problem of estimating the
state of a linear stochastic system by using measurements
or sensor outputs that are linear functions of the state. We
suppose that the stochastic systems can be represented by
the types of plant and measurement models (for continuous and discrete time) shown as equations in Table 2, with
dimensions of the vector and matrix quantities. The measurement and plant noise vk and wk , respectively, are assumed to be zero-mean Gaussian processes, and the initial
value xo is a Gaussian random variable with known mean
x0 and known covariance matrix Po . Although the noise sequences wk and vk are assumed to be uncorrelated, this
restriction can be removed, modifying the estimator equations accordingly.
The objective of statistical optimization is to nd an estimate of the n state vector xk represented by x k , a linear
function of the measurements zi , . . . , zk , which minimizes
the weighted mean-squared error
or its discrete-time equivalent. This implementation automatically minimizes the expected risk associated with
any quadratic loss function of the estimation error.
The statistical performance of the estimator can be predicted a priori (i.e., before it is actually used) by solving
the nonlinear differential (or difference) equations used
in computing the optimal feedback gains of the estimator. These are called Riccati equations, named in 1763 by
Jean le Rond DAlembert (17171783) for Count Jacopo
Francesco Riccati (16761754), who had studied a secondorder scalar differential equation, although not the form
that we have here (13, 14). Kalman gives credit to Richard
S. Bucy for the discovery that the Riccati differential equation serves the same role as the Wiener-Hopf integral equation in dening optimal gains. The Riccati equation also
arises naturally in the problem of separation of variables
in ordinary differential equations and in the transformation of two-point boundary value problems to initial value
problems (15). The behavior of their solutions can be shown
analytically in trivial cases. These equations also provide
a means for verifying the proper performance of the actual
estimator when it is running.
THE LINEAR OPTIMAL ESTIMATOR IN DISCRETE TIME
(KALMAN FILTER)
Suppose that a measurement has been made at time tk and
that the information that it provides is to be applied in
updating the estimate of the state x of a stochastic system
at time tk . It is assumed that the measurement is linearly
related to the state by an equation of the form zk = Hxk + vk ,
where H is the measurement sensitivity matrix and vk is
the measurement noise.
The optimal linear estimate is equivalent to the general (nonlinear) optimal estimator if the random variables
x and z are jointly Gaussian. Therefore, it sufces to seek
an updated estimate x k (+) (observation zk is included in the
estimate) that is a linear function of the a priori estimate
and the measurement z:
Equations Table 2 and for x k (+) from Eq. (9) into Eq. (10),
then we will observe from Eqs. 1 and 2 that the data z1 , . . . ,
zk do not involve the noise term wk . Therefore, because the
random sequences wk and vk are uncorrelated, it follows
that Ewk zT i = 0 for 1 i k.
Nonlinear Filters
and
Let
because vk is a white noise sequence uncorrelated with wk
white noise i = 1, . . . k 1. Then Eq. (13) can be reduced to
the form
and therefore
Let
Equation (16) can be satised for any given xk if
Nonlinear Filters
Therefore,
Let
arrows show the successive values assumed by the variables, with the annotation (in parentheses) on the arrows
indicating which input variables dene the indicated transitions.
Note that each variable assumes two distinct values at
each discrete time: its a priori value corresponding to the
value before the information in the measurement is used
and the a posteriori value corresponding to the value after
the information is used.
At the beginning of the design phase of a measurement
and estimation system, when neither real nor simulated
data are available, just the covariance calculations can
be used to obtain preliminary indications of lter performance. Covariance calculations consist of solving the estimator equations with steps 13, repeatedly. It is important
to notice that the covariance calculations are independent
of data zk . Covariance calculations will involve the plant
noise matrix Q, measurement noise matrix R, state transition matrix , measurement matrix H, and initial covariance matrix Po .
Step 4 of the Kalman lter implementation [computation of x k (+)] can be implemented only for state vector propagation where simulator or real data sets are available.
In the design tradeoffs, the covariance matrix update
(steps 1 and 3) should be checked for symmetry and positive deniteness. Failure to attain either condition is a sign
that something is wrong. One possibility is that it is an illconditioned problem. In order to overcome ill-conditioning,
another equivalent expression for Pk (+) is called the Joseph
form, as shown Eq. (32)
Note that the right-hand side of this equation is the summation of two symmetric matrices. The rst of these is positive denite and the second is nonnegative denite, thereby
making Pk (+) a positive denite matrix. Other techniques
are described in the implementation methods to alleviate
the ill-conditioning.
k and Pk (+) are not that useful (817).
Other forms for K
It can be shown that state vector update, Kalman gain,
and error covariance equations represent an asymptotically stable system; therefore, the estimate of state x k becomes independent of the initial estimate x o , Po as k is increased. It is also obvious that the Kalman gain and error
covariance equations are independent of the observation.
These equations are used for covariance analysis purposes.
THE CONTINUOUS-TIME OPTIMAL ESTIMATOR
(KALMAN-BUCY FILTER)
Analogous to the discrete-time case, the continuous-time
random process x(t) and the observation z(t) are given by
Nonlinear Filters
Nonlinear Filters
Substituting Eq. (50) into Eq. (51) and taking the limit as
t 0, we obtain the desired result
with
initial
condition
x (0).
Equations
<xref
target="W1020-mdis-0051 W1020-mdis-0054"/>, and <xref
target="W1020-mdis-0055"/> dene the continuous-time
Kalman estimator, which is also called the Kalman-Bucy
lter (119).
The Wiener lter is dened for stationary systems in
continuous time, and the Kalman lter is dened for either stationary or nonstationary systems in either discrete
time or continuous time, but nite state dimension. To
demonstrate the connections on problems satisfying both
sets of constraints, like the continuous-time Kalman Bucy
estimator Eqs. <xref target="W1020-mdis-0051"/>, <xref
target="W1020-mdis-0054"/>, and <xref target="W1020mdis-0055"/>, letting F, G, H be constants; the noises be
stationary (Q and R constants), and the lter reach steady
state (P constant). That is, as t , then P(xt) 0. The
Riccati differential Eq. (54) becomes the algebraic Riccati
equation for continuous time systems. Taking the Laplace
with zero process noise and with the mean xN 0 as the initial
condition would be a nominal trajectory for a discrete-time
system.
The word perturbation has been used by astronomers
to describe a minor change in the trajectory of a planet
(or any free-falling body) due to secondary forcessuch as
those produced by other gravitational bodies. Astronomers
had learned long ago that the actual trajectory can be accurately modeled as the sum of the solution of the two-body
problem (which is available in closed form) and a linear
dynamic model for the perturbations due to the secondary
forces. This technique also works well for many other nonlinear problems, including the problem at hand. In this
case, the perturbations are due to the presence of random
process noise and errors in the assumed initial conditions.
If the function f in the previous example is continuous,
then the state vector xk at any instant on the trajectory will
vary smoothly with small perturbations of the state vector
xk1 at the previous instant. These perturbations are the
result of off-nominal (i.e., off-mean) values of the random
variates involved. These random variates include the initial value of the state vector x0 , the process noise wk , and
(in the case of the estimated trajectory) the measurement
noise vk .
If f is continuously differentiable, then the inuence of
the perturbations on the trajectory can be represented by a
Taylor series expansion about the nominal trajectory. The
likely magnitudes of the perturbations are determined by
Nonlinear Filters
the variances of the variates involved. If these perturbations are sufciently small relative to the higher-order coefcients of the expansion, then we can obtain a good approximation by ignoring terms beyond some order. (However,
we must usually evaluate the magnitudes of the higherorder coefcients before making such an assumption.)
Let the symbol denote perturbations from the nominal
Let h be sufciently differentiable, then the measurement can be represented by a Taylor series:
or
or
a n n constant matrix.
Nonlinear Filters
and
The models and implementation equations of the extended Kalman lter are summarized in Table 6. The last
three equations in this table are the same as those for the
standard Kalman lter, but the other equations are noticeably different from those of the linearized Kalman lter
in Table 5.
It has been said that modeling is the hardest part of
Kalman ltering. This is especially true when there are
nonlinearities in the physical equation that must be linearized. Developing a good Kalman lter model is partly
art and partly science. As a general rule, we look for
models that are simple enough to be implemented but, at
the same time, still represent the physical situation with
a reasonable degree of accuracy (3).
THE MATRIX RICCATI DIFFERENTIAL EQUATION
In order to implement a Kalman lter, the Riccati equation
must be solved. This section presents a brief discussion of
solution methods for the Riccati differential equation for
the Kalman-Bucy lter. A more thorough treatment of the
Riccati equation can be found in Ref. 21.
Transformation to a Linear Equation
The matrices have the same general form as for linearization about a nominal trajectory, except for the evaluations
of the partial derivatives:
and
Linearized and Extended Kalman Filters. The block diagram of Fig. 4 shows the data ow of the estimator linearized about a nominal trajectory of the state dynamics.
Note that the operations within the dashed box have no
inputs. These are the computations for the nominal trajectory. Because they have no inputs from the rest of the estimator, they can be precomputed ofine. The models and implementation equations for the linearized discrete Kalman
lter are summarized in Table 5. Note that the last three
equations in this table are identical to those of the standard Kalman lter.
10
Nonlinear Filters
Nonlinear Filters
or taking the differences of the left-hand sides and substituting P for AB1
11
12
Nonlinear Filters
where
Consequently,
and P0 is a constant symmetric position denite n n matrix, which is the solution of the algebraic matrix Riccati
equation
Nonlinear Filters
13
against roundoff errors. Numerically stable implementation methods are called square root ltering because they
use factors of the covariance matrix of estimation uncertainty or its inverse, called the information matrix.
Numerical solution of the Riccati equation tends to be
more robust against roundoff errors if Cholesky factors of
a symmetrical nonnegative denite matrix P is a matrix
C such that CCT = P. Cholesky decomposition algorithms
solve for C that is either upper triangular or lower triangular. Another method is modied Cholesky decomposition.
Algorithms solve for diagonal factors and either a lower triangular factor L or an upper triangular factor U such that
P = UDu UT = LDL LT where DL and Du are diagonal factors
with nonnegative diagonal elements. Another implementation method uses square root information lters that use a
symmetric product factorization of the information matrix
P1 .
Alternative Kalman lter implementations use these
factors of the covariance matrix (or its inverse) in three
types of lter operations: (1) temporal updates, (2) observation updates, and (3) combined updates (temporal and
observation). The basic algorithm methods used in these
alternative Kalman lter implementations fall into four
general categories. The rst three of these categories are
concerned with decomposing matrices into triangular factors and maintaining the triangular form of the factors
through all the Kalman ltering operation. The fourth
category includes standard matrix operations (multiplication, inversion, etc.) that have been specialized for triangular matrices. These implementation methods have succeeded where the conventional Kalman lter implementations have failed (3, 31).
Even though uses are being explored in virtually every discipline, research is particularly intense on successful implementation of Kalman ltering to global positioning systems (GPS), inertial navigation systems (INS), and
guidance and navigation. GPS is a satellite-based system
that has demonstrated unprecedented levels of positioning accuracy, leading to its extensive use in both military
and civil arenas. The central problem for GPS receivers is
the precise estimation of position, velocity, and time based
on noisy observations of satellite signals. This provides an
ideal setting for the use of Kalman ltering. GPS technology is used in automobile, aircraft, missiles, ships, agriculture, and surveying.
In 1995 the United States began development of the
Wide Area Augmentation System (WAAS) under the auspices of the Federal Aviation Administration (FAA) and
the Department of Transportation (DOT), to provide precision approach capability for aircraft. Without WAAS, ionospheric disturbances, satellite clock drift, and satellite orbit errors cause too much error in the GPS signal for aircraft to perform a precision landing approach. Additionally,
signal integrity information as broadcast by the satellites
is insufcient for the demanding needs of public safety in
aviation. WAAS provides additional integrity messages to
aircraft to meet these needs.
WAAS includes a core of approximately twenty-ve wide
area ground reference stations (WRS) positioned throughout the United States which have precisely surveyed coordinates. These stations compare the GPS signal measure-
14
Nonlinear Filters
BIBLIOGRAPHY
1. R. E. Kalman, A new approach to linear ltering and prediction problems, ASME J. Basic Eng., 82: 3445, 1960.
2. A. Gelb et al., Applied Optimal Estimation, Cambridge, MA:
MIT Press, 1974.
3. M. S. Grewal, A. P. Andrews, Kalman Filtering Theory & Practice Using MATLAB (Second Edition), New York, NY: Wiley &
Sons, 2001 (8th printing).
4. B. D. O. Anderson, J. B. Moore, Optimal Filtering, Englewood
Cliffs, NJ: Prentice Hall, 1979.
5. A. E. Bryson, Jr., Y. C. Ho, Applied Optimal Control, Waltham,
MA: Blaisdell, 1969.
Nonlinear Filters
31. G. J. Bierman, Factorization Methods for Discrete Sequential
Estimation, New York: Academic, 1977.
32. M. S. Grewal, L. R. Weill, and A. P. Andrews, Global Positioning Systems, Inertial Navigation, and Integration (Second Edition), New York, NY: Wiley & Sons, 2007.
MOHINDER S. GREWAL
California State University,
Fullerton Fullerton, CA,
92634
15
16
Nonlinear Filters
209
210
techniques, which need not be constrained by physical boundaries of subsystems, but may lead to efficient solutions of
large problems utilizing modern computer architectures. With
that in mind, our first objective will be to describe several
decomposition schemes that can either be used as preconditioners for decentralized control design, or can serve to speed
up computations involving the control of large-scale systems
(usually through the use of parallel processing).
To efficiently decompose a large-scale system, it is generally convenient to represent it in the form of a graph. Depending on whether or not the graph is weighted, we can
roughly distinguish between two broad classes of decomposition algorithms. In problems where weights are not assigned
to edges of the graph, decomposition schemes typically exploit
topological properties such as structure and sparsity to obtain
an appropriate partitioning (18). Whenever possible, it is
also useful to incorporate any existing information regarding
the physical attributes of the system (such as hierarchical
structures or repetitive blocks that are built into the design).
Decomposition algorithms of this type are commonly applied
for solving large sparse systems of linear equations, and a
number of them have been successfully utilized in parallelizing control-related computational problems.
A conceptually different class of algorithms arises when
the system is viewed as a weighted graph. In this case the
partitioning strategy changes significantly, since we can now
utilize edge weights to identify weakly coupled subsystems
and establish hierarchical relationships between them. Decomposition schemes based on this approach have found numerous applications not only in parallel computing, but also
in decentralized control design. In the following sections particular attention will be devoted to the epsilon decomposition
algorithm (911), due to its efficiency and simplicity; in this
context, we will also examine the concept of overlapping, and
its application to the control of large-scale systems.
Our second objective in this paper will be to provide the
motivation and describe the basic ideas and techniques for
decentralized control of dynamic systems. The accumulated
experience in controlling large complex systems suggests
three basic reasons for using decentralized control structures:
dimensionality, information structure constraints, and uncertainty (6). By decomposing a system of large dimension into
subsystems, a designer can devise decentralized strategies for
solving control problems that would be either impractical or
impossible to solve using a single centralized controller. Furthermore, in a large complex system where databases are developed around the plant with distributed sources of data, a
need for fast control actions in response to local inputs and
perturbations dictates use of distributed (that is, decentralized) measurement and control structures.
A restriction on what and where the information is delivered in a large system is a common structural constraint in
building controllers and estimators. A good example is the
standard automatic generation control in power systems,
where the decentralized schemes are used to reduce the cost
of communication that would be demanded by a centralized
control strategy spread over distant geographical areas.
In modeling and control of large systems, it has been long
recognized that models of subsystems can be obtained with
increasing levels of accuracy and versatility. The essential uncertainty resides in the interconnections between the subsystems, since these interconnections are often poorly known in
x1
y1
x2
y2
x3
y3
(1)
(2)
1
0.05
A = 0.01
1
0.1
0.1
0.1
2
(3)
y1
x2
y2
x3
y3
Figure 2. Subgraph B.
x1
y1
x3
y3
x2
y2
211
VA = AV
Figure 3. Components of B.
(8)
where
A =
---------
---------
1
2 0
0
0 0.5
AD = 0.1
AC = - 0
(4)
- - - - -0- - - - -1- -
- - - - - -2- - - -0- ,
0
0 1
0.1 1
0
(k = 0, 1, . . .)
1
0
V =
0
0
(5)
---------
1
0.05 0
AD = -0.01
(6)
- - - - - - - -1- - - - - -0-
0
0
2
we obtain I A1
D A2 1, and the process diverges; on the
other hand, the partitioning in Eq. (4) obtained by epsilon
decomposition results in rapid convergence.
Overlapping Epsilon Decompositions. Given a matrix A and
a particular choice of parameter , there is no guarantee that
A can be permuted into the form shown in Eq. (2). The obvious remedy in such cases is to repeat the decomposition with
a larger value of ; alternatively, we can use the concept of
overlapping. The following example illustrates the basic ideas
behind overlapping epsilon decomposition.
Example 2. Consider the following matrix
A =
0
1
0
1
0
0
1
0
(10)
Scaling. Another issue that arises in the practical application of epsilon decompositions is that elements of the matrix
A can widely vary in size. In such cases it may not be possible
to find a meaningful value for , and row scaling needs to be
utilized to obtain a more uniform distribution of element values. This process is demonstrated by the following example.
Example 3. Let matrix A be
10
A = 0.1
4.5
defined as
5
0.3
0.2 0.02
1
100
(11)
x1
x1
y1
x2(1)
x2
y2
x3
x3
y3
x2(2)
Figure 4. Subgraph B.
(9)
now has an
It is easily verified that the expanded matrix A
epsilon decomposition in the sense of Eq. (2).
(7)
0
0
and
0
0
(12)
y1
y2(1)
y3
y2(2)
.
Figure 5. Expanded subgraph B
212
---------
1
0.5 0
AD = -0.5
- - - - - - -1- - - - 0
- - ,
0
0
1
0
0
0.3
(13)
A1
D AC
=
AD AC
1
(14)
Structural Decompositions
By their very nature, epsilon decompositions are ideally
suited for iterative solutions of Eq. (1). In contrast, there is
an entire class of decompositions that are aimed at solving
Eq. (1) directly; in this approach, no matrix entries are discarded, and the decompositions are designed to achieve certain desirable structures. A structure that is of particular interest in the analysis and simulation of large-scale systems is
the bordered block-diagonal (BBD) form shown in Fig. 6. The
appeal of this structure lies in its inherent potential for parallel computation, a feature that has been widely exploited in
different areas of engineering.
Numerous algorithms have been developed for permuting
a matrix into the BBD form, based on diverse concepts ranging from node clustering (2) and diakoptics (1,13) to various
forms of graph dissection (3,4,7). Despite their obvious differences, all these methods have a common goal in identifying a
minimal border that induces a block-diagonal structure in the
rest of the matrix. The border size is an important issue in
the solution of the system in Eq. (1), since the computation
time depends heavily on the number of nonzero elements in
the border.
In this section, we will briefly describe three methods for
obtaining BBD structures. We begin our analysis with the
classic nested dissection algorithm of George and Liu (3).
Nested Dissection
Nested dissection is a typical representative of BBD decomposition methods. It is relatively simple algorithm, in which the
matrix is assumed to be structurally symmetric and can
therefore be represented by an undirected graph. The following example illustrates the decomposition procedure.
Example 4. Let us consider the structurally symmetric matrix
1 2
0
1
2B
BB
3 B0
B0
4B
B
5B
BB
6 @
7
3 4
6 7
0
0
0
0
0
0
0
0
0
1
CC
CC
CC
CC
CC
A
(15)
5
1
Level 0
6
4
7
Level 1
Level 2
0
0
0
0
Level 3
Level 4
6 3
----------------------
-------------------
(16)
-----------------------0 0
The nested dissection algorithm was found to be very successful for matrices with a regular structure (such as those
arising in the numerical solution of partial differential equations using finite elements). However, this decomposition is
much less effective for systems such as electric circuits, where
the matrix structure is typically highly irregular (4). Several
different solutions where proposed to alleviate this problem,
dating back to the work of Sangiovanni-Vincentelli et al. (2).
We now describe two recent decomposition algorithms, which
have been successful over a wide range of matrix structures.
Decompositions Using Eigenvectors of Graphs. The idea of
eigenvector decompositions was introduced by Pothen et al.
(5), and is based on the Laplacian matrix of a graph. This
matrix is defined as
QDA
7 2
0
1
5 0 0 0 0
1B
BB 0 0 0 CCC
6 B 0 0 0 C
B0- - - -0- - - -0- - - -- - - -- - -- - - CC
3B
B
C
4B
BB0 0 0 CCC
7 @ 0 0 0 0A
2
213
(17)
where A is the adjacency matrix of the graph, and D is a diagonal matrix whose entries represent vertex degrees. It can be
shown that matrix Q is always positive semidefinite, with at
least one zero eigenvalue; the smallest positive eigenvalue of
Q is denoted 2, and the corresponding eigenvector is denoted
by X2. The decomposition procedure can now be summarized
as follows:
1. Compute eigenvector X2, and determine its median component xl.
2. Partition the vertices of the graph in the following way:
for any vertex i, if xi xl, set i A; otherwise, i B.
In this way, the vertices will be partitioned into two
approximately equal sets, A and B.
3. All the edges connecting sets A and B constitute an edge
separator, H. The objective now is to find a minimal vertex cover for H (that is, the minimal number of vertices
that need to be removed so that all edges in set H are
removed). This vertex cover constitutes the separator,
which appears in the border of the BBD structure.
4. Repeat steps 13 on the remaining components after
the separator is removed.
Decompositions based on eigenvectors of graphs were
found to be effective and applicable to a wide range of matrix
structures. However, for large matrices computing the second
eigenvector can be difficult, if not impossible. This consideration has motivated the development of the balanced BBD
decomposition (11,16), which is described next.
214
u1
u2
1000
2
k
l
2000
3000
1000
2000
3000
4000
5000
Figure 10. A model for the US electric power network (5300 5300).
To introduce the decentralized control problem, let us consider the two inverted pendulums interconnected by a spring
as shown in Fig. 12. The control objective is to keep the pendulums upright by using the inputs u1 and u2. After linearization around the equilibrium state 1 2 0, the equations
of motion are
ml 2 1 = mgl1 ka2 (1 2 ) + u1
ml 2 2 = mgl2 ka2 (2 1 ) + u2
(18)
-----------------
5000
-----------------
4000
0
1
0
0
0
0
2
g
1
ka2
ka
0
0
0
2
2
2
ml
ml
l
ml
S:
x = - - - - - - - - - - - - - - - - - - - - - - - - - - - - - x + - - - - - - - - - - - - u
0
0
0
1
0
ka2
ka2
g
0
0
0
ml 2
ml 2
l
ml 2
(19)
In choosing the feedback control laws to achieve the objective, the information structure constraint is essential: Each
input u1 and u2 can depend only on local states x1 (1, 1)T
and x2 (2, 2)T of each individual pendulum. In other words,
u1 = u1 (x1 ),
u2 = u2 (x2 )
(20)
u2 = kT2 x2
(21)
k2 = (k21 , k22 )T
(22)
need to be selected to stabilize S, that is, to keep the pendulums in the upright position. When the two pendulums are
considered as two subsystems
0 1
0
S1 :
x1 =
x +
u
0 1
1
(23)
0 1
0
S2 :
x2 =
x2 +
u2
0
2000
3000
4000
5000
0
1000
2000
3000
4000
5000
By choosing the decentralized control of Eq. (21), we effectively intend to stabilize the interconnected system S by stabilizing the two subsystems in Eq. (23). This turns out to be
a robust control strategy, since it can produce an overall
closed-loop system that can remain stable despite the essential uncertainty about the height a of the spring. This uncertainty is represented by the normalized interconnection parameter e, which can take any value between 0 and 1. When
an interconnected system is stabilized in this way, it is said
to be connectively stable (6).
By using the control law of Eq. (21) in Eq. (23) we obtain
the closed-loop subsystems as
0
x1 =
k11
0
x2 =
k21
S 1 :
S 2 :
1
x
k12 1
1
x
k22 2
(25)
1
0
0
0
x +e
x +e
k12 1
0 1
1
0 0
0
x1 + e
x2 + e
0
k22
:
S
0
x1 =
k11
0
x2 =
k21
Rn = Rn 1 Rn 2 Rn N
Rm = Rm 1 Rm 2 Rm N
Rl = Rl 1 Rl 2 Rl N
x = AD x + BD u + AC x + BC u
y = CD x + CC x
0
x
0 2
0
x
0 2
(26)
y = Cx
xi = Ai xi + Bi ui +
yi = Ci xi +
(Aij x j + Bij u j )
jN
Cij x j ,
AD = diag{A1 , A2 , . . ., AN }
BD = diag{B1 , B2 , . . ., BN }
iN
(33)
CD = diag{C1 , C2 , . . ., CN }
and the coupling block matrices are
AC = (Aij ),
BC = (Bij ),
CC = (Cij )
(34)
x = AD x + BD u
y = CD x
(35)
which is obtained from Eq. (32) by setting the coupling matrices to zero.
Important special classes of interconnected systems are input (BC 0) and output (CC 0) decentralized systems, where
inputs and outputs are not shared among the subsystems. Inputoutput decentralized systems are described as
x = AD x + BD u + AC x
y = CD x
(36)
as an interconnection
S:
(32)
where
S:
x = Ax + Bu
(31)
S:
(30)
215
(28)
u = KD x
jN
(37)
of N subsystems
Si :
xi = Ai xi + Bi ui
yi = Ci xi ,
iN
KD = diag{K1 , K2 , . . ., KN }
(29)
where xi(t) ni, ui(t) mi, yi(t) li are the state, input,
and output of the subsystem Si at a fixed time t , all ma-
(38)
which implies that each subsystem Si has its individual control law
ui = Ki xi ,
iN
(39)
216
with a constant gain matrix Ki. The control law u of Eq. (37),
which is equivalent to the totality of subsystem control laws
of Eq. (39), obeys the decentralized information structure constraint requiring that each subsystem Si is controlled using
its locally available state xi. The resulting closed-loop system
is described as
S :
x = (AD BD KD )x + AC x
(40)
yi and inputs ui, but are collectively responsible for the stabilization of S. The closed-loop system is
x
A BKDC BHD x
(46)
S&CD :
=
z
z
GD C
FD
The basic result of Ref. 17 states that the closed-loop system
S&CD is stabilizable by decentralized controllers Ci if and
only if the set of decentralized fixed modes
D =
(A BKDC) =
z i = Fi zi + Gi yi
ui = Hi zi Ki yi ,
(41)
iN
z = FD z + GD y
where
u=
y=
S&CD :
x
AD BD KDCD + AC
=
z
GDCD
BD HD
FD
x
(44)
z
S:
x = Ax +
B i ui
iN
yi = C i x,
(43)
(45)
iN
B i KiC i
(47)
iN
C j
1
Cj
= .2
.
.
C jNP
lies in the open left half plane. This result includes the interconnected system of Eq. (44) in an obvious way.
A simple characterization of decentralized fixed modes was
provided in Ref. 18. For any subset I i1, . . ., iP of the
index set N , let I C j1, . . ., jNP denote the complement of
I in N , and define
B I = [B i , B i , . . ., B i ],
z=
(42)
u = HD z KD y
K1 ,...,KN
KD
Ci :
C I c
(48)
ui =
hTi zi
kij yi
(50)
jJ i
where J i j : kij 1.
If we choose a gain matrix K to conform with the feedback
structure K, then the set
K =
(A BKC)
(51)
STABILIZATION
It has been common practice to use decentralized control in
the stabilization of large interconnected systems. Each subsystem is stabilized independently using local feedback, and
stability of the overall system is established using the MatrosovBellman concept of vector Liapunov functions. While a
vector Liapunov function may offer more flexibility and computational simplicity than a scalar one, it remains inherently
conservative. The use of vector functions, however, has been
justified by the presence of uncertainty in the interconnections, which can cause a breakup of the system during operation along the subsystem boundaries. The method of vector
Liapunov functions is a natural tool for making decentrally
controlled systems robustly stable to interconnection failures,
that is, connectively stable (6).
Vector Liapunov Functions
Let us assume that the plant is governed by linear time-invariant equations
S:
can be conveniently specified as the set of fixed modes with
respect to the decentralized feedback structure constraint defined by K. It is fairly easy to show (6) that the system S can
be stabilized by the constrained controller CK if and only if
K is contained in the open left half plane. To characterize
K as in Eq. (49), let us consider the index sets I M 1,
2, . . ., M and replace I C by J iI C J i, where I C now
refers to the complement of I in M .
In order to formulate graph-theoretic conditions for the existence of structurally fixed modes, let us recall that two systems are said to be structurally equivalent if they have the
same graph. Then, a system S is said to have structurally
fixed modes with respect to a given feedback structure K if
every system structurally equivalent to S has fixed modes
with respect to K. Structurally fixed is a generic concept:
Having structurally fixed modes is a property of a class of
systems sharing the same graph. Most importantly, if a system has no structurally fixed modes, then it either has no
fixed modes, or the fixed modes can be removed by arbitrarily
small perturbations of system parameters. This means that if
a system has no structurally fixed modes for a given K, then
(generically) it can be stabilized by a controller with a gain K
and structure K, decentralized structure being a special case.
It was shown in Ref. 21 that a system S has no structurally
fixed modes with respect to a feedback pattern K if and only
if:
217
xi = Ai xi + Bi ui +
N
eij Aij x j ,
iN
(52)
j=1
which are an obvious derivative of Eq. (28) save for the insertion of the interconnection parameters eij [0, 1]. The parameters are coefficients of the N N interconnection matrix
E (eij), which are used to model the uncertain strength of
interconnections.
To stabilize S we use the decentralized control laws of Eq.
(39) to stabilize each individual closed-loop system
:
S
i
xi = (Ai Bi Ki )xi ,
iN
(53)
This is presumably a relatively easy task, because the subsystems have low dimensions. Stability of each decoupled closedloop system follows from a subsystem Liapunov function
vi : ni . These individual functions are then stacked up
to form a vector Liapunov function v(x) [v1(x1), v2(x2), . . .,
vN(xN)]T. Finally, the vector Liapunov function is used to form
a scalar Liapunov function V : n of the form
V (x) = d T v
(54)
218
:
S
xi = A i xi +
N
for the system S of Eq. (27). By applying the epsilon decomposition to M and regrouping the variables, we get S as
eij Aij x j ,
iN
(55)
j=1
S:
xi = A i xi + B i u i +
N
(56)
wij =
eij ij ,
i = j
(57)
eij =
1 if S j acts on Si
0 if S j does not act on Si
(58)
S :
Epsilon Decompositions
Epsilon decompositions are ideal preconditioners for the stabilization of large-scale systems using decentralized control
and vector Liapunov functions (6,9). To see this, assume that
C 0 and form the augmented matrix
B
0
xi = A i xi +
N
A ij x j ,
iN
(62)
j=1
j = i
wij =
i ,
i= j
(63)
ij , i = j
(59)
A
M=
0
iN
c=W d
(61)
i eii ii , i = j
Aij x j + Bij u j
j=1
j = i
i Ai BiKi.
where A
Taking the total time derivative of V(x) with respect to S,
after lengthy but straightforward computations (6), we get
V (x) d T W z
(64)
xi = Ai xi + Bi
S:
ui +
N
eij Dij x j
iN
(65)
j=1
xi = Ai xi + bi ui +
eij Aij x j ,
iN
(66)
jN
(60)
mij =
n,
(67)
(mij 1) < 0
(68)
iI
jJ
xi = A i xi + hi (t, x),
iN
(69)
hi (t, x)
N
eij ij
x j
,
iN
(70)
219
ADAPTIVE CONTROL
Decentrally stabilizable systems with suitable interconnection structures can (always) be stabilized by using local states
or outputs, often employing high-gain feedback. How high the
gains should be depends on how strong the interconnections
are. More often than not, it is difficult, or impossible, to predict and limit the size of the coupling among the subsystems,
implying that fixed-gain controllers may be incapable in stabilizing the system. In such cases, one has to use adaptive controllers, which can adjust the gains to values needed for overall stability (66).
Let us consider a single-input, single-output version of S,
j=1
S:
xi = Ai xi + bi ui + Pi vi
yi = cTi xi
w i = Qi xi ,
(71)
iN
iN
(72)
220
hi (t, w)
N
ij
w j
(73)
j=1
where ij are nonnegative, but unknown, numbers. The matrices Ai and vectors bi and ci defining the subsystem Si are not
specified, except for the fact that pairs (Ai, bi) are controllable
and that pairs (Ai, ci) are observable. The uncertainty about
the triples (Ai, bi, ci) compounds the essential uncertainty
about the overall system S caused by our ignorance of the
interconnections.
The control objective is to force each state xi(t) of Si to track
the state xmi(t) of the corresponding reference model
Mi :
iN
(74)
Pi = bi pTi ,
iI
qi cTi ,
i J
Qi =
(75)
iN
(76)
iN
(77)
:
S
iN
(78)
S:
(79)
---------
------
-------
0
A11 A12 A13
B11
(80)
are decomposed along the dashed lines defining the two overlapping subsystems. By using a linear transformation
x = V x
where V is the n n matrix
I1
0
V =
0
0
0
I2
I2
0
(81)
0
0
0
I3
(82)
x = A x + Bu
:
S
(83)
B = V B + N
(84)
and
0
0
0
I1
U = 0 12 I2 21 I2 0
0
0
0
I3
1
1
A
2 A12
0
2 12
0
1
A
12 A22
2 22
M=
1
1
0 2 A22
2 A22
1
0 12 A32
2 A32
N=0
0
0
(85)
------------
------------
0
A13
0
A11 A12
B11
A
0
0
A
0
A
:
S
x = - - - -21- - - - - -22- - - - - - - - - - - - -23- x + - - - - - - - - - - - u (86)
A21
0
0
0
A22 A23
0
B32
A31
0
A32 A33
22
-------
21
-------
x + - - - - - - - - - - u
(87)
S:
x =
0
B
A
A
22
221
OPTIMIZATION
Despite considerable efforts and a large number of new results in the theory of large complex systems, the fundamental
problem of optimal decentralized control has remained unsolved for over two decades. The simple reason has been the
fact that the decentralized information structure constraints
have not been successfully incorporated into any of the standard optimization frameworks. Neither Pontryagins maximum principle nor Bellmans dynamic programming can handle the lack of complete state observation, a difficulty
recognized by Fleming (99) as far back as the late sixties. For
this reason, there have been a large number of results relying
on pragmatic suboptimality concepts, which have capitalized
on effective solutions of robustness issues in the suboptimality framework (6).
The standard practice has been to optimize each decoupled
subsystem using linear quadratic (LQ) control laws. Then,
suboptimality of the interconnected closed-loop system, which
is driven by the union of the locally optimal LQ control laws,
is determined with respect to the sum of the quadratic costs
chosen for the subsystems. Under relatively mild conditions,
suboptimality implies stability. Furthermore, the degree of
suboptimality, which is computed with respect to the globally
optimal union of decoupled subsystems, can serve as a measure of robustness with respect to a wide spectrum of uncertainties residing in both the subsystems and their interactions. It has been shown (100) how the classical measures of
gain and phase margins, as well as the gain reduction tolerance, can be incorporated in the decentralized LQ control of
large interconnected systems.
Recently, a solution to the optimal decentralized stochastic
control problem has been offered in the classical optimization
framework of Lagrange (101) relying on the constrained optimization approach proposed in Ref. 102. The principal idea is
to redefine the information structure constrains as differential equations, and attach Lagrange multipliers to each constraint to obtain sufficient as well as necessary conditions for
optimality of decentralized control laws. The multipliers are
functions of time and state, and, because of the role they play
in the optimization process, they are termed the Lagrange
Liapunov multipliers. The sufficient conditions are formulated in terms of the HamiltonJacobi-like equations and are
proved by using the method of global optimization (103). In
contrast, the necessary conditions of optimality, which are de-
222
S:
dxi =
n
Aij (t)x j
(88)
dt + Ci (t, x) dwi ,
iN
j = i
J=
N t1
1
E
[xT Q (t)xi + uTi Ri (t)ui ] dt
2 i=1 t 0 i i
(89)
iN
(90)
iN
(91)
223
51. M. Ikeda, Decentralized H controller design for large-scale systems, Proc. 35th IEEE CDC, Kobe, Japan, 1996, pp. 16.
52. H. H. Rosenbrock, Computer-Aided Control System Design, New
York: Academic Press, 1974.
53. M. Araki and O. I. Nwokah, Bounds for closed-loop transfer
functions of multivariable systems, IEEE Trans. Autom. Control,
20: 666670, 1975.
54. W. H. Bennett and J. S. Baras, Decomposition and decentralized
control system design: A review of frequency domain methods,
Proc. 24th IEEE CDC, Ft. Lauderdale, 1985, pp. 18281835.
55. Y. Ohta, D. D. Siljak, and T. Matsumoto, Decentralized control
using quasi-block diagonal dominance of transfer function matrices, IEEE Trans. Autom. Control, 31: 420430, 1986.
56. J. H. Seo, A frequency domain method for decentralized stabilization based on quasi-block diagonal dominance, Int. J. Control,
64: 4159, 1996.
57. A. N. Gundes and C. A. Desoer, Algebraic Theory of Linear Feedback Systems with Full and Decentralized Compensators, Heidelberg: Springer-Verlag, 1990.
zguler, Linear Multichannel Control: A System Matrix
58. A. B. O
Approach, UK: Prentice-Hall Int., 1994.
59. V. Manousiouthakis, On the parametrization of all decentralized stabilizing controllers, Sys. Control Lett., 21: 397403, 1993.
60. D. D. Sourlas and V. Manousiouthakis, Best achievable decentralized performance, IEEE Trans. Autom. Control, 40: 1858
1871, 1995.
61. M. Jamshidi, Large-Scale Systems: Modeling, Control, and Fuzzy
Logic, Upper Saddle River, NJ: Prentice-Hall, 1997.
62. J. Lunze, Feedback Control of Large-Scale Systems, Englewood
Cliffs, NJ: Prentice-Hall, 1992.
63. H. Tamura and T. Yoshikawa, Large-Scale Systems Control and
Decision Making, New York: Marcel Dekker, 1990.
64. C. T. Leondes (ed.), Control and Dynamic Systems, Orlando, FL:
Academic Press, 1986, vol. 22, 1985; vols. 23, 24.
65. M. S. Mahmoud, M. F. Hassan, and M. G. Darwish, Large-Scale
Control Systems, New York: Marcel Dekker, 1985.
66. D. T. Gavel and D. D. Siljak, Decentralized adaptive control:
42. R. J. Veillette, J. V. Medanic, and W. R. Perkins, Design of reliable control systems, IEEE Trans. Autom. Control, 37: 290
304, 1992.
224
tions on time-invariant and time-varying systems, ASME J. Dynam. Syst. Meas. Control, 116: 456473, 1994.
76. A. Datta, Performance improvement in decentralized adaptive
control: A modified model reference scheme, IEEE Trans. Autom.
Control, 38: 17171722, 1993.
DRAGOSLAV D. S ILJAK
ALEKSANDAR I. ZECEVIC
Santa Clara University
441
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
442
(y p Am y p )T Qi (y p Am y p ) + uT Ru dt,
Qi > 0
REFERENCE MODEL
Reference Model with Inputs
The use of a reference model for controls can be traced to
aircraft systems. Often, the situation therein is such that the
controls designer is sufficiently familiar with the plant to be
controlled and its desired properties; thus by choosing the
structure and parameters of a reference model suitably, its
outputs can be used as the desired plant response. While in
principle such a model can be linear or nonlinear, considerations of analytical tractability have made linear reference
models more common in practice.
Explicit and Implicit Model-Following
Two methods that have been studied extensively in this context include explicit and implicit model-following methods (2),
both of which include the use of a reference model described
by the homogeneous differential equation
ym = Am ym
(1)
(2)
Ie =
0
[(y p ym )T Qe (y p ym ) + uT Ru] dt
ym = Cm xm
(3)
where r is a bounded piecewise continuous function, determine the control input u(t) for all t t0 so that
lim |y p (t) ym (t)| = 0
443
C( )
(4)
2. Determine a differentiator-free controller C() parameterized by a vector (t) m, which generates u, such
that for a constant value *, the transfer function
of the plant together with the controller is equal to the
transfer function of M.
3. Determine a rule for adjusting (t) such that the closedloop system is stable and Eq. (4) is satisfied.
When a disturbance is present, such an asymptotic output
tracking may not be possible if very little information is available about the disturbance. In such a case, the goal of MRAC
is to minimize the error between yp and ym as much as possible.
Since stability is vital to the satisfactory operation of any
dynamic system and since in general adaptive systems are
nonlinear, one of the major difficulties in designing adaptive
systems is ensuring their stability properties. This often
serves as a guideline while solving the MRAC problem
stated above.
Error Model Approach
The solution of the MRAC problem is often significantly facilitated by an error model approach. This approach consists of
studying the relationship between two kinds of errors commonly present in any adaptive system: (1) the tracking error
e between the plant output and the model output and (2) the
parameter error between the estimated adaptive parameter
and its desired value. If the evolution of the error e is determined by the differential equation
e(t)
= f 1 (e(t), (t), t)
(5)
then the MRAC problem can be formulated as the determination of the adaptive law
(t) = f 2 (e(t), t)
(6)
444
in such a way as to ensure closed-loop stability and asymptotic tracking. Focusing attention directly on the error rather
than on the actual response of the plant or the reference
model enables the designer to concentrate on the essential
features of the problem and determine the adaptive law by
inspection. Such an approach has facilitated the design of
many adaptive systems both in the disturbance-free case as
well as when disturbances and modeling errors are present.
Z p (s)
R p (s)
kp
Wm (s)[ T ]
km
(8)
1 = e1 + e2
e2 = T Wm (s) Wm (s) T
Defining the filtered sensitivity function as where
= Wm (s)
one can show that the underlying error model, when kp is
known, is simplified from Eq. (8) to
Zm (s)
Wm (s) = km
Rm (s)
1 = T .
u = T (t)(t)
1 =
1 + u
2 =
2 + y p
= [r, 1T , y p , 2T ]T
= [k, 1T , 0 , 2T ]T
where (n1)(n1) is asymptotically stable, with det(sI
) Zm(s), (, ) is controllable, n is the control parameter to be adjusted appropriately so that (4) is achieved, and
is a sensitivity function which essentially estimates the state
of the system on-line. The requisite adaptive law, assuming
that km 0, is given by
= sign(k p ) e1 ,
e1 =
> 0
(7)
(9)
1 + T
(10)
445
matic tuning, as well as practical aspects of control implementation and applications, can be found in Ref. (15).
PARAMETER IDENTIFICATION
BIBLIOGRAPHY
e = Ae + bT ,
.
= e1
e1 = h T e
(11)
(12)
where
u( )u ( ) d I
T
t t0
(13)
for some constants t0, T0, and . Several statements equivalent to (13) can be given, one of which is that, for every unit
vector w in m, we obtain
1
T0
t+T0
t
|uT ( )w| d 0
6. A. S. Morse, Structure and design of linear model following systems, IEEE Trans. Autom. Control, 18: 346354, 1973.
7. S. H. Wang and C. A. Desoer, The exact model matching of linear
multivariable systems, IEEE Trans. Autom. Control, 17: 347
349, 1972.
8. G. Stein and M. Athans, The LQG/LTR procedure for multivariable feedback control design, IEEE Trans. Autom. Control, 32:
105114, 1987.
9. D. R. Mudgett and A. S. Morse, Adaptive stabilization of linear
systems with unknown high frequency gains, IEEE Trans. Autom. Control, 30: 549554, June 1985.
10. K. S. Narendra and A. M. Annaswamy, Stable Adaptive Systems,
Englewood Cliffs, NJ: Prentice-Hall, 1989.
11. M. Krstic, I. Kanellakopoulos, and P. V. Kokotovic, Nonlinear design of adaptive controllers for linear systems, IEEE Trans. Autom. Control, 39: 738752, 1994.
u=
1 + T
t t0
(14)
ANURADHA M. ANNASWAMY
Massachusetts Institute of
Technology
MULTIVARIABLE SYSTEMS
51
MULTIVARIABLE SYSTEMS
MULTIVARIABLE LINEAR SYSTEMS
Introduction
With the development of miniaturized, cheap, sensor, and actuator technology, many of todays control problems must coordinate the actions of multiple actuators, based on multiple
output signals from diverse sensors. Such systems, illustrated
in Fig. 1, are referred to as multi-input, multi-output (MIMO)
systems. An important class of MIMO systems is linear
MIMO systems, whose relationship between input and output
signals is represented by linear transformations.
This article introduces the basic concepts for analyzing
linear time-varying and time-invariant MIMO systems for
continuous time input and output signals. Extensions of the
concepts in this article to discrete time signals are straightforward and are found in the references at the end. The chapter discusses input-output and state-space models of linear
MIMO systems and introduces the concepts of controllability
and observability for state-space models. It also discusses
modal analysis for state-space models of time invariant
MIMO systems, MIMO poles and zeros, and singular-value
analysis for characterizing the frequency response of linear,
time-invariant MIMO systems.
Input-Output Models
Input-output models capture the essential relationships between inputs to a system and the outputs of that system. Instead of focusing on the internal representation and operation
of a system, input-output models represent these internal effects implicitly within a transformation from inputs to outputs, as illustrated in Fig. 1. In this section, we review results
on input-output models of linear MIMO systems in continuous time.
Consider a system with input u and output y, where the
relationship between input and output is denoted by the map
y1
System
...
um
...
u1
yp
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
52
MULTIVARIABLE SYSTEMS
y(t) =
H(t, )u( ) d
(5)
Another important property of the example is that the impulse response depends only on the difference t . This
property is known as time invariance.
Definition. The input-output system Eq. (1) is time-invariant if
H(t, ) = H(t , 0) H(t )
(6)
H(t )u( ) d =
H( )u(t ) d
(7)
H(t, )u( ) d
(1)
uk (t) = 0, k = j
(2)
then
yi (t) = hi j (t t0 ), i = 1, . . ., p
(3)
H(t, )
(t ) + 0.5 e(t ) u1 (t )
=
(t ) + e2(t ) u1 (t )
0
e3(t ) u1 (t )
so that
u ( )
y1 (t)
H(t, ) 1
=
d
y2 (t)
u1 ( )
In contrast with input-output models, state-space models provide an explicit representation of the internal operations of
a system, leading to the transformation between input time
functions and output time functions. The state of a system at
a given time provides a complete summary of the effects of
past inputs to the system, which is sufficient to uniquely characterize system output responses to future inputs. This article
focuses on linear MIMO systems where the state takes values
in a finite-dimensional space of dimension n. Such models are
analyzed with concepts from linear algebra and matrix differential equations.
The general form of a state-space model for a linear system
with m inputs u(t) Rm and p outputs y(t) is given by the
matrix differential equation
d
x(t) = A(t)x(t) + B(t)u(t)
dt
(8)
(9)
1
0
0
1 0
A = 0 2
0 ;
B = 1 0
0
0 3
0 1
0.5 0 0
1 0
C=
;
D=
0
1 1
1 0
(4)
MULTIVARIABLE SYSTEMS
(10)
starting from arbitrary initial conditions x(t0) at time t0. Under the assumption that the matrix A(t) is piecewise continuous, there exists a unique continuous solution of Eq. (10)
given by
x(t) = A (t, t0 )x(t0 )
(14)
(t )
e
A (t, ) = 0
0
(11)
53
0
e
0
0
2(t )
e3(t )
A (t, ) = eA(t )
0.5e(t )
e2(t )
0
e3(t )
i=0
(At)i
i!
=
(t, )
t A
= A(t)A (t, ).
Given the state transition matrix A(t, ), the general solution of Eq. (8) is written as
x(t) = A (t, t0 )x(t0 ) +
t0
t t0 (12)
y(t) =
(13)
0.5e(t )
e2(t )
e3(t )
u( ) d +
(t ) + 0.5e(t )
(t ) + e2(t )
0
e3(t )
1
1
0
u(t)
0
u( ) d
(15)
(16)
54
MULTIVARIABLE SYSTEMS
d
x(t) = x(t) + u(t)
dt
y(t) = u(t)
Controllability and Observability. The concepts of controllability and observability of state-space models characterize the
degree to which inputs and outputs determine the internal
state trajectory of a state-space model. This section presents
an overview of these concepts for linear, state-space models.
For a more detailed exposition of these concepts see CONTROLLABILITY AND OBSERVABILITY.
Consider a linear system with a state-space model whose
matrices are A(t), B(t), C(t), D(t), which are assumed to be
continuous functions of time. The system is said to be controllable on the interval [t0, t1] if, given any initial state x0 at t
t0 and any desired final state x1 t t1, it is possible to
specify a continuous input u(t), t0 t t1 so that if x(t0)
x0, then x(t1) x1. The system is observable on the interval
[t0, t1] if, given knowledge of u(t), t0 t t1 and y(t), t0 t
t1, the initial state x(t0) (and thus the entire state trajectory
x(t), t [t0, t1]) is uniquely determined.
Conditions for verifying controllability and observability
are determined from the explicit representation of the trajectories of state-space models in Eqs. (11) and (13). Using Eq.
(11), controllability is equivalent to finding a control u(t), t
[t0, t1], which solves
t
1
x(t1 ) A (t1 , t0 )x(t0 ) =
A (t1 , )B( )u( ) d
(17)
t0
=
t1
t0
A (t1 , )B( )B
( )TA (t0 , )
WC1 (t0 , t1 ) x(t0 ) A (t0 , t1 )x(t1 )
= A (t1 , t0 )WC (t0 , t1 )WC1 (t0 , t1 )[x(t0 ) A (t0 , t1 )x(t1 )]
= x(t1 ) A (t1 , t0 )x(t0 )
t1
t0
In this example, the state x does not affect the output y. Thus,
the system is internally unstable although it is BIBO stable.
t0
t1
z=
t0
t1
zT z =
t0
implying that z 0.
The controllability Gramian has many properties. It is
symmetric, positive-semidefinite for all t1 t0, and satisfies
the following matrix differential equation:
d
W (t, t1 ) = A(t)WC (t, t1 ) + WC (t, t1 )AT (t) B(t)BT (t)
(19)
dt C
WC (t1 , t1 ) = 0
Direct integration of this matrix equation is preferred to the
integral expression for numerically computing the controllability Gramian.
In the special case that the matrices A, B do not depend
on time, there is a simple algebraic test for controllability.
Define the matrix
Mc (A, B) = (B AB A2 B
An1 B)
t > t0
(20)
W0 (t0 , t1 ) =
t1
t0
The observability Gramian is again symmetric, positivesemidefinite, and satisfies the matrix differential equation
d
W (t, t1 ) = AT (t)W0 (t, t1 ) W0 (t, t1 )A(t) C(t)T C(t)
(21)
dt 0
W0 (t1 , t1 ) = 0
The system is observable on [t0, t1] if and only if W0(t0, t1)
is invertible. If the system is observable, then, in the absence
MULTIVARIABLE SYSTEMS
t1
t0
T ( , t0 )CT ( )y( ) d
In the special case where the matrices A, C are independent of time, observability is determined from the matrix
55
Mo (A,C) = Mc (AT , CT )T
The state-space model is observable if and only if the matrix
Mo(A, C) has rank n.
Theorem (3). There exists a state space realization of dimension n for the weighting pattern G(t, ) if and only if there
exists a p n matrix matrix function H(t) and an n m
matrix function F(t), both continuous for all t, such that
(22)
t0
d
x(t) = A(t)x(t) + B(t)u(t)
dt
y(t) = C(t)x(t) + D(t)u(t)
(23)
where x(t) Rn, u(t) Rm, y(t) Rp and A(t), B(t), C(t), D(t)
are continuous matrices.
The converse of the question is straightforward. Statespace models correspond to causal input-output models of the
form
t
t0
for all t, .
y(t) =
y(t) =
G(t, ) = H(t)F ( )
t t0
The proof of this theorem is straightforward. The invertibility properties of the state transitional matrix guarantee
that, for a state-space model, its input-output relationship is
factored as
C(t)(t, )B( ) = C(t)(t, 0)(0, )B( )
so that H(t) C(t)(t, 0), F(t) (0, t)B(t). Conversely, given
H(t), F(t), the state-space model
x(t)
= F(t)u(t)
y(t) = H(t)x(t)
(24)
56
MULTIVARIABLE SYSTEMS
d
x(t) = Ax(t) + Bu(t)
dt
y(t) = Cx(t) + Du(t)
(25)
uH
i vj = 0
Eigenvalues and Eigenvectors. Let Cn denote the space of ndimensional, complex-valued vectors. Consider a vector v
Cn of dimension n, expressed in terms of its real and imaginary parts as
whenever i j.
Now consider a matrix M Cnn with n distinct eigenvalues i, i 1, . . ., n, and, with corresponding left and right
eigenvectors ui, vi, i 1, . . ., n. It can be shown that the n
eigenvectors vi form a basis for the space Cn. In this case, the
matrix M is represented in terms of these eigenvalues and
eigenvectors in a dyadic expansion as
M=
a1
b1
v1
v = ... = ... + j ... = a + jb
vn
an
bn
where ai, bi, i 1, . . ., n are real-valued. Denote by v the
Hermitian of the vector v, defined as the complex conjugate of
the transpose of v:
vH = aT jbT
Given two vectors u, v C , the inner product of u and v
is defined as the complex-valued scalar product u, v uHv.
The Euclidean norm of a vector v Cn is given by
n
n
(26)
v2 = vH v =
i vi uH
i
i=1
ki vi uH
i
i=1
resulting from the orthogonality property of left and right eigenvectors corresponding to different eigenvalues.
System Modes. Consider the LTI state-space model in Eq.
(25). In the absence of inputs u(t), the response of the system
is determined by the initial condition x(0) and the system matrix A, as
x(t) = eAt x(0)
(a2i
b2i )
i=1
t2
+
2
n
n
n
(it)2
vi uH
=
(i t)0 vi uH
(i t)vi uH
i +
i +
i +
2
i=1
i=1
i=1
eAt = (At)0 + At + A2
e i t v i u H
i
i=1
Mv = v
x(t) =
and
uH M = uH
In the special case where the matrix M is real-valued (M
Rnn), if is an eigenvalue of M with a nonzero imaginary
part, then its complex conjugate is also an eigenvalue of M.
Because the previous equations are linear any multiple of
an eigenvector is also an eigenvector. Thus, eigenvectors can
be scaled to have any nonzero magnitude. In the rest of this
e i t vi [uH
i x(0)]
i=1
This is interpreted as follows: The initial condition x(0) is decomposed into its contributions along n different system
modes using the left eigenvectors. The ith system mode is defined as eitvi and has its own characteristic exponent i. When
the initial condition x0 corresponds to a right eigenvector vi,
the state response x(t) eitvi is focused along the same direction vi.
MULTIVARIABLE SYSTEMS
e i t [(Cvi )(uH
i x(0)]
i=1
n
(Cvi )(uH
i B)
i=1
e i (t ) u( ) d + Du(t)
1
1
2
A = 0 2
1
0
0 3
1 0
B = 1 0
0 1
0.5 0 0
C=
0
1 1
1 0
D=
1 0
The eigenvalues of A are 1 1, 2 2, 3 3. A set of
left and right eigenvectors is given by
The term uiHB indicates how the control action affects the ith
mode. Similarly, the term Cvi shows how much the ith mode
affects the system output y(t). Thus, modal analysis of LTI
systems decomposes the performance of MIMO systems into
a superposition of n independent modes which are exited by
the input signals and initial condition.
Based on Eq. (27), one can derive intuitive conditions for
controllability and observability of LTI systems using the system modes. In particular, note that the ith mode is uncontrollable if uiHB 0, because the input has no effect on the ith
mode trajectory. Thus, controllability requires that uiHB 0
for all modes i 1, . . ., n. Similarly, the ith mode does not
affect the output if Cvi 0. In this case, an initial condition
of vi yields an identical output to an initial condition of 0 and
thus is unobservable. Observability requires that Cvi 0 for
all modes i 1, . . ., n.
1
1
0
u1 = 1 , v1 = 0 ; u2 = 1 ,
1.5
0
1
1
0
0.5
v1 = 1 ; u3 = 0 , v1 = 1
0
1
1
57
y(t) =
H(t )u( ) d
(27)
x(t)est dt
(28)
(29)
k = 1, . . ., p
k=1
58
MULTIVARIABLE SYSTEMS
=0
0 0
C D
For a given transmission zero sk, the generalized eigenvector
associated with that transmission zero provides the initial
condition and input directions xk, uk which yield zero output
at that input frequency.
Example. Consider the MIMO LTI system described by
state-space matrices
1
0
0
A = 0 2
0
0
0 3
1 0
B = 1 0
0 1
0.5 0 0
C=
0
1 1
1 0
D=
1 0
The transfer function matrix for this two-input, two-output
system is given by
s+1.5
0
s+1
H (s) = s+3
1
s+2
s+3
Because A is diagonal, the poles of the system are easily determined as 1, 2, 3. Solving the generalized eigenvalue
problem, one obtains two transmission zeros, at frequencies
1.5 and 2. In particular, the input
1
u(t) =
e1.5t
4.5
with initial condition
2
x(0) = 2
3
MULTIVARIABLE SYSTEMS
0
x(0) = 1
1
sk kT = kT A + kT C
0 = kT B + BTk D
Consider an LTI MIMO system, specified by its transfer function matrix H (s). Assume that the system is bounded-input,
bounded-output stable, with no initial conditions. The transfer function H (s) can be interpreted as the complex gain of
the linear system in response to bounded inputs of the form
est. That is, if the input is defined as u(t) ves0t for t 0 for
some complex number s0 with nonpositive real part and some
direction vector v Rm, the output y(t) is given by
y(t) = H (s0 )ves 0 t
The frequency response of the system is the set of transfer
functions H ( j) for all frequencies R. Thus, the frequency response of the system defines the outputs corresponding to sinusoidal inputs of the form e jt.
In single-input, single-output (SISO) systems, the transfer
function is a scalar. Thus, the frequency response is characterized by the complex-valued function H ( j), which is represented by a magnitude and phase. In contrast, the frequency
response of MIMO systems is a complex, matrix-valued function of the frequency, which has a range of gains, depending
on the direction a of the sinusoidal input. To understand how
to represent this effect, it is useful to review some concepts of
gains for complex-valued matrices.
Complex Matrices and Gains. At a specific frequency, the
transfer function matrix H ( j) is a complex-valued matrix of
dimension p m. Denote by Cpm the space of complex-valued
matrices of dimension p m. Any matrix M Cpm, is decomposed into its real and imaginary parts, as
M = A + jB
where A, B Rpm. In a manner similar to a vector, the Hermitian of a matrix is defined as the complex conjugate of its
transpose, that is,
M H = AT jBT
d
x(t) = AT x(t) + CT u(t)
dt
y(t) = BT x(t) + DT u(t)
59
(30)
M2 = maxMv2
v 2
60
MULTIVARIABLE SYSTEMS
(Mu)H Mu = uH (M H M)u = uH u
.
=
..
0
0
..
.
k
0k(mk)
0 ( pk)(mk)
1
0
VHu =
..
.
0
i = 1, . . ., k
0
2
..
..
.
.
0
0 ( pk)k
MULTIVARIABLE SYSTEMS
Mu2 m 0
If the rank of M is m, the lower bound is strictly positive.
When the number of outputs is less than the number of inputs (p m), M must have a nontrivial null space, and thus
the lower bound is always 0.
Singular-Value Representation of MIMO Frequency Response. Now consider the MIMO frequency response of a
bounded-input, bounded-output stable system with transfer
function matrix H (s). Assume that the number of inputs m is
less than or equal to the number of outputs p. When the input
vector is a complex exponential of the form u(t) aejt, the
output vector is given by
y(t) = H ( j)u(t) = H ( j)ae jt = be jt
for some complex vector b H ( j)a. A useful characterization of the MIMO frequency response is provided in terms of
bounds on the gain of complex matrix H ( j) as a function
of frequency.
For each frequency , the singular-value decomposition of
H ( j) is obtained as
10
0
Gain (dB)
10
20
30
40
50
60
100
101
102
Frequency (rad/s)
s+3
BIBLIOGRAPHY
H ( j) = U ( j)()V H ( j)
70
101
61
103
DAVID CASTANON
Boston University
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright 2007 John Wiley & Sons, Inc.
made to operate satisfactorily with the addition of a simple controller in the error channel, which is shown by the
transfer function Gc (s) in Fig. 1. Typical performance criteria, which the system may be required to meet, are that it
is stable, has zero steady-state error and a good response to
a step input, suitably rejects disturbances, and is robust to
parameter variations. Although one reason for using feedback control is to reduce sensitivity to parameter changes,
specic design techniques can be used to ensure that the
system is more robust to any parameter changes. If the process to be controlled is strongly nonlinear, then a nonlinear
controller will have to be used if it is required to have essentially the same step response performance for different
input step amplitudes. Some control systemsfor example, simple temperature control systemsmay work in a
limit cycle mode, so that in these instances the designer is
required to ensure that the frequency and amplitude variations of the controlled temperature are within the required
specications.
In the next section, we examine in some detail various
approaches which can be used for investigating the analysis and design of nonlinear systems. The rst topic discussed is the phase plane method, which can normally only
be used to investigate second-order systems. It is a useful
technique, since it can be used when more than one nonlinearity exists in the system and since many control problems, such as position control systems, can be modeled approximately by second-order dynamics. As mentioned previously, one specication for the design may be that the
system must be stable. For linear systems, assessment of
stability is a simple problem, but this is not the case for a
nonlinear system, even when it is as simple as that shown
in Fig. 1. Several absolute stability criteria exist for checking whether the system of Fig. 1 will be stable, and these
are discussed in more detail later. The criteria presented
are easy to use; and the circle criterion in particular, being
basically an extension of the Nyquist criterion, is easy to
implement and follow. A disadvantage, however, is that all
these criteria only produce sufcient conditions so that if
the condition is violated the system may still be stable.
To try to obtain an estimate of the possibility of this
being the situation, the describing function method has
been used by engineers for many years. The difculty with
the describing function approach, which approximates the
nonlinearity by its gain to a sinusoidal input, is that the
results are not exact. It does, however, enable the designer
to obtain more insight into the situation, and, of course, the
ideas can often be further checked by simulation. The describing function approach can also be helpful for system
design in terms of shaping the frequency response of the
system to produce a more stable situation or for indicating possible nonlinear effects which can be added in the
controller to counteract the nonlinear effects in the plant.
Describing functions for other than a single sinusoid can
be obtained, and these allow some of the more complex
aspects of the behavior of nonlinear systems to be investigated. These include, for example, synchronization and
subharmonic generation as well as estimating more accurately the frequency content of any limit cycle. Relay-type
characteristics are often introduced in control system to
provide economic designs or to produce variable structure
systems. First, a method for the determination of limit cycles in relay systems is presented. This is an interesting
approach, since it allows the exact evaluation of a limit cycle and also an exact determination of whether it is stable
or not. The method in this sense is unique, since exact limit
cycle data for systems with any order dynamics containing
a relay can be obtained.
As with the design of linear control systems, the issue of
robustness to unmodeled dynamics and parameter uncertainty is also pertinent in the nonlinear control area. One
such robust technique is the so-called variable structure or
sliding mode approach. Variable structure control systems
(VSCS) are characterized by a set of feedback control laws
and an associated decision rule or switching function. This
decision rule has as its input some measure of the current
system behavior and produces as an output the particular
feedback control law which should be used at that instant
in time. The resulting variable structure system (VSS) may
be regarded as a combination of subsystems, where each
subsystem has a xed control law which is valid for specied regions of system behavior. One of the advantages of
introducing this additional complexity into the system is
the ability to combine useful properties of each of the subsystems into a very exible closed-loop control strategy. Indeed, it will be seen that a VSS may be designed to possess
new properties which are not present in any of the composite structures. Utilization of these natural ideas began
in the late 1950s in the Soviet Union and formed the foundations for signicant contributions to the area of robust
nonlinear control.
Of particular interest in the area of VSS is the so-called
sliding mode behavior, where the control is designed to
drive and then constrain the system state and lie within
a neighborhood of the switching function. There are two
signicant advantages with this approach to controller design. First, the dynamic behavior of the system may be
tailored by the particular choice of switching function. Second, the closed-loop response becomes totally insensitive to
changes in certain plant parameters and will completely
reject a particular class of external disturbances. This invariance property clearly renders sliding mode control a
strong candidate for robust control. In addition, the ability
to specify performance directly makes sliding mode control
attractive from the design perspective. This is seen from
the wide exposure of sliding mode control to many applications areas including robotics, aerospace, and automotive
industries.
The sliding mode design approach involves two stages.
The rst consists of the design of an appropriate switching
function to ensure that the system behavior during sliding
motion satises the required design specications. This is
termed the existence problem. In the simplest case, this
will be seen to amount to the design of a linear full-state
feedback controller for a particular subsystem. The second
design stage is concerned with the selection of a control law
which will make the switching function attractive to the
system state. This is termed the reachability problem. It is
important to note that this control law is not necessarily
discontinuous in nature.
and if this is rearranged as two rst-order equations, choosing the phase variables as the state variablesthat is,
x1 = x, x2 = xit can be written as
where x10 and x20 are the initial values of x1 and x2 . Since
Eq. (6) describes a parabola, which for the special case
of K = 0 has the solution x2 = x20 , it is easy to calculate the systems response from any initial condition (x10 ,
x20 ) in the phase plane. Figure 4 shows the response from
(4.6, 0) with = 1 and K = 1.25. The initial parabola meets
the rst switching boundary at A; the ensuing motion is
horizontalthat is, at constant velocityuntil the second
switching boundary is reached at B. The resulting parabola
meets the same switching boundary again at C, at which
point motion from either side of the switching line through
C will be directed toward C, so that the resulting motion is
a sliding motion. Responses from any other initial conditions are obviously easy to nd, but, from the one response
shown, several aspects of the systems behavior are readily apparent. In particular, the system is seen to be stable
since all responses will move inward, possibly with several
overshoots and undershoots, and will nally slide down a
switching boundary to 1. Thus a steady-state error of unit
magnitude will result from any motion.
When the velocity-feedback signal saturatesthat is,
when |x2 | > hthe input signal to the relay is x1 h. The
switching boundaries change to those shown in Fig. 5, but
the equations describing the motion between the boundaries remain unaltered. Therefore for a large step input
the response will become more oscillatory when the velocity saturates. When the hysteresis is nite then the switch-
where the nonlinearity n(x) lies in the sector (0, k). The
theorem has the simple graphical interpretation shown in
where
and
where
where
The function p(x) is the amplitude probability density function of the input sinusoidal signal and is given by
ideal saturation characteristic shown in Fig. 9 the nonlinearity output waveform y() is as shown in the same gure.
Because of the symmetry of the nonlinearity the fundamental of the output can be evaluated from the integral
over a quarter period so that
a
n
The integral n =
x p(x) dx is known as the nth moment of the probability density function; and for the sinusoidal distribution with p(x) = (1/)(a2 x2 )1/2 , n has the
value
n (x)(a2 x2 )1/2 dx
a1 = (4/a)
if
n(0) = 0
as before.
Therefore a1 = (4/a) 34 a4 = 3a3 /4, as before.
Saturation Nonlinearity. The DF can also be found by taking the nonlinearity input as a sin , in which case for the
valued or whether G(j) is available from a transfer function G(s) or as measured frequency response data. Typically the functions G(j) and N(a) are plotted separately
on Bode, Nyquist, or Nichols diagrams. Alternatively, stability criteria such as the HurwitzRouth or root locus plots
may be used for the characteristic equation
Thus
For the alternative approach, one must rst obtain the inphase and quadrature nonlinearities which are shown in
Fig. 11. Using Eqs. 23 and 24, one obtains
Figure 11. In-phase and quadrature characteristics for the relay of Fig. 10.
evaluated from
where Ni (a) is known as the incremental describing function (IDF). For a single-valued nonlinearity, Ni (a) can be
10
to have a root with a positive real part; that is, K > 1. G(j)
has 180 phase shift when = 1 when its gain is K. Thus
the DF solution for the amplitude of the limit cycle is given
by
which results in
giving
As K is increased, because of the shape of the nonlinearity, the limit cycle becomes more distorted. For example, if
K = 2.4, the DF solution gives = 1 and a = 2.10, whereas
if four harmonics are balanced, which requires a computer
program, the limit cycle frequency is 0.719 and the amplitudes of the fundamental, third, fth, and seventh harmonics at the input to the nonlinearity are 2.515, 0.467, 0.161,
and 0.065, respectively.
As the DF approach is a method for evaluating limit cycles, it is sometimes suggested that it cannot be used to
guarantee stability of a feedback system, since instability
may be exhibited by a signal in the system becoming unbounded, not oscillatory. It is, however, known for the autonomous feedback system of Fig. 1 that if the symmetric
odd, single-valued nonlinearity n(x) is sector-bounded such
that k1 x < n(x) < k2 x for x > 0 and n(x) tends to k3 x for large
x, where k1 < k3 < k2 , then the nonlinear system is either
stable or possesses a limit cycle, provided that the linear
system with gain K replacing N is stable for k1 < K < k2 .
Thus for this situation, which is often true in practice, the
nonexistence of a limit cycle guarantees stability.
Accuracy of the Describing Function
Since the DF method is an approximate analytical approach, it is desirable to have some idea of its accuracy.
Unfortunate consequences may result if a system is predicted to be stable, and in practice this turns out not to be
the case. Although many attempts have been made to nd
solutions for this problem, those that have been obtained
either are difcult to apply or produce results which are
often as conservative as the absolute stability criteria discussed earlier.
Since, as has already been shown, the C(a) locus of a
sector bounded, single-valued nonlinearity is the diameter of the circle in the circle criterion, errors in the DF
method are related to its inability to predict a phase shift
which the fundamental may experience in passing through
the nonlinearity, rather than an incorrect magnitude of the
gain. When the input to a single-valued nonlinearity is a
sinusoid together with some of its harmonics, it is easy
11
2h 1
{sin(nt)cos[n(t t1 )]
where RP and IP denote the real and imaginary parts, respectively. The IP expressions give two nonlinear algebraic
equations which, if they have solutions, yield the unknown
parameters t and T of possible limit cycles. Using these
solution values, the corresponding relay input waveform
x(t) can be found, from which the RP conditions can be
checked, as can the continuity conditions
n=1(2)
12
where S Rmn is of full rank. This will dene the switching function. It should be noted that the choice of S need
not be restricted to a hyperplane and more general, nonlinear, possibly time-varying switching functions may be
chosen.
If there exists a nite time ts such that the solution to
Eq. (61) satises
where B2 Rmm and is full rank. The transformation matrix exists as B is full rank and can be readily found via QR
decomposition. By using the coordinate transformation x
T1 x then the nominal linear system can be written as
The hyperplane design problem can therefore be considered to be one of choosing a state feedback matrix M to
prescribe the required performance to the reduced order
subsystem dened by (A11 , A12 ). It can be shown that controllability of the nominal (A, B) pair is sufcient to guarantee controllability of the (A11 , A12 ) pair.
13
and thus
14
The dynamics in the sliding mode are thus completely invariant to the signal f. The system behavior will be determined entirely by Eq. (69) when sliding despite the presence of such matched uncertainty. Any unmatched uncertainty will affect the dynamics of Eq. (69), but such unmatched effects can be minimized by designing M such that
the subsystem of Eq. (69) is maximally robust.
In order to illustrate the theoretical concepts of VSCS,
consider the double integrator system
where x1 = y and x2 = y. This system is already in an appropriate regular form for sliding mode design. Consider
application of a negative feedback law
15
ate p(x) is used. When the input x is a sinusoidal or Gaussian signal, however, it can also be shown that the error
signal between the nonlinearity and linear gain outputs
that is, n(x) Keq xis uncorrelated with the input x (6).
Typically, when dealing with Gaussian inputs to a simple
nonlinear feedback system, the mean-squared values of the
signals at various points in the loop can be calculated approximately using the RDF for the nonlinearity.
In many feedback systems it may be necessary to take
account of bias, as well as other signals, due to constant input or disturbance signals or because of asymmetry in nonlinear characteristics. In this case the nonlinearity, again
using the minimum mean-squared error denition, may
be modeled by two DFs, one for the bias, , and one for the
other input (6, 7). When the other input is considered as a
sinusoid of amplitude a, then the two DFs for the singlevalued nonlinearity n(x) are given by
and
the former being the DF for the sinusoid and the latter for
the bias. Here p(x) = 1/(a2 x2 )1/2 for the sinusoidal signal.
Use of this DF allows, amongst other possibilities, for the
determination of limit cycles with bias. For example, if in
Fig. 1 the input r(t) has a constant value R, then balancing
the bias and fundamental of the limit cycle gives the two
equations
mean, is given by
a formula which can be used for a single-valued nonlinearity with sinusoidal or Gaussian inputs when the appropri-
The equations can be solved to nd the bias and sinusoidal amplitude a of the limit cycle at the input to the
nonlinearity.
The above approach can be used in principle to obtain a
DF representation for a nonlinearity whose input consists
of any number of uncorrelated signals, but for practical
reasons the approach is difcult to justify for more than
two or possibly three components. A difculty in applying
such multiple input describing functions is understanding
the errors which are caused by neglecting not only higher
harmonics of the input signals but also cross-modulation
products which may be produced at frequencies lower than
those in the nonlinearity input signal.
Reasonably successful use, however, of the DF approach
for two related frequency sinusoidal inputs to a nonlinearity has been achieved to give results of some value in
control system design. This requires consideration of two
inputs such as a cos t and b cos(3t + ) so that the describing functions for the two signals become functions of
the three parameters a, b and , not just a and b (6). Analytically, results can only be obtained for simple nonlinearities such as a cubic; but by using computational methods, other characteristics can be considered (11). This procedure has been used to investigate subharmonic oscillations and synchronization phenomena when the feedback
loop of Fig. 1 has an input r(t) which is sinusoidal, and it
16
17
BIBLIOGRAPHY
1. K. S. Narendra, J. H. Taylor, Frequency Domain Criteria for
Absolute Stability, New York: Academic Press, 1973.
2. V. M. Popov, Absolute stability of nonlinear control systems of
automatic control, Autom. Remote Control, 22: 857858, 1962.
18
DEREK ATHERTON
SARAH SPURGEON
School of Engineering
University of Sussex Brighton,
England, BN1 9QT
Department of Engineering
University of Leicester
Brighton, England, BN1 9QT
519
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
520
(1)
y = h(x)
(2)
(3)
yd = h(xd )
(4)
Then the tracking control problem consists in finding, if possible, a smooth control law u such that for any bounded initial
conditions (t0, x0), x0 x(t0),
lim y(t)
, t
lim (y(t) yd (t)) = 0
(5)
(6)
y = Cx
(8)
(9)
(10)
where L is an output-injection matrix. Note that the estimation error x x x dynamics has the form
x = (A LC)x
u = K x
(12)
where x follows from Eq. (9). To find an answer to this question we write the overall controlled system plus observer:
x
A BK
=
0
x
BK
x
A LC x
(13)
BK
sI A + LC
sI A + BK
p(s) = det
0
or, equivalently, the roots of
(7)
521
(11)
It is clear that one can find an L such that Eq. (11) be asymptotically stablethus x 0 as t if the pair (A, C) is
detectable (10).
At this point we have successfully designed an asymptotically stabilizing feedback and a asymptotically converging observer. The natural question which arises now is whether it
is possible to use the state estimates in the feedback Eq. (7)
(14)
(x0 )
= 0
(15)
z = S(x),
(16)
such that the closed-loop system [Eqs. (14)(15)] in the z coordinates is a controllable system and without loss of generality
may be assumed to be in the Brunovski form:
0
z1
.
.
d
.=
.
dt .
.
0
zn
1
..
.
z1
0
.
. .
. .
+ . v
.
0
1 ..
1
0
zn
..
(17)
522
z 1 =
S1 (x)
S1 (x)
(x) x =
(x) f (x)
x
x
S1 (x)
+
(x) g(x)u = z2 = S2 (x)
x
(18)
Defining Lf S(x) S(x)/x f(x) as the directional or Lie derivative of function S(x) in the direction of f, we obtain from Eq.
(18) that
S2 (x) = L f S1 (x)
(19)
0 = Lg S1 (x)
(20)
i = 1, . . . , n 1
(21)
0 = Lg Si (x),
i = 1, . . . , n 1
(22)
(23)
(25)
(26)
[X1 , X2 ](x) =
n2
i (x)ad if g(x)
1=0
(27)
x2
mgl sin x k (x x )
1
1
3
I
f (x) =
x4
k
(x1 x3 )
J
0
0
g(x) =
0
1/J
(28)
One may verify that conditions (1) and (2) for feedback linearization are fulfilled about the equilibrium x0 0. In order to
find the linearizing coordinate change [Eq. (16)] and feedback
[Eq. (15)], we first solve Eq. (24). Using Eq. (28), this set of
partial differential equations reads as
S1 (x)
S1 (x)
S1 (x)
=
=
=0
x2
x3
x4
having as a (nonunique) nontrivial solution
z1 = S1 (x) = x1
(29)
(30)
z3 = S3 (x) = L f S2 (x) =
mgl
k
sin x1 (x1 x3 )
I
I
(31)
z4 = S4 (x) = L f S3 (x) =
mgl
k
x cos x1 (x2 x4 )
I 2
I
(32)
IJ
k
(x) = (Lg S4 (x))1 L f S4 (x)
IJ mgl
mgl
k
=
sin x1 x22 +
cos x1 +
k
I
I
I
k
k
mgl
k
+ (x1 x3)
+ +
cos x1
I
I
J
J
T
T
H(T ) H(0) +
i u(t)2 dt +
o y(t)2 dt
0
0
stored energy
dissipated
where qp n is the vector of link positions (generalized position coordinates), D(qp) D(qp) 0 is the inertia matrix,
C(qp, qp) is the Coriolis and centrifugal forces matrix, g(qp) is
the gravitational vector, and u n is the vector of control
torques.
The tracking control problem for system [Eq. (33)] is to
make the link position qp follow a desired trajectory yd(t)
qpd(t). The computed torque controller is a feedback linearization approach which, since it was proposed in Ref. 6, has become very popular. The control law is given by
u = D(q p )v + C(q p , q p )q p + g(q p )
(34)
u(t)y(t)dt
0
523
(35)
supplied
H(T ) +
i v(t)2 dt <
v(t)i(t) dt
0
o i(t)2 dt
0
that is, we can recover less energy than what was fed to the
circuit. Formally, the definition of passivity we will use is the
following (4):
Definition 1. Let T 0 be any. A system with input u
n
n
and output y L 2e
defines a passive operator : u y if
L 2e
there exists a such that
u(t)T y(t) dt
(36)
u(t)T y(t) dt o
y(t)2 dt +
(37)
T
0
u(t)T y(t) dt i
u(t)2 dt +
(38)
524
1
T (q, q)
= q T D(q)q
2
(40)
L (q, q)
=Q
q
(41)
where L (q, q)
T(q, q) V(q) is the Lagrangian function.
We assume that the external forces, Q n, are composed
only of potential forces (derived from a time-invariant potential V(q)) u n and dissipative forces F (q)/q, hence
Q=u
F (q)
F (q)
q
2
q
F (q)
q c 2
q
(46)
q j
qi
qk
(47)
With these definitions of matrices D(q) and C(q, q) the following properties hold:
P1. The matrix D(q) is positive definite, and the matrix
(q) 2C(q, q) is skew symmetric, that is,
N(q, q) D
N N. Moreover, there exist some positive constants dm and dM such that
dm I < D(q) < dM I
(48)
kc > 0
(49)
(50)
(51)
(52)
F (q)
=u
q
(43)
(44)
1
2
ci jk (q)q i
An EL system is underdamped if
q T
n
(42)
as q
col[qp qc] where we call qp the undamped coordinates
and call qc the damped ones. With this notation we can distinguish two classes of systems: An EL system with parameters
[Eq. (39)] is said to be a fully damped EL system if ( 0)
q T
(39)
where q n are the generalized coordinates and n corresponds to the number of degrees of freedom of the system.
We focus our attention on fully actuated EL systemsthat is,
systems for which there is a control input available for each
generalized coordinate. Moreover, we assume that the kinetic
energy function is of the form
Ck j (q, q)
ci jk (q)
d
dt
where the entries of the matrix C(q, q) are called the Coriolis
and centrifugal forces; the kjth entry is
(45)
(53)
1. D(q)
D p (q p )
0
0
, where Dc (qc ) Rn c n c .
Dc (qc )
F (q)
2. q T
525
Tc (qc , q c ) Vc (qc , q p ) Fc (q c )
Dc (qc )q c + D c (qc )q c
+
+
=0
qc
qc
q c
(54)
Note that the potential energy of the controller depends on
the measurable output qp, and therefore qp enters into the
controller via the term Vc(qc, qp)/qc. On the other hand, the
feedback interconnection between plant and controller is established by
u=
Vc (qc , q p )
q p
(55)
, Tp (q p , q p ) + Tc (qc , q c ),
V (q) , Vp (q p ) + Vc (qc , q p ),
F (q)
, F p (q p ) + Fc (q c )
T (q, q)
Fc (q c )
q c 2
q c
526
(56)
= (qc + Bq p )
(57)
Tc (qc , q c ) = 0, Fc (q c ) =
(60)
1 T 1 1
q B A q c
2 c
= (q )T q +
p
p
1
Vc (qc , q p ) = (qc + Bq p )T B1 (qc + Bq p )
2
The controller above injects the necessary damping to achieve
asymptotic stability and its action has the following nice passivity interpretation. First, we recall that the EL plant Eq.
(41) defines a passive operator u qp. On the other hand,
the controller Eq. (54) defines a passive operator qp Vc(qp,
qc)/qp. These properties follow, of course, from the passivity
of EL systems established in Proposition 1. It suffices then to
choose u as in Eq. (55) to render the closed-loop passive.
(58)
= KI q p (0) = 0 Rn
(59)
2q p
1 + 2q p 2
(61)
(62)
q p
= = (q pd )T q p +
1 + q p
(63)
1 + q p
0
Note that the choice KP KP KI, with KI 1/ (qpd)
(qpd), yields the controller implementation
u = KP q p KD q p +
=
KI
q p
,
1 + q p
(64)
(0) = 0 Rn
(65)
(0) = 0 R
(66)
n
(67)
(0) = 0 Rn
(68)
(69)
q c = A(qc + Bq p )
(70)
= qc + Bq p
(71)
4dM
B>
I
dm
KP > (4kg + 1)I
(72)
(73)
col[qp, qp, ,
where kg is defined by Eq. (49), and define x
g(qpd)]. Then, given any (possibly arbitrarily large) initial
condition x(0), there exist controller gains that ensure
limt x(t) 0.
527
(74)
(75)
e = w + Ld (e e)
(76)
w = L p (e e)
(77)
Then for any set of bounded initial conditions (t0, x(t0)) we can
always find sufficiently large gains Kp, Kd, Lp, and Ld such
t t0 + t1
528
uniform ultimate boundedness result of Proposition 8 is of local nature because the controller gains depend on the initial
conditions x(t0); nevertheless, it is important to remark that
the Proposition states that for any set of finite initial conditions there always exist control gains . . .; that is, the result
is semiglobal.
However, even without the knowledge of bounds over the
plants parameters, the closed loop system can be made uniformly ultimately bounded by selecting the control gains sufficiently large. Hence there is no need to quantify these
bounds a priori.
It is also important to remark that the linear control
scheme [Eq. (75)(77)] allows quick response in an online implementation, due to its simplicity. Since this control scheme
completely ignores the system dynamics, however, the conditions of Proposition 8 (see [30] for details) may require that
Kd and Ld be large to obtain an acceptable tracking performance. Such high gain implementations are not always desirable in practical applications. For this reason it may be
profitable to add model-based compensation terms to the control input, when available. See, for instance, Ref. 17 and references therein.
in closed loop with Eq. (78) is GAS. Now, let us suppose that
only x2 and x3 are measurable. In this particular case, a nonlinear observer for x1 can be easily designed using the Lyapunov control approach (4). Motivated by the control law Eq.
(78)
x2 = x1 + ax2
(79)
x3 = x3 (x1 b) + c + u
(80)
(81)
u = x3 (x1 b) c k1 x1 k2 x2 k3 x3
(82)
1 2
x + x22 + x23 + x21 +
x2 (x3 x1 )
2 1
(83)
where x1 x1 x1 is the estimation error and 0 is sufficiently small to ensure that V(x) is positive definite and radially unbounded. Then, let V(x) be a Lyapunov function candidate for system Eq. (78) in closed loop with the control law
Eq. (82). We proceed now with the design of a reduced observer x1 f(x2, x3, x1). First, evaluating the time derivative
of V(x, x1) along the trajectories of the closed loop Eqs. (78),
(82) we get after some bounding
V (x, x1 ) 1 x21 2 x22 2 x23 + (x3 +
x2 )(k1 + x3 )x1 + x1 x1
x1 = (x3 +
x2 )(k1 + x3 )
the Lyapunov function becomes negative semidefinite. A simple analysis using the KrasovskiiLaSalles invariance principle shows that the closed-loop system is GAS. Note, moreover,
that the observer can be implemented as
x1 = (x3 +
x2 )(k1 + x3 ) (x2 + x3 )
without measurement of x1.
In the case when more variables are unmeasurable, one
may think that a similar procedure leads to the design of an
observer for the unavailable variables. Unfortunately, this
seems not the case when only x2 or x3 are considered available
for measurement. Moreover, the lack of a physical interpretation for the Rossler system makes this task more difficult.
The lesson one can take from this illustrative example is that
observer-based schemes become complicated even if the system itself is feedback-linearizable. The lack of (physical) passivity properties hampers the use of passivity-based control.
CONCLUSION
We have briefly illustrated different control design methods
for nonlinear systems. We derived necessary and sufficient
conditions to solve the local feedback linearization problem
and illustrated this approach on the flexible joints robots case.
We focused our attention into a special class of second-order systems, the EL systems. However, the Lagrangian formulation applies to all fields in which variational principles
can be used to model the plant in question; hence this class
includes a wide number of physical systems such as robot manipulators.
We saw that the EL class has some nice energy properties
which can be exploited by using passivity-based control. The
goal of this methodology is to design a controller and a dy-
529
ANTONIO LORIA
University of California
at Santa Barbara
HENK NIJMEIJER
University of Twente
19
Murphy Software
Company
1000 Town Center, Suite
1950
Southfield, MI 48075
Tel: 248-351-0900
Fax: 248-351-0906
ProModel Corporation
1875-T S. State Street,
Suite 3400
Orem, UT 84097-8075
Tel: 801-223-4600
Fax: 801-226-6046
Scientific and Engineering
Software
4301-T Westbank Drive
Austin, TX 78746-6546
Tel: 512-328-5544
Fax: 512-327-6646
Wolfram Research, Inc.
100-T Trade Center Drive,
Suite 600
Champaign, IL 61820-4853
Tel: 217-398-0700
Fax: 217-398-0747
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
20
j
s-plane
s1
C
s1
A
B
or
|c(t)|
|r(t )|| g( )| d
(3)
|r(t)| R
STABILITY CONDITIONS
(4)
|c(t)| R
| g( )| d
(5)
(6)
then
| g( )| d C <
(7)
| g( )| d Q <
(8)
For Eq. (8) to hold, the integral of the absolute value of g(t)
must be finite.
The Laplace transform may be used to show the relationship between the roots of the characteristic equation and Eq.
(8). For g(t)
L[ g(t)] =
g(t)est dt = G(s)
(9)
0
Where s is a complex number having real and imaginary components on the s-plane as s j. Taking the absolute
value on both sides of the equation yields
st
|G(s)| =
g(t)e dt
|g(t)est | dt
(10)
0
But since
|est | = |e t |
r(t)
g(t)
(11)
c(t)
| g(t)e t | dt
(12)
s1j
4
7
j
5
5
s1j
7
4
j
5
5
s 1 j
7
4
j
13
13
s 1 j
7
4
j
13
13
2
F(s)
s5j
5 j4
s5j
5 j4
s3j
3 j4
s3j
3 j4
s 3 j
4
3
j
25
25
s 3 j
3
4
j
25
25
s 5 j
5
4
j
41
41
s 5 j
5
4
j
41
41
s 7 3j
4
7
j
3
3
s 7 3j
7
4
j
3
3
s 7 3j
21
12
j
65
65
s 7 3j
21
12
j
65
65
F(s)
R| g(t)| dt =
| g(t)| dt
0
21
F(s)
4
F(s)
(13)
(14)
Note that Eq. (14) does not satisfy the BIBO relation because
the equation is not bounded as in Eq. (8). Hence, to satisfy
the BIBO stability, the roots of the characteristic equation or
the poles of G(s) must lie on the left side of the j-axis. A
system is classified to be marginally stable if the first-order
poles, poles of conjugate pairs, lie on the j-axis. However,
multiple-order poles or repeating conjugate pairs of poles represent an unstable system. In addition, a system is classified
as unstable if more than one pole exists at the origin.
Another definition that is worth mentioning and that helps
in the understanding of the Nyquist criterion is the steadystate error. Steady-state error is the difference between the
input and the output as t for a given test input.
The steady state errors are generally described for three
main types of test inputs: the step, the ramp, and the parabolic. Often, control systems are subjected to these inputs to
test their ability to give the required outputs. Usually, these
test inputs are in the form of electrical signals that have defined waveforms. For example, the parabolic input has a constant second derivative, which represents acceleration with
respect to accelerating targets. In general, the output of any
system can be represented by the sum of the natural response
and the forced response. In relation to the Nyquist stability
criterion, a steady state error can be calculated from the
closed-loop transfer function of the system M(s) or the loop
transfer function G(s)H(s).
F (s) =
(s + zi )
i=1
P
(15)
(s + pk )
k=1
F (s)
(s
+
z
)
(s
+
pk )
i
i=1
k=1
(16)
s+1
(s + 2)(s + 3)
(17)
s2 2s + 1
(s + 2)2 (s + 3)2
(18)
Calculate
THE PRINCIPAL ARGUMENT
The Nyquist criterion is based on a theorem using the theory
of complex variables, which leads to the principal argument.
The principal argument may be presented in a number of
ways. Here, two approaches will be presented.
CASE 1. In this case, a rigorous mathematical approach may
be employed by using theories of contours and mappings in
complex planes. Assume that F(s) is a function of s and singlevalued, that is, for each point in the s-plane there exist a corresponding point, including infinity, in the F(s)-plane, and the
function consists of a finite number of poles and a finite number of zeros in the s-plane. Now, suppose that there is an arbi-
F (s) =
(19)
s2 2s + 1
A
B
C
F (s)
=
=
+
+
F (s)
(s + 1)(s + 2)(s + 3)
(s + 1)
(s + 2)
(s + 3)
(20)
22
3.00
s-plane
s4
2.00
s3
1.00
s1
s2
0.
Re
1.00
F(s) = s + 4
s-4
2.00
3.00
7.00
5.00
3.00
1.00
1.00
3.00
5.00
7.00
f (s) ds =
1
ds = ln(s + a)|
(s + a)
(23)
Im
The right-hand side of this equation can be evaluated by substituting values of s(t) a ej yielding
8.00
F(s)-plane
6.00
F2
4.00
2.00
0.
(24)
F3
F1
F4
Re
2.00
4.00
6.00
F (s)
(s + 1)
(s + 2)
(s + 3)
k=
9.00
(21)
and verifying the general Eq. (16). Equations (16) and (21)
indicate that the zeros of F(s) appear as the denominators of
the new equation with positive signs. The poles of F(s) still
appear in the denominators, but they have negative signs.
After having stated this important point, we can turn our
attention to Cauchys theorem. Although Cauchys theorem
may be of great mathematical interest, the intention of this
article is not to discuss the intricacies involved in the theorem
but rather to use it to illustrate relevant points in order to
establish a solid understanding of Nyquist stability criteria.
Cauchys theorem is based on the complex integrals, mapping of points and closed contours between the s-plane and
1
ds = 2kj
(s + a)
(25)
On the contrast to Eq. (22) where f(s) is analytic in the splane, for f(s)) having a singular point in the contour in the
s-plane, the resulting closed contour is no longer zero but
equal to multiples of 2j. This indicates that the contour on
the f(s)-plane containing all the values of s on the s-plane goes
through k number of revolutions around the origin. The number of revolutions around the origin depends on the number
of times the point a is encircled in the s-plane. The encirclement k now can be expressed as
8.00
7.00 5.00 3.001.00
f (s) ds =
1
2 j
f (s) ds =
1
2 j
1
ds
(s + a)
(26)
f (s) =
Z
P
1
1
F (s)
=
F (s)
(s + zi ) k=1 (s + pk )
i=1
(27)
where f(s) can be viewed as the residues of F(s). Now, substitute this Eq. (27) into Eq. (25)
f (s) ds =
Z
i=1
1
ds
(s + zi )
P
k=1
1
ds
(s + pk )
(28)
Making use of the similarities between Eq. (23) and Eq. (28)
will yield the following conclusions. There will be Z number
of revolutions around the origin, one for each zi for k 1.
R(s) +
E(s)
G(s)
23
C(s)
H(s)
f (s) ds = Z P
(29)
(30)
Figure 4. Block diagram of a closed-loop system. A closed-loop system has a forward path transfer function G(s) and a feedback path
transfer function H(s). The relation between the input and output can
be expressed in terms of these two terms in the form of a system
transfer function to be used in analysis and design of systems.
NG (s)
DG (s)
(31)
H(s) =
NH (s)
DH (s)
(32)
and
then
G(s)H(s) =
where
N number of encirclements of the origin made by the
path F,
Z number of zeros of F(s) encircled by the path s,
P number of poles of F(s) encircled by the path s.
G(s) =
NG (s)NH (s)
DG (s)DH (s)
(33)
(34)
NG (s)DH (s)
G(s)
=
1 + G(s)H(s)
DG (s)DH (s) + NG (s)NH (s)
(35)
(s + z1 )(s + z2 ) (s + zZ )
, m, n {1, 2, . . .}
(s + p1 )(s + p2 ) (s + pP )
(36)
This is equivalent to
(s) = |(s)|\(s)
(37)
24
j
s-plane
s1
p2
z3
p2
z 3
p3
p
3
1
2 j
z
1
p
1
p1
1
f (s) ds =
2 j
F (s) ds
1
=
F (s)
2 j
d[ln F (s)] = Z P
(40)
z1
z2
z2
1
2 j
d[ln F (s)] =
1
2 j
where
|(s)| =
|s + z1 ||s + z2 | |s + zZ |
|s + p1 ||s + p2 | |s + pP |
(38)
(39)
j
+j
s = j
s-plane
RHP
LHP
RHP
LHP
s-plane
s = j
(a)
= Poles or zeros
R = Infinite radius
(b)
NZP
N0
N0
N0
Direction of
s-Plane Locus
Encircles
of the
Origin
Direction of
Encirclement
Clockwise
Counterclockwise
Clockwise
Counterclockwise
Clockwise
Counterclockwise
N
N
N
N
0
0
Clockwise
Counterclockwise
Counterclockwise
Clockwise
None
None
25
(42)
because F(s) and 1 F(s) always have the same poles. The
result is similar to the one derived earlier in the discussion.
The other stability requirements are that for the closed-loop
stability
Z1 = 0
(43)
(44)
(45)
26
(46)
(47)
Now the Nyquist criterion may also be stated in the following manner.
For a closed-loop system to be stable, the Nyquist plot of
F(s) must encircle the critical point (1, j0) as many times
as the number of poles of F(s) that lie in the right half of
the s-plane. The encirclements, if any, must be in the
clockwise direction when the Nyquist path is defined in
the counterclockwise sense.
1 = 2 N1,1 = 2 (Z1 P1 )
(50)
2 = 2 N1,2 = 2 (Z1 P P1 )
(51)
and
(48)
(49)
2 = 211 12 + 13
(53)
where
(54)
1 + 2 = 2 (2Z1 P 2P1 )
(55)
11 = (Z1 0.5P P1 )
(56)
since 13 is zero,
and
N1,2 = Z1 P P1
(52)
and
1 = 211 + 12 + 13
hence
Equation (56) means that the net angle traversed by the phasor from the (1, j0) point to the F(s) Nyquist plot corresponding to the positive j-axis of the s-plane excluding any of the
small indentations, that is
The number of zeros of 1 F(s) in the right half of the splane minus the sum of half the poles on the j-axis and
the number of poles of F(s) in the right half of the s-plane
multiplied by radians.
This means that the Nyquist plot can be constructed corresponding to s 0 to s j portion of the Nyquist path. For
an unstable closed-loop system, the number of roots of the
characteristic equation that fall in the right half of the splane can be determined via Eq. (55).
As mentioned earlier, a closed-loop system is stable only if
Z1 is equal to zero.
Hence,
11 = (0.5P + P1 )
(57)
SOLUTION. To obtain the Nyquist plot, rearrange this equation, substitute s j, and assign the nominal value, K 1,
G( j)H( j) =
(58)
(1 + j)(1 + 0.05 j)
j(1 + 5 j)(1 + 2.5 j)
(59)
1 + 2 1 + (0.05)2
|G( j)H( j)| =
1 + (5)2 1 + (2.5)2
(60)
and
(61)
Now Nyquist contour may be applied by substituting values of from zero to on the positive part of the imaginary
axis on the s-plane. By avoiding the pole located on the origin
and substituting small positive values on the positive and
negative side of zero, it is possible to observe the following
features:
0+
NYQUIST DIAGRAMS
The Nyquist analysis is based on the assumption that the
control systems are linear; hence, the dynamic performances
are described by a set of linear differential equations. Because
of the nature of feedback control systems, the degrees of numerator of the loop transfer function F(s) G(s)H(s) is always
less than or equal to the degree of the denominator. All the
Nyquist diagrams presented here are based on these two assumptions.
As explained previously, when plotting Nyquist diagrams,
it is sufficient to assign values for the complex variable s on
the j-axis avoiding possible poles and zeros on the imaginary
axis. The frequency response of G(s)H(s) can be determined
by substituting s j and by finding the imaginary and complex components of G( j)H( j). Alternatively, G( j)H( j)
can be written in polar form, and magnitudes and angles are
determined for plotting on a polar graph paper. These techniques will be illustrated in the following examples.
27
and
and
and
and
G( j)H( j) 90
G( j)H( j) 90
G( j)H( j) 90
G( j)H( j) 90
0.1
0.4
0.5
1.0
2.0
4.0
90
124
9.0
175
0.79
181.6
0.55
188.9
0.1
182.8
0.021
174.2
0.005
90
0
28
0.1
0.4
0.5
1.0
2.0
4.0
5.03
7.46
0.787
0.069
0.55
0.018
0.099
0.015
0.021
0.0001
0.005
0.0005
0
0
1 + 2 1 + (0.05)2
|G( j)H( j)| = K
1 + (5)2 1 + (2.5)2
(62)
(1 + 2s)
s(s 1)
(63)
Im
=0
G(s)H(s)-plane
R
1
Re
= 0+
Figure 7. The Nyquist plot of Example 1. As the Nyquist path on
the s-plane traverses in the clockwise direction, the corresponding
path of G(s)H(s) traverses in the same direction on the G(s)H(s)-plane.
Because the Nyquist plot does not encircle the critical 1 j0 point,
this system is stable.
(1 + j2)
j( j 1)
1 + 42
G( j) =
1 + 2
(64)
(65)
and
G( j) = tan1 2 90 tan1 /(1)
(66)
Lets consider the nominal value of K. Observing the extreme values for in the clockwise direction and starting
from 0 gives
(1 + j2)
32 + j( 23 )
(1 + j2)
= 2
=
(67)
j( j 1)
j
2 + 4
29
0.1
0.4
0.6
1.0
4.0
10.0
G( j)H( j)
G( j)H( j)
90
107.0
10.15
150.5
2.97
171.16
2.23
198.4
1.58
248.8
0.49
261.4
0.02
270
0
The graph crosses the real axis when the imaginary part is
equal to zero, that is ( 23) 0 or 1/ 2 rad/s. The
intersection can be calculated by substituting the value of
in the real component of G( j)H( j) as
3 (1/2)
32
= 2
=
2
4
+
(1/2) + (1/4)
(68)
There are many examples of Nyquist plots in control engineering books (15). Figure 9 illustrates typical Nyquist plots
of some of the selected control systems.
Stability Margins
In the preceding examples, we have demonstrated that the
Nyquist plots of the loop transfer function, G( j)H( j), depends on the values of K. This is illustrated in Fig. 10. As the
K is increased or decreased, as the case may be, at a certain
value the locus passes through 1 j0 point. At this point,
the system exhibits sustained oscillations. As K increases further, the system becomes unstable.
Generally, oscillations increase as the locus of G( j)H( j)
gets closer to the 1 j0 point. The closeness of the locus to
the critical point is measured by the stability margins usually
expressed in the form of phase and gain margins, as illustrated in Fig. 10. These margins indicate relative stability
and hence help the design of control systems to achieve desired responses. The gain and phase margins may be defined
as follows.
Gain margin is the amount of gain that can be allowed to
increase before the closed loop system becomes unstable.
Im
=0
G(s)H(s)-plane
Re
= 0+
Figure 8. The Nyquist plot of Example 2. The open-loop transfer
function of this control system has a pole on the RHP; hence, the
system is open-loop unstable. However, the Nyquist plot encircles the
critical 1 j0 point once in the counterclockwise direction, indicating that there are no closed-loop poles on the RHP. Therefore, the
system is stable.
1
dB
|G( jc )H( jc )|
(69)
(70)
Nyquist diagrams
0.6
K
G(s)H(s) =
(s1 + 1)
0.2
0.4
0.6
0.5
0
Real axis
0.5
0
0.2
0
2
4
0.6
6
0.5
0
Real axis
0.5
8
4
K
(s 1+1)(s2+1)(s3+1)
0.4
0.8
1
G(s)H(s) =
0.2
2
4
Real axis
(c)
(b)
(a)
Nyquist diagrams
Nyquist diagrams
1.5
1.5
1
Imaginary axis
0.8
1
Imaginary axis
Imaginary axis
0.4
G(s)H(s) =
K
(s 1+1)
0
0.5
0
0.5
1
1.5
1.5
1.5
2
2
K
(s 1+1)(s2+1)
0.5
G(s)H(s) =
0.5
2
0 0.5
2 1.5 1 0.5
Real axis
(d)
1.5
(e)
Nyquist diagrams
Im
10
8
6
Imaginary axis
Imaginary axis
G(s)H(s) =
K
(s1 + 1)(s2 + 1)
0.6
0.4
0.2
Nyquist diagrams
8
0.8
Imaginary axis
30
G(s)H(s) =
K (s z +1)
(s 1+1)(s2+1)
4
2
0
2
Re
4
6
8
10
2 1.5 1 0.5
0 0.5
Real axis
(f)
1.5
2
(g)
10
31
Im
Im
G(s)H(s)-plane
G(s)H(s)-plane
Unit circle
Re
1
PM
GM
Re
=0
Increasing
Increasing K
Figure 10. Phase and gain margins. The closeness of the locus to
the critical 1 j0 point is measured by the margins. Good gain and
phase margins are obtained for K1. As K increases, both gain and
phase margins become zero (for K2) indicating critical stability. Further increase in K leads to unstable conditions. Note the changes in
the gain and phase margins for varying K.
(71)
where T is the time delay. The term esT does not introduce
any additional poles or zeros within the contour. However, it
adds a phase shift to the frequency response without altering
the magnitude of the curve. This is because
Figure 11. The effect of time delays. Pure time delays do not introduce any extra poles and zeros into the system. However, the magnitude is equal to unity for all frequencies, the phase ( T) affects
the stability. For large values of time delay T, the system may be unstable.
G(z)
C(z)
=
R(z)
1 + GH(z)
Im z
G(z)H(z)-plane
(72)
(73)
0.0231
1
= 3.14
Re z
=0
Figure 12. Nyquist plot of Example 3. The Nyquist path on the zplane must have small indention at z 1 on the unit circle. The
Nyquist plot of path of GH(z) in the GH(z)-plane intersects the negative real axis at 0.231 when 3.14 rad/s. For stability, the value
of K must be less than 4.33.
32
Im
G(s)H(s)-plane
Inverse Nyquist is simply the reciprocal of the complex quantity in the Nyquist plot. They find applications particularly
in multiple loop and multivariable systems where graphical
analysis may be preferred.
The Nyquist stability criterion applied to inverse plots can
be stated as a closed loop system stable, if the encirclement
of the critical 1 j0 point by the 1/G(s)H(s) is in the counterclockwise direction for a clockwise Nyquist path in the splane. As in the case of a normal Nyquist, the number of encirclements must equal the number of poles of 1/G(s)H(s) that
lie in the right half of the s-plane.
Inverse Nyquist plots is particularly useful in the analysis
of multi-inputmulti-output control systems. In the multivariable feedback control systems, the relations between inputs and outputs may be expressed in matrix form as
Re
1.5
G(s) =
s3 + 2s2 +2s+1
1
4s+1
C(s) = [I + KG(s)H(s)]1G(s)KR(s)
x
H(s) = I
Figure 13. Examples of Nyquist plots of multivariable systems. The
Nyquist plot for multivariable systems carries similar information as
in the single-inputsingle-output systems. The number and the direction of encirclements of the critical 1 j0 point conveys the message
about the stability. But rigorous mathematical analysis is necessary
because matrices are involved.
(75)
0.632z
(z 1)(z 0.368z)
(74)
HALIT EREN
BERT WEI JUET WONG
Curtin University of Technology
150
(1)
where x and z are state coordinates representing small betatron oscillations of the beam, is the betatron wave number,
is an independent variable (azimuth), and n is the field index of refraction. For proper operation, the field index of refraction should satisfy 0 n 1. However, for this range of
n, the beam focusing is often unacceptable. Feedback is usually not possible due to the difficulty of measuring x and z.
Feedforward also has similar measurement difficulties.
In such cases, a natural question is whether control is possible for such unmeasureable systems. For many systems,
one alternative is open-loop oscillatory control, sometimes referred to as vibrational control (not to be confused with vibration control where the idea is to reduce vibrations). Open-loop
oscillatory control is a fairly recently developed control methodology that does not require measurements of states or disturbances. Instead, zero mean periodic excitation is used to
modify the plant behavior in such a way that control is
achieved as the result of the systems natural response to the
excitation. For example, oscillations in the cyclic accelerator
can be introduced by appropriately focusing and defocusing
sectors of the magnetic lens. This causes a suppression of
betatron oscillations and thereby makes the focus beam more
acceptable. An early heuristic description of this phenomena
was given by Livingston (1), but it was not until 1980 that the
heuristically controlled azimuth accelerator was explained in
the context of open-loop oscillatory control in (2).
Definition 1. Open-loop Oscillatory Control. The utilization of periodic (or almost periodic) control laws, without the
use of measurements in order to induce a desired dynamic
response in a system is referred to as open-loop oscillatory
control or vibrational control.
The simplicity of open-loop oscillatory control synthesis is
offset by the difficulty added in introducing explicit time dependence in the state system models. In order to simplify the
analysis, open-loop oscillatory control algorithms may restrict
the control action so the controlled system admits a small parameter. One way to obtain a small parameter is to introduce
periodic excitation whose frequency is an order of magnitude
larger than the highest system natural frequency. The small
parameter will then arise as the result of a rescaling of time.
For such systems, the time-varying open-loop controlled system can be approximated by the behavior of a time-invarient
averaged equation, to which the usual analytical techniques
for time-invariant systems may be applied. This result forms
the basis of classical averaging theory in applied mathematics
and dynamical systems. Within the context of forced mechanical systems and averaging, energy methods and a quantity
called the averaged potential provide the most direct method
of analysis. In the absence of distinct system time or length
scales, the local stability of an equilibrium or periodic orbit
can be studied by the analysis of the linearized systems first
return map, or monodromy matrix, obtained through Floquet theory.
One of the most compelling examples of open-loop oscillatory control is the stabilization of the simple pendulums inverted equilibrium by high frequency vertical oscillation of
the pendulums suspension point. This discovery is usually
attributed to Bogoliubov (3,4) and Kapitsa (5), although earlier references to similar phenomena exist (6). More recent
accounts of this stabilization may be found in (7,8,9,10),
151
152
where eigenvalues locations are typically symmetric with respect to the imaginary axis. Coron (34) has shown the existence of a time-varying feedback stabilizer for systems whose
averaged versions have eigenvalues on the imaginary axis.
Additional interest in this design derives from the observation that it provides a method of smooth feedback stabilization for systems which Brockett (35) had previously shown
were not stabilizable by smooth, time-invariant feedback.
Stability of a system is concerned with the asymptotic behavior of the system. Often it is important to study trajectories of systems as steady-state behavior is being approached.
Analysis of such trajectories when there is an oscillatory control input is a difficult task. The oscillatory control is usually
designed to be high frequency. As a result, the controlled system is composed of a fast zero average oscillatory trajectory
superimposed on a slow trajectory. Therefore, the designer
must attempt to control the slow part of the trajectory and
ignore (or filter out) the high frequency component.
One disadvantage of open-loop oscillatory control is its accuracy. It is well known that driving a nonlinear system with
a periodic signal generally excites an array of resonances, and
under appropriate conditions chaos in the homoclinic tangles
of unstable resonances [See (36) for a complete exposition on
this topic]. While subharmonic resonances and chaos tend to
be suppressed at high forcing frequencies, 1 : 1 resonances
(primary resonances, or periodic orbits), whose averages correspond to fixed points of an averaged representation of the
dynamics, persist. If a stable 1 : 1 resonance has no association with a fixed point of the time-varying system (i.e., it
arises through a bifurcation), it is called a hovering motion.
These high frequency responses limit the utility of open-loop
oscillatory control when control accuracy is important.
PROBLEMS IN OPEN-LOOP OSCILLATORY CONTROL
Classes of Systems
This section considers systems of ordinary differential equations, with inputs, of the form
x = f (x, u)
(2)
(3)
x =
1
T
x (t) dt
w = q( y)
(4)
w = h(t, y)
(5)
153
where, once again, w approximates x. The oscillatory openloop control results in a superposition of fast oscillatory trajectories on slow trajectories. The slow dynamics are represented by y, and h can be a fast periodic function. In either of
the above two cases, it is hoped to find oscillatory control input u(t) such that the transient performance of w meets desired objectives. Since the state equation of the approximations are time-invariant, the analysis becomes simpler. In
fact, even though Eq. (5) is time-varying, it is only the output
equation which explicitly depends on t. Therefore, many of
the well established tools can be applied directly to the state
equation. In particular, when P(y) Ay B, then the eigenvalues of matrix A help determine the qualitative features of
transient behavior.
Steering and Path Planning for Kinematically
Nonholonomic Systems
An application of open-loop oscillatory control which lies
largely outside the boundaries of this chapter is the use of
periodic functions in path generation for so-called kinematically nonholonomic systems. Such systems include wheeled
vehicles such as the unicycle, autonomous wheeled robots,
and cars with or without trailers. More generally, kinematically nonholonomic systems are systems which possess nonintegrable constraints, and typically the state equations do not
include dynamic effects, such as torques, accelerations, and
forces. Since this class of problems does not involve the articles central themes of stabilization and improvement of transient performance, only a brief description is given.
Consider the special case of Eq. (2) in which
f (x, u) =
m
ui gi (x)
i=1
154
Open-loop
oscillatory input
Unstable plant
(autonomous)
Stable average
plant
Controlled system
by selecting the (t) in Eq. (3) to be periodic zero average signals, such as sinusoidal inputs or zero average square waves.
The frequency of the input is selected to be large, or equivalently, as Fig. 1 shows, the period is small. The periodic system, in the form of Eq. (3) can then be transformed into the
form of Eq. (1.1) in Appendix 1, where turns out to be proportional to the period. At this point, the transformed system
can be averaged. If the averaged system has a uniformly
asymptotically stable equilibrium point, then this implies
that there will be a uniformly asymptotically stable periodic
orbit of the transformed time-varying system in the vicinity
of the equilibrium point. The final criteria for vibrational stabilization is that the periodic orbit satisfying Eq. (3) remain
in the vicinity of xs (even though a transformation is used
prior to averaging). This is the reason for introducing the
definition of x*, which is the average value of the periodic
solution of Eq. (3).
What follows is a step-by-step procedure for the analysis
of open-loop oscillatory control laws by the classical method
of averaging. A brief introduction to the topic of averaging
and references to more comprehensive accountings may be
found in Appendix A.1. Many summaries of this procedure
detailed in this section can also be found in the literature
(e.g. see Refs. 13,24,25,37). The following discussion is based
on (37).
Assume that f in Eq. (3) has a special structure so that Eq.
(3) can be rewritten as
dx
= f 1 (x(t)) + f 2 (x(t), (t))
dt
(6)
where f 1(x(t)) f 1(0, x(t)) and the function f 2(x(t), (t)) is linear with respect to its second argument. Additionally, assume
that (t) is periodic of period T (0 T 1) and of the form
(t) u
(t), where 2/T, and u ( ) is some fixed period2 function. Since the primary interest is high frequency forcing, the usual implication is that the amplitude of (t) is
large. It is possible, however, that u
( ) has small amplitude,
making the amplitude of (t) small also.
Then Eq. (6) can be rewritten as
dx
= f 1 (x(t)) + f 2 (x(t), u(t))
dt
(7)
form Eq. (1.1) in Appendix 1. To make this desired transformation, consider the so called generating equation given as
dx
= f 2 (x(t), u(t))
dt
Suppose that this generating equation has a period T general
solution h(t, c), for some u
( ) and t t0, where h : n
n
n
and c is uniquely defined for every initial condition
x(t0) n.
Introduce into Eq. (7) the Lyapunov substitution x(t)
h(t, z(t)) to obtain
1
h(t, z(t))
dz
=
f 1 (h(t, z(t))
dt
z
(8)
(9)
dy
= P(y( ));
d
P(c) =
1
2
2
0
h( , c)
c
1
f 1 (h( , c)) d
(10)
(11)
*(t) satisfying Eq. (8), in the vicinity of ys that is asymptotically stable also. Furthermore, T is known to be equal to 2/
. Since the transform x(t) h(t, z(t)) is a homeomorphism,
there will exist an asymptotically stable T-periodic solution to
Eq. (7) given by x*(t) h(t, *(t)). Equation (2) is said to be
T
vibrationally stabilized provided that x* 1/T 0 x*(t)dt remains in the vicinity of xs.
Example 1. Oscillatory stabilization of scalar differential
equations. Consider the scalar linear differential equation
x(n) + (a1 + u1 (t)) x(n1) + . . . + (an + un (t)) x = 0
(12)
i = 1, 2, . . ., n
(13)
(14)
where
i =
k2 ki
cos(2 i )
2
i = 1, 2, . . ., n
155
x =
1.3
0
+
1.6
0
0.6
0.8
1
u(t) x
0
(15)
0.6
dx
=
dt
0.8
1.3
0
x+
1.6
0
cos(t)
x
0
(16)
x =
cos(t)
x
0
0
0
1
dz
=
d
0
sin( )
1
0.6
0.8
1.3
1
1.6
0
sin( )
z( ) (17)
1
0.6
dy
=
dt
0.8
1.3 0.4 2
y(t)
1.6
(18)
(19)
156
the averaged equation, then for sufficiently large forcing frequencies there exists an asymptotically stable periodic orbit
near the inverted equilibrium. A simple linearization of the
averaged equation reveals that the stability condition for the
inverted equilibrium is given by 22 2Ig/m.
R(t)
x1 = 0,
x2 =
m
sin t sin x1
I
x1 = c1 = h1 (t, c),
x2 =
m
cos t sin c1 + c2 = h2 (t, c)
I
x1 = z 1 ,
x2 =
m
cos t sin z1 + z2
I
2
m
mg
sin z1
cos2 cos z1 sin z1
z2 =
I
I
m
mb
b
z cos cos z1 +
+
cos sin z1 z2
I 2
I2
I
Therefore the averaged equations, given by Eq. (11), are
y1 = y2 ,
y2 =
1
2
m
I
2
cos y1 sin y1
mg
b
sin y1 y2
I
I
Remark 3. Simple nonquantitative experiments demonstrating the stabilization described in this example are not difficult to build and work remarkably well. Such an experiment
is shown in Fig. 3. In this experiment, the rotary motion of a
dc motor is rectified to periodic linear motion by the mechanism shown in the left frame of Fig. 3. Note that by virtue of
the construction, the forcing amplitude is fixed and the forcing frequency can vary. It is observed that when current is
applied to the motor, the inverted equilibrium is unstable until the forcing frequency reaches a critical frequency at which
the inverted equilibrium experiences a bifurcation which renders it stable, as depicted in the right frame of Fig. 3. The
inverted equilibrium is then stable for all higher frequencies.
Remark 4. To this point, the main goal has been to use averaging as a means of studying the local stability properties of
periodically excited systems. Under certain conditions, however, the averaged system gives far more information about
the global structure of the periodically excited system. As essentially a perturbation technique, averaging theorems as
found in (36,47,48) give no clues as to how large the small
parameter can be perturbed off zero before the averaged
dynamics fail to describe the forced dynamics. For sufficiently large, a variety of undesirable nonlinear effects arise,
such as subharmonic resonance and stochasticity, which are
not captured in any way by the simple averaging of the nonautonomous dynamics. Because of the inherent difficulty of
the analysis, theory for the prediction of nonlinear effects in
this range has been slow to emerge. A later section briefly
illustrates some of the features of periodically excited systems
exhibit when is allowed to vary.
Averaging for Mechanical Systems
Recently, interest has emerged in using high frequency oscillatory forcing to control the dynamics of mechanical systems.
The typical applications setting is a controlled Lagrangian
system where only some of the degrees of freedom are directly
controlled:
d L
L
=u
dt q 1
q1
(20)
d L
L
=0
dt q 2
q2
(21)
157
1 T
q M (q, x)q + vT A (q, x)q Va (q; x, v)
2
(22)
If
L (q1 , q 1 ; q2 , q 2 ) =
1
(q T1 , q T2 )
2
M11
T
M12
M12
M22
q 1
V (q1 , q2 )
q 2
1
(p A T v)T M 1 (p M T v) + Va
2
(23)
H (q, p)
1 T 1
1
p M p vT AM 1 p + vT AM 1 (M 1 )1 M 1 A T v
2
2
1 T
1
+ v AM 1 A v vT AM 1 (M 1 )1 M 1 A T v + V
2
2
1
1
1
T
= (M p M A v)T (M 1 )1 (M 1 p M 1 A T v)
2
=
averaged potential
(24)
The averaged potential given in Eq. (24) is an energy-like
function of the generalized coordinates q which is abbrevi-
158
ated V A(q). A complete understanding of the relationship between the dynamics of nonautonomous Hamiltonian systems
of Eq. (23) and the appropriate counterparts for averaged
Hamiltonian S of Eq. (24) does not presently exist. There are
very broad classes of such systems, however, for which it is
possible to prove the validity of the following:
q0
Averaging Principle for Periodically Forced Hamiltonian Systems. The dynamics associated with Eq. (23) under periodic
forcing (x(t), v(t)) are locally determined in neighborhoods of
critical points of the averaged potential V A(q) as follows:
If q* is a strict local minimum of V A( ), then provided
the frequency of the periodic forcing (x( ), v( )) is sufficiently high, the system will execute motions confined to
a neighborhood of q*.
If (q, p) (q*, 0) is a hyperbolic fixed point of the corresponding averaged system (i.e., the Hamiltonian system
determined by Eq. (24)), then there is a corresponding
periodic orbit of the forced system such that the asymptotic stability properties of the fixed point (q*, 0) of the
averaged system coincide with the asymptotic stability
properties of the periodic orbit for the forced system.
x(t)
cos
=
sin t
y(t)
sin
where prescribes the direction of the oscillatory motion, and
is the frequency. This system is illustrated in Fig. 4. If, as
in the last example, the pendulum has total mass m and inertia I about its hinge point, the motion under this oscillatory
forcing is described by a second order differential equation
I m2 cos( ) sin t + mg sin = 0
(25)
(m )2
cos2 ( ) mg cos
4I
When /2, the hinge of the pendulum undergoes the vertical oscillation described in the previous example. The averaged potential has two or four critical points in the interval
[0, 2) depending on whether or not 22 is less than or larger
than 2Ig/m. Clearly the equilibrium is a strict local
minimum of the averaged potential if and only if 22
2Ig/m. According to the theory of averaging presented in (7)
and (8), the pendulum will execute stable motions confined to
a neighborhood of this equilibrium for sufficiently large values of . This also recovers the result obtained in the previous example.
Remark 5. This example illustrates nonclassical behavior in
the case /2. For this case there will be, for sufficiently
large values of , strict local minima of the averaged potential
which are not equilibrium points of the nonautonomous Eq.
(25). Nevertheless, the pendulum will still execute motions
confined to neighborhoods of such local minima. For details
on this type of emergent behavior, see (33) and (10).
Remark 6. The strategy behind the very simple (open loop)
control designs associated with the averaged potential (and
more generally with systems having oscillatory control inputs) is to produce robustly stable emergent behavior which
is related to the critical point structure of the averaged potential. The design method for control laws in this category involves designing the averaged potential functions themselves
by means of appropriately chosen inputs. The guiding theory
for this approach remains very much under development.
Floquet Theory
Another body of theory used in the study of the stability of
equilibriums and periodic orbits of systems controlled by
open-loop oscillatory inputs is Floquet theory. As described in
Appendix 2, the central idea behind the theory is that the
local stability of an equilibrium or periodic orbit may be deter-
159
0.500
x = A(t)x
where A(t) A1 on 0 t t, and A(t) A2 on t t T,
such that A1A2 A2A1. Then the monodromy matrix M can be
obtained by computing the state transition matrix on the
interval [0, T]; that is
t
0 A 1 dt
T
t
A 2 dt
0.000
0.000
10.000
=M
While in mechanical problems the assumption of piecewise
constant forcing is somewhat nonphysical, such approximations are often useful when the forcing frequency is sufficiently large. Piecewise constant forcing is often not a problem in the analysis electrical and electronic systems, where
such forcing is common.
Example 5. Oscillatory stabilization of a simple pendulum:
Floquet theory. In the previous examples, it was shown that
the inverted pendulum may be stabilized with high frequency
vertical oscillation by averaging the time-varying equations
and studying the stability of the inverted equilibrium with
the averaged system. In this example, stabilization is inferred
by linearizing the pendulum dynamics about the inverted
equilibrium and studying the eigenvalues of the monodromy
matrix. To facilitate the calculation, assume that the pendulum is forced by square wave forcing, that is, in Eq. (19)
(t) 2u(t) where u(t) u(t 1/) is a square wave
R
which switches periodically between 1 and 1. Also, assume
that the damping term b 0. This assumption simplifies the
following calculations and more importantly allows us to
show that pendulum stabilization does not require dissipation. Linearizing Eq. (19) about the inverted equilibrium gives
rise to the linear system
0
1
1
1
= 1
2
2
[mg m2 u(t)] 0
I
Because the input u(t) is piecewise constant, the state transition matrix (t, 0) over one period of u(t) may be computed
as follows:
(0, 1/)
= (0, 1/2)(1/2, 1)
1
1
sin 1
sinh 2
cosh 2
cos 1
2
1
2
2
2
2
=
1
2
1
2
1 sin
2 sinh
cos
cosh
2
2
2
2
2
2
1
1
1
cosh
+
sinh
sin
cos
2
2 1
2
2
=
2
1
1
cosh
+ 2 cos
sinh 2
1 sin
2
2
2
2
1
2
1
1
sinh
+
cos
sin 1 cosh 2
2
2
2
1
2
2
2
1
1
1
1
cosh
sinh
sin
cos
2
2
2
2
2
2
2
1 sin 1 sinh 2 + 2 cos 1 cosh 2 + 1 = 0
1
2
2
2
2
2
and the usual stability condition is that both eigenvalues
must lie within the unit disk in the complex plane. Given the
form of the characteristic polynomial, an equivalent condition
for stability is trace (0, 1/) 2, which in this case may be
written
2
2
1
1
cosh
+
sin 1 sinh 2 < 2
2 cos
2
2
1
2
2
2
The boundary between regions of stability and instability may
be approximated by expanding the trigonometric functions in
Taylor series around zero and solving a truncated inequality
for in terms of . Note beforehand that in doing so, it has
been implicitly assumed that 1 /2 0 and 2 /2 which implies is large. For the present example, the stability regions
displayed in Fig. 5 have been obtained by numerically calculating 1 and 2 over the indicated ranges of and .
Remark 7. Note that in the absence of damping, (0, 1/) is
an area-preserving map. This fact implies that 1 and 2 are
constrained to either lie on the real axis such that 2 1/ 1,
or lie on the unit disk in the complex plane. In this case, stability results are fairly weak (lack of asymptotic stability), but
are typical of Hamiltonian, and more generally, conservative
systems. In the presence of dissipation, the eigenvalues may
occur as conjugate pairs inside the unit disk, implying asymptotic stability.
Remark 8. Note that results obtained through Floquet theory are strictly local results. In the case where the controlled
system is nonlinear, proofs of stability and instability are
strictly for the equilibrium or periodic orbit, and no information is given about the asymptotic behavior of solutions
nearby. This remark is not surprising for neutrally stable
equilibria, but it is in fact sometimes true for unstable equilibria as well. As an example, the reader is referred to the
parametrically excited pendulum example later in this article,
160
d2x
+ 2 [1 (n + K(, ))] x = 0
d 2
d2z
+ 2 [n + K(, )] z = 0
d 2
(26)
dx1
d
dx2
d
dx3
d
dx4
d
= x2
= 2 (1 n) x1 + 2 sin( ) x1
(27)
= x4
= 2 nx3 2 sin( ) x3
d y
+ 2 (1 n + 0.5 2 )y = 0
d 2
d 2
+ 2 (n + 0.5 2 ) = 0
d 2
(28)
H(c)
1
T
h(, c) d
(29)
and T is the period of h(t, ). This is done by approximating x(t) H(y(t)) and once again, analyzing the
transient behavior of y (making this technique more
typical of Eq. (4)). Since the fast dynamics in h are averaged out, this technique introduces some error. On the
other hand, since H does not explicitly depend on t, the
analysis becomes simpler.
In either of the two methods, controlling y(t) in Eq. (11) governs how to control x(t).
Example 7. Transient behavior of the vertically oscillating
pendulum. Referring back to Example 3 and using the same
notation, it is possible to now approximate the trajectories of
, the angular position, and , the pendulums angular velocity. The averaged equations of the system are
y1 = y2
1
y2 =
2
where
1
2
= y2
m
I
mg
b
sin y1 y2
cos y1 sin y1
I
I
x1 = z 1
x2 =
m
cos t sin z1 + z2
I
m
cos t sin y1 + y2
I
RESONANCE PHENOMENA IN
PERIODICALLY FORCED SYSTEMS
1.5
1.5
Averaged system
Method 1
Averaged system
Original system
0.5
Theta
Theta
m
cos s sin y1 + y2 ds
I
Resonances arise as the natural result of subjecting any system to periodic excitation. The resonances produced vary
H(y)
0
0.5
1
H(y)
0.5
0
0.5
2.6
2.8
3
Theta
3.2
3.4
2.6
2.8
3
Theta
3.2
3.4
1.5
Method 1
Original system
0.5
Theta
Theta
Method 1
Original system
1.5
1
0.5
0
0
0.5
0.5
1
2.6
2.8
161
3
Theta
3.2
3.4
10
15
20
Figure 6. A comparison of the behaviors of the averaged/Method 2 system, the Method 1 system,
and the original system. In these plots, m g I 1, b 0.5, 0.2, and 10.
25
162
m
F0cos t
g
q
m
(30)
q +
+
sin
sin q = 0
2
2
1.5
1.5
0.5
0.5
q
0.5
0.5
1.5
1.5
0
q
1.5
1.5
0.5
0.5
0.5
1.5
1.5
2
0
q
1.5
1.5
0.5
0.5
q
0.5
1.5
1.5
2
0
q
0
q
0
q
0.5
0.5
163
0
q
Figure 8. Poincare maps showing separatrix splitting and resonance bands for the parametrically excited pendulum. The phase portrait for the unperturbed simple pendulum is shown in
the upper left frame, with Poincare maps of the forced system to the right and underneath.
Forcing parameters used in the Poincare maps are indicated in each plot. Note that the Poincare
maps of the forced system more closely resemble the averaged phase portrait as the forcing
frequency becomes large.
Remark 10. The example gives somewhat anecdotal evidence that choosing a sufficiently high forcing frequency
tends to suppress the negative features of periodic excitation.
This has also been found to be the case in the cart and pendulum problem described in Examples 3, 4, and 5 [see (10)], the
vertically forced rotating chain (33), and in an entire class of
164
(1.2)
where f 0 : n n is defined
f 0 (y)
1
T
f (s, y) ds
0
ACKNOWLEDGMENTS
Brad Lehman gratefully acknowledges the support of the National Science Foundation through an NSF President Faculty
Fellowship, grant CMS 9596268. John Baillieul would like to
express gratitude for support from the United States Air
Force Office of Scientific Research under grant F49620-96-10059.
APPENDIX 1. CLASSICAL AVERAGING THEORY
The classical method of averaging was originally developed
for periodic systems of the form
x = f (t, x)
x = A(t) x
(2.1)
(1.1)
165
(t + T ) = (t)M
BIBLIOGRAPHY
1. M. S. Livingston, High Energy Accelerators, New York: WileyInterscience, 1954.
2. S. M. Meerkov, Principle of vibrational control: theory and applications, IEEE Trans. Autom. Control, AC-25: 755762, 1980.
13. B. Lehman, Vibrational control of time delay systems. In J. Wiener and J. K. Hale, (eds.), Ordinary and Delay Equations, Pitman
Research Notes in Mathematical Series 272. Harlow, Essex, England: Longman Scientific & Technical, 1992.
14. B. Lehman, I. Widjaya, and K. Shujaee, Vibrational control of
chemical reactions in a CSTR with delayed recycle, J. Math.
Anal. & Appl., 193: 2859, 1995.
15. J. E. Bailey, Periodic operation of chemical reactors: A review,
Chem. Eng. Com., (1): 111124, 1973.
16. J. Fakhfakh and J. Bentsman, Experiment with vibrational control of a laser illuminated thermochemical system, ASME J. Dyn.
Syst., Meas. Control, 110: 353360, 1990.
17. S. M. Osovets, Dynamic methods of retention and stabilization of
plasma, Soviet Phys. - Usp., 17 (2): September 1974.
18. D. Tilbury, R. Murray, and S. Sastry, Trajectory generation for
the n-trailer problem using goursat normal form, IEEE Trans.
Autom. Control, AC40: 802819, 1995.
19. J. L. Speyer, A second variational theory for optimal periodic processes, IEEE Trans. Autom. Control, AC-29-2: 138148, 1984.
20. J. L. Speyer, Nonoptimality of steady-state cruise for aircraft.
AIAA J., 14 (11): 16041610, 1976.
21. S. Weibel and J. Baillieul, Averaging and energy methods for robust open-loop control of mechanical systems. To appear in Essays on Mathematical Robotics. Berlin: Springer-Verlag, 1998.
22. S. Weibel and J. Baillieul, Oscillatory control of bifurcations in
rotating chains. In Proc. 1997 ACC, Albuquerque, 27132717,
1997.
23. J. Baillieul, S. Dahlgren, and B. Lehman, Nonlinear control designs for systems with bifurcations and applications to stabilization and control of compressors. In Proc. 34th CDC, New Orleans,
30623067, 1995.
25. B. Lehman, J. Bentsman, S. V. Lunel, and E. I. Verriest, Vibrational control of nonlinear time lag systems with bounded delay:
averaging theory, stabilizability, and transient behavior, IEEE
Trans. Autom. Control, AC-39: 898912, 1994.
27. J. Bentsman and K.-S. Hong, Vibrational stabilization of nonlinear parabolic systems with neumann boundary conditions, IEEE
Trans. Autom. Control, AC-36: 501507, 1991.
28. J. Bentsman and K.-S. Hong, Transient behavior analysis of vibrationally controlled nonlinear parabolic systems with neumann
boundry conditions, IEEE Trans. Autom. Control, AC-38: 1603
1607, 1993.
29. S. Lee, S. Meerkov, and T. Runolfsson, Vibrational feedback control: zero placement capabilities, IEEE Trans. Autom. Control,
AC-32: 604611, 1987.
30. K. Shujaee and B. Lehman, Vibrational feedback control of time
delay systems. In Proc. 34th IEEE CDC, 936941, 1995.
31. K. Shujaee and B. Lehman, Vibrational feedback control of time
delay systems, IEEE Trans. Autom. Control, AC-42: 15291545,
1997.
166
OPERATIONAL AMPLIFIERS
34. J. M. Coron, Global asymptotic stabilization for controllable systems without drift, Math. Control, Signals, Syst., 5: 295312,
1992.
55. W. Magnus and S. Winkler, Hills Equation. Tracts of Mathematics. vol. 20, New York: Interscience Publishers, 1966.
B. LEHMAN
S. WEIBEL
Northeastern University
J. BAILLIEUL
Boston University
42. N. E. Leonard and P. S. Krishnaprasad, Motion control of driftfree, left invariant systems on lie groups, IEEE Trans. Autom.
Control, 40 (9): 1995.
43. H. J. Sussman and W. Liu, Lie bracket extension and averaging:
The single bracket case. In Z. Li and J. F. Canny (eds.), Nonholonomic Motion Planning. 109147. Boston: Kluwer Academic Publishers, 1993.
44. H. J. Sussman and W. Liu, Limits of highly oscillatory controls
and approximation of general paths by admissible trajectories. In
Proc. 30th IEEE CDC, December 1991.
45. S. M. Meerkov and M. Tsitkin, The effectiveness of the method
of vibrational control for dynamic systems described by a differential equation of order n, Automation and Remote Control (translation of Avtomatika i Telemekhanika), 525529, 1975.
46. R. Bellman, J. Bentsman, and S. M. Meerkov, Stability of fast
periodic systems, IEEE Trans. Autom. Control, AC-30: 289291,
1985.
47. J. K. Hale, Ordinary Differential Equations. Texts and Monographs in Pure and Applied Mathematics, Malabar, FL: Robert
E. Krieger Publishing, 1969.
48. J. A. Sanders, On the fundamental theory of averaging, SIAM J.
Math. Anal., 14: 110, 1983.
49. D. Seto and J. Baillieul, Control problems in superarticulated
mechanical systems, IEEE Trans. Autom. Control, AC-39-12:
24422453, 1994.
50. M. Fliess, Generalized controller canonical forms for linear and
nonlinear dynamics, IEEE Trans. Autom. Control, AC-35-9: 994
1001, 1990.
51. A. J. Lichtenberg and M. A. Lieberman, Regular and Chaotic Dynamics. Volume 38 of Applied Mathematical Sciences. Berlin:
Springer Verlag, 1992.
52. S. Wiggins, Introduction to Applied Nonlinear Dynamical Systems
and Chaos. In Texts in Applied Mathematics. vol. 2, Berlin:
Springer Verlag, 1990.
ERATING SYSTEMS.
364
OPTIMAL CONTROL
OPTIMAL CONTROL
Optimal control theory is concerned with the development of
techniques that allow one to control physical phenomena described by dynamical systems in such a manner that a predescribed performance criterion is minimized. The principal
components of an optimal control problem are the mathematical model in the form of a differential equation, a description
of how the control enters into this system, and a criterion
describing the cost.
The start of optimal control theory, as a mathematical discipline, dates back to the mid 1940s. The increasing interest
in and use of methods provided by optimal control theory is
linked to the rise of the importance of mathematical models
in many diverse areas of scienceincluding chemistry, medicine, biology, management, and financeand to ever increasing computing power, which allows the realization of optimal
control strategies for practical systems of increasing difficulty
and complexity. While optimal control theory has its roots in
the classical calculus of variations, its specific nature has necessitated the development of new techniques. In contrast
with general optimization problems, whose constraints are
typically described by algebraic equations, the constraints in
optimal control problems are given by dynamical systems.
The dynamic programming principle, the Pontryagin maximum principle, the HamiltonJacobiBellman equation, the
Riccati equation arising in the linear quadratic regulator
problem, and (more recently) the theory of viscosity solutions
are some of the milestones in the analysis of optimal control
theory.
Analyzing an optimal control problem for a concrete system requires knowledge of the systems-theoretic properties
of the control problem and its linearization (controllability,
stabilizability, etc.). Its solution, in turn, may give significant
additional insight. In some cases, a suboptimal solution that
stabilizes the physical system under consideration may be the
main purpose of formulating an optimal control problem,
while an exact solution is of secondary importance.
In the first section we explain some of the concepts in optimal control theory by means of a classical example. The following sections describe some of the most relevant techniques
in the mathematical theory of optimal control.
Many monographs, emphasizing either theoretical or control engineering aspects, are devoted to optimal control theory. Some of these texts are listed in the bibliography and
reading list.
d2
y(t) + mg sin y(t) = u(t),
dt 2
t>0
(1)
and
d
y(0) = v0
dt
x1 (t)
x2 (t)
=
x2 (t)
g sin x1 (t) + u(t)
(2)
with initial condition x(0) x0 (y0, v0) R2, where we assume m 1. In general, a control system is written in the
form
d
x(t) = f (t, x(t), u(t)),
dt
x(0) = x0
(3)
with state vector x(t) Rn, control input u(t) Rm, and f :
R1 Rn Rm Rn. If f is independent of time t [f f(x, u)],
then the system is said to be autonomous.
Next we formulate a sample control system associated to
Eq. (1). For that purpose, note that the stationary solutions
to the uncontrolled system, which are characterized by f(x,
u) 0 for u 0, are given by (0, 0) and (, 0). Our objective
is to regulate the state x(t) R2 to the stationary state (, 0).
Thus a control u must be determined that steers the system
described by Eq. (1) from the initial state x0 to the vertical
position (y ) or into its neighborhood (inverted-pendulum
problem). This objective can be formulated as an optimal control problem: minimize the cost functional
tf
J(x, u) =
0
(4)
tf
min
0
(5)
subject to Eq. (3), over u L2(0, tf; Rm) with u(t) U a.e. in
(0, tf), where U is a closed convex set in Rm describing con-
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
OPTIMAL CONTROL
365
straints that must be observed by the class of admissible controls. In the terminology of the calculus of variations, Eq. (5)
is called a Bolza problem. The special cases with f 0 0 and
with g 0 are referred to as the Lagrange and the Mayer
problem, respectively. If tf 0 is finite, then Eq. (5) is called
a finite-time-horizon problem. In case g 0 and tf , we
refer to Eq. (5) as an infinite-time-horizon problem. The
significance of the latter is related to the stabilization of
Eq. (3).
If Eq. (5) admits a solution u*, we refer to it as the optimal
control, and the associated state x* x(u*) is the optimal
trajectory. Under certain conditions, the optimal control can
be expressed as a function of x*, that is, u*(t) K(t, x*(t)) for
an appropriate choice of K. In this case u* is said to be given
in feedback or closed-loop form.
If the final time tf itself is a free variable and f 0 1, then
Eq. (5) becomes the time optimal control problem.
For certain analytical and numerical considerations, the
treatment of the fully nonlinear problem (5) can be infeasible
or lengthy. In these cases, a linearization of the nonlinear
dynamics around nominal solutions will be utilized.
at x (0, 0) and
Linearization
and B = f u (x, u)
(6)
A=
0
g
1
0
and B =
0
1
1
0
and B =
0
1
(6b)
0
g
1
x(t) +
0
0
u(t)
1
(8)
d
x(t) = Ax(t)
dt
is asymptotically stable if and only if all eigenvalues of the
matrix A satisfy Re 0. For example, for the matrices in
Eqs. (6a) and (6b) we have
det(I A) = 2 + g = 0 = { gi}
(marginal stability)
and
(instability)
x(0) = x0 x
0
g
det(I A) = 2 g = 0 = { g}
d
z(t) = Az(t) + Bv(t) + r(z(t), v(t))
dt
d
x(t) = Ax(t) + Bu(t),
dt
A=
(6a)
K = (0
) R12
366
OPTIMAL CONTROL
subject to Eq. (7), where Q, G Rnn are symmetric nonnegative definite matrices and R Rmm is symmetric and positive
definite. The optimal solution u*( ) to Eq. (9) is given in feedback form by
u (t) = R1 Bt Px (t)
(10)
Existence of Optimal Controls
the initial and terminal times. Also we shall not systematically discuss the bangbang principle, which states that, for
certain control systems with controls constrained to lie in a
convex compact set, the optimal controls are achieved in the
extremal points of the admissible control set.
x (0) = x0
and the symmetric nonnegative definite matrix P Rnn satisfies the matrix Riccati equation
At P + PA PBR1 Bt P + Q = 0
(11)
J(x, u)
(16)
x(t) f (t, x, u)
E(x, u) = dt
=0
(t f , x(t f ))
tf
J(x, u) =
(12)
(17)
and
u K = {u(t) U a.e. in (0, t f )}
(18)
(13)
control constraints
u(t) U
(a closed set in Rm )
(14)
and (t f , x(t f )) = 0
(15)
with (t f , x(u)(t f )) = 0
(19)
OPTIMAL CONTROL
367
f )), (t f ))
(H(t f , x (t f ), u (t f ), (t
(20)
= 0 (1, 0, . . ., 0) + (0, x (x (t f ))
for some (0, ) R Rp. Here we set x(x*(tf))
p
i1 i grad i(x*(tf)).
4. If one can ascertain that 0 0 (normality), then without loss of generality we can set 0 1, and conditions
23 of Theorem 1 can be equivalently expressed as
d
(t) = Hx (t, x , u , )
dt
(t f ) = x (x (t f ))
(21)
3. Transversality:
f )), (t f )) T f
(H(t f , x (t f ), u (t f ), (t
where Ttf is the tangent space to the manifold described by
(t, x) 0 at (tf, x*(tf)).
An admissible triple (x*, u*, *) that satisfies the conclusions of Theorem 1 is called an extremal element. The function x* is called the extremal trajectory, and u* is called the
extremal control.
Remarks
1. The maximum principle provides a necessary condition
for optimality. It is simple to find examples illustrating
x1 (0) = c > 0
=
0
x3 (t) dt
368
OPTIMAL CONTROL
J=
[u(t) 1]x(t) dt
(22)
u (t) =
0 on (T 1, T]
Lagrange Multiplier Rule
subject to
d
x(t) = u(t)x(t),
dt
H = 0 (u 1)x + ux
d
= Hx = 0 (u 1) u
dt
and the transversality condition implies that (T) 0. Since
(0, (t)) 0 on [0, T], it follows that 0 0. Thus normality
of the extremals holds, and we set 0 1. The maximum
condition implies that
[1 u (t)]x (t) + (t)x (t)u (t) (1 u)x (t) + (t)x (t)u
for all u [0, 1]
Since necessarily x*(t) 0, the sign of (t) 1 determines
u*(t), that is,
u (t) =
1 if (t) 1 > 0
d
= (1 )u 1,
dt
Theorem 2. Assume that (x(u*), u*) minimizes the cost functional in Eq. (16) subject to Eqs. (17) and (18) and that the
regular point condition
0 int{E
(x(u ), u )(h, v u ) : h HL1 (0, t f , Rn ) and v K}
(23)
holds. Then there exists a Lagrange multiplier (, ) L2(0,
tf, Rn) Rp such that
(Lu , u u ) 0
tf
0
x (t) = ce ,
tf
t
ft
and (t) = e
(t )
d
h dt
dt
+ x (x (t f ))h(t f ) = 0
[ f x0 (s, x , u ) f x (s, x , u )] ds
on [, ]
for all h HL1 (0, tf; Rn). An integration-by-parts argument implies that
for all u K
(T ) = 0
(24)
0 if (t) 1 0
u (t) = 1,
,R n )R p
Further define HL1 (0, tf, Rn) as the set of functions in H1(0, tf,
Rn) that vanish at t 0.
We have the Lagrange multiplier rule:
+ x (x (t f )) +
d
h(t) dt = 0
dt
and thus
tf
(t) =
t
[ f x0 (s, x , u ) + f x (s, x , u )] ds x (x (t f ))
OPTIMAL CONTROL
369
pose we introduce additional scalar components for the dynamical system and for the target constraint by
d
x
= 0 and p+1 (x) = g(x) txn+1
dt n+1
where x (x1, . . ., xn) as before, and the augmented cost
functional is
f0 = f 0 (t, x, u) + xn+1
We find
Huu (t) 0
for t [0, t f ]
for t [0, t f ]
(25)
Q(t ) = (x(t ))
on ker x (x(t f ))
xx
f
f
(26)
tf
for all (x, u) satisfying Eq. (13), Eq. (15), and (x, u) (x*,
u*)L(0,tf ;Rnm) .
tf
0
tf
=
0
d
(t))
(t) = Hx (t, x (t), u (t),
dt
d
(t) = 0 ,
dt n+1
n+1 (0) = 0
(t f ) = gx (x (t f ))
Bolza Problem
Here we discuss the maximum principle for a Bolza type problem, where the cost functional of Eq. (12) is replaced by
min
tf
min
tf
J(x, u) =
0
with g : Rn R. Augmenting the system (13), the Bolza problem can be expressed as a Lagrange problem. For this pur-
(27)
subject to
d
x(t) = Ax(t) + h(t, u(t)),
dt
x(t) = x0
370
OPTIMAL CONTROL
for all u U
(28)
x(0) = x0
u (t) = R1 Bt t (t)
[ f x0 (t, x )x + h0 (t, u )] dt
(t f )x(t f )
tf
0
[ f x0 (t, x )x
(t f ) = xt (t f )G
d
d
d
(x) =
x + x = ( f x0 A)x + (Ax + h)
dt
dt
dt
= f x0 (, x )x + h(, u)
tf
(30)
where f(t) L2(0, tf; Rm) represents a disturbance or an external force, Q, G Rnn are symmetric and nonnegative matrices, and R Rmm is symmetric and positive definite. This
problem is referred to as the finite-time-horizon linear quadratic regulator problem. The Hamiltonian H is given by
d
(t) = (t)A + xt (t)Q,
dt
we have
x(0) = x0
(t f )x (t f )
d
x(t) = Ax(t) + Bu(t) + f (t),
dt
(t f ) = gx (x (t f ))
d
x(t) = Ax + h(t, u),
dt
d
x(t) = Ax(t) BR1 Bt p(t) + f (t),
x(0) = x0
dt
d
p(t) = At p(t) Qx(t),
p(t f ) = Gx(t f )
dt
(31)
+ h (t, u)] dt
0
gx (x (t f ))[x(t f ) x (t f )] +
tf
0
tf
f x0 (t, x )(x
x ) dt
Note that (x) (x*) x(x*)(x x*) for all x, x* for any
convex and C1 function . Since g, f 0 are convex, we have
g(x(t f )) +
tf
g(x (t f )) +
tf
min
[ f 0 (t, x ) + h0 (t, u )] dt
tf =
tf
1 dt
(32)
d
x(t) = Ax(t) + Bu(t)
dt
x(0) = x0
(33)
J(x0 , u)
t f
1
[xt (t)Qx(t) + ut (t)Ru(t)] dt + xt (t f )Gx(t f )
=
2
0
ui [1, 1] for 1 i m,
min
(29)
and x(t f ) = 0
OPTIMAL CONTROL
and x1 (t) =
(t + c1 )2
+ c2
2
Thus, the orbit is on the manifold x1 x22 /2 c2 oriented upwards. Similarly, for u* 1 the orbit is on the manifold
x1 x22 /2 c2 oriented downwards. Since the optimal controls have at most one switch and the orbits must terminate
at (0, 0), it follows that the optimal control u* is given in
feedback form by
d
= A
dt
371
1 if (x1 , x2 ) is above S
1 if (x1 , x2 ) is below S
and hence
ui (t) = sign gi (t), for 1 i m
where g(t) : (t)B eA(tft)B. We claim that g(t) is nontrivial.
In fact, if g(t) 0 for some t [0, tf], then, since (A, B) is
controllable, 0 and (t) 0. We have
H(t f ) = 0 + (t f )[Ax(t f ) + Bu(t f )] = 0 = 0
and thus (0, (t)) is trivial if g(t) 0. This gives a contradiction to Theorem 1.
In the remainder of this subsection we consider a special
linear control system (rocket sled problem) and provide its
solution. Let y(t), the displacement of a sled with mass 1 on a
friction-free surface be controlled by an applied force u(t) with
constraint u(t) 1. By Newtons second law of motion,
(d2 /dt2)y(t) u(t). If we define x1(t) y(t) and x2(t)
(d/dt)y(t), then the state x(t) col(x1(t), x2(t)) satisfies Eq. (33)
with
A=
0
0
1
0
and B =
0
1
We observe that the system (A, B) is a single-input controllable system that is marginally stable. This implies existence of
an optimal control u*. From the above discussion it follows
that u* must satisfy
u (t) = sign 2 (t)
tf
J(s, y; u) =
min
(34)
subject to
d
x(t) = f (t, x(t), u(t)),
dt
x(s) = y,
u(t) U
(35)
where U is a closed convex set in Rm. Under appropriate conditions on f, Eq. (35) has a unique solution x x(t; (s, y)), and
moreover x C(s, tf; Rn) depends continuously on (s, y) [0,
tf] Rn and u L1(s, tf; Rm). As discussed in the preceding
section, sufficient conditions on f 0 and g, which guarantee the
existence of an optimal pair (x*, u*) for each (s, y) [0, tf]
Rn, are well known. We define the minimum-value function
V(s, y) by
V (s, y) = min J(s, y; u)
uK
min
f (t, x(t), u(t)) dt + V ( , x( )) : u U on [s, ]
0
= V (s, y)
and 2 (t) = 1 (T t) + 2
for some nonzero (1, 2). Hence the optimal control assumes at most the values 1 and 1 (bangbang control) and
it has at most one switch between these two values. Assume
(36)
J(s, y; u) =
s
(37)
372
OPTIMAL CONTROL
and thus
V (s, y)
min
s
f 0 (t, x(t), u(t)) dt + V ( , x( )) : u Uad on [s, ]
V (s, y)
which implies Eq. (36).
Suppose that V is continuously differentiable. Then V satisfies the so-called HamiltonJacobiBellman (HJB) equation:
Vt (s, y) + min[ f (s, y, u)Vx (s, y) + f 0 (s, y, u)] = 0
uU
(38)
u =
u on ( , t f )
V (s, y) =
(39)
Hence we have
d
d
V (t, x(t)) = Vt + Vx x(t) = Vt + f (t, x(t), u(t)) Vx
dt
dt
(Vt + fVx ) dt
Now, since V(s, y) J(s, x; u), the equation above Eq. (39)
implies
tf
V (s, y)
s
or equivalently,
V ( , x( )) = V (s, y) +
x(s) = y
has a solution x*(t) C1(s, tf; Rn) for each (s, y) [0, tf]
Rn, then the feedback solution u*(t) (t, x*(t)) is
optimal, that is, V(s, y) J(s, y; u*).
u on (s, )
(40)
d
V (t, x (t)) = Vt (t, x (t)) + f (t, x (t), u (t))Vx (t, x (t))
dt
= f 0 (t, x (t), u (t))
and thus
lim
s +
1
s
tf
V (s, y) =
s
(41)
OPTIMAL CONTROL
V (x) = g(x) on x
uU
(47)
Vt H(t, x, Vx ) = 0
(42)
373
p(t f ) = gx (x (t f )) (43)
d
Vx
d
Vx (t, x (t)) =
(t, x (t)) + Vxx (t, x (t)) x (t)
dt
t
dt
= Hx (t, x, Vx (t, x (t)))
+ Vxx (t, x (t))H p (t, x, Vx (t, x (t)))
+ Vxx (t, x (t)) f (t, x (t), u (t))
x (t, x (t), u (t), Vx (t, x (t)))
=H
where Vxx 2V/xi xj Rnn. Here we have used the fact
that
with V (1) = 0
H p (t, x, p) = f (t, x, u)
(48)
at (t0 , x0 )
and
t H(t, x, x ) 0
Hx (t, x, p) = f x (t, x, u)
t p f x0 (t, x, u)
t
(t, x, u, p) over U. We
where u (t, x, p) U maximizes H
observe that Eq. (43) represents the adjoint equation of the
maximum principle with adjoint variable given by Vx( ,
x*).
Next let us set U Vx. Then U satisfies
Ut (t, x) + f (t, x, (t, x, U (t, x)))
Ux (t, x) Hx (t, x, U (t, x)) = 0
(44)
d
x(t) = f (t, x, u(t))
dt
d
p(t) = Hx (t, x, p(t))
dt
d
V (t) = f 0 (t, x, u(t))
dt
It is clear that a C1 solution to Eq. (42) is a viscosity solution, and if v is a viscosity solution of Eq. (42) and Lipschitz
continuous, then vt H(t, x, vx) 0 a.e. in (0, tf) Rn. The
viscosity solution concept is derived from the vanishing viscosity method illustrated by the following theorem.
Theorem 6. Let V(t, x) C1,2((0, tf) Rn) be a solution to
the viscous equation
Vt H(t, x, V ) + V = 0,
V (t f , x) = g(x)
(49)
(45)
are the characteristic equations of the first order partial differential equations (PDEs) (42) and (44).
Viscosity Solution Method
In this section we discuss the viscosity solution method for
the HJB equation. For motivation, we first consider the exit
time problem
min J(y, u) =
f 0 (x(t), u(t)) dt + g(x( ))
(46)
0
at (t0 , x0 )
at (t0 , x0 )
x = 0,
xx
0
at (t, , x )
374
OPTIMAL CONTROL
It follows that
t H(t, x, x ) + t H(t, x, x ) +
at (t , x )
V (1) = V (1) = 0
(50)
uU
(51)
dP
(t) + At (t)P(t) + P(t)A(t)
dt
P(t)B(t)R1 (t)Bt (t)P(t) + Q(t) = 0
(53)
for some
(t)
D+
x V (t, x (t))
d
v(t) = [A BR1 Bt P(t)]t v(t) = P(t) f (t),
dt
(54)
v(t f ) = 0
1
h (p) =
|p|2
|p|
if |p| < 1
1
2
if |p| 1
and
We find
hp (p)
if |p| < 1
p/|p| if |p| 1
uU
= a(x) p h (b(x)t p)
(55)
OPTIMAL CONTROL
where the symmetric matrix P(t) and the feedforward v(t) satisfy Eqs. (53) and (54). The matrix K(t) R1BtP(t) describing
the control action as a function of the state is referred to as
the feedback gain matrix. The solution in the form of Eq. (55)
can be derived from the dynamical programming principle in
the preceding subsection, but here we prefer to give an independent derivation based on the two-point boundary value
problem (31). Since this equation is affine, we can assume
that
p(t) = P(t)x(t) + v(t)
(56)
d
dt
d
x (t) = [A BR1 Bt Pt (t)]x (t) BR1 Bt v(t) + f (t) (57)
f
dt
where v(t) is a solution to Eq. (54), and if we set p(t)
P(t)x*(t) v(t), then the pair (x*(t), p(t)) is a solution to Eq.
(31). Thus, the triple (x*(t), u*(t), pt(t)) satisfies the maximum principle. From Theorem 3 or from Eq. (58) below, it
follows that the feedback solution (55) is optimal.
The formula (56) is called the Riccati transformation. It
transforms the TPBV problem (31) into a system of initial
value problems backwards in time. Moreover, the feedback
solution given by Eq. (55) is unique. This follows from the fact
that for arbitrary u L2(0, tf; Rm), multiplying Eq. (57) from
the left by [Ptf (t)x*(t)]t using Eqs. (53) and (54), and integrating the resulting equation on [0, tf], we have
tf
0
1
2
tf
+
0
1
2
(60)
x(0) = x0
This problem need not admit a solution. For example, consider the system
d
x(t) =
dt
1
0
1
x(t) +
1
1
u(t)
0
and let Q I and R be arbitrary. Then there exists no admissible control u L2(0, ; Rm) such that J(x0, u) is finite, unless x2(0) 0.
Under the assumption that for each x0 Rn there exists at
least one admissible control such that J(x0, u) is finite (finitecost condition), it can be shown that the optimal control is
given in feedback form by
(61)
The control
(58)
J(x0 , u ) =
J(x0 , u) =
B P(t) + Q x(t)
1
2
min
u (t) = R1 Bt P x (t)
d
v(t) + [A BR1 Bt P(t)]t v(t) + P(t) f (t) = 0
dt
J(x0 , u) = J(x0 , u ) +
375
u (t) = R1 Bt P x (t)
is the unique solution to the LQR problem (60), and
J(x0 , u ) = 12 xt0 P x0 =
min
J(x0 , u)
uL 2 (0,;R m )
(59)
Conversely, if there exists a nonnegative symmetric solution P to Eq. (61), then for all x0 Rn there exists an
376
OPTIMAL CONTROL
Thus, Ptf (0) Ptf (0) for tf tf. The assumption on the existence of admissible controls shows that eitPtf (0)ei and (ei
ej)tPtf (0)(ei ej) are monotonically nondecreasing and bounded
with respect to tf. Here ei denotes the ith unit vector in Rn.
Defining P limtf Ptf (0), it follows that P is symmetric, is
nonnegative, and moreover satisfies the steady-state equation
Eq. (61). It can then be argued that the feedback control
u*(t) R1BtPx*(t) is the unique optimal solution to the
LQR problem (60). To prove the last assertion of part 1, suppose that P is a nonnegative symmetric solution to Eq. (61).
Let x(t) be the solution to (d/dt)x(t) (A BR1BtP)x(t) with
x(0) x0, and let u(t) R1BtPx(t). Then
d t
[x (t)Px(t)] = 2xt (t)P(A BR1 Bt P)x(t)
dt
= xt (t)(Q + PBR1 Bt P)x(t)
Integration of this equation on [0, tf] implies
tf
0
and thus J(x0, u) x0t Px0 and x0t Px0 x0t Px0 for every
x0 Rn.
To verify part 2, note that
(A BR1 Bt P )t P + P (A BR1 Bt P )
+ Q + P BR1 Bt P = 0
It can be shown that (A BR1BtP, (Q PBR1BtP)1/2) is
detectable (observable) if (A, Q1/2) is detectable (observable).
Hence, it follows from the Liapunov criterion (5) that A
BR1BtP is asymptotically stable and moreover that P is
positive definite if (A, Q1/2) is observable. For the proof of
uniqueness we refer to the literature (see, e.g., Ref. 5).
In the following theorem we consider the LQR theory with
external forcing.
Theorem 9. Consider the infinite-time-horizon problem
min
1
2
[x (t)Qx(t) + u (t)Ru(t)] dt
t
subject to
d
x(t) = Ax(t) + Bu(t) + f (t),
dt
x(0) = x0
v() = 0
Proof. From Theorem 8 it follows that A BR1BtP is asymptotically stable and thus v L2(0, ; Rn) and u*(t)
L2(0, ; Rm). Since (A, Q1/2) is detectable, there exists a matrix
G Rnn such that A GQ is asymptotically stable. For arbitrary admissible controls u L2(0, ; Rn)
x(t) = e(AGQ)t x0 +
H=
A
Q
BR1 Bt
At
(62)
d
x(t) = Ax(t) + Bu(t) + f (t),
dt
|x(t)|2 dt <
x(0) = x0 and
0
OPTIMAL CONTROL
377
v() = 0
and
and
I
P
0
I
I
P
0
I
A WP
0
A WP
Q At P
W
At + PW
W
At
=
A WP
P(A WP)
W
At
Theorem 11
1. Let Q, W be symmetric n n matrices. Solutions P to
the algebraic Riccati equation AtP PA PWP Q
0 coincide with the set of matrices of the form P
VU1, where the n n matrices U [u1, . . ., un], V
[v1, . . ., vn] are composed of upper and lower halves of
n real Schur vectors of the matrix
H=
A
Q
W
At
and U is nonsingular.
2. There exist at most n eigenvalues of H that have negative real part.
3. Suppose [u1, . . ., un] are real Schur vectors of H corresponding to eigenvalues 1, . . ., n, and i j for 1
i, j n. Then the corresponding matrix P UV1 is
symmetric.
4. Assume that Q, W are nonnegative definite and (A,
Q1/2) is detectable. Then the solution P is symmetric and
nonnegative definite if and only if Re k 0, 1 k
n.
Proof. We prove part 1. Let S be a real Schur form of H, that
is, HU US with UtU I and
S=
Thus
U11
H
U21
where
S11
0
S12
S22
U11
U21
S11
and
I
P
0
I
I
P
0
I
A WP
0
W
At + PW
J(x0 , u) =
(63)
subject to
d
x(t) = f (x(t), u(t)),
dt
x(0) = x0 ,
u(t) U
(64)
for all u U
(65)
We assume that
(66)
U11
U21
Thus
admits a unique minimizer over U, denoted by (x, p). Finally, we assume that l, h, g, and f are sufficiently smooth
with l and g bounded from below.
378
OPTIMAL CONTROL
Discrete-Time Approximation
We consider the discretized problem
min
J N (uN ) =
N
[l(xk ) + h(uk )] t + g(xN )
(67)
k=1
Proof. First, we show that xN and pN are uniformly Lipschitzian in N. It will be convenient to drop the tilde in the notation for these sequences. In the case that U is bounded, we
proceed by taking the inner product of Eq. (68) with xk and
employing Eq. (65):
1
(|xk |2
2
subject to
xk xk1
= f (xk , uk ) and uk U
t
1kN
(68)
xk xk1
= f (xk , uk )
t
pk+1 pk
= f x (xk , uk )t pk + lx (xk )
t
uk = (xk , pk ) U
(69)
xk xk1
(t tk1 )
t
and
pk+1 pk
(t tk1 )
t
d
x(t) = f (x(t), u(t)),
x(0) = x0
dt
d
p(t) = f x (x(t), u(t))t p(t) + lx (x(t)),
dt
u(t) = (x(t), p(t)) U
pN)N1
such that Eq. (69) holds. Let u
N denote the step function defined by u
N(t) uk on (tk1, tk), 1 k N, and let xN
and pN be the piecewise linear functions defined by
p N (t) = pk +
1
(|xk |2
2
1
(|pk |2
2
xN (t) = xk1 +
p(T ) = gx (x(T ))
(70)
J N (uN ) J N (vN )
for all v
OPTIMAL CONTROL
for the construction of the feedback synthesis K to the problem (63)(64), which is still under investigation.
As we discussed in the Section titled Dynamic Programming Principle and HamiltonJacobyBellman Equation,
the optimal feedback law is given by K(t, x(t)) KT(t, x(t))
(x(t), Vx(t, x(t)), where V(t, x) is the solution to HJB Eq. (38)
and we stress the dependence of K on T. Let us assume that
f(x, u) f(x) Bu, and h /2u2. Then
1
KT (t, x(t)) = Bt Vx (t, x(t))
x0 Rn
(72)
G(x, x j ) j
with
G(xi , x j ) j = K(xi ),
1iM
j=1
(73)
The Greens function interpolation (73) has the following variational principle: Consider
min
Rn
|K(x)|2 dx
subject to K(xi ) = i ,
x0
(74)
over Wad, where x max(0, x). Then we set K(x) (x, Wx).
ACKNOWLEDGMENTS
K(x)
=
(71)
379
1iM
K. Itos research was supported in part by AFSOR under contracts F-49620-95-1-0447 and F-49620-95-1-0447. K. Kunischs research was supported in part by the Fonds zur Forderung der wissenschaftlichen Forschung, SFB Optimization
and Control.
BIBLIOGRAPHY
1. R. Bellman, Dynamic Programming, Princeton, NJ: Princeton
Univ. Press, 1957.
2. W. H. Fleming and R. W. Rishel, Deterministic and Stochastic Control, New York: Springer-Verlag, 1975.
3. M. G. Crandall and P. L. Lions, Viscosity solutions of Hamilton
Jacobi equations, Trans. Amer. Math. Soc., 277: 142, 1983.
4. W. Fleming and M. Soner, Controlled Markov Processes and Viscosity Solutions, Berlin: Springer-Verlag, 1993.
5. W. M. Wonham, Linear Multivariable Control: A Geometric Approach, Berlin: Springer-Verlag, 1974.
6. D. L. Russell, Mathematics of Finite-Dimensional Control Systems,
Theory and Design, New York: Marcel Dekker, 1979.
7. H. T. Banks and K. Ito, A numerical algorithm for optimal feedback gains in high dimensional linear quadratic regulator problems, SIAM J. Control Optim., 29: 499515, 1991.
Reading List
V. M. Alekseev, V. M. Tikhomirov, and S. V. Fomin, Optimal Control,
New York: Plenum, 1987.
M. Athans and P. Falb, Optimal Control, New York: McGraw-Hill,
1963.
L. D. Berkovitz, Optimal Control Theory, New York: Springer-Verlag, 1974.
A. E. Bryson and Y. Ho, Applied Optimal Control, New York: Wiley,
1975.
M. R. Hestenes, Calculus of Variations and Optimal Control, New
York: Wiley, 1963.
H. Hermes and J. L. LaSalle, Functional Analysis and Time Optimal
Control, New York: Wiley, 1969.
H. Knobloch, Higher Order Necessary Conditions in Optimal Control
Theory, Berlin: Springer-Verlag, 1981.
H. Kwakernaak and R. Sivan, Linear Optimal Control Systems, New
York: Wiley-Interscience, 1972.
B. Lee and L. Marcus, Foundations of Optimal Control Theory, New
York: Wiley, 1967.
L. W. Neustadt, Optimization: A Theory of Necessary Conditions,
Princeton, NJ: Princeton Univ. Press, 1976.
KAZUFUMI ITO
North Carolina State University
KARL KUNISCH
Karl-Franzens-Universitat Graz
380
PERIODIC CONTROL
59
PERIODIC CONTROL
The fact that a periodic operation may be advantageous has
been well known to humankind since time immemorial. All
farmers know that it is not advisable to grow the same product repeatedly in the same field because the yield can be improved by rotating crops. So, cycling is good.
More recently, similar concepts have been applied to industrial problems. Traditionally, almost every continuous industrial process was set and kept, in the presence of disturbances, at a suitable steady state. However, there are
circumstances under which a periodic time-varying action
proves to be better. This observation germinated in the field
of chemical engineering where it was seen that the performance of a number of catalytic reactors improved by cycling;
see the pioneering contributions in Refs. 13. Unfortunately,
as pointed out in Ref. 4, periodic control was still considered
too advanced in the industrial control scenario, in that the
steady-state operation is the norm and unsteady process behaviour is taboo. Its use was therefore confined to advanced
(aerospace or classified) applications, such as those treated in
Refs. 5 and 6. Today, however, the new possibilities offered
by current control technology, together with the theoretical
developments of the field, have opened the way for using periodic controllers in place of the traditional stationary ones. In
fact, the term periodic control takes a wider significance in
the contemporary literature. In addition to the control problems that arise when operating a plant periodically, periodic
control also includes all situations where either the controller
or the plant is a proper periodic system. One of the reasons
behind such an extension is the possible improvement of the
performances, in terms of stability and robustness, of plants
described by time-invariant models, when using a periodic
controller (see Ref. 7).
The diffusion of digital apparatuses in control has also
contributed to the increasing importance of periodic control
because computer-controlled systems are often based on sample-and-hold devices for output measurements and input updating. In multivariable control, it may also be necessary, for
technological or economical reasons, to adopt different sampling and/or hold intervals for the various actuators or transducers. For example, certain variables may exhibit a much
slower dynamic than others so that different sampling interJ. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
60
PERIODIC CONTROL
B(t + T ) = B(t)
C(t + T ) = C(t),
D(t + T ) = D(t)
PERIODIC CONTROL
and matrices
f ( , v)
A(t) =
,
= ,v= v
h( , v)
C(t) =
,
= ,v= v
f ( , v)
B(t) =
v = ,v= v
h( , v)
D(t) =
v = ,v= v
u1 (t)
x(t)
= Ax(t) + B
u2 (t)
u (t)
y1 (t)
= Cx(t) + D 1
y2 (t)
u2 (t)
yi (k) = y(ky i ),
u j (t) = u j (k),
y(k)
= y(k),
u=v~
v
Periodic regime
~~~
u,x,y
j = 1, 2
u(k)
= u(k)
= N(k)y(k)
N(k) =
tu j (k) = ku j
= f( ,v)
= h ( ,v)
t [ku j , ku j + u j ),
n1 (k)
0
ni (k) =
1
0
if k is a multiple of y i /
otherwise
s (k)
S(k) = 1
0
0
s2 (k)
with
s j (k) =
y = ~
0
n2 (k)
with
Time-invariant
nonlinear system
v
i = 1, 2
where
61
0 if k is a multiple of u j /
1 otherwise
t [k, k + )
where the fast updated signal u(k) is obtained from the slow
one u
(k) according to the holding selector mechanism
u
u(k)
= S(k)v(k) + (I S(k))u(k)
Periodic linear
system
Figure 1. Linearization around a periodic orbitthe dynamics of the
nonlinear system in the orbit vicinity is governed by a linear periodic system.
62
PERIODIC CONTROL
~
u
^
u
S (.)
Holding selector
(period 12)
u
ZOH
A,B,C,D
Zero-order
holder
System
u1
y^
N (.)
Sampler
( )
~
y
Sampling selector
(period 10)
y1
y
u
u2
y2
~
u
i (slow updating)
~
y i (slow sampling)
^
u
i (fast updating)
Figure 2. A multirate sampled-data system with two inputs and two outputs. The symbol
denotes the clock period. The first output signal y1 is sampled at rate y1 2 and the second y2
at rate y2 5. Hence, the sampling selector is a periodic discrete-time system with period
Ty 10. Moreover, the first input signal u1 is updated at rate u1 3 and the second u2 at rate
u2 4. The holding selector is a periodic discrete-time system with period Tu 12. The period
of the global system is therefore T 60.
The overall multirate sampled-data system is a discretetime periodic system with state
x(k)
x(k)
=
v(k)
where
A(k)
=
B(k)
=
and equations
u(k)
x(k
+ 1) = A(k)
x(t)
+ B(k)
x(k)
y(k)
= C(k)
+ D(k)
u(k)
A
0
0
e
0
d BS(k)
S(k)
e A d B(I S(k))
I S(k)
C(k)
= N(k)[C DS(k)]
D(k)
= N(k)D(I S(k))
PERIODIC CONTROL
in continuous time. Therefore, the state solution with a generic initial state x() and input function u( ) is
t
j= +1
Input-Output Representation
Another mathematical representation of periodic systems is
based on a direct time-domain relationship between the input
and the output variables without using any intermediate latent variables. In discrete time, this leads to a model of the
form
TIME-INVARIANT REPRESENTATIONS
Reversibility. In continuous time, there is no analytic expression for the transition matrix. However, its determinant
can be worked out from the so-called Jacobi formula. In other
words,
The simplest way to achieve stationarity is to resort to a sample-and-hold procedure. Indeed, with reference to a continuous or a discrete-time periodic system, suppose that the input
is kept constant over a period, starting from an initial time
point . That is,
u(t) = u(k),
t [kT + , kT + T + )
As seen before, periodicity is often the result of ad hoc operations over time-invariant systems. On the other hand, in periodic systems analysis and control, a major point is to address
the backward problem of finding a way to transform a periodic system into a time-invariant one. In such a way, we can
resort to the results already available in the time-invariant
realm.
Sample and Hold
63
trace[A( )] d
64
PERIODIC CONTROL
where
( ) =
T +
A (T + , )B( ) d
T
+ 1
A (T + , i + 1)B(i)
in continuous time
in discrete time
i=
t [kT + , kT + T + )
B( + T 1)]
H = [C( )
A ( + 1, ) C( + 1) . . .
A ( + T 1, )C( + T 1) ]
E = {(E )ij },
(E )ij =
v( )
..
v ( ) = . ,
v( + 1)
v ( + 1) =
, . . .,
..
.
..
v ( + T 2) =
,
v( + T 2)
..
v ( + T 1) =
v( + T 1)
where is any element, typically set to zero. Obviously, the
previous pattern repeats periodically for the other periods.
This signal transformation is used for the input, output, and
state of the system. Then, we can relate the cyclic input to
the cyclic state by means of a time-invariant state-equation
and the cyclic output to the cyclic state via a time-invariant
transformation. In this way, we obtain an nT-dimensional
time-invariant system with mT inputs and pT outputs.
Finally, the frequency-lifted reformulation is based on the
following considerations. For a discrete-time (vector) signal
v(t), let V(z) be its z-transform. Now, one can associate with
V(z) the frequency augmented vector Vf (z) as follows:
F = A ( )
G = [A ( + T, + 1)B( ) A ( + T, + 2)B( + 1) . . .
x (k + 1) = F x (k) + G u (k)
i, j = 1, 2, . . ., T
i< j
D( + i 1)
i= j
V (z)
V (z)
V (z 2 )
V f (z) =
..
T 1
V (z
)
where e2j/T. By applying this procedure to the z-transforms of the input and output signals of the periodic system,
it is possible to establish an input-output correspondence described by a matrix transfer function; see Ref. 33. Such a
PERIODIC CONTROL
uk e s k t
kZ
65
Consider now the Fourier series for the periodic matrix coefficients. That is,
A(t) =
Ak e jkt
kZ
and similarly for B(t), C(t), and D(t), and plug the expansions
of the signals x(t), u(t), x(t) and the matrices A(t), B(t), C(t),
D(t) into the system equations. By equating all terms at the
same frequency, we obtain an infinite-dimensional matrix
equation of the following kind:
X = (A
A N )X
X + BU
sX
Y = CX + DU
where X , U , and Y , are doubly infinite vectors found with the
harmonics of x, u and y respectively, organized in the following fashion:
X
..
A =
..
.
A0
A1
A2
A3
A4
..
.
..
.
..
.
..
.
..
.
A1
A0
A1
A2
A3
..
.
A2
A1
A0
A1
A2
..
.
A3
A2
A1
A0
A1
..
.
A4
A3
A2
A1
A0
..
.
..
.
where
sk = s + jk
The quantity T 2/ is the named period of the EMP signal. The class of EMP signals is a generalization of the class
of T-periodic signals. As a matter of fact, an EMP signal with
s 0 is just an ordinary time-periodic signal. Indeed, as it is
easy to verify, an EMP signal is such that
u(t + T ) = u(t),
= esT
x(t) =
xk e s k t
kZ
x(t)
=
s k xk e s k t
kZ
y(t) =
kZ
yk e s k t
N = blkdiag{ jkI},
kZ
66
PERIODIC CONTROL
e AT
A ( ) =
AT
in continuous time
in discrete time
region (i.e., the left half plane in continuous time and the unit
disk in discrete time) and nevertheless the system is unstable. Notable exceptions in continuous time are slowing-varying matrices or high-frequency perturbed matrices, see Refs.
38 and 39, respectively.
A celebrated stability condition can be formulated in terms
of the so-called Lyapunov equation.
There are two possible formulations of such an equation,
known as filtering Lyapunov equations and control Lyapunov
equation, reflecting the fact that Lyapunov equations may
arise in the analysis of both filtering and control problems. In
continuous time, the filtering Lyapunov equation takes the
form
P(t)
= P(t)A(t) + A(t)P(t) + Q(t)
and the control Lyapunov equation is
P(t)
= A(t) P(t) + P(t)A(t) + Q(t)
where Q( ) is a periodic [Q(t T) Q(t), t] and positive
definite [xQ(t)x 0, t, x 0] matrix.
It turns out that the continuous-time periodic system is
stable if and only if the Lyapunov equation (in any of the
two forms above) admits a (unique) periodic positive-definite
solution P( ).
An analogous result holds in discrete time, by making reference to
P(t + 1) = A(t)P(t)A(t) + Q(t)
P(t) = A(t) P(t + 1)A(t) + Q(t)
as filtering and control Lyapunov equations, respectively. As
before, Q( ) is periodic and positive definite.
The Lyapunov stability theorem can be expressed in a
more general form by referring to positive semidefinite matrices Q( ) provided that further technical assumptions on the
pair (A( ), Q( )) are met with; see Ref. 40 for more details on
the theoretical aspects and Ref. 41 for the numerical issues.
It is useful to point out that the above Lyapunov stability
condition can also be stated in a variety of different forms. In
particular, it is worth mentioning that one can resort to the
Lyapunov inequality, i.e., in continuous time
P(t)
> A(t)P(t) + P(t)A(t)
(filtering)
P(t)
> A(t) P(t) + P(t)A(t) (control)
and in discrete time
P(t + 1) > A(t)P(t)A(t)
(filtering)
PERIODIC CONTROL
Cyclostationary Processes
In the stochastic realm, the so-called cyclostationary processes
are well suited to deal with pulsatile random phenomena and
are the subject of intense investigation in signal processing;
see Ref. 42. Specifically, a stochastic process with periodic
mean and covariance function (t, ) satisfying the biperiodicity condition (t T, T) (t, ) is said to be a cyclostationary process. In particular, its variance (t, t) is T-periodic.
The periodic Lyapunov equation serves as a fundamental
tool in the analysis of these processes. Assume that the initial
state x(0) of the system is a random variable with zero mean
and covariance matrix P0, and assume also that the input of
the system is a white-noise process, independent of x(0), with
zero mean and unitary intensity. Then (21), under the stability assumption, the state of the periodic system asymptotically converges to a zero mean cyclostationary process with
variance (t, t), which can be computed via the periodic filtering Lyapunov equation by letting Q(t) B(t)B(t) and P(0)
P0. It turns out that
lim { (t, t) P(t)} = 0
Floquet Theory
One of the long-standing issues in periodic systems is
whether it is possible to find a state-coordinate transformation leading to a periodic system with constant dynamic matrix. In this way, the eigenvalues of such a dynamic matrix
would determine the modes of the system. With reference to
linear differential equations, this issue was considered by various mathematicians of the nineteenth century. Among them,
a prominent role was played by the French scientist Gaston
Floquet (18471920) who worked out a theory to solve linear
homogeneous periodic systems, which is now named after
him (43).
This theory can be outlined in a simple form as follows. If
S( ) is a T-periodic invertible state-space transformation,
(t)
x(t) S(t)x(t), then, in the new coordinates, the dynamic A
is given by
1
in continuous time
S(t)A(t)S(t)1 + S(t)S(t)
A(t) =
1
S(t + 1)A(t)S(t)
in discrete time
The Floquet problem is then to find S(t) (if any) in order to
(t) A
.
obtain a constant dynamic matrix A
In continuous time, it can be shown that such a transformation S( ) does exist, and the Floquet problem can be
can be obtained by solving eAT A(),
solved. Indeed, A
where is any given time point. The appropriate transformation S( ) is simply given by
S(t) = e A(t ) A ( , t)
67
S(t + 1) = AS(t)A(t)
x(T ) = x(0)
S(t)
= AS(t)
S(t)A(t)
J=
1
T
g(x(t), u(t)) dt
0
68
PERIODIC CONTROL
Uk e
k=
jkt
2
=
T
Uk (k)Uk
k=
1
T
y(t) y(t) dt
u(t) u(t) dt 1
t [kT, kT + T )
PERIODIC CONTROL
x(kT + T ) = Ac x(kT )
where
Ac = e
AT
A(T )
BF ( )C d
F (t) = B eA
(T t )
e A(T ) BB e A
(T )
69
x(t)
= (A(t) + B(t)K(t))x(t) + B(t)S(t)v(t)
in continuous time and
x(t + 1) = (A(t) + B(t)K(t))x(t) + B(t)S(t)v(t)
1
d
J=
in continuous time
in discrete time
k=0
70
PERIODIC CONTROL
(t) and D
(t) are to be tuned by the designer. In this
where C
way, the performance index can also be written in the (perhaps more popular) form
tf
{x( ) Q( )x( )
()C
(), S() D
(), and R()
()C
where Q() C
()D
(). We will assume for simplicity that the problem is
D
nonsingular [i.e., R() 0, ].
This problem is known as the linear quadratic (LQ) optimal control problem. To solve it, the auxiliary matrix equation
P(t) + P(t)A(t)
P(t)
= A(t)
P(t)B(t)R(t)1B(t) P(t) + Q(t)
is introduced. This is the well-known differential Riccati equation, in one of its many equivalent forms. More precisely, because the coefficients are periodic, the equation is referred to
as the periodic differential Riccati equation.
Let ( , tf) be the backward solution of the periodic Riccati
equation with terminal condition (tf, tf) Ptf . Assuming that
the state x( ) can be measured, the solution to the minimization problem can be easily written in terms of ( , tf) as follows
tf
z( ) z( ) d
z(t) = C(t)x(t)
+ D(t)u(t)
where
A(t)
= A(t) B(t)R(t)1 S(t),
Q(t)
= Q(t) S(t) R(t)1 S(t)
u( ) = o ( , tf )x( )
where
o ( , tf ) = R( )1 [B( ) ( , tf ) + S( )]
tf
tf
PERIODIC CONTROL
If the state is not accessible, we must rely instead on the measurable output
y(t) = C(t)x(t) + D(t)u(t)
A first task of the controller is then to infer the actual value
of the state x(t) from the past observation of y( ) and u( ) up
to time t. This leads to the problem of finding an estimate
x(t) of x(t) as the output of a linear system (filter) fed by the
available measurements. The design of such a filter can be
carried out in a variety of ways, among which it is worth mentioning the celebrated Kalman filter, the implementation of
which requires the solution of another matrix Riccati equation with periodic coefficients. When x(t) is available, the control action is typically obtained as
u( ) = Ko ( )x(
)
J = x(tf ) Pt x(tf ) +
f
tf
tf T
{x( ) Q( )x( ) + u( ) u( )} d
This condition is usually referred to as cyclomonotonicity condition. Now, consider the periodic extension Pe( ) of ( , tf)
Pe (t + kT ) = (t, tf ),
71
t (tf T, tf ],
integer k
Optimal periodic
gain
x^
Periodic Kalman
filter
(Filtering Riccati eq.)
72
PERIODIC CONTROL
5. J. L. Speyer, On the fuel optimality of cruise, AIAA J. Aircraft,
10: 763764, 1973.
6. G. Sachs and T. Christopoulou, Reducing fuel consumption of
subsonic aircraft by optimal cycling cruise, AIAA J. Aircraft, 24:
616622, 1987.
7. P. P. Khargonekar, K. Poolla, and A. Tannembaum, Robust control of linear time-invariant plants using periodic compensation,
IEEE Trans. Autom. Control, AC-30: 10881096, 1985.
8. M. A. Dahleh, P. G. Vulgaris, and L. S. Valavani, Optimal and
robust controllers for periodic and multirate systems, IEEE
Trans. Autom. Control, 37: 9099, 1992.
9. M. Araki, Recent developments in digital control theory, 12th
IFAC World Cong., Sidney, vol. 9, 251260, 1993.
10. W. Johnson, Helicopter Theory, Princeton, NJ: Princeton University Press, 1981.
11. D. P. Schrage and D. A. Peters, Effect of structural coupling of
parameters on the flap-lag forced response of rotor blades in
forward flight using Floquet theory, Vertica, 3 (2): 177185,
1979.
12. R. M. McKillip, Periodic control of individual blade control helicopter rotor, Vertica, 9 (2): 199225, 1985.
13. D. Teves, G. Niesl, A. Blaas, and S. Jacklin, The role of active
control in future rotorcraft, 21st Eur. Rotorcraft Forum, Saint
Petersburg, Russia, III.10.117, 1995.
14. W. Wiesel, Perturbation theory in the vicinity of a stable periodic orbit, Celestial Mechanics, 23: 231242, 1981.
15. J. Hauser and C. C. Chunt, Conserse Lyaupov functions for exponentially stable periodic orbits, Systems and Control Letters,
23: 2734, 1994.
16. C. C. Chung and J. Hauser, Nonlinear H control around periodic orbits, Syst. Control Lett., 30: 127137, 1997.
17. G. O. Guardabassi, A. Locatelli, and S. Rinaldi, Status of periodic optimization of dynamical systems, J. Opt. Theory Appl., 14
120, 1974.
18. E. Noldus, A survey of optimal periodic control of continuous
systems, Journal A, 16: 1116, 1975.
19. G. Guardabassi, The optimal periodic control problem, Journal
A, 17: 7583, 1976.
20. S. Bittanti and G. O. Guardabassi, Optimal periodic control and
periodic system analysis: An overview, 25th IEEE Conf. Decision
Control, pp. 14171423, 1986.
21. S. Bittanti, Deterministic and stochastic periodic systems, in S.
Bittanti (ed.), Time Series and Linear Systems, Lecture Notes in
Control and Information Sciences, vol. 86, pp. 141182, Berlin:
Springer-Verlag, 1986.
22. S. Bittanti and P. Colaneri, Analysis of discrete-time linear periodic systems, in C. T. Leondes (ed.), Control and Dynamic Systems, vol. 78, San Diego: Academic Press, 1996.
23. A. Marzollo, Periodic Optimization, vols. 1, 2, New York:
Springer Verlag, 1972.
24. V. A. Yakubovich and V. M. Starzhinskii, Linear Differential
Equations with Periodic Coefficients, New York: Wiley, 1975.
25. T. Kaczorek, Two-Dimensional Linear Systems, New York:
Springer Verlag, 1995.
26. S. Bittanti, P. Bolzern, and G. O. Guardabassi, Some critical
issues on the state representation of time-varying ARMA models, IFAC Symp. Identification System Parameter Estimation,
York, pp.15791583, 1985.
27. P. Colaneri and S. Longhi, The minimal realization problem for
discrete-time periodic systems, Automatica, 31: 779783, 1995.
28. P. T. Kabamba, Control of linear systems using generalized
sampled-data hold functions, IEEE Trans. Autom. Control, 32:
772783, 1987.
PERIODIC CONTROL
54. D. S. Bernstein and E. G. Gilbert, Optimal periodic control: The
-test revisted, IEEE Trans. Autom. Control, 25: 673684, 1980.
55. J. Speyer and R. T. Evans, A second variational theory of optimal periodic processes, IEEE Trans. Autom. Control, 29: 138
148, 1984.
56. Q. Wang and J. Speyer, Optimal periodic control: A general theory of necessary conditions, SIAM J. Control and Optimization,
15: 717746, 1990.
57. Q. Wang, J. Speyer, and L. D. Dewell, Regulators for optimal
periodic processes, IEEE Trans. Autom. Control, 40: 1777
1778, 1995.
58. S. Lee, S. Meerkov, and T. Runolfsson, Vibrational feedback control: Zero placement capabilities, IEEE Trans. Autom. Control,
32: 604611, 1987.
59. S. Bittanti, G. Fronza, and G. Guardabassi, Periodic optimization of linear systems under control power constraints, Automatica, 9 (1): 269271, 1973.
60. S. Bittanti and G. Guardabassi, Optimal cyclostationary control:
A parameter-optimization frequency domain approach, VIII
IFAC World Congr., 6: 191196, 1981.
61. V. L. Syrmos, C. T. Abdolloh, D. Doreto, and K. Grigoriedis,
Static output feedbackA survey, Automatica, 33: 125137,
1997.
62. V. Blondel and M. Gevers, Simultaneous stabilization of three
linear systems is rationally undecidible, Math. Control, Signals
Syst., 6: 135145, 1994.
63. B. D. O. Anderson and J. Moore, Time-varying feedback laws
for decentralized control, IEEE Trans. Autom. Control, 26: 1133
1139, 1981.
64. A. B. Chammas and C. T. Leondes, On the design of linear timeinvariant systems by periodic output feedback, Int. J. Control,
27: 885903, 1978.
65. B. A. Francis and T. T. Georgiou, Stability theory for linear
time-invariant plants with periodic digital controllers, IEEE
Trans. Autom. Control, 33: 820832, 1988.
66. H. M. Al-Rahmani and G. F. Franklin, Linear periodic systems:
Eigenvalues assignment using discrete periodic feedback, IEEE
Trans. Autom. Control, 34: 99103, 1989.
67. A. B. Chammas and C. T. Leondes, Pole assignment by
piecewise constant output feedback, Int. J. Control, 44: 1661
1673, 1986.
68. J. L. Willems, V. Kucera, and P. Brunovski, On the assignment
of invariant factors by time-varying feedback strategies, Syst.
Control Lett., 5: 7580, 1984.
69. T. Kaczorek, Pole placement for linear discrete-time systems by
periodic output feedbacks, Syst. Control Lett., 6: 267269, 1985.
70. A. W. Olbrot, Robust stabilization of uncertain systems by periodic feedback, Int. J. Control, 45 (3): 747758, 1987.
71. R. Ortega and G. Kreisselmeier, Discrete-time model reference
adaptive control for continuous time systems using generalized
sample data hold functions, IEEE Trans. Autom. Control, 35:
334338, 1990.
72. K. L. Moore, S. P. Bhattacharyya, and M. Dahleh, Capabilities
and limitations of multirate control schemes, Automatica, 29:
941951, 1993.
73. A. Feuer and C. A. Goodwin, Generalized sampled-data functions: Frequency domain analysis of robustness, sensitivity, and
intersample difficulties, IEEE Trans. Autom. Control, 39: 1042
1047, 1994.
74. A. Feuer, C. A. Goodwin, and M. Salgado, Potential benefits of
hybrid control for linear time-invariant plants, Amer. Control
Conf., Albuquerque, 1997.
73
75. D. Aeyels and J. Willems, Pole assignment for linear time-invariant systems by periodic memoryless output feedback, Automatica, 28: 11591168, 1992.
76. H. Chapellat and M. Dahleh, Analysis of time-varying control
strategies for optimal disturbance rejection and robustness,
IEEE Trans. Autom. Control, 37: 17341745, 1992.
77. P. Brunovski, A classificaton of linear controllable systems, Kybernetika, 3: 173187, 1970.
78. M. Kono, Pole-placement problem for discrete-time linear periodic systems, Int. J. Control, 50 (1): 361371, 1989.
79. V. Hernandez and A. Urbano, Pole-placement problem for discrete-time linear periodic systems, Int. J. Control, 50: 361
371, 1989.
80. P. Colaneri, Output stabilization via pole-placement of discretetime linear periodic systems, IEEE Trans. Autom. Control, 36:
739742, 1991.
81. O. M. Grasselli and S. Longhi, Pole placement for nonreachable
periodic discrete-time systems, Math. Control, Signals Syst., 4:
437453, 1991.
82. P. Colaneri and V. Kucera, The model matching problem for discrete-time periodic systems, IEEE Trans. Autom. Control, 42:
14721476, 1997.
83. O. M. Grasselli and S. Longhi, Robust tracking and regulation
of linear periodic discrete-time systems, Int. J. Control, 54: 613
633, 1991.
84. J. R. Canabal, Periodic geometry of the Riccati equation, Stochastics, 1: 432435, 1974.
85. E. C. Bekir and R.S. Bucy, Periodic equilibria for matrix Riccati
equations, Stochastics, 2: 1104, 1976.
86. H. Kano and T. Nishimura, Periodic solutions of matrix Riccati
equations with detectability and stabilizability, Int. J. Control,
29: 471481, 1979.
87. S. Bittanti, P. Colaneri, and G. Guardabassi, Periodic solutions
of periodic Riccati equations, IEEE Trans. Autom. Control, 29:
665667, 1984.
88. V. Hernandez and L. Jodar, Boundary problems for periodic Riccati equations, IEEE Trans. Autom. Control, 30: 11311133,
1985.
89. M. A. Shayman, On the phase portrait of the matrix Riccati
equation arising from the periodic control problem, SIAM J.
Control Optim., 23: 717751, 1985.
90. S. Bittanti, P. Colaneri, and G. Guardabassi, Analysis of the
periodic Riccati equation via canonical decomposition, SIAM J.
Control Optim., 24: 11381149, 1986.
91. C. E. de Souza, Riccati differential equation in optimal filtering
of periodic non-stabilizing systems, Int. J. Control, 46: 1235
1250, 1987.
92. S. Bittanti, P. Colaneri, and G. De Nicolao, The difference Riccati equation for the periodic prediction problem, IEEE Trans.
Autom. Control, 33: 706712, 1988.
93. S. Bittanti, A. J. Laub, and J. C. Willems (eds.), The Riccati
Equation, Berlin: Springer Verlag, 1991.
94. J. J. Hench and A. J. Laub, Numerical solution of the discretetime periodic Riccati equation, IEEE Trans. Autom. Control, 39:
11971210, 1994.
95. D. W. Clarke, C. Mohtadi, and P. C. Tuffs, Generalized predictive control: part 1The basic algorithm; part 2Extensions
and interpolators, Automatica, 23: 137160, 1987.
96. R. R. Bitmead, M. R. Gevers, I. R. Petersen, and R. J. Kaye,
Monotonicity and stabilizability properties of solutions of the
Riccati difference equation: Propositions, lemmas, theorems, fallacious conjectures and counterexamples, Syst. Control Lett., 5:
309315, 1986.
74
S. BITTANTI
P. COLANERI
Politecnico di Milano
446
PID CONTROL
PID CONTROL
PID control strategies are by far the most widely employed of
all strategies used in the automatic feedback control of industrial systems. The acronym PID stands for proportional, integral, derivative. Although the central concept of PID control
can be gleaned from an examination of how these three terms
are blended to form a control signal, the intelligent application of PID control in any given case requires an understanding of linear and nonlinear properties of the system that is
being controlled, and practical PID controllers incorporate
features such as bumpless transfer and anti reset windup.
The proportional term of the PID controller forms a part
of the control signal based on a proportional gain times a system error. The derivative term of the PID controller forms a
part of the control signal based on a gain times the rate of
change of a system error. Similarly, the integral term of the
PID controller forms a part of the control signal based on a
gain times the integral of a system error. This integral term
basically forces the error from a nonzero value (a value that
would exist without integral action present) to a zero value
the integral term in effect resets the error to zero. Because of
this phenomenon, the integral term of the PID controller is
sometimes called the reset term. This terminology is especially prevalent in older literature; Eckman, for example, uses
reset consistently in place of integral (1).
Much of this article is devoted to practical issues of PID
controllers. Several alternative forms of PID algorithms are
examined, and their relative merits are discussed. Rules for
tuning the parameters of the controller are considered, the
literature base devoted to this topic is exceedingly large. Tuning rules are concerned with ways of assigning controller parameters (proportional gain, integral gain, and derivative
gain) to achieve good performance based on various process
model assumptions. That there are a variety of tuning rules
should come as no surprise; generally there are tradeoffs to be
made in alternative performance criteria and different tuning
rules weight some performance criteria more than others.
Tuning rules tend to emphasize system performance in the
neighborhood of an operating point, and therefore generally
are based on linearized models of the process being controlled.
Tuning rules are but one aspect of PID controller design.
Of equal importance is how the system performs when large
input changes occur. For most systems, the control signal input to the system is limited by maximum and minimum
bounds; when the control signal is saturated at one of these
bounds for any length of time, the error between the desired
system output and the actual system output can be quite
large. Within the PID controller, this error signal is being integrated, and unless some mechanism is employed to control
the integration process (either by selectively turning it off and
on or by selectively forcing the integrator output to track an-
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
PID CONTROL
P
r(t)
e(t)
I
kpe(t)
+
ki tedt +
+
d(t)
+
u(t)
+
c(t)
Process
kdde/dt
447
(1)
(2)
Assuming Ks 0, it would appear that ess could be made arbitrarily small by making kp arbitrarily large; however, for any
given system, an upper limit ku on kp invariably exists such
that the closed loop system is unstable for kp ku, where the
ultimate value ku depends on the system dynamics.
The Integral Term
The steady-state error described above generally can be reduced to zero by the inclusion of the integral term. With ki
0, kp 0, and kd 0, the controller of Fig. 1 is a PI controller.
In this case, consider what would happen to the integrator
output if it started at 0 at t 0, and the error signal e was
constant at e ess 0; the integrator output would then
equal kiesst, but this increases indefinitely as t increases! The
logical conclusion is that ess must be zerothe output of the
integrator is part of the control signal u(t), and as u(t) increases, so also does the controlled output c(t) to the point
where rss css ess 0 thereby inhibiting further increases
in the output of the integrator.
The Derivative Term
In Fig. 1, if ki 0 with kp 0 and kd 0, the controller is a
PD controller. If all three gain terms are nonzero, the controller is a three-term PID controller. The kd term plays the role
of an anticipatory or predictive element. That is, even if the
e(t1) is zero, if e(t1) de(t1)/dt is large, the implication is that
the error for t t1 is going to increase. By including kde in
the control, the controlled output is forced to increase sooner
than it would otherwise, with the goal of reducing the anticipated future error. This action is useful in compensating for
dynamic lags that invariably are present in the process being
controlled. However, if significant measurement noise is present in the measured output signal, the kd term can have a
detrimental effect on performance (the derivative of rapidly
changing high-frequency noise can be extremely large even if
the magnitude of the noise is small). Because of this, and because the output of any real circuit is band-limited, practical
implementation of the kd term requires the use of a frequencydependent differentiator, as is described later in this article.
448
PID CONTROL
Performance Objectives
The values of kp, ki, and kd should be selected to provide good
system response under design conditions, and to reduce
harmful effects of process changes that lead to off-design conditions. System response is often described by step response
characteristics such as: (1) the 1090% rise time resulting
from a step change in the reference input; (2) the percent
overshoot in response to a step input; (3) the peak time associated with the peak overshoot in response to a step input; and
(4) the time required for the system output to settle to within
2% of its final value in response to a step input. Of interest
are response characteristics caused by process disturbances
(represented by d(t) in Fig. 1) in addition to responses activated by reference inputs r(t). Note that for constant disturbance values, with d(t) dss, the reason given in the previous
Subsection on the integral term for ess being 0 still holds, and
the controlled output c(t) approaches the steady-state reference input rss independent of the dss value, assuming of course
that practical upper and lower bounds on the output of the
integrator are not exceeded.
If the reference input, in addition to the error signal is
available to the controller, a modified reference input may be
beneficial. For example, if a controlled output exhibits too
much overshoot in response to a step change in a reference
input, a rate-limited version of the reference input can be generated and inserted in place of the actual reference input.
Conversely, if a controlled output reacts too slowly in response to a change in a reference input, a rate-enhanced version of the reference input can be generated and inserted in
place of the actual reference input. In some applications, future values of a reference input are available to the controller;
in machine-tool control, for example, a desired contour to be
traversed is known in advance. This look-ahead or preview
information can be used to reduce the error between the reference input and the controlled output.
ALTERNATIVE PID FORMS
The control action of Fig. 1 is expressed as
u(t) = kp e(t) + ki
t
0
e() d + kd
de(t)
dt
(3)
(s)
k
= kp + i + kd s
, UE(s)
s
(4)
The above form of the PID terms is called the parallel form.
An alternative form to that of Eq. (4) is called the standard
or noninteracting form and is characterized by
1
+ Td s
Gc (s) = K 1 +
(5)
Ti s
with the obvious relationships kp K, ki K/Ti, and kd
KTd. Although these two forms are equivalent, the standard
form of Eq. (5) uses different terminology: the Ti constant is
called the integral time; and the Td constant is called the de-
(7)
Whereas the feedback system of Fig. 1 has one degree of control freedom, that of Fig. 2 has two degrees of control freedom:
(1) the kp, ki, and kd gains can be selected to achieve desired
goals regarding closed-loop stability, disturbance rejection,
and sensitivity reduction; and (2) the kp and kd gains can be
selected to achieve performance goals associated with the re-
PD
R(s)
E(s)
D(s)
ki
s
+
+
U(s)
+
PD
C(s)
G(s)
+
+
N(s)
PID CONTROL
ep
ed
449
k ps
1 + ps
+
ki
1
s
u(t)
Integrator as
part of actuator
k ps 2
(1 + ds2)
kd s
1 + s
(8)
(s)
GPID (s)
=
, UN(s)
1 + G (s)G(s)
(9)
PID
where
GPID (s) = kp +
kd s
ki
+
s
1 + s
(10)
(11)
(12)
(13)
Equations (12) and (13) correspond to the two-degree-of-freedom control loop of Fig. 2.
A potential problem with the velocity PID form is that the
second-derivative of an error signal must be approximated;
for the same reasons described in the preceding section, this
process leads to difficulties in the presence of significant measurement noise. Also, with the integrator external to the controller, it is essential that ki not be zero in Fig. 3 so that no
attempt is made to cancel the 1/s of the integrator by the s
inherent in a velocity form of the PD controller. This means
that the velocity PID of Fig. 3 has the integral term as an
essential term of the controller. However, if the control signal
strom and Hagglund
u(t) is available as a measured signal, A
(2) show how to use it to obtain a valid PD form from the
velocity PID form.
The velocity form of the PID controller does have some advantages: (1) if the integrator in the actuator is implemented
in a way that limits the output of the integrator at the saturation bound (when either the upper or lower saturation bound
is intercepted), then no integrator windup can occur, and no
special anti-reset-windup strategy need be implemented; (2)
if the control is transferred from automatic to manual or visa
versa, the signal supplied to the actuator is of an incremental
control form, and the resulting transfer in control generally
will not cause a large swing in the process outputthe transfer of control will be bumpless; and (3) when the velocity PID
is implemented digitally (in which case it is called an incremental PID algorithm), the PID values being accumulated
(corresponding to those at the summation point in Fig. 3) are
incremental in nature, and often can be accommodated by a
450
PID CONTROL
umax umin
kp
(14)
Alternatively, we say in this case that the error is in the proportional band if emin e(t) emax.
General Case of Proportional Band
When all three terms are present in the PID controller, but
d(t) 0 in Fig. 1, the control signal u(t) will be out of saturation only if the instantaneous e(t) r(t) c(t) satisfies
umin ki
e() d kd e(t)
Conditional integration schemes tend to be heuristic and application dependent. A given approach may work reasonably
well in one application, only to fail in another. Because of
their use of logic, conditional integration schemes are readily
incorporated in digital implementations of PID. One approach
is to condition the integration on both kpe(t) and kde(t): if
either one of these values exceeds preassigned bounds, the
integration process is suspended and the output of the integrator is reset to a desirable level (perhaps zero); on the other
hand, if both values are less then their respective bounds,
the integration process is uninhibited; and in either case, the
control signal u(t) can be formed on the basis of the three
PID terms.
Anti-reset-windup for Two-degree-of-freedom Control
A basic tracking implementation of anti-reset windup is depicted in Fig. 4. The output from the PD and PD block is
q(t) = kp r(t) kp c(t) + kd r(t)
kd c(t)
(16)
(15)
Thus, the proportional band varies in a complicated way, depending both on the integral and the derivative of the error.
u=
umax, w umax
u ,wu
min
min
ANTI-RESET-WINDUP
The control signal of a given system can be in saturation for
long periods of time for many reasons. For example, if a large
step increase in the reference input occurs at t 0, the error
signal for some time after t 0 also will be large, and its
direct effect on the control signal will be to force it into saturation until the system output is able to match to some degree
the reference input. Although the effective control signal is in
saturation, the integrator in the controller accumulates the
area under the large e(t) curve, unless restricted to do otherwise. If the integrator output is not restricted in some way, it
takes a relatively long time for the integrator output to reduce to a normal level; it does not even start to decrease until
after the controlled output c(t) has overshot the desired reference input value [causing e(t) to turn negative and allowing
the output of the integrator to start to decrease]. Thus it is
essential that the integrator output be managed in some effective way whenever the control signal is in saturation. It is
of interest to note that windup is not limited to PID control-
PD & PD'
e=rc
ki
(17)
q
+
+
1
s
kt
+
p +
PID CONTROL
e=rc
PD
w
+
+
1
1 + sTi'
1
U (s) = 1 +
sTi
Q(s)
BUMPLESS TRANSFER
When a feedback loop is switched from an automatic mode of
control to a manual mode of operation or vice versa, it is often
important to avoid large instantaneous jumps in the control
action. Certain forms of PID implementation lend themselves
in a natural way to smooth transitions from one mode of control to another. Velocity (incremental) PID implementations
supply incremental values to be accumulated by an integrating actuator, so large transitions in effective control are not
incurred when the source of the incremental changes is
switched.
For a general parallel implementation of PID control,
tracking can be used to obtain bumpless transfer in addition
to anti-reset-windup. Figure 6 is a diagram of such a system.
The relationship between w and u in this diagram is that
given by Eq. (17). When the switch is in the manual mode, as
shown, the human operator supplies a value em to drive the
system: if the nonlinear block is not saturated, the control u
increases (or decreases) at a rate given by kmem; at the same
time, the signal uPID tracks um because of the upper tracking
loop, with tracking gain kt; and when the nonlinear block is
saturated, the lower tracking loop insures that the output of
the lower integrator will be maintained at either umax or umin,
as appropriate. When the switch to automatic control is made,
the uPID value starts with the most recent tracked value of
um plus any additional component supplied by the PD terms
operating on error signals involving r(t) and c(t). Similarly,
when the switch is in the PID position, the manual mode output tracks the control u(t) supplied by the PID unit because
of the presence of the lower tracking loop, also with tracking
gain kt.
Considerations of bumpless transfer also apply when parameters in the controller are changed abruptly. Such
changes are especially easy to make in digital implementations of PID control. Real-time changes in the controller parameters can be motivated by monitored changes in the dynamics of the system being controlled. When these changes
are done automatically, the resulting control system is classified as an adaptive control system. The particular PID parameter that is most readily conditioned for bumpless transfer is
(18)
r
451
PD & PD'
(19)
(Closed-loop
feedback)
e=rc
ki
+
+
1
s
+
+
uPID
kt
(Manual)
em
km
+
+
1
s
+
um
kt
Figure 6. A general analog PID controller featuring anti-resetwindup and bumpless transfer.
452
PID CONTROL
Digital PID control often is implemented by having the system signals of interest [r(t), c(t), and e(t)] sampled periodically
with sample period T; the sampled input signals are supplied
to a microprocessor through an analog-to-digital converter
(ADC), and the microprocessor is programmed to supply the
control signal u(t) to the system using a digital-to-analog converter (DAC). Usually u(t) appears to the input of the process
being controlled as a sampled-and-held signal with sample
period T. Within the microprocessor, the linear aspects of PID
control are based on the following calculations (using a parallel form):
u(kT ) = Kp e(kT ) + Ki
k
m=0
(20a)
u(k) = Kp e(k) + Ki
k
(20b)
m=0
z
z1
+ Kd
U (z)/E(z) = Kp + Ki
z1
z
b0 z2 + b1 z + b2
z(z 1)
(24)
This falls naturally into the incremental PID form. If the actuator of the control loop contains the integrator, then during
each sample period, the digital microprocessor simply has to
perform the calculations given on the right-hand side of Eq.
(24), normalize the result by dividing by T, and send the resulting approximate derivative value through the DAC to the
integrating actuator. Of course a slight modification of the
above is required if two-degree-of-freedom control is employed; namely, the value V,
V = (Kp Kp )r(k) + (Kd Kd )[r(k) r(k 1)]
(25)
(21)
(23)
(22)
where b0, b1, and b2 are related to Kp, Ki, and Kd in a straightforward way (to be described). Essentially all digital PID
z-transform transfer functions can be based on Eq. (22), which
has two poles, one at z 0 and one at z 1. The two zeros
of Eq. (22) can be either two real zeros or a pair of complex
conjugate zeros located to achieve performance objectives.
Analogous Implementation of Digital PID
The general continuous PID implementation of Fig. 6 can be
converted to digital form by using the following approximations. Assume that the lower portion of the figure is left unchanged, but that uPID is supplied from the output of a DAC
and hold circuit. The derivative action is approximated by ap-
(26a)
(26b)
PID CONTROL
1.
2.
3.
4.
5.
6.
7.
8.
9.
Initialize e1 e2 u1 0.
Compute temp u1 b1e1 b2e2.
Wait for the sample period.
At the sample period, obtain e r c using appropriate
A/D converters.
Compute u temp b0e.
If u umax, assign u umax, or if u umin, assign u
umin.
Output u to the DAC.
Assign in proper order e2 e1, e1 e, and u1 u.
Return to step (2).
Note that the above steps are arranged to minimize the number of calculations required between the instant that e is obtained from the ADC and the instant that u is sent to the
DAC. This process minimizes the computational time delay
that is introduced into the control loop; computational time
delay can have a detrimental effect on system performance.
Also, for two-degree-of-freedom implementations, the terms
from Eq. (25) would need to be included appropriately in the
code.
Signal Property Considerations
Both control signals and measured process signals are constrained in a variety of ways. Common examples of constrained control signals are (1) signals that are continuously
adjustable between saturation levels; (2) control signals that
can be assigned only a few valuesas in the case of a relay
with dead zone, with possible output values umax, 0, or umax;
and (3) control signals that are pulse-width modulated. Even
in case (1) above, when the control is supplied by the output
of a DAC, the finite word length of the DAC may have a noticeable quantization effect for small steady-state error signals. In case (2) above, when the control is supplied by an
electromechanical relay (once a common practice, but now
largely displaced by power electronics), a correct balance of P
and D terms is needed to avoid excessive switching of the
relay; in all-electronic implementations of such systems, rapid
switching actually may be included intentionally to achieve a
sliding mode of control, and nonlinear control strategies involving nonlinear functions of e and e are appropriate. Case
(3) above is especially prevalent in electronic drives for dc motors. This is because a power amplifier dissipates the least
energy if it is either full on or full off. To reduce heat buildup
in the amplifier, the control is implemented as follows: if the
desired control action from t kT to t (k 1)T is uo where
umax uo umax, the effectively equivalent pulse-width control supplied is
umaxsign(uo ), kT t < kT + T
(27)
u(t) =
0, kT + T t < (k + 1)T
453
(29a)
or
Ghp (z) = (1 z1 )Z
g( ) d
(29b)
1, uo > 0
sign(uo ) = 0, uo = 0
1, u < 0
(28)
(30)
with closed-loop poles of the system being roots of the characteristic equation. The PID gains can be adjusted to obtain desirable pole locations inside the unit circle of the z plane, and
a variety of other digital controller design methods can be employed.
454
PID CONTROL
kd s
ki
+
s
1 + s
(31)
(33)
1.6z 0.4
z + 0.2
(34)
of the process being controlled. There are many representative sources that can be consulted for details on a wide variety
of alternative tuning rules (2,47). Common tuning rules are
described in this section. Most of these approaches give excellent results if the process being controlled is well damped and
is dominantly second order. Systems that are essentially first
order can often be tuned well with PI control only. Systems
with large time delay can be controlled with PID control in
combination with a Smith predictor (8). From a frequency response viewpoint, the PID integral term supplies 90 of phase
lag at low frequencies, and the practical derivative term supplies somewhat less than 90 of phase lead at some intermediate to high frequencies. Thus, if the system requires more
than 90 of phase-lead compensation over some frequency
range to achieve good performance, PID control alone will not
be adequate.
For lightly damped oscillatory open-loop systems, one
method that is tempting (but should be approached with caution) is to place the zeros of the PID controller at the pole
locations of the process, to cancel the oscillatory poles. This
approach should be avoided if the PID controller employed
has one degree of freedom only: although the response mode
of the canceled poles will not be excited by reference inputs,
disturbance inputs will excite the oscillatory mode in the controlled output, and the PID controller with its zero gain at
the oscillatory frequency will not supply damping. For twodegree-of-freedom systems, the PID gains within the loop can
be assigned to add damping to the oscillatory mode, whereas
the independent PD factors associated with the reference input can be adjusted to provide blocking zeros. Sensitivity of
oscillatory pole-zero cancellations with respect to parameter
changes also is of concern (9).
ZieglerNichols Tuning
A widely applied tuning rule, the ultimate gain rule, was developed in 1942 by Ziegler and Nichols (10). This rule is based
on experimentation with the process to be controlled. First,
with ki and kd set to zero in Eq. (3), the proportional gain kp
is increased until the system starts to oscillate; the value of
kp that starts the system oscillating is denoted as ku and is
called the ultimate gain. The period of the oscillation, Tu, also
is recorded. As an alternative procedure, rather than determining ku and Tu by forcing the actual system into instability, if the frequency response G( j) is available, it can be used
to obtain both ku and Tu analytically, as follows. First, the
frequency u at which the angle of G( ju) 180 is determined, and the corresponding value of G( ju) Au is obtained. The value of ku satisfies the characteristic equation
1 + ku G( ju ) = 0
(36)
(37a)
ki = 1.2ku /Tu
(37b)
kd = 3ku Tu /40
(37c)
and
PID CONTROL
S( j)
Many variations of the ultimate gain rule have been developed. One popular one credited to Harriott in (4) is as follows.
First, with ki and kd set to zero in Eq. (3), kp is adjusted until
the step-response of the closed loop exhibits a decay ratio of
0.25 (the percent overshoot of the second peak in the step
response is one-fourth of the percent overshoot of the first
peak in the step response). Let kc denote the critical value of
kp corresponding to this 0.25 decay ratio. Also, let Tc denote
the difference between the time of occurrence of the second
peak and that of the first peak. Next, assign ki and kd on the
basis that
ki = 1.5kc/Tc
(38a)
kd = kc Tc /6
(38b)
455
1
, 1 + G ( j)G(
j)
(43)
(44)
and
And finally, while conducting additional closed-loop step-response tests, adjust all three gains by the same percentage
until desirable overshoot conditions are achieved.
More Advanced Tuning Methods
When the transfer function G(s) of Fig. 2 is known and is a
reasonably good model of the process being controlled, an
array of tuning methods are available, depending on the form
of G(s). For example, variations of the ZieglerNichols methods are based on cases where G(s) assumes forms such as
G(s) =
K0 esT0
0 s + 1
(39)
G(s) =
K0 esT0
s
(40)
or
b1 s + b2
s2 + a 1 s + a 2
(41)
(42)
CONCLUSION
In his classical paper on governors, Maxwell (11) in 1868
clearly expressed the difference in effects produced by proportional and integral control. However, there is no one person
that can be credited with having invented PID control. Many
of the early developments, those involving anti-reset-windup
for example, were company proprietary and therefore were
not available in the open literature. With the advent of computer control in the 1960s, many of the traditional PID techniques were reassessed and translated into the digital control algorithms.
In this article, the basics of PID control have been described. Operational features of both linear PID terms and
nonlinear characteristics have been examined. Digital PID algorithms have been described and have a special significance
in the modern era of digital control.
Today, PID controllers can be purchased from many manu strom and Hagglund (2) for a recent listing of
facturers [see A
companies and features available]. In addition to being used
for SISO control, PID controllers are used in a variety of other
applications. In cascade control, one control variable is generated on the basis of several measured variables. In selector
control one actuator can be driven from a selected PID unit,
and then automatically switched to a different PID unit based
on a max/min selector to control some other process signal if
it starts to deviate from a desired range. In ratio control two
outputs are controlled with one of the outputs required to be
a fixed percentage of the other output. In split-range control
and multiactuator control there may be fewer measured signals than actuators, and the amount of control from each actuator has to be automatically balanced to achieve the desired
overall process performance goals. Feedforward control of disturbances, when they can be measured, often leads to vastly
improved disturbance rejection. Automatic gain scheduling of
PID parameters can be based on measured process conditions.
And interactions of coupled control loops can be reduced by
properly designed coupled PID controllers. In many cases, as-
456
PIECEWISE-LINEAR TECHNIQUES
DONALD A. PIERRE
Montana State University
PROCESS CONTROL
Process control was initially developed from the practices of process industries such as paper, steel, and chemical
manufacturing. Beginning in the early part of the century, it became more widely practiced starting in the
1940s. Applications now include processes in microelectronics manufacturing, such as bulk crystal growth,
chemical vapor deposition, and etching. Process control is important because it provides a means for improving
process yield and reducing variation, and also enables new processing capabilities to manufacture unique
engineered materials and structures.
Process control provides an approach to developing an automatic control system for complex, nonlinear,
distributed parameter systems involving mechanical, transport, electrical, material, and chemical attributes.
The term process control is used broadly to cover a variety of functions including supervisory control, failure
detection, and the lower level control that determines the required inputs to achieve a desired objective. It
is applicable to large plants incorporating many subprocesses and for single processes. This article primarily
focuses on the issues associated with developing the lower level of control for a single process.
A major goal of process control is to automatically determine the process input settings to achieve the
desired output condition (referred to as the command signal) while minimizing the variation of the output
from the desired level (e.g., error). Variations are caused by external inputs called disturbances and variation
of the plant, such as aging. Performance is evaluated in terms of dynamic aspects (how fast the system needs
to respond) and magnitude of the error. A powerful conceptual approach used in control system analysis is to
evaluate the output error as a function of command and disturbance signals, which provides a useful basis for
designing compensators to meet performance specifications.
Taguchis concept of the process variability loss function provides an important basis for evaluating the
value of process control (1). Given target quality objectives, deviation from that target should be characterized
by some loss function. Thus, for a quadratic loss function, the narrower the distribution, the greater value the
process provides. One way manufacturers deal with process variations is to divide (or grade) production output
into various quality classes, a practice common throughout semiconductor manufacturing. This approach,
however, actually adds cost because sorting provides no additional value. In addition, there is the loss associated
with the value that could have been attained if the production had met a narrower distribution. In this context,
process control is valuable if used to reduce variation, but the cost of the solution must be balanced against the
return.
There are two major control options: feedforward and feedback. Feedback utilizes an error signal formed
by comparing the actual output to the desired value, which is then used to determine a corrective adjustment
to the plants inputs. Feedback corrects for unknown variations, but it is reactive, acting only after an error has
occurred. Feedforward control utilizes process knowledge to determine the input values required to achieve the
desired output without process measurements. Such process knowledge can be informal, such as that of the
equipment operator, or expressed formally in a mathematical model. Feedforward is anticipatory for known
conditions, but does not compensate for unknown factors, such as disturbances or plant variations. Feedforward
control is typically used in cases where a cost-effective sensor is not available and a robust model is available.
1
PROCESS CONTROL
Such an approach has been used to control wafer temperature distribution in tube furnaces where it is not
practical to measure the wafers temperature distribution.
Feedback systems are implemented either utilizing measurements in real-time or after a process is
completed. Real time implies that the output (or related process states) are measured as the process is running
and the inputs are adjusted on the fly to compensate for errors. If the process variations or commands are slow
relative to processing time (i.e., the batch time), then real-time control is not necessarily needed. In this case,
run-to-run control is used where the results are measured after the process is completed. If the disturbances
have long-term behavior relative to process batch time, run-to-run control works well assuming that there is
only one independent disturbance. In cases where variations are fast relative to process timescales, closed-loop
control provides significant performance advantages. For example, in spot welding, the thermal dynamics of
the process vary significantly throughout the welding cycle, so that development of a dynamic measurement
controller provides significant performance improvements (2). Similarly, real-time control also works well to
eliminate time-varying disturbances. It also is advantageous if one wants to build in process flexibility (i.e.,
the ability to change process output levels), even for slower processes, when process modeling is not accurate
enough to meet the performance objectives.
It is not always possible to directly measure the variable related to the process objective. An alternative
strategy is to identify and control an intermediate or secondary process variable related to the primary variable.
In crystal growth, for example, dislocation density is an important material objective for optoelectronic devices
which is not feasible to measure in realtime. Even if it were measurable, it might not be desirable because,
once a dislocation has been introduced in the crystal matrix, it cannot be removed. However, by controlling the
crystals thermal gradients, dislocations are prevented from forming in the first place. A related strategy is to
use cascade control structures wherein disturbance paths are identified and local loops eliminate their effect
before they propagate to the primary control objectives.
Other approaches to process control include minimizing input variations [the objective of statistical
process control (SPC)], thereby eliminating external disturbances, or selecting the operating regime so that the
output is insensitive to the disturbance (the later approach is Taguchis design of experiments technique). The
disadvantage of SPC is that it requires continued effort to track down all disturbances, whereas feedback control
operates automatically. The disadvantage of the design of experiments is that it requires an extensive set of
experiments on the production equipment, which could significantly affect production. Taguchis technique,
however, utilizes existing equipment without additional capital or development costs. In some cases, it is
beneficial to combine these techniques. For example, design of experiments is used to determine the desired
operating point when number of input parameters are to be set, and local feedback is used to eliminate
disturbances and plant variation. Feedback control, however, can be implemented only if there is a manipulable
input that compensates for the specific disturbance.
Developing a process control system requires addressing a number of issues beyond the design of the
control algorithm (i.e., the equations that map measurement information to the input levels). Important
issues include specification of the relevant process objectives (both technical and economic), identifying the
constraints posed by the equipment and process physics, determining which control structure best achieves
the desired performance objectives, and considering opportunities for improving the process capabilities by
changing the system design. Answering these questions typically requires a good understanding of the relevant
process physics in addition to the relevant control principles. Modeling is useful in answering these questions,
but should be performed from a control perspective. Such a perspective seeks to understand the system by
identifying its fundamental dynamics and the limitations they pose, its disturbances, objectives, and available
manipulable inputs.
One major control structure issue stems from the fact that more than one input typically affects the
desired output(s). Thus, it is important to understand which inputs can achieve the best performance. One
way to evaluate the performance of alternatives is utilizing frequency-domain techniques to quantify gain and
bandwidth for alternative inputs. Additionally, one should evaluate if some outputs are more difficult to control
PROCESS CONTROL
independently. For example, in a welding process, it is difficult to independently control the width of the heat
affected zone and the size of the weld nugget. These problems are related to the inherent coupling of a process,
and it may not be physically possible to achieve all the objectives. Thus, an important objective of the analysis
is to determine such inherent conflicts.
A control perspective is also useful in selection of the operating regime, i.e., the specification of the nominal
operating conditions for all inputs and process settings or trajectories. Important factors include how fast the
process proceeds, the significance of different disturbances, and which aspects of the process physics dominate.
The operating regime is selected from a basic understanding of the dominant process physics and/or from
formal optimization techniques, such as design of experiments.
Consideration should also be given to redesign of the process and/or alternative control structures (i.e.,
combinations of feedback, feedforward, or cascade) to improve performance. Important examples of the benefit
of integrating system and control design include eliminating known disturbances, adding new actuators to
achieve independent control of an additional output, and sizing actuators to meet the performance specification.
Development of the control system requires identification of the fundamental dynamics of the process.
Important issues to address are factors that fundamentally limit control performance or that need to be
compensated for through system or control design. Some of these features might suggest changes in the
operating regime selected to avoid the problem, such as if there are poorly damped modes. Other features, such
as nonlinearities, need to be identified and considered in designing the control algorithm.
where do is an unknown disturbance that acts on the plant and is reflected at the output, di is a disturbance
that affects a process input, and n is a noise signal that corrupts the measurement of the output. (These
PROCESS CONTROL
relationships are generalized for MIMO systems by matrix algebra.) The output disturbance do indicates that
there is an effect on the output but the functional form is not known, whereas an input disturbance di represents
variations of the manipulated input or additional external inputs whose functional input/output mapping is
known.
The achievable control performance is analyzed in terms of the error where e = r y. The performance of
the two structures is expressed as
Equation (4) reveals that feedback control reduces process variation, that is |e| can be made small. In contrast,
process variation for feedforward control is reduced only by eliminating disturbances and plant variation (see
Eqs. (3) and (5)).
Analysis of the steady-state performance to step inputs provides insight into feedback control performance.
This can be done by utilizing the final value theorem, which for step inputs corresponds to evaluating the
magnitude of the corresponding closed-loop transfer functions [T = gk/(1 + gk), S = 1/(1 + gk)] at steady state
(i.e., s = 0.). In general, making |gk| 1 achieves good performance for following commands and rejecting
disturbances because this makes |S| 1. Performance with respect to insensitivity to noise, however, is poor
because |T| 1. Note that several dynamic factors limit the maximum gain that can be used, thereby limiting
achievable performance (see the following section).
One can consider the feedback controller k as an amplifier because u = kem (where em is the measured
error) and u is a power-level signal. Viewing k as an amplifier is consistent with requiring |gk| 1 to achieve
good performance. A simple version of such a controller is a proportional gain (i.e., P-type controller). More
complex forms with varying gains for different frequency ranges can achieve better performance. In cases where
PROCESS CONTROL
u is not an electrical signal, the controller can consist of two parts: the control algorithm and the actuator that
manipulates the physical input u.
The achievable performance relative to variation of the process and/or model error, g, is given by
where g = m + g and m is the nominal process or model. Thus, feedforward is sensitive to the relative model
error, but feedback control compensates for such error or process variation, provided S is made small.
The different approaches to reducing process variation are analyzed by linearizing the output about the
nominal operating point (W o , yo ) and expressing the output variation as
where the nonlinear mapping of the material inputs w and manipulable inputs u is given by y = f (w, u). Classical
SPC and Taguchi are considered feedforward approaches wherein SPC seeks to identify the assignable causes
of the variation and to eliminate them, thus making |w| small. Taguchi seeks to identify the operating regime
(i.e., values of wo , uo ) whereby the output is insensitive to the variations, that is |f /w| 1. The merit of
this method is that it requires no additional capital expense. However, because it is based on experimentally
mapping the input/output space, it might be quite expensive to run the experiments on the production line.
The required condition for implementing a feedback solution is that the plant gain g = f /u = 0. Thus, in some
cases there are disturbances or operating conditions that cannot be corrected for by feedback.
Closed-Loop Frequency and Dynamic Analysis. Frequency analysis yields important insight into
designing an appropriate process control system and into factors that pose important performance limitations.
It is also used to analyze the performance of more complex inputs and nonlinear systems linearized about a
nominal operating point. The output response of a linear system g(s) to a sinusoidal input u(t) = sin t is given
by
where |g(j)| is the magnitude of the complex transfer function evaluated at frequency and is a phase shift.
The frequency characteristics of a system can be visualized by plotting the magnitude of g(s) as a function of
frequency (s = j) on a log-log plot, known as a Bode plot (Fig. 2). An important characteristic is the range of
frequencies with uniform amplification called the bandwidth BW . The Laplace domain provides a basis for
relating the characteristics of the Bode plot to the time-domain characteristics of the process. For example,
the open-loop bandwidth approximates the open-loop dynamic response as BW dominant 1 dominant where
dominant is the dominant pole of the system, and dominant is the dominant time constant of the system (3).
Frequency analysis is useful for analyzing the error magnitude and dynamics for closed-loop systems
and for compensator design. Important design insight is obtained by comparing the magnitude plots of |gk|,
sensitivity S(s), and closed-loop T(s) transfer functions (Fig. 2). The closed-loop bandwidth of T (CLBW ) is
bounded by the frequency that |gk| 1. Since the closed-loop bandwidth characterizes the range of frequencies
PROCESS CONTROL
for good command following, disturbance rejection, insensitivity to plant variations, and speed of response, it
is used as a primary design variable. Thus K(s) is selected to achieve the desired crossover frequency for gk.
The desired performance of the system is expressed in terms of the range of frequencies where r, d, and n have
significant energy (i.e., large magnitude).
Noise and stability pose important performance limitations because S and T cannot be made small in
the same frequency range. Thus, good performance cannot be achieved if the power spectrums for desired
command following and disturbances overlap the noise spectrum. In addition, the Nyquist stability criterion
poses a conservative magnitude bound on |gk| relative to the uncertainty of the open-loop process dynamics as
|T| < |g/m|. Thus, CLBW is limited by model uncertainty, and an accurate process model is needed through the
desired closed-loop bandwidth. (Similar logic can also be used for determining the model accuracy required to
achieve performance objectives for a model-based feedforward control approach.)
The frequency-domain analysis also provides a context for selecting appropriate actuators. Each process
typically has a variety of inputs that might be used as the manipulable variable for real-time control. From a
general nonlinear representation of the process, y = f (u1 , u2 , . . .) where ui is the ith input of l different inputs,
one can linearize the system about the nominal operating point yo , uo i . Because at least p inputs are needed
to independently control p outputs, the first design decision is which p inputs achieve the best performance.
For single-input, single-output (SISO) systems, the question reduces to Which of the actuators achieve the
best performance? The comparison is made in terms of the open-loop bandwidth, reflecting how fast that input
affects output, and the open-loop gain, which indicates how large an impact input has on output.
Inputs with the greatest bandwidth and gain typically result in the best performance. However it is not
necessarily true that all high-bandwidth systems have large gain, thus possibly requiring a tradeoff analysis.
High-bandwidth systems require less gain to meet the desired closed-loop bandwidth than slower systems,
thus typically they have a larger robustness margin. In addition, they may have less model error over the
same frequency range because the dominant pole is at a higher frequency. Similarly, high-gain systems require
less controller gain to achieve a desired error performance level, thereby improving the robustness bounds.
Alternative actuators are compared by plotting the Bode plots for each input. Because the inputs are of
different units, the transfer functions should be scaled so they can be compared on a consistent basis. For
example, normalizing by the nominal value of each input expresses each input as a fractional variation of the
nominal input.
In selecting actuators, one should also evaluate limitations posed by dynamic issues such as nonminimum
phase characteristics including time delays, open-loop instabilities, and right-half-plane (RHP) zeros (4). This
factors fundamentally limit achievable closed-loop performance. For example, a RHP zero causes defective
transient responses in that the initial response is in the direction opposite to the final steady-state. In practice,
the closed-loop bandwidth is limited to the bandwidth of the RHP zero.
PROCESS CONTROL
MIMO Systems. Additional issues arise for MIMO systems because of cross-coupling between inputs
and outputs. This coupling causes more than one output when a single input is varied (Fig. 3). MIMO systems
are represented by y = G u, where G is a p l matrix mapping l inputs to p outputs. The frequency analysis
is extended to MIMO systems by utilizing the spectral norms and principal gains of the transfer matrix which
are calculated by a singular value decomposition. For a p p matrix, there are p singular values, i , which are
positive scalars (3,4). Thus, for |u| = 1, (G)|y| (G) where , are the maximum and minimum singular
values of G(s). (G)|y| (G)(G)|y| (G)
There are a variety of control design algorithms for MIMO systems that compensate for cross-coupling.
However, there are some systems, where the coupling makes it difficult to independently control all outputs.
Mathematically, a large
This problem is characterized a large condition numbers for G, that is, 2 (G) = /1.
condition number indicates that the matrix is close to loosing rank, that there are output directions which entail
large control efforts. Such large inputs might saturate the actuators, which could result in loss of stability or
inability to achieve the desired output value. On a practical basis, trying to design a system to operate under
these conditions requires actuators which, for the most part are not used significantly, except to reach sensitive
directions. Another analysis method that provide similar insight is the relative gain array (RGA) (3).
An important solution for poorly conditioned systems is to seek alternative system designs that improve
the ability to control the process. Determining the reason for the poor conditioning is helpful and is obtained
from the singular vectors of the process or by decomposing the transfer matrix into column vectors (15). Poor
conditioning results from several actuators with similar effects on the outputs or an input that does not have the
same magnitude of impact as the others. Alternative designs can be proposed once the nature of the limitation
is determined.
Because both i and 2 depend on the input/output scaling, all variables should be scaled consistently.
Use of dimensionless variables compensate for the different units used for each input and output. Different
normalization methods are used such as scaling about the nominal operating point, or defining perturbations
that reflect relevant engineering scales, such as tolerances, saturation limits, or error limits. For example,
chemical engineers normalize outputs by transmitter spans and inputs by appropriate valve gains (3).
Other Control Structures. Two important variations of control structure use alternative measurements of the primary process output. In processes where it is not practical or possible to directly measure the
process output, a secondary variable related to the primary objective is used for feedback. For example, in
bulk crystal growth, dislocation density is important because it affects the electro-optical characteristics of the
material, but it cannot be measured in real time. Because dislocations are related to temperature gradients,
one can instead control the temperature gradients to prevent the dislocation from being introduced into the
PROCESS CONTROL
crystal matrix. In this example, there is an added benefit because a process state upstream of the output is
controlled by feedback. Because feedback control is reactive, it only takes control action after an error has
occurred. In the case of dislocations, however, once a dislocation has been introduced into the crystal matrix, it
cannot be eliminated. Thus by controlling the upstream variable (e.g., the temperature gradients), one prevents
the defect from occurring.
A related option is to utilize a cascade control structure (Fig. 4) (3). If a disturbance or a known process
variation can be identified and measured, it is possible to close the loop around this variable to ensure that the
variation does not propagate downstream to affect the primary objective. The benefit of this practice is that
it results in better overall performance for a level of desired robustness because it significantly reduces the
action that the primary control loop needs to perform.
Control Algorithms. There are a variety of methods for designing the control algorithm, and the reader
is referred to the other related articles for specific details. For linear systems, important design methods include
classical design methods such as PID for SISO systems, and more advanced techniques such as optimal control,
robust control, H infinity, and model predictive (3) which can be used for both SISO and MIMO systems. These
methods require a dynamic process model which can be either derived from first principles or experimentally
using system identification techniques. Alternatively, adaptive control techniques do not require a system
model.
For systems that vary in time or with a changing parameter, gain scheduling is used to interpolate between
control designs developed for different conditions. Where there are significant distributions of disturbances or
noise, stochastic control design can be used. Some systems have characteristics that require special approaches.
These include nonlinearities and distributed parameter systems (which entail spatial variations). Other control
approaches provide important advantages such as neural networks for nonlinear systems, fuzzy logic that
responds to different conditions, and non-real-time techniques based on statistics and designed experiments
(4,5).
PROCESS CONTROL
because of different loading (i.e., the number of wafers processed each in run), the spatial characteristics of
heat transfer modes (such as natural convection effects that differ significantly for vertical and horizontal
furnaces), and end effects. Important process objectives include the rate of heat-up/cool down (which affect
production rate without providing value), maintaining thermal uniformity, and achieving the process setpoint temperature. Because of sensor limitations, it is difficult to directly measure the wafer temperature.
Thermocouples used in the furnace include spike thermocouples, located near the heater elements, and profile
thermocouples, located near the wafer edges. Thus, tube-furnaces are controlled by secondary measurements.
The process is a coupled MIMO problem because each furnace segment affects its neighbors. Traditionally,
decoupled proportional-integral-differential (PID) loops control each zone, which does not take into account
the cross talk of neighboring elements. This coupling limits how tightly each loop is tuned, because unmodeled
dynamics cause it to become unstable or introduce a disturbance into the neighboring region.
Several model-based control schemes have been developed to overcome the measurement limitations.
These schemes are essentially a hybrid of feedback and feedforward, where the model is infers the wafer
thermal distribution based on the process model and measurement data and designs a MIMO controller.
Because many of the primary heat transfer coefficients dependent on operating temperature, wafer loading,
geometry, and material properties that are difficult to measure directly and/or highly variable, the model has
a number of coefficients that must be empirically determined with instrumented dummy wafers.
A MIMO control approach enables more aggressive control action because the otherwise ignored interactions between zones is now taken into accout. Development and application of such a process has been
undertaken by a team combining the manufacturer (Motorola), the furnace vendor (Silicon Valley Group), and
the control company (Voyan) (5). Implementation has reduced process cycle time by 18% and the fractional
variation in standard deviation (defined as the standard deviation normalized by the set-point temperature)
was reduced from 1.67 to 0.77% at high temperatures (950 C).
Rapid Thermal Processing. Rapid thermal processing (RTP) technology enables fast processing for a
variety of semiconductor manufacturing applications including diffusion, oxidation, chemical vapor deposition,
and nitridation. Tungsten-halogen lamps heat the entire wafer surface, thereby minimizing processing times
and achieving novel structures by avoiding the slow thermal relaxation that occurs with slow ramp rates.
RTPs commercial prospects are in low-volume production of application-specific integrated circuits (ASIC),
shortening the development time for new products, and possibly competing with large conventional fabricating
processes because of the reduced processing time. Typical operating characteristics include ramp rates on
the order of 50 C/s from ambient up to 1100 C, constant temperature for 15 minutes, followed by a rapid
cool down rate. Temperature uniformity across the wafer is a critical performance objective. For example, the
sensitivity of coating thickness variations to temperature variation for oxidation or CVD deposition of films
can be calculated from the deposition rate relation Rdep = k exp (E/kT) because these processes are typically
thermally activated. Normalized sensitivities for the growth of polysilicon layers are 2.5%/ C of variation.
Real-time control is critical in several aspects of RTP. Although the ramp rate is determined by the
lamp power flux, achieving uniform thermal distribution in both steady-state and during the ramps is difficult
because of heat transfer variations across the wafer. The different view factors, wafer fixturing, and the
variations in radiation and convective heat transfer over the range of temperatures and pressures prevent use
of a single open-loop thermal actuator design to achieve uniform temperature for all conditions. Even a single
closed-loop controller will not work. Process nonlinearity is characterized by gain values and time constants
that vary by an order of magnitude over the operating regime (7).
System design is important in achieving good control performance for RTP. Independently driven thermal
actuators are desirable to compensate for heat transfer variations across the wafer. In a joint TIStanford
design, three independently controllable concentric rings of actuators were used (6). The maximum input
power was 2 kW, 12 kW, and 24 kW (starting with the inner ring), consistent with the increased loss toward
the outer portion of the wafer. Even though reflectors were used, there is some actuator overlap because each
portion of the wafer sees more than one bank of lamps. Thus, the MIMO aspects of the problem must be
10
PROCESS CONTROL
Fig. 5. RTP comparison of open-loop (dashed line) and closed-loop (solid line) operation for two 24 wafer lots (7).
considered at both the system design and control design levels. At the system design level, it is appropriate to
evaluate the condition number of the system, that is, 2 for the DC transfer matrix. In the first design, 2 110,
indicating that there are some output directions that would result in large actions in one or more actuators
(1 = 12.2, 2 = 1.1, 3 = 0.11). This might not be a significant problem because the maximum singular value
corresponds to a condition of uniform temperature, which is the desired command signal. It does indicate that
the lamps could be wired in two independent banks. However, it is difficult to compensate for a disturbance
corresponding to the worst condition. To solve this problem, a baffle was designed to change the heat fluxes
of the actuators to the wafer, thereby reducing the condition number to 23. Note that a low condition number
does not necessarily mean that the actuators are decoupled.
Making accurate measurements is a significant challenge for RTP control. Thermocouples are not desirable because they require contact with the wafer and entail some amount of lag. Pyrometers are noncontact,
but are sensitive to variations in surface emissivity related to temperature and film characteristics. Acoustic
sensors, based on the temperature sensitivity of the wafers elasticity, are being developed that avoid these
problems.
One solution for optical pyrometer drift is to use run-to-run control to estimate the sensor error (7). Figure
5 compares open-loop RTP control with closed-loop. The open-loop varies widely, indicating that conventional
run-to-run control does not perform well because there are no long-term trends. The closed-loop control, in
contrast, has a ramp component, indicative of a systematic sensor error. The run-to-run measurement to detect
this drift would be film thickness, because it is strongly temperature-dependent.
A number of different design approaches have been used to develop the control algorithm (6). Utilization
of independent PID loops does not result in good performance because of the cross talk between inputs. The
large range of operating temperatures requires a design technique to handle the related nonlinearities. The
control design must also deal with the MIMO nature of the process and the significant modeling uncertainties.
One approach is to incorporate both a feedforward and feedback approach implemented with internal model
control (IMC). The process model is obtained from first principles or by system identification. To compensate
for nonlinear variations, gain scheduling is implemented by parameterizing the model coefficients as functions
PROCESS CONTROL
11
of temperature. A linear quadratic Gaussian (LQG) design was also developed on the basis of linear models
obtained from system identification. LQG easily accommodates the MIMO nature of the design problem and
can also be augmented to include integrated error states ensuring zero steady-state error. Results achieved are
control of ramp temperature distribution to 5 C and 0.5 C during hold.
Etching. Etching is a process that transfers patterns to a wafer by removing material not covered by a
mask. Researchers at the University of Michigan have applied real-time control to reactive ion etching (RIE),
which utilizes plasma ions to assist the etching process (8,9). A bias voltage accelerates the ions toward the
wafer surface, enhancing etching normal to the surface. The process chemistry is quite complex including both
gas phase and surface reactions. Etching process objectives include selectivity, the ability to etch only the
desired layer and not other material; uniformity of etch rate over the wafer; anisotropy, the ability to etch the
layer primarily in the vertical direction to prevent overhangs under the mask; and the etch depth.
The process can be divided into two coupled parts: plasma generation and etching. Inputs to the process
include power, gas flow rates, and pressure set point where the pressure is regulated by varying a throttle
valve upstream of a vacuum pump. Fluorine concentration [F] and bias voltage are outputs of the plasma
process and inputs to the etch process. The conventional control practice is to set the power input and flow
rate levels and use closed loop control to maintain chamber pressure. These variables are far away from
the physics that determine the etch characteristics, suggesting that regulating intermediate plasma variables
might yield better performance. A cascaded control structure is used with an inner loop around the plasma
generator and the outer loop using the etch rate to determine the set points of the inner loop. If no real-time etch
rate measurements are available, run-to-run control is used to adjust the set-point values of the intermediate
variables. The function of the closed-loop control of plasma variables is to reject plasma generation process
disturbances. Such disturbances include variations in the RF power and matching network, variations in
the performance of the mass flow controllers, aging of the chamber, due to polymers deposited on the inside
surfaces, and loading variations that occur when the amount of surface area varies (because of the number of
wafers or mask variation).
Experiments indicate that etching has important dynamic characteristics that could also benefit from
closed loop control. At the end of the etch, less material is left so that the etch rate increases. In some cases,
this results in etching the sidewalls, resulting in undercutting. Thus, whereas the conventional practice is to
utilize simple end-point detection by monitoring when new species are introduced into the gas phase, additional
benefits can be achieved by etch rate control. There can also be initial transients due to contaminant variations,
which change total etch time. Having dynamic control capability enables compensation for both of these factors.
Because the plasma generator is nonlinear, either local linear controllers can be developed to operate at
a set point or a strategy to compensate for the nonlinearities can be applied. Important design limitations are
imposed by both sensor and actuator considerations. Because of the poor signal-to-noise ratio of the fluorine
concentration sensor, the bandwidth is limited to less than 1 radian/second. Some of the actuators, such as the
throttle valve, are highly nonlinear, so that the regulation range is restricted for use in linear local controllers.
To minimize steady-state error, free integrator dynamics were added to the controller. Both PID and LQG/LTR
methods are used. The resulting performance of real-time control for maintaining plasma states is shown
in Fig. 6, which indicates superior rejection of load variability and a more uniform etch rate throughout the
process (8).
It is also desirable to control etch uniformity across the wafer surface, requiring a multivariable control
approach. If real-time measurement of etch rate and uniformity output is not practical, a postprocess measurement run-to-run controller is used, built from a response surface mapping of the local plasma states to
the etch outputs (9). Development of the feedback control algorithm requires a more general approach because
the MIMO controller entails significant change in the operating regime. One approach is to consider that the
process dynamics are linear and are coupled to a nonlinear static map of the inputs to output magnitudes
(e.g., a Hammerstein model 9a). One can design a linear dynamic controller that operates in conjunction with
a static controller to compensate for the nonlinearity.
12
PROCESS CONTROL
Fig. 6. RIE etch rate responses for different loading conditions: (a) open-loop, (b) closed-loop (8).
Experimental evaluation reveals that the pure run-to-run approach has results nearly comparable to
the cascade approach for uniformity. This suggests that there may be a lack of good control authority in the
selected operating regime and/or that the disturbances are more related to the etch process and, therefore, are
not compensated for by the plasma real-time controller.
Chemical Vapor Deposition. Chemical vapor deposition (CVD) enables the manufacture of coatings
with a variety of properties used in semiconductor production. Different coating structures and compositions
PROCESS CONTROL
13
are achieved by proper choice of precursors and deposition conditions which to date are primarily determined
empirically. There are numerous CVD variations in terms of different precursors, energy sources, and reactor design. Important processing objectives include composition, structure, thickness, and uniformity of the
deposited coating. Process examples used to provide insight into the application of control include control of
metallorganic CVD (MOCVD) and deposition of TiN.
MOCVD uses liquid precursors in a bubbler to bring a gas phase of the metallorganic molecules into the
reactor. Both III-V and II-VI compounds have been grown for applications, such as light emitting diodes, laser
structures, and photodiodes. Such structures need precise tailoring of the spatial composition of the coating,
which must be repeatable run-to-run. Two approaches employed to control coating composition include (1)
control of the actuators and (2) feedback control from a measurement of composition. The first approach will
work if the system is operating in a regime where mass transport limits growth but fails if chemical reactions
interfere.
Gaffney et al. (10) focus on improving the performance of the input actuators, relying on a feedforward
control approach to meet the composition objectives. This approach works if there are no process disturbances
(other than inlet material variations) and there are no significant dynamics to the process relative to the desired
deposition structures. Utilization of measurements related to the state of the bubbler enable direct feedback
control of this actuator with a simple control law. Evaluation of the performance is established by measuring
the optoelectronic characteristics of superlattices grown with five alternating layers of 3000 GaInAs with
200 layers of InP.
Warnicks et al. (11) approach is to measure the composition and thickness of the coating in real time with
an ellipsometer. Because of the range of compositions desired to grow Alx Ga1 x , the input-output mapping is
nonlinear. The dopant rate dynamics are approximated as a simple integration of feed material inputs with a
time delay to represent the transport delay, without detailed representation of process states and/or coupling. A
Hammerstein model is used because the linear dynamics are assumed to be invariant to the nonlinearities. This
could be valid for a specific operating point, for example, if the bulk pressure and temperature of the process
are held constant but there are variations in composition. A nonlinear mapping is proposed that decouples
the two outputs (composition and thickness) and linearizes the input-output mapping. Thus, linear control
techniques are used for the two resulting decoupled linear systems. Free integrators are incorporated to insure
asymptotic performance of the closed-loop system, in spite of the significant parametric modeling error that
might exist. The bound on transient errors is established by the achievable closed-loop bandwidth. The most
significant limitation is that the accuracy of the measurements is not clearly established, thereby introducing a
possible DC error in all control efforts. The transient performance characteristics of the closed-loop system are
also limited by the dead time. Because the processes are operated in the mass transport regime, the time delay
also corresponds to the rate of growth, and limits how sharp a transition is made. In contrast, the feedforward
approach of actuator control does not have such a limitation.
Major problems with each of these approaches suggest that a combination of methods could achieve better
performance. The feedforward aspect would overcome the limitations posed by the time delay, whereas feedback
would compensate for parameter uncertainties and unmodeled process physics. The combined control system
could result in tighter control of spatial composition, thereby enabling faster growth rates with acceptable
error.
To generalize control of CVD, one should consider the control structure design question. Gevelber et al.
(12,13) evaluated alternative actuator performance capabilities for CVD, where, for many applications, control
of grain size in addition to uniformity is critical. A mechanistic model of nucleation and growth suggests that
independent control of temperature and deposition rate are required to control grain size. To determine which
of the available actuators should be utilized, both the steady-state gain (Fig. 7) and dynamic characteristics
(Fig. 8) of each actuator are evaluated in terms of controlling deposition rate (12,13). This analysis is conducted
as a function of temperature because it strongly parameterizes changes in the dominant physics. Several
input variables are ruled out because they have poor features like sudden reduction of actuator effectiveness
14
PROCESS CONTROL
Fig. 7. Steady-state normalized gains of Rdep for various inputs for TiN CVD (13).
(partial pressure of TiCl4 ). Analysis of the dynamic features are obtained from an analysis of the linearized
system. One of the actuators has right-half-plane zero dynamics (pressure) that would limit the achievable
control performance. In the case presented, a clear choice is given solely in terms of gain because the open-loop
dynamics are similar. Note, however, that the dynamics vary significantly with temperature.
Selection of the appropriate operating regime must account for the variety of conflicting performance
objectives, such as maximizing growth rate while minimizing nonuniformities, while considering the variations
in the dominant process physics. Modeling analysis provides useful insight into answering these questions and
should be directed to obtaining information about the dominant phenomena scale. Thus, for this example,
because the dominant time delay scales with the reactor volume and flow rate Q, as T D /Q, one can adjust
the operating point (here the total flow rate Q), to minimize the fundamental limitation posed by the time delay
(12). In hot-wall batch reactors, the reactant concentration varies throughout the reactor, resulting in varying
growth rate from point to point. A ratio indicating the sensitivity is developed to determine which operating
regime enables one to meet uniformity requirements while maximizing growth rates.
Crystal Growth. Growth of bulk crystals by the Czochralski process is one of the first steps in wafer fabrication. The process begins by bringing a seed crystal to thermal equilibrium with the molten semiconductor,
and pulling the crystal from the melt. Heat transfer is a major phenomenon that determines growth characteristics. Important material properties that determine the crystal quality for different applications include
dislocation density and dopant distribution. Si crystals are currently grown dislocation-free for commercial use
with diameters of 200 mm. Some crystals, however, are more difficult to grow. For example, GaAs has a lower
thermal conductivity than Si and a lower critically resolved shear stress (the stress at which a dislocation is
PROCESS CONTROL
15
Fig. 8. (a) Bode and (b) 2% step response of Rdep for various inputs for TiN CVD, T = 1300K (12).
formed). Thus temperature gradients are higher, resulting in greater thermal stress and higher dislocation
densities.
Dislocations are formed at the melt-crystal interface and also formed and multiplied in the crystal. The
diameter is determined at the interface between the crystal and melt, whereas the dopant distribution is
determined by the melt flow and the interface. Because the nature of this coupling, it cannot be guaranteed
that any arbitrary set of material objectives can be achieved. For example, the interfacial shape that minimizes
dislocation formation might not be optimal for other defects and dopant distribution, suggesting that a tradeoff
16
PROCESS CONTROL
analysis is required. To the extent that the detailed physics can be modeled, the analysis can be formalized.
However, because much of the important physics is not analytically tractable, important insight is obtained
from understanding the physics and experimental investigations.
Feedback systems were developed in the 1960s for automatic diameter control. Several different diameter
measuring schemes are used including optical and weight measurement. Although there are a number of
actuators for the process including heater power, pulling rates, rotation rates of the crystal and crucible
position, and possibly an external magnetic field applied to minimize melt turbulence, typically only heater
power and pulling rate are used for diameter control. However, because the other material objectives are
coupled to the interface, they could adversely be affected by the action of the diameter servo loop.
Examination of the process indicates that several disturbances act on the growth interface (14). Because
most pullers are run with a fixed charge, the reduction in melt level changes the heat flux and, therefore, the
growth dynamics. In addition, as the crystal grows longer, it experiences a changing thermal environment.
Thus, a cascaded control structure with an inner melt temperature loop can eliminate the melt disturbance.
Direct compensation for the changing thermal environment is more difficult unless one considers designing an
independent thermal actuator. The benefit of such an actuator is even more significant if it is desired to control
the diameter and also the interface shape. Gevelbers analysis of the MIMO nature of this problem reveals that
the system is poorly conditioned without such an actuator (15).
Modeling the dominant process physics reveals a right-half-plane zero associated with the weight measurement (14). The modeling provides a parametric expression of the zeros location relative to the process poles
which is used to determine when there are significant performance limitations. The parametric description of
the pole location is also used to understand when the system is underdamped, resulting in growth variations.
This explains, for example, why it is more difficult to grow GaAs crystals and is used to help select the operating
regime. Modeling also suggests important design consequences. For example, although a bottom heater does
not achieve better control authority over the interface shape, it does result in a significantly faster system that
may yield performance advantages.
Conventional control loops on pullers have been implemented with PID logic. Rivera and Seider have
proposed a model predictive controller (MPC) and control structure to help integrate the different process
objectives (16). A cascaded control structure accounts for the different timescales. One loop is the bulk controller
that manipulates the heater power inputs. The objectives of this loop include pulling velocity, thermal stresses,
and the melt dominant distribution. The other loop controls the radius by manipulating the pulling velocity.
MPC coupled with models of the process achieves control solutions which directly take into account constraints
to meet the other objectives.
BIBLIOGRAPHY
1. M. S. Phadke, Quality Engineering Using Robust Design, Englewood Cliffs, NJ: Prentice-Hall, 1989.
2. K. Haefner et al., Real time adaptive spot welding control. In D. E. Hardt (ed.), Control Methods for Manufacturing
Processes. New York: The Dynamic Systems and Controls Division, ASME, November 27December 2, 1988, pp. 5162.
3. W. L. Luyben, Process Modeling, Simulation and Control for Chemical Engineers, 2nd ed., New York: McGraw-Hill,
1990.
4. J. M. Maciejowski, Multivariable Feedback Design, Reading, MA: Addison-Wesley, 1989.
5. C. Porter et al., Improving furnaces with model-based temperature control. Solid State Technology, 39 (11): 119131,
1996.
6. T. Kailath et al., Control for advanced semiconductor device manufacturing: A case history. In W. S. Levine (ed.), The
Control Handbook, Boca Raton, FL: CRC Press, 1996.
7. C. Schaper, T. Kailath, Thermal model validation for rapid thermal chemical vapor deposition of polysilicon. J. Electrochem. Soc., 143 (1): 374381, 1996.
PROCESS CONTROL
17
8. O. D. Patterson, P. P. Khargonekar, Reduction of loading effect in reactive ion etching using real time closed loop control,
J. Electrochem. Soc., 144 (8): 28662871, 1997.
9. M. Hankinson et al., Combined Real-Time and Run-to-Run Control of Etch Depth and Spatial Uniformity in Plasma
Etching, J. Electrochem. Soc., 144 (7): 24732479, 1997. (a) L. Ljung,, System Identification, Englewood Cliffs, NJ:
Prentice-Hall, 1987.
10. M. S. Gaffney et al., Control of III-V epitaxy in a metalorganic chemical vapor deposition process: impact of source flow
control on composition and thickness, J. Crystal Growth, 167 (12): 816, 1996.
11. S. C. Warnick, M. A. Dahleh, Feedback control of MOCVD growth of submicron compound semiconductor films. IEEE
Trans. Control Syst. Technol., 6: 6271, 1998.
MICHAEL A. GEVELBER
Boston University
602
While robustness is a concept of universal significance, robustness analysis for control systems is the study of whether
a system, however designed, can meet specified stability and
performance goals in the presence of uncertainty within a
prescribed range.
Uncertain plants can be modeled in various ways. In particular, models can be stochastic or purely deterministic. In
robust control, uncertain systems are typically modeled deterministically, as bounded sets of system models. A property is
then said to hold robustly if it holds for every model in the
set. The simplest case is that of unstructured uncertainty: the
model set consists of all systems in a certain neighborhood
(e.g., all transfer functions lying within a certain distance
of a distinguished nominal system). Such a description is
particularly appropriate to account for unmodeled dynamics.
One rather typical example is the modeling of flexible structures. It is well known that, in general, a flexible structure
cannot be accurately represented by a finite-dimensional system. For control design purposes, however, we desire, and
most often are compelled to find, an approximate finite-dimensional model with a relatively low order. In doing so, a
common practice is to include in the nominal model a small
number of dominant modes in the low-frequency range and
to treat the high-frequency modes as modeling uncertainty.
Evidently, this description is also appropriate for modeling
errors resulting from model reduction, or from any frequency
response truncation. Moreover, it can be used to cover, albeit
in a conservative way, parameter variations. The latter
amounts to drawing up a frequency response envelope to describe the range of parameter variation in frequency domain.
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
603
604
Thus each plant in the model set corresponds to the selection of one element in each of the uncertain blocks. The nominal plant corresponds to the choice 0 for all elementary uncertain blocks. Note that the assumption that all the
uncertainty balls have unity radius is made at no cost: any
size information can be embedded into known blocks (e.g.,
connected in cascade with the uncertain blocks). It turns out
that, for uncertain block diagrams of this type, the transfer
function (or transfer function matrix) between any two nodes
(or tuples of nodes) in the block diagram is given by a linear
fractional transformation (LFT). Similarly, when such an uncertain plant is connected with a feedback compensator, the
transfer function between any two nodes will also be an LFT.
LFTs have attractive mathematical properties that can be
used to advantage at the modeling, analysis, and synthesis
stages. More on LFTs can be found, e.g., in Ref. 7.
Robust Stability
It should be intuitively clear that block diagrams of the type
just described can always be redrawn in the form of an
M loop as depicted in Fig. 1 (external input and outputs
have been left out). Here M corresponds to the nominal system, which is comprised of closed-loop transfer functions as
elements and has an input channel and an output channel for
each elementary uncertain block, and is a block-diagonal
matrix whose diagonal blocks are the elementary uncertain
blocks. For its generality, the M loop paradigm has found
wide acceptance in robust control (see, for example, Refs. 2
and 48).
Throughout most of this article, we will assume that the
nominal system, or equivalently M, is linear and time-invariant, as are all instances of the uncertainty blocks, equivalently, of . We will also assume that M and all instances of
are in H . In this case, an immediate payoff of the LFT
uncertainty description and the ensuing representation of the
system via the M loop is the following strong form of the
Small Gain Theorem, a necessary and sufficient condition for
well-posedness and stability of the M loop, in the case
where consists of a single uncertainty block, ranging over
the unit ball in H .
Small Gain Theorem. Let M H . Then the M loop is
well-posed and BIBO stable for all H with 1 if
and only if M 1.
As alluded to earlier, the H norm of a causal, stable, continuous-time transfer function matrix M is defined as
M = sup (M( j))
Uncertain Systems
Throughout this article, we consider model sets P with the
following property: P can be represented by a block diagram
with some of the blocks being fully known systems, and others
being elementary uncertain blocks. The latter are elementary sets, namely, unit balls in simple vector spaces. For
example, some uncertainty blocks might be the real interval
[1, 1] and others might be the unit ball of H , the set of
transfer function matrices (linear time-invariant systems)
that are causal and bounded-input bounded-output (BIBO)
stable; the size or H -norm of H is defined to be the
supremum of its largest singular value over the imaginary
axis (continuous time) or unit disk (discrete time).
Figure 1. M loop.
As an example consider a model set of the output multiplicative uncertainty type. Specifically, let
namics, consider now the model set (1 w1c)Pa with Pa explicitly given as
P = (I + w)P0
where P0 is the transfer matrix of a linear, time-invariant
nominal model, w is a scalar weighting transfer function,
and ranges over the unit ball in H . The weight w is introduced to account for the fact that the amount of uncertainty
is usually frequency-dependent; in particular, system dynamics are often poorly known at high frequencies. Suppose that
a feedback controller K has been tentatively selected to stabilize the nominal system P0. (We use the negative feedback
convention, i.e., the loop transfer function is KP0.) Isolating
from the nominal closed-loop system, we obtain an M
loop with
M = wKP0 (I + KP0 )1
Since K stabilizes P0, M is stable.
As a note of interest, we expect a keen connection between
the Small Gain Theorem and the classical Nyquist criterion.
Indeed, this can be best observed by examining single-input/
single-output systems. In such case P and K are scalar, and
thus so is M. Since both M and are stable, Nyquists criterion implies that the M loop is stable whenever the Nyquist plot of M does not encircle the critical point 1 j0.
Clearly, this will be the case for every satisfying ( jw) 1
for all w if and only if M( jw) 1 holds at all frequencies
(including ).
The Small Gain Theorem suggests that one way to obtain
a robustly stable system, or more generally to obtain a robust
design, is to make sure that the H norm of a certain system
transfer function is small enough. This has triggered an entire field of research known as H design, which is discussed
elsewhere in this encyclopedia. The focus of the present article is the case when is block-diagonal (i.e., when the uncertainty model consists of several blocks or, in other words,
when the uncertainty is structured). Typically, two types of
uncertainty blocks are considered in the literature: (i) the set
of real, constant scalar multiples of the identity, with the scalar having magnitude no larger than one, and (ii) the set of
causal and BIBO stable (H ) transfer function matrices, with
H -norm no larger than one. The latter corresponds to unmodeled dynamics. The former, on the other hand, is used to
represent parametric uncertainty, particularly when a same
uncertain parameter affects more than one coefficients in a
transfer function. For example, concurrent variation as a
function of temperature (e.g., dilation) of multiple quantities
in a mechanical system can result in such a block. This description is more general than the simpler scalar nonrepeated blocks.
Examples with structured uncertainty arise with plants
modeled as being affected by uncertainty at more than one
physical location, e.g.,
P = (I + w1 1 )P0 (I + w2 2 )
where both input and output uncertainty are accounted for.
Another instance arises in the context of the robust performance problem, discussed in a later section. For an example
including both parametric uncertainty and unmodeled dy-
605
Pa (s) =
1
sa
w2 P0 (I + P0 K)1
M=
w1 P0 (I + P0 K)1
r 0
=
0 c
w2 KP0 (I + P0 K)1
w1 KP0 (I + P0 K)1
606
with
r := {diag(1r Ik , . . ., mr r Ik m ) : ir R}
1
c :=
C :=
C
{diag(1C , . . ., m
)
C
ic
ic
C}
{ 0 : MDM H + GM H MG 2 D < 0}
k m +m +i k m +m +i
r
c
r
c
}
inf
DD + ,GG
{diag(1c Ik
, . . ., mc c Ik m +m c )
m r +1
r
1
otherwise. It can be checked that for structures simply consisting of one full complex block as in the Small Gain Theorem, (M(j) becomes the largest singular value of M(j),
and M is thus equal to M.
Given a matrix M and a block structure , computing
(M) is generally not an easy task. Indeed, this computation
is known to be NP-hard, even when is simplified to a structure containing only full complex blocks. Thus estimates of
(M), e.g., upper and lower bounds on (M), are often used
instead. These, as well as other properties of , are discussed next.
Let U be the set of unitary matrices in and D be the set
of nonsingular matrices that commute with every . The
latter consist of block-diagonal matrices with scalar multiples
of the identity in correspondence with full complex blocks
(C), and with arbitrary blocks in correspondence with those
constrained to be scalar multiples of the identity (r, c). Then
the following result holds:
Here the superscript H indicates complex conjugate transpose, Q is the subset of consisting of matrices whose complex blocks are unitary, D is the subset of D consisting of
Hermitian positive definite matrices, G is the subset of D
consisting of skew-Hermitian matrices (i.e., GH G) with
zero blocks in correspondence with repeated real blocks in ,
and the sign indicates that the matrix expression is constrained to be negative definite. The lower bound in condition
(2) is always equal to (M). The upper bound is never greater
than that in condition (1) (it reduces to it when G 0 is imposed) and, as was the case for condition (1), can be computed
by solving an LMI problem. See, for example, Section 8.12 in
Ref. 8.
For the class of problems where the matrix M has rank
one, Young (13) showed that the right-hand side in inequalities [Eq. (2)] is equal to (M). In that case, Chen et al. (14)
obtained an explicit formula for (M). Let M baH, where a
and b are column vectors. Let also a and b be partitioned into
subvectors ai and bi compatibly with . For i 1, . . ., mr
mc, let i aiHbi. Moreover, define
m
r +m c
|i | +
i=m r +1
Then,
(M) = inf
x
m
ai 2 bi 2
|Re(i ) + x Im(i )| +
1+x
!
2
(3)
Re(l )
Re(1 )
Im(1 )
Im(l )
Re(k )
,
Im(k )
l
k = 1, . . ., l
x 0 =
;
x k =
DD
i=1
xk =
U U
m r +m c +m C
i=m r +m c +1
(1)
(2)
i=1
|Im(i )|
l
i=1
|Im(i )|
k
l
i=1 |Im(i )|
i=k+1 |Im(i )|
l
k
2
i=1 |Im(i )|
i=k+1 |Im(i )|
1
M
607
(4)
1
M
where now denotes the structured singular value corresponding to the augmented block structure diag(kpkp, )
(i.e., the block-structure corresponding to the actual uncertainty, augmented with a full complex block).
For an example of a typical robust performance problem,
consider an uncertain plant described by the multiplicative
uncertainty model set
P = (I + w1 )P0
with a fixed feedback controller K. It is desired to determine
whether w2S 1 for all possible in a possibly structured
unit uncertainty ball, where S is the sensitivity function (i.e.,
using the negative feedback convention, S (I PK)1).
Here w1 and w2 are stable transfer functions introduced for
frequency-weighting purposes. For simplicity, w1 and w2 are
assumed to be scalars. Using the transformation just outlined, we obtain
w2 (I + P0 K)1
M=
w1 (I + KP0 )1 K
w2 (I + P0 K)1 P0
w1 (I + KP0 )1 KP0
(5)
608
(7)
The problem of concern is to determine if the polynomial family P is robustly Hurwitz stable, by which we mean that every member in P is Hurwitz stable. We shall assume that the
coefficients ak(q) are continuous functions of q. Furthemore,
we assume that an(q) 0 for all q Q (i.e., all polynomials
in P have the same degree). For control system analysis, it is
typical to restrict the polynomial family P to the following
classes, arranged by order of increased complexity.
1. P a: the coefficients ak(q) are affine functions of q. For
example,
p(s, q) = s2 + (q1 + 2q2 + 3)s + (4q1 + 5q2 + 6)
2. P m: the coefficients ak(q) are multiaffine functions of q.
For example,
p(s, q) = s3 + (2q1q2 + 2q1 q3 + q3 + 1)s2
+ (4q2q3 + 5)s + (q1 q2 q3 + 1)
3. P p: the coefficients ak(q) are multivariate polynomials
in q. For example,
n
k=0
ak (q)sk ,
(6)
(8)
k=0
!1/ p
m
|qi |
, 1 p<
q p =
i=1
max |qi |,
p=
1im
(9)
where the qks and the qks are fixed. Kharitonovs original
treatment of the interval polynomial problem is of an algebraic nature. He constructed four extremal members of P ,
K1 (s) = q0 + q1 s + q2 s2 + q3 s3 + q4 s4 + q5 s5 +
K2 (s) = q0 + q1 s + q2 s2 + q3 s3 + q4 s4 + q5 s5 +
K3 (s) = q0 + q1 s + q2 s2 + q3 s3 + q4 s4 + q5 s5 +
K4 (s) = q0 + q1 s + q2 s2 + q3 s3 + q4 s4 + q5 s5 +
later dubbed Kharitonov polynomials. In a remarkable fashion, Kharitonov showed that the stability of the entire interval polynomial family can be ascertained by testing merely
these four.
Kharitonovs Theorem. The interval polynomial is Hurwitz
stable if and only if K1(s), K2(s), K3(s), and K4(s) are all Hurwitz stable.
Subsequent development showed that for polynomials of degree 5, 4, and 3, the test can be further simplified, requiring
checking only 3, 2, and 1 of the four Kharitonov polynomials,
respectively.
Dasgupta (24) gave a geometrical interpretation to Kharitonovs Theorem in frequency domain, which sheds light on
why such a puzzling result would hold. The interpretation
makes use of the concept of value set. The idea is to examine
the image of the polynomial family when s lies on the boundary of the stability region, namely
p( j, Q) = {p( j, q) : q Q}
609
Because p(s, q0) is stable and by assumption an(q) never vanishes on Q, we can conclude from the continuity properties of
the zeros of polynomials that if some member in P is not
Hurwitz stable, then some polynomial in P must have a zero
on the imaginary axis. Thus the entire family P is Hurwitz
stable if and only if
0
/ p( j, Q),
that is, the value set never contains the origin. This is a
rather straightforward fact known as early as in 1929 and is
generally referred to in the literature as the zero exclusion
principle. Note, in particular, that for a polynomial p(s, q) in
the family described by Eq. (11), for any q Q, the real and
imaginary parts of p( j, q) are respectively given by
Re p( j, q) = q0 q2 2 + q4 4
and
Im p( j, q) = j(q1 q3 2 + q5 4 )
Thus the real part (resp. imaginary part) of p( j, q) depends
only on the parameters with even (resp. odd) subscript. For
an interval polynomial, therefore, it becomes clear that
Re K1 ( j) = Re K2 ( j) Re p( j, q) Re K3 ( j) = Re K4 ( j)
Im K1 ( j) = Im K4 ( j) Im p( j, q) Im K3 ( j) = Im K2 ( j)
Because the four Kharitonov polynomials are themselves in
P , it follows that, at each , the value set is a rectangle in
the complex plane with vertices Ki( j), as depicted in Fig. 4.
As increases, the rectangle evolves in a continuous fashion. Its dimensions vary, but its edges remain parallel to the
axes, and the relative positions of the Ki( j)s do not change.
Now, a polynomial p(s) of degree n with positive coefficients
is Hurwitz stable if and only if the phase of p( j) monotonically increases from 0 to n/2 as goes from 0 to . This is
best seen by noting that p(s) can always be factorized as
p(s) an(s si) in terms of its zeros si, and by considering
the phase of each factor separately. When this is understood,
Kharitonovs Theorem becomes self-evident. Indeed, if the
Kis are Hurwitz stable, then the phase of each Ki( j) increases monotonically from 0 to n/2 and the entire Kharitonov rectangle rotates around the origin by a total angle of
n/2 without ever touching it. Stability of P then follows from
the zero exclusion principle.
Kharitonovs Theorem opened an era epitomized by the
search for so-called vertex results in robust stability analysis,
as manifested by the construction of the four Kharitonov polynomials at the vertices of the hyperrectangle Q. This theme
Im
K2(j)
K3(j)
K1(j)
K4(j)
Re
610
persisted in much of the subsequent work. Although Kharitonovs Theorem is without question a milestone in control theory, and perhaps a measured intellectual triumph in general,
we should note that the interval polynomial family is nevertheless a very special instance of an affine uncertainty class,
and hence the theorem has a rather limited scope in application. The quest thus continues.
Edge Theorem
In light of Kharitonovs Theorem, it is tempting to contemplate the possibility that a similar vertex result would hold
for the general affine polynomial family with Q a hyperrectangle. This, unfortunately, turns out to be false; a counterexample can be readily constructed to demonstrate the opposite.
What can be said about a polynomial family P in class P a?
Bartlett et al. (25) provided an answer. In what is now known
as the Edge Theorem, they took the bounding set Q to be a
convex polytope in m. Let qi, i 1, . . ., l, be the vertices of
Q. Then, it is well known that Q can be represented as a
convex hull of the vertices. That is,
Q = conv{q1 , . . ., ql } =
l
i=1
i q i :
i = 1, i 0
i=0
Xr () = (0r + (2 2 )r + (4 4 )r + )1/r
Yr () = (1r + (3 2 )r + (5 4 )r + )1/r
R() = 0 q00 2 q02 2 + 4 q04 4 +
I() = 1 q01 3 q03 2 + 5 q05 4 +
Note that R() Re(p( j, q0)) and I() Im(p( j, q0)).
Tsypkin-Polyak Criterion. The polynomial family P of Eq.
(10) is Hurwitz stable if and only if p(s, q0) is Hurwitz stable,
|R()| p |I()| p
Xr ()
Yr ()
> 1,
(0, )
(12)
M(s) = .
p(s, q0 )
p(s, q0 )
1
To see this, observe that p(s, q) in Eq. (8) is the characteristic
polynomial of the M loop of Fig. 1 with diag(q1
q10, . . ., qm qm0 ). Indeed,
m
pk (s)
p(s, q)
(qk q0k ) =
det(I M(s)) = 1 +
0
p(s, q )
p(s, q0 )
k=0
611
Yr ()/|I()|
if R() = 0
if I() = 0
Xr ()/|R()|
p (M( j)) =
()Y
()
X
r
r
p
p
p
p 1/ p
otherwise
which leads to a similar condition for robust Hurwitz stability. This condition is slightly more general than, but essentially replicates, the graphical criterion by Tsypkin and Polyak. Note that for p , the polynomial family becomes an
interval polynomial, and the stability condition reduces to
checking whether p(s, q0) is Hurwitz stable, q0 1, and
X1 () Y1 ()
,
< 1,
(0, )
min
|R()| |I()|
A little thought reveals that the latter is equivalent to determining whether one of the four conditions Re(K1( j)) 0,
Re(K3( j)) 0, Im(K1( j)) 0, and Im(K3( j)) 0 holds.
Clearly, this is further equivalent to the requirement that the
rectangular value set in Fig. 4 never contains the origin.
Extensions
There is an immense body of literature devoted to polynomial
stability problems. Various extensions to Kharitonovs Theorem have been obtained. They generally fall into the categories of vertex results and frequency-sweeping conditions, consisting of delicate studies and intricate technical details. We
summarize some of the highlights next. A recent and comprehensive account can be found in the books by Barmish (26)
and by Bhattacharyya et al. (36).
612
(13)
Here A, B, and C are known matrices of appropriate dimensions. The matrix A is assumed to be stable. The system uncertainty is represented by a set of allowed values for the
real matrix , which may be unstructured or structured. Typical perturbation classes considered in the literature are as
follows, arranged in increasing order of generality. In all
cases, 0 is given.
Unstructured Perturbation. The set consist of all real
matrices with spectral norm less than a given number:
U = { real : () }
Element-by-Element Perturbation. Each element in
varies in a given interval. Let rij 0 be given. The set
is defined as
r11 11 r1m 1m
.
..
..
E = real : =
..
.
. ,
rn1 n1 rnm nm
i, j
613
There essentially exists no result for the structured stability radius other than those already known for . For the unstructured stability radius, much of the early work was devoted to derivation of bounds. One representative example is
Re[G(s)]
Im[G(s)]
r(A, B, C)
1
C(sI A)1 B
(14)
0
i= j
max{|a ij |, |aij |}
W = [wij ],
wij =
i
= j
|aii |
614
|P|R + RT |P|
2
< 1/
(16)
(17)
r11
.
R = ..
rn1
..
.
r1n
..
.
rnn
k
1
(|PEi + EiT P|)
2 i=1
< 1/
(18)
where, again, P 0 is the unique solution to Eq. (17). Subsequent developments led to further extensions for problems
with even more detailed uncertainty descriptions. For example, the ks may be allowed to vary in asymmetric intervals.
Moreover, because rather obviously any interval matrix can
be represented alternatively in the form of A : E,
these conditions can be applied to determine the Hurwitz stability of an interval matrix as well.
Yet another issue clearly of interest is whether it is possible to derive vertex versions of these sufficient conditions.
Boyd and Yang (44) examined stability problems for matrix
polytopes. Specifically, they postulated the uncertainty description
A = conv{A1 , . . ., Ak }
A sufficient condition for A to be Hurwitz stable can be easily
found to be the existence of a P 0 such that
i = 1, . . ., k
(19)
(15)
k
1
(PEi + EiT P)
2 i=1 i
< 1/
(20)
615
BIBLIOGRAPHY
1. J. Ackermann, Robust Control Systems with Uncertain Physical
Parameters, London: Springer-Verlag, 1996.
616
2. J. Doyle, Analysis of feedback systems with structured uncertainties, IEE Proc., 129-D (6): 242250, 1982.
3. V. L. Kharitonov, Asymptotic stability of an equilibrium position
of a family of systems of linear differential equations, Differentialnye Uraveniya, 14 (11): 14831485, 1978.
4. M. G. Safonov, Stability margins of diagonally perturbed multivariable feedback systems, IEE Proc., 129-D: 251256, 1982.
5. A. Packard and J. C. Doyle, The complex structured singular
value, Automatica, 29 (1): 71109, 1993.
6. W. S. Levine (ed.), The Control Handbook, Boca Raton, FL: CRC
Press, 1996.
7. K. Zhou, J. C. Doyle, and K. Glover, Robust and Optimal Control,
Upper Saddle River, NJ: Prentice-Hall, 1996.
8. K. Zhou and J. C. Doyle, Essentials of Robust Control, Upper Saddle River, NJ: Prentice-Hall, 1998.
9. A. L. Tits and V. Balakrishnan, Small- theorems with frequency-dependent uncertainty bounds, Math. Control. Signals,
Syst., in press, 1998.
10. J. M. Krause, Structured singular value analysis of multidimensional system stability, IEEE Trans. Autom. Control, 34 (6): 638
639, 1989.
11. J. Chen and H. A. Latchman, Frequency sweeping tests for stability independent of delay, IEEE Trans. Autom. Control, 40 (9):
16401645, 1995.
12. M. K. H. Fan, A. L. Tits, and J. C. Doyle, Robustness in the presence of mixed parametric uncertainty and unmodelled dynamics,
IEEE Trans. Autom. Control, 36 (1): 2538, 1991.
13. P. M. Young, The rank one mixed problem and Kharitonovtype analysis, Automatica, 30: 18991911, 1994.
14. J. Chen, M. K. H. Fan, and C. N. Nett, Structured singular value
and stability of uncertain polynomials, Part 1: The generalized ,
Syst. Control Lett., 23 (1): 5365, 1994.
15. B. Kouvaritakis and H. Latchman, Necessary and sufficient stability criterion for systems with structured uncertainties: The
major principle direction alignment principle, Int. J. Control, 42
(3): 575598, 1985.
16. J. Chen, M. K. H. Fan, and C. N. Nett, Structured singular values
with non-diagonal structures, Part 1: Characterizations, IEEE
Trans. Autom. Control, 41 (10): 15071511, 1996.
17. J. Chen, M. K. H. Fan, and C. N. Nett, Structured singular values
with non-diagonal structures, Part 2: Computations, IEEE Trans.
Autom. Control, 41 (10): 15111516, 1996.
18. A. L. Tits, V. Balakrishnan, and L. Lee, Robustness under
bounded uncertainty with phase information, IEEE TAC, in press.
19. M. Khammash and J. B. Pearson, Robust disturbance rejection in
l1-optimal control systems, Syst. Control Lett., 14: 93101, 1990.
20. M. Khammash and J. B. Pearson, Performance robustness of discrete-time systems with structured uncertainty, IEEE Trans. Autom. Control, 36 (4): 398412, 1991.
21. J. Shamma, Robust stability with time-varying structured uncertainty. IEEE Trans. Autom. Control, 39 (4): 714724, 1994.
22. A. Megretski, Necessary and sufficient conditions of stability: A
multiloop generalization of the circle criterion, IEEE Trans. Autom. Control, 38 (5): 753756, 1993.
23. K. Poolla and A. Tikku, Robust performance against time-varying
structured perturbations, IEEE Trans. Autom. Control, 40 (9):
15891602, 1995.
24. S. Dasgupta, Kharitonovs theorem revisited, Syst. Control Lett.,
11: 381384, 1988.
25. A. C. Bartlett, C. V. Hollot, and L. Huang, Root locations of an
entire polytope of polynomials: It suffices to check the edges,
Math. Control, Signals, Syst., 1: 6171, 1988.
26. B. R. Barmish, New Tools for Robustness of Linear Systems, New
York: Macmillan, 1994.
ROUNDOFF ERRORS
48. O. Toker, On the conservatism of upper bound tests for structured singular value analysis, Proc. 35th Conf. Decision Control,
Kobe, Japan: 1996, pp. 12951300.
49. M. Fu, The real is hardly approximable, IEEE Trans. Autom.
Control, 42 (9): 12861288, 1997.
50. G. E. Coxson and C. L. DeMarco, The computation complexity of
approximating the minimal perturbation scaling to achieve instability in an interval matrix, Math. Control, Signals Syst., 7 (4):
279292, 1994.
51. S. Boyd et al., Linear Matrix Inequalities in System and Control
Theory, Studies in Applied Mathematics, Philadelphia: SIAM,
1994, vol. 15.
52. A. Megretski and A. Rantzer, System analysis via integral quadratic constraints, IEEE Trans. Autom. Control, 42 (6): 819
830, 1997.
53. R. F. Stengel and L. R. Ray, Stochastic robustness of linear timeinvariant systems, IEEE Trans. Autom. Control, 36 (1): 8287,
1992.
54. B. R. Barmish and C. M. Lagoa, The uniform distribution: A rigorous justification for its use in robustness analysis, Proc. 35th
IEEE Conf. Decision Control, 1996, pp. 34183423.
55. R. Tempo, E. W. Bai, and F. Dabbene, Probabilistic robustness
analysis: Explicit bounds for the minimum number of samples,
Proc. 35th IEEE Conf. Decision Control, 1996, pp. 34243428.
56. X. Chen and K. Zhou, On the probabilistic characterization of
model uncertainty and robustness, Proc. 36th IEEE Conf. Decision
Control, 1997, pp. 38163821.
JIE CHEN
University of California
ANDRE L. TITS
University of Maryland
617
SELF-TUNING REGULATORS
A self-tuning regulator (STR) is a controller that automatically nds its parameters in the control law. Another name
or synonym is self-adjusting controller. STR is a class of
adaptive controllers used when the process to be controlled
has constant but unknown parameters. However, STR can
also be used in an adaptive context.
BASIC IDEA
Controller Design Procedure
The design of a controller contains several steps:
1. Finding specications for the closed-loop system
2. Determination of a model for the process to be controlled
3. Decision on a design method
4. Calculation of the parameters in the controller
In many cases, it is desirable to automate these steps. This
is the idea behind adaptive and self-tuning regulators.
The specications for the closed-loop system depend on
such things as quality constraints on the controlled variable, available magnitude (power) of the control signal, and
nonlinearities of the system to be controlled. This implies
that the specications are determined by the process engineer at the start of the design procedure. The specications often lead to a natural choice of the design method.
For instance, if the main specication is to keep the process output constant and if the disturbances are occasional
large disturbances, then the design procedure can be a
method that as quickly as possible eliminates the inuence of the disturbance. The choice of specications and
design method is thus usually made by the designer of the
control loop. In STRs, as well as in adaptive controllers,
steps 2 and 4 above are automatically taken care of by the
controller.
The structure of a self-tuning controller is best described
from the block diagram in Fig. 1. The self-tuning regulator
consists of two closed loops. The rst loop is a conventional
controller feedback-loop consisting of the process and the
controller where the output of the process is measured and
compared with the desired output (reference signal) of the
closed-loop system. The mismatch between the reference
and output signals is used to compute the control action
that is sent to the process. The controller has parameters
that determine its properties. These parameters are determined by the second loop in the STR, the updating loop.
In Fig. 1, the updating loop has two main blocks. The
rst block is an estimator, which determines a mathematical model of the process based on the measured inputs
and outputs. The second block carries out the design of the
controller. This block uses the process model and the specications to determine the controller parameters that then
are sent to the controller.
It is necessary that the controller feedback-loop be
closed all the time to take care of the inuence of disturbances and changes in the reference signal. The updating
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright 2007 John Wiley & Sons, Inc.
Self-Tuning Regulators
Linear STRs are not appropriate to use when the process is very nonlinear. The updating mechanism will then
not be sufciently fast. In such cases the nonlinearities
should be built into the process model and the controller.
or
ALGORITHMS FOR SELF-TUNING CONTROL
This section describes in more detail how STRs are constructed. It also gives the main properties of STRs. To describe the algorithms we need to specify the process model,
the specications, the controller, the estimator, and the design method. We will use discrete-time models for the process and the controller since most implementations of STRs
are done using computers. It is, however, also possible to
derive continuous-time STRs.
Process Model
The process is described as a sampled-data linear system.
The process is also assumed to have a single input and a
single output. The model is given as a difference equation
Self-Tuning Regulators
The controller has a feedback part dened by the polynomials R*(q1 ) and S*(q1 ) and a feedforward part dened
by R*(q1 ) and T*(q1 ). Using Eq. (4) on the process of Eq.
(2) gives the closed-loop system
Estimator
Estimation of process models can be done in many different ways. Summaries of methods and their properties
can be found in Ljung (6), Soderstrom and Stoica (7), and
Johansson (8). Here only the recursive least squares (RLS)
method will be discussed. Dene the vectors
This is called the recursive least squares algorithm. The estimate at time k is obtained as an update of the estimate at
time k 1. The correction term depends on the latest process output, which is compared with the predicted output
based on the parameter estimate at time k 1. The matrix
P(k) can be interpreted as an estimate of the uncertainty of
the parameter estimate at time k. The statistical interpretation can be made rigorous by making assumptions about
the disturbances that are acting on the system.
The recursive least squares method is well suited for
process parameter estimation when there are no disturbances or when a white noise process is added to the righthand side of Eq. (2). For other noise or disturbance assumptions, there are variants of the recursive least squares
method that can be used.
where is the forgetting factor. Since the weights are exponentially decaying, the resulting algorithm is called recursive least squares with exponential forgetting. The updating formulas are only slightly modied into
Self-Tuning Regulators
Since B+ is canceled, it must be a factor of Ac . The closedloop characteristic polynomials is thus of the form
Also, linear quadratic gaussian controllers can be interpreted as solving a special form of the diophantine equa om and Wittenmark (9).
tion, see Astr
Design of Self-Tuning Regulators
The design of STRs can be summarized by the following
procedure:
Specications. Determine the class of controller by determining the specications on the closed-loop system.
Estimation. Estimate the process parameters using, for
instance, the recursive least squares algorithm of
Eq. (7).
Design Procedure. Determine the controller parameters using the estimated process parameters as if
they are the correct ones. The controller design is
usually reduced to the solution of an equation such
as the diophantine Eq. (8).
Control. Update the parameters of the controller, for instance, in the form of the controller in Eq. (4).
Example 1: Indirect Deterministic Self-Tuning Regulator Assume that the open-loop process is described by
the continuous-time system
Self-Tuning Regulators
The estimated parameters are thus the same as the controller parameters, and the design step has been eliminated.
Self-Tuning Regulators
Elimination of Disturbances
We will now see how the inuence of disturbances can be
reduced by introducing integrators and by using feedforward.
Introduction of Integrators. Consider the process
Figure 4. The accumulated loss function V(k) when the direct
self-tuning algorithm (full) and the optimal minimum variance
controller (dashed) are used on the process in Example 2.
Figure 5. The controller gain s 0 /I 0 when the self-tuning algorithm is used (full). The gain of the optimal minimum controller
is shown as a dashed line.
The new model has the equation error e(k) instead of v(k).
The process model can now be estimated from Eq. (17).
Based on the estimated model the controller design is done
by solving the diophantine equation
The controller now contains the factor Ad , which will eliminate the inuence of the disturbance v.
In the direct minimum variance self-tuning algorithm
an integrator can be introduced by changing the model of
Eq. (16) and estimating the controller parameters from
Self-Tuning Regulators
results have lately been rened and extended [see Well om and Wittenmark (5)].
stead and Zarrop (4) and Astr
One important theoretical aspect is the inuence of unmodeled dynamics. Unmodeled dynamics are present if the
estimator is trying to t a too-simple model to the data. The
unmodeled dynamics may cause severe stability problems,
which must be avoided by introducing counter measures
such as careful ltering of the signals in the STR. This type
of problem has successfully been analyzed using averaging
theory.
It is important that a controller is robust against assumptions and choices of controller parameters. Much theoretical research has been devoted to make STRs and
adaptive controllers more robust. This work has resulted
in practical rules of thumb for their implementation (see
Ref. 5). Robust design methods are complementary to selftuning and adaptive control. In robust control one xed
controller is designed to cope with a variety of processes.
By using tuning and adaptation the parameters of the controller are instead tuned to adjust to the present process
dynamics.
PRACTICAL ISSUES AND IMPLEMENTATION
The rst part of the controller is the feedback from the measurement y(k) and the second is the feedforward from the
measurable disturbance vm (k). Feedforward is, in general,
very useful in STRs because to make effective feedforward,
it is necessary to have a good model of the process. By combining the measurement of the disturbance and the selftuning property of the controller it is possible to eliminate
much of the disturbance before it reaches the output of the
process.
SOME THEORETICAL PROBLEMS
The previous section described the basic ideas of STRs.
Self-tuning regulators are inherently nonlinear. The nonlinearities are due to the estimation part and the changing
parameters in the controller. This makes the analysis of
STRs very difcult. The STRs contain two feedback loops
and it is necessary to investigate the stability and convergence properties of the closed-loop systems. This is a difcult question because of the interaction between the two
feedback loops. One way to circumvent this problem is to
make a time separation between the two loops. The controller loop is assumed to be fast compared to the updating
loop. This makes it possible to use averaging theory to analyze the updating loop on a much longer time-scale. This
approach has made it possible to derive results concerning
stability and convergence of STRs.
om and Wittenmark (2) showed how to characterize
Astr
the stationary properties of STRs, that is, the properties if
and when the parameter estimation has converged. The
algorithms were used in a number of applications before
several of the theoretical problems were solved. Goodwin,
Ramadge, and Caines (3) gave the rst results showing
when the algorithm converges and that the closed-loop system remains stable during the estimation phase. These
Some problems in the implementation of STRs are discussed briey in this section. Self-tuning regulators as well
as adaptive controllers will run unattended on the processes. It is therefore very important that there be a good
safety net around the self-tuning algorithm.
There are many aspects of STR implementation that
are important for implementations of digital controllers in
om and Wittenmark (9)]. Some important
general [see Astr
issues for STRs are
Self-Tuning Regulators
BIBLIOGRAPHY
1. R. E. Kalman, Design of self-optimizing control systems.
ASME Trans., 80: 468478, 1958.
om, B. Wittenmark, On self-tuning regulators. Auto2. K. J. Astr
matica, 9: 185199, 1973.
3. G. C. Goodwin, P. J. Ramadge, P. E. Caines, Discrete-time multivariable adaptive control, IEEE Trans. Autom. Control, AC25: 449456, 1980.
4. P. E. Wellstead, M. B. Zarrop, Selftuning Systems: Control and
Signal Processing, Chichester, U.K.: Wiley, 1991.
om, B. Wittenmark, Adaptive Control, 2nd ed. Read5. K. J. Astr
ing, MA: Addison-Wesley, 1995.
6. L. Ljung, System IdenticationTheory for the User.
Englewood Cliffs, NJ: Prentice-Hall, 1987.
7. T. Soderstrom, P. Stoica, System Identication. Hemel Hempstead, U.K.: Prentice-Hall International, 1988.
8. R. Johansson, System Modeling and Identication. Englewood
Cliffs, NJ: Prentice-Hall, 1993.
om, B. Wittenmark, Computer-Controlled Systems,
9. K. J. Astr
3rd ed. Englewood Cliffs, NJ: Prentice-Hall, 1997.
10. K. F. Gauss, Theoria motus corposum coelestium, (1809). English translation, Theory of the Motion of the Heavenly Bodies.
New York: Dover, 1963.
BJORN
WITTENMARK
Lund University
Lund, Sweden
302
(1)
(2)
+
u
+
v
Figure 2. RC circuit.
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
x(t0 ) = x0
(3a)
z = g(x, z, , t),
z(t0 ) = z0
(3b)
where x Rns, z Rnf, and 0 is the small singular perturbation parameter. The parameter represents the small time
constants and other parasitics to be neglected in the slow
time-scale analysis.
The rate of change of z in Eq. (3b) is of order 1/ , that is,
z O(1/ ), which means that z exhibits a fast transient.
When this fast transient settles, the longer-term behavior of
x and z is determined by the quasi-steady-state equation
g(x, z, 0, t) = 0
(4)
where the bar indicates that this equation is obtained by setting 0 in Eq. (3b). This equation will make sense only if
it has one or several distinct (isolated) roots
z = (x, t)
(5)
for all x and z of interest. If this crucial requirement is satisfied, for example, when det(g/z) 0, then we say that system (3) is a standard model.
The substitution of Eq. (5) into Eq. (3a) results in the reduced model
x = f (x, (x, t), 0, t),
x(t0 ) = x0
303
f (x, y + (x, t), , t),
t
x
y(t0 ) = z0 (x0 , t0 )
(7)
dy
d
1
dy
=
; hence,
=
dt
d
dt
dy
= g(x, y + (x, t), , t)
f (x, y + (x, t), , t),
d
t
x
y(0) = z0 (x0 , t0 ) (8)
In the fast time-scale , the variables t and x are slowly
varying because t t0 and x x(t0 ). Setting 0
freezes these variables at t t0 and x x0 and reduces Eq.
(8) to the autonomous system
dy
= g(x0 , y + (x0 , t0 ), 0, t0 ),
d
y(0) = z0 (x0 , t0 )
(9)
(6)
Tikhonovs Theorem
If Eq. (4) has several distinct roots as shown in Eq. (5), then
each of them leads to a distinct reduced model as shown in
Eq. (6). The singular perturbation analysis determines which
of these models provides an O() approximation of the slow
phenomenon in system (3).
When, and in what sense, will x(t), z(t) obtained from Eqs.
(6) and (5) be an approximation of the true solution of system
(3)? To answer this question, we examine the variable z,
which has been excluded from the reduced model in Eq. (6)
by z (x, t). In contrast to the original variable z, which
starts at t0 from a prescribed z0, the quasi-steady state z is
not free to start from a prescribed value, and there may be a
large discrepancy between z(t0) (x0, t0) and the prescribed
initial state z0. Thus, z(t) cannot be a uniform approximation
of z. The best we can expect is that the approximation z
z(t) O() will hold on an interval excluding t0, that is, for
t [tb, tf ] where tb t0. On the other hand, it is reasonable to
expect the approximation x x(t) O() to hold uniformly
for all t [t0, tf ] because x(t0) x(t0). If the error z z(t) is
indeed O() over [tb, tf ], then it must be true that during the
initial (boundary-layer) interval [t0, tb] the variable z approaches z. Let us remember that the speed of z can be large
since z g/ . In fact, having set 0 in Eq. (3b), we have
made the transient of z instantaneous whenever g 0. It is
(10)
(11)
(12a)
+ O()
z = (x(t), t) + y(t/)
(12b)
304
(13)
which (x, ) must satisfy for all x of interest and all (0,
*]. This is a partial differential equation, which, in general,
is difficult to solve. However, its solution can be approximated
by the power series
(x, ) = 0 (x) + 1 (x) + 2 2 (x) +
(21)
where the functions 0(x), 1(x), . . ., can be found by equating the terms with like powers in . To this end, we expand
f and g as power series of
f
(x) + (22)
z 1
g
g(x, 0 (x) + 1 (x) + ) = g(x, 0 (x)) + 1 (x) + (23)
z
f (x, 0 (x) + 1 (x) + ) = f (x, 0 (x)) +
(15)
W
g(x, y + (x, t), 0, t) c3 y2
y
(16)
over the domain of interest, where c1 to c3 are positive constants independent of (x, t).
Slow Manifold
In the state-space Rns Rnf of (x, z), the equation g(x, z, 0, t)
0 forces x and z to lie in an ns-dimensional quasi-steadystate manifold M, explicitly described by Eq. (5). It can be
shown that, under the conditions of Tikhonovs Theorem,
there exists an * 0 such that for all (0, *], the system
in Eq. (3) possesses an integral manifold M that is invariant:
whenever x(t0), z(t0) M, then x(t), z(t) M for all t [t0,
tf ]. The slow manifold M is in the -neighborhood of the quasisteady-state manifold M. We will characterize M in the special case when f and g in the system in Eq. (3) do not depend
on t and :
x = f (x, z)
(17a)
z = g(x, z)
(17b)
(25)
For the standard model, det(g/z) 0, that is, g/z is nonsingular, so that
1
g
0 (x)
f (x, 0 (x))
1 (x) =
(26)
z
x
This recursive process can be repeated to find the higherorder terms in Eq. (21).
The fast off-manifold variable is
y = z (x, )
(27)
x = f (x, (x, ) + y)
(28a)
y = g(x, (x, ) + y)
f (x, (x, ) + y)
x
(28b)
(18)
d
(x, ) =
x
z =
dt
x
(19)
x2
(20)
(24)
x(0)
x1
Figure 3. Trajectory converging to a slow manifold.
A11 A12 L
0
=
y
0
A22 + LA12 y
B1 H(B2 + LB1 )
+
u
B2 LB1
x(t0 ) = x0
(29a)
z = A21 x + A22 z + B2 u,
z(t0 ) = z0
(29b)
(30)
(31)
2
1
2
L() = A1
22 A21 + A22 A21 (A11 A12 A22 A21 ) + O( )
(34)
(35)
L0 = A1
22 A21
(36)
can be used.
Although the fast variable y in the triangular system in
Eq. (32) is now decoupled from the slow variable x, the slow
variable x is still driven by the fast variable y. To remove this
influence, the change of variables
= x H()y
(37)
(t0 )
x0 H(z0 + Lx0 )
=
y(t0 )
z0 + Lx0
(40)
(41)
H0 = A12 A1
22
(42)
(33)
Note that Eq. (31) is the slow manifold condition in Eq. (20)
for linear systems.
Given that A22 is nonsingular, the implicit function theorem implies that Eq. (31) admits a solution L() for sufficiently small. Furthermore, an asymptotic expansion of the
solution to Eq. (31) is given by
(39)
305
(t0 ) = x0
(43)
1
where A0 A11 A12A1
22 A21 and B0 B1 A12A22 B2. The fast
subsystem is approximated to O() by
y = A22 y + B2 u,
y(t0 ) = z0 A1
22 A21 x0
(44)
Thus as 0, the slow eigenvalues of Eq. (29) are approximated by (A0), and the fast eigenvalues are approximated
by (A22)/ . It follows that if Re(A0) 0 and Re(A22) 0,
then there exists an * 0 such that (29) is asymptotically
stable for all (0, *]. Furthermore, if the pair (A0, B0) and
the pair (A22, B2) are each completely controllable (stabilizable), then there exists an * 0 such that Eq. (29) is completely controllable (stabilizable) for all (0, *].
EXAMPLES
Example 1. An RLC Circuit
To complete our introductory example, we represent the RLC
circuit in Eq. (1) using the state variables x v and z v:
x = z
z = z
(45a)
1
(x u)
RC
(45b)
RC x = x + u
(38)
(46)
y(t0 ) = z(t0 ) +
1
(x(t0 ) u(t0 ))
RC
(47)
Speed (rad/s)
Current (A)
306
52.6
52.5
52.4
52.3
52.2
0.1
0.2
0.3
Time (s)
0.4
0.5
0.1
0.2
0.3
Time (s)
0.4
0.5
3.45
3.40
J = Ki TL
(48a)
Li = K Ri + u
(48b)
where i, u, R, and L are the armature current, voltage, resistance, and inductance, respectively, J is the combined moment of inertia of the motor and the load, is the angular
speed, TL is the load torque, and K is a motor design constant
such that Ki and K are, respectively, the motor torque and
the back emf (electromotive force).
We consider the case in which the electrical time constant
e L/R is much smaller than the mechanical time constant
m JR/K2 (3). Defining e / m, x , and z i, we rewrite
Eq. (48) as
1
R
z TL
m K
J
1
1
K
x
u
z+
z =
m R
m
m R
x =
(49a)
(49b)
dy
= y,
d
y(0) = z(0)
u(0) Kx(0)
R
(53)
(50)
(51)
(54)
which, when substituted in Eq. (49a), yields the slow mechanical subsystem
m x = x +
m
1
u
T
K
J L
(55a)
z = [z + (1 + t)x] z [z (1 + t)]
Figure 4. dc motor.
(56)
(55b)
where the initial conditions are x(0) 1 and z(0) z0. Setting
0 results in
0 = [z + (1 + t)x] z [z (1 + t)]
(52)
x = x2 (1 + t)/z
TL
z = (1 + t)x;
z = 0;
z = 1+t
(57)
(58)
x(0) = 1
y(0)
= z0 + 1
ks
bs
mu
dy
= [y + (1 + t)x] y [ y (1 + t)]
d
dy
= [ y + (1 + t) + (1 + t)x][y + (1 + t)] y
d
(62)
(63)
has the unique solution x(t) 1/(1 t) for all t [0, 1).
Notice that x(t) has a finite escape time at t 1. However,
Tikhonovs Theorem still holds for t [0, tf] with tf 1. The
boundary-layer system, with t 0 and x 1,
y(0)
= z0 1
Tire
kt
dr
Road
surface
Reference
As in the first case, it can be shown that the origin is exponentially stable uniformly in (x, t). The reduced system
x = x2 , x(0) = 1
du
(61)
full model (solid curves) and two for each reduced model
(dashed curves).
Example 4. A Quarter-Car Model
A simplified quarter-car model is shown in Fig. 7, where ms
and mu are the car body and tire masses, ks and kt are the
spring constants of the strut and the tire, and bs is the
damper constant of the shock. The distances ds, du, and dr are
the elevations of the car, the tire, and the road surface, respectively. From Newtons Law, the balance of forces acting
on ms and mu results in the modeling equations
ms ds + bs (ds du ) + ks (ds du ) = 0
(65a)
(65b)
(64)
=
ks /ms
=
kt /mu
0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2
Time (s)
k s mu
k t ms
(66)
ds du
,
x=
ds
2
0
ds
(60)
dy
= (y + 2)(y + 1)y,
Car
body
ms
(59)
has the unique solution x(t) et for all t 0. The boundarylayer system with t 0 and x 1 is
dy
= y(
y 1)(y 2),
d
307
(du dr )/
z=
du
(67)
(68a)
z = A21 x + A22z + B2 u
(68b)
308
where
0
=
0
0
B1 =
0
0
1
1
, A12
,
ks /ms bs /ms
bs /ms
0
0
0
1
, A22 =
,
A21 =
ks /ms bs /ms
ks /ms bs /ms
1
(69)
B2 =
0
A11 =
and
k s ms
k t mu
(70)
(71a)
VT = h(x)
(71b)
where x is the machine state vector including the flux variables and the scalar output VT is the generator terminal voltage. Here we focus on the exciter system which, from Fig. 8,
is described by
TM E = E KM KG Efd + KMVR
Efd = VB (x)E
(72a)
(72b)
KM
1 + sTM
Efd
Synchronous
machine
VB(x)
VT
1
KMV B
1
(73)
V
(75)
TM
VB (x)
TM R
which yields, as 0, the quasi-steady state
E fd =
1
V
KG R
(76)
d
= s
(77a)
dt
2H d
1
1
1
1
= Tm +
d q + d Ed + q Eq (77b)
s dt
Lq
Ld
Lq
Ld
Tqo
KG
=
Tdo
VR
dEq
dt
Ld Ld
Ld
E
d + Efd
q
Ld
Ld
Lq Lq
dEd
Lq
= Ed
q
dt
Lq
Lq
(77c)
(77d)
Ra
Ra
1 dd
= d + Eq +
q + V sin
s dt
Ld
Ld
s
(77e)
1 dq
Ra
Ra
= q Ed
+ V cos
s dt
Lq
Lq
s d
(77f)
d = V cos ,
q = V sin
(78)
yq = q + V sin
(79)
dyd
=
yq ,
yd (0) = d (0) V cos (0)
dt
s
dyq
= yd , yq (0) = q (0) + V sin (0)
dt
s
(80a)
(80b)
x = z
z = x + z
(81a)
1 3
3z
(81b)
(82)
(83)
and the eigenvalue condition of Tikhonovs Theorem is satisfied because z2 1. Therefore, the branches AB and CD are
attractive, that is, trajectories converging to these two
branches will remain on them, moving toward either the
point B or C. However, the root on the branch BC is unstable
because z2 1; hence, this branch of the slow manifold is repulsive.
Figure 9 shows vertical trajectories converging toward AB
and CD, because 0. The mechanism of two interior (rather
3
A
2
1
z
(Ld, Lq) are the d- and q-axis voltages, flux linkages, opencircuit time constants, synchronous reactances, and transient
reactances, respectively, Ra is the stator resistance, Tm is the
input mechanical torque, Efd is the excitation voltage, and s
is the system frequency.
In the model shown in Eq. (77), the slow variables are ,
, Ed, and Eq, and the fast variables are d and q. The singular perturbation parameter can be defined as 1/s.
If the stator resistance is neglected, that is, Ra 0, it can
be readily verified that the slow manifold condition in Eq. (20)
gives the exact slow invariant manifold
309
2
3
2 1.5 1 0.5
D
0
x
0.5
D
1
1.5
STABILITY ANALYSIS
We consider the autonomous singularly perturbed system in
Eq. (17). Let the origin (x 0, z 0) be an isolated equilibrium point and the functions f and g be locally Lipschitz in a
domain that contains the origin. We want to analyze stability
of the origin by examining the reduced and boundary-layer
models. Let z (x) be an isolated root of 0 g(x, z) defined
in a domain D1 Rn that contains x 0, such that (x) is
continuous and (0) 0. With the change of variables y
z (x), the singularly perturbed system is represented in
the new coordinates as
x = f (x, y + (x))
y = g(x, y + (x))
f (x, y + (x))
x
(84a)
(84b)
310
x D1
(85)
(x, y) D1 D2
(86)
(x, y) D1 D2
(87)
0<d<1
(88)
where the constant d is to be chosen. Calculating the derivative of along the trajectories of the full system in Eq. (84),
we obtain
d W
V
f (x, (x)) +
g(x, y + (x))
= (1 d)
x
y
V
[ f (x, y + (x)) f (x, (x))]
+ (1 d)
x
W
W
+d
f (x, y + (x))
x
y x
(89)
1 (x)
2 ( y)
T
(1 d)1
d)1 12 d2
12 (1
12 (1
d)1 12 d2
d 2
1 (x)
2 ( y)
(92)
1 2
1 + 1 2
(94)
(95a)
(95b)
(96)
W
f (x, y + (x)) = y(x3 + y) 1 2 + 22
y
(97)
Note that W/x 0. Hence, Eqs. (90) and (91) are satisfied
with 1 2 1. Therefore, the origin is asymptotically
stable for * 0.5. Because all the conditions are satisfied
globally and (x, y) (1 d)V(x) dW(y) is radially unbounded, the origin is globally asymptotically stable for
0.5.
The preceding two-time-scale stability analysis can be extended to the nonautonomous singularly perturbed system in
Eq. (3). For sufficiently smooth f, g, and , it can be shown
that if the origin of the reduced system is exponentially stable, and the origin of the boundary-layer system is exponentially stable, uniformly in (t, x), then the origin of the full system in Eq. (3) is exponentially stable for sufficiently small .
The same conditions ensure the validity of Tikhonovs Theorem for all t t0. Note that the earlier statement of Tikhonovs Theorem, which does not require exponential stability of
the reduced system, is valid only on a compact time interval
[t0, tf] for a given tf.
COMPOSITE FEEDBACK CONTROL
The impetus for the systematic decomposition of the slow and
fast subsystems in singularly perturbed systems can be
readily extended to the separate control design of the slow
and fast dynamics. As will be shown, the crucial idea is to
compensate for the quasi-steady state in the fast variable.
(98a)
z = g(x, z, u)
(98b)
(99)
(100)
(101)
(102a)
(102b)
(103)
311
(111)
where
1
G1 = (Im + G2 A1
22 B2 )G0 + G2 A22 A21
(112)
(104)
(105)
(106)
A1
22 (A21 x
+ B22 G0 x)]
, G2 y
(114)
(115)
qs (t) = C0 + D0 us
(116)
where
with
C0 = C1 + C2 A1
22 A21 ,
D0 = C2 A1
22 B2
(117)
and
qf (t) = C2 y
(118)
From the subsystems in Eqs. (43) and (44) and the decomposition in Eq. (115), the linear quadratic regulator problem in
Eq. (113) can be solved from two lower-order subproblems.
(107)
(108)
Find the slow control us for the slow subsystem in Eqs. (43)
and (116) to minimize
1 T
Js (us ) =
(qs qs + uTs Rus ) dt, R > 0
2 0
(119)
1 T T
=
( C0 C0 + 2uTs DT0 C0T + uTs R0 us ) dt
2 0
(109)
y = (A22 + B2 G2 )y
(110)
312
applied to the system in Eq. (29) achieves an O() approximation of Jopt, that is,
where
R0 = R + DT0 D0
(120)
If the triple (C0, A0, B0) is stabilizable and detectable (observable), then there exists a unique positive-semidefinite
(positive-definite) stabilizing solution Ks of the matrix Riccati
equation
T
1 T
T
0 = Ks (A0 B0 R1
0 D0 C0 ) (A0 B0 R0 D0 C0 ) Ks
T
1 T
+ Ks B0 R1
0 B0 Ks C0 (I D0 R0 N0 )C0
(121)
T
R1
0 (D0 C0
(129)
In Theorem 1.1, an asymptotic expansion exists for the solution to the matrix Riccati equation associated with the full
linear regulator problem. Theorem 1.3 is one of the robustness results with respect to fast unmodeled dynamics,
that is, if the fast dynamics is asymptotically stable, a feedback control containing only the slow dynamics would not destabilize the fast dynamics.
APPLICATIONS TO LARGE POWER SYSTEMS
+ BT0 Ks )
= G0
(122)
1 T
(qf qf + uTf Ruf ) dt, R > 0
2 0
1 T T
=
(y C2 C2 y + uTf Ruf ) dt
2 0
Jf (uf ) =
(123)
If the triple (C2, A22, B2) is stabilizable and detectable (observable), then there exists a unique positive-semidefinite
(positive-definite) stabilizing solution Kf of the matrix Riccati
equation
0 = Kf A22 AT22 Kf + Kf B2 R1 B2 Kf C2T C2
(124)
(125)
In this section, we analyze a large power system as an example of time-scales arising in an interconnected systems. A
power system dispersed over a large geographical area tends
to have dense meshes of power networks serving heavily populated areas and many fewer transmission lines interconnecting these urban centers. When such a system is subject to a
disturbance, it is observed that groups of closely located machines would swing coherently at a frequency that is lower
than the frequency of oscillation within the coherent groups.
Singular perturbation techniques have been successfully applied to these large power networks to reveal this two-timescale behavior (9).
Consider the linearized electromechanical model of an
n-machine power system in the second-order form with damping neglected
M = K
where Rn is the machine rotor angle vector, M is the diagonal matrix of machine inertias, and K is the stiffness matrix
determined by the network impedances. Assume that the system in Eq. (130) has r tightly connected areas, with the connections between the areas being relatively fewer. In this
case, we decompose K into
K = K I + K E
Theorem 1.
1. If the triples (C0, A0, B0) and (C2, A22, B2) are stabilizable
and detectable (observable), then there exists an * 0
such that for all (0, *], an optimal control exists
for the linear regulator problem (113) with an optimal
performance Jopt.
2. The composite control in Eq. (111)
1
T
T
uc = [(I R1 BT2 Kf A1
22 B2 )R0 (D0 C0 + B0 Ks )
1 T
+ R1 BT2 Kf A1
22 A21 ]x R B2 Kf z
(126)
applied to the system in Eq. (29) achieves an O( ) approximation of Jopt, that is,
2
(127)
(128)
(130)
(131)
where K I is the stiffness matrix due to the impedances internal to the areas, and K E is the stiffness matrix due to the
impedances external to the areas and scaled by the small parameter that represents the ratio of the external to the internal connection strength.
For illustrative purposes, we let r 2, with r1 machines in
area 1 and r2 machines in area 2. Arranging the machines in
area 1 to appear first in , K I has a block-diagonal structure
K I = block-diag(K1I , K2I )
(132)
(134)
(135)
where
1r 1
1
G=
0
Ir
1 1
0
1r 1
2
0
Ir
(136)
2 1
K EUa + (GM
K I G+ + GM
(137a)
1
K E G+ )d (137b)
313
The composite control has also been applied to solve optimal regulator problems for nonlinear singularly perturbed
systems (15). Recently, composite control results for H optimal control of singularly perturbed systems have been obtained (16, 17). The composite control idea can also be used
to establish the time scales in a closed-loop system induced
by a high-gain control (18).
Singular perturbation methods also have significant applications in flight control problems (19, 20). For example, twopoint boundary-value problems arising from trajectory optimization can be solved by treating the fast maneuvers and
the slower cruise dynamics separately.
The slow coherency and aggregation technique is also applicable to other large-scale systems, such as Markov chains
(21) and multi-market economic systems (22). These systems
belong to the class of singularly perturbed systems in the nonstandard form, of which an extended treatment can be found
in Ref. 23.
A topic not covered here is the filtering and stochastic control of singularly perturbed systems with input noise. As
0, the fast dynamics will tend to a white noise. Although
the problem can be studied in two-time-scales, the convergence of the optimal solution requires that the noise input in
the fast dynamics be either colored or asymptotically small
(24).
Another topic not covered is the control of distributed parameter systems possessing two-time-scale properties. Averaging and homogenization techniques are also a class of twotime-scale methods (25). More developments are expected in
this area.
In more complex singularly perturbed systems, jump behaviors may arise not only at the end points but also in interior layers. Reference 27, beside being an introductory text to
singular perturbations, contains a detailed treatment of such
phenomena. It also contains a historical development of singular perturbations.
The singular perturbation results presented in this article
represent developments over a period of three decades and
contribute to the advances of modern control theory. As new
control problems are proposed and new applications are discovered for systems with time-scales, we expect that singular
perturbation methods will also be extended accordingly to
provide simpler models to gain useful design insights into
these new problems.
BIBLIOGRAPHY
1. P. K. Kokotovic, H. K. Khalil, and J. OReilly, Singular Perturbation Methods in Control: Analysis and Design, London: Academic
Press, 1986.
2. P. K. Kokotovic and H. K. Khalil (eds.), Singular Perturbations
in Systems and Control, New York: IEEE Press, 1986.
3. A. E. Fitzgerald, C. Kingsley, and A. Kusko, Electric Machinery,
3rd ed., New York: McGraw-Hill, 1973.
4. IEEE Committee Report, Excitation system models for power system stability studies, IEEE Trans. Power Appar. Syst., PAS-100:
494509, 1981.
5. P. V. Kokotovic and P. W. Sauer, Integral manifold as a tool for
reduced-order modeling of nonlinear systems: A synchronous machine case study, IEEE Trans. Circuits Syst., 36: 403410, 1985.
6. H. K. Khalil, Nonlinear Systems, 2nd ed., Upper Saddle River,
NJ: Prentice-Hall, 1996.
314
SKIN EFFECT
PETAR V. KOKOTOVIC
University of California, Santa
Barbara
JOE H. CHOW
Rensselaer Polytechnic Institute
HASSAN K. KHALIL
Michigan State University
At first contact, symbolic software is reminiscent of university-level engineering analysis and calculus. It
offers the flexibility, excitement, and expectation of computing exactly with a completely abstract representation
of the problem at hand. Unfortunately, it shares the same drawback, in that the problems that can be dealt with
are limited in their complexity. Nevertheless, using symbolic computation, software computations that involve
cumbersome algebraic expressions that are far beyond error-prone manual manipulations become routine. As
a result, symbolic computation software has significantly pushed the boundaries of what can be regarded as an
analytically tractable mathematical problem, and so it has become an invaluable tool in the control designers
toolbox for modeling, design, and analysis of control systems.
A plethora of symbolic computation algorithms targeting control system analysis and design are available. Symbolic computation is useful in almost every step of control system modeling and the analysis and
controller design of linear and nonlinear control systems. In particular, symbolic manipulation of bond graphs
and multidimensional Laplace transforms, which are important tools in the modeling of nonlinear control
systems, are facilitated by computer algebra. The analysis of qualitative and quantitative system properties
under parameter variations is supported by symbolic computation through sensitivity and bifurcation analysis.
Robust control problems in either linear or nonlinear systems are facilitated by computer algebra software.
The important geometric approach to nonlinear control design (e.g., computing a linearizing feedback or determining zero dynamics) is enabled by commercially available symbolic software. A more elaborate overview of
other symbolic software algorithms used in modeling, analysis, and design of control systems is provided in 2.
In this article, polynomial control system analysis and design are targeted, and in particular three important methods are discussed: Grobner bases, differential algebra (in particular, Ritts algorithm), and quantifier
elimination. In essence, these methods deal with polynomials. Computation of Grobner bases is implemented in
most commercial symbolic software packages such as Maple and Mathematica. On the other hand, Ritts algorithm is not as yet commercially available, but software implementations may be obtained from the specialists
working in the field. Quantifier elimination has been implemented in Mathematica.
Polynomial systems are a natural generalization of linear systems; they possess a nice algebraic structure,
and their analysis is amenable to computer algebra. Moreover, as polynomials are universal approximators,
polynomial control models can be used as valid models for almost any physical system. Furthermore, the
restriction to polynomial systems is not severe, since any nonlinear function that satisfies a polynomial differential or algebraic equation, such as trigonometric functions, can be handled by introducing artificial states;
see 3 and 4 for more details.
The state-space model of a polynomial control system is given by
Here x n , y p , u m represent respectively the state, the observed output, and the manipulated control
input of the system. The operator is either the derivative x(t) = dx(t)/dt for continuous-time systems (when
t ) or the forward difference x(t) = x(t + 1) for discrete-time systems (when t ). The symbols , , ,
and
denote respectively the sets of real, rational, natural, and complex numbers. The vector functions f and
h are assumed to have entries that are polynomial functions in all their variables. Moreover, for computational
reasons the polynomials are assumed to have rational coefficients. Continuous-time (or discrete-time) linear
systems
form a special subclass of polynomial control systems (1). A wide range of systems can be adequately modeled
using Eq. (1). Indeed, systems as diverse as biochemical reactors, grain dryers, satellites, robots, and airplanes
have been successfully modeled using polynomial systems (1) (see, e.g., 5).
In the context of polynomial control systems, problems like determining equilibria, estimating a domain
of attraction of a stable equilibrium, and locating approximately periodic solutions, as well as testing controllability and observability properties, reduce naturally to the analysis of sets of algebraic polynomial equations.
Such problems can be addressed using the Grobner basis method. In any modeling exercise, the elimination
of variables from sets of differential and algebraic equations plays a crucial role. This problem also arises
naturally in the context of determining if two state representations are equivalent. Differential algebra and
Ritts algorithm are very useful in this context. A range of different control problems can be formulated as
questions involving quantifiers there exists and for all, and this naturally leads to the so-called quantifier
elimination algorithms.
The Grobner
Basis Method
Central objects in the Grobner basis theory are polynomial ideals and affine varieties (6). Let p1 ,. . ., ps be
multivariate polynomials in the variables x1 ,. . . , xn whose coefficients are in the field k. For the collection of
ordered n-tuples of elements of k the notation kn is used. The variables x1 ,. . . , xn are considered as place
markers in the polynomial. The notation p1 ,. . . , ps k[x1 ,. . . , xn ] is adopted. The affine variety (or variety)
defined by the s polynomials p1 ,. . . , ps is the collection of all solutions in kn of the system of equations
For instance, a straight line, a parabola, an ellipse, a hyperbola, and a single point are all examples of varieties
in 2 . The polynomial ideal I that is generated by p1 ,. . . , ps is a set of polynomials obtained by combining these
polynomials through multiplication and addition with other polynomials: formally,
The polynomials pi , i = 1,. . . , s, form a basis for the ideal I. A very useful interpretation of a polynomial ideal
I is in terms of the equations (3). On multiplying pi by arbitrary polynomials gi k[x1 ,. . . , xn ] and adding the
products, the implication of Eq. (3) is that f = g1 p1 + + gs ps = 0, and of course f I. Hence, the ideal I =
p1 ,. . . , ps contains all the polynomial consequences of the equations (3).
A notion at the core of the Grobner basis method is that of the monomial ordering (a monomial is a
polynomial consisting of a single term), since it introduces an appropriate extension of the notion of the leading
term and the leading coefficient familiar from univariate polynomials to multivariate polynomials. One can
define many different monomial orderings [lexicographic, graded lexicographic, graded reverse lexicographic,
etc. (6)], but to be specific, consider the so-called lexicographic, or lex, ordering. Let , be two n-tuples of
integers [ = (1 ,. . ., n ), = (, . . . , n ) n . The n-tuple is said to succeed (in lex ordering), denoted as
, if in the vector difference = (1 1 ,. . ., n n ) the leftmost nonzero entry is positive. Obviously, it
is possible to define n! different lex orderings for polynomials in n variables. For the polynomial f = 2x1 3 x2 x3 3
+ 4x1 3 x3 5 , using the lex ordering x1 x2 x3 the monomial x1 3 x2 x3 3 succeeds the monomial x1 3 x3 5 , as the
multidegrees of the monomials satisfy (3,1,3) (3,0,5). With this ordering, the leading coefficient and the
leading term are respectively LC(f ) = 2 and LT(f ) = 2x1 3 x2 x3 3 . Using the alternative lex ordering x1 x2 x1 ,
the leading term becomes LT(f ) = 4x1 3 x3 5 .
In general, an ideal I does not have a unique basis, but given any two different bases p1 ,. . . , ps and
g1 ,. . . , gm of I, the varieties V(p1 ,. . . , ps ) and V(g1 ,. . . , gm ) are equal. In other words, a variety only depends
on the ideal generated by its defining equations. Some bases of an ideal may be simpler than other bases.
Intuitively, if all the polynomials in a given basis of an ideal have a degree that is lower than the degree of
any other polynomial in the ideal with respect to a particular monomial ordering, then this basis is in some
sense the simplest basis. In particular, a Grobner basis of an ideal (for a given monomial ordering) has such a
property and can be thought of as the simplest or canonical basis. Given an ideal I and a monomial ordering,
denote the set of leading terms of elements of I as LT(I). The ideal generated by elements of LT(I) is denoted
LT(I). In general, the ideal generated by the leading terms of a particular ideal I is not the same as the
ideal generated by the leading terms of the polynomials in a basis for that particular ideal I. A Grobner basis
is a special basis for which this property holds, and it is formally defined as the set of polynomials g1 ,. . . ,
gm for which LT(I) = LT(g1 ),. . . , LT(gm ). When computing Grobner bases, the user specifies a monomial
ordering; different monomial orderings produce different Grobner bases. Given a monomial ordering, the two
most important properties of Grobner bases are:
(1) Every ideal I k[x1 ,. . . , xn ], other than the trivial ideal 0, has a Grobner basis. Furthermore, any Grobner
basis of an ideal I is a basis for I.
(2) Given an ideal I k[x1 ,. . . , xn ], other than the trivial ideal 0, a Grobner basis of I can be computed in a
finite number of algebraic operations.
The first algorithm for computation of Grobner bases, published in the 1960s, is attributed to B. Buchberger (6). Since then a number of improvements have been reported and the algorithm has been implemented in most commercial symbolic software packages. Buchbergers algorithm generalizes two well-known
algorithms: Gauss elimination for sets of multivariate linear algebraic equations, and Euclids algorithm for
computing the greatest common divisor of a set of univariate polynomials.
Solving Sets of Multivariate Polynomial Equations. Grobner bases facilitate solving a set of multivariate polynomial equations (3) in the same way as the Gauss elimination algorithm facilitates solving a
set of linear algebraic equations. Indeed, in a given lex ordering the Grobner basis has a triangular structure
reminiscent of the triangular structure in Gauss elimination. If the Grobner basis is given by {1}, then the
system of polynomial equations has no solution.
Equilibria for Polynomial Systems. Consider a polynomial system without inputs x(t) = f (x(t)). A state
x n is called an equilibrium of this system if x(0) = x implies that x(t) x t. Equilibria for polynomial
systems are therefore obtained as solutions of a set of polynomial equations of the form f (x) = 0 or f (x) = x for
continuous-time or discrete-time systems respectively. Grobner bases facilitate finding all equilibria.
Example: The equilibria of the polynomial system
which is hard to solve. The Grobner basis for the ideal p1 , p2 , p3 using the lexicographic ordering x1 x2 x3
is computed using Maple:
By construction the equations gi = 0, i = 1, 2, 3, 4, have the same solutions as Eq. (6). But obviously the
polynomials gi have a much simpler structure than the polynomials pi . Indeed, g4 depends only on x3 , and this
allows g4 = 0 to be solved numerically for x3 . The solutions can be substituted into g2 = 0 and g3 = 0 to obtain
polynomials in x2 only that can then be solved numerically. This process of back-substitution can be continued
until all solutions are found.
Observe that a basis for an ideal does not have a fixed number of basis polynomials, and indeed the simpler
Grobner basis contains more polynomials than the original basis. As may be observed, it even contains more
polynomials than there are variables.
Periodic Solutions for Polynomial Systems. Consider the system f (dn y/dtn , . . . , dy/dt, y, t) = 0, where
y can be thought of as the output of a closed-loop control system. Let f be a polynomial in all its variables.
To approximately compute a periodic solution y(t) of this system, the method of harmonic balancing can be
used. It can be shown that under mild conditions an approximate periodic solution implies the existence of
an exact periodic solution close to the approximate one (7). A truncated Fourier series can be considered
as a candidate approximate periodic solution y(t) = k = N N Ck e jkt , ck , where ck and c k are complex
conjugates. Postulating that y(t) is a solution leads to a set of polynomial equations, because f is polynomial.
Using the Grobner basis method with lex ordering, this set of equations can be solved for the unknowns c0 ,
c1 ,. . . , cN and to obtain y(t).
In practice the actual solution may not be so important, and it often suffices to know that a periodic
solution exists and to know its oscillation frequency . This can be achieved by finding the Grobner basis of
the polynomial equations with any lex ordering where is the lowest-ranking variable.
Example: Consider the van der Pol equation (8) y a (1 by2 ) y + y = 0, and postulate a periodic solution y(t)
= , where ck = ckr + cki j, k = 0, 1, 2, 3, and j = . Substituting y(t) into the van der Pol differential equation
and equating the coefficients on the left-hand and right-hand sides of the equation leads to
and c0 = 0, c2 = 0, and c1i = 0. The coefficients a and b are regarded as variables. To simplify the computations,
p1 is divided by c1 , and p2 by ac1 . A Grobner basis of the ideal p1 , p2 , p3 , p4 with the lex ordering c3 c3i
c3 b a contains seven polynomials of nontrivial complexity (42 pages of Maple output). It contains
one polynomial in and a only:
Obviously, for a = 0 the van der Pol oscillator is a simple harmonic oscillator with frequency = 1. In the
presence of nonlinear damping a = 0, the solution of g7 = 0, which continues = 1, can be expanded as
Grobner
Bases and Elimination. Grobner bases can be used to eliminate some variables of interest
from a set of polynomial equations, as illustrated by Example 1, where x1 and x2 are eliminated from g4 in Eq.
(7). The variables that are ranked higher in the lex ordering are eliminated first. This feature of Grobner bases
can be used to obtain an estimate of the domain of attraction of equilibria for polynomial systems.
Estimation of the Domain of Attraction. If the closed-loop control system is stable, it is of interest to
determine those initial conditions from which the solutions converge to a specified equilibrium. The collection
of all the initial conditions that converge to the same equilibrium is called the domain of attraction of that
particular equilibrium. Estimating domains of attractions is a hard problem. One way of obtaining a (conservative) estimate for the domain of attraction is to use Lyapunov functions. It is a standard result in Lyapunov
theory that if x = 0 is an equilibrium point for the continuous-time system d/dt x = f (x), if D n is a domain
containing x = 0, and if W : D is a continuously differentiable function such that W(0) = 0 and for all x D
{0} one has W(x) > 0 and W/x f (x) < 0, then x = 0 is asymptotically stable. Given such a Lyapunov function,
consider the sets = {x n : W/xf (x) < 0} and Bd = {x n : W(x) d}. If Bd for some d > 0, then
the set Bd is an estimate for the domain of attraction. For polynomial systems with a polynomial Lyapunov
function W, Grobner bases can be used to compute Bd systematically. Indeed, it is feasible to construct the
largest such set by finding d such that Bd is as large as possible and still inside . For polynomial systems
with polynomial Lyapunov functions, W(x) d and W/x f (x) are polynomials, and hence the boundaries of
the sets Bd and are varieties. At the points where the varieties V(W d) and V (W/xf (x)) touch each other,
the gradients of W and
variables is obtained:
computing a Grobner basis for the above system of equations, where the variable d has the least rank in the
lex ordering, a polynomial equation in d only is obtained. The least positive solution to this equation is the
best value of d for which Bd , and this yields in turn the best estimate for the domain of attraction that
could be obtained with the particular Lyapunov function W.
with the Lyapunov function W(x) = 3x2 1 + 4x1 x2 + 4x2 2 . The polynomials (8) are in this case
and using the lexicographic ordering x2 x1 d, the Grobner basis of the ideal p1 , p2 , p3 , p4 is computed.
It contains the polynomial g(d) = 4d4 147d3 + 786d2 + 2048d. The approximate solutions of g(d) = 0 are
1.9223, 0, 8.9657, 29.707. The smallest positive value of d for which there is a solution to the system of
equations pi = 0 is 8.9657. The corresponding estimate for the domain of attraction is therefore {x 2 : W(x)
8.9657}.
Equality of Ideals. A special type of a Grobner basis can be used to decide if two ideals are the same
or not. This is the so-called reduced Grobner basis. A reduced Grobner basis for an ideal I is a Grobner basis
G for I such that LC(p) = 1 for all p G, and for all p G, no monomial of p lies in LT(G {p}). The main
property of reduced Grobner bases is that given an arbitrary ideal I = 0 and a particular monomial ordering,
I has a unique reduced Grobner basis. Hence, two ideals J 1 and J 2 are the same if and only if their reduced
Grobner bases G1 and G2 are the same (G1 and G2 must be computed with the same monomial ordering).
Most commercial computer algebra systems, such as Maple and Mathematica, provide finite algorithms for
computing reduced Grobner bases.
Analysis of Discrete-Time Polynomial Systems. Fundamental control-theoretic concepts such as controllability or observability can be reduced to the problem of computing maximal control-invariant varieties.
This is well known for linear systems (computation of control-invariant subspaces), but it holds equally well
for discrete-time polynomial systems of the form x(t + 1) = f (x(t),u(t)). For such systems, a variety V is controlinvariant if f (V,u) V for all possible control inputs u. The computation of the maximal control-invariant
subset of a given variety V can be completed in finitely many operations. Consider the defining ideal of the
variety V; say J 1 = g1,1 ,. . ., g1,m1 , where gj,k [x]. If the variety corresponding to J 1 were control-invariant,
then g1,k f (x,u) 0 for all u and all k = 1,. . . , m1 . The polynomials g1,k f (x,u) can be viewed as polynomials
in u with coefficients in [x]. Denote the collection of all these coefficients as g2,k , k = 1, 2,. . . , m2 , and the
corresponding ideal as J 2 = g2,1 ,. . . , g2 ,m2 . Invariance would imply that J 1 = J 2 , and if this is not the case,
then obviously J 1 J 2 and the corresponding varieties satisfy V 1 V 2 . This process can be continued to
construct an ascending chain of ideals (or descending chain of varieties) J 1 J 2 J 3 . This chain must
terminate in finitely many steps (6). That is, there exists an integer N such that J N 1 J N = J N+1 =
. The variety V(J N ) is the maximal control-invariant subset of V(J 1 ). The check whether J k = J k+1 can be
completed via the computation of reduced Grobner bases for J k and J k+1 .
as said to be observable if for each pair of initial states = , there exists an integer N and an input sequence U N
= {u(0),. . . , u(N1)} such that the solutions starting at and produce different outputs after N steps, that is
h(x(N, , U N )) = h(x(N, , U N )). The polynomial h(x(N, , U N )) h(x(N, , U N )) can be regarded as a polynomial
in elements of U N with coefficients in [, ]. A chain of ideals J k is constructed using these coefficients, and
at each step the condition J k = J k+1 is tested. It can be shown that if J N for some N has a reduced Grobner
basis {1 1 ,. . . ,n n }, then J N+1 = J N and the system is observable. The above-discussed algorithm
for computing invariant sets streamlines these computations and allows a systematic determination of the
integer N.
Example: Consider the simple Wiener system (9)
The system consists of a linear dynamical system and a quadratic static output nonlinearity. The ideal J 1 =
1 2 1 2 and the output equation (12) are used to construct the following ideals:
Using the lex ordering 1 2 1 2 , the reduced Grobner basis for J 4 is G4 = {1 1 , 2 2 }, and therefore
the system (12) is observable with N = 4.
A Brief Overview of the Literature and Related Problems. Grobner bases are useful for a range of other
control problems, such as the inverse kinematic problem and motion planning in robotics (6), the computation
of the switching surfaces in the solution of time optimal control problems (10,11), identifiability, inputoutput
equivalence of different state-space realizations, normal forms and zero dynamics (12), analysis of hybrid
control systems, computation of limit cycles for discrete-time polynomial systems, observability of continuoustime polynomial systems, and forward accessibility of discrete-time polynomial systems.
In control design for linear systems, the linear functional observer problem, the model-matching problem,
the deterministic identification problem, and the disturbance decoupling problem play a central role. All these
problems can be seen as the characterization of a maximally control-invariant subspace of a given subspace
of the state space (2). In control theory this problem is known as the cover problem. The cover problem can be
solved via a computation of elementary divisors of matrix pencils, which in turn leads to multilinear equations
where the unknowns are the basis vectors of the invariant subspace. The Grobner basis method facilitates the
analytic solution of this problem. The elegance of this approach stems from the fact in one single algorithm all
possible (typically infinitely many) control-invariant subspaces can be explicitly characterized. The degrees of
freedom established in this process can then be further exploited to optimize other desirable system properties,
such as sensitivity.
We have focused on the use of Grobner bases for commutative rings over infinite fields, that is, finding
Grobner bases in the context of polynomials with rational coefficients. Two important related areas are Grobner
bases for commutative rings over finite fields and Grobner bases in the context of noncommutative rings. The
former is of interest in the context of discrete-event dynamic systems (13,14) and coding theory in communications. The latter is useful for control-theoretic problems involving polynomial matrices (15). In particular, in
15 a new extra term in the expansion of the classical state-feedback optimal control problem in the singular
perturbation form was obtained using noncommutative Grobner bases.
An important drawback of Buchbergers algorithm is that even with the best known versions of the
algorithm, it is easy to generate examples for which the computation of a Grobner basis requires inordinate
amounts of computer resources (time and/or memory). The main bottlenecks are that the total degrees of the
intermediate polynomials that the algorithm generates may be quite large and that the coefficients of the
Grobner basis may be complicated rational numbers. This may be the case even if the original ideal generators
are polynomials of small degrees with small integer coefficients. In general, the intermediate polynomials
observed in the computation of a Grobner basis can have total degrees of the order of 22d , where d is the
total degree of the ideal generators. Although this appears to be a rather negative result, it typifies the
worst-case scenario. It appears that the running time and the storage requirements seem to be much more
manageable on average. It is important to emphasize that different monomial orderings may produce very
different computational times, and some experimentation with the ordering may yield significant reductions
in computation time.
and
10
where denotes is ranked lower than. Any ranking is possible provided it satisfies two conditions:
for all variables u and y, all nonnegative integers and v, and all positive integers . The highest-ranking
variable or derivative of a variable in a differential polynomial is called the leader. The ranking of variables
gives a ranking of differential polynomials. They are simply ranked as their leaders. If two have the same
leader, they are considered as polynomials in their leader and the one of lower degree is ranked lower.
Let A, B be two differential polynomials, and let A have the leader v. Then B is said to be reduced with
respect to A if there is no derivative of v in B and if B has lower degree than A when both are regarded as
polynomials in v. A set
of differential polynomials is called autoreduced if all the Ai are pairwise reduced with respect to each other.
Normally autoreduced sets are ordered so that A1 ,. . ., Ap are in increasing rank. Autoreduced sets are ranked
as follows. Let A = A1 ,. . . , Ar and B = B1 ,. . . , Bs be two ordered autoreduced sets. A is ranked lower if either
there is an integer k, 0 k min(s, r), such that
A characteristic set for a given set of differential polynomials is an autoreduced subset such that no other
autoreduced subset is ranked lower.
The separant SA of a differential polynomial A is the partial derivative of A with respect to the leader,
while the initial I A is the coefficient of the highest power of the leader in A. If a differential polynomial f is not
reduced with respect to another differential polynomial g, then either f contains some derivative of the leader
ug of g or else f contains ug to a higher power. In the former case one can differentiate g a suitable number (say
) of times and perform a pseudodivision to remove that derivative, giving a relation
where S is the separant of g and R does not contain the highest derivative of ug that is present in f . If f contains
ug to a higher power, a pseudodivision of f by g can be performed to obtain
11
Observability. In control theory a system is called observable if it is possible to compute the state from
inputs and outputs. Using differential algebra, the observability question can be answered in a constructive
manner for continuous-time polynomial systems (1).
Example: Consider the system
which is in state-space form. Suppose an inputoutput description (i.e., a description directly relating u and y)
is called for. The original set of equations is equivalent to
From this construction it follows that p = 0 whenever the equations of (19) are satisfied. The last equation of
(19) can be replaced by p = 0 to get the system description
From (Eq. 20) it follows that every solution of Eq. (19) also solves Eq. (21). If, moreover, it is known that x2 =
0, then the converse is also true, and Eqs. (19) and (21) are equivalent. It is now possible to form
to conclude that, provided ux2 = 0, Eq. (19) is equivalent to the following system description:
12
The leftmost equation is an inputoutput relation, while the middle and right equations show how x1 and x2
can be computed from the input and output. This establishes the observability, and more.
Using the terminology of differential algebra, the state-space form is an autoreduced set under the ranking
Autoreduced sets thus give a generalization of several of the standard forms for dynamic systems.
as said to be globally identifiable if, given input and output measurements u(t) and y(t) over an interval t
[0,T], there is only one constant value of that satisfies Eq. (26). Global identifiability of polynomial systems
(1) can be checked in a constructive manner using the tools from differential algebra. Moreover, conditions on
u and y under which the system is not globally identifiable can also be obtained, as is illustrated below.
Example: Consider the system
where is a constant parameter. Can be identified by observing the output y? Using the notation p = y +
2 y + 2 y, compute the expression
It follows readily that is uniquely determined by y (and its derivatives), provided that along the output
trajectory
The Algorithms of Ritt and Seidenberg. The concepts of the previous section were introduced in the
1930s by the American mathematician Ritt, (17). He devised an algorithm by which it is possible to start with
an arbitrary number of differential polynomials, introducing a suitable ranking of the variables and performing
successive operations of the type described by Eqs. (14), (15) to arrive at an equivalent representation in the
form of an autoreduced set of equations. This set has the property that an f belonging to the original differential
polynomials can be reduced to zero using (Eqs. 14) and (15) for different gs belonging to the set. If certain
factorization conditions are met, then Ritt showed that the generated autoreduced set is a characteristic set,
not only for the explicitly generated polynomials, but also for an infinite set of polynomials constituting the
13
so-called differential ideal. Ritt used his algorithmic procedure to produce an extensive theory for polynomial
systems of differential equations. Later a more systematic algebraic treatment was given by Kolchin (18). The
theory was also extended to partial differential equations (18) and difference equations (19).
From the viewpoint of practical calculations, a major drawback of Ritts algorithm is the factorization it
requires, since this is a task of high computational complexity. Seidenberg (20) has proposed an algorithm for
deciding the solvability of systems of equations and inequations. The algorithm uses (Eqs. 14), (15) repeatedly
to eliminate variables and reduce the problem to the single-variable case. As in the simple example above,
equivalence of the original and final sets of equations requires the separant S and initial I of (Eqs. 14), (15) to
be nonzero. To handle this, Seidenbergs algorithm splits the problem into different subproblems, considering
the cases S = 0, S = 0 (or I = 0, I = 0) together with the original equations and inequations. By repeated
splitting a tree of subproblems is generated. Each subproblem is finally reduced to a single-variable problem
for which solvability can be decided. In recent years several variations of Ritts and Seidenbergs algorithms
have been implemented in symbolic manipulation languages like Maple or Mathematica.
Other Control Problems. Differential-algebra calculations can be used to compute the zero dynamics
of a system. This dynamics is obtained by restricting the output to be zero. By adding the equation y = 0 and
using Ritts algorithm with a suitable ranking, it is possible to obtain a description of the resulting dynamics. In
a similar manner a regulator description can be obtained. Suppose a differential equation describing the desired
response from reference signal to output is given. Using this equation together with the system dynamics as
the input to Ritts algorithm, it is possible to arrive at an equation containing the control signal together with
the reference and output signals. This equation can be interpreted as a description of the regulator, although
it can be difficult to implement due to its implicit nature in the general case.
Quantier Elimination
Quantifier elimination is a method for rewriting formulae that include quantifiers such as for all () and there
exists () in an equivalent form without the quantifiers and the quantified variables. According to a theorem
by Tarski (21), it is always possible to eliminate the quantified variables in formulae consisting of logical
combinations of multivariate polynomial equations and inequalities. Tarski provided a constructive proof in
the late 1940s, but the corresponding algorithm has such complexity that it is impractical for most problems.
In the mid 1970s Collins (22) presented a new method, the so-called cylindrical algebraic decomposition, which
exhibits much better complexity. Since then the algorithmic development has made significant progress (23,24,
25). Nevertheless, cylindrical algebraic decomposition and quantifier elimination are known to be inherently
complex (26). It is therefore of importance to identify classes of problems for which the computational complexity
is much lower, and for which specialized algorithms can be developed (27,28). Implementations of cylindrical
algebraic decomposition and quantifier elimination are available in for example Mathematica (29). Depending
on the specific nature of the problem posed, different algorithms are used in order to minimize the computational
burden. The algorithmic developments are ongoing.
Introductions to cylindrical algebraic decomposition and quantifier elimination can be found in Ref. 30.
An extensive bibliography covering early papers can be found in Ref. 31, and a survey of the algorithmic
development of cylindrical algebraic decomposition is given in Ref. 32.
An early application of quantifier elimination techniques in control theory was made by Anderson et al.
(33). This contribution predates the introduction of efficient computational tools, and so it was of theoretical
interest only. With the availability of the experimental software QEPCAD (23), the number of papers related to
control increased. Now, as these methods are implemented in widely available symbolic computation software
packages like Mathematica, the control practitioner can explore their potential.
Systems of Real Algebraic Equations and Inequalities. In the context of polynomial equations
with real coefficients, inequalities arise naturally, for example to express when a quadratic polynomial has
14
real solutions. The need of using not only equations but also inequalities in describing computational problems
triggered the development of methods for doing symbolic computations with so-called semialgebraic sets. A
semialgebraic set is a generalization of an algebraic set or variety, as it is the solution set of a system of real
algebraic equations and inequalities.
An algebraic expression in variables {x1 ,. . . , xn } is an expression constructed with {x1 ,. . . , xn } and
rational numbers, using addition, multiplication, rational powers, and algebraic numbers and functions. A
system of real algebraic equations and inequalities in variables {x1 ,. . . , xn } is a logical combination of equations
and inequalities with both sides being algebraic expressions in {x1 ,. . . , xn }. The notation (and), (or),
(implies) for Boolean operators is used.
The following is an example of a system of real algebraic equations and inequalities in the variables a, b,
x, y:
Alternatively, a semialgebraic set can be characterized as a set obtained by finitely many unions, intersections,
and complementations of sets of the form {x | f (x) 0}. Here f is a multivariate polynomial with rational
coefficients. Semialgebraic sets are thus closed under projection, union, and intersection. As Grobner bases
provide a way to replace a system of multivariate polynomials by a simpler equivalent set, there are systematic
ways to simplify a system of real algebraic equations and inequalities. It can be shown that the set of solutions of
any system of real algebraic equations and inequalities in variables {x1 ,. . . , xn } can be written as a disjunction
of a finite number of cylindrical parts of the form
In the above expression, stands for one of <, , and =; and f i and gi are either , , or algebraic expressions
in the variables {x1 ,. . . , xi i } that are real-valued for all tuples of real numbers {a1 ,. . . , ai i } satisfying
The method of rewriting a real algebraic system as a disjunction of the above form is called cylindrical algebraic
decomposition; see Refs. 22,30,32. Observe the triangular nature of the resulting system.
Example: An example of a script in Mathematica for computing a cylindrical algebraic decomposition of an
ellipsoid in 3 is
15
Here the functions f i and gi can be expressed in terms of simple algebraic operations and radicals. More
typically the full generality offered by algebraic functions is required:
In Mathematica algebraic functions are represented by special Root objects. In the above case the placeholder
variable #1 is a function of x and y. The last argument enumerates the branches of the algebraic function,
which in this case are the first two roots according to Mathematicas enumeration scheme of the 12 possible
complex roots.
A quantified system of real algebraic equations and inequalities in variables {x1 ,. . . , xn } is an expression of
the form
where the Qi s are the quantifiers or , and S is a system of real algebraic equations and inequalities in
{x1 ,. . . , xn ; y1 ,. . . , ym }. According to Tarskis theorem the solution set of quantified systems of real algebraic
equations and inequalities is a semialgebraic set, which means that QS always can be rewritten as an equivalent
expression without any quantifiers and quantified variables. The process of eliminating quantified variables
from such systems is called quantifier elimination. The yi s in QS are called bounded variables, and the xi s free
variables. Problems without free variables are called decision problems.
Example: The proof of the well-known inequality between the arithmetic and the geometric mean of two real
positive numbers can be formulated as a decision problem:
where Q is either or , var is a single variable or list of variables, and cond gives additional conditions
on the variables such as their domain and/or semialgebraic constraints. The command to perform quantifier
elimination in Mathematica is called Resolve.
16
Projection of semialgebraic sets is equivalent to quantifier elimination on existential quantifiers. Consider for
example the projection of the ellipsoid in Example 7 onto the xy plane:
Set inclusion for semialgebraic sets is another useful property that can be formulated as a decision problem: A
B is the decision problem xA B:
Sets of Equilibria. As illustrated before, Grobner bases are useful to determine equilibria for polynomial systems. In polynomial control systems it is of importance to describe the steady-state solutions: the set of
reachable equilibria, for admissible input values. Characterizing the set of reachable equilibria in state space
is a quantifier elimination problem.
Example: Add a control input to the last equation in Example 1. Let this input be constrained to the interval
1 u 1. Compute the projection of the set of corresponding equilibria onto the x1 x2 plane.
Stability. To determine the local stability of equilibria of dynamic systems the stability tests of Routh
and Hurwitz (34) may be used. These stability tests give explicit conditions on the coefficients of the characteristic equation of the Jacobian of the dynamical system at equilibrium in terms of a number of polynomial
inequalities. Hence the RouthHurwitz test can be handled using cylindrical algebraic decomposition and
quantifier elimination. A less well-known, but equivalent stability test is the LienardChipart criterion (34),
which has the advantage over the Routh criterion that it involves polynomial inequalities of lower degree.
Given a polynomial in [s] a0 sn + a1 sn 1 + + an 1 s + an , ai > 0, the LienardChipart criterion states
that all its zeros have strictly negative real parts if and only if the following inequalities holds:
17
where Di , i = 1,. . . , n, are the so-called Hurwitz determinants of order i, which are defined as follows:
For zero input u = 0 the zero state is an equilibrium. Linearizing the system around (x,u) = (0,0) and computing
the characteristic polynomial of the resulting state transition matrix gives
Compute the solution set corresponding to the LineardChipart stability inequalities for this polynomial:
Example: Consider the problem of stabilizing in a unit-feedback scheme an unstable system G, using a lead
compensator F that is parametrized by the parameters B and b, where
18
or
depending on the choice of the order for b and B. Using quantifier elimination, questions like For which values
of b can it be guaranteed that for all values of 5 < B < 10 the closed loop system is stable? can be answered:
Nonlinear Tracking. Tracking a particular output is an important control objective. There are many
control design methods that deal with this problem, mainly based on optimization techniques. However, most
(analytic) methods do not take the ever present constraints on control signals or states into account. As a
consequence the resulting control system must be validated, often through extensive simulations. For a large
class of nonlinear tracking problems, quantifier elimination can be used to decide if there exists a solution to a
given tracking problem subject to the constraints of interest.
Consider the continuous-time polynomial control system
where x n , u m and each component of f is a polynomial in all its variables. Furthermore, the input value
u(t) is required to belong to some semialgebraic set U:
called the admissible control set. Let be a rationally parametrized curve in the state space, i.e.,
where g :
is a rational function and the orientation of is defined by increasing values of s.
For the state of the system (27) to follow the curve , there has to be an admissible control u such that
the vector field f (xp , u) is tangent to at each xp on . Since the curve is parametrized in terms of s, this
19
where > 0 ensures that the state follows in only one direction. The tracking problem is thus equivalent to
deciding if the following formula is true or false:
The above problem becomes a quantifier elimination problem if f or g include some free (design)
parameters.
Example: Consider the polynomial control system
and the curve = {x 2 |x1 = s x2 = s3 + 1.5s2 0 s 1}. Is it possible to steer the system
state along using control signals constrained to 1 u 1? Using Mathematica, this can be resolved as
follows:
20
Consider now a = 1, and compute the limits on u such that the above tracking problem has a solution:
For the tracking problem to have a solution it suffices that 0.98077 u 0.98077.
Multiobjective Feedback Design. Many objectives in linear system design are formulated in terms
of frequency-domain inequalities. Let the plant be represented by G(s, p) and the controller by F(s, q). Both G
and F are assumed to be scalar-valued rational functions of the Laplace variables s, real plant parameter p,
and real controller parameter q. It is now possible to write many robust design problems in a form suitable for
quantifier elimination.
Stability. The unit-negative-feedback closed-loop system consisting of plant and controller is asymptotically stable if and only if all zeros of the rational function 1 + G(s, p)F(s, q) have strictly negative real part.
This can be converted into polynomial inequalities in p and q using the LineardChipart criterion.
Tracking Error. The tracking error at the output of the unit-negative-feedback loop of plant and controller
is governed by the so-called sensitivity transfer function
Acceptable levels of the tracking error can be specified by the inequality |S(iw)| < T , 0 w w1 , which can be
rewritten as a real algebraic inequality in the variables w, p, q, and T . Through quantifier elimination it can
be verified that for the class of systems p P a single stabilizing controller q exists that meets the sensitivity
objective.
In essence any linear control design question formulated in frequency domain leads naturally to a quantifier elimination or cylindrical algebraic decomposition problem. More importantly, symbolic manipulation
software is up to the task of systematically approaching these control design questions in practical and moderately complex situations.
Related Problems and Other Control Applications. There is a vast array of problems in control
that can be posed as quantifier elimination problems or systems of multivariate polynomial inequalities,
whose solutions can be described by cylindrical algebraic decomposition. For example, quantifier elimination
has been used to investigate stability, stabilizability, and controllability of discrete-time polynomial systems
(35,36), stability and stabilizability of switched polynomial systems and unstable zero dynamics via switching
(30,37), frequency-domain design (38), and multiobjective robust control design (39). Quantifier elimination
has also been used for design of nonlinear control systems for nonlinear aircraft dynamics (40) and for robust
nonlinear feedback design (41).
21
The generality of the quantifier elimination method comes at the cost of very high computational complexity. Quantifier elimination software tools are not a panacea allowing one to approach every control problem
by brute force. The challenge in successfully applying quantifier elimination software tools lies in formulating
the problems in such a way that their inherent structure is maximally exploited. Most importantly, instead of
applying quantifier elimination to a complete problem, there are often simplifications that can be carried out
so that only a small problem (with far fewer variables) has to be handled by quantifier elimination. Control
designers should start focusing on providing the right heuristics and tools, built on their experience, in order to
make general-purpose quantifier elimination tools better attuned to control problems. In parallel, algorithm designers now focus on developing specialized methods as components of a general quantifier elimination package
in order to exploit structure and to handle complexity. For instance, in Mathematica (25) there are a modified
simplex linear optimization algorithm for linear systems of equations and inequalities with inexact or rational
coefficients, linear quantifier elimination for equations and inequalities with exact coefficients according to 27,
preprocessing by linear equation solvers, Grobner bases, special care and simplification for variables appearing
at most linearly in the systems, a simplified cylindrical algebraic decomposition algorithm for generic solutions
(24), and quantifier elimination by partial cylindrical algebraic decomposition (23) for the general case.
Conclusion. The possibilities offered through symbolic (and numeric) software in order to address
control problems are as varied and exciting as the control problems themselves. This article just presents the
tip of an iceberg. Dealing with complexity in control design is the main issue. For control algorithm developers
there is now more than ever a compelling need to focus on reliable software algorithms that allow nonspecialists
to approach a control problem with confidence. Software tools that provide information to the user about the
likelihood of success and the difficulties in a particular solution are called for to make this a reality.
BIBLIOGRAPHY
1. F. Cucker L. Blum S. Smale M. Shub Complexity and Real Computation, Berlin: Springer-Verlag, 1998.
2. N. Munro (ed.) Symbolic Methods in Control System Analysis and Design, Institution of Electrical Engineers, 1999.
3. L. A. Rubel M. F. Singer A differentially algebraic elimination theorem with applications to analog computability in
the calculus of variations, Proc. Amer. Math. Soc. 94: 653658, 1985.
4. P. Lindskog Methods, algorithms and tools for system identification based on prior knowledge, PhD thesis 436, Department of Electrical Engineering, Linkoping University, Sweden, 1996.
5. D. Nesic Dead-beat control for polynomial systems, PhD thesis, RSISE, Australian National University, Canberra,
Australia, 1996.
6. J. Little D. Cox D. OShea Ideals, Varieties and Algorithms, Berlin: Springer-Verlag, 1992.
7. A. Mees Dynamics of Feedback Systems, Chichester: Wiley, 1981.
8. K. Forsman Constructive commutative algebra in nonlinear control, PhD thesis, Department of Electrical Engineering,
Linkoping University, Linkoping, Sweden, 1997.
9. D. Nesic A note on observability tests for general polynomial and simple WienerHammerstein systems, Syst. Control
Lett., 35: 219227, 1998.
10. S. T. Glad An algebraic approach to bangbang control, European Control Conference, ECC 95, Rome, 1995, Vol. 4, pp.
28922895.
11. T. Georgiou U. Walther A. Tannenbaum Computational algebraic geometry and switching surfaces in optimal control,
Proc. Conf. Decision Control, Phoenix, AZ, 1999, pp. 47244729.
12. H. Fortell Algebraic approaches to normal forms and zero dynamics, PhD thesis, Department of Electrical Engineering,
Linkoping University, Linkoping, Sweden, 1995.
13. R. Germundsson Symbolic systems, PhD thesis, Department of Electrical Engineering, Linkoping University,
Linkoping, Sweden, 1995.
14. J. Gunnarsson Symbolic methods and tools for discrete event dynamic systems, PhD thesis, Department of Electrical
Engineering, Linkoping University, Linkoping, Sweden, 1997.
15. F. D. Kronewitter III Non-commutative computer algebra, PhD thesis, University of California, San Diego, 2000.
22
16. M. Fliess S. T. Glad An algebraic approach to linear and nonlinear control, in H. L. Trentelman and J. C. Willems (eds.),
Essays on Control: Perspectives in the Theory and Its Applications, Boston: Birkhauser,
1993.
17. J. F. Ritt Differential Algebra, New York: American Mathematical Society, 1950.
18. E. R. Kolchin Differential Algebra and Algebraic Groups, New York: Academic Press, 1973.
19. R. M. Cohn Difference Algebra, Huntington, NY: R. E. Krieger, 1979.
20. A. Seidenberg An elimination theory for differential algebra, University of California Publications in Mathematics,
New Series, 1956, pp. 3166.
21. A. Tarski A Decision Method for Elementary Algebra and Geometry, 2nd ed. University of California Press, 1948.
22. G. E. Collins Quantifier elimination for real closed fields by cylindrical algebraic decomposition, Second GI Conf.
Automata Theory and Formal Languages, Kaiserslauten, Lecture Notes in Comput. Sci., 33 Berlin: Springer-Verlag,
1975, pp. 134183.
23. G. E. Collins H. Hong Partial cylindrical algebraic decomposition for quantifier elimination, J. Symbolic Comput. 12:
299328, 1991.
24. S. McCallum An improved projection for cylindrical algebraic decomposition, in B. F. Caviness and J. R. Johnson
(eds.), Quantifier Elimination and Cylindrical Algebraic Decomposition, Monographs in Symbolic Computation, Berlin:
Springer-Verlag, 1998, pp. 242268.
25. A. Strzebonski Solving algebraic inequalities with Mathematica version 4, Mathematica J.7(4): 525541, 2000.
26. S. Basu R. Pollack M.-F. Roy On the combinatorial and algebraic complexity of quantifier elimination, Assoc. Comput.
Mach., 43(6): 10021045, 1996.
27. R. Loos V. Weispfenning Applying linear quantifier elimination, Comput. J. 5(36): 450461, 1993.
28. A. Strzebonski An algorithm for systems of strong polynomial inequalities, Mathematica J. 4(4): 7477, 1994.
29. S. Wolfram The Mathematica Book, 4th ed., Champaign, IL: Wolfram Media, Cambridge University Press, 1998.
30. M. Jirstrand Constructive methods for inequality constraints in control, PhD thesis, Department of Electrical Engineering, Linkoping University, Linkoping, Sweden, May 1998.
31. D. S. Arnon A bibliography of quantifier elimination for real closed fields, J. Symbolic Comput. 5(12): 267274, 1988.
32. G. E. Collins Quantifier elimination by cylindrical algebraic decompositiontwenty years of progress, in B. F. Caviness
and J. R. Johnson (eds.), Quantifier Elimination and Cylindrical Algebraic Decomposition, Berlin: Springer-Verlag,
1998.
33. B. Anderson N. Bose E. Jury Output feedback stabilization and related problemssolution via decision methods, IEEE
Trans. Autom. Control, AC-20: 5365, 1975.
34. F. R. Gantmacher Matrix Theory, Vol. II, Chelsea: New York, 1960.
35. D. Nesic I. M. Y. Mareels Stabilizability and stability for implicit and explicit polynomial systems: a symbolic computation approach, Eur. J. Control 5: 3243, 1999.
36. D. Nesic I. M. Y. Mareels Dead-beat controllability of polynomial systems: symbolic computation approaches. IEEE
Trans. Autom. Control 43: 162175, 1998.
37. D. Nesic M. Jirstrand Stabilization of switched polynomial systems, in IMACS Conf. on Applications of Computer
Algebra (ACA98), Prague, 1998.
38. P. Dorato W. Yang C. T. Abdallah Quantifier elimination approach to frequency domain design, in S. Tarbouriech and G.
Garcia (eds.), Control of Uncertain Systems with Bounded Inputs, Lecture Notes in Control and Information Sciences,
227, Berlin: Springer-Verlag, 1997, pp. 165172.
39. P. Dorato W. Yang C. Abdallah Robust multi-objective feedback design by quantifier elimination, J. Symbolic Comput.
24(2): 153159, 1997.
40. M. Jirstrand Nonlinear control system design by quantifier elimination, J. Symbolic Comput. 24(2): 137152, 1997.
41. P. Dorato D. Famularo C. T. Abdallah W. Yang Robust nonlinear feedback design via quantifier elimination theory, Int.
J. Robust Nonlinear Control, 9: 817822, 1999.
D. NES IC
I. M. Y. MAREELS
The University of Melbourne
S. T. GLAD
Linkoping University
M. JIRSTRAND
MathCore AB
464
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
The singular value decomposition (SVD) goes back to the beginning of this century. In a paper of Beltrami (1) it was
shown for the first time that any n n matrix A can be diagonalized via orthogonal row and column transformations. For
the more general case of an n n complex matrix A, the
result says that there exist unitary matrices U and V of dimension n n and a real diagonal matrix diag1, . . .,
n such that
A = UV
where 1 n 0. If A is real then U and V are also
real. This implies indeed that U and V diagonalize A since
A A = Im , B B = Im
Moreover,
A B = = diag{1 , . . ., m }
and from this diagonal form one can define the canonical
angles between the spaces A and B as cos i i, i 1,
. . ., m.
The second important property is that the singular value
decomposition yields a direct construction of best lowerrank approximation to a given matrix A. Let us indeed rewrite the SVD in its dyadic form:
U AV =
A=
The decomposition is in nature close to an eigenvalue decomposition, which was well known at the time. But this new
decomposition is also very different since singular values are
always positive real, whereas eigenvalues are in general complex. Also, the transformations are unitary in this decomposition, whereas in the eigenvalue decomposition they are just
nonsingular, and hence can be quite badly conditioned (see
the next section for a more detailed discussion). The use of
this new decomposition was not apparent from the very beginning, but nowadays it has become an invaluable tool in several application areas such as statistics, signal processing,
and control theory.
The first important property that was observed is the perturbation result for the singular values of a matrix A. If A
is a small perturbation of the matrix A, then its singular values i are perturbed by an amount that can be bounded by
the norm of the perturbation A. The fact that the sensitivity of singular values to perturbations is rather low makes
them a good candidate for measuring certain variations in an
observed phenomenon or in a model for it, and this is also its
principal use in engineering applications. More formally, one
can show that several matrix norms can actually be expressed
in terms of its singular values. The most important ones are
the 2-norm and the Frobenius norm:
Ax2
.
A2 = max
= 1 (A),
x = 0 x2
.
AF =
|ai, j |2 =
i, j
r
i2
n
i ui vi
i=1
r
i ui vi
i=1
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
special properties of the original mapping since now it is decoupled in a set of scalar equations. We will see that in the
context of dynamical systems this coordinate system plays a
fundamental role in what are called balanced realizations
(see the section entitled Balanced Realization).
(1)
where
0
Rmn
0
r
=
0
(2)
NUMERICAL BACKGROUND
The importance of the SVD is strongly tied to numerical aspects. For this reason, we first give a very brief discussion of
numerical stability and conditioning, which play a very important role in the study of numerical algorithms. For more
details we refer to standard textbooks such as Refs. 3 and 4.
We also choose the example of the singular value decomposition to introduce the relevant concepts.
Let A be an arbitrary m n matrix. Then it is well known
that there always exist unitary matrices U and V such that
U A =
R
, A V = C0
0
A=
r
i ui vi
(3)
i=1
(4)
A A = V T V
where T and T are clearly diagonal. Hence the left singular vectors are the eigenvectors of AA*, the right singular vectors are the eigenvectors of A*A, and the nonzero singular
values are the square roots of the nonzero eigenvalues of both
AA* and A*A. Deriving Theorem 1 from these connections is
in fact quite simple, but we refer to Ref. 2 for the details.
(5)
Ir
Im U AV = Im
,
0
465
Ker U AV = Im
Inr
Ir
Im A = Im U
= Im U1 ,
0
Ker A = Im V
Inr
= Im V2
(6)
Hence, the computed decomposition does not correspond exactly to the given matrix A but rather to a perturbed version
A A. When using the SVD algorithm available in the literature (5,6), this perturbation A can be bounded by
A cA A = cA A
(7)
where is the machine accuracy and cA is some known polynomial expression in m and n (3). Very often, this is a rough
upper bound and one prefers to replace cA by some statistical
estimate cA, usually close to 1. The error A induced by this
algorithmcalled the backward error because it is interpreted as an error on the datathus has roughly the same norm
as the input error in generated when reading in the data A
in the computer. When such a bound exists for the perturbation A induced by a numerical algorithm, it is called backward stable. We can make this definition more rigorous by
considering a function X f(A) with data A and solution X.
466
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
.
[ f (A)] = lim sup
0
AA=
X X
,
X = f (A),
X = f (A) (8)
Notice that we have not specified what norms are used in this
definition, but in principle one can use different norms in the
data and solution spaces (7). From this definition it is clear
that the condition number [f(A)] is some sort of derivative
of the function X f(A) that we want to compute. When
[f(A)] is infinite, the problem of determining X f(A) from
A is ill posed (as opposed to well posed). When [f(A)] is finite
and relatively large (or relatively small, the problem is said
to be ill conditioned (or well conditioned). Further details can
be found in Ref. 7.
It is important to note that backward stability is a property of an algorithm, while conditioning is associated with a
problem and the specific data for that problem. The errors in
the result depend on both the stability of the algorithm used
and the conditioning of the problem solved. A good algorithm
should therefore be backward stable since the size of the errors in the result is then mainly due to the condition of the
problem and not due to the algorithm. An unstable algorithm,
on the other hand, may yield a large error even when the
problem is well conditioned.
We point out that if f(A) has a Taylor expansion around
A, then we can write
f (A) = f (A) + x f (A)(X X ) + O(X X 2 )
(9)
(10)
This is derived from the variational properties of singular values (2,3) and leads to the following theorem.
Theorem 2. Let A be an arbitrary m n complex matrix
and A an arbitrary perturbation of the same dimensions.
Then the corresponding SVD of A and A A satisfy the
following strict bounds:
2 A2 ,
F AF
(11)
. 0
A=
0
0
0
a
;
0
0
1
0
.
A = A + A =
0
0
. 0
A =
0
0
0
c
s
0
0
s
c
0
0
0
0
1
0
0
0 0
0
1
0
a
0
0
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
(Im A, Im A)
1
= = 21
In the subsections that follow, we survey a number of problems from systems and control theory that rely heavily on
the singular value decomposition. We shall only discuss the
numerical aspects here; for the system theoretical background, we refer the reader to the systems and control literature.
Impulse Response Realization
Let H(z) be an p m transfer function of a discrete-time
causal system, and let its impulse response be given by
H(z) =
Hl
..
.
..
.
..
..
.
H2
=
.
.
.
Hk
H2
H k,l
H1
..
yk = Cxk + Duk
(15)
when the impulse response sequence Hi is given. In the scalar case, this problem is related to the Pade approximation
problem, for which fast methods exist (see Refs. 16 and 17 for
a survey). In Ref. 16, it is shown that the Pade approach is in
fact unstable, and it is better to consider a more general technique based on matrix decompositions of the Hankel matrix:
(16)
Hk+l1
Hi = CAi1 B,
i = 1, . . ., k + l 1
where
.
C l = [B AB
Al1 B],
(17)
This implies that Hk,l has at most rank n and a simple argument proves that Hk,l will have exactly rank n. Since determining the order of the system requires a rank determination, it is natural to use here the SVD Eq. (4):
H k,l = U1 nV1T
(13)
. CA
Ok =
...
CAk1
(12)
Here, x(t) is an n-vector of states, u(t) is an m-vector of controls or inputs, and y(t) is an p-vector of outputs. The standard discrete-time analog of Eq. (12) takes the form
xk+1 = Axk + Buk
(14)
H k,l = C l O k
Hi zi
i=0
x(t)
= Ax(t) + Bu(t)
..
467
R = n1/2V1T ,
H k,l = L R
(18)
468
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
L1
C1
=
,
L=
L2
C2
R = [B1
R2 ] = [R1
B2 ]
(19)
B = B1
..
..
..
.
zl
..
.
..
.
z 2
=
.
.
.
zk
z2
..
H 1:k,l
z1
Tx3
y2
Txl
yl1
TAT 1 TB Tx1
=
D
u1
CT 1
Tx2
u2
Txl1
ul1
(25)
x2
xl ]
(26)
zk+l1
xk+2
xk+l ]
(27)
where
X Tk+1,l ] = Im [H
H T1:k,l ] Im [H
H Tk+1:2k,l ]
Im [X
ui
zi =
yi
(20)
We start by noting that the problem would be much simpler if the sequence of states xk would be known as well. From
Eq. (13), rewritten as
A
xk+1
=
yk
C
xk
uk
(21)
x3
y2
xl
yl1
A
=
C
B
D
x1
u1
x2
u2
xl1
ul1
A
C
B
D
H 1:k,l
H k+1:2k,l
=
I
0
A
11
A21
Q
A12
03
01
02 V T
A33
(29)
(22)
where
E=
(28)
x2
y1
or
Tx2
y1
AR1 = R2
So the problem is solved as soon as the states xi are determined. But those depend on the choice of coordinates chosen
for the state-space model. Replace indeed xi by xi Txi; then
Eq. (22) becomes the related equation
x2 x3
A B
x1 x2 xl1
xl
=
(24)
y1 y2 yl1
C D u1 u2 ul1
(23)
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
T
AT11
A21
Im AT12 Im 0
0
0
In
= Im 0
AT33
0
In
H T1:k,l ) Im(H
H Tk+1;2k,l ) = V Im 0
Im(H
0
Balanced Realization
In the preceding section we pointed out that the realization
problem from an impulse response or from input/output data
is only defined up to a state-space transformation T, which in
principle can be chosen arbitrarily. Is there a particular coordinate system that should be chosen for some reason, and if
so, how can we construct it?
We develop here the concept of balanced realization that is
based on the singular value decomposition and has several
appealing properties. For this we first need to define the controllability Gramian Gc and observability Gramian Go of a system. For the continuous-time system equation (12) these are
defined as follows:
.
Go (T ) =
(Ce At )T (Ce At ) dt
0
(30)
K 1
.
Go (K) =
(CAi )T (CAi )
(31)
k=0
(32)
.
Gc (T ) =
469
Go = LT2 L2
(33)
where L1 and L2 are both lower triangular. One then computes the singular value decomposition of the upper triangular matrix LT1LT2:
LT2 LT1 = UV T
(34)
(35)
and
T
T
1/2
=
G c = TGc T T = 1/2U T L1
1 (L1 L1 )L1 U
T
1
1/2
=
G o = T T Go T 1 = 1/2V T LT
2 (L2 L2 )L2 V
.
E =
(36)
470
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
A
C
B
D
A 11
.
=
A21
C
1
A 12
A
22
C 2
B 1
B 2
(37)
11, B
1, D will be a good approxima 1, C
and the subsystem A
tion of the original system in the sense that it keeps the
states that were responsible for the largest part of the energy
transfer (i.e., the largest singular values). A more formal approximation measure is the so-called H norm, which can be
bounded as a function of the balanced singular values if the
balancing was based on infinite-horizon Gramians. This socalled balanced truncation problem is also related to that of
optimal approximation in the Hankel norm (26,27) and has
led to a renewed interest in the partial realization problem
and related topics (28).
Finally, we point out here that the realization algorithm
described in the section entitled Impulse Response Realization for discrete-time systems in fact constructs immediately
a balanced realization. Let the matrices Cl and Ok be defined
as in Eq. (17). Then
C l C Tl =
l1
.
(Ai B)(Ai B)T = Gc (l),
i=0
O Tk O k =
k1
(38)
.
(CAi )T (CAi ) = Go (k)
i=0
1/2
n
V1T ,
O k = U1
1/2
n
(39)
C lC Tl =
O Tk O k =
1/2
T
n V1 V1
1/2 T
n U1 U1
1/2
n =
1/2
=
n
n,
(40)
rank O n = n
A1,k+1
X1 A1,1 A1,2 A1,k
..
..
0
X2
A2,2
.
.
(42)
. ..
.
.
.
.
.
..
..
..
..
..
= .
..
.
.
.
Xk Ak,k
Ak,k+1
Ak+1,k+1
C n = [B A B . . . A n1 B]
X1:1
..
0
X1:2
.
.
.
.
..
..
= ..
0
..
..
.
.
X1:k
0
0
..
.
..
.
..
.
..
.
(43)
(41)
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
V T AV
CV
Ao
= 0
0
Ao
(44)
Co
A
C
B
D
A11
0
0
0
A12
A22
0
C2
A13
A23
A33
C3
B1
B2
0
D
(45)
and where the subsystem A22, B2, C2, D is minimal, that is,
both observable and controllable (reachable). Moreover, the
, B
, D and A22, B2, C2, D are equal.
, C
transfer functions of A
This form is closely related to the Kalman decomposition and
for its construction we refer to Ref. 31.
In addition to the controllable and unobservable subspaces
of a system, there are other spaces that play a fundamental
role in the control of systems modeled as Eqs. (12) and (13).
Two other fundamental objects in the so-called geometric system theory (32) are the supremal (A, B)-invariant and controllability subspaces contained in a given subspace. As shown in
Refs. 31 and 33 they can also be computed via a matrix recurrence based on a sequence of SVDs constructing an orthogonal basis for the relevant spaces. The role of the SVD in these
staircase algorithms is not only the reliable rank determination of the subsequent steps, but at the same time the singular values allow one to assess the sensitivity of the computed
bases (31).
Robustness in Systems and Control
In the last decade, there has been a significant growth in the
theory and techniques of robust control. These developments
mainly center around two concepts: H (34) and the structured singular value (35). They both provide a framework for
synthesizing robust controllers for linear systems, in the
sense that they achieve a desired system performance in the
presence of a significant amount of uncertainty in the system.
In this section, we first focus on H techniques. The H
norm of a stable rational transfer matrix H(s) (continuoustime) or H(z) (discrete-time) is defined as
.
H(s) = sup max[H( j)],
.
H(z) = sup max [H(e j )]
[0,2 ]
(46)
where max[ ] denotes the largest singular value of a (complex) matrix. We explain how this quantity comes about by
starting from a basic robustness problem. Consider the homo-
471
geneous systems
x = Ax,
xk+1 = Axk
(47)
which are assumed to be stable against Cs, that is, the eigenvalues of the matrix A are in a region Cs of the complex plane,
which is the open left-half plane for a continuous-time system
and the open unit disk for a discrete-time system, respectively.
The complex stability radius measures the robustness of
system stability for complex perturbations. This radius is defined as the norm of the smallest complex perturbation such
that the perturbed matrix A BC becomes unstable (where
B nm, C pn, and hence mp). For A BC to be
unstable, it must have at least one eigenvalue in the complement of Cs. It is important to note that although Cs is convex,
the set of all Cs-stable matrices St M: (M) Cs, where
(M) denotes the spectrum of M, is nonconvex, as well as its
complement Unst of Cs-unstable matrices. The stability radius r therefore measures the distance of a stable matrix A
to the nonconvex set Unst.
By continuity of the spectrum of a matrix versus perturbations on its entries, the stability radius is clearly equal to the
distance from a stable matrix A to an optimal matrix A
BC lying on Unst. Indeed, when a matrix A BC passes
the boundary Unst, at least one of its eigenvalue must also
cross Cs. The boundary Unst in the matrix space describes
matrices with at least one eigenvalue in Cs. Therefore, the
stability radius can be written as
rC (A,B,C) = inf{2 : (A + BC) Cs = }
Consider a parametrization of the boundary Cs by a real
variable , such as Cs j, or Cs ej, [0,
2]. The stability radius can then be rewritten as
(48)
the second equality resulting from the stability of the initial
matrix A (I A is invertible for Cs) and from the fact
that det(I XY) 0 det(I YX) 0. The following classical result allows us to simplify this expression considerably
and is based on the singular value decomposition M
UV*.
Theorem 5. One has the inequality
1
inf {2 : det(I M) = 0} max
[M]
C nn
C s
1
472
SOFTWARE FOR CONTROL SYSTEM ANALYSIS AND DESIGN, SINGULAR VALUE DECOMPOSITION
C s
(M) = inf 2
(0,1]
Re M
1 Im M
Im M
Re M
!
(50)
This becomes now an optimization problem in two real parameters and , but in Ref. 39 it is shown that the function
Eq. (50) is unimodal in . In Ref. 40 an efficient algorithm is
then derived to compute this real stability radius based on
ideas of Ref. 36.
The computation of the real stability radius can be considered a special case of the more general structured stability
radius. Structured singular values (35) have been introduced
to provide a perturbation bound for structured uncertainties
in control system analysis and design. Therefore, the structured singular value approach can be viewed as a complement
for the H approach. In a linear system with multiple independent norm-bounded perturbations, it is always possible by
rearranging the system to isolate the perturbations as a single large block-diagonal perturbation . Then, denoting the
transfer function from the collective outputs of the perturbation to their inputs by M(s), the stability problem reduces to
ensuring that det(I M) 0 at all frequencies and for all
allowable . Notice that is again not arbitrary anymore and
therefore Theorem 5 does not apply. The largest singular
value of that theorem has thus to be replaced by the so-called
largest structured singular value, which is defined as follows.
Let the set of allowable perturbations be denoted by D
nn and be defined as
D = { = block diag(1r Ik , . . ., rp Ik p , 1c Ik
1
p+1
, . . ., qc Ik p+q ,
(51)
min
max ()
otherwise
{D : det(IM )=0}
Computing D (M) is a difficult numerical problem. It is a nonconvex optimization problem and its complexity can be nonpolynomial for certain norms (see Ref. 41). One approach,
which is computationally rather demanding, is to formulate
the problem as a nondifferentiable convex optimization problem involving the maximum singular value of a matrix obtained from M. A more efficient scheme is given in Ref. 42
and uses several smooth optimization problems that do not
involve any eigenvalue or singular value computations. The
computational complexity of the problem of computing
D (M) has prompted several researchers to look for bounds
that are easier to compute (43,44).
CONCLUDING REMARKS
In this paper we have given several uses of the singular value
decomposition in analysis and design problems of systems
and control. We have considered computational issues and
useful properties of this decomposition as well, such as diagonalization, norms, and sensitivity. The list given here is far
from complete. Closeness problems (45,46) and canonical
forms (31,47) are just a few examples. We expect the number
of applications to grow also in the future because of the serious interdisciplinary effort that is under way between the
communities of the numerical linear algebra field on the one
hand and of the systems and control field on the other hand.
BIBLIOGRAPHY
1. E. Beltrami, Sulle funzioni bilineari, Giornali di Mathematiche,
11: 98106, 1873.
2. P. Lancaster and M. Tismenetsky, The Theory of Matrices, 2nd
ed., New York: Academic Press, 1985.
3. G. H. Golub and C. F. Van Loan, Matrix Computations, 2nd ed.,
Baltimore, MD: Johns Hopkins Univ. Press, 1989.
4. N. Higham, Accuracy and Stability of Numerical Algorithms, Philadelphia, PA: SIAM, 1995.
5. G. H. Golub and W. Kahan, Calculating the singular values and
pseudo-inverse of a matrix, SIAM J. Numer. Anal., 2: 205224,
1965.
6. G. H. Golub and C. Reinsch, Singular value decomposition and
least squares solutions, Numer. Math., 14: 403420, 1970.
7. J. R. Rice, A theory of condition, SIAM J. Numer. Anal., 3: 287
310, 1966.
8. J. H. Wilkinson, The Algebraic Eigenvalue Problem, Oxford, England: Oxford University Press, 1965.
9. E. G. Kogbetliantz, Solution of linear equations by diagonalization of coefficient matrix, Quart. Appl. Math., 13: 123132, 1955.
10. C. Paige and P. Van Dooren, On the quadratic convergence of
Kogbetliatzs algorithm for computing the singular value decomposition, Lin. Alg. Appl., 77: 301313, 1986.
11. J.-P. Charlier and P. Van Dooren, On Kogbetliantzs SVD algorithm in the presence of clusters, Lin. Alg. Appl., 95: 135160,
1987.
12. J. Demmel and W. Kahan, Accurate singular values of bidiagonal
matrices, SIAM J. Sci. Stat. Comput., 11: 873912, 1990.
13. V. Fernando and B. Parlett, Accurate singular values and differential qd algorithms, Numer. Math., 67: 191229, 1994.
14. G. Golub, K. Slna, and P. Van Dooren, Computing the Singular
Values of Products and Quotients of Matrices, in M. Mooren and
B. DeMoor (eds.), SVD in Signal Processing III, Algorithms, Architectures and Applications, Amsterdam: Elsevier, 1995.
SOFTWARE HOUSES
15. T. F. Chan, Rank revealing QR factorizations, Lin. Alg. Appl., 88/
89: 6782, 1987.
16. L. S. de Jong, Towards a formal definition of numerical stability,
Numer. Math., 28: 211220, 1977.
17. A. Bultheel, Recursive algorithms for the matrix Pade problem,
Math. Comput., 35: 875892, 1980.
18. B. C. Moore, Principal component analysis in linear systems:
Controllability, observability, and model reduction, IEEE Trans.
Autom. Control., AC-26: 1731, 1981.
19. H. Zeiger and A. McEwen, Approximate linear realizations of
given dimension via Hos algorithm, IEEE Trans. Autom. Control,
AC-19: 153, 1974.
20. S. Kung, A New Identification and Model Reduction Algorithm
via Singular Value Decompositions, Proc. 12th Asilomar Conf. Circuits Syst. Comp., 1978, pp. 705714.
21. M. Moonen et al., On- and off-line identification of linear statespace models, Int. J. Control, 49: 219232, 1989.
22. R. Patel, A. Laub, and P. Van Dooren, Numerical Linear Algebra
Techniques for Systems and Control, Piscataway, NJ: IEEE, 1993.
23. M. T. Heath et al., Computing the singular value decomposition
of a product of two matrices, SIAM J. Sci. Stat. Comput., 7: 1147
1159, 1986.
24. C. Mullis and R. Roberts, Synthesis of minimum roundoff noise
fixed point digital filters, IEEE Trans. Circuit Syst., CAS-23: 551
562, 1976.
25. M. Gevers and G. Li, Parametrizations in Control, Estimation and
Filtering Problems, London: Springer, 1993.
26. V. Adamjan, D. Arov, and M. Krein, Analytic properties of
Schmidt pairs for a Hankel operator and the generalized SchurTakagi problem, Mat. USSR Sbornik, 15: 3173, 1971.
27. K. Glover, All optimal Hankel norm approximations of linear
multivariable systems, and their L error bounds, Int. J. Control,
39: 11151193, 1984.
28. A. Antoulas, New results on the algebraic theory of linear systems: The solution of the cover problems, Lin. Alg. Appl., 50: 1
45, 1983.
29. T. Kailath, Linear Systems, Englewood Cliffs, NJ: Prentice-Hall,
1980.
30. C. Paige, Properties of numerical algorithms related to computing controllability, IEEE Trans. Autom. Control, AC-26: 130
138, 1981.
31. P. Van Dooren, The generalized eigenstructure problem in linear
system theory, IEEE Trans. Autom. Control, AC-26: 111129,
1981.
32. W. M. Wonham, Linear Multivariable Control: A Geometric Approach, 2nd ed., New York: Springer, 1979.
33. V. C. Klema and A. J. Laub, The singular value decomposition:
Its computation and some applications, IEEE Trans. Autom. Control, AC-25: 164176, 1980.
34. G. Zames, Feedback and optimal sensitivity: Model reference
transformations, multiplicative semi-norms, and approximate inverses, IEEE Trans. Autom. Control, AC-26: 301320, 1981.
35. J. C. Doyle, Analysis of feedback systems with structured uncertainties, IEE Proc., Pt. D, 129: 242250, 1982.
36. S. Boyd, V. Balakrishnan, and P. Kabamba, A bisection method
for computing the H norm of a transfer matrix and related problems, Math. Control Signals Syst., 2: 207219, 1989.
37. N. A. Bruinsma and M. Steinbuch, A fast algorithm to compute
the H norm of a transfer matrix, Syst. Control Lett., 14: 287
293, 1990.
38. R. Byers, A bisection method for measuring the distance of a stable matrix to the unstable matrices, SIAM J. Sci. Stat. Comput.,
9: 875881, 1988.
473
39. L. Qiu et al., A formula for computation of the real stability radius, Automatica, 31: 879890, 1995.
40. J. Sreedhar, P. Van Dooren, and A. L. Tits, A Level-Set Idea to
Compute the Real Hurwitz-Stability Radius, Proc. 34th IEEE
Conf. Dec. Control, New Orleans, 1995, WA05.
41. O. Toker and H. Ozbay, On the NP-Hardness of the Purely Complex Computation, Analysis/Synthesis, and Some Related Problems in Multi-Dimensional Systems, Proc. 1995 Amer. Control
Conf., 1995, pp. 447451.
42. M. K. H. Fan and A. L. Tits, Characterization and efficient computation of the structured singular value, IEEE Trans. Autom.
Control, AC-31: 734743, 1986.
43. P. M. Young, M. P. Newlin, and J. C. Doyle, Practical Computation of the Mixed Problem, Proc. 1992 Amer. Control Conf.,
Chicago, 1992, pp. 21902194.
44. C. T. Lawrence, A. L. Tits, and P. Van Dooren, A Fast Algorithm
for the Computation of an Upper Bound on the -norm, Proc.
13th IFAC World Congr., San Francisco, 1996, vol. H, pp. 5964.
45. R. Eising, Between controllable and uncontrollable, Syst. Control
Lett., 4: 263264, 1984.
46. D. Boley, Computing rank-deficiency of rectangular matrix pencils, Syst. Control Lett., 9: 207214, 1987.
47. M. M. Konstantinov, P. Hr. Petkov, and N. D. Christov, Invariants and canonical forms for linear multivariable systems under
the action of orthogonal transformation groups, Kybernetika, 17:
413424, 1981.
341
342
where
FTI International, Inc.
Hashikma St. Ind. Zone
P.O.B. 87 Kfar Saba,
Israel
Tel: 052-959152-4
Fax: 052-959162
Kuntz Automation
Engineering
402 Goetz Street, Dept. 7
Santa Ana, CA 92707
Tel: 714-540-7370
Fax: 714-540-6287
Eliminating the error e(t) and rearranging Eqs. (1) and (2)
gives the closed-loop gain
Munck Automation
Technology
161-T Enterprise Drive
Newport News, VA 23603
Tel: 804-887-8080
Fax: 804-887-5588
Portech Pathfinder
Operations
1610 Fry Avenue, Dept. T
Canon City, CO 81212
Tel: 800-959-0688
Fax: 719-269-1157
Fata Automation
37655 Interchange Drive
Farmington, MI 48335
Tel: 810-478-9090
Fax: 810-478-9557
Once the system performance is stated as differential equations in the time domain, Laplace transforms are commonly,
used for frequency analysis. Some examples of Laplace transforms are given in Table 2.
A control system can be an open loop or closed loop. Figure
1 illustrates an open-loop system in which a controller controls the process without using any feedback. In this case, the
output is not compared with the input; therefore, deviations
of the output from the desired value cannot be automatically
corrected. This method finds limited application since it does
not lead to fully automatic systems.
In a closed-loop system, the actual output is compared with
a desired reference input by a suitably arranged feedback
mechanism. Figure 2 illustrates a single-inputsingle-output
feedback control system. In this system, a prescribed relationship of one system variable to another is maintained by comparing the two and using the difference as a means of control.
Using system modeling and mathematical representations, a
closed loop control system with a single-input single-output
may be represented as shown in Fig. 3. In this case, the relationship between the input and output of the single-input
single-output system can be expressed as
e(t) = r(t) c(t)H
c(t) = e(t)G
(1)
(2)
M=
M(s) =
G
c(t)
=
r(t)
1 + GH
G(s)
C(s)
=
R(s)
1 + G(s)H(s)
(3)
343
1
s
1
s2
n!
s n1
eat
1
sa
eat ebt
ba
1
(s a)(s b)
n
1 2
2n
s 2 n s 2n
ent sin(n1 2 t)
sin(n t)
n
s 2 2n
cos(n t)
s
s 2 2n
1
1 2
n t
1 2
2n
s(s 2 n s 2n)
2
1 cos(n t)
2n
s(s 2n)
1 et/T
1
s(1 Ts)
1
[sin(n t) n t cos(n t)]
2n
s2
(s 2n)2
1 2T (1 2T )et/T
1
s 2(1 Ts)2
output relationship as
L
K i=1 (s + zi )
C(s)
= U
M(s) =
R
2
2
2
2
R(s)
k=1 (s + pk )
j=1 (s + 2 j s + a j s + j + j )
(4)
BOUNDED-INPUTBOUNDED-OUTPUT STABILITY
A system is stable if a bounded input yields to a bounded
output. When the closed-loop transfer function of a linear system is expressed as a Laplace transform, the stability may be
defined in terms of the locations of the poles of the system in
the complex plane or the s-plane. An s-plane is shown in Fig.
6 indicating the right half plane and left half plane. Take a
single-inputsingle-output system and express the input and
Desired
response
Reference
input or
Controller
Process
Output
Comparator
Controller
Process
Output
Measurement
desired response
Figure 1. An open-loop control system. An open-loop control system
does not compare the output and the input; therefore, any deviations
between the two cannot be corrected automatically. This system is
applicable only in simple cases in which the process characteristics
are fully known and the outputs from the desired values are not all
very important.
Figure 2. A closed-loop control system. In a single-inputsingle-output system, the output of the system is compared with the input,
and any deviation between the two is corrected by suitably designed
controllers. However, the use of feedback can lead to instability.
Closed-loop control arrangements are used extensively in automatic
processes and devices.
344
r(t) +
e(t)
c(t)
2.0
1.8
1.6
1.4
1.2
c(t)
Figure 3. Block diagram of a closed-loop system. A closed-loop system may be represented mathematically by a forward loop transfer
function G and a feedback loop transfer function H. The relation between the input and output can be expressed in terms of these two
terms in the form of a system transfer function, which is extremely
useful in system design and analysis.
1.0
Step
input
0.8
0.6
c(t) =
U
(Ak ePk ) +
k=1
R
B e t sin( t)
j=1
(5)
Left-Half-Plane Poles
To clarify the important concept of locations of poles on the splane, as used in stability analysis, see Eq. (4). As we can see
in this equation, the poles of the closed-loop system may be
real or complex, and simple or repeated. It is often convenient
to plot the poles and zeros of the closed-loop transfer function
on the s-plane. The s-plane can be considered to be in three
parts, the right half plane (RHP), the left half plane (LHP),
and the pure imaginary axis or j-axis. If a pole lies inside
the open LHP, then the pole has a negative real part. If it lies
inside the closed RHP, then it has a positive or zero repeated
real part.
Consider poles 1/(s p)n on the s-plane. For a real p, a
portion of the time domain response of the system will be proportional to
c(t) 1/n! t n1 e pt
(6)
If p 0, it lies on the RHP, and its response increases exponentially. If it is at the origin, p 0 and simple, its response
is a step function. When p 0 and repeated with multiplicity
n 2, then its response approaches infinity as t . If the
poles are in the LHP or p 0, then the response ept approaches zero as t .
Therefore,
1. The time response of a pole, simple or repeated, approaches zero as t if and only if the pole lies inside
the open-loop LHP or has a negative real part.
Inputs
(desired
responses)
Controllers
Process
Outputs
Measurement
Figure 4. A multivariable control system. Many control systems
have multiple inputs and multiple outputs. In these cases, using the
multivariable control theory and matrix approach is applicable. In
the design and analysis, most of the theories developed for singleinputsingle-output systems can still be used.
0.4
0.2
0
6
7
nt
10 11 12 13
2. The time response of a pole approaches a nonzero constant as t if and only if the pole is simple and
located at s 0.
As indicated earlier, in order to obtain a bounded response to
a bounded input, the poles of the closed-loop system must be
in the left-hand portion of the s-plane. That is, pk 0 and
j 0 so that the exponential terms epkt and ejt decay to
zero as the time goes to infinity. A necessary and sufficient
condition is that a feedback system is stable if all the poles of
the system transfer function have negative real parts. If the
characteristic equation has simple roots on the imaginary
axis ( j) with all other roots on the left half plane, the steady
state output is sustained oscillations for a bounded input. If
the input is sinusoidal with a frequency equal to the magni-
Imaginary axis
j
LHP
stable
s-plane
RHP
unstable
Real axis
(7)
Suppose that the transfer function of the overall control system is M(s); if M(s) is not stable the system cannot follow any
reference signals. If M(s) is stable, in order for the system to
be asymptotically stable, it is an additional requirement that
the system be capable of following all inputs. This is important because, in some cases, the output may be excited by
nonzero initial conditions such as noise or disturbance. As a
result, the stability conditions may be generalized as follows.
1. The system is stable if the natural response approaches
zero as t .
2. The system is unstable if the natural response grows
without bound as t .
3. The system is marginally stable or marginally unstable
if the natural response neither grows nor decays as t
.
4. The system is stable if bounded inputs result in
bounded outputs.
5. The system is unstable if bounded inputs result in unbounded outputs.
Here, in order to explain asymptotic stability and to lay a firm
background for the following theories, a rigorous mathematical approach may be introduced. To observe the natural response h(t) of a linear time-invariant system, a dirac delta
function (impulse) (t) may be applied to the input to give the
system internal energy upon which to act. The dirac delta
function has a rectangular shape with a height of 1/ and a
width of . is made vanishingly small so that the function
has infinite height and zero width and unit area. The ensuing
response is the natural response. Its Laplace transform is
identical to the transfer function of the system, which can be
written in the general partial fraction form as
H(s) =
n
F (s, pi , ri )
i=1
F (s, pi , ri ) =
345
K1
K2
Kr
+
+ +
(s pi )r i
(s pi )
(s pi )r i 1
(8)
where there are n sets of poles, located at s pi, each of multiple of ri.
The impulse response may be written in terms of the system poles by taking the inverse Laplace transform of H(s).
The general expression is
h(t) =
n
f (t, pi , ri )
(9)
i=1
pt
L = lim (t r1 |e pt |)
t
= lim
t r1
(10)
where p j. The limit is in the infinity divided by infinity indeterminate form. Applying LHopitals rule r 1 times
results in
(r 1)!
t r1 e t
(r 1)!
= lim
e t
t
r1
|L| = lim
(11)
L=
(r 1)!
lim
r1
( , r1, t )(0, 0, )
e t
1
= ek
1
=C
(12)
L=
lim
(r 1)!
( , t )(0, )
= lim
r1
(r 1)!
r1
e t
ek
(13)
346
Stability
ROUTHHURWITZ CRITERION
To determine the stability of a system, we need to know
whether any poles are located in the RHP. It is always possible to calculate these pole locations by direct computational
methods, but it is not necessary. For determining system stability, it is enough just to know whether there are any poles
on the RHP or not. This can be investigated using the Routh
Hurwitz criterion.
A generalized nth-order characteristic polynomial may be
represented as
P(s) =
n
sn1
sn2
sn3
..
.
..
.
s1
a i si
i=0
n
=k
s0
(14)
(s + pi )
i=0
(15)
an
an1
an1 an2 an an3
an1
bn1an3 an1 bn3
=
bn1
bn1 =
cn1
..
.
..
.
xn1
a0
an2
an4
an3
an5
an1 an4 an an5
an1
bn1 an5 an1 bn5
=
bn1
bn3 =
bn5 =
cn3
cn5
..
.
..
.
..
.
..
.
xn3
To simplify manual calculation of the table, it is useful to note
that multiplication of any row by a positive constant will not
affect the end result.
(16)
s4
s3
s2
s1
2
0
which confirms the fact that there are two poles on the right
half plane.
Special Cases
In forming Rouths array, there are three special cases that
need further consideration.
First Case. The first column of a row is zero, but the rest of
the row is not entirely zero.
Take the following polynomial as an example.
P(s) = s5 + s4 + 2s3 + 2s2 + 3s + 4
(18)
When the zero appears, replace it with a variable , to complete the table. Then take the limit as 0, both from above
and below, to determine if there are any sign changes.
0+
0
s5
s4
s2
1
2+
s1
s0
4 2
2 + 1
The table shows that there are two changes of sign in the first
column, regardless of whether approaches zero from above
or below in this case. Consequently, there are two roots in the
right half plane.
The poles are located at s1 0.6672 1.1638j, s2
0.5983 1.2632j and at s3 1.1377, confirming the
result.
Second Case. A whole row consists of zeros only.
When an entire row of zeros is encountered in row sm, an
auxiliary polynomial of order m 1 is formed by using the
sm1 row as the coefficient and by skipping every second power
of s. The row containing zeros is then replaced with the coef-
12
16
12
(19)
s3 3
s2 1
s1 2
s0 2
347
A(s) = 3s2 + 12
(20)
A (s) = 6s
(21)
Auxillary
polynomial
s5
s4
s2
8
8
16
s0
16
16
Derivative
Final table
entry
16
16
16
16
16
16
s + 8s + 16
4
4s + 16
2
4s + 16s
3
8s
8
16
348
Im
=
R=
Re
NYQUIST CRITERION
s-plane
(22)
jv G(s)-plane
where
N is the number of encirclements of the origin by 2,
Z is the number of zeros of L(s) encircled by 1, and
P is the number of poles of L(s) encircled by 1.
A positive N indicates that 2 and 1 both travel in the same
direction (i.e., clockwise or counterclockwise), whereas negative N indicates opposite directions. Some examples of contours 2 and 1 and encirclements are given in the section
dedicated for Nyquist. Interested readers should refer to this
section.
The Nyquist Contour
Consider the transfer function of a closed-loop system
C(s)
G(s)
=
R(s)
1 + G(s)H(s)
G(s)
=
1 + L(s)
(23)
(a)
s-plane
(b)
when the contour is mapped onto the L(s) plane. The direction
of the contour has been arbitrarily drawn as clockwise.
If the Nyquist contour is mapped onto the 1 L(s)-plane,
we would find that the resulting contour would encircle the
origin N Z P times (in the clockwise sense). It should be
emphasized that Z is the variable under investigation because it concerns the poles of the closed-loop system. The requirement for stability is that 1 L(s) contain no zeros on
the right half plane, or Z 0. That is to say, if the (clockwise)
Nyquist contour were mapped onto the 1 L(s)-plane, it is a
requirement for closed loop stability that the resulting contour encircle the origin counterclockwise exactly the same
number times as the number of poles of L(s) in the RHP.
The contour shown in Fig. 8 skips to the right around jaxis poles. Consequently, these j-axis poles are not considered to be right-half-plane poles. It is perfectly feasible for the
contour to skip to the left around these poles, in which case
they should be included in the count of right-half-plane poles.
It is emphasized that the poles of L(s) are easily obtainable.
A further refinement of the Nyquist criterion is that it is
unnecessary to plot the contour on the 1 L(s)-plane and
observe the number of encirclements about the origin. The
plot on the L(s) plane is in fact identical to that of the 1
L(s)-plane, except that it is shifted left by one unit. It will
therefore suffice to plot the contour on the L(s)-plane and observe the number of encirclements about the Cartesian point
(1, 0).
Simplified Nyquist Plot
Suppose that the function 1 L(s) contains P poles on the
RHP, P poles on the j-axis, and Z zeros on the right half
plane and that two contours are to be mapped onto the L(s)plane.
1. The Nyquist contour skips to the right around the P
poles on the j-axis. When mapped on the L(s)-plane, it
is found to encircle the Cartesian point (1, 0) point
N1 times.
2. The Nyquist contour skips to the left around the P
poles on the j-axis. When mapped on the L(s)-plane, it
is found to encircle the Cartesian point (1, 0) point
N2 times.
Each contour may be considered to consist of three sections,
each contributing a certain number of turns about the
Cartesian point (1, 0).
1. The section consisting of , excluding the
skips around j-axis poles. Because of symmetry about
the real axis in both the contour and the location of
poles and zeros, this section may be divided into two
halvesthe positive imaginary axis and the negative
imaginary axis, each contributing NA turns.
2. The section consisting of the infinite semicircle, contributing NB turns. If the order of the numerator of L(s) is
less than or equal to the denominator, then as s ,
L(s) corresponds to a point on the real axis or an encirclement of the origin. The contribution to the number of
turns about the Cartesian point (1, 0) in either case
is, therefore, NB 0.
349
(24)
N2 = 2NA + NB NC
(25)
(26)
2Z 2P P
4
P
= ZP
NA =
or
(27)
= P+
P
2
(28)
G(s)H(s) = L(s)
30
(s + 1)(s + 2)(s + 3)
30
= 3
s + 6s2 + 11s + 6
=
(29)
The Nyquist plot of L(s) of Eq. (29) can be obtained in a number of ways (e.g., polar plots) by substituting s j. By calculating the real and imaginary components of L(j), the Ny-
350
kNG DH
C(s)
=
R(s)
DG DH + kNG NH
Imag axis
2
1
0
1
2
3
4
2
Real axis
Figure 9. A typical Nyquist plot. This is the plot of a third-order
system and hence it traces three quadrants. The curve cuts the real
axis on the negative side. If the gain is increased sufficiently, the
curve will encircle the 1 point hence indicating instability. This
means that at least one of the roots of the characteristic equation,
poles of the closed loop system, will be on the right half of the s-plane.
(32)
(33a)
(33b)
i=1
k
s(s + 2)(s + 5)
k
= 3
s + 7s2 + 10s
T (s) =
(34)
(31)
(30)
1
k
(35)
Transfer function
Nyquist plot
351
Root locus
Imag axis
Imag axis
K
(s 1 + 1)
0
1
1
0
Real axis
Imag axis
Imag axis
Real axis
K
(s 1 + 1)(s 2+ 1)
0
Real axis
Imag axis
K
(s 1 + 1)(s 2+ 1) (s 3 + 1)
Imag axis
Real axis
1
0
0
Real axis
Imag axis
K
s(s 1 + 1)
Imag axis
Real axis
1
1
0
Real axis
Imag axis
K
s(s 1 + 1)(s 2 + 1)
Imag axis
Real axis
As k approaches zero, the magnitude of the loop transfer function becomes infinite, corresponding to a pole. For k becoming
infinitely large, the loop transfer function becomes infinitesimally small, corresponding to a zero.
Actually, inspection of Eq. (31) reveals that the poles of
H(s) and zeros of G(s) never actually appear as poles of the
+
G(s)
0
Real axis
Real axis
R(s)
1
1
C(s)
H(s)
Figure 11. Block diagram of a closed-loop system with variable k. In
many systems, one of the parameters of the system is varied to
achieve the desired response. In this case, the variation of k in the
forward path will relocate the roots of the characteristic equation on
the s-plane. The suitable locations of the roots lead to appropriate
system design.
352
10
8
6
d
G(s)H(s) = 0
ds
arg[G(s)H(s)] = (1 + 2n)
Imag axis
4
2
0
2
4
6
8
10
10
0
2
Real axis
10
Figure 12. The root locus of a system with a characteristic equation 1 ks(s 2)(s 5). This is a typical example of root locus.
Roots start from poles of the characteristic equation when k 0 and
approaches zeros as k . In this example, all three zeros are at .
At some value of k, the root loci crosses the imaginary axis to the
RHP, thus indicating unstable conditions. With the aid of root locus
a suitable value of k can be determined to locate the roots at the
desired points on the s-plane.
(38)
1
k
(39)
(36a)
(36b)
Imag axis
(2n + 1)
vu
v
u
i=1 pi
i=1 zi
A =
vu
(37b)
For real values of s, Eq. (37a) implies Eq. (37b), but for
complex s, there is no such implication.
If there are n poles involved in a breakaway point,
then there are always n branches entering and n
branches leaving. The angle between the entering and
leaving branches is given by
n =
(37a)
2
0
2
4
6
6
0
Real axis
Figure 13. Breakaway points. The root locus plotted for G(s)H(s)
(s 1)(s 3)(s2 4s 8) shows typical breakaway points at which
multiple roots meet and then diverge. The breakaways generally occur on the real axis, but they may occur anywhere. In this example,
a breakaway has happened on the real axis where as two others have
taken place on the s-plane, in which the corners of the two root loci
have met.
R(s)
1
(s + 1)(s + k)
353
ing that there is a known solution x1, then the equation may
be reduced to a first-order linear equation by letting
C(s)
x = x1 +
1
u
(44)
(45)
dx1
1 du
1
2
+ P(t) x1 +
dt
u dt
u
= Q(t) x21 +
1
2x1
+ 2
u
u
(46)
(40)
1
s2 + s + 1
s+1
1+k 2
s +s+1
G(s)
=
1 + kG(s)H(s)
C(s)
=
R(s)
(41)
The equation was first developed and applied by Count Riccati and Jacopo Francesco in the early eighteenth century.
In recent times, the Riccati equation finds wide application,
particularly in the area of optimal control and filtering. In
these applications, the matrix Riccati equation depicts a system of Riccati equations given by
X(t)
+ X(t)A(t) D(t)X(t) B(t)X(t) + C(t) = 0
(47)
(48)
which is a linear first-order differential equation and is simple to solve particularly if P(t), Q(t), and R(t) are constants,
as would the case if Eq. (41) were to describe a time-invariant system.
The Matrix Riccati Differential Equation
Consider the dynamic system given by the state-space description as
dx
+ P(t)x = Q(t)x2 + R(t)
dt
+ R(t)
(43)
x(t)
= Fx(t) + Gu(t)
y(t) = Hx(t)
(49)
where x(t) is the state matrix (n by 1), u(t) is the control matrix (q by 1), and y(t) is the matrix (p by 1) of output variables
to be controlled and F, G, and H are matrices of appropriate
dimensions that characterize the system.
In the optimal control applications, the object of optimal
control is to find u(t) over an interval t [t1, t2] such that
some cost function is optimized. One of the popularly used
cost function is the quadratic cost function that can be generalized by
F (t1 , t2 , T ) =
t2
t1
(51)
P(t)
+ P(t)F + F P(t) P(t)GG P(t) + H H = 0
with the terminal condition that P(t2) T.
(52)
354
X(t)
+ X(t)A A X(t) X(t)BX(t) + C = 0
(53)
U(t)
A
=
C
V(t)
B
A
U(t)
V(t)
(54)
(56)
As an example, consider
dx
+ 2x x2 + 1 = 0
dt
XA + A X XBX + C = 0
Hxj = xj + xj1
(57)
(60)
B
A
ui
vi
and
U = [u1 un ]
and
(61)
(64)
A
H=
C
(59)
(63)
Hx1 = x1
x(t) =
V = [v1 vn ]
then the solutions of Eq. (63) are given by X VU1.
b2
c
a
b
2
(66)
b
ac
2
b2 4ac
=
2
(67)
1
w = b
(68)
2a
a
The solution of Eq. (65) is then found as
2a 2a
b b2 4ac
=
2a
x=
1
x1 = b
2a
0
x2 = 1
(65)
H=
355
(69)
(70)
e = f(xe ) = 0
X
(72)
LYAPUNOV STABILITY
Lyapunov studied the question of motion, basing his theories
on the nonlinear differential equations of motion. His equations for linear motion are equivalent to Rouths criterion.
The Lyapunovs theorem determines the stability in the small
region about an equilibrium point. The stability in the large
may be determined by the global stability techniques. In this
article, stability in the small will be introduced, and some references will be made to global stability methods.
Lyapunov considers the stability of general systems described by ordinary differential equations expressed in the
state-variable form as
= f(x)
X
(71)
Unstable
x=0
xo
Asymptotically
stable
Stable
Figure 15. Lyapunov stability criteria. The equilibrium state xe is
stable if for every 0, there exists a 0, where depends only
on , such that the trajectory remains within a given neighborhood of
the equilibrium point. Likewise, an equilibrium state xe of a dynamic
system is unstable if there exists an , such that a corresponding
value cannot be found.
356
(73)
All eigenvalues of A have negative real parts if for any symmetric positive definite matrix N, the Lyapunov equation
AT M + AM = N
(74)
(75)
(76)
V = dxT Px/dt
= xT Px + xT Px
= xT (FT P + PF)x
= xT Qx
(77)
where
Q = (FT P + PF)
(78)
For any positive Q, the solution of P of the Lyapunov equation is positive if and only if all the characteristic roots of F
have negative real parts. That is, if a system matrix F is
given, it is possible to select a positive Q, solve the Lyapunov
equation in n(n 1)/2 unknowns, and test to see if P is positive by looking at the determinants of the n principal minors.
From this, the stability may be determined from the equations without either solving them or finding the characteristic roots.
The study of nonlinear systems is vast, here only the basic
principles of the methods have been discussed. Also, Lyapunov methods are applied in many diverse areas of control engineering; therefore, it is impossible to cover them all here.
Interested readers should refer to the reading list given at
the end of this article.
ROBUST STABILITY
Consider a linear time-invariant feedback system with a
plant transfer function G(s) and a compensator with Gc(s) cascaded as shown in Fig. 16. In many applications, the plant
model will not accurately represent the actual physical system because of (1) unmodeled dynamics and time delays, (2)
changes in equilibrium points, (3) nonlinear characteristics of
the plant, (4) noise and other disturbance inputs, and (5) parameter drift. The aim of a robust system is to assure that
performance is maintained in spite of model inaccuracies and
parameter changes.
The closed-loop transfer function of the system in Fig. 16
may be written as
M(s) =
G(s)Gc (s)
C(s)
=
R(s)
1 + G(s)Gc (s)
(79)
R(s) +
1
M/M
=
G/G
1 + G(s)Gc (s)
(80)
C(s)
Gc(s)
G(s)
H(s)
Figure 16. Block diagram of a closed-loop control system for robust
stability analysis. In the many mathematical representation of systems, a full account of all affected parameters may not be taken into
consideration because of unmodeled dynamics and time delays. Also,
during the operations, the equilibrium points may change, parameters may drift, and noise and disturbances may become significant.
The aim of a robust system is to assure that performance is maintained in spite of model inaccuracies and parameter changes.
G(s) H(s)-plane
EXPONENTIAL STABILITY
Imaginary
0
1
Real
(t, ) = e (t, )
Figure 17. An example of a closed-loop system resulting from parameter drifts in Nyquist plots. This diagram indicates that because
of uncertainties in modeling and changes in parameters the gain and
phase margins may be altered. These alterations may lead to unstable conditions if these margins are close to critical values.
As can be seen from Eq. (80), the sensitivity function has the
same characteristic equation [1 G(s)Gc(s)] as the closed-loop
transfer function M(s). For sensitivity to be small, it is necessary to have a high value for loop gain L(s) G(s) Gc(s). The
high gain is obtained at high frequencies of L( j). But as we
know, high gain could cause instability and poor responsiveness of M(s). Now, the design problem becomes a matter
of selecting Gc(s) such that the closed-loop sensitivity is small,
and the closed-loop transfer function has a wide bandwidth.
At the same time, the desired gain and phase margins must
be achieved.
The stability of the control system depends on the openloop transfer function L(s) G(s)Gc(s). Because of the uncertainties outlined here, the transfer function may be written
as
L(s) = Gc (s)[G(s) + G(s)]
357
(81)
For system stability, Nyquists stability condition must always be satisfied. That is the 1 point must not be encircled
by the L( j) under any circumstances. An example of uncertainty in a typical Nyquist plot resulting from G(s) is illustrated in Fig. 17. To guarantee stability, a safe gain and
phase margin must be ensured. Many methods are available
to deal with the robustness of the system including classical
methods linked to the use various compensators and PID controllers. The H technique is one method that finds extensive
application in robust control design and analysis.
In general, feedback reduces the effect of disturbances and
moderate modeling errors or parameter changes in the control
system. In the presence of disturbances and sensor noises,
systems are designed such that they keep the tracking errors
and outputs small for disturbance inputs. In order to achieve
this, the sensitivity to modeling errors and sensor noise must
be made small, thus making the system robust. In this case,
the plant output will follow any reference input asymptotically even if there are variations in the parameters of disturbance and noise. Briefly, it can be said that the system is
more robust if it can tolerate larger perturbations in its parameters.
(82)
HALIT EREN
ALEXANDER LOWE
Curtin University of Technology,
Australia
390
as time becomes large. This article deals with the latter notion of stability. An asymptotically stable response is the basis of a steady-state behavior whereby other responses asymptotically approach the steady-state response. A steady-state
response can be as simple as a constant (time-invariant) response, or it can be a periodic one. These, as well as other
more complicated steady-state responses, are described in the
first section.
In our study of dynamical systems, we model the system
by a finite number of coupled first-order ordinary differential
equations
x1 = f 1 (t, x1 , . . ., xn )
x2 = f 2 (t, x1 , . . ., xn )
..
.
..
.
xn = f n (t, x1 , . . ., xn )
where xi denotes the derivative of xi with respect to the time
variable t. We call the variables x1, x2, . . ., xn the state variables. They represent the memory that the dynamical system
has of its past. They are usually chosen as physical variables
that represent the energy-storing elements. For example, in
an RLC electrical circuit, the state variables could be voltages
across capacitors and currents through inductors, while in a
spring-mass-damper mechanical system, the state variables
could be positions and velocities of moving masses. We usually use vector notation to write the above equations in a compact form. Define
x1
f 1 (t, x)
x2
f 2 (t, x)
x = . , f (t, x) = .
..
..
xn
f n (t, x)
and rewrite the n first-order differential equations as an ndimensional first-order vector differential equation
x = f (t, x)
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
STEADY-STATE BEHAVIOR
The steady-state behavior of a system is described by the asymptotic nature of solutions as time becomes large. For nonlinear systems, this will depend on the system as well as on
the initial conditions provided. The possible types of steadystate behavior are more varied than one might think; they
include the well-known constant time behavior (asymptotically stable equilibria) and periodic time behavior (asymptotically stable limit cycles), as well as more complicated behaviors, such as multiperiodic behavior (asymptotically stable
tori) and chaos (strange attractors). We begin with some
definitions that are general enough to capture this range of
possibilities, and then provide some examples.
The steady-state behavior of a system takes place on a subset of the state space called an attractor. The key ingredients
for defining an attractor A are the following: (1) If a solution
is started in A , it never leaves A . That is, A is an invariant
set, defined by saying that for each x(0) x0 A , x(t) A
t. (2) Solutions started sufficiently close to A will approach
A as t . That is, A is locally attractive. (3) A feature of
an attractor A is that it contains a solution that comes arbitrarily close to every point in A at some time. This implies
that A is minimal in the sense that there are no subsets of
A that satisfy conditions (1) and (2).
The domain of attraction (or region, or basin, of attraction)
for an attractor A is defined to be the set of initial conditions
in the state space that are asymptotic to A as t . It can
(at least formally) be constructed by considering a neighborhood UA of A that is used in proving its asymptotic stability,
and taking t0x(t) x(0) UA . This simply starts solutions
in a neighborhood in which one knows they will approach A
in forward time and runs time backward. In this way all solutions that will approach A are collected. Except in simple
problems it is impossible to determine the domain of attraction, although parts of it can often be estimated using Lyapunov methods, as described in the next section. It is important to realize that a nonlinear system may possess multiple
attractors of various types. The domains of attraction for different attractors must, of course, be distinct. Typically, differ-
391
(1)
which is expressed in polar coordinate form. This two-dimensional system has an invariant set (r, ): r 1, [0, 2),
a circle, which attracts all solutions except for the unstable
equilibrium at r 0, and on which solutions wind around
with constant speed leading to a period T 2/. Furthermore, since all points on the circle are visited by every solution started on it, it satisfies all the conditions for an attractor. Such a closed trajectory can, of course, exist in
higher-order systems as well. Another example of an asymptotically stable periodic attractor is offered by the simple linear equation
x = x + cos(t)
(2)
=
Figure 1. Limit cycle in a cylindrical state space.
392
1 = 1 ,
2 = 2
(3)
functions of x, defined for all x in a domain D n that contains the origin x 0. Suppose the origin is an equilibrium
point of x f(x); that is, f(0) 0. Our goal is to characterize
and study the stability of the origin. There is no loss of generality in taking the equilibrium point at the origin, for any
equilibrium point x 0 can be shifted to the origin via the
change of variables y x x.
The equilibrium point x 0 of x f(x) is stable, if for each
0, there is () 0 such that x(0) implies that
x(t) , for all t 0. It is said to be asymptotically stable if
it is stable and can be chosen such that x(0) implies
that x(t) approaches the origin as t tends to infinity. When
the origin is asymptotically stable, the domain of attraction is
defined as the set of all points x such that the solution of x
f(x) that starts from x at time t 0 approaches the origin as
t tends to . When the domain of attraction is the whole
space n, we say that the origin is globally asymptotically
stable.
Lyapunovs Method
In 1892, Lyapunov introduced a method to determine the stability of equilibrium points without solving the state equation.
Let V(x) be a continuously differentiable scalar function defined in D. A function V(x) is said to be positive definite if
V(0) 0 and V(x) 0 for every x 0. It is said to be positive
semidefinite if V(x) 0 for all x. A function V(x) is said to be
negative definite or negative semidefinite if V(x) is positive
definite or positive semidefinite, respectively. The derivative
(x)
of V along the trajectories of x f(x) is given by V
n
i1(V/xi)xi (V/x)f(x), where V/x is a row vector whose
ith component is V/xi.
Lyapunovs stability theorem states that the origin is stable
if there is a continuously differentiable positive definite function
(x) is negative semidefinite, and it is asymptotiV(x) so that V
(x) is negative definite. A function V(x) satiscally stable if V
fying the conditions for stability is called a Lyapunov function.
The surface V(x) c, for some c 0, is called a Lyapunov
surface or a level surface. Using Lyapunov surfaces, Fig. 2
makes the theorem intuitively clear. It shows Lyapunov
surfaces for decreasing constants c3 c2 c1 0. The condi 0 implies that V(x(t)) decreases along the trajectory
tion V
x(t). Therefore, when a trajectory crosses a Lyapunov surface
V(x) c, it moves inside the set V(x) c and can never come
0, the trajectory moves from one Lyapuout again. When V
nov surface to an inner Lyapunov surface with a smaller c.
As c decreases, the Lyapunov surface V(x) c shrinks to the
origin, showing that the trajectory approaches the origin as
0, we cannot be
time progresses. If we only know that V
c3
c2
V(x) = c1
c1 < c2 < c3
Figure 2. Level surfaces of a Lyapunov function.
sure that the trajectory will approach the origin, but we can
conclude that the origin is stable since the trajectory can be
contained inside any neighborhood of the origin by requiring
the initial state x(0) to lie inside a Lyapunov surface contained in that neighborhood.
(x) is only negative semidefinite, we may still be
When V
able to conclude asymptotic stability of the origin if we can
(x)
show that no solution can stay identically in the set V
0, other than the trivial solution x(t) 0. Under this condition, V(x(t)) must decrease toward 0, and consequently x(t)
0 as t . This extension of the basic theorem is usually
referred to as the invariance principle.
Lyapunov functions can be used to estimate the domain of
attraction of an asymptotically stable origin, that is, to find
sets contained in the domain of attraction. If there is a Lyapunov function that satisfies the conditions of asymptotic stability over a domain D and if V(x) c is bounded and contained
in D, then every trajectory starting in V(x) c remains in
V(x) c and approaches the origin as t . Thus, V(x)
c is an estimate of the domain of attraction. If the Lyapunov
function V(x) is radially unbounded that is, x implies
that V(x) , then any point x n can be included in the
bounded set V(x) c. Therefore, the origin is globally asymptotically stable if there is a continuously differentiable, radially
unbounded function V(x) such that for all x n, V(x) is posi (x) is either negative definite or negative
tive definite and V
semidefinite but no solution can stay identically in the set
(x) 0 other than the trivial solution x(t) 0.
V
Lyapunovs method is a very powerful tool for studying the
stability of equilibrium points. However, there are two drawbacks to the method of which the reader should be aware.
First, there is no systematic method for finding a Lyapunov
function for a given system. In some cases, there are natural
Lyapunov function candidates like energy functions in electrical or mechanical systems (see Example 2). In other cases, it
is basically a matter of trial and error. Second, the conditions
of the method are only sufficient; they are not necessary. Failure of a Lyapunov function candidate to satisfy the conditions
for stability or asymptotic stability does not mean that the
origin is not stable or asymptotically stable.
Example 1. Consider the second-order system
x1 = x1 + x21 x2 ,
x2 = x1 x2
P=
a
b
b
d
393
by
a
Q=2
0.5d
0.5d
d
x2 = a sin x1 bx2
394
x1
V (x) =
a sin y dy +
1
2
1
2
x22
x2 = g1 (x1 ) g2 (x2 )
Linear Systems
The linear time-invariant system x Ax has an equilibrium
point at the origin. The equilibrium point is isolated if and
only if det(A) 0. Stability properties of the origin can be
characterized by the locations of the eigenvalues of the matrix A. Recall from linear system theory that the solution of
x Ax for a given initial state x(0) is given by x(t)
exp(At)x(0) and that for any matrix A there is a nonsingular
matrix P (possibly complex) that transforms A into its Jordan
form; that is,
P1 AP = J = block diag[J1 , J2 , . . ., Jr ]
where Ji is the Jordan block associated with the eigenvalue
i of A. Therefore,
exp(At) = P exp(Jt)P1 =
mi
r
t k1 exp(it)Rik
(4)
i=1 k=1
where g1( ) and g2( ) are locally Lipschitz functions and saty
isfy gi(0) 0, ygi(y) 0 y 0, i 1, 2, and 0 g1(z) dz
, as y . The system has an isolated equilibrium point
at the origin. The equation of this system can be viewed as a
generalized pendulum equation with g2(x2) as the friction
term. Therefore, a Lyapunov-function candidate may be taken
x
as the energylike function V(x) 01 g1(y) dy x22, which is
2
positive definite in and radially unbounded. The derivative
of V(x) along the trajectories of the system is given by
V (x) = g1 (x1 )x2 + x2 [g1 (x1 ) g2 (x2 )] = x2 g2 (x2 ) 0
(x) is negative semidefinite. Note that V
(x) 0 imThus, V
plies x2g2(x2) 0, which implies x2 0. The only solution that
can stay identically in the set x 2x2 0 is the trivial
solution x(t) 0. Thus, by the invariance principle, the origin
is globally asymptotically stable.
PA + AT P = Q
(5)
f
1
x1
f
=
f
x
2
x1
f1
0
x2
=
a
cos
x1
f2
x2
1
b
0
f
A=
=
x x=0
a
1
b
f
A =
x x
1 = , x 2 =0
0
=
a
1
b
395
z = z + ay 2
(6)
z = A2 z + g2 ( y, z)
(8)
396
A center manifold for this system is simply a smooth invariant manifold of the form z h(y) with h(0) 0 and
(h/y)(0) 0. Under some smoothness conditions on the nonlinear terms in Eq. (8) (that are inherited from the original
equation), a local center manifold exists, although it is not in
general unique. The power of the center manifold is that it
can be used to reduce the dimensionality of the problem, as
follows. By restricting the system dynamics to the center
manifold one obtains y A1y g1(y, h(y)), referred to as the
reduced system. The center manifold theorem states that if the
origin of the reduced system is asymptotically stable (unstable),
then the origin of the original system x f(x) is likewise asymptotically stable (unstable).
The construction of the center manifold can be carried
out as follows. We take the time derivative of z h(y) to obtain z [h(y)/y]y. Equation (8) is used to substitute in for
z and y, and z is replaced everywhere by h(y). This leads to
A2h(y) g2(y, h(y)) [h(y)/y][A1y g1(y, h(y))]. This equation for h(y), which must satisfy the conditions h(0) 0 and
h/y(0) 0, is generally impossible to solve. However, since
only local information is needed for stability considerations,
an approximation for h(y) can be obtained by assuming a series expansion for h(y), substituting it into the equation, and
matching coefficients, as demonstrated in the forthcoming Example 5. Once the expansion form for h(y) is determined and
the expanded version of the reduced equation is in hand, various techniques can be employed for determining the stability
of the reduced system. In general, this task is made much
easier due to the lower dimensionality of the reduced system.
Example 5. Consider the system given in Eq. (6). Here A1
0, A2 1, g1(y,z) yz y3, and g2(y,z) ay2. The center
manifold is assumed to be of the form h(y) c1y2 c2y3
. This is substituted into the equation for h(y) and expanded in powers of y, and the coefficients of y2, y3, etc., are
gathered and solved. This leads to the result that c1 a and
c2 0. Therefore, h(y) ay2 O(y4). [We use the notation
f(y) O(yp) when f(y) kyp for sufficiently small y.] The
reduced system is given by taking the equation for y and
simply replacing z by the expansion for h(y), resulting in y
(a 1)y3 O(y5). Thus, the conclusion is reached that for
a 1 0, x 0 is an asymptotically stable equilibrium point
for system Eq. (6), while for a 1 0 the origin is unstable.
For a 1 0 no conclusions regarding stability can be drawn
without considering higher-order expansions.
Nonautonomous Systems
Suppose the origin x 0 is an equilibrium point of the nonautonomous system x f(t, x); that is, f(t, 0) 0 for all t 0.
For nonautonomous systems we allow the Lyapunov-function
candidate V to depend on t. Let V(t, x) be a continuously differentiable function defined for all t 0 and all x D. The
derivative of V along the trajectories of x f(t, x) is given by
(t, x) V/t (V/x) f(t, x). If there are positive-definite
V
functions W1(x), W2(x), and W3(x) such that
W1 (x) V (t, x) W2 (x),
V (t, x) W3 (x)
(9)
for all t 0 and all x D, then the origin is uniformly asymptotically stable, where uniformly indicates that the
definition of stability and the convergence of x(t) to zero are
x2 = x1 x2
V 2 x 21 + 2 x1 x2 2 x 22 = xT
2
1
1
x = xT Qx
2
The matrix Q is positive definite. Hence the origin is uniformly asymptotically stable. Since all inequalities are satisfied globally and x12 x22 is radially unbounded, the origin is
globally uniformly asymptotically stable.
BIFURCATION THEORY
The term bifurcation, strictly speaking, refers to the splitting
of a whole into two parts. While this is relevant to its meaning in dynamical systems, it has taken on a much broader
definition. The general bifurcation problem deals with qualitative changes in system behavior as parameters are varied
in a quasistatic manner. The simplest example is the case
when a parameter is varied in such a manner that the real
part of an eigenvalue of the Jacobian at an equilibrium point
changes sign, corresponding to the change in stability of the
equilibrium. There are two generic ways in which such transitions can occur. The first is a real eigenvalue passing through
zero; the second is a complex conjugate pair of eigenvalues
passing through the imaginary axis. While linearization and
the center-manifold theory allow one to determine the stability of the equilibrium point, the larger question looms as to
what changes take place near the equilibrium through such a
transition. Questions such as these are the basic motivation
behind bifurcation theory.
Consider the system x f(x, ), where represents a system parameter. The technical definition of a bifurcation is as
follows: A bifurcation is said to occur at 0 if the state space
for 0 is not topologically equivalent to that for 0. It
is said that 0 is the bifurcation value of the parameter. The
y = y + zy y 3 ,
z = z + ay2
(10)
397
y
a >1
a <1
398
STACK SIMULATION
the equilibrium persists through the instability, a transcritical bifurcation, shown in Fig. 4, generally occurs. When a special symmetry in the problem exists, the pitchfork bifurcation
encountered in the preceding example can occur. These cover
the generic bifurcations involving a single eigenvalue going
through zero. When a complex conjugate pair of eigenvalues
passes through the imaginary axis, something more interesting happens. First, the center-manifold reduction for this
problem leads to a dynamical system consisting of two equations. When expressed in polar coordinates these are r
r ar3 and br2 . The radial variable,
representing the amplitude of the oscillation, undergoes a
pitchfork bifurcation (although r 0 is meaningless) that is
super- (a 0) or sub- (a 0) critical as is increased through
zero. The angular variable simply describes rotation at a
nominal frequency of with a small amplitude-dependent
shift arising from nonlinear effects. This is the Hopf bifurcation. The result of this bifurcation is shown in Fig. 5, and it
results in the birth of stable limit cycles as an asymptotically
stable equilibrium goes unstable (for a 0) or in the merging
of an asymptotically stable equilibrium with an unstable limit
cycle (for a 0). This completes the list of bifurcations that
generically occur as a single parameter is varied and an equilibrium point changes stability.
Center-manifold and normal-form methods also exist for
analyzing the bifurcations that occur when limit cycles
change stability. These problems can be handled by defining
and using a Poincare map near the limit cycle and studying
the generic bifurcations of fixed points of maps. It is found
that limit cycles can undergo saddle-node, transcritical, and
pitchfork bifurcations that are essentially the same as shown
in Figs. 3 and 4. The analogy to the Hopf bifurcation, the
Neimark-Sacker bifurcation, involves a limit cycle changing
stability and merging with a two-dimensional torus around it
(in a super- or subcritical manner). However, this situation is
complicated by the fact that resonances can occur between the
two frequencies of the torus, and this can lead to secondary
nonlinear phenomena. In addition, limit cycles have an additional bifurcation, the period-doubling bifurcation, in which a
limit cycle changes stability and merges with another limit
cycle that has a period twice that of the original. Again, this
can be super- or subcritical in nature.
Before leaving the topic of local bifurcations it should be
noted that if several parameters are being simultaneously
varied, one can encounter situations in which several eigenvalues simultaneously have zero real parts. Similarly, one
may find that some critical terms in a given normal form
change sign as parameters are varied (for example, the parameter a in the Hopf bifurcation). The bifurcation analysis
of such problems is extremely rich and often exceedingly complicated, but some classifications along these lines have been
carried out for two and three parameter systems.
Global bifurcations result in qualitative changes in the
state space that cannot be described as local to any equilib-
HASSAN K. KHALIL
STEVEN W. SHAW
Michigan State University
357
358
design methodology for nonlinear tracking controllers for actuator saturation effects mitigation, is presented.
LYAPUNOV STABILITY CRITERION
Stability is a very important characteristic of a system. If a
system is linear and time-invariant, several criteria such as
the Nyquist and Routhian stability criteria may be applied to
determine stability. However, for nonlinear and time-varying
systems, the Lyapunov direct method is applicable. Its advantage is that it is not necessary to solve the system equation,
which is generally very difficult.
The system with zero input is defined by
x = f(x, t)
(1)
x = f(x, u, t)
(2)
(3)
t t0
(4)
(5)
(6)
An equivalent condition for an LTI system is that all eigenvalues (3) are located in the left-half plane. In that case, the
transient response associated with those eigenvalues decrease with time, and the output response therefore approaches the particular solution which is determined by the
input.
Asymptotic Stability
b
c
x(0)
x0
it represents asymptotic stability. Trajectory c leaves the region and therefore indicates instability.
When the region includes the entire state space, the
definitions of Lyapunov and asymptotic stability are said to
apply in a global sense. The stability or instability of a linear
system is global because any initial state yields the same
stability determination. A stable linear system is globally
asymptotically stable.
The technique for linerarizing nonlinear differential equations in the neighborhood of their singularities or equilibrium
points is presented in the section entitled Linearization (Jacobian Matrix). The validity of determining stability of the
unperturbed solution near the singular points from the linearized equations was developed independently by Poincare
and Lyapunov in 1892. Lyapunov (4,7) designated this as the
first method. This stability determination is applicable only in
a small region near the singularity and results in stability in
the small. The section on linearization considers Lyapunovs
second method, which is used to determine stability in the
large. This larger region may include a finite portion, or sometimes the whole region, of the state space.
359
f (x, y) =
n
n
i=1 j=1
a i j xi y j
a11
a
= [x1 x2 xn ] 21
an1
a12
a22
an2
y1
a1n
y
a2n 2
ann
yn
(11)
= xT Ay = x, Ay
V =
n
n
a i j xi x j
(12)
i=1 j=1
where all aij are real. This is the special case of Eq. (11)
where x y. The quadratic form V can therefore be expressed
as the inner product
QUADRATIC FORMS
Some of the techniques used in determining stability of control systems and for optimizing their response utilize scalar
functions expressed in quadratic form. The necessary background for expressing functions in quadratic form is developed in this section. Then some important properties of quadratic forms are presented.
Conjugate Matrix. The elements of a matrix may be complex quantities. For example, a matrix A may have the elements aij ij jij. A conjugate matrix B has elements with
the same real component and with imaginary components of
the opposite sign; that is, bij ij jij. This conjugate property is expressed by
B = A
(7)
(8)
(9)
V (x) = xT Ax = x, Ax
1
3
0
= x T 3
4 10 x = xT Ax
0 10
5
(14)
(10)
(13)
x
x
(16)
Principal Minor. A principal minor of a matrix A is obtained by deleting any row(s) and the same numbered col-
360
1 = |a11 |
123
a
11
= a21
a31
12
a12
a22
a32
a
= 11
a21
a13
a23 12n = |A|
a33
a12
a22
(17)
1 1
A = 1 1
1 1
1
0
(18)
The leading principal minors are evaluated to check for positive definiteness:
1 1 1
1 1
= 0,
1 = 1,
12 =
=
1
1
1
==0
123
1 1
1 1 0
The conditions for positive definiteness are not satisfied. Next
the remaining principal minors are evaluated to check for
positive semidefiniteness.
1 1
1 1
= 1,
13 =
=
23
1 0 = 1
1 0
2 = 1,
3 = 0
x2
Since the principal minors are not all nonnegative, the matrix A is not positive semidefinite. Similarly, A is not negative
definite or negative semidefinite; therefore A is indefinite.
Example 2. Transform A in Eq. (18) to the diagonal Jordan
form and check its definiteness. The characteristic equation
is
1
1
1
|II A| = 1
1 1 = 3 22 2 = 0
(19)
1
1
The eigenvalues are 1 2.732, 2 0.732, and 3 0. Thus
the matrix A is indefinite since
2.732
= 0
0
0
0.732
0
0
0
A linear system with no forcing function (an autonomous system) and with A 0 has only one equilibrium point x0. A
nonlinear system, on the other hand, may have more than
one equilibrium point. This is easily illustrated by considering
the unity-feedback angular position-control system shown in
Fig. 2. The feedback action is provided by synchros which generate the actuating signal e sin(i o). With no input,
i 0, the differential equation of the system is
o + ao + K sin o = 0
(20)
x2 = K sin x1 ax2
(21)
The slope of the trajectories in the phase plane (25) is obtained from
x2
K sin x1 ax2
=
x1
x2
(22)
f( e)
K
s(s + a)
Saddle point
Saddle point
x1
Separatix
in a Taylor series about one of the equilibrium points x0. Assuming that x* is restricted to a small neighborhood of the
equilibrium point, the higher-order terms in the Taylor series
may be neglected. Thus, the resulting linear variational state
equation is
f1
f1 f1
x
xn
1 x2
f2 f2
f
2
x1 x2
xn
x =
x = Jx x
(23)
fn
fn fn
x1 x2
xn
x= x 0
N=
Node
x1 = x2 ,
361
Synchro
Figure 2. A nonlinear feedback control system.
0
1 x1
x1
= Jx x
x = =
(24)
K a x1
x2
For these linearized equations the eigenvalues are 1,2
a/2 (a/2)2 K. This equilibrium point is stable and is
either a node or a focus, depending upon the magnitudes of a
and K. A node denotes an overdamped response and a focus
denotes an underdamped response about the equilibrium
point. For motion about the equilibrium point x1 , x2 0,
the state equation is
0
1
x =
x
(25)
K a
The eigenvalues of Jx are 1,2 a/2 (a/2)2 K. Thus, one
eigenvalue is positive and the other is negative, and the equilibrium point represents an unstable saddle point. The motion
around the saddle point is considered unstable since every
point on all trajectories, except on the two separatrices, moves
away from this equilibrium point. A phase-plane portrait for
this system can be obtained by the method of isoclines (25).
From Eq. (22) the isocline equation is x2 K sin x1 /(N a).
The phase portrait is shown in Fig. 3. Note that the linearized
362
equations are applicable only in the neighborhood of the singular points. Thus, they describe stability in the small.
In analyzing the system performance in the vicinity of
equilibrium, it is usually convenient to translate the origin of
the state space to that point. This is done by inserting x
x0 x* into the original equations. With the origin at x0, the
variations from this point are described by x*.
SECOND METHOD OF LYAPUNOV
The second, or direct, method of Lyapunov provides a means
for determining the stability of a system without explicitly
solving for the trajectories in the state space. This is in contrast to the first method of Lyapunov, which requires the determination of the eigenvalues from the linearized equations
about an equilibrium point. The second method is applicble
for determining the behavior of higher-order systems which
may be forced or unforced, linear or nonlinear, time-invariant
or time-varying, and deterministic or stochastic. Solution of
the differential equation is not required. The procedure requires the selection of a scalar energy function V(x), which
is tested for the conditions that indicate stability. When V(x)
successfully meets these conditions, it is called a Lyapunov
function. The principal difficulty in applying the method is in
formulating a correct Lyapunov function because, for asymptotically stable systems, the failure of one function to meet
the stability conditions does not mean that a true Lyapunov
function does not exist. This difficulty is compounded by the
fact that the Lyapunov function is not unique. Nevertheless,
there is much interest in the second method (10,11).
In order to show a simple example of a Lyapunov function,
consider the system of Fig. 2 represented by Eq. (20) which
has multiple equilibrium points at o 0, o n, where n is
an integer. The proposed function is the sum of the kinetic
and potential stored energies, given by
V (o , o ) = 12 o2 + K(1 cos o )
(26)
ellipsoid, in the state space with its center at the origin. The
entire state space is filled with such nonintersecting closed
surfaces, each representing a different positive value of V(x).
(x) is negative for all points in the state space, it
When V
means that all trajectories cross the closed surfaces from the
outside to the inside and eventually converge at the equilibrium point at the origin.
A qualitatively correct Lyapunov function is shown in Fig.
4 for a second-order system. The function V(x1, x2) x12 x22
is positive definite and is represented by the paraboloid surface shown. The value V(x1, x2) ki (a constant) is represented
by the intersection of the surface V(x1, x2) and the plane z
ki. The projection of this intersection on the x1, x2 plane is a
closed curve, an oval, around the origin. There is a family of
such closed curves in the x1, x2 plane for different values of
ki. The value V(x1, x2) 0 is the point at the origin; it is the
innermost curve of the family of curves representing different
levels on the paraboloid.
The gradient vector of V(x) is defined by
V (x)
x1
V (x)
V (x)
xn
(x) along any trajectory is
The time derivative V
V (x) dx1
V (x) dx2
V (x) dxn
V (x) =
+
+ +
x1 dt
x2 dt
xn dt
(29)
= [V (x)]T x = V, x
(27)
of o, except o 0. Note that the slope d o /do of the phaseplane trajectories [see Eq. (22)] is infinite along the line represented by o 0, except at the equilibrium points o n.
Thus, the line o 0 does not represent an equilibrium, except at o n, and it is not a trajectory in the state plane.
Since the energy stored in the system is continuously decreasing at all points except at the equilibrium points, the equilibrium at the origin and at even multiples of o n is asymp is NSD.
totically stable. This V
This example demonstrates that the total system energy
may be used as the Lyapunov function. When the equations
of a large system are given in mathematical form, it is usually
difficult to define the energy of the system. Thus, alternate
Lyapunov functions must be obtained. For any system of order n, a positive, constant value of the proper positive definite
Lyapunov function V(x) represents a closed surface, a hyper-
z = V(x)
x2
V(x) = const = k1
x1
0
V(x)
State plane trajectory
cally stable, and V(x) is a proper Lyapunov function. The concepts described above may be summarized in the following
theorem, which provides sufficient, but not necessary, conditions for stability.
Theorem 1. Lyapunov asymptotic stability. A system x
f(x, t), where f(0, t) 0 for all t, is asymptotically stable in
the vicinity of the equilibrium point at the origin if there exists a scalar function V(x) such that:
1. V(x) is continuous and has continuous first partial derivatives in a region S around the origin.
2. V(x) 0 for x 0. [V(x) is PD.]
3. V(0) 0.
(x) 0 for x 0. [V
(x) is ND.]
4. V
Conditions 13 ensure that V(x) is positive definite. Therefore, V(x) k is a closed surface within the region S. Condi (x) is negative definite, and thus any tration 4 means that V
jectory in S crosses through the surface V(x) k from the
outside to the inside for all values of k. Therefore, the trajectory converges on the origin where V(0) 0.
(x) 0 for x 0 can be relaxed in
The condition that V
Theorem 1 under the proper conditions. Condition 4 can be
(x) 0; that is, V
(x) is negative semidefinite.
changed to V
(x) is not
This relaxed condition is sufficient, provided that V
equal to zero at any solution of the original differential equation except at the equilibrium point at the origin. A test for
(x) 0 into the
this condition is to insert the solution of V
state equation x f(x) to verify that it is satisfied only at the
equilibrium point. Also, if it can be shown that no trajectory
can stay forever at the points or on the line, other than the
0, then the origin is asymptotically staorigin, at which V
ble. This is the case for the system of Fig. 2 as described at
the beginning of this section. For linear systems there is only
one equilibrium point which is at the origin; therefore it is
(x) to be negative semidefinite. Theorem 1 may
sufficient for V
also be extended so that it is applicable to the entire state
space. In that case the system is said to have global stability,
or stability in the large. Including these conditions results in
the following theorem:
Theorem 2. Lyapunov global asymptotic stability. A system
is globally asymptotically stable if there is only one stable
equilibrium point and there exists a scalar function V(x)
such that:
1. V(x) is continuous and has continuous first partial derivatives in the entire state space.
2. V(x) 0 for x 0.
3. V(0) 0.
4. V(x) at x .
(x) 0.
5. V
(x) 0, except at x 0, or any locus in the
6. Either V
(x) 0 is not a trajectory of the
state space where V
system.
any xi , then V(x) ki is a closed curve for any ki. Conditions 5 and 6 mean that V(x) is continuously decreasing along
any trajectory in the entire plane and ensures that the system
is asymptotically stable. In order to check for global stability,
it is necessary to select V(x) so that conditions 14 are all satisfied.
Finding a proper Lyapunov function V(x) means that the
system is stable, but this is just a sufficient and not a necessary condition for stability. The fact that a function V(x) has
not been found does not mean that it does not exist and that
the system is not stable.
Application of the Lyapunov Method to Linear Systems
The second method of Lyapunov is applicable to time-varying
and nonlinear systems. However, there is no simple general
method of developing the Lyapunov function. Methods have
been developed for many such systems, and the literature in
this area (4) is extensive. The remaining applications in this
chapter are restricted to linear systems. Since linear systems
may have only one equilibrium point, the stability or instability is necessarily global in nature. The RouthHurwitz stability criterion is available for determining the stability of linear
systems. However, the necessity of obtaining the characteristic polynomial can be a disadvantage for higher-order systems. Thus, the following material is presented to develop familiarity with the more general second method of viewing
stability. Confidence in the second method of Lyapunov can
be developed by showing that the results are identical to
those obtained with the RouthHurwitz method.
The approach presented is to select a V(x) which is positive
(x). The coefficients of V(x) and
definite and to evaluate V
those restraints on the system parameters are then deter (x) negative definite or negative semimined which make V
definite. Consider the linear, unity feedback system presented
in Fig. 5 with r(t) 0.
Example
G(s) =
K
,
s(s + a)
a>0
(30)
(31)
When phase variables with x1 e are used, the state equations are
x1 = x2 ,
x2 = Kx1 ax2
(32)
R(s) +
1
1
1
p x2 + p x2 = xT Px
2 1 1 2 2 2
2
E(s)
G(s)
Y(s)
Conditions 13 ensure that V(x) is positive definite. Condition 4 is satisfied when V(x) is positive definite (i.e., it is
closed) in the entire state space. When V(x) goes to infinity as
363
(33)
364
p1 p2 K
0
T
2
= xT p p K
x = x Nx
1
2
ap2
2
(34)
(35)
(p1 p2 K)2
0
4
(36)
(41)
0
A=
a0
1
a1
(42)
p12
p22 ,
p11
P= p
12
2
N=
0
0
0
(43)
Lure Method
Another design approach is the use of a procedure developed
by Lure (cited in Ref. 12). Consider the unexcited system represented by the state equation in which A is of order n:
x = Ax
(37)
(38)
1
=
,
a0
p22
1
=
,
a0 a1
(44)
p11
a + a21
= 0
a0 a1
The necessary conditions for P to be positive definite are obtained by applying the Sylvester theorem:
p11 =
a0 + a21
> 0,
a0 a1
a0
>0
a20 a21
(45)
(39)
(40)
The second equation requires that a0 0. Using this condition in the first equation produces the necessary condition
a1 0. These are obviously the same conditions that are obtained from the RouthHurwitz conditions for stability, as
shown in a later section.
Krasovskii (cited in Ref. 4) has shown that a similar approach may be used with a nonlinear system. However, only
sufficient conditions for local asymptotic stability in the vicinity of an equilibrium point may be obtained. For the Lyapu-
j
s = jd
s = f + jg
3
s = b + jc
2
s=a
s=0
3
1
2*
s = b jc
s = jd
s plane
s=e
6*
s = f jg
4*
(x)
nov function V(x) xTPx, the time derivative is V
xTNx, where
N = JxT P + PJx
(46)
aw sw + aw1 sw1 + + a1 s + a0
P(s)
R(s) =
R(s)
Q(s)
sn + bn1sn1 + + b1 s + b0
P(s)
(47)
R(s)
=
(s s1 )(s s2 ) (s sk ) (s sn )
Y (s) =
where R(s) is the system input and the values of the poles s1,
s2, . . ., sn may be real or occur in conjugate complex pairs.
The stability and the corresponding response of a system
can be determined from the locations of the poles of the response transform Y(s) in the s plane. The possible positions of
the poles are shown in Fig. 6, and the responses associated
with the poles are given in Table 1. These poles are the roots
of the characteristic equation Q(s) 0.
1
22*
3
44*
5
66*
Form of
Response
Ae at
Ae bt sin(ct )
A
A sin(dt )
Ae et
Ae ft sin(gt )
Characteristics
Damped exponential
Exponentially damped sinusoid
Constant
Constant-sinusoid
Increasing exponential (unstable)
Exponentially increasing sinusoid
(unstable)
(48)
365
sn | b n
bn2
bn4
bn6
| bn1
bn3
bn5
bn7
| c1
c2
c3
| d1
d2
n1
n2
n3
s1 | j 1
s0 | k 1
(49)
366
The constants c1, c2, c3, and so on, in the third row are evaluated as follows:
bn1 bn2 bn bn3
bn1
(50)
bn bn5
b b
c2 = n1 n4
bn1
(51)
c1 =
This pattern is continued until the rest of the cs are all equal
to zero. Then the d row is formed by using the sn1 and sn2
rows. These contants are
c1 bn3 bn1 c2
c1
(52)
bn1 c3
c b
d2 = 1 n5
c1
(53)
d1 =
(54)
10
152
s4 |
72
240
s3 |
62
88
s2 |
70.6
240
s1 |
122.6
s0 |
240
s |
s |
5
(57)
(58)
(55)
s4 |
2
s5 |
(56)
s4 |
s3 |
s2 |
s1 |
s |
(59)
There are two changes of sign; therefore there are two roots
of x in the right-half s plane. The number of roots of s with
positive real parts is also two. This method does not work
when the coefficients of Q(s) and of Q(x) are identical.
Method 2
Q1 (s) = Q(s)(s + 1) = s5 + 2s4 + 3s3 + 4s2 + 7s + 5
(60)
The reader may obtain the Routhian array which has two
changes of sign in the first column, so there are two zeros of
Q(s) with positive real parts. Thus, the same result is obtained by both methods. An additional method is described in
Ref. 20.
Theorem 3. When all the coefficients of one entire row are
zero, the procedure is as follows:
1. The auxiliary equation can be formed from the preceding row, as shown below.
367
(66)
s4 |
11
s3 |
18
s |
s2 |
18
s2 |
s1 |
(61)
18
(62)
The presence of a zero row for s1 indicates that there are roots
that are the negatives of each other. The next step is to form
the auxiliary equation from the preceding row, which is the
s2 row. The highest power of s is s2, and only even powers of s
appear. Therefore the auxiliary equation is s2 9 0. The
roots of this equation are s j3. These are also roots of the
original equation. The presence of imaginary roots indicates
that the output includes a sinusoidally oscillating component.
For a sinusoidal input with the frequency corresponding to
the imaginary root, the response is unbounded. Thus, the system with imaginary roots is considered unstable.
To complete the Routhian array, the auxiliary equation is
differentiated and is
2s + 0 = 0
(63)
s1 |
s0 |
(64)
(65)
K(s + a1 ) (s + ah ) (s + aw )
sm (s + b1 )(s + b2 ) (s + bc ) (s + bu )
(67)
w
K h=1 (s zh )
K(s z1 ) (s zw )
=
u
sm (s p1 ) (s pu )
sm c=1 (s pc )
(68)
where indicates a product of terms. The degree of the numerator is w and that of the denominator is m u n. The
zh are the zeros and the pc are the poles of G(s)H(s). In the
form shown in Eq. (68), with the coefficients of s all equal to
unity, the K is defined as the loop sensitivity.
The underlying principle of the root locus is that the poles
of the control ratio C(s)/R(s) are related to the zeros and poles
of the open-loop transfer function G(s)H(s) and to the loop
R(s) +
E(s)
G(s)
C(s)
B(s)
H(s)
Figure 7. Block diagram of a feedback system.
368
N1 (s)
,
D1 (s)
H(s) =
N2 (s)
D2 (s)
(69)
Thus
C(s)
A(s)
G(s)
= M(s) =
=
R(s)
B(s)
1 + G(s)H(s)
(70)
|sm | |s p1 | |s pu |
|s z1 | |s zw |
(79)
(80)
where
D D + N1 N2
B(s) 1 + G(s)H(s) = 1 2
D1 D2
(72)
From Eqs. (71) and (72) it is seen that the zeros of B(s) are
equal to the poles of M(s), and they determine the form of the
closed-loop system response. The degree of the numerator of
B(s) is equal to m u. Therefore B(s) has n m u finite
zeros. The roots of B(s) 0, which is the characteristic equation of the closed-loop system, must satisfy the equation
B(s) 1 + G(s)H(s) = 0
(73)
The procedure for obtaining the root locus is to first put the
open-loop transfer function G(s)H(s) into the form shown in
Eq. (68). The poles and zeros are then plotted in the s
j plane. Then the angle condition given in Eq. (80), for both
K 0 and K 0, is used to obtain the complete root locus.
Then the root locus can be calibrated in terms of the loop
sensitivity K by using the magnitude condition given in Eq.
(79). The dominant roots are those that contribute the most to
the overshoot and the settling time of the closed-loop system
response. Selection of the dominant complex roots on the root
locus is based on the characteristics of a simple second-order
system with a unit step input (8). The peak overshoot Mp is
related to the damping ratio by
M p = 1 + exp
1 2
(81)
K(s z1 ) (s zw )
= 1
sm (s p1 ) (s pu )
(74)
|G(s)H(s)| = 1
h = 0, 1, 2, . . .
(76)
For K 0:
Magnitude condition:
|G(s)H(s)| = 1
(77)
h = 0, 1, 2, . . .
(82)
(83)
Angle condition:
\G(s)H(s) = h360,
= cos1
(75)
Angle condition:
\G(s)H(s) = (1 + 2h)180,
(78)
real poles and zeros to the right is an even number (including zero).
Rule 3. The root locus starts (K 0) at the open-loop poles
and terminates (K ) at the open-loop zeros or at
infinity.
Rule 4. The angles of the asymptotes of the root locus
branches that end at infinity are determined by a
K1
s(s/25 + 1)(s2 /2600 + s/26 + 1)
65,000K1
=
s(s + 25)(s2 + 100s + 2600)
K
= 4
s + 125s3 + 5,000s2 + 65,000s
(84)
G(s) =
n
o =
c=1 Re(pc )
j=1
pj =
n
where K 65,000K1.
Specification: Find C(s)/R(s) with 0.5 for the dominant
roots (roots closest to the imaginary axis).
1. The poles of G(s)/H(s) are plotted on the s plane in Fig.
8. The values of the poles are: s 0, 25, 50 j10,
50 j10. The system is completely unstable for K
0. Therefore, this example is designed only for the condition K 0.
2. There are four branches of the root locus. Since there
are no zeros, all branches end at infinity for K .
3. The locus exists on the real axis between 0 and 25.
4. The angles of the asymptotes are
=
(1 + 2h)180
= 45 , 135
4
0 25 50 50
= 31.25
4
jw
K
K
s3
55.9 + j18
50 + j10
= 0.5
s1
K=0
K=0
o
50 j10
25
sa
60
K=0
s2
K=0
s4
(86)
j=1
can be used to find one real or two complex roots. Factoring known roots from the characteristic equation can
also simplify the work of finding the remaining roots.
s plane
6.6 + j11.4
rj
(87)
(85)
369
G(s)H(s) =
65, 000K1
s(s + 25)(s2 + 100s + 2600)
370
14. With a unit step input, R(s) 1/s, the output response
can be obtained from a CAD program (8,21):
dK
= 4s3 + 375s2 + 10,200s + 65,000 = 0
ds
(90)
C(s)
65,000K1
FREQUENCY RESPONSE
= 4
(88)
3
R(s)
s + 125s + 5,100s2 + 65,000s + 65,000K1
The objective in using feedback is threefold. The first purpose
The Routhian array for the denominator of C(s)/R(s), is to achieve stability and/or a specified degree of stability
which is the characteristic polynomial, is
margin. The second goal is to reduce the sensitivity to parameter variation. The third goal is disturbance rejection. The
s4 |
1
5100
65,000K1
previous section introduces the root locus method of design.
The closed-loop poles are assigned in the left-half plane so
s3 |
1
520
(After division by 125)
that the performance is stable, and they are located on the
1
14.2K1 (After division by 4580)
s2 |
root locus in positions that best achieve the desired performance criteria. This section introduces the frequency res1 | 520 14.2K1
sponse methods that ensure closed-loop system stability by
0
s |
14.2K1
application of the Nyquist criterion and meet desired figures
of merit. The gain of the forward transfer function is the adPure imaginary roots exist when the s1 row is zero.
justable parameter which is used in the design. The feedback
This occurs when K1 520/14.2 36.6. The auxiliary
control system is shown in Fig. 7. For a linear time-invariant
2
equation is s 14.2K1 0. The imaginary roots are
(LTI) system, the open-loop transfer function in the s domain
s j14.2K1 j22.8.
is shown in Eq. (68), and the closed-loop transfer function is
9. Additional points on the root locus are found by locat- represented by Eqs. (70) and (72).
ing points that satisfy the angle condition
The forward frequency transfer function is written in the
generalized
form:
\s + \(s + 25) + \(s + 50 j10)
+ \(s + 50 + j10) = (1 + 2m)180
(89)
G( j) =
=
(92)
(93)
Re[G( j)] = 0
270
Type 3
0+
K0
1 + j0
180
=+
0+
=0
Type 2
B(s) =
0+
Type 1
90
Figure 9. A summary of direct polar plots of different types of
systems.
D1 D2 + N1 N2
=0
D1 D2
(94)
Type 0
371
P(s)
(s Z1 )(s Z2 ) (s Zn )
=
(s p1 )(s p2 ) (s pn )
Q(s)
(95)
where Z1, Z2, . . ., Zn are the zeros and p1, p2, . . ., pn are the
poles. The poles pi are the same as the poles of the open-loop
transfer function G(s)H(s) and include the s term for which
p 0, if it is present.
372
O
Q
s plane
Z1
p2
(s Z1)
(s p
2)
(s)
Z4
p4
p1
(s
p3
(s Z3)
Z3
)
p3
(s Z2)
p3
Q
Z2
(97)
For a stable system, B(s) can have no zeros ZR in the righthalf s plane. It is therefore concluded that, for a stable system, the net number of rotations of B(s) about the origin must
be counterclockwise and equal to the number of poles PR that
lie in the right-half plane. In other words, if B(s) experiences
a net clockwise rotation (i.e., if N is negative), this indicates
that ZR PR, where PR 0, and thus the closed-loop system
is unstable. If there are zero net rotations, then ZR PR and
the system may or may not be stable, according to whether
PR 0 or PR 0.
j
s=+j
O
s = re j
s plane
Q
+
(96)
s=j
Figure 11. The contour that encloses the entire right-half s plane.
270
373
270
1 + j0
180
0
1 + j0
B(s)
G(s) H(s)
90
90
(a)
(b)
K1
s(1 + T1 s)(1 + T2 s)
(98)
K
K1
K
K
= 1j = 1 e j = 1 e j
s
e
(99)
= 0
1
2
7 6
j0 +
s = e j
5
4
3
12
Q
j0
1 + j0
B(s)
G(s)H(s)
j
(a)
=+
=
= 0+
7
(b)
Figure 13. (a) The contour Q which encircles the right-half s plane.
(b) Complete plot for Eq. (98).
374
Km
Km
Km
Km
= m jm = m e jm = m e jm
sm
e
(100)
G(s)H(s) =
1 0
G(s)H(s)
=+
B(s)
Radial
line
PR = 0
N=0
ZR = 0
Stable
system
1 + j0
dB
(102)
(101)
375
Phase
margin
(+)
Lm G ( j ) (dB)
Lm (dB)
1
a
1
0 dB
(+)
360
270
Gain
margin (+)
0 Angle
90
180
c
Gain
margin, Lma
(+)
90
135
180
G( j)
225
270
Phase
margin,
(+)
=
(a)
(b)
(c)
Figure 15. Frequency response plots: (a) Polar plot, (b) Nichols plot, and (c) Bode plots of
G( j).
is stable. The log-magnitudeangle contour for a non-minimum-phase system does not close; thus it is difficult to determine the value of N. For these cases the polar plot is easier
to use to determine stability.
It is not necessary for minimum-phase systems to obtain
the complete log-magnitudeangle contour to determine stability. Only that portion of the contour is drawn representing
G( j)H( j) for the range of values 0 . The stability
is then determined from the position of the curve of
G( j)H( j) relative to the (0 dB, 180) point. In other
words, the curve is traced in the direction of increasing frequencythat is, walking along the curve in the direction of
increasing frequency. The system is stable if the (0 dB, 180)
point is to the right of the curve. This is a simplified rule of
thumb which is based on Nyquists stability criterion for a
minimum-phase system.
A conditionally stable system is one in which the curve
crosses the 180 axis at more than one point. Figure 17
shows the transfer-function plot for such a system with two
stable and two unstable regions. The gain determines
whether the system is stable or unstable.
Gain Margin and Phase Margin Stability from the Nichols Plot
The absolute stability of an LTI closed-loop system can be
determined by applying the Nyquist stability theorem, using
the representation of the open-loop transfer function G( j) as
a polar plot or as Nichols plot (log magnitude versus angle).
Some measures of degree of stability can be expressed in
terms of gain margin and phase margin. The following quantities are used to express these stability measures:
Gain Crossover. This is the point on the plot of the transfer function at which the magnitude of G( j) is unity
[Lm G( j) 0 dB]. The frequency at gain crossover is
called the phase-margin frequency .
Phase Margin. This is 180 plus the negative trigonometrically considered angle of the transfer function at the
gain-crossover point. It is designated as the angle ,
which can be expressed as 180 , where
G( j ) is negative.
Phase Crossover. This is the point on the plot of the transfer function at which the phase angle is 180. The fre-
This portion of
contour represents
the semicircle of
radius of contour Q
= 0+
= 0
0+
+ dB
+ dB
0 dB
270
180
90
+90
+180
+270
dB
= +180
G(s) Contour dB
This portion of contour represents the semicircle
of infinite radius r of contour Q
=
Figure 16. The log-magnitudeangle contour for
the minimum-phase system of Eq. (98).
(a)
Lm (dB)
Lm (dB)
0
180
Lm (dB)
Lm (dB)
376
180
180
(b)
(c)
180
(d)
(103)
(104)
The Bode plots consist of two components. One plot is for the
log magnitude [Lm G( j) 20 log G( j)] versus log , and
the second plot is for the angle of G( j) versus log . The
Bode plots are often used to represent the open-loop transfer
function. The log magnitude of G( j) converts the operations
of multiplication and divisoin to addition and subtraction, respectively. Each factor of a transfer function [see Eq. (91)],
plotted versus log , has a distinctive characteristic which is
readily recognized.
Constant Km. The Lm Km is a constant which is independent of frequency. Thus, it is a horizontal line and
serves to raise or lower Lm G( j), when it is larger than
or smaller than unity, respectively.
j Factor. The Lm G( j) has a positive slope of 20 dB per
decade when this factor appears in the numerator. Its
slope is negative when it appears in the denominator of
G( j). The transfer function type is related to the low
frequency slope of the Lm plot. A zero slope indicates a
Type 0 system, a slope of 20 dB/decade indicates a
Type 1 system, and so on.
1 jT Factor. The Lm(1 jT) has a corner frequency
at cf 1/T. At frequencies below cf the asymptote is
the 0 dB line. Above cf the asymptote is a straight line
that passes through 1/T at zero dB, with a slope of
20 dB/decade. When (1 jT) appears in the numerator of G( j), the slope of the asymptote is positive.
When it appears in the denominator, the slope of the
asymptote is negative.
Quadratic Factor [1 2j/n ( j/n)2]. The log magnitude of the quadratic factor has a corner frequency at
cf n. At frequencies below cf the asymptote is the 0
dB line. Above cf the asymptote is a straight line that
passes through cf at 0 dB with a slope of 40 dB/decade.
When the quadratic factor is in the denominator, the
slope of the asymptote is negative. When it appears in
the numerator, the slope of the asymptote is positive.
Because this quadratic factor has the damping ratio
as an additional variable, there is a family of log magnitude plots which depend on the value of .
The gain margin and the phase margin angle can be determined from the Bode plots, as shown in Fig. 15(c). Adjusting
the gain results in raising or lowering the log magnitude
curve, without changing the angle curve of G( j). This permits the designer to change the phase margin frequency ,
the phase margin angle , and the gain margin. For a stable
closed-loop system, it is necessary for both the phase margin
angle and the gain margin to be positive.
Experimental Determination of Transfer Function
The log-magnitudephase-angle diagram is of great value
when the mathematical expression for the transfer function
1
,
2 1 2
m = n 1 2 2
(105)
M p = 1 + exp
,
d = n 1 2
(106)
1 2
377
4
,
n
tp =
n 1 2
(107)
When G( j) is plotted in a polar (Nyquist) plot, there exists a family of constant M closed-loop contours, where M
C( j)/R( j). These contours are a set of circles with specified
centers and radii (8):
x0 =
M2
,
M2 1
y0 = 0,
M
r0 = 2
M 1
(108)
378
379
Limiter
r
kr
uc
Plant = (A, B)
kx
Figure 18. Current control system with open-loop limiter.
CONTROL CONCEPT
Figures 1821 show examples of nonlinear control system architectures for linear systems with control constraints.
Figure 18 shows the typical nonlinear control system architecture currently used in (flight control) practice: A simple
open-loop limiter is used to restrict the reference signal and
help mitigate the adverse effects of actuator saturationcaused windup. Figure 19 is a general nonlinear feedback
control system architecture where the nonlinear controller,
G, computes the commanded control signal based on the current plant state and the current, and possibly past, values of
the reference signal. The controller is cognizant of the saturation level of the downstream actuator which is located at the
plant input. Hence, the controller-generated signal, uc, may
attain its limits, and by doing so the plant is driven to its
full capacity. Figure 20 shows a specific example of a general
nonlinear controllers architecture. A nonlinear dual-loop control system is envisaged where the controllers nonlinearity is
confined to a feedback limiter, N, which performs a nonlinear
scaling of the reference signal. The inner loop controller is
linear and the nonlinear element N is contained in the outer
loop, as illustrated in Fig. 20. Furthermore, the inner-loop linear controller is the result of a linear design, wherein it is
assumed that the actuator operates in the linear regime.
Hence, the inner loop is a linear control system. The nonlinear element N in the outer loop is a (pilot) command limiter
which employs feedback. The nonlinear element N scales the
reference signal such that the commanded control signal uc
does not violate the actuator imposed control constraints in
the inner loop: The feedback (pilot) command limiter generates a modified reference signal r which drives the inner loop
in such a way that saturation in the downstream actuator is
precluded, and strictly linear action ensues in the inner loop.
The outer loops nonlinearity N renders transparent the actuators nonlinearity in the inner loop, and therefore the feedback control system shown in Fig. 20 is equivalent to the simpler feedback control system of Fig. 21, where the inner loop
is linear. In conclusion, in Figs. 20 and 21 the linear inner
loop controller is designed to meet small signal tracking and
stability specifications, and the nonlinear element N in the
outer loop is designed to yield good tracking performance and
uc
Plant = (A, B)
x(0) = x0 ,
k = 0, 1, . . ., N 1
x
Figure 19. Nonlinear control system.
380
kr
uc
Plant = (A, B)
x
kx
y(k + 1) = Cx(k + 1)
u (i) = ki x x0 + ki r r1 ,
J(u) =
N
[Q(r(k + 1) y(k + 1))2 + Ru2 (k)]
k=0
The Q and R weights establish the tracking performance/control effort tradeoff in the inner (linear) loop and determine
the small signal performance of the control system. Fullstate feedback is assumed.
For a given reference sequence r, LQ optimal control returns the optimal control time history u* [u*0 , u*1 , . . .,
u*N1], obtained within each RHW as a linear function of the
complete reference vector r [r1, r2, . . ., r N], and the initial
plant state, x0, where r1 is the currently commanded reference
signal. Now, polynomial extrapolation and interpolation
yields r2, . . ., rN, the reference signals prediction, linear in
the data. Specifically, if, for example, a simple ZOH extrapolation strategy is used, it is shown in Ref. 40 that the predicted
reference vector [r2, . . ., r N] is linear in r1; that is, r is linear
in r1, which ultimately yields the optimal control sequence
[u*0 , u*1 , . . ., u*N1] linear in r1, the current pilot-demanded
reference signal, and x0. Note that r1 is the actual pilot de-
i = 0, 1, . . ., N 111.1
kr k0 r
so that
u (0) = kx x0 + kr r1
(110)
Both kx and kr are provided by the solution of the finite horizon LQ optimal control problem. The row vector kTx Rn corresponds to the gain derived from the solution, over the finite
optimization horizon, of the Riccati difference equation; this
same Riccati equation is also associated with the solution of
the finite horizon regulation problem. Moreover, one can deviate from the optimal tracking control solution in the RHW
and instead, once kx has been determined, choose the gain kr
such that asymptotic tracking of a step reference command r
is enforced. Of course, if integral action is used to obtain a
type-one system, then asymptotic tracking is achieved for
any kr.
kr
(109)
uc = u
Plant = (A, B)
x
kx
i = 0, 1, . . ., N 1
1
(u ui1 )
T i
1
(k x + ki r r1 ki1 x x0 ki1 r r1 )
=
T i x 0
1
[(ki x ki1 x )x0 + (ki r ki1 r )r1 ]
=
T
u i
r1
min
1iN1
,
u1 kx x0 + T u max
kr
(111)
where u*1 is the control signal from the last window. Hence,
actuator constraints are easily transformed into constraints
on the current reference signal. Saturation avoidance is guaranteed, provided that the reference signal at time now, r1,
satisfies the above inequalities.
Actuator displacement and rate constraints impose constraints on the inner closed-loops exogenous reference signal.
Thus, an additional outer control loop is employed which performs a nonlinear modification of the reference signal from
the pilot, such that the downstream actuator constraints in
the inner loop are satisfied. Now, if r1 causes the violation of
any of the constraints, it may be optimally scaled such that
the modified reference signal, r1, satisfies all constraints;
here, optimally scaled means that in the outer loop r1 is
chosen so that, for example, r1 r1 is minimized, subject to
r1 satisfying the actuator constraints.
In accordance with the receding-horizon modus operandi,
at each time step the optimization window is shifted forward
in time, a new state is attained, and upon receipt of a new
reference signal from the pilot the reference signal extrapola-
381
tion algorithm is reapplied and the open-loop LQ optimal control problem is solved from the beginning. Hence, r1 (or r1) is
the only reference value actually tracked; and within each
RHW, u*0 is the only control which is actually applied to the
system, so that feedback action is achieved. Now, the windows local time instant 0 corresponds to the current time
k. Thus, at time k, the optimal control signal u*k satisfies
u (k) = kx x(k) + kr r(k + 1)
and since the nonlinear element N in the outer loop causes
the exogeneous reference signal rk1 to be replaced by rk1,
which is then sent to the inner loop, the actual optimal control signal is
u (k) = kx x(k) + kr r (k + 1)
(112)
382
systems. The discrete-time reference governor (DTRG) developed in Ref. 32 is a nonlinear dynamical element, that is, it
is a first-order lag filter with a variable bandwidth parameter, [0, 1], which scales the reference signals increments
so that the controlled systems constraints are not violated.
These constraints characterize the maximal output admissible set. By forming a modified reference signal such that the
discrete-time systems state update satisfies these constraints, both saturation avoidance and BIBO stability are
enforced: BIBO stability also requires that the maximal output admissible set be bounded. In the sequel, the DTRGbased approach to tracking control is discussed. More importantly, the construction of the maximal output admissible set
is presented. This is a critical step in the development of more
general BIBO tracking control systems, in both the discretetime and the continuous-time settings.
The LTI discrete-time system
x(0) = x n ,
u(t) ,
(114)
t = 0, 1, 2, . . . (= I + )
(115)
Let Acl A Bkx and Ccl C Dkx. Then, Eqs. (114) and
(115) become
x(t + 1) = Acl x(t),
x(0) = x,
t I+
(116)
Hence, the saturation avoidance problem has been transformed into a feasibility problem consisting of an unforced
LTI discrete-time system with an output constraint. In Ref.
31 the maximal output admissible set associated with system
(116) is defined as the set of all initial states x n such that
the unforced closed-loop linear systems response does not violate the system output constraints for all t I. Thus, the
maximal output admissible set is the largest set of initial
states, x n, that is positively invariant w.r.t. system [see
Eq. (116)]. Evidently, the maximal output admissible set is
O (Acl , Ccl , Y ) = {x n : Ccl Atcl x Y
t I+ }
(117)
Example 1. The Output Constraint Set Y 0. The maximal output admissible set is then the subspace of unobservable states that corresponds to the pair (Acl, Ccl), namely,
O = (N(Ccl ))A
cl
i = 1, 2, . . ., s}
(118)
(119)
Although Y is a polyhedron, Eqs. (117) and (119) each represent an infinite number of constraints. However, if there exists a finite t*i I for each inequality in (119) for which the
constraints associated with the ith inequality constraint are
inactive for t t*i , let t* 1 i s maxt*i . Then O(Acl, Ccl,
Y) is characterized by a finite number (st*) of inequality constraintsin this case, O(Acl, Ccl, Y) is said to be finitely determinedand determination of whether a particular initial
state vector, x, is an element of O(Acl, Ccl, Y) involves evaluation of a finite set of linear inequalities. Moreover, it may
transpire that the constraints associated with one or more of
the inequalities, f i, that define Y are inactive for all t I.
Thus, let S* denote the set of inequality constraints that are
active for some t I, namely, S* 1, 2, . . ., s. Then,
O(Acl, Ccl, Y) may be written as
O (Acl , Ccl , Y ) = {x n : f i (Ccl Atcl x) 0,
t = 0, 1, . . ., ti , and i S }
(120)
In the discrete-time case under consideration, sufficient conditions for the existence of a finite t*, namely, sufficient conditions for the finite determination of O, are as follows: (1) Acl
is asymptotically stable, (2) 0 int(Y), (3) Y is bounded, and
(4) the pair (Acl, Ccl) is observable. Here, int( ) denotes the
interior of the set.
The concept of maximal output admissible sets is only applicable to unforced systems. Thus, it is not directly applicable to the tracking problem without some modification. In Ref.
32 the concept of maximal output admissible sets is adapted
for use with the tracking problem by using a nonlinear element in the DTRG which has first-order dynamics; that is, a
first-order lag filter with a variable bandwidth parameter
[0, 1] is used to prefilter the exogenous reference signal. The
closed-loop system state vector is then augmented with the
prefilter state, namely, the modified reference signal, r. The
end result is an augmented system whose exogenous input
x(t) n , u(t) m
(121)
(122)
(123)
(124)
where (r(t), xg(t)) [0, 1]; and in Eq. (119) r(t) is replaced by
the filters output r(t) (the modified reference signal). Then,
the augmented state is
r
xg =
x
(125)
and the augmented system dynamics and the output constraints are given by
xg (t + 1) = Ag xg (t) + Bg (r(t), xg (t)) r(t) [I
y(t) = Cg xg (t) Y
0]xg (t)
(126)
where
Ag =
I
Bcl
0
,
Acl
I
Bg =
,
0
Ccl ]
(127)
The set of statically admissible states, Xs, for the closedloop system defined by Eq. (123), is the set of initial states for
which there exists a statically admissible reference signal
such that the ensuing trajectory does not violate the systems
state and control constraints for all time. That is,
Xs = {x n : r Rs
Cg = [Dcl
383
s.t.
y(t) Y
t I+ }
(128)
(129)
384
0]xg ) ,
i = 1, . . ., s
(130)
and f j ([Ho
0]xg ) , j = 1, . . ., s}
(131)
Now, Xgs is characterized by a finite set of inequality constraints. Moreover, the upper limit imposed on (r(t), xg(t))
[0, 1] by each inequality, so that xg(t 1) Xgs, is given by a
simple formula. Thus, if xg(0) Xgs, and at each time-step we
choose (r(t), xg(t)) [0, 1] such that it satisfies the minimum
of the upper limits imposed by all inequality constraints in
Eq. (131), then xg(t) Xgs for all t I.
The finite determination of the maximal statically admissible (invariant) set is exclusively an artifact of discrete-time
dynamics. The maximal output admissible invariant sets of
continuous-time systems are not polyhedral.
(138)
Y = {y : f i (y) 0, i = 1, 2}
(139)
(132)
(133)
acl = a + bkx
(134)
a
a
kx = cl
b
b
1
0
Ag =
9.995 104 0.999
1
Bg =
,
Cg = [1 3]
0
x = acl (x r )
(144)
H0 = d + c(1 ad )1 bd = 1.9985
(145)
(135)
where
and
a
u = kx x cl r
b
(143)
(142)
r (k)
xg (k) =
x(k)
and
1 u 1
(141)
The concept of a statically admissible set and the discretetime reference governor (DTRG) are illustrated. Specifically,
a tracking controller is synthesized for a constrained scalar
control system. The continuous-time control system is given
by
x(0) = xo ,
(140)
Combining the closed-loop system and reference governor dynamics results in the augmented second-order system
x = ax + bu,
(137)
(136)
Then,
O = {x : f i (Cg Akg xg ) 0, k = 0, . . ., ki , i S ,
f j ([H0
0]xg ) , j = 1, 2}
(146)
0, k 0, . . ., k*i , i S*) deals with saturation of the transient response, and the second set ( fj([H0, 0]xg) , j 1, 2)
deals with saturation in steady state. Now, Algorithm 3.2 of
Ref. 31 may be used to determine k*i and S*. In this case S*
is empty. That is, the inequalities associated with the transient saturations are inactive. This should be expected because the closed-loop system of Eqs. (137) and (138) is a stable
first-order (over damped) system. Thus, Eq. (146) becomes
O = { f j ([H0
0]xg ) , j = 1, 2}
(148)
where
min 1, 0.95 H0 r (k)
H0 (r(k) r (k))
1 (k) =
1,
and
min 1, 0.95 + H0 r (k)
H0 (r (k) r(k))
2 (k) =
(147)
385
(151)
While the above reference governor avoids saturation and affords BIBO stability, it is somewhat conservative in that it
restricts the feasible reference signal, r, such that at all
times it is always statically admissible. The BIBO stable
tracking controller developed in Refs. 4042 employs a static
nonlinearity N; it is nevertheless related to Gilberts DTRG
(32) in that it also includes static admissibility of the modified reference signal in the feasibility criteria. However, there
are certainly cases where a modified exogenous reference input that is not statically admissible over a finite time interval
would not necessarily result in saturation. Thus, as discussed
in the next section, the achievable tracking performance can
be enhanced, while at the same time the constraint violation
is avoided and BIBO stability is guaranteed.
STATIC ADMISSIBILITY AND INVARIANCE
The scalar constrained control system is revisited and the application of statically admissible and more general (maximal
output admissible) invariant sets to achieve saturation avoidance and BIBO stability enforcement is demonstrated. The
construction of these invariant sets is a crucial step in the
design of tracking control laws. Insights into the critical role
played by the above-mentioned invariant sets permit focusing
on design for tracking performance, while at the same time
guaranteeing the BIBO stability of the control system. More-
Control Concept 1: Choose r such that r r is minimized subject to the control constraints. Assuming kx 0,
acl 0, and b 0 results in the explicit (nonlinear) control
law
b
bkx
b
bkx
r
x+
r
x
if
a
a
a
a
cl
cl
cl
cl
bk
b
bkx
b
x
x
if
r
x
r (x, r) =
acl
acl
acl
acl
b
bkx
b
bkx
if
r
x+
a x+ a
a
a
cl
cl
cl
cl
(152)
and substituting Eqs. (152) into (135) results in the closedloop system
bkx
b
bkx
b
x+
r
x
acl (x r) if
a
a
a
a
cl
cl
cl
cl
bkx
b
if
r
x
x = ax + b
(153)
acl
acl
bkx
b
if
r
x+
ax b
acl
acl
The saturation avoidance control law (152) does not guarantee BIBO stability in the case of an open-loop unstable plant.
However, the desired BIBO stability can be obtained through
an additional invariance requirement.
A bounded set XI X x: x is invariant with respect
to the system given by Eqs. (132), (133), and (153) if and only
if on the boundary of XI, xx 0. Thus, if a bounded invariant
set is characterized as
!
a
XI = x : r such that 1 u = kx x cl r 1 and xx 0
b
then, by restricting x to x XI, BIBO stability is guaranteed.
First, consider the case of an open-loop unstable plant, a 0.
For (bkx /acl)x b/acl r (bkx /acl)x b/acl, the result is r
r, and x acl(x r). Hence, r x results in x 0, and r x
results in x 0. Second, for r (bkx /acl)x b/acl we have
x ax b. In this case x b/a results in x 0, and x
b/a results in x 0. Finally, for r (bkx /acl)x b/acl we
have x ax b. Therefore, x b/a results in x 0, and
x b/a results in x 0. This is summarized in Fig. 22, where
the directions of the arrows represent the sign of x in the
Cartesian product space, V, defined by
V = {v 2 : v = [r, x]T , r, x }
From Fig. 22 and the above discussion it is clear that if the
control law allows x b/a the system will diverge due to the
constraint on the control signal, u. Also, if the system ever
386
kx x
b
acl
acl
r=1
b
b
a
From Eq. (135) and the assumption that acl 0, this results in
x=r
b
a
1
kx
1
kx
b
a
a
b
acl
(154)
a
r (x, r) =
b
+
bk
bkcl
a x+
cl
bkx
b
bkx
b
x+
r
x
acl
acl
acl
acl
b
b
and + x
a
a
b
b
if x > and r
a
a
b
b
if x < + and r +
a
a
bkx
b
b
if r >
x
and x
acl
acl
a
bkx
b
b
if r
x+
and x +
acl
acl
a
(155)
if
b
acl
b
acl
bkx
b
bkx
b
x+
r
x
acl
acl
acl
acl
bkx
b
x
, if r > x
a
a
cl
cl
b
bkx
x+
, if r < x
a
a
cl
cl
b
b
if r = x and x
r = r
a
a
bkx
b
b
if r = x and x <
a x a
cl
cl
bkx
b
b
if r = x and x >
a x+ a
a
cl
(156)
(157)
cl
ax + b
ax b
x = 0
ax + b
ax b
if r > x
if r < x
b
b
x
a
a
b
if r = x, and x <
a
b
if r = x, and x >
a
if r = x, and
(158)
bkx
b
b
x
if r > x, and x
acl
acl
a
b
b
bkx
if r < x, and x +
a x+ a
a
cl
cl
b
b
(159)
r = r, r = x
if + x
a
a
b
b
if r > x, and x >
a
a
b
b
+
if r < x, and x < +
a
a
Now, x XI is enforced, the system will not become stuck
at x b/a, and BIBO stability is achieved.
The scalar controlled output variable y must track the exogeneous reference signal r. The point of departure is LQT.
Hence, a tracking linear control law is specified; the simplest
such control law is u kxx krr, where kr 1/cA1
cl b and
Acl A bkx. Hence,
0.5
DTRG
CC1
CC2
0.45
0.4
387
0.35
r(t) = h(t 0.5) h(t 1.5)
u = kx x
0.3
1
r
cA1
b
cl
x 0.25
x = Acl x
0.15
1
br
cA1
b
cl
0.1
0.05
0
0.5
1.5
2
2.5
Time (sec)
3.5
cA1
b
cl
r1
1
A1 br, rs r rs }
cA1
b cl
cl
BIBO stabilty enforcement requires the construction of certain invariant sets in the state space.
The maximal statically admissible set Xs is characterized
as follows: The rest state which corresponds to the constant
reference signal r is
xr =
1
A1
cl br,
cA1
b
cl
rs r rs
1 u 1
y = cx
1 + r kx x 1 + r
r(t) = 0.3h(t 0.5) 0.3h(t 1.5)
0.4
(160)
0.45
(161)
Now, fix r, rs r rs, and determine the largest set contained in the slab in Eq. (161) which is invariant under Eq.
(160); denote this maximal output admissible set by Xs(r). Finally, the maximal statically admissible set Xs is
DTRG
CC1
CC2
0.35
0.3
Xs = 1r1 Xs (r)
x 0.25
0.2
0.15
0.1
0.05
0
b1
kx A1
cl
0.5
1.5
Time (sec)
2.5
Remark. The solution of well-posed inner-loop optimal control problems automatically renders the pair (A, kx) observable. Hence, the set Xs will be bounded and BIBO stability
can be enforced.
In general, the maximal statically admissible set Xs is compact and convex. However, Xs is not polyhedral; that is, it is
388
0 1
0
x(t)
=
x(t) +
u(t)
2 1
2
It is now fairly easy to synthesize the outer-loop nonlinearity N, namely, the nonlinear control law r r (x, r). For
example, consider the Control Strategy 1: r is chosen to minimize r r, subject to the saturation avoidance constraint
1 kx x
provided that
y(t) = [1 0]x(t)
1 u(t) = [9
2.5]x(t) + 8r(t) 1
+ PAcl = Q
where Q is a somewhat arbitrary, real, symmetric, and positive definite matrix. The ellipsoid
1
T
Es = x | x Px
kx P1 kTx
is statically admissible. Furthermore, Xr Es, provided that
2
1 T
T
1 T
(cA1
cl b) (kxP kx )(b (Acl ) b).
1
r 1
cA1
b
cl
xT Px <
1
kx P1 kTx
xT Px =
1
kx P1 kTx
2
bT Px xT Qx
cA1
b
cl
0.5
1.5
1
X2
0.5
X2
0.5
0.5
1
0.5
0.5
1.5
1.5
X1
0.5
0
X1
(a)
(b)
Figure 25. (a) Maximal statically admissible set. (b) Comparison of Xs and the suboptimal
ellipse.
0.5
1.5
389
390
39. N. J. Krikelis and S. K. Barkas, Design of tracking systems subject to actuator saturation and integrator windup, Int. J. Control,
39: 667683, 1984.
40. R. B. Miller, A New Approach to Manual Tracking Flight Control
with Amplitude and Rate Constrained Dynamic Actuators, Ph.D.
dissertation, Air Force Institute of Technology, March 1997.
41. R. B. Miller and M. Pachter, Manual tracking control with amplitude and rate constrained actuators, in Proc. 1996 Conf. Decision
Control, December 1996, pp. 31593164.
42. R. B. Miller and M. Pachter, Manual control with saturating actuators, in Proc. 1997 Eur. Control Conf., July 14, 1997.
43. R. B. Miller and M. Pachter, Maneuvering flight control with actuator constraints, AIAA J. Guidance, Control Dynamics, 20 (4):
729734, 1997.
44. C. Moussas and G. Bitsoris, Adaptive constrained control subject
to both input-amplitude and input-velocity constraints, in Proc.
IEEE Int. Conf. Syst., Man Cybern., Vol. 4, 1993, pp. 601606.
45. M. Pachter and D. H. Jacobson, Observability with a conic observation set, IEEE Trans. Autom. Control, AC-24: 632633, 1979.
46. M. Pachter, P. R. Chandler, and M. Mears, Control reconfiguration with actuator rate saturation, in Proc. Amer. Control Conf.,
June 1995, pp. 34953499.
47. M. Pachter, P. R. Chandler, and L. Smith, Velocity vector roll
control, in 1996 AIAA Guidance, Navigation Control Conf., San
Diego, CA, July 1996, AIAA paper No. 96-3867.
48. M. Pachter, P. R. Chandler, and L. Smith, Maneuvering flight
control, in 1997 Amer. Control Conf., Albuquerque, NM, June
46, 1997.
49. M. Pachter and R. B. Miller, Manual flight control with saturating actuators, IEEE Control Syst. Magazine, 18: 1019, 1997.
50. V. M. Popov, The Hyperstability of Control Systems, New York:
Springer, 1973.
51. A. A. Rodriguez and Y. Wang, Saturation Prevention Strategies
for an Unstable Bank-To-Turn (BTT) Missile: Full Information,
Ph.D. dissertation, Department of Electrical Engineering, Arizona State University, Tempe, AZ.
52. A. G. Tsirukis and M. Morari, Controller design with actuator
constraints, in Proc. Conf. Decision Control, 1992, pp. 26232628.
J. J. DAZZO
M. PACHTER
Air Force Institute of Technology
(AFIT/ENG)
529
(1)
where L : Rp R1 represents some loss function to be minimized, represents the p-dimensional vector of adjustable parameters, and C Rp represents a constraint set defining the
allowable values for the parameters . (Note that a maximization problem can be trivially recast as the minimization problem in Eq. (1) by applying a minus sign to the objective function.) Our focus in this article will be problems for which
represents a vector of continuous parameters; this is in contrast to discrete problems such as how many items X do we
need to optimize performance? Discrete optimization is a
large subject unto itself, and will not be considered in any
detail here (see, e.g., Ref. 1 for a detailed discussion of this
subject from a deterministic perspective). Further, we are inJ. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
530
L
1
.
.
g ( )
.
L
p
Then for local unconstrained optimization (i.e., C Rp), a
necessary condition for optimization when L is a continuously
differentiable nonlinear function is that * satisfies
g ( ) = 0
(2)
(3)
531
(4)
532
changes, this is an indication that the iterate is near the solution, while if signs are not changing we have an indication
that the iterate is far from the solution. This forms the basis
for an adaptive choice of the gain ak, whereby a larger gain is
used if there are no sign changes and a smaller gain is used
if the signs change frequently. Kesten (32) established a.s.
convergence with such a scheme. A multivariate extension of
the Kesten idea is described in Delyon and Juditsky (33), including theoretical justification of the extension via a.s. convergence of the iterates.
Ljung and Soderstrom (34, Sec. 2.4, Sec. 3.4, and Chap. 4)
and Wei (35) discuss stochastic analogs of the Newton
Raphson search in the context of parameter estimation for
particular (possibly linear) models. [These represent extensions of the first paper on adaptive Hessian estimates for scalar problems (100).] In so doing, Refs. 34 and 35 demonstrate,
via the method of minimizing prediction error, that a batch
version of the problem of finding a * satisfying Eq. (2) (where
all the data are processed simultaneously) can be converted
to the recursive form (where the data are processed one at a
time) using only the current (instantaneous) input. The scalar
gain ak is then replaced in their formulation by a matrix that
approximates the (unknown) true inverse Hessian matrix corresponding to the current data points contribution to the
loss function.
Ruppert (36) describes an approach where the Hessian is
estimated by taking finite differences of a gradient measurement. The gradient used by Ruppert differs slightly from the
standard RM gradient in Eq. (3) in that he converts the basic
problem from one of minimizing L() to one of minimizing
L/2 (note that this yields the same * when there is a
unique minimum). Spall (102) presents a more efficient approach to general Hessian estimation based on the simultaneous perturbation idea discussed below.
Iterate Averaging. An important and relatively recent development in SA is the concept of iterate averaging. Like
many good ideas, this one is simple and, in some problems,
can be effective. The idea was jointly described by Polyak and
Juditsky (37) and Ruppert (Ref. 12, based on a 1988 internal
technical report). There are several variations, but the basic
idea is to replace k as our current best estimate of * after
k iterations with the average
k (k + 1)1
k
(5)
j=0
533
convergence that previously was possible only with knowledge of the Hessian H(*), as discussed under Adaptive estimation and second-order algorithms above. This result is
available for any valid gain satisfying the above-mentioned
additional condition. Hence, in principle, iterate averaging
greatly reduces one of the traditional banes of SA implementationnamely, choosing the gain sequence in some optimal
way.
Variations on the above basic iterate averaging formulation are readily available. One obvious variation is to not average in the first few iterations, but rather start the average
at some N 0 or else use only a sliding window of the last
n N (say) measurements. In practical finite-sample implementations, such modifications are likely to help, since the
first few iterations tend to produce the poorest estimates. In
the sliding-window approach, formal asymptotically optimal
normality can be retained if the window length grows with
time (see, e.g., Ref. 38). A further modification to the basic
approach is to use the averaged value k (together with k) in
a modified form of the RMSA iteration [instead of k alone on
the right-hand side of the basic form (3)]. This is referred to
as the feedback approach in Ref. 39, and can be shown to
sometimes yield further improvement.
In practice, however, the results on iterate averaging are
more mixed than the above would suggest. Numerical studies
by the author have shown that a well-chosen ak sequence will
yield results superior to that possible by averaging (see also
Ref. 40, p. 57, and Ref. 41 for some cautionary notes). Once
again, this appears to be a consequence of the finite-sample
properties of practical problems. More study is required to
understand the full capabilities of iterate averaging in
practice.
Time-Varying Loss Functions. A final generalization of the
RobbinsMonro recursion that we discuss is one where the
loss function (and corresponding gradient with respect to )
varies with k. This problem is treated in Refs. 42 and 43. The
basic idea is that, while the loss function may change shape
with k, it is assumed that the underlying minimum * is either constant for all k or fixed in the limit as k (even
though the loss function may change shape indefinitely). The
two references mentioned above show a.s. convergence in
these cases for the scalar- setting; the proofs have to be
changed from those commonly seen in SA to accommodate the
time-varying loss.
Figure 1 depicts a time-varying loss function Lk() for a
scalar parameter. This figure shows a case where the loss
function and the minimizing parameter value both change
with time, but one where the optimal parameter value converges to a limiting point * . This is a situation for which the
above-mentioned theory would apply.
Applications of RobbinsMonro Stochastic Approximation
RMSA has been applied in a large number of engineering
(and other) problems, often under another name. This subsection provides a summary of several applications together with
some references for further study.
Neural network (NN) training via the well-known backpropagation algorithm has long been recognized as an application of the RMSA algorithm, two of the publications discussing this connection being Ref. 9 and Ref. 44 (pp.
534
L1( )
L2( )
L3( )
Lk+1( )
Lk( )
Time index, k
The final application we mention here is in image restoration. Here the task is to recover a true image of a scene from
a recorded image of the scene where the recorded image is
typically corrupted by noise and/or otherwise degraded from
the original image. The essential reason for using RMSA
rather than more conventional signal processing and deconvolution techniques is the ability to adapt to nonlinearities in
the process. Some recent references that discuss implementations of RMSA in the context of image restoration are Refs.
50 and 51. Abreu et al. (50) are concerned with removing impulse noise that may be due, for example, to noisy sensor or
transmission errors in collecting the image data. They note
that nonlinear methods have often proven superior to linear
methods for this problem. Cha and Kassam (51) describe an
approach to image restoration based on a type of NN called
the radial basis function network (e.g., Ref. 6, Sec. 9.5), where
RMSA is used to train the weights of this form of function
approximator. Their use of the radial basis function network
is as a spatially invariant filter taking measurement data as
input and producing estimates of the pixel-by-pixel gray levels in the image.
STOCHASTIC APPROXIMATION WITH GRADIENT
APPROXIMATIONS BASED ON FUNCTION MEASUREMENTS
Introduction and Contrast of Gradient-Based
and Nongradient Algorithms
There has been a growing interest in stochastic optimization
algorithms that do not depend on direct gradient information
or measurements. Rather, these algorithms are based on an
approximation to the gradient formed from (generally noisy)
measurements of the loss function. This interest has been motivated, for example, by problems in the adaptive control and
statistical identification of complex systems, the optimization
of processes by large Monte Carlo simulations, the training of
recurrent neural networks, the recovery of images from noisy
sensor data, and the design of complex queuing and discreteevent systems.
Overall, such algorithms exhibit certain convergence properties of the RobbinsMonro gradient-based algorithms considered above while requiring only loss-function measurements. A main advantage of such algorithms is that they do
not require the detailed knowledge of the functional relationship between the parameters being adjusted (optimized) and
the loss function being minimized that is required in gradient-based algorithms [in particular, they do not need the
Y() term in the RM recursion of Eq. (3)]. Such a relationship
can be notoriously difficult to develop in some areas (e.g., nonlinear feedback controller design or system optimization via
large-scale simulation), while in other areas (such as highdimensional statistical parameter estimation) there may be
large computational saving in calculating a loss function
rather than a gradient. In contrast, the approaches based on
gradient approximations require only conversion of the basic
output measurements to sample values of the loss function,
which does not require full knowledge of the system input
output relationships. Examples of approximation-based methods using loss-function measurements only are given below;
such methods include, as an early prototype, the Kiefer
Wolfowitz finite-difference SA algorithm (52).
Because of the fundamentally different information needed
in implementing these gradient-based (RM) and gradient-free
535
(6)
where gk( k) is the estimate of the gradient L/ at the iterate k based on the above-mentioned measurements of the loss
function. Under appropriate conditions, the iteration in Eq.
(6) will converge to * in some stochastic sense, usually a.s.
(see, e.g., Ref. 53 or 15). Typical convergence conditions are
similar to those mentioned above for the RMSA algorithm.
The essential part of Eq. (6) is the gradient approximation
gk( k). We discuss below two forms that have attracted attention. We let y( ) denote a measurement of L( ) at a design
level represented by the dot [i.e., y( ) L( ) noise], and ck
be some (usually small) positive number [if the noise has
mean 0, then y( ) Q( ) as defined in the section on RMSA].
One-sided gradient approximations involve measurements
y( k) and y( k perturbation) while two-sided gradient approximations involve measurements of the form y(k perturbation). The two general forms of gradient approximations
are:
Finite Difference (FD) (52,58). Each component of k is perturbed one at a time, and corresponding measurements
y( ) are obtained; each component of the gradient estimate is formed by differencing the corresponding y( )
values and then dividing by a difference interval. This
is the standard approach to approximating gradient vectors and is motivated directly from the definition of a
gradient as a vector of p partial derivatives, each constructed as the limit of the ratio of a change in the function value over a corresponding change in one component of the argument vector. Typically, the ith
component of gk( k) (i 1, 2, . . ., p) for a two-sided FD
approximation is given by
gki (k ) =
y(k + ck ei ) y(k ck ei )
2ck
y(k + ck k ) y(k ck k )
2ck ki
where the distribution of the user-specified random perturbations for SP, k (k1, k2, . . ., kp)T, satisfies conditions mentioned in the next section.
The algorithm [Eq. (6)] with one of the gradient approximations will be referred to as FDSA or SPSA, as appropriate.
[An approach in the same spirit as SPSA, called random di-
536
as k
(7)
537
(8)
Normalized L ( )
1.2
160
65,920
0.8
SPSA
0.4
FDSA
20
40
Iterations
60
80
538
measurement is recommended as a check on algorithm behavior). The algorithm is implemented with two simple parallel
recursions: one for and one for the Hessian of L(). The recursion for is a stochastic analog of the well-known NewtonRaphson algorithm of deterministic optimization. The
recursion for the Hessian matrix is simply a recursive calculation of the sample mean of per-iteration Hessian estimates
that are formed using SP-type ideas.
x
(9)
P(energy = x) = cT exp
cb T
SIMULATED ANNEALING
Enew Ecurr
(10)
exp
cb T
539
level increases from 0, in contrast with the sudden large degradation in approaches such as SAN that use explicit decision
criteria based on loss-function measurements). The primary difficulty arises in the critical decision step 3, where
curr is changed or not changed according to the value of
L(new) L(curr) together with the random sampling associated with the Metropolis criterion. With noisy function measurements, the value of , now equal to y(new) y(curr) instead of L(new) L(curr), will be altered from its underlying
true value according to noise-free function measurements. In
fact, even a modest level of noise will frequently alter the sign
of , which is likely to change the decision regarding the acceptance or rejection of the new point new. The obvious means
by which one can attempt to cope with noise is to average
several function measurements y( ) at each of the values
curr and new when performing the comparison at step 3. However, this may dramatically increase the cost of optimization
(specifically, the total number of function evaluations required), since a large amount of averaging will often be required to effectively remove the noise, especially in the region
around a local or global minimum where the function may be
relatively flat. An alternative way is to alter the acceptance
criterion to accept the inherent errors that will be made with
the noisy measurements. Hence we replace the criteria 0
or 0 with or , where is the function measurement noise standard deviation (or, more likely, an estimate of it) and the number of multiples of the standard
deviation that will be tolerated. The rationale for such a
change is that, while we are willing to accept a new that temporarily increases the loss function (a basic aspect of SAN),
we are less willing to forgo a new that decreases the loss function. Changing the unconditional acceptance criterion from
0 to will allow for a greater number of cases where
L(new) L(curr) even though y(new) y(curr) due to the noise.
With 2, one can be sure (through the Chebyshev inequality of probability) that most such cases will be caught, although at the expense of letting some additional new values
that increase the loss function be accepted.
Evaluation of Simulated Annealing
Although many positive studies regarding SAN have been reported in the literature (e.g., Refs. 9294), the author is unaware of any formal theoretical analysis comparing SAN with
the gradient-free SA algorithms discussed in the preceding
two sections. (Note that some connections of SAN with SA
have been explored in Refs. 95 and 96. The basic idea there
is, beginning with the basic SA recursion using either gradients or function measurements, to add a Monte Carlo-generated Gaussian noise term bkWk to the recursion, where bk
0 and Wk is the Gaussian random vector. This term is similar
to SAN in that it will force the iterate out of local minima
under some conditions that are established in the cited papers.) However, Fabian (97), Chin (103), Styblinski and Tang
(62), and the author (together with several students in a graduate class he taught) have conducted some numerical comparisons. Fabian (97), Chin (103), and Styblinski and Tang (62)
compare different forms of SAN against several random and
deterministic search algorithms for problems with noise-free
measurements of the loss function. The SAN algorithms generally compared poorly with the competing algorithms in
these two papers. The authors studies involved four different
loss functions: p 2 and p 20 fourth-order polynomials,
540
and the Example 1 (p 2, fourth-order polynomial) and Example 6 (p 10, trigonometric) functions in Ref. 62. The studies of the author considered both noise-free and noisy measurements of the four loss functions and compared SAN with
simple random search mentioned in Ref. 7 (pp. 186189) and
the global version of SPSA outlined in Ref. 80. Although SAN
was sometimes superior to the random search method, the
global SPSA method was significantly more efficient than
SAN in all cases considered, and, even more dramatically,
was sometimes convergent when SAN seemed unable to obtain a solution anywhere near the true optimal (this was most
apparent in the Example 6 problem of Ref. 62, which had a
very large number of local minima, one of which always
seemed to form a trap for SAN). In order to have a fair comparison in performing these studies, the algorithm coefficients
(e.g., the gains for the SA algorithms and the decay rate and
other coefficients mentioned above for SAN) were tuned to approximately optimize performance for the competing algorithms.
Since one must be careful in drawing conclusions beyond
the specific cases in any numerical analysis, the above should
not indicate that SAN is always a poor performer. Rather,
these results should be a cautionary note in view of some of
the positive results reported elsewhere. Certainly, SAN does
have a role in the area of global optimization, as evidenced by
its popularity and positive performance in some challenging
problems. Furthermore, the above studies were for problems
with smooth (differentiable) loss functions, and SAN (unlike
most SA algorithms) has the ability to deal with nonsmooth
loss functions.
CONCLUDING REMARKS
This article has surveyed several important algorithms with
stochastic optimization. The focus was on algorithms within
the stochastic approximation and simulated annealing
classes. Broadly speaking the gradient-based (Robbins
Monro) SA algorithms are most appropriate when enough
prior information is available about the system so that it is
possible to determine the gradient of the observed loss function. It is expected that when such information is available,
this type of algorithm may be the most efficient means for
optimization within the stochastic setting, especially when
coupled with one of the algorithm acceleration techniques
such as iterate averaging or second-order methods. However,
it was also pointed out that in many applications, it is difficult
or impossible to obtain the gradient information needed for
the application of the gradient-based methods. Then one is
faced with carrying out the optimization using only the (possibly noisy) measurements of the loss function itself. In that
context, we emphasized three different approaches: finite-difference and simultaneous perturbation SA (both of which are
based on approximating the unknown gradient vector) and
simulated annealing. The FDSA method is the oldest and
best-known of these approaches, and can work in a variety of
applications. The newer SPSA approach will, however, usually be more efficient than FDSA, with the gain in efficiency
being approximately equal to the dimension of the vector
being optimized (i.e., the SPSA algorithm will use only onepth the number of function evaluations to obtain the same
level of statistical accuracy as FDSA). Unlike the SA algorithms, simulated annealing is focused on global (as opposed
541
542
JAMES C. SPALL
The Johns Hopkins University
STOCHASTIC SYSTEMS
STOCHASTIC SYSTEMS
Many control systems in practice are subject to imperfectly
known disturbances which may be taken as random. Such
disturbances have been ignored in the study of deterministic
control problems. In many control systems some unknown parameters occur. The system behavior depends on the parameters and the fact that the value of the parameters is unknown
makes the system unknown. Some crucial information concerning the system control is not available to the controller,
and this information should be learned during the systems
performance. The problem described is the problem of adaptive control. The adaptive control problem can be considered
the identification problem and the control problem. A solution
to an adaptive control problem will be understood as a solution to the identification problem and to the control problem.
We focus on control and adaptive control of continuous-time
linear stochastic systems. The theory of adaptive control of
continuous-time linear stochastic systems has been recently
developed as an important application of the stochastic control theory in engineering, biology, economics, finance, and
manufacturing systems. Continuous-time linear systems are
an important and commonly used class of systems (1,2). It is
assumed that the models evolve in continuous time rather
than discrete time because this assumption is natural for
many models and it is important for studying discrete time
models when the sampling rates are large and for analyzing
numerical round-off errors. Stochastic systems are described
by linear stochastic differential equations.
The general approach to adaptive control described here
exhibits a splitting or separation of the problems of identification of the unknown parameters and adaptive control. Maximum likelihood, least squares, or weighted least squares estimators are used to identify the unknown constant
parameters (3,4,5). These estimates are given recursively and
are strongly consistent, which means that the family of estimates converges to the true value of the parameter with the
probability one. It turns out that for some cases the weighted
least squares estimator is strongly consistent whereas the
least squares estimator is not. The adaptive control constructed by the so-called certainty equivalence principle, that
is, the optimal stationary control, is computed by replacing
the unknown true parameter values by the current estimates
of these values. Because the optimal stationary controls are
continuous functions of the unknown parameters, the selftuning property is verified, which means that asymptotically
adaptive control using the estimate of the unknown parameter is as good as optimal control if we know the system. It is
also shown that the family of average costs using the control
from the certainty equivalence principle converges to the optimal average cost. This verifies the self-optimizing property.
In describing a stochastic control model, the kind of information available to the controller at each instant plays an
important role. Several situations are possible:
(1) The controller has no information during the systems
operation. In this case a function of time is chosen as a control. These controls are often called open loop as distinct
from closed-loop or feedback controls, in which the actual
value of control input at time t is a function of the observation
at time t (6). The distinction between open-loop and feedback
control is fundamental. An open-loop control is deterministic
whereas a feedback law determines a random control process.
543
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
544
STOCHASTIC SYSTEMS
I (U ) = Ey
s
y = W (s)
X (0) = x0 Rn
dW (t)
Z(0) = z0
STOCHASTIC SYSTEMS
J[U ()] = E
E[X (t)|Fs ]
a.s.
(2)
> X (s)
a.s.
(3)
>
PRELIMINARY RESULTS
We shall assume that (, F , P) is a probability space and
that F t, t 0 is a nondecreasing, right-continuous family of
sub--algebras of F . We assume that the filtration (F t) is
complete with respect to P, that is, each F t contains all the
P-null sets of F .
6 X (s)
where G is the running cost of inventory/backlog and production and 0 is the discount rate. The problem is to find an
admissible control U( ) that minimizes J[U( )].
545
(1)
a.s.
(5)
and E[Y] .
The Wiener Process Used for Modeling the Noise
Definition 5. The stochastic process (W(t), t 0) is a
Brownian motion (or a Wiener process) if
(1) W(0) 0;
(2) for t 0, W(t s) W(t) is a random variable that is
normally distributed with mean zero and variance 2s where
2 is a constant; and
(3) the process has independent increments (i.e., if 0 t1
t2 tn, then W(t2) W(t1), W(t3) W(t2), . . .,
W(tn) W(tn1) are mutually independent random variables).
If 2 1, then the process is called standard Brownian
motion.
The following result provides important information about
the quadratic variation of the Wiener process:
546
STOCHASTIC SYSTEMS
n1
lim
n
[W (ti+1
) W (tin )]2 = T
a.s.
(6)
(t) = j
i=0
t [t j , t j+1 )
(11)
and
lim E
n1
2
n
[W (ti+1
)
W (tin )]2
=0
(7)
i=0
n
i1
n
i
(s) dW (s) =
that is
(W(t ) (W(t )) converges to t both almost
surely and in L2.
n1
i0
j1
i [W (ti+1 ) W (t j )] + j [W (t) W (t j )]
i=0
(12)
Note that the integral is linear and continuous (almost
surely) in t. We want to preserve these properties and at the
same time we want to extend the class of functions that we
can integrate. To do this, we need the following result:
Proposition 11. Let (t), t [0, T] be a process adapted
to F t, t [0, T] such that
(13)
lim P
(s) dW (s)
(15)
then using Proposition 11, the family Yn converges to a process Y uniformly in probability. Note that the uniform convergence and the continuity of sample paths for Yn imply the
continuity of the sample paths of the process Y. Now we can
define the stochastic integral as follows:
(9)
(16)
Stochastic Integrals
In this section we define the stochastic integral
(s) dW (s)
Yn (t) =
(14)
(8)
If (W(t), t 0) is a Brownian motion process with continuous sample paths and we let F t be the -algebras generated
by [W(s), s 0], it follows from the independence of increments property of Brownian motion that W(t), F tt [0, T]
is a square integrable martingale with continuous sample
paths. The converse of this statement is true:
(10)
(17)
(18)
The unbounded variation property of Brownian motion sample paths described in the previous section does not allow us
to use the definition of the RiemannStieljes integral. Nevertheless by proceeding similarly we can define stochastic inte-
and
t
0
STOCHASTIC SYSTEMS
547
We mentioned previously that we shall present the differential rule that we must use in stochastic calculus. This rule
was introduced by Ito.
(19)
(26)
(27)
and
T
0
f 2 (t) dt = a.s.
(20)
Then
f (t) dW (t)
lim
= 0 a.s.
(21)
(28)
(29)
and
f (t) dt
2
Then
An interesting feature of the stochastic integral is that it
does not satisfy the ordinary calculus rules of integration. A
standard example of this situation is the fact that
W (s) dW (s) =
0
t
W 2 (t)
2
2
(22)
k
1
F f (t) f j (t) + Fk+1 dt
dY (t) =
Fi dXi (t) +
2 i, j=1 ij i
i=1
X (0) = X0
Fij =
2F
[X (t), . . ., Xk (t), t]
Xi X j 1
(24)
We have the following result about the existence and uniqueness of the solution:
Theorem 13. If a(t, x) and (t, x) are continuous in t and satisfy the Lipschitz condition in x uniformly in t, that is, for
every t [0, T]
|a(t, x) a(t, y)| + | (t, x) (t, y)|
6 K|x y|,
Fk+1 =
F
[X (t), . . ., Xk (t), t]
t 1
i, j 1, . . ., k.
THE FILTERING PROBLEM (7)
a[u, X (u)] du +
0
F
[X (t), . . ., Xk (t), t]
Xi 1
(23)
Fi =
and
In the previous section we introduced the concept of stochastic integration. Now we can use it to define stochastic differential equations. In fact we say that X(t), t [0, T] is a
solution for the stochastic differential equation
X (t) = X0 +
(30)
where
k
KR
(25)
(31)
548
STOCHASTIC SYSTEMS
then
where
dZ = c[t, X (t)] dt + [t, X (t)] dW (t)
(33)
Z(0) = 0
X (t) is Ft -measurable,
G2 (t)
dS
= 2F (t)S(t) 2 S2 (t) + C2 (t),
dt
D (t)
S(0) = E{(X0 E[X0 ])2 }
(35)
(40)
(34)
= inf{E[|X (t) = Y |2 ] : Y Y }
where
1
t
2
2
S(t) = S1
+
X
(s)
(s)
ds
0
Y := {Y :
Rn ; Y L2 (P) and Y is Ft -measurable}, (36)
where L2(P) L2(, P) and (, F , P) is a probability space.
One of the most important results is the following:
d(t)
This is the basis for the general FujisakiKallianpurKunita equation of filtering theory. See, for example, Ref. 34.
Here we concentrate on the linear case, which allows an
explicit solution in terms of a stochastic differential equation
(t) (the KalmanBucy filter):
for X
In the linear filtering problem the system and observation
equations have the following form:
(linear system)
(37)
+ X (t)2 (t)2 dt
0
0
t
1
2
2
d S0 +
X (s) (t) ds (t)
(38)
(linear system)
F (t),C(t) R
G(t), D(t) R
X (t)S(t)
[dZ(t) = X (t)(t)]
(t)2
(39)
so
+
0 S1
0
S1
+
0
X (s)2 (t)2 ds
This estimate coincides with the maximum likelihood estimate in the classical estimation theory if S1
0 0. See (22).
For more information about estimates of drift parameters
in diffusion, see the next section.
Theorem 16. The Multidimensional KalmanBucy Filter.
The solution
X (t) = E[X (t)|Ft ]
STOCHASTIC SYSTEMS
(45)
where
S(t) := E [X (t) X (t)][X (t) X (t)]T Rnn
W
+ min AsU W + L(s, X , U ) = 0
s
(47)
dS
T
T
T
= FS + SF DG (DDT )1 GS + CC
dt
S(0) = E {X (0) E[X (0)]}{X (0) E[X (0)]}T
The condition on D(t) mr is now that D(t)D(t)T is invertible
for all t and that [D(t)D(t)T]1 is bounded on every bounded
t-interval.
A similar solution can be found for the more general situation
where X(t) , Z(t) and B(t) [W(t), V(t)] is (n m)dimensional Brownian motion with appropriate dimensions
on the matrix coefficients. See (34), which also treats the nonlinear case.
For various applications of filtering theory see Ref. 33.
n
549
(49)
(41)
h0
Tt+h Tt
= At
h
(43)
1
f
2 f
(x) + 2 (t, x) 2 (x)
x
2
x
>
J u (s, x) = E s,x
t1
T
T
T
X (t) C(t)X (t) + U (t) D(t)U (t) dt + X (t1 ) RX (t1 )
s
(42)
(44)
(A t, t 0) is called the infinitesimal generator of the semigroup (Tt, t 0). Consider the problem of controlling the solu-
s 6 t1
(51)
550
STOCHASTIC SYSTEMS
(53)
t < t1
(55)
t1
W (s, x) = xT S(s)x +
s < t1
(56)
t1
a(s) =
dX (t) = A()X (t) + BU(t) dt + dW (t)
(54)
tion of the unknown parameters and adaptive control. Maximum likelihood (or equivalently least squares) estimates are
used to identify unknown constant parameters. These estimates are given recursively and are strongly consistent. The
adaptive control is usually constructed by the so-called certainty equivalence principle, that is, the optimal stationary
controls are computed by replacing the unknown true parameter values by the current estimates of these values. Because
the optimal stationary controls are continuous functions of
the unknown parameters, the self-tuning property is verified.
The family of average costs using the control from the certainty equivalence principle converges to the optimal average
cost. This verifies the self-optimizing property.
A model for the adaptive control of continuous-time linear
stochastic systems with complete observations of the state is
described by the following stochastic differential equation:
tr( T S)t dt
1
lim sup J(X0 , U, , t)
t
t
(57)
i Ai
(58)
(60)
p
i=1
(59)
(61)
where
J(X0 , U, , t) =
QX (s), X (s) + PU (s), U (s) ds
(62)
and t (0, ), X(0) X0, Q L (n); P L (m) are selfadjoint and P1 exists; [X(t), t ] satisfies Eq. (59); and
[U(t), t ] is adapted to (F t, t ). It is well known (33)
that, if is known, then there is an optimal linear feedback
control such that
U (t) = KX (t)
(63)
where K P1B*V and V is the unique, symmetric, nonnegative-definite solution of the algebraic Riccati equation
VA + AV VB P1 BV + Q = 0
(64)
6 t X (t)
(65)
STOCHASTIC SYSTEMS
where [K(t), t 0] is an L (n, m)-valued process that is uniformly bounded and there is a fixed 0 such that (K(t), t
0) is measurable with respect to (Xu, u t ) for each t
and [K(t), t [0, )] is a deterministic function. For such
an adaptive control, it is elementary to verify that there is a
unique strong solution of Eq. (59). The delay 0 accounts
for some time required to compute the adaptive control law
from the observation of the solution of Eq. (59).
Let (U(t), t 0) be an admissible adaptive control and let
[X(t), t 0] be the associated solution of Eq. (59). Let A (t)
[aij(t)] and A (t) [aij(t)] be L (p)-valued processes such that
t
aij (t) =
Ai X (s), A j X (s) ds
where (X(t), t 0) is the solution of Eq. (59) with the admissible adaptive control [U(t), t 0] and 0 K and V is the
solution of the algebraic Riccati Eq. (64) with 0. Then
lim inf
T
1
J(X0 , U, 0 , T )
T
> tr V
a.s.
(71)
a.s.
(72)
a ij (t) =
551
aij (t)
lim
aii (t)
1
J(X0 , U, 0 , T ) = tr V
T
a.s.
(73)
The previous results can be combined for a complete solution to the stochastic adaptive control problem Eqs. (59, 61).
a.s.
(74)
Theorem 18. Let (K(t), t 0) be an admissible adaptive feedback control law. If (5.A15.A4) are satisfied and 0 A ,
the interior of A , then the family of least squares estimates
[ (t), t 0] where (t) is the minimizer of Eq. (66), is strongly
consistent, that is,
(67)
P 0 lim (t)
= 0 = 1
Introduction
The family of estimates [ (t), t 0] can be computed recursively because this process satisfies the following equation
d(t)
The linear, Gaussian control problem with an ergodic, quadratic cost functional is probably the most well-known ergodic
control problem. Because the optimal control is easily computed and the existence of an invariant measure follows directly from the stability of the optimal system, it is a basic
problem to solve for stochastic adaptive control. For discretetime linear systems it has been studied extensively, especially
for ARMAX models (see Refs. 6, 28, 3740 for many references). Although this adaptive control problem has been less
studied for continuous-time linear systems, it is nonetheless
an important problem as a model for physical systems that
naturally evolve in continuous time and as an approximation
for discrete-time sampled systems when the sampling rate is
high.
The adaptive control problem is solved using only the natural assumptions of controllability and observability. The
weighted least squares scheme is used to obtain the conver-
(69)
(70)
lim
1
J(X0 , U, 0 , T ) = tr V
T
a.s.
(75)
552
STOCHASTIC SYSTEMS
(76)
where X(0) X0, X(t) n, U(t) m, [W(t), t 0] is an pvalued standard Wiener process, and [U(t), t 0] is a control
from a family specified subsequently. The random variables
are defined on a fixed, complete probability space (, F , P),
and there is a filtration (F t, t 0) defined on this space that
is specified subsequently. It is assumed that A and B are unknown.
The following assumption is made:
(A1) (A, B) is controllable.
To describe the identification problem in a standard form,
let
= [A B]
(77)
(1)
>
(85)
t 0
(2)
(86)
lim (t) =
(3)
a.s.
(87)
and
X (t)
(t) =
U (t)
and
(78)
(t, x) = [A(t, x)
B(t, x)]
(79)
F (t, x) = det
n1
(80)
(81)
G(t, x) = det
1
f [r(t)]
t
r(t) = e +
|(s)|2 ds
a(t) =
(82)
(83)
>0
i=0
and f .
F = f | f : R+ R+ , f is slowly increasing
dx
< for some c
and
x f (x)
c
(84)
n1
A (t, x)C CA (t, x)
iT
i=0
The linear transformation C is known. For the adaptive control problem, C Q1/2
where Q1 determines the quadratic
1
form of the state in the cost functional.
A random search method ensures uniform controllability
and observability of a family of estimates. Let (n, n ) be
a sequence of independent, identically distributed M (n
m,n)-valued random variables that is independent of [W(t),
t 0] so that each random variable n is uniformly distributed in the unit ball or the unit sphere for a norm of the
matrices. Define a sequence of M (n m,n)-valued random
STOCHASTIC SYSTEMS
If (t) is given by
T
(t) = A(t) B(t)Q1
2 B (t)R(t)
0 = 0
k = k
k1
if f (k, k )
otherwise
> (1 + ) f (k, k1 )
and
T
() = A() B()Q1
2 B ()R()
(t) = k
then
lim (t) = ()
a.s.
T
U (t) = Q1
2 B (t)R(t)X (t)
(88)
(89)
(91)
where k and
553
1
T
T
0
(90)
T
AT (t)R(t) + R(t)A(t) R(t)B(t)Q1
2 B (t)R(t) + Q1 = 0
a.s.
(92)
a.s.
554
STOCHASTIC SYSTEMS
a.s.
where
= [A B]
Remark. It should be noted that the conditions of this proposition are satisfied if D 0 so that the identification of deterministic systems is included in this result.
Because the family of estimates [ (t), t 0] is strongly consistent, the self-optimality of the diminishingly excited lagged
certainty equivalence control Eq. (92) can be verified.
Theorem 24. Let (A1) and (A2) be satisfied for the stochastic
system
dX (t) = AX(t) dt + BU (t) dt + D dW (t)
with the cost functional Eq. (90) where A and B are unknown.
If the admissible adaptive control
Lk = Q1
2 B (k)R(k)
is used, then
a.s.
where R is the unique, positive, symmetric solution of the algebraic Riccati equation
A R + RA RBQ1
2 B R + Q1 = 0
and
J(U ) = lim sup
T
1
T
T
0
a.s.
BIBLIOGRAPHY
1. V. Borkar, Optimal Control of Diffusion Processes, New York: Wiley, 1989.
2. T. E. Duncan and B. PasikDuncan, Adaptive control of continuous-time linear stochastic systems, Math. Control, Signals, Syst.,
3: 4560, 1990.
STOCK MARKETS
31. P. Bertrand, Adaptive control of partially observed linear stochastic systems, Stochastics, 54: 2151, 1995.
32. T. E. Duncan, B. PasikDuncan, and L. Stettner, Adaptive control of a partially observed discrete time Markov processes, Appl.
Math. Optim., 37: 269293, 1998.
33. R. E. Kalman and R. S. Bucy, New results in linear filtering and
prediction theory, Trans. ASME, Ser. D, J. Basic Eng., 83: 95
107, 1961.
34. G. Kallianpur, Stochastic Filtering Theory, New York: Springer
Verlag, 1980.
35. G. Yin and Q. Zhang (eds.), Mathematics in Stochastic Manufacturing Systems, Lecture Notes in Applied Mathematics, Providence:
Amer. Math. Soc., 1997, vol. 33.
36. H. J. Kushner, Stochastic Stability and Control, New York: Academic Press, 1967.
37. P. E. Caines, Linear Stochastic Systems, New York: Wiley, 1988.
38. G. Goodwin, P. Ramadge, and P. Caines, Discrete time stochastic
adoptive control, SIAM J. Control Optim., 19: 829853, 1981.
strom, and P. R. Kumar (eds.), Adaptive Con39. G. C. Goodwin, K. A
trol, Filtering and Signal Processing, IMA Volumes in Mathematics and Its Applications 74, New York: Springer, 1995.
40. G. Chen, G. Chen, and S. H. Hua, Stochastic Control Systems,
Boca Raton, FL: CRC Press, 1995.
BOZENNA PASIKDUNCAN
University of Kansas
555
SYSTEM IDENTIFICATION
SYSTEM IDENTIFICATION
The process of going from observed data to a mathemathical
model is fundamental in science and engineering. In the control area, this process has been termed system identification,
and the objective is then to find dynamical models (difference
or differential equations) from observed input and output signals. Its basic features are, however, common with general
model building processes in statistics and other sciences.
System identification covers the problem of building models of systems when insignificant prior information is available and when the systems properties are known, up to a few
parameters (physical constants). Accordingly, one talks about
black-box and gray-box models. Among black-box models,
there are familiar linear models such as ARX and ARMAX,
and among nonlinear black-box models we have, for example,
artificial neural networks (ANN).
THE PROBLEM
The area of system identification begins and ends with real
data. Data are required to build and to validate models. The
result of the modeling process can be no better than what
corresponds to the information contents in the data.
Look at two data sets:
Example 1 An Unstable Aircraft
Figure 1 shows some results from test flights of the new
Swedish aircraft, JAS-Gripen, developed by SAAB Military
Aircraft AB, Sweden. The problem is to use the information
in these data to determine the dynamical properties of the
aircraft for fine-tuning regulators, for simulations, and so on.
Of particular interest are the aerodynamical derivatives.
Example 2 Vessel Dynamics
Figure 2 shows data from a pulp factory. They are collected
from one of the buffer vessels. The problem is to determine
the residence time in the vessel. The pulp spends about 48 h
total in the process, and knowing the residence time in the
different vessels is important in order to associate various
portions of the pulp with the different chemical actions that
have taken place in the vessel at different times. (The -num-
263
20
40
60
80
100
120
140
160
180
20
40
60
80
100
120
140
160
180
20
40
60
80
100
120
140
160
180
20
40
60
80
100
120
140
160
180
Figure 1. Results from test flights of the new Swedish aircraft JASGripen, developed by SAAB Military Aircraft AB, Sweden. From
above (a) Pitch rate, (b) Elevator angle, (c) Canard angle, (d) Leading
edge flap.
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
264
SYSTEM IDENTIFICATION
Output #1
25
20
15
10
5
0
100
200
300
400
500
600
Input #1
20
15
y(t|
) = g(, Zt1 )
(1)
10
0
100
200
300
400
500
600
y(t) y(t|
(2)
80
60
40
20
0
0
100
200
300
400
500
600
200
150
100
50
0
0
100
200
300
400
500
600
The Model
We shall generally denote the systems input and output at
time t by u(t) and y(t), respectively. Perhaps the most basic
relationship between the input and output is the linear difference equation
y(t) + a1 y(t 1) + . . . + an y(t n)
= b1 u(t 1) + . . . + bm u(t m)
OUTLINE
The system identification procedure is characterized by four
basic ingredients:
1. The observed data
2. A set of candidate models
3. A criterion of fit
4. Validation
(3)
We have chosen to represent the system in discrete time, primarily since observed data are always collected by sampling.
It is thus more straightforward to relate observed data to discrete time models. Nothing prevents us, however, from working with continuous time models: we shall return to that
later.
In Eq. (3), we assume the sampling interval to be one time
unit. This is not essential but makes notation easiser.
A pragmatic and useful way to see Eq. (3) is to view it as
a way of determining the next output value given previous
observations
y(t) = a1 y(t 1) . . . an y(t n)
+ b1 u(t 1) + . . . + bm u(t m)
(4)
SYSTEM IDENTIFICATION
(5)
T
(6)
265
Once the vectors (t) are defined, the solution can easily be
found by modern numerical software, such as MATLAB, see
Ref. 2.
Example 3 First-Order Difference Equation
Consider the simple model
y(t) + ay(t 1) = bu(t 1)
y(t) = T(t)
This gives us the estimate according to Eqs. (5), (6) and (13)
To emphasize that the calculation of y(t) from past data [Eq.
(4)] indeed depends on the parameters in , we shall rather
call this calculated value y(t) and write
y(t|
) = T(t)
(7)
1
a N
y2 (t 1) y(t 1)u(t 1)
=
y(t 1)u(t 1) u2 (t 1)
b N
y(t)y(t 1)
y(t)u(t 1)
(8)
(9)
where
V (, ZN ) =
N
1
(y(t) y(t|
))2
N t=1
N
1
=
(y(t) T (t) )2
N t=1
(10)
(11)
0=
N
d
2
VN (, ZN ) =
(t)(y(t) T (t) )
d
N t=1
(t)y(t) =
t=1
N
(t) T (t)
which gives
N
(12)
t=1
or
N =
N
t=1
1
(t) (t)
T
N
t=1
(t)y(t)
(13)
266
SYSTEM IDENTIFICATION
This is a two input (v2 and r) and one output model and corresponds to choosing
(t) = [y(t 1) v2 (t 1) r(t 1)]T
in Eq. (7).
that E N 0, since e has zero mean. The estimate is consequently unbiased. Here, E denotes mathematical expectation.
We can also form the expectation of N TN, that is the covariance matrix of the parameter error. Denote the matrix within
brackets by RN. Take expectation with respect to the white
noise e. Then, RN is a deterministic matrix, and we have
2
y(t|
)
2
(14)
(15)
where e(t) is a white noise sequence with variance , but otherwise unknown. (That is, e(t) can be described as a sequence
of independent random variables with zero mean values and
variances .) Analogous to Eq. (7), we can write this as
y(t) = T (t)0 + e(t)
(16)
N =
=
N
1
(t) T (t)
t=1
N
N
t=1
(t) T (t)
t=1
PN =
(t) T (t)0 +
t=1
N
(t)e(t)
t=1
N = N 0 =
N
t=1
1
(t) T (t)
(20)
Starting from Eq. (3), there is actually another, quite different, way to approach the calculation of good values of ai and
bi from observed data in Eq. (8).
Equation (3) describes a linear, discrete-time system with
transfer function
or
1
R
N
(t)y(t)
1
N
(19)
This will be the covariance matrix of the input, that is, the
i j-element of R is
1
R
N N
N
G(z) =
(t)e(t)
(17)
t=1
(21)
(assuming n m)
Here, z is the z-transform variable, and one may simply
think of the transfer function G as a shorthand notation for
the difference in Eq. (3).
SYSTEM IDENTIFICATION
(22)
y(t|
) = G(q)u(t)
B(q)
u(t)
A(q)
(23)
(26)
See Eqs. (21)(22). Based on Eq. (26) we can predict the next
output from previous measurements either as in Eq. (23)
y(t|
)=
267
B(q)
u(t)
A(q)
(27)
) = (1 A(q))y(t) + B(q)u(t)
(28)
(29)
)=
b
u(t) = b
(a)k1 u(t k)
q+a
k=1
(30)
Then, we can add y(t) to both sides of the equation and rearrange to obtain
or
y(t
+ 1| ) + ay(t|
) = bu(t)
(24)
(31)
N = arg min
N
1
[ y(t) y(t|
)]2
N t=1
(25)
W (q, ) = 1 + w1 q1 + w2 q1 + w2 q2 + . . .
so that 1 W(q, ) actually contains a delay. We thus obtain
the predictor
y(t|
(32)
(33)
268
SYSTEM IDENTIFICATION
0 ) = W (q, )v(t)
(34)
(37)
(38)
(39)
1
B(q)
, H(q, ) =
A(q)
A(q)
G(q, ) =
C(q)
B(q)
, H(q, ) =
A(q)
A(q)
(41)
(35)
That is, we assume the system (plant) dynamics and the noise
model to have common poles and no numerator dynamics for
the noise. Its main feature is that the predictor y(t) will be
linear in the parameters according to Eq. (11) or Eq. (7).
We can make Eq. (40) more general by allowing also numerator dynamics. We then obtain the parameterization
(40)
B(q)
, H(q, ) = 1
F (q)
(42)
C(q)
B(q)
, H(q, ) =
F(q)
D(q)
(43)
This particular model parameterization is known as the BoxJenkins (BJ) model, since it as suggested in the well-known
book by Box and Jenkins (6).
It differs from the ARMAX-model in Eq. (42) in that it assigns different dynamics (poles) to the noise characteristics
from the inputoutput properties. It is thus better suited for
cases where the noise enters late in the process, such as
measurement noise. See Fig. 3.
One might wonder why we need all these different model
parameterizations. As has been mentioned in the text, each
has its advantages which can be summarized as follows
ARX. Gives a linear regression, very simple to estimate
ARMAX. Gives reasonable flexibility to the noise description, assumes that noise enters like the inputs
OE. Concentrates on the inputoutput dynamics
BJ. Very flexible, assumes no common characteristics between noise and inputoutput behavior.
Physically Parameterized Linear Models
So far, we have treated the parameters only as vehicles to
give reasonable flexibility to the transfer functions in the general linear model in Eq. (38). This model can also be arrived
at from other considerations.
SYSTEM IDENTIFICATION
e
u
y(t)
y(t)
x(t) =
ARX
e
u
B
F
0 1
x =
x+
0 a
y = (1 0) x + v
OE
269
0
u
b
(45)
C
D
B
F
0
0
1
a
B( ) =
0
b
(46)
C = (1 0)
The parameterization reflects our insight that the system contains an integration, but is in this case not directly derived
from detailed physical modeling. Basic physical laws would,
in this case, have given us how depends on physical constants, such as resistance of the wiring, amount of inertia,
friction coefficients, and magnetic field constants.
A( ) =
ARMAX
a
b
BJ
(44a)
(44b)
(47)
where
Here, x(t) is the state vector and typically consists of physical
variables (such as positions and velocities, etc). The state
space matrices A, B, and C are parameterized by the parameter vector , reflecting the physical insight we have into the
process. The parameters could be physical constants (resistance, heat transfer coefficients, aerodynamical derivatives,
etc.) whose values are not known. They could also reflect
other types of insights into the systems properties.
Example 6 An Electric Motor
Consider an electric motor with the input u being the applied
voltage and the output y being the angular position of the
motor shaft.
A first but reasonable approximation of the motors dynamics is as a first-order system from voltage to angular velocity,
followed by an integrator:
G(s) =
b
s(s + a)
A( ) = e
A( )T
,
B( ) =
eA( ) B( ) d
(48)
(49)
where e(kT) is white noise. The step from Eqs. (47) to (49)
is really a standard Kalman filter step: x will be the one-step
ahead predicted Kalman states. A pragmatic way to think
about it is as follows: In Eq. (47), the term v(kT) may not be
white noise. If it is colored, we may separate out that part of
v(kT) that cannot be predicted from past values. Denote this
part by e(kT): it will be the innovation. The other part of
270
SYSTEM IDENTIFICATION
(50)
where
(t) = (Z t1 )
Let the dimension of be d. As before, we shall call this vector the regression vector, and its components will be referred
to as the regressors. We also allow the more general case that
the formation of the regressors is itself parameterized:
(t) = (Z t1 , )
with
(51)
We are, thus, back at the basic linear model in Eq. (38). The
parameterization of G and H in terms of is, however, more
complicated than the ones we discussed earlier.
The general estimation techniques, model properties [including the characterization in Eq. (85)], algorithms, etc.
apply exactly as described in the section on general parameter estimation techniques.
From these examples, it is also quite clear that non-linear
models with unknown parameters can be approached in the
same way. We would then typically arrive at a structure
x(t)
= f (x(t), u(t), )
y(t) = h(x(t), u(t), ) + v(t)
(52)
) = g(, Z t1 )
(56)
which we for short write (t, ). For simplicity, the extra argument will, however, be used explicitly only when essential for the discussion.
The choice of the nonlinear mapping in Eq. (53) has thus
been reduced to two partial problems for dynamical systems:
1. how to choose the nonlinear mapping g() from the regressor space to the output space (i.e., from Rd to Rp)
2. how to choose the regressors (t) from past inputs and
outputs
The second problem is the same for all dynamical systems,
and it turns out that the most useful choices of regression
vectors are to let them contain past inputs and outputs, and
possibly also past predicted/simulated outputs. The regression vector will thus be of the character in Eq. (6). We now
turn to the first problem.
Nonlinear Mappings: Possibilities
Now, let us turn to the nonlinear mapping
g(, )
(57)
which for any given maps from Rd to Rp. For most of the
discussion, we will use p 1; that is, the output is scalarvalued. At this point, it does not matter how the regression
vector (1, . . ., d)T was constructed. It is just a vector
that lives in Rd.
It is natural to think of the parameterized function family
as function expansions:
g(, ) =
k gk ()
(58)
(53)
We shall consequently allow quite general nonlinear mappings g. This section will deal with some general principles
for how to construct such mappings, and will cover artificial
neural networks as a special case. See Refs. 11 and 12 for
recent and more comprehensive surveys.
Nonlinear Black-Box Structures
Now, the model structure family in Eq. (53) is really too general, and it turns out to be useful to write g as a concatenation
of two mappings: one that takes the increasing number of
past observations Zt1 and maps them into a finite dimensional vector (t) of fixed dimension and one that takes this
vector to the space of the outputs:
y(t|
) = g(, Z t1 ) = g((t), )
(55)
(54)
SYSTEM IDENTIFICATION
(59)
(61)
1
1 + ex
(63)
271
gk () =
d
(kj ( j kj ))
(64)
j=1
k )
(65)
2
k
= T k
(66)
with k as a possibly k-dependent positive definite matrix of dilation (scale) parameters. In simple cases, k
may be just scaled versions of the identity matrix.
3. Ridge Construction. Let be any single-variable function. Then for all k Rd, k R, a ridge function is
given by
gk () = gk (, k , k ) = (kT + k ), Rd
(67)
The ridge function is thus constant for all in the subspace Rd : Tk constant. As a consequence, even
if the mother basis function has local support, the basis functions gk will have unbounded support in this
subspace. The resulting basis could be said to be semiglobal, but the term ridge function is more precise.
Approximation Issues. For any of the described choices, the
resulting model becomes
g(, ) =
en
k (k ( k ))
(68)
k=1
272
SYSTEM IDENTIFICATION
1
(d/ p)
(69)
It thus increases exponentially with the number of regressors. This is often referred to as the curse of dimensionality.
Connection to Named Structures. Here we briefly review
some popular structures; other structures related to interpolation techniques are discussed in Refs. 11 and 12.
Wavelets. The local approach corresponding to Eqs. (58,
and 65) has direct connections to wavelet networks and
wavelet transforms. The exact relationships are discussed in Ref. 11. Loosely, we note that via the dilation
parameters in k, we can work with different scales simultaneously to pick up both local and not-so-local variations. With appropriate translations and dilations of a
single suitably chosen function (the mother wavelet), we can make expansion Eq. (58) orthonormal. This
is discussed extensively in Ref. 12.
Wavelet and Radial Basis Networks. The choice in Eq. (61)
without any orthogonalization is found in both wavelet
networks (13) and radial basis neural networks (14).
Neural Networks. The ridge choice in Eq. (67) with given
by Eq. (63) gives a much-used neural network structure
through the one hidden layer feedforward sigmoidal net.
Hinging Hyperplanes. If instead of using the sigmoid
function we choose V-shaped functions (in the form of
a higher-dimensional open book), Briemans hinging
SYSTEM IDENTIFICATION
) = g(, Z t1 )
(70)
choose
L(q) = 1
t = 1, 2, . . ., N
(71)
(72)
(73)
(74)
as N
(78)
That is, as more and more data become available, the estimate converges to that value * that would minimize the expected value of the norm of the filtered prediction errors.
This is in a sense the best possible approximation of the true
system that is available within the model structure. The expectation E in Eq. (78) is taken with respect to all random
disturbances that affect the data, and it also includes averaging over the input properties. This means, in particular, that
* will make y(t*) a good approximation of y(t) with respect
to those aspects of the system that are enhanced by the input
signal used.
The second basic result is the following one. If (t, *) is
approximately white noise, then the covariance matrix of N
is approximately given by
E(N )(N )T
(75)
(77)
where
where
N
1
VN (, ZN ) =
(F (t, ))
N t=1
(76)
Model Quality
The MLE has several nice statistical features and, thus, gives
a strong moral support for using the outlined method. Another pleasing aspect is that the method is independent of the
particular model parameterization used (although this will affect the actual minimization procedure). For example, the
method of back propagation often used in connection with
neural network parameterizations amounts to computing N
in Eq. (74) by a recursive gradient method. We shall deal with
these aspects later.
),
273
[E (t) T (t)]1
N
(79)
where
= E 2 (t, )
(t) =
d
y(t|
)| =
d
(80)
(81)
274
SYSTEM IDENTIFICATION
N
1
(t) T (t)
N t=1
L(q)
[ y(t) G(q, )u(t)]
H(q, )
L(q)
=
[(G0 (q) G(q, ))u(t) + v(t)]
H(q, )
+
G0 (ei ) G(ei , ) 2
(87)
FN = EV (N )
(88)
The
The
The
The
FN V N ( ) 1 +
u () L(ei ) 2
d
H(ei , ) 2
(85)
v () L(ei ) 2 / H(ei , ) 2 d
= arg min
Here, expectation E is over the data properties [i.e., expectation over Z with the notation Eq. (8)]. Recall that expectation also can be interpreted as sample means as in Eq. (82).
Before we continue, let us note the very important aspect
that the fit V will depend not only on the model and the true
system but also on data properties, like input spectra, possible
feedback, etc. We shall say that the fit depends on the experimental conditions.
The estimated model parameter N is a random variable
because it is constructed from observed data that can be described as random variables. To evaluate the model fit, we
then take the expectation of V( N) with respect to the estimation data. That gives our measure
(84)
V ( ) = E y(t) y(t|
The rather remarkable fact is that if the two last data properties coincide, then, asymptotically in N (4, Chap. 16),
F (t, ) =
(82)
The results in Eqs. (77) through (81) are general and hold for
all model structures, both linear and nonlinear ones, subject
only to some regularity and smoothness conditions. They are
also fairly natural and will give the guidelines for all user
choices involved in the process of identification. See Ref. 4 for
more details around this.
(86)
This shows clearly that the fit can be affected by the choice of
prefilter L, the input spectrum u, and the noise model H .
*
dim
N
(89)
(90)
and let
= E y(t) y0 (t|t 1)
(91)
SYSTEM IDENTIFICATION
bias error, that is, the discrepancy between the true predictor
and the best one available in the model structure. Under the
same assumptions as above, Eq. (89) can be rewritten as
FN + W ( ) +
dim
N
(92)
(93)
(i+1) = (i) i R1
i gi
(94)
(95)
where
(t, ) =
y(t|
R( ) = VN ( ) =
N
d 2VN ( )
1
=
(t, ) T (t, )
d 2
N t=1
N
1
2
+
( y(t) y(t|
)) 2 y(t|
)
N t=1
(98)
The true Newton direction will, thus, require that the second
derivative
2
y(t|
)
2
be computed. Also, far from the minimum, R() need not be
positive semidefinite. Therefore, alternative search directions
are more common in practice:
Gradient Direction. Simply take
(99)
Ri = Hi =
N
1
(t, (i) ) T (t, (i) )
N t=1
(100)
(101)
Search Directions. The basis for the local search is the gradient
N
1
dVN ( )
=
( y(t) y(t|
)) (t, )
d
N t=1
(97)
where
Ri = I
2
VN ( ) =
The three terms constituting the model error then have the
following interpretations:
275
(96)
276
SYSTEM IDENTIFICATION
then the output after possible transients have faded away will
be
for t = T, 2T, 3T, . . .
Ryu ( )ei
(106)
y(t) = y0 cos(t + ),
yu () =
If the system is driven by the input [see Eq. (102)] for a certain u0 and 1 and we measure y0 and from the output signal, it is possible to determine the complex number G(ei1T)
using Eqs. (104)(105). By repeating this procedure for a
number of different , we can get a good estimate of the frequency function G(eiT). This method is called frequency analysis. Sometimes, it is possible to see or measure u0, y0, and
directly from graphs of the input and output signals. Most of
the time, however, there will be noise and irregularities that
make it difficult to determine directly. A suitable procedure
is then to correlate the output with cos t and sin t.
(107)
1
U ()
N N
(108)
UN () =
N
u(t)eit
(109)
t=1
(110)
u () + v ()
(111)
(112)
It is easy to see how the transfer function G(ei) and the noise
spectrum v() can be estimated using these expressions, if
only we have a method to estimate cross spectra.
(103)
Estimation of Spectra. The spectrum is defined as the Fourier transform of the correlation function. A natural idea
would then be to take the transform of the estimate
y0 = G(eiT ) u0
(104)
= arg G(eiT )
(105)
N
1
R N
y(t)u(t )
yu ( ) =
N t=1
where
(113)
SYSTEM IDENTIFICATION
That will not work in most cases, though. The reason could
N
yu
be described as follows: The estimate R
() is not reliable for
large since it is based on only a few observations. These
bad estimates are mixed with good ones in the Fourier
transform, thus creating an overall bad estimate. It is better
to introduce a weighting, so that correlation estimates for
large lags carry a smaller weight:
N
yu () =
i
R N
yu () w ()e
(114)
about
k <
(115)
G N (ei ) =
N
yu ()
N
u ()
(116)
N
N
v () = y ()
2
N
yu ()
N
u ()
(117)
(118)
v ()
N u ()
(119)
N
2v ()
Var
v () 0.7
N
(120)
1
k
1 + cos
2
w (k) = 0
w (k) =
radians/time unit
=
277
and
(121)
We assume that we have no insight into the particular structure, and we would just estimate any matrices A, B, C, and
D, that give a good description of the input-output behavior
of the system. This is not without problems, among other
things because there are an infinite number of such matrices
that describe the same system (the similarity transforms).
The coordinate basis of the state-space realization, thus,
needs to be fixed.
Let us for a moment assume that not only are u and y
measured, but also the sequence of state vectors x. This
would, by the way, fix the state-space realization coordinate
basis. Now, with known u, y, and x, the model in Eq. (121)
becomes a linear regression: the unknown parameters, all of
278
SYSTEM IDENTIFICATION
the matrix entries in all the matrices, mix with measured signals in linear combinations. To see this clearly, let
Y (t) =
=
(t) =
x(t + 1)
y(t)
A
C
B
D
x(t)
u(t)
DATA QUALITY
w(t)
e(t)
E(t) =
(122)
k = {1, 2, . . ., n}
x(t) = L
y(t
+ 1|t)
..
.
y(t
+ n|t)
(123)
where n is the model order (the dimension of x). See also Appendix 4.A in Ref. 4. We could then form these predictors and
select a basis among their components:
Another case where the input is too simple is when it is generated by feedback such as
u(t) = Ky(t)
(125)
(124)
SYSTEM IDENTIFICATION
If we let the input be filtered white noise, this gives information on how to choose the filter. In the time domain, it is often
useful to think like this:
Use binary (two-level) inputs if linear models are to be
built; this gives maximal variance for amplitude-constrained inputs.
Check that the changes between the levels are such that
the input occasionally stays on one level so long that a
step response from the system has time, more or less, to
settle. There is no need to let the input signal switch so
quickly back and forth that no response in the output is
clearly visible.
Note that the second point is really just a reformulation in the
time domain of the basic frequency domain advice: let the input
energy be concentrated in the important frequency bands.
A third basic piece of advice about experiment design concerns the choice of sampling interval.
Construct
experiment,
collect data
Data
Filter
data?
Polish and
present data
Data
Fit model
to data
Choose model
structure
Model
Validate
model
Data
not OK
Model structure
not OK
No
279
Accept
model?
Yes
(126)
N2
for which the criterion is evaluated:
and validation data, Zval
2)
FN = VN (N , ZN
val
1
(127)
Here, VN is the criterion in Eq. (75). Then, FN will be an unbiased estimate of the measure FN, defined by Eq. (88), which
was discussed at length in the previous section. The procedure would the be to try out a number of model structures
and choose the one that minimizes FN1.
Such cross validation techniques to find a good model
structure have an immediate intuitive appeal. We simply
check if the candidate model is capable of reproducing data
it hasnt yet seen. If that works well, we have some confidence
in the model, regardless of any probabilistic framework that
might be imposed. Such techniques are also the most commonly used ones.
A few comments could be added. In the first place, one
could use different splits of the original data into estimation
280
SYSTEM IDENTIFICATION
and validation data. For example, in statistics, there is a common cross validation technique called leave one out. This
means that the validation data set consists of one data point
at a time but successively applied to the whole original set.
In the second place, the test of the model on the validation
data does not have to be in terms of the particular criterion
[see Eq. (127)]. In system identification, it is common practice
to simulate (or predict several steps ahead) the model using
the validation data and then visually inspect the agreement
between measured and simulated (predicted) output.
Estimating the Variance ContributionPenalizing the Model
Complexity. It is clear that the criterion in Eq. (127) has to
be evaluated on the validation data to be of any use; it would
be strictly decreasing as a function of model flexibility if evaluated on the estimation data. In other words, the adverse effect of the dimension of shown in Eq. (92) would be missed.
There are a number of criteria, often derived from entirely
different viewpoints, that try to capture the influence of this
variance error term. The two best known ones at Akaikes
Information Theoretic Criterion, AIC, which as the form (for
Gaussian disturbances)
2 dim
V N (, ZN ) = 1 +
N
N
1
2 (t, )
N t=1
(128)
mercial packages for identification available, such as Mathworks System Identification Toolbox (2), Matrixxs System
Identification Module (23) and PIM (24). They all have in
common that they offer the following routines:
Handling of data, plotting, and so on: Filtering of data,
removal of drift, choice of data segments, etc.
Non-parametric identification methods: Estimation of covariances, Fourier transforms, correlation- and spectralanalysis, etc.
Parametric estimation methods: Calculation of parametric estimates in different model structures.
Presentation of models: Simulation of models, estimation
and plotting of poles and zeros, computation of frequency
functions, and plotting Bode diagrams, etc.
Model validation: Computation and analysis of residuals
((t, N)). Comparison between different models properties, etc.
The existing program packages differ mainly in various user
interfaces and by different options regarding the choice of
model structure according to C above. For example, MATLABs Identification Toolbox (2) covers all linear model structures discussed here, including arbitrarily parameterized linear models in continuous time.
Regarding the user interface, there is now a clear trend to
make it graphically oriented. This avoids syntax problems
and relies more on click and move, at the same time as tedious menu-labyrinths are avoided. More aspects of CAD
tools for system identification are treated in Ref. 25.
How to Get to a Good Model?
It follows from our discussion that the most essential element
in the process of identificationonce the data have been recordedis to try out various model structures, compute the
best model in the structures using Eq. (38), and then validate
this model. Typically, this has to be repeated with quite a few
different structures before a satisfactory model can be found.
While one should not underestimate the difficulties of this
process, the following simple procedure to get started and
gain insight into the models could be suggested:
1. Find out a good value for the delay between input and
output, for example, by using correlation analysis.
2. Estimate a fourth order linear model with this delay
using part of the data, and simulate this model with the
input and compare the models simulated output with
the measured output over the whole data record. In
MATLAB language, this is simple
z = [y u];
compare(z,arx(z(1:200,:),[4 4 1]));
If the model/system is unstable or has integrators, use prediction over a reasonable large time horizon instead of simulation.
Now, either of two things happen:
The comparison looks good. Then, we can be confident
that with some extra worktrying out different orders
SYSTEM IDENTIFICATION
20
40
60
80
100
120
140
160
281
180
Clearly, this advice does not cover all the art of identification,
but it is a reasonable first approximation.
6
4
2
0
2
4
6
8
0
20
40
60
80
100
120
140
160
180
Figure 5. Dashed line: actual pitch rate. Solid line: 10 step ahead
predicted pitch rate, based on the fourth order model from canard
angle only.
100
200
300
400
500
600
Figure 7. Dashed line: -number after the vessel, actual measurements. Solid line: simulated -number using the input only and a
fourth order linear model with delay 12, estimated using the first 200
data points.
282
SYSTEM INTERCONNECTS
9. D. Brillinger, Time Series: Data Analysis and Theory, San Francisco: Holden-Day, 1981.
10. N. Draper and H. Smith, Applied Regression Analysis, 2nd ed,
New York: Wiley, 1981, 2nd ed.
11. J. Sjoberg et al., Nonlinear black-box modeling in system identification: A unified overview, Automatica, 31 (12): 16911724,
1995.
6
4
2
12. A. Juditsky et al., Nonlinear black-box modeling in system identification: Mathematical foundations, Automatica, 31 (12): 1724
1750, 1995.
13. Q. Zhang and A. Benveniste, Wavelet networks, IEEE Trans.
Neural Networks, 3: 889898, 1992.
14. T. Poggio and F. Girosi, Networks for approximation and learning. Proc. IEEE, 78: 14811497, 1990.
6
0
4
6
8
50
100
150
200
250
300
350
400
16. J. E. Dennis and R. B. Schnabel, Numerical Methods for Unconstrained Optimization and Nonlinear Equations, Englewood Cliffs,
NJ: Prentice-Hall, 1983.
17. J. Rissanen, Basis of invariants and canonical forms for linear
dynamic systems, Automatica, 10: 175182, 1974.
18. H. Akaike, Stochastic theory of minimal realization, IEEE Trans.
Autom. Control, AC-19: 667674, 1974.
19. P. V. Overschee and B. DeMoor, Subspace Identification of Linear
Systems: Theory, Implementation, Application, Boston, MA:
Kluwer, 1996.
20. W. E. Larimore, System identification, reduced order filtering
and modelling via canonical variate analysis, Proc. 1983 Amer.
Control Conf., San Francisco, 1983.
21. H. Akaike, A new look at the statistical model identification,
IEEE Trans. Autom. Control, AC-19: 716723, 1974.
22. J. Rissanen, Modelling by shortest data description, Automatica,
14: 465471, 1978.
23. MATRIXx Users Guide, Santa Clara, CA: Integrated Systems
Inc., 1991.
24. I. D. Landau, System Identification and Control Design Using
P.I.M. Software, Englewood Cliffs, NJ: Prentice-Hall, 1990.
25. L. Ljung, Identification of linear systems, in E. D. Linkens (ed.),
CAD for Control Systems, New York: Marcel Dekker, 1993, Chap.
6, pp. 147165.
LENNART LJUNG
BIBLIOGRAPHY
strom and T. Bohlin, Numerical identification of linear
1. K. J. A
dynamic systems from normal operating records, IFAC Symp.
Self-Adapt. Syst., Teddington, England, 1965.
2. L. Ljung, The System Identification Toolbox: The Manual, Natick,
MA: MathWorks Inc., 1995, 4th ed.
3. L. Ljung and T. Glad, Modeling of Dynamic Systems, Englewood
Cliffs, NJ: Prentice-Hall, 1994.
4. L. Ljung, System IdentificationTheory for the User, Englewood
Cliffs, N.J: Prentice-Hall, 1987.
5. T. Soderstrom and P. Stoica, System Identification, London: Prentice-Hall Int., 1989.
6. G. E. P. Box and D. R. Jenkins, Time Series Analysis, Forecasting
and Control, San Francisco: Holden-Day, 1970.
7. J. Schoukens and R. Pintelon, Identification of Linear Systems: A
Practical Guideline to Accurate Modeling, London: Pergamon,
1991.
8. L. Ljung and T. Soderstrom, Theory and Practice of Recursive
Identification, Cambridge, MA: MIT Press, 1983.
Linkoping University
592
ROBUST CONTROL
ROBUST CONTROL
The role of feedback in automatic control systems is to exploit
evolving real-time measurements to obtain more precise control over system behavior. A robust control system is one that
performs within specified tolerances despite uncertain variations in the controlled plant within given bounds. Robust control theory is that branch of mathematical system theory concerned with the design and analysis of robust feedback
control systems.
The block diagram in Fig. 1 represents a typical feedback
control system consisting of an uncertain plant and a controller. The plant might be an aircraft or missile, a robotic
lunar surface explorer, a chemical process, an automobile engine, or a nuclear power plant. It might be something as small
as the magnetic read-head positioning system in a computer
disk drive or as large as the global economy. It could be almost any complex dynamic system. The controller, on the
other hand, is typically a computer or a microprocessor,
though it could be simply the thermostat in a home heating
control system or the mechanical linkage in the eighteenthcentury flyball governor invented by James Watt for controlling steam engines. From a control theorists perspective the
uncertain plant and the controller are simply mathematical
relationships, for example, differential equations. Robust control theory is focused on the quantitative analysis of the consequences of plant uncertainty. The need for robust control
arises when a control system design must be based on an inexact mathematical model of the true physical plant.
d
Command
+
r
Error
e
K(s)
Control
u
Controller
G(s)
Output
y
Plant
he showed how to design uncertainty-tolerant controllers using classical root-locus and frequency-response methods.
Unfortunately, the significance of Horowitzs contribution
was not immediately noticed by control researchers. A gap
between mathematical control theory and control engineering
practice emerged in the 1960s and early 1970s, as the mathematical theory of feedback control increasingly decoupled
from engineering practice. Time-tested root-locus and frequency-response techniques were regarded by researchers as
ad hoc and simplistic. They sought to pose feedback design
as a mathematical optimization. But, explicit mathematical
representations of robustness issues were not incorporated in
their mathematical representations of the problems of feedback control. In those early days, mathematical control research generally presumed that available mathematical models were sufficiently accurate. The profound significance of
this omission was not immediately evident.
In the 1960s, mathematical proofs of optimality were
widely regarded as sufficient evidence that the emergent, but
as yet untried, modern mathematical theories would produce
superior results. But in the early 1970s, researchers began to
be jolted as early attempts to apply mathematical optimization theories to the design of complex multivariable feedback
controllers for military aircraft and submarines resulted in
some surprising failures. In one unpublished design study
carried out at Systems Control Inc. under the supervision of
D. Kleinman with MITs Michael Athans as consultant, a linear quadratic Gaussian (LQG) controller for a Trident submarine caused the vessel to unexpectedly surface in nonlinear
simulations involving moderately rough seas. In another example, a 1977 journal paper describing the disappointing results
of an LQG control design study for the F-8C Crusader aircraft
concluded euphemistically with the observation that The
study has pinpointed certain theoretical weaknesses . . . as
well as the need for using common sense pragmatic techniques
to modify the design based on pure theory. A lack of attention
to robustness was quickly identified as the dominant factor in
these failures.
In 1976, the modern field of robust control theory was
born. The term robustness was introduced into the control
theory lexicon in papers by E. J. Davison and by M. Safonov
and M. Athans. Mathematical control theorists in Michael
Athans MIT laboratory began to refocus their attention on
methods for analyzing and optimizing the robustness of feedback control systems. Researchers sought to salvage as much
as possible of the mathematical theory developed in the
1960s. They worked to link linear optimal control theory to
classical root-locus/pole-placement methods. In 1978 researcher Gunter Stein summarized the mood of the time, explaining that his goal was to make modern control theoretical methods work. Researchers were beginning to reexamine
classical frequency-response methods and to take a closer look
at Horowitzs work on uncertainty-tolerant single-loop feedback design. They were seeking multivariable generalizations
of Horowitzs ideas that would mesh well with the sophisticated mathematics of modern optimal control theory.
Though attempts at direct generalizations of Horowitzs
ideas to more complex multivariable feedback systems with
several feedback control loops proved unfruitful, mathematical control researchers quickly discovered that multivariable
robustness analysis methods could be developed based on the
so-called small-gain theorem of nonlinear stability theory.
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
ROBUST CONTROL
Along with the Lyapunov, ZamesSandberg and other nonlinear stability theories, this now became the core of the emergent mathematical theory of robust control. Prior to 1976, the
Lyapunov and ZamesSandberg theories were generally regarded as esoteric tools for nonlinear stability analysis. Despite some suggestive early remarks by stability theory pioneers G. Zames and V. Popov, by 1975, the robustness
implications of nonlinear stability theories had yet to be developed. Today, many still regard the mid 1976 change of focus of mathematical control theory from optimal control to
robustness as a revolutionary paradigm shift, though, unlike
many scientific revolutions, the shift occurred with remarkably little controversy.
In 1977 Safonov showed that the general problem of robust
stability analysis is equivalent to computing a topological separation of graphs of feedback operators. Lyapunov, conic sector, positivity small-gain and other nonlinear stability theories emerge as special cases. Other researchers, such as J. C.
Doyle, G. Stein and N. R. Sandell exploited H. H. Rosenbrocks multivariable version of the Nyquist stability criterion
to develop a simplified, and hence more widely accessible, linear explanation of multivariable robustness criteria, singular-value (-plots), and structured singular-value (-plots)
frequency-response plots.
By 1981, the concept of singular-value, loop shaping and
the closely related concept of mixed sensitivity became central
to understanding robust control. For a multivariable control
system with loop transfer function matrix L(s), design specifications were posed in terms of the sensitivity matrix
S(s) [I L(s)]1 and the complementary sensitivity matrix
T(s) L(s)[I L(s)]1. From these definitions, it follows that
there is the fundamental constraint
S(s) + T (s) = I
(1)
593
W1 (s)S(s)
W2 (s)K(s)S(s)
W3 (s)T (s)
Here W1, W2, W3 are frequency-dependent weights specified by
the designer, L(s) G(s)K(s) is the loop transfer function,
K(s) is the controller, and G(s) is the plant to be controlled.
SOURCES OF UNCERTAINTY
Uncertainty is always present to some degree in mathematical models of physical systems. It arises from ignorance of
physical details, from approximations, or from unpredictable
external effects, such as noise or weather changes.
Models derived from idealized physical laws and theoretical principles are uncertain. This includes, for instance,
Newtons force law, the second law of thermodynamics,
Kirchhoff s current law, Boyles ideal gas law, Maxwells
electromagnetic equations, and the law of supply and demand
from economic theory. Though some laws are very good predictors of behavior, all such laws are idealizations with fundamental limits of accuracy. Moreover, applications of such laws
and principles usually entail assumptions that further degrade the models accuracy of prediction.
Approximations produce modeling uncertainty. Linearized
models are uncertain, and neglected nonlinear distortion becomes uncertainty. Engineers designing stability augmentation systems for aircraft, robot manipulator arms, or similar
mechanical systems employ a rigid-body approximation neglecting flexible bending modes, thereby introducing uncertainty into the mathematical models of these systems. Deliberate truncation of modes to simplify a model is called model
reduction. Uncertainty arises even when all flexible modes
are included in models, because the numerical values in the
equations for all modes are seldom known with precision.
Mathematical models derived from experimental data are
also uncertain. Knowledge of the past is uncertain and predictions of the future are always uncertain. The approximate
curve-fitting processes implicit in most algorithms for system
identification mean that experimentally identified models
usually can not exactly reproduce even the past data. And
even when they do, forecasts of future behavior derived from
such models are still never guaranteed to be accurate.
STABILITY AND ROBUSTNESS
It turns out that the effects of even very small model uncertainties are amplified many times by a feedback system.
Thus, it is vital that control engineers quantify the tolerance
of their designs to such uncertainties. Because modeling uncertainty is usually a dominant factor in determining how
precisely it is possible to control a plants behavior, an engineer wishing to optimize control performance must quantify
ROBUST CONTROL
the intrinsic tradeoff between uncertainty tolerance and control precision, that is the intrinsic tradeoff between robustness and performance.
Examples of extreme sensitivity to small effects abound in
nature. For example, a ball balanced on the top of a hill is
unstable. Left undisturbed, it would, in principle, remain forever at rest on the mountain peak. Yet, a disturbance smaller
than the flapping of a butterfly wing causes the ball to plummet. Such extreme sensitivity to small disturbances is called
instability. Unstable systems are never robust.
On the other hand, a ball at rest in the cusp of a valley is
stable and is not dislodged even by a major earthquake. Further, so long as gravity remains an attractive force, then balls
at rest in valleys remain stably at rest. Not only is the ball
stable, but it remains stable for any large nonnegative variation in the gravitational constant. A ball in a valley is said to
possess the property of stability robustness.
But even the property of stability itself can be extremely
sensitive to small effects. If one had a knob which could reduce the Earths gravitational attraction until gravitational
attraction became negative, then balls in valleys would no
longer have stability robustness. As the pointer on the knob
approached zero, the property of stability itself would be extremely sensitive to the position of the knob. With the knob
set a little bit above zero, a ball in a valley would remain
stably at rest. But with the pointer on the knob ever so
slightly below zero, attraction would become repulsion, and
stability robustness would no longer hold. The ball would be
launched into space in response to the slightest change in the
knobs setting. The property of stability for a ball in a valley
is thus not robust to negative variations in the gravitational
constant.
Of course, although there is little prospect that gravity will
reverse itself, there are many situations encountered by engineers in which stability is not robust to variations in physical parameters.
Robustness in Classical Single-Loop Control
Although the term robustness was not generally used in control theory until the 1970s, robustness has always concerned
feedback control system designers. For example, consider the
flyball governor system invented by James Watt in 1788 to
regulate the rotational velocity of a steam engine. A sketch of
Watts invention is shown in Fig. 2. The flyball governor device is coupled to the main shaft of a steam engine via a pulley. It maintains a set value for the rotational velocity of the
shaft by opening the steam valve when velocity decreases and
closing it when it increases.
A simplified linear model for a steam engine feedback system incorporating a flyball governor is given by the following
differential equations:
d2
d
y(t) + 2 y(t) + y(t) = Ks u(t)
dt 2
dt
d
Flyball governor:
u(t) + u(t) = Kg (r(t) y(t))
dt
Steam engine:
Steam
594
K
r + c1 et + c2 e t sin(t + )
1+K
ROBUST CONTROL
595
MULTIVARIABLE ROBUSTNESS
In the 1960s and 1970s as increasingly complex control systems with multiple feedback loops were deployed, engineers
typically evaluated the robustness of these systems against
variations one-loop-at-at-a-time. That is, they examined the
effects of varying each feedback while holding all the other
gains fixed at their nominal values. This proved unwise, because a system may tolerate large, even infinite, variations in
individual feedback gains, yet may be destabilized by even
very small simultaneous variations in several feedback gains.
The term multivariable robustness refers to the tolerance of
feedback systems to such simultaneous variations.
As early as 1971, H. H. Rosenbrock and M. Athans urged
greater attention to the issue of simultaneous variations in
several feedback gains. And Rosenbrock even developed crude
methods for designing robust controllers for such systems,
provided that the loops were approximately decoupled, a condition that Rosenbrock called diagonal dominance. But, for
many, the telling example that made the case for multivariable robustness theory was the following very simple example
produced by J. C. Doyle in 1978.
1 + 1
u
s 100
1
s2 + 100
1 + 2
Multiplicative
uncertainties
Plant G0 (s)
Unstable
1
Stable
Unstable
1
Unstable
Figure 3. Doyles 1978 two-loop feedback example shows that oneloop-at-a-time stability robustness analysis is unwise. In this case,
the feedback system remains stable for large variations in either of
the two gain perturbations 1 and 2 when the other is fixed to be
zero, but is unstable for small simultaneous variations.
If one closes either one of the two feedback loops around this
system, the open-loop transfer function in the other loop is
simply given by
g(s) =
1
s
whence the system remains stable if the gain in the other loop
assumes any nonnegative value. Moreover, this implies that
gain margin is infinite and the phase margin is 90 in each
loop, when evaluated one-loop-at-a-time. The system seems to
be extraordinarily robust when examined one-loop-at-a-time.
However, this is not the case when gains in both loops are
varied simultaneously, as the stability diagram in Fig. 3 indicates.
Stable
1
Multivariable Robustness
Consider a two-loop multivariable feedback system with nominal loop transfer function
s 100
10(s + 1)
1
G0 (s) = 2
s + 100 10(s + 1)
s 100
10(s + 1)
10(s + 1) s 100
Uncertainty
An important aspect of robust control theory is quantifying
the difference between two transfer function matrices, say
G(s) and G0(s). Typically, G0(s) is a nominal mathematical
model, and G(s) is the corresponding true system and the
difference is called uncertainty. The uncertainty may be
viewed as additive or multiplicative, or it may be regarded in
terms of more exotic linear fractional transformation (LFT) or
normalized coprime matrix fraction description (MFD) transformations. Useful quantitative measures of size are singular
values, H norms, and the gap metric.
Additive and Multiplicative Uncertainty. An additive uncertainty is a matrix, say A(s), which produces the true system
G(s) when added the nominal model G0, that is
(2)
where
P
P(s) = 11
P21
P12
P22
(3)
596
ROBUST CONTROL
where [N0(s), D0(s)] and [N(s), D(s)] are stable normalized coprime left MFDs of G0(s) and G(s), respectively. The gap metric has attracted much interest from theorists because it has
a rich mathematical structure with many subtle symmetries
and geometric interpretations.
G0(s)
K(s)
K(s)
G0(s)
and denoted
G(s)
,P
A = UV
N(s)
N(s)
=
Q(s)
D(s)
D(s)
(4)
is an r r diagonal matrix
and
0
1 0
0
0
2
=.
..
.
.
.
N(s)
U (s) =
D(s)
N(s)
N0 (s)
i ui vi
i=1
0
0
..
Ax
x
x = x21 + x22 + + x2n
The H norm of a stable transfer function matrix, say
P(s), is defined in terms of the largest singular value of its
frequency response
P
, sup
1 [P( j)]
For an unstable transfer function P(s), the H-norm by definition is infinite, that is, whenever P(s) has unstable poles
P
.
ROBUST CONTROL
A
2 max{1 (A), 1 (B)}
max{1 (A), 1 (B)} 1
B
(5)
0
0
0
1
0
0
0
2
= .
..
..
.
.
.
.
Fictitious
uncertainty
n
1 1
, D
n - 1
Plant P(s)
P11
P12
P21
P22
n
( j)]
inf 1 () | det[I Ty 1 u 1 ( j)] = 0
u1
, Km [Ty u
y1
with 1 and, perhaps, subject to an additional constraint D where D is a specified subset of the set of diagonal matrices with H-norm less than one.
Closely related to the canonical robust control problem are
the concept of the multivariable stability margin and the
structured singular value. Consider the system in Fig. 5.
Denote by Ty1u1(s) the closed-loop transfer function from u1 to
y1 with the uncertainty removed. Loosely, Ty1u1(s) is the
closed-loop transfer function that the diagonal uncertainty
sees as it looks back into the rest of the system. Given a stable transfer function matrix Ty1u1(s), the multivariable stability margin Km(Ty1u1) associated with Ty1u1(s), by definition, is
the size of the smallest destabilizing D. The structured
singular value [Ty1u1(s)], by definition, is the reciprocal of
the multivariable stability margin. More formally, for any
stable transfer function Ty1u1( j),
1
[Ty 1 u 1 ( j)]
Uncertainty
597
y2
u2
Controller
K(s)
The problem of choosing the feedback K(s) so that the foregoing H inequality holds is known as the H control problem.
To gain a sense of the tolerance of uncertainty at each frequency, control engineers plot singular value Bode plots, such
as in Fig. 6. The number 1/ 1[Ty1u1( j)] is interpreted as a
lower bound on the size Km( j) of the smallest destabilizing
uncertainty ( j). Hence, singular values provide a convenient upper bound on the structured singular value, namely,
[Ty 1 u 1 ( j)] 1 [Ty 1 u 1 ( j)]
(6)
598
ROBUST CONTROL
matrix
10
103
Peak value:
976.93 =
18.89 =
1.08 =
Matlab function:
sigma.m (complex full-matrix )
ssv.m with perron option (complex )
ssv.m with muopt (real )
102
0
d2
0
0
..
.
0
0
..
.
dn
101
, D diagonal
inf 1 (DTD1 )
100
10-1
d1
0
D=
..
.
100
101
Hz
Computation of via LMIs. In practice, the foregoing optimization over D is usually reduced to an equivalent optimization
problem involving a linear matrix inequality (LMI), namely
max
Figure 6. Singular value Bode plots are widely used to evaluate robustness, but other less-conservative upper-bounds on (Ty1u1) are better when the uncertainty is structured. Shown here are Bode plots
of several upper-bounds computed using the Matlab Robust Control
Toolbox functions for a particular Ty1u1.
subject to
X T X T 0
The greatest value of for which this optimization is feasible
yields (T) , and the optimal diagonal scaling D X1/2.
Solutions to LMI optimization problems are computed via
standard algorithms. An LMI-based D, G-scaling technique
extends the LMI concept to further tighten the upper bound
on when some of the uncertainties i are known to be real
constants. For example, the Matlab Robust Control Toolbox
function muopt.m computes this way.
Kharitonov Theorem
A system is stable if and only if all its poles have negative
real parts. The characteristic polynomial associated with a
transfer function Ty1u1(s) is any polynomial whose roots are its
poles. One very simple but useful test for robustness allows
evaluating the stability of a system whose characteristic polynomial has several uncertain real coefficients. Consider the
polynomial p(s) ansn an1sn1 a1 a0. Further,
suppose that each of the coefficients ai (i 1, . . ., n) is a real
number which varies in the range ai ai ai. The Kharitonov theorem (6) gives necessary and sufficient conditions for
all of the roots of the polynomial p(s) to be robustly stable
(i.e., have negative real parts) for all values of the coefficients
in the specified ranges. The Kharitonov theorem says that robust stability is assured if, and only if, four specific polynomials are stable. The four Kharitonov polynomials are obtained
by setting all of the even and all of the odd coefficients to
their minimal values and maximal values, respectively. Specifically, the four Kharitonov polynomials are
p1 (s) =
a i si +
a i si +
ai si +
i odd
a i si
a i si
i even
i odd
p4 (s) =
i even
i odd
p3 (s) =
a i si
i even
i odd
p2 (s) =
ai si +
i even
a i si
ROBUST CONTROL
p2 (s) = 3s + 5s + 3s + 5
2
, K (s)min
sup [Ty u
stabilizing
1 1
( j)]
(7)
, K (s)min
sup 1 [Ty u
stabilizing
1 1
( j)]
D12 D12 = I
D21 D21 = I
D11 = 0
D22 = 0
(8)
599
D12C1 = 0
B1 D21 = 0
(11)
u2 = F (I QP)1 x
x = (A + QC1T C1 )x + B2 u2 + H(y2 C2 x)
or, equivalently,
opt
, K (s)min
Ty u
stabilizing
1 1
.
(9)
(10)
Optimal H controllers possess the property that their closedloop singular-value frequency-response is completely flat, that
is
opt = Ty 1 u 1 = 1 [Ty 1 u 1 ( j)] for all
This flatness of optimal H controllers is called the all-pass
property, analogous to the flat response of all-pass filters that
arise in circuit theory and signal processing.
H Control Riccati Equations. Suppose that the plant P(s) in
Fig. 5 is given in state-space form as
x
A
y1 = C1
y2
C2
B1
D11
D21
B2
x
D12 u1
D22
u2
2. 1((D12
)D11) 1 and 1(D11(D21)) 1 where (D) denotes an orthogonal matrix whose column span is
null(D).
3. Neither of the Hamiltonians Eqs. (1213) has any
purely imaginary eigenvalues; that is, Imag[i(HP)] 0
i and Imag[i(HQ)] 0 i. (This is necessary and sufficient for the two Riccati equations to have stabilizing
solutions that is, solutions such that A (B1B1
B2B2)P and A Q(C1C1 C2C2) have only eigenvalues
with negative real parts.)
4. Both A B2F and A HC2 are stable (i.e., both have
all eigenvalues with strictly negative real parts). (Note:
In theory, this condition holds if and only if the Riccati
solutions P and Q are both positive-semidefinite matrices, but in practice semidefiniteness cannot be checked
reliably with a finite precision computation.)
600
ROBUST CONTROL
for all i.
D
P(s)
0
0
I
P11
P21
P12
P22
D
0
1
0
I
(14)
Step 2. Compute the solution K(s) to the H optimization problem Eq. (9) for the diagonally scaled
plant Eq. (14) using the -iterative technique,
and set
Ty 1 u 1 = P11 + P12 K(I P11 K)2 P21
Step 3. At each frequency , compute the diagonal scaling D( j) that minimizes 1(DTy1u1D1).
Step 4. Using a curve-fitting method, find a polynomial
transfer function D(s) that approximates the
frequency response D( j) computed in Step 3.
Step 5. Go to Step 1.
There are no theoretical guarantees that this algorithm
converges to a joint minimum in K(s) and D(s). In fact, it may
not even find a local minimum. But, at least it improves the
controller, which is enough for it to be useful in some engineering applications. Refinements on this algorithm have recently appeared in which D, G-iteration handles real uncertainties with less conservatism. Also, a technique for
bypassing the curve fitting in Step 4 is available. The optimal
D(s) is a prespecified order computed via an LMI-related technique [see (22)].
Ty 1 u 1
W1 (s)S(s)
= W2 (s)K(s)S(s)
W3 (s)T (s)
Figure 7 shows how a conventional control system is augmented with weights W1, W2, W3 to produce an augmented
plant P(s) such that the closed-loop Ty1u1 has the indicated
form, namely,
W1 W1 G
0
W2
P(s) =
0 W3 G
G
I
Let us suppose (as is often the case in applications of the theory) that the control-signal weight W2 is absent, so that Ty1u1
reduces to
Ty 1 u 1
W1 (s)S(s)
=
W3 (s)T (s)
W1 ( j)
1
1 [S( j)]
opt
opt
W3 ( j)
ROBUST CONTROL
601
G(s)
y1a
W2(s)
y1b
W3(s)
y1c
u2
y2
Controller
K(s)
dB
1(L)
1(T)
Performance
bound W1
0
1
1(S)
n(L)
y1
+
u1
W1(s)
1
W3
Robustness
bound
Figure 8. The mixed-sensitivity weights W1(s) and W3(s) provide engineers with the means to directly specify the desired shape of the
compensated loop transfer function L(s). The Bode plot of L(s) is sandwiched between the plot of W1(s) and the plot of 1/W3(s). Loop-shaping
is achieved by manipulating the weights W1(s) and W3(s).
BIBLIOGRAPHY
1. O. Mayr, The Origins of Feedback Control, Cambridge, MA: MIT
Press, 1970.
2. M. G. Safonov, Stability and Robustness of Multivariable Feedback
Systems, Cambridge, MA: MIT Press, 1980. Based on the authors
PhD Thesis, Robustness and stability aspects of stochastic multivariable feedback system design, MIT, 1977.
602
MICHAEL G. SAFONOV
University of Southern California
603
604
Thus each plant in the model set corresponds to the selection of one element in each of the uncertain blocks. The nominal plant corresponds to the choice 0 for all elementary uncertain blocks. Note that the assumption that all the
uncertainty balls have unity radius is made at no cost: any
size information can be embedded into known blocks (e.g.,
connected in cascade with the uncertain blocks). It turns out
that, for uncertain block diagrams of this type, the transfer
function (or transfer function matrix) between any two nodes
(or tuples of nodes) in the block diagram is given by a linear
fractional transformation (LFT). Similarly, when such an uncertain plant is connected with a feedback compensator, the
transfer function between any two nodes will also be an LFT.
LFTs have attractive mathematical properties that can be
used to advantage at the modeling, analysis, and synthesis
stages. More on LFTs can be found, e.g., in Ref. 7.
Robust Stability
It should be intuitively clear that block diagrams of the type
just described can always be redrawn in the form of an
M loop as depicted in Fig. 1 (external input and outputs
have been left out). Here M corresponds to the nominal system, which is comprised of closed-loop transfer functions as
elements and has an input channel and an output channel for
each elementary uncertain block, and is a block-diagonal
matrix whose diagonal blocks are the elementary uncertain
blocks. For its generality, the M loop paradigm has found
wide acceptance in robust control (see, for example, Refs. 2
and 48).
Throughout most of this article, we will assume that the
nominal system, or equivalently M, is linear and time-invariant, as are all instances of the uncertainty blocks, equivalently, of . We will also assume that M and all instances of
are in H . In this case, an immediate payoff of the LFT
uncertainty description and the ensuing representation of the
system via the M loop is the following strong form of the
Small Gain Theorem, a necessary and sufficient condition for
well-posedness and stability of the M loop, in the case
where consists of a single uncertainty block, ranging over
the unit ball in H .
Small Gain Theorem. Let M H . Then the M loop is
well-posed and BIBO stable for all H with 1 if
and only if M 1.
As alluded to earlier, the H norm of a causal, stable, continuous-time transfer function matrix M is defined as
M = sup (M( j))
Uncertain Systems
Throughout this article, we consider model sets P with the
following property: P can be represented by a block diagram
with some of the blocks being fully known systems, and others
being elementary uncertain blocks. The latter are elementary sets, namely, unit balls in simple vector spaces. For
example, some uncertainty blocks might be the real interval
[1, 1] and others might be the unit ball of H , the set of
transfer function matrices (linear time-invariant systems)
that are causal and bounded-input bounded-output (BIBO)
stable; the size or H -norm of H is defined to be the
supremum of its largest singular value over the imaginary
axis (continuous time) or unit disk (discrete time).
Figure 1. M loop.
As an example consider a model set of the output multiplicative uncertainty type. Specifically, let
namics, consider now the model set (1 w1c)Pa with Pa explicitly given as
P = (I + w)P0
where P0 is the transfer matrix of a linear, time-invariant
nominal model, w is a scalar weighting transfer function,
and ranges over the unit ball in H . The weight w is introduced to account for the fact that the amount of uncertainty
is usually frequency-dependent; in particular, system dynamics are often poorly known at high frequencies. Suppose that
a feedback controller K has been tentatively selected to stabilize the nominal system P0. (We use the negative feedback
convention, i.e., the loop transfer function is KP0.) Isolating
from the nominal closed-loop system, we obtain an M
loop with
M = wKP0 (I + KP0 )1
Since K stabilizes P0, M is stable.
As a note of interest, we expect a keen connection between
the Small Gain Theorem and the classical Nyquist criterion.
Indeed, this can be best observed by examining single-input/
single-output systems. In such case P and K are scalar, and
thus so is M. Since both M and are stable, Nyquists criterion implies that the M loop is stable whenever the Nyquist plot of M does not encircle the critical point 1 j0.
Clearly, this will be the case for every satisfying ( jw) 1
for all w if and only if M( jw) 1 holds at all frequencies
(including ).
The Small Gain Theorem suggests that one way to obtain
a robustly stable system, or more generally to obtain a robust
design, is to make sure that the H norm of a certain system
transfer function is small enough. This has triggered an entire field of research known as H design, which is discussed
elsewhere in this encyclopedia. The focus of the present article is the case when is block-diagonal (i.e., when the uncertainty model consists of several blocks or, in other words,
when the uncertainty is structured). Typically, two types of
uncertainty blocks are considered in the literature: (i) the set
of real, constant scalar multiples of the identity, with the scalar having magnitude no larger than one, and (ii) the set of
causal and BIBO stable (H ) transfer function matrices, with
H -norm no larger than one. The latter corresponds to unmodeled dynamics. The former, on the other hand, is used to
represent parametric uncertainty, particularly when a same
uncertain parameter affects more than one coefficients in a
transfer function. For example, concurrent variation as a
function of temperature (e.g., dilation) of multiple quantities
in a mechanical system can result in such a block. This description is more general than the simpler scalar nonrepeated blocks.
Examples with structured uncertainty arise with plants
modeled as being affected by uncertainty at more than one
physical location, e.g.,
P = (I + w1 1 )P0 (I + w2 2 )
where both input and output uncertainty are accounted for.
Another instance arises in the context of the robust performance problem, discussed in a later section. For an example
including both parametric uncertainty and unmodeled dy-
605
Pa (s) =
1
sa
w2 P0 (I + P0 K)1
M=
w1 P0 (I + P0 K)1
r 0
=
0 c
w2 KP0 (I + P0 K)1
w1 KP0 (I + P0 K)1
606
with
r := {diag(1r Ik , . . ., mr r Ik m ) : ir R}
1
c :=
C :=
C
{diag(1C , . . ., m
)
C
ic
ic
C}
{ 0 : MDM H + GM H MG 2 D < 0}
k m +m +i k m +m +i
r
c
r
c
}
inf
DD + ,GG
{diag(1c Ik
, . . ., mc c Ik m +m c )
m r +1
r
1
otherwise. It can be checked that for structures simply consisting of one full complex block as in the Small Gain Theorem, (M(j) becomes the largest singular value of M(j),
and M is thus equal to M.
Given a matrix M and a block structure , computing
(M) is generally not an easy task. Indeed, this computation
is known to be NP-hard, even when is simplified to a structure containing only full complex blocks. Thus estimates of
(M), e.g., upper and lower bounds on (M), are often used
instead. These, as well as other properties of , are discussed next.
Let U be the set of unitary matrices in and D be the set
of nonsingular matrices that commute with every . The
latter consist of block-diagonal matrices with scalar multiples
of the identity in correspondence with full complex blocks
(C), and with arbitrary blocks in correspondence with those
constrained to be scalar multiples of the identity (r, c). Then
the following result holds:
Here the superscript H indicates complex conjugate transpose, Q is the subset of consisting of matrices whose complex blocks are unitary, D is the subset of D consisting of
Hermitian positive definite matrices, G is the subset of D
consisting of skew-Hermitian matrices (i.e., GH G) with
zero blocks in correspondence with repeated real blocks in ,
and the sign indicates that the matrix expression is constrained to be negative definite. The lower bound in condition
(2) is always equal to (M). The upper bound is never greater
than that in condition (1) (it reduces to it when G 0 is imposed) and, as was the case for condition (1), can be computed
by solving an LMI problem. See, for example, Section 8.12 in
Ref. 8.
For the class of problems where the matrix M has rank
one, Young (13) showed that the right-hand side in inequalities [Eq. (2)] is equal to (M). In that case, Chen et al. (14)
obtained an explicit formula for (M). Let M baH, where a
and b are column vectors. Let also a and b be partitioned into
subvectors ai and bi compatibly with . For i 1, . . ., mr
mc, let i aiHbi. Moreover, define
m
r +m c
|i | +
i=m r +1
Then,
(M) = inf
x
m
ai 2 bi 2
|Re(i ) + x Im(i )| +
1+x
!
2
(3)
Re(l )
Re(1 )
Im(1 )
Im(l )
Re(k )
,
Im(k )
l
k = 1, . . ., l
x 0 =
;
x k =
DD
i=1
xk =
U U
m r +m c +m C
i=m r +m c +1
(1)
(2)
i=1
|Im(i )|
l
i=1
|Im(i )|
k
l
i=1 |Im(i )|
i=k+1 |Im(i )|
l
k
2
i=1 |Im(i )|
i=k+1 |Im(i )|
1
M
607
(4)
1
M
where now denotes the structured singular value corresponding to the augmented block structure diag(kpkp, )
(i.e., the block-structure corresponding to the actual uncertainty, augmented with a full complex block).
For an example of a typical robust performance problem,
consider an uncertain plant described by the multiplicative
uncertainty model set
P = (I + w1 )P0
with a fixed feedback controller K. It is desired to determine
whether w2S 1 for all possible in a possibly structured
unit uncertainty ball, where S is the sensitivity function (i.e.,
using the negative feedback convention, S (I PK)1).
Here w1 and w2 are stable transfer functions introduced for
frequency-weighting purposes. For simplicity, w1 and w2 are
assumed to be scalars. Using the transformation just outlined, we obtain
w2 (I + P0 K)1
M=
w1 (I + KP0 )1 K
w2 (I + P0 K)1 P0
w1 (I + KP0 )1 KP0
(5)
608
(7)
The problem of concern is to determine if the polynomial family P is robustly Hurwitz stable, by which we mean that every member in P is Hurwitz stable. We shall assume that the
coefficients ak(q) are continuous functions of q. Furthemore,
we assume that an(q) 0 for all q Q (i.e., all polynomials
in P have the same degree). For control system analysis, it is
typical to restrict the polynomial family P to the following
classes, arranged by order of increased complexity.
1. P a: the coefficients ak(q) are affine functions of q. For
example,
p(s, q) = s2 + (q1 + 2q2 + 3)s + (4q1 + 5q2 + 6)
2. P m: the coefficients ak(q) are multiaffine functions of q.
For example,
p(s, q) = s3 + (2q1q2 + 2q1 q3 + q3 + 1)s2
+ (4q2q3 + 5)s + (q1 q2 q3 + 1)
3. P p: the coefficients ak(q) are multivariate polynomials
in q. For example,
n
k=0
ak (q)sk ,
(6)
(8)
k=0
!1/ p
m
|qi |
, 1 p<
q p =
i=1
max |qi |,
p=
1im
(9)
where the qks and the qks are fixed. Kharitonovs original
treatment of the interval polynomial problem is of an algebraic nature. He constructed four extremal members of P ,
K1 (s) = q0 + q1 s + q2 s2 + q3 s3 + q4 s4 + q5 s5 +
K2 (s) = q0 + q1 s + q2 s2 + q3 s3 + q4 s4 + q5 s5 +
K3 (s) = q0 + q1 s + q2 s2 + q3 s3 + q4 s4 + q5 s5 +
K4 (s) = q0 + q1 s + q2 s2 + q3 s3 + q4 s4 + q5 s5 +
later dubbed Kharitonov polynomials. In a remarkable fashion, Kharitonov showed that the stability of the entire interval polynomial family can be ascertained by testing merely
these four.
Kharitonovs Theorem. The interval polynomial is Hurwitz
stable if and only if K1(s), K2(s), K3(s), and K4(s) are all Hurwitz stable.
Subsequent development showed that for polynomials of degree 5, 4, and 3, the test can be further simplified, requiring
checking only 3, 2, and 1 of the four Kharitonov polynomials,
respectively.
Dasgupta (24) gave a geometrical interpretation to Kharitonovs Theorem in frequency domain, which sheds light on
why such a puzzling result would hold. The interpretation
makes use of the concept of value set. The idea is to examine
the image of the polynomial family when s lies on the boundary of the stability region, namely
p( j, Q) = {p( j, q) : q Q}
609
Because p(s, q0) is stable and by assumption an(q) never vanishes on Q, we can conclude from the continuity properties of
the zeros of polynomials that if some member in P is not
Hurwitz stable, then some polynomial in P must have a zero
on the imaginary axis. Thus the entire family P is Hurwitz
stable if and only if
0
/ p( j, Q),
that is, the value set never contains the origin. This is a
rather straightforward fact known as early as in 1929 and is
generally referred to in the literature as the zero exclusion
principle. Note, in particular, that for a polynomial p(s, q) in
the family described by Eq. (11), for any q Q, the real and
imaginary parts of p( j, q) are respectively given by
Re p( j, q) = q0 q2 2 + q4 4
and
Im p( j, q) = j(q1 q3 2 + q5 4 )
Thus the real part (resp. imaginary part) of p( j, q) depends
only on the parameters with even (resp. odd) subscript. For
an interval polynomial, therefore, it becomes clear that
Re K1 ( j) = Re K2 ( j) Re p( j, q) Re K3 ( j) = Re K4 ( j)
Im K1 ( j) = Im K4 ( j) Im p( j, q) Im K3 ( j) = Im K2 ( j)
Because the four Kharitonov polynomials are themselves in
P , it follows that, at each , the value set is a rectangle in
the complex plane with vertices Ki( j), as depicted in Fig. 4.
As increases, the rectangle evolves in a continuous fashion. Its dimensions vary, but its edges remain parallel to the
axes, and the relative positions of the Ki( j)s do not change.
Now, a polynomial p(s) of degree n with positive coefficients
is Hurwitz stable if and only if the phase of p( j) monotonically increases from 0 to n/2 as goes from 0 to . This is
best seen by noting that p(s) can always be factorized as
p(s) an(s si) in terms of its zeros si, and by considering
the phase of each factor separately. When this is understood,
Kharitonovs Theorem becomes self-evident. Indeed, if the
Kis are Hurwitz stable, then the phase of each Ki( j) increases monotonically from 0 to n/2 and the entire Kharitonov rectangle rotates around the origin by a total angle of
n/2 without ever touching it. Stability of P then follows from
the zero exclusion principle.
Kharitonovs Theorem opened an era epitomized by the
search for so-called vertex results in robust stability analysis,
as manifested by the construction of the four Kharitonov polynomials at the vertices of the hyperrectangle Q. This theme
Im
K2(j)
K3(j)
K1(j)
K4(j)
Re
610
persisted in much of the subsequent work. Although Kharitonovs Theorem is without question a milestone in control theory, and perhaps a measured intellectual triumph in general,
we should note that the interval polynomial family is nevertheless a very special instance of an affine uncertainty class,
and hence the theorem has a rather limited scope in application. The quest thus continues.
Edge Theorem
In light of Kharitonovs Theorem, it is tempting to contemplate the possibility that a similar vertex result would hold
for the general affine polynomial family with Q a hyperrectangle. This, unfortunately, turns out to be false; a counterexample can be readily constructed to demonstrate the opposite.
What can be said about a polynomial family P in class P a?
Bartlett et al. (25) provided an answer. In what is now known
as the Edge Theorem, they took the bounding set Q to be a
convex polytope in m. Let qi, i 1, . . ., l, be the vertices of
Q. Then, it is well known that Q can be represented as a
convex hull of the vertices. That is,
Q = conv{q1 , . . ., ql } =
l
i=1
i q i :
i = 1, i 0
i=0
Xr () = (0r + (2 2 )r + (4 4 )r + )1/r
Yr () = (1r + (3 2 )r + (5 4 )r + )1/r
R() = 0 q00 2 q02 2 + 4 q04 4 +
I() = 1 q01 3 q03 2 + 5 q05 4 +
Note that R() Re(p( j, q0)) and I() Im(p( j, q0)).
Tsypkin-Polyak Criterion. The polynomial family P of Eq.
(10) is Hurwitz stable if and only if p(s, q0) is Hurwitz stable,
|R()| p |I()| p
Xr ()
Yr ()
> 1,
(0, )
(12)
M(s) = .
p(s, q0 )
p(s, q0 )
1
To see this, observe that p(s, q) in Eq. (8) is the characteristic
polynomial of the M loop of Fig. 1 with diag(q1
q10, . . ., qm qm0 ). Indeed,
m
pk (s)
p(s, q)
(qk q0k ) =
det(I M(s)) = 1 +
0
p(s, q )
p(s, q0 )
k=0
611
Yr ()/|I()|
if R() = 0
if I() = 0
Xr ()/|R()|
p (M( j)) =
()Y
()
X
r
r
p
p
p
p 1/ p
otherwise
which leads to a similar condition for robust Hurwitz stability. This condition is slightly more general than, but essentially replicates, the graphical criterion by Tsypkin and Polyak. Note that for p , the polynomial family becomes an
interval polynomial, and the stability condition reduces to
checking whether p(s, q0) is Hurwitz stable, q0 1, and
X1 () Y1 ()
,
< 1,
(0, )
min
|R()| |I()|
A little thought reveals that the latter is equivalent to determining whether one of the four conditions Re(K1( j)) 0,
Re(K3( j)) 0, Im(K1( j)) 0, and Im(K3( j)) 0 holds.
Clearly, this is further equivalent to the requirement that the
rectangular value set in Fig. 4 never contains the origin.
Extensions
There is an immense body of literature devoted to polynomial
stability problems. Various extensions to Kharitonovs Theorem have been obtained. They generally fall into the categories of vertex results and frequency-sweeping conditions, consisting of delicate studies and intricate technical details. We
summarize some of the highlights next. A recent and comprehensive account can be found in the books by Barmish (26)
and by Bhattacharyya et al. (36).
612
(13)
Here A, B, and C are known matrices of appropriate dimensions. The matrix A is assumed to be stable. The system uncertainty is represented by a set of allowed values for the
real matrix , which may be unstructured or structured. Typical perturbation classes considered in the literature are as
follows, arranged in increasing order of generality. In all
cases, 0 is given.
Unstructured Perturbation. The set consist of all real
matrices with spectral norm less than a given number:
U = { real : () }
Element-by-Element Perturbation. Each element in
varies in a given interval. Let rij 0 be given. The set
is defined as
r11 11 r1m 1m
.
..
..
E = real : =
..
.
. ,
rn1 n1 rnm nm
i, j
613
There essentially exists no result for the structured stability radius other than those already known for . For the unstructured stability radius, much of the early work was devoted to derivation of bounds. One representative example is
Re[G(s)]
Im[G(s)]
r(A, B, C)
1
C(sI A)1 B
(14)
0
i= j
max{|a ij |, |aij |}
W = [wij ],
wij =
i
= j
|aii |
614
|P|R + RT |P|
2
< 1/
(16)
(17)
r11
.
R = ..
rn1
..
.
r1n
..
.
rnn
k
1
(|PEi + EiT P|)
2 i=1
< 1/
(18)
where, again, P 0 is the unique solution to Eq. (17). Subsequent developments led to further extensions for problems
with even more detailed uncertainty descriptions. For example, the ks may be allowed to vary in asymmetric intervals.
Moreover, because rather obviously any interval matrix can
be represented alternatively in the form of A : E,
these conditions can be applied to determine the Hurwitz stability of an interval matrix as well.
Yet another issue clearly of interest is whether it is possible to derive vertex versions of these sufficient conditions.
Boyd and Yang (44) examined stability problems for matrix
polytopes. Specifically, they postulated the uncertainty description
A = conv{A1 , . . ., Ak }
A sufficient condition for A to be Hurwitz stable can be easily
found to be the existence of a P 0 such that
i = 1, . . ., k
(19)
(15)
k
1
(PEi + EiT P)
2 i=1 i
< 1/
(20)
615
BIBLIOGRAPHY
1. J. Ackermann, Robust Control Systems with Uncertain Physical
Parameters, London: Springer-Verlag, 1996.
616
2. J. Doyle, Analysis of feedback systems with structured uncertainties, IEE Proc., 129-D (6): 242250, 1982.
3. V. L. Kharitonov, Asymptotic stability of an equilibrium position
of a family of systems of linear differential equations, Differentialnye Uraveniya, 14 (11): 14831485, 1978.
4. M. G. Safonov, Stability margins of diagonally perturbed multivariable feedback systems, IEE Proc., 129-D: 251256, 1982.
5. A. Packard and J. C. Doyle, The complex structured singular
value, Automatica, 29 (1): 71109, 1993.
6. W. S. Levine (ed.), The Control Handbook, Boca Raton, FL: CRC
Press, 1996.
7. K. Zhou, J. C. Doyle, and K. Glover, Robust and Optimal Control,
Upper Saddle River, NJ: Prentice-Hall, 1996.
8. K. Zhou and J. C. Doyle, Essentials of Robust Control, Upper Saddle River, NJ: Prentice-Hall, 1998.
9. A. L. Tits and V. Balakrishnan, Small- theorems with frequency-dependent uncertainty bounds, Math. Control. Signals,
Syst., in press, 1998.
10. J. M. Krause, Structured singular value analysis of multidimensional system stability, IEEE Trans. Autom. Control, 34 (6): 638
639, 1989.
11. J. Chen and H. A. Latchman, Frequency sweeping tests for stability independent of delay, IEEE Trans. Autom. Control, 40 (9):
16401645, 1995.
12. M. K. H. Fan, A. L. Tits, and J. C. Doyle, Robustness in the presence of mixed parametric uncertainty and unmodelled dynamics,
IEEE Trans. Autom. Control, 36 (1): 2538, 1991.
13. P. M. Young, The rank one mixed problem and Kharitonovtype analysis, Automatica, 30: 18991911, 1994.
14. J. Chen, M. K. H. Fan, and C. N. Nett, Structured singular value
and stability of uncertain polynomials, Part 1: The generalized ,
Syst. Control Lett., 23 (1): 5365, 1994.
15. B. Kouvaritakis and H. Latchman, Necessary and sufficient stability criterion for systems with structured uncertainties: The
major principle direction alignment principle, Int. J. Control, 42
(3): 575598, 1985.
16. J. Chen, M. K. H. Fan, and C. N. Nett, Structured singular values
with non-diagonal structures, Part 1: Characterizations, IEEE
Trans. Autom. Control, 41 (10): 15071511, 1996.
17. J. Chen, M. K. H. Fan, and C. N. Nett, Structured singular values
with non-diagonal structures, Part 2: Computations, IEEE Trans.
Autom. Control, 41 (10): 15111516, 1996.
18. A. L. Tits, V. Balakrishnan, and L. Lee, Robustness under
bounded uncertainty with phase information, IEEE TAC, in press.
19. M. Khammash and J. B. Pearson, Robust disturbance rejection in
l1-optimal control systems, Syst. Control Lett., 14: 93101, 1990.
20. M. Khammash and J. B. Pearson, Performance robustness of discrete-time systems with structured uncertainty, IEEE Trans. Autom. Control, 36 (4): 398412, 1991.
21. J. Shamma, Robust stability with time-varying structured uncertainty. IEEE Trans. Autom. Control, 39 (4): 714724, 1994.
22. A. Megretski, Necessary and sufficient conditions of stability: A
multiloop generalization of the circle criterion, IEEE Trans. Autom. Control, 38 (5): 753756, 1993.
23. K. Poolla and A. Tikku, Robust performance against time-varying
structured perturbations, IEEE Trans. Autom. Control, 40 (9):
15891602, 1995.
24. S. Dasgupta, Kharitonovs theorem revisited, Syst. Control Lett.,
11: 381384, 1988.
25. A. C. Bartlett, C. V. Hollot, and L. Huang, Root locations of an
entire polytope of polynomials: It suffices to check the edges,
Math. Control, Signals, Syst., 1: 6171, 1988.
26. B. R. Barmish, New Tools for Robustness of Linear Systems, New
York: Macmillan, 1994.
ROUNDOFF ERRORS
48. O. Toker, On the conservatism of upper bound tests for structured singular value analysis, Proc. 35th Conf. Decision Control,
Kobe, Japan: 1996, pp. 12951300.
49. M. Fu, The real is hardly approximable, IEEE Trans. Autom.
Control, 42 (9): 12861288, 1997.
50. G. E. Coxson and C. L. DeMarco, The computation complexity of
approximating the minimal perturbation scaling to achieve instability in an interval matrix, Math. Control, Signals Syst., 7 (4):
279292, 1994.
51. S. Boyd et al., Linear Matrix Inequalities in System and Control
Theory, Studies in Applied Mathematics, Philadelphia: SIAM,
1994, vol. 15.
52. A. Megretski and A. Rantzer, System analysis via integral quadratic constraints, IEEE Trans. Autom. Control, 42 (6): 819
830, 1997.
53. R. F. Stengel and L. R. Ray, Stochastic robustness of linear timeinvariant systems, IEEE Trans. Autom. Control, 36 (1): 8287,
1992.
54. B. R. Barmish and C. M. Lagoa, The uniform distribution: A rigorous justification for its use in robustness analysis, Proc. 35th
IEEE Conf. Decision Control, 1996, pp. 34183423.
55. R. Tempo, E. W. Bai, and F. Dabbene, Probabilistic robustness
analysis: Explicit bounds for the minimum number of samples,
Proc. 35th IEEE Conf. Decision Control, 1996, pp. 34243428.
56. X. Chen and K. Zhou, On the probabilistic characterization of
model uncertainty and robustness, Proc. 36th IEEE Conf. Decision
Control, 1997, pp. 38163821.
JIE CHEN
University of California
ANDRE L. TITS
University of Maryland
617
403
Two basic and desirable features of an approximation theory are the existence of:
a constructive methodology for obtaining reduced-complexity models
an appropriate quantization of the approximation error,
in other words, the estimation of some measure of the
error between the given high-complexity model and the
derived reduced-complexity model(s)
The importance of the second item cannot be overemphasized:
In approximating a system, one wishes to have some idea of
what has been eliminated.
Often an additional desirable feature consists in
looking for reduced-complexity models within a specified
class, e.g. the class of stable systems
For instance, in case the nominal model is stable, if the intended use of a reduced-complexity model is in open-loop, it
is imperative that the latter also be stable.
We will deal with linear, time-invariant, discrete- and continuous-time, finite-dimensional systems described by convolution sums or integrals. In addition, we will only consider the
approximation of stable systems. Most of the approximation
methods availablefor example, Pade approximationfail to
satisfy the requirements listed above. We will discuss the theory of Hankel-norm approximation and the related approximation by balanced truncation. The size of systemsincluding
error systemsis measured in terms of appropriately defined
2-norms, and complexity is measured in terms of the (least)
number of state variables (i.e., first-order differential or difference equations) needed to describe the system.
This approach to model reduction has an interesting history. For operators in finite-dimensional spaces (matrices),
the problem of optimal approximation by operators of lower
rank is solved by the SchmidtMirsky theorem (see Ref. 1).
Although this is not a convex problem and consequently conventional optimization methods do not apply, it can be solved
explicitly by using an ad hoc tool: the singular value decomposition (SVD) of the operator. The solution involves the truncation of small singular values.
In the same vein, a linear dynamical system can be represented by means of a structured (Hankel) linear operator in
appropriate infinite dimensional spaces. The AdamjanArov
Krein (AAK) theory (see Refs. 2 and 3) generalizes the
SchmidtMirsky result to dynamical systems, that is to structured operators in infinite-dimensional spaces. This is also
known as optimal approximation in the Hankel norm. The original setting of the AAK theory was functional analytic. Subsequent developments due to Glover (4) resulted in a simplified
linear algebraic framework. This setting made the theory
quite transparent by providing explicit formulae for the quantities involved.
The Hankel norm approximation problem can be addressed and solved both for discrete- and continuous-time systems. However, the following is a fact: The discrete-time case
is closer to that of finite-dimensional operators and to the
SchmidtMirsky result. Therefore, the intuitive understanding of the results in this case is more straightforward. In the
continuous-time case, on the other hand, while the interpreta-
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright # 1999 John Wiley & Sons, Inc.
404
:=
x21 + + x2n
:= sup
2-ind
x = 0
Ax 2
x 2
(1)
Ax 22
x A Ax
max (A A)
=
x 22
x x
where ( )* denotes complex conjugation and transposition. By
choosing x to be the eigenvector corresponding to the largest
eigenvalue of A*A, the above upper bound is attained, and
hence the 2-induced norm of A is equal to the square root of
the largest eigenvalue of A*A:
2-ind
max (A A)
:=
trace(A A) =
A2i, j
AX
2-ind
k+1 (A)
(5)
X :=
k
i ui vi
(6)
X (1 , . . ., k ) :=
A A = V SV V , AA = USU U
k
(i i )ui vi ,
0 i k+1, i = 1, . . ., k
i=1
(7)
where U, V are (square) orthogonal matrices of size n, m, respectively (i.e., UU* In, VV* Im). Furthermore SV, SU are
diagonal, and assuming that n m we have
SV = diag(12 , . . ., n2 , 0, . . ., 0),
i i+1 0
12 + + n2
i=1
n
(3)
A=
i, j
SU = diag(12 , . . ., n2 ),
(2)
405
(4)
i=1
:=
406
:=
+ X (1)
2
F
+ X (0)
2
F
+ X (1)
2
F
The space of all p m sequences having finite 2-norm is denoted by 2pm(). Given the linear map A : X Y, where X, Y
are subspaces of 2(), the norm induced by the 2-norm in the
domain and range is the 2-induced norm, defined by Eq. (1);
again Eq. (2) holds.
Linear, Discrete-Time Systems
A linear, time-invariant, finite-dimensional, discrete-time system is defined as follows:
:
tZ
A
C
B
D
R(n+ p)(n+m)
(8)
(9)
t
=
..
.
h(0)
h(1)
h(2)
h(3)
..
.
h(0)
h(1)
h(2)
..
.
S
A=
A+
0
B=
B+
B
C (C+ C )
where
+ D + C (zI A )1 B = H (z)
h (t )u( ), t Z (10)
h(0)
h(1)
..
.
0
A
0 . 1 0.
BB .. CC BB . .
BBy(2)CC BB
BBy(1)CC = BB
BB y(0) CC BB
BB y(1) CC BB
@
A @
1 i p, 1 j m
h2 := h2+ + h2
S : u y,
pi j (z)
t
1
,
H (z) :=
h (t)z = C(zI A) B + D =
qi j (z)
t=0
h(0)
..
.
..
1
CC
CC
CC
CC
CC
A
This is equivalent to trading the lack of stability (poles outside the unit disk) for the lack of causality (h2 is nonzero for
negative time). Thus a system with poles both inside and outside of the unit disk (but not on the unit circle) will be interpreted as a possibly antistable 2 system 2, by defining the
impulse response to be nonzero for negative time. Consequently h2 2(). The matrix representation of the corresponding convolution operator is
0 . 1
BB .. CC
BBu(2)CC
BBu(1)CC
BB u(0) CC
BB u(1) CC
@
A
..
.
(11)
0.
BB . .
BB
B
S2 = B
BB
BB
@
..
.
h2+ (0)
h2+ (1)
h2+ (2)
h2+ (3)
..
.
..
.
h2 (1)
h2+ (0)
h2+ (1)
h2+ (2)
..
.
..
.
h2 (2)
h2 (1)
h2+ (0)
h2+ (1)
..
.
..
.
h2 (3)
h2 (2)
h2 (1)
h2+ (0)
..
.
1
CC
C
C
C
C
C
CC
C
A
..
.
(13)
= u
(14)
|z| = 1
2-ind
:= S
2-ind
= sup
u = 0
Due to the equivalence between the time and frequency domains (the Fourier transform is an isometric isomorphism),
this norm can also be defined in the frequency domain. In
particular,
2-ind
2-ind
2-ind
:= S
2-ind
= H
H : m (Z) p (Z+)
u y+ ,
where y+ (t) =
1
=
h (t )u ( ),
t0
S u 2
u 2
(15)
407
(16)
H is called the Hankel operator of . Its matrix representation in the canonical bases is
y(0)
y(1)
y(2)
..
.
=
h(1)
h(2)
h(3)
..
.
h(2) h(3)
h(3) h(4)
h(4) h(5)
..
..
.
.
..
.
u(1)
u(2)
u(3)
..
.
H
(17)
Thus the Hankel operator of maps past inputs into future
outputs. It has a number of properties given next. The first
one for single-input single-output systems is due to Kronecker.
Proposition 2. Given the system defined by Eq. (8), the
rank of H is at most n. The rank is exactly n if, and only if,
the system is reachable and observable. Furthermore, if
is stable, H has a finite set of nonzero singular values.
In order to compute the singular values of the Hankel operator, we define the reachability matrix R and the observability matrix O :
R(A, B) = [B AB A2 B
],
O (C, A) = [R(A , C )]
408
P := RR =
At BB (A )t ,
Q := O O =
t0
(A )t CCAt
t0
(18)
The quantities P and Q are n n symmetric, positive semidefinite matrices. They are positive definite if and only if is
reachable and observable. The grammians are (the unique)
solutions of the linear matrix equations:
AP A + BB = P ,
A QA + CC = Q
(19)
2
= i2 ui RR
OO Rui = i Rui
Q
(20)
q
ri = n
(21)
i=1
k+1 (H )
2-ind
H H
2-ind
= k+1 (H )
(24)
< k ()
:= 1 ()
( )
(23)
H H ui = i2 ui R O ORui
P
2-ind
(22)
e
Figure 1. Construction of approximants.
k+1 ()
k+1 ()
stable approximant of in the 2-induced norm of the convolution operator (i.e., the norm).
(b) We are given a stable system and seek to compute an
approximant in the same class (i.e., stable). In order to
achieve this, the construction given above takes us outside
this class of systems, since the all-pass dilation system has
poles both inside and outside the unit circle. In terms of matrices, we start with a system whose convolution operator S
is a (block) lower triangular Toeplitz matrix. We then compute a (block) Toeplitz matrix S , which is no longer lower
triangular, such that the difference S S is unitary. It
then follows that the lower left-hand portion of S , which is
the Hankel matrix H , has rank r and approximates the Hankel matrix H , so that the 2-norm of the error satisfies
Eq. (25).
(c) The suboptimal and optimal approximants can be constructed using explicit formulae. For continuous-time systems, see the section entitled Construction Formulas for
Hankel-Norm Approximants.
CONSTRUCTION OF APPROXIMANTS
The purpose of this section is to present, and to a certain extent derive, formulas for suboptimal and optimal approximants in the Hankel norm. Because of Theorem 3, all we need
is the ability to construct all-pass dilations of a given system.
To this goal, the first subsection is dedicated to the presentation of important aspects of the theory of linear, continuoustime systems; these facts are used in the second subsection.
The closely related approach to system approximation by balanced truncation is briefly discussed in the section entitled
Balanced Realizations and Balanced Model Reduction.
Linear, Continuous-Time Systems
L 2 Linear Systems. For continuous-time functions, let
(25)
(26)
L n (I ) := { f : I Rn , I R}
Frequent choices of I : I , I or I . The 2-norm
of a function f is
In case k1() ,
k+1 () =
409
:=
tI
f (t) 22 dt
1/2
f L n (I )
=
L2n (I ) := { f L n (I ), f
< }
=
F
:=
t I
F (t)
2
F
1/2
dt
410
:
dx(t)
= Ax(t) + Bu(t)
,
dt
y(t) = Cx(t) + Du(t)
tR
A
=
C
B
D
(n+ p)(n+m)
(27)
In analogy to the discrete-time case, we define the continuous-time unit step function I(t) ( I(t) 1 for t 0, and zero
otherwise) and the delta distribution . The impulse response
of this system is
h (t) = CeAt BI(t) + (t)D
(28)
A+
0
0
A
B=
B+
B
S : u y,
C = (C+
C )
h2 := h2+ + h2
2-ind
:= S
h2 (t) := C e
tR
(30)
(29)
B+ I(t) + (t)D
B I(t)
2-ind
= H
(31)
eAt BB eA t dt,
Q :=
eA t CCeAt dt
(32)
A t
h (t )u( ) d ,
where
h2+ (t) := C+ e
P :=
A+ t
,
A Q + QA + CC = 0
(33)
(34)
A Q + QA + CC = 0
QB + C D = 0
(35)
AP + P A + BB = 0,
2-ind
A Transformation Between Continuous- and Discrete-Time Systems. As mentioned in the introductory paragraphs of this article, the formulas for the construction of approximants will
be given for the continuous-time case because they are generally simpler than their discrete-time counterparts. Thus, in
order to apply the formulas to discrete-time systems, the system will have to be transformed to continuous-time first, and
at the end the continuous-time optimal or suboptimal approximants obtained will have to be transformed back to discretetime approximants.
One transformation between continuous- and discrete-time
systems is given by the bilinear transformation z (1
s)/(1 s) of the complex plane onto itself. The resulting relationship between the transfer function Hc(s) of a continuoustime system and that of the corresponding discrete-time system Hd(z) is
Hc (s) = Hd
1 + s
1s
A
C
B
,
D
d :=
F
H
G
J
A Q + QA + CC = 0
u y+ , where y+ (t) =
= +
Lemma 2. Let the distinct Hankel singular values of the stable system be i, i 1, . . ., q; following holds: 2-ind
2(1 q).
c :=
D D = 2 Im
The solutions P and Q of the Lyapunov equations
411
h (t )u ( ) d ,
t0
(36)
H is called the Hankel operator of . Unlike in the discretetime case, however (see Eq. (17)), H has no matrix representation.
It turns out that just as in Eq. (20), it can be shown that
the nonzero singular values of the continuous-time Hankel
operator are the eigenvalues of the product of the two grammians:
i2 (H ) = i (H H ) = i (PQ)
(37)
Discrete-Time
A, B, C, D
z
1s
1s
A (F I)1(F I)
B 2(F I)1G
C 2H(F I)1
D J H(F I)1G
s
z1
z1
F (I A)(I A)1
G 2(I A)1B
H 2C(I A)1
J D C(I A)1B
F, G, H, J
412
K = diag(
...
, 1, 1, 1)
of appropriate size. Given a polynomial a with real coefficients, the polynomial c* is defined as c(s)* : c(s). This
means that
c = K c
The basic construction given in Theorem 3 hinges on the construction of an -all-pass dilation of . Let
H (s) =
p(s)
,
q(s)
H (s) =
p(s)
q(s)
(38)
This polynomial equation can be rewritten as a matrix equation involving the quantities defined above:
T(q))
q
p
=0
(39)
The solution of this set of linear equations provides the coefficients of the -all pass dilation system . Furthermore, this
system can be solved for both the suboptimal i and the
optimal i cases. We will illustrate the features of this
approach by means of a simple example. For an alternative
approach along similar lines, see Ref. 9.
Example. Let be a second-order system, that is, n 2. If
we normalize the coefficient of the highest power of q, that
is, q2 1, we obtain the following system of equations:
0
p2 q2
p1 + q1
p0 q0
0
0
0
p2 + q2
p1 q1
p0 + q0
W ()
q2
q1
q0
0
0
0
q2
q1
q0
0
0
0
q2
q1
q0
q 1
q 0
p 2
p 1
p 0
p2 + q2
p1 q1
p0 + q0
0
0
This can be solved for all which are not roots of the equation
det W() 0. The latter is a polynomial equation of second
degree; there are thus two values of , 1, and 2, for which
the determinant of W is zero. It can be shown that the roots of
this determinant are the eigenvalues of the Hankel operator
H ; since in the single-input single-output case H is selfadjoint (symmetric), the absolute values of 1 and 2, are the
singular values of H . Thus both suboptimal and optimal approximants can be computed this way. (See also the first example of the section entitled Examples.)
State-Space Construction for Square Systems: Suboptimal
Case. In this section we will discuss the construction of ap-
B
D
A
C
=
A
B
B
A
C
e =
C
(40)
According to the last characterization of Proposition 7, e is all-pass iff the corresponding grammians and D matrices satisfy
= 2 Im
D
D
Pe Qe = 2 I2n ,
(41)
p
?
?
?
q
?
?
?
1
0
0
1
p
1 pq
q
1 pq
q(1 pq)
1
p
1 pq
=
1
0
0
1
(42)
(43)
Pe =
P
Q
Qe =
Q
In
In
1 P
(44)
413
)
A = A + C (CP DB
B = QB + C D
(45)
)
C = (CP DB
has exwhere * : (*)1. There remains to show is that A
actly k stable eigenvalues. From the Lyapunov equation for
1P . By construction [see Eq.
Q e it also follows that Q
(43)], has k positive and n k negative eigenvalues; the
(i.e., in (Q
) k, 0, n k). Furthersame holds true for Q
, A
is observable, because otherwise A has
more, the pair C
eigenvalues on the j axis, which is a contradiction to the
original assumption that A is stable. Then by Proposition 6,
, which completes the
is equal to that of A
the inertia of Q
proof.
Corollary 1. The system has dimension n and the dilated
system e has dimension 2n. The stable subsystem has
dimension k.
State-Space Construction for Square Systems: Optimal Case.
The construction of the previous section can be extended to
include the optimal case. This section presents the formulae
which first appeared in Section 6 of Ref. 4.
In this case we need to construct the all-pass dilation e
for equal to the lower bound in Eq. (25), namely k1().
Thus, k1 is an eigenvalue of the product of the grammians
P Q of multiplicity, say r. There exists a basis change in the
state space such that
P =
Ir k+1
0
0
P2
Q=
Ir k+1
0
0
Q2
(46)
The balancing transformation Eq. (60) discussed in the section entitled Balanced Realizations and Balanced Model Reduction accomplishes this goal.
Clearly, only the pair P 2 and Q 2 needs to be dilated. As in
the suboptimal case, explicit formulae for can be obtained
by using Eq. (35) of Proposition 7. Partition A, B, and C conformally with P and Q :
A=
A11
A21
A12
A22
B=
B1
B2
C = (C1
C2 )
UU = Im
Using Eqs. (45) we construct an all-pass dilation of the subsystem (A22, B21, C12); solving the corresponding equations we
obtain
= U
D
k+1
B = Q2 B2 + C2 D
)( )1
C = (C2 P2 DB
2
2
A = A22 + C2C
(47)
414
2
where 2 : k1
P 2Q 2 (nr)(nr). Hence unlike the subop is not arbitrary. In the single-input single-outtimal case, D
is completely determined by the above relationput case, D
ship (is either 1 or 1) and hence there is a unique optimal
approximant. This is not true for systems with more than one
input and/or more than one output. Furthermore, in the onestep reduction case (i.e., k1 n) the all-pass dilation system
has no antistable part; that is, it is the optimal Hankelnorm approximant.
where
C :=
A :=
Q :=
C
0
0
B
A
0
0
A
In
P
Ip
0
0
Im
R2n2n
R2n2n ,
(s) =
:=
11
21
I
12
22
1 (
)
2 (
)
(s)
:= (s)
Im
=
C(sI A)1 1 B
Im B (sI + A )1 Q 1 B
(49)
(50)
(52)
D0
i = 1, . . ., n rq
(53)
H1 (s) + + Hq (s)
q
1 1
P C
p + C(sI A)
B (sI + A )1 C
i (q ) = i (),
H (s) D0
R( p+m)( p+m)
q)
q := (
R( p+m)2n
Q
In
J :=
(48)
(51)
1 + + q
2(1 + + q )
(54)
1 + + q
(55)
) + + nk1 (
)
1 (
+ D0
k+1
)
k+1 () + 1 (
(56)
2(k+1 + + n )
(57)
The first upper bound in Eq. (56) above is the tightest; but
both a D term and the singular values of the antistable
part of the all-pass dilation are needed. The second upper
bound in Eq. (56) is the second-tightest; it only requires
knowledge of the optimal D term. Finally the upper bound in
Eq. (57) is the least tight among these three. It is, however,
the most useful, since it can be determined a priori, that is,
before the computation of the approximants, once the singular values of the original system are known.
Remark 4. (a) In a similar way a bound for the infinity norm
of suboptimal approximants can be obtained. Given , let
be an all-pass dilation satisfying Eq. (25). Then, similarly to
Eq. (56), the following holds:
+
r+1
2( + r+1 + + n )
=
= q
(59)
= T QT 1 P Q
= T (PQ)T 1
Q
A model reduction method which is closely related to the Hankel-norm approximation method is approximation by balanced truncation. This involves a particular realization of a
linear system given by Eq. (27), called balanced realization
(the D matrix is irrelevant in this case). We start by explaining the rational behind this particular state-space realization.
The Concept of Balancing. Consider a stable system with
positive definite reachability and observability grammians P
and Q . It can be shown that the minimal energy required for
the transfer of the state of the system from 0 to some final
state xr is
xr P 1 xr
415
(58)
Similarly, the largest observation energy produced by observing the output, when the initial state of the system is xo and
(60)
416
A=
B=
A11
A21
A12
A22
B1
B2
The systems
i :=
S=
S1
0
0
S2
,
(61)
C = (C1 C2 )
A
ii
Bi
Ci
i = 1, 2
2(k+1 + + q )
(62)
2 trace(S2 )
Remark 5. (a) The last part of the above theorems says that
if the neglected singular values are small, then the Bode plots
of and 1 are guaranteed to be close in the H norm. The
difference between part 3 for continuous- and discrete-time
systems above is that the multiplicities of the neglected singular values do not enter in the upper bound for continuoustime systems.
(b) Proposition 8 implies that the bilinear transformation
between discrete- and continuous-time systems preserves balancing (see also A Simple Discrete-Time Example below).
(c) Let hank and bal be the reduced-order systems obtained
by one step Hankel-norm approximation, and one step balanced truncation, respectively. It can be shown that hank
bal is all-pass with norm q; it readily follows that
bal
2q
and
bal
2q
1
1
21
1 + 2
A=
1
1
1 + 2
22
C = (1 1),
D=0
B=
1
,
1
We wish to compute the suboptimal Hankel-norm approximants for 1
2. Then we will compute the limit of this
family for 2 and 1, and we will show that the system obtained is indeed the optimal approximant. From Eq.
(43), 2I2 S2 diag(2 12, 2 22); the inertia of is
1, 0, 1; furthermore, from Eq. (45) we obtain
A =
B =
C =
1
21 ( + 1 )
2
(1 + 2 )( + 1 )
1
2
1
+ 1
,
1
+ 2
1
(1 + 2 )( + 2 )
2
22 ( + 2 )
C =
2 1
22 (1 + 2 )
2 1
21 (1 + 2 )
0
1
1 + 2
1
22
B =
2 1
,
21 (1 + 2 )
d,2 =
,
d,1 =
1 2
,
21 (1 + 2 )
=
D
2
B = 2 1 ,
C=
1
,
1 + 2
=
D
2
. If we choose it to
Equations (45) depend on the choice of D
be , the limit still exists and gives a realization of the optimal system which is equivalent to A, B, C, D given above.
Finally, if 1, after a pole-zero cancellation, we obtain
the following reachable and observable approximant:
A =
3 55
10
B = 1 2 ,
C =
1
,
1 + 2
=
D
1
c =
A
C
G2
J2
F1
H1
G1
J1
0
=
=
0
0
0
0
1 2 2
2 2
2
2 2
+
1
2
2
1 + 2 2
(63)
F2
H2
B
=
D
c,2 =
z2 + 1
z3
5+1
2
Notice that d,2 is balanced, but has singular values which are
different from 1 and 2. d,1 is also balanced since G1 H2,
but its grammians are not equal to 1.
Let c denote the continuous-time system obtained from
d by means of the bilinear transformation described in the
section entitled A Transformation Between Continuous- and
Discrete-Time Systems
H(z) =
1 =
P = Q = S = diag(1 , 2 , 3 ),
51
3 =
2 = 1,
2
0
,
0
F G
d :=
=
H J
0
0
0
3 5+5
1/4
,
=
=5
,
10
=
D
A =
417
c,1 =
A2
C2
B2
D2
A1
C1
B1
D1
1 2 2
2
+
=
1 2 2
+
+
2
2
2
d,2 =
d,1 =
F2
H 2
G 2
J
F1
H 1
G 1
J1
0
=
=
3 /2
( + )3 /2 2
( + )3 /2 2
2 ( 2 + )2 /(1 + 2 )
418
The conclusion is that d,2 and d,1 are balanced and different
from d,2 and d,1. It is interesting to notice that the singular
value of d,1 is 2 (1 1)/ 5, while that of d,1 is 1; 2
satisfies 2 2 1. Furthermore, the singular values of
d,2 are 52 /4, 52 /4 which satisfy the following interlacing
inequalities:
3 <
01
BB0
BB
= B1
BB0
@
0
1
0
0
..
.
1
0
0
0
1
C
CC
CC
CC
A
.
0
0
0
0
Hd (z) Hd,2 (z) = 3
1
0
1
0
1
0
3
5
1
0
0
0
2
0
0
1
0
01
BB 0
BB
B3
=B0
BB 2
B@3
0
1
0
r 1
3 C
5 C
C
0 C
r CCA
1
1
(64)
1 + s
1s
2(s3 s2 + s 1)
(s + 1)3
z
z 2 3
1 3 z 2
z3 (z2 3 )
0
3
0
32
0
..
.
3
0
32
0
33
32
0
33
0
34
0
32
0
33
0
1
C
CC
CC
CC
CC
A
.
..
is
In this particular case the 3 3 principal submatrix of H
also an optimal approximant of the corresponding submatrix
of H .
0r
BB 15
0
BB 0
0
B r
3 B
@ 3
..
r 1
3 C
5 C
C
0 C
r CCA
1
z+1
0r
BB 15
1
B 0
0 =B
BBr
0
@
z 1
(s2 1)
2
(1 3 )s + 21 s + (1 3 )
0
3
0
3
0
0r
BB 15
3
B 0
0
=B
BBr
32
@ 3
0
1
0
1 + 32
0
0
r 1
3 C
5 C
C
0 C
r CCA
1
0r
BB 15
0
B 0
0 B
B r
0 B
@ 3
0
3
0
0
1
0
r 1
3 C
5C
C
0 C
r CCA
1
(65)
1 + 3 z2
z3 (z2 3 )
1
1 3 z 2
+ 2
z
z(z2 3 )
A Higher-Order Example
In our last example, we will approximate four well-known
types of analog filters by means of balanced truncation and
Hankel norm approximation. These are:
1.
2.
3.
4.
B Butterworth
C1 Chebyshev 1:
C2 Chebyshev 2:
E Elliptic:
HSV Butterworth N = 20
HSV Chebyshev 1 N = 20
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
10
15
20
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
10
Butterworth
9.9998e-01
9.9952e-01
9.9376e-01
9.5558e-01
8.2114e-01
5.6772e-01
2.9670e-01
1.1874e-01
3.8393e-02
1.0402e-02
2.3907e-03
4.6611e-04
7.6629e-05
1.0507e-05
1.1813e-06
1.0622e-07
7.3520e-09
3.6804e-10
1.1868e-11
1.8521e-13
15
20
Chebyshev 1
9.5000e-01
9.5000e-01
9.5000e-01
9.4996e-01
9.4962e-01
9.4710e-01
9.3362e-01
8.8268e-01
7.5538e-01
5.5133e-01
3.3740e-01
1.8209e-01
9.8132e-02
6.3256e-02
5.2605e-02
5.0362e-02
5.0036e-02
5.0003e-02
5.0000e-02
5.0000e-02
10
15
) 9 ()
+ + 11 (
20
HSV Elliptic N = 20
1
)
Bound 11 := 9 () + 1 (
HSV Chebyshev 2 N = 20
1
419
+ + 20 () =: Bound 12
)
Bound 21 := 11 () + 1 (
) 11 ()
+ + 9 (
+ + 20 () =: Bound 22
Finally, Figure 4 shows the amplitude Bode plots of the
ROMs obtained by Hankel and balanced reductions.
Chebyshev 2
8.9759e-01
7.3975e-01
5.1523e-01
3.0731e-01
1.6837e-01
9.5205e-02
6.3885e-02
5.3342e-02
5.0643e-02
5.0103e-02
5.0014e-02
5.0002e-02
5.0000e-02
5.0000e-02
5.0000e-02
5.0000e-02
5.0000e-02
5.0000e-02
5.0000e-02
5.0000e-02
10
15
20
Elliptic
9.8803e-01
9.8105e-01
9.6566e-01
9.3602e-01
8.8469e-01
8.0582e-01
6.9997e-01
5.7677e-01
4.5166e-01
3.3873e-01
2.4576e-01
1.7413e-01
1.2135e-01
8.3613e-02
5.7195e-02
3.9014e-02
2.6729e-02
1.8650e-02
1.3613e-02
1.0869e-02
CONCLUSION
The computational algorithms that emerged from balancing
and Hankel-norm model reduction have found their way to
software packages. The Matlab toolboxes, robust control toolbox (13) and -toolbox (14), contain m-files which address the
approximation problems discussed above. Two such m-files
are sysbal and hankmr; the first is used for balancing and
balanced truncation, while the second is used for optimal
Hankel-norm approximation (including the computation of
the anti-stable part of the all-pass dilation).
We will conclude with a brief discussion of some articles
which were written since Glovers seminal paper (4), namely,
Refs. 9, 12, 1520.
Hankel-norm approximation or balanced truncation can be
applied to stable systems. The approximants are guaranteed
to be close to the original system, within the bounds given in
the section entitled Error Bounds of Optimal and Suboptimal
Approximants; the important aspect of this theory is that
these bounds can be computed a priorithat is, before computing the approximant. The bounds are in tems of the H
norm which is a well-defined measure of distance between
420
10
10
10
10
20
20
30
30
40
40
50
101
100
101
50
101
Chebyshev 2: FOMROM
10
10
10
20
20
30
30
40
40
101
100
101
Elliptic: FOMROM
10
50
100
101
50
101
100
hank
bal
2( 9 + . . . + 20)
Butterworth
Chebyshev 2
0.0383
0.0506
0.0388
0.1008
0.0779
0.0999
0.1035
1.2015
11
hank
Chebyshev 1
Elliptic
0.3374
0.2457
0.4113
0.2700
bal 2( 11 + . . . + 20)
0.6508
0.3595
1.9768
1.5818
101
The second paper (9) authored by Fuhrmann uses the authors polynomial models for a detailed analysis of the Hankel
operator with given (scalar) rational symbol (i.e., single-input
single-output systems). Furhmanns computations follow a
different approach than the one used in Ref. 4. The Schmidt
pairs (singular vectors) are explicitly computed and hence the
optimal Hankel approximants are obtained. This analysis
yields new insights into the problem; we refer to Theorem 8.1
on page 184 of Ref. 9, which shows that the decomposition of
the transfer function in the basis provided by the Schmidt
pairs of the Hankel operator provides the balanced realization
of the associated state space representation. Another advantage is that it suggests the correct treatment of the polynomial equation [Eq. (38)], which yields a set of n linear equa-
Optimal D
9
hank
Bound 11
Bound 12
Butterworth
Chebyshev 2
0.0384
0.1010
0.0384
0.0506
0.0389
0.1008
0.0390
0.5987
0.0517
0.6008
Optimal D
11
hank
Bound 21
Bound 22
0.2988
0.2441
0.3374
0.2458
0.4113
0.2700
0.5030
0.3492
0.9839
0.7909
Chebyshev 1
Elliptic
10
10
20
20
30
30
40
40
102
101
100
Chebyshev 1: ROM
(Hankel and Balanced)
101
BIBLIOGRAPHY
1. G. W. Stewart and J. Sun, Matrix Perturbation Theory, New York:
Academic Press, 1990.
102
Chebyshev 2: ROM
(Hankel and Balanced)
101
100
101
Elliptic: ROM
(Hankel and Balanced)
10
10
20
20
30
30
40
421
40
101
Figure 4. Analog filter approximation: Bode plots of the reducedorder models obtained by balanced truncation (dashdot curves) and
Hankel-norm approximation (continuous curves).
7. A. C. Antoulas, Lectures on optimal approximation of linear systems, Draft, Dept. Elec. Comp. Eng., Rice Univ., Houston, TX,
1998.
102
101
100
101
102
101
100
422
23. R. Ober, On Stieltjes functions and Hankel operators, Syst. Control Lett., 27: 275278, 1996.
24. A. C. Antoulas, Approximation of linear operators in the 2-norm,
Linear Algebra Appl., special issue on Challenges in Matrix Theory, 1998.
A. C. ANTOULAS
Rice University
VOLTAGE-TO-FREQUENCY CONVERTERS
INTRODUCTION
The Voltage-to-Frequency Converters (VFCs) belong to the
category of the oscillator circuits. The function of an oscillator is to produce periodic time-varying signals with
only dc excitations 1 . These signals may have quite diverse waveform shapes such as triangular, square, impulsive, exponential, sinusoidal, etc. For some oscillator applications, the shape of the generated signal is an important
design criteria. For instance, in the production of periodic
sweep voltages for display systems; or in the generation of
quasi-sinusoidal signals for testing and instrumentation
purposes. In other applications, the important design criteria is the signal frequency, while the spectral purity of
the signal remains secondary. This is for instance the case
of the oscillators used to generate clock signals for timing
purposes. And this is also the case of VFCs.
The function of VFCs is to convert the value of an analog input voltage or current into an output frequency; i.e.
they must generate signals of any shape whose frequency
is a faithful representation of the analog input data. Thus,
they also belong to the category of the data converter circuits, whose function is to convert the information among
different domains with minimum degradation of the information itself. From this point of view, the VFCs have some
similarities with the Analog-to-Digital Converters (ADCs),
whose function is to convert an analog data into a digital word with the minimum error possible. In the case of
VFCs the domain of the transformed information is the frequency of a signal, instead of a digital word, and the aim
is to obtain the smallest possible error in the voltage-tofrequency mapping. The design criteria for VFCs include
the following (14):
(1)
(2)
J. Webster (ed.), Wiley Encyclopedia of Electrical and Electronics Engineering. Copyright 2007 John Wiley & Sons, Inc.
Voltage-to-Frequency Converters
Figure 1. N-like shaped characteristics necessary to implement oscillator circuits and associated piecewise-linear approximation.
Figure 2. Conceptual realization of the basic oscillator model in Eq. (3) (a) and LC realizations for x = v, y = RiL , y = L/R (b) and for x =
RiL , y = vC , x = L/R, y = CR (c).
dx
= [y (x)]
d
dy
1
= x
d
(4)
where = y /x .The two particular cases of interest occur for extreme values of the parameter ; they are respectively caalled the harmonic oscillator casewhich cor-
(6)
Voltage-to-Frequency Converters
Figure 3. Trajectories, signals and spectrum for the oscillator model in the harmonic oscillation
case for x = 10, y = 0.1 and = 1. The three drawings at the top are for NE = 1 and NI = 1, those
in the center row are for NE = 1 and NI = 0.2, the drawings at the bottom were obtained for NE
= 1 and NI = 3. We see that the oscillation frequency is practically the same in the three cases
(basically set by 0 = x y ). We also observe that the larger NI , the larger the signal amplitude
and distortion, while the dynamic evolution toward the limit cycle is faster.
achieved through electronic tuning of the time constants x and y For the LC circuits of Figure 2(b) and
(c), the frequency can be made voltage-controlled by
resorting to the use of tunable capacitors or inductors
(1524).
On the one hand, for given values of x , y , NE and
, the larger NI , the larger the signal amplitude and
the signal distortion. Also, the faster the dynamic evolution towards the limit cycle from the inside. On
the other hand, for given values of x , y , NI and ,
the smaller NE , the smaller the signal distortion, although the speed of the dynamic evolution towards
the limit cycle from the inside does not change with
NE . Finally, the value of inuence either the signal
amplitude, but has no inuence on either the signal
distortion or the speed of the evolution towards the
limit cycle.
Voltage-to-Frequency Converters
Figure 4. (a) Representation of the basic oscillator model in Figure 2(a) as the interconnection of
a linear dynamic block and a nonlinear block and (b) general harmonic oscillator model with these
blocks. The phase shift of the linear block is null at the oscillation frequency, and the oscillation
occurs at the amplitude for which the loop gain is just unity.
Figure 2(a) can be redrawn as depicted in Figure 4(a), consisting of the interconnection of a linear dynamic block
with transfer function in the s-domain given by
1
H(s) =
s x
X(s)
= 2
Z(s)
s + s NE
+
x
1
x y
(7)
The phase shift of the linear block must be frequencydependent and null at a single frequency = 0 ,
which is the osocillation frequency. The sharper the
phase-vs-frequency characteristics of the linear block
around this zero-phase frequency, the larger the accu-
Gm R
1 + sRC
(8)
1
tan
{3 + (1)N }
2RC
2N
(9)
(10)
Voltage-to-Frequency Converters
Figure 5. (a) Ring oscillator structure and its linearized model as a specic example of Figure 4.
(b) Simple typical circuit realizations for the tunable transconductors.
Figure 6. The limit cycle in the relaxation case ts with the nonlinear function (x) in (a) as
depicted in (b), where we assume that the horizontal transitions occur instantaneously. The bottom
drawing (c) shows the limit cycle, the x(t) waveform and its spectrum for x = 0.1, y = 10, = 1
and NE = NI = 1.
meaning that the shape of the limit cycle ts with the nonlinear function (x), drawn with a thick solid trace in Figure 6(a).
This piece of information has to be complemented with
the analysis of the solution trajectories of Eq. (4) the different linear pieces of (x) : Let us consider rst the inner
y() [y(0)]e NI
x()
(11)
Voltage-to-Frequency Converters
for y(0) < NI x(0). The arrows drawn with thin solid lines
in Figure 6(a) indicate the direction and sense of the trajectories in this inner region. From this picture, it is readily
inferred that the nonlinearity inner piece is not part of the
limit cycle.
Let us now consider the outer pieces, where y =
NE x , where the minus is for the left-hand piece, the
plus for the right-hand one, and is the absolute value
of the ordinate at the origin. Analysis obtains a fast transient across a quasi-horizontal trajectory followed by a slow
asymptotic behavior whose trajectory is described by
1
(
)
1
NE
x()
[y(0) ( )]e
NE
(12)
1
(
)
x() [y(0) ( )]e NE
Note that this slow asymptotic behavior results in trajectories which t exactly with the corresponding outer piece
of the nonlinear function.
As a summary of these considerations, the limit cycle
can be built as shown in Figure 6(b), where we assume that
the horizontal transitions occur instantaneously. Thus, the
oscillation frequency is basically determined by the time
invested in making the asymptotic evolutions on the outer
pieces of the nonlinearity, which can be calculated from Eq.
(12) as
Tc = 2y NE ln
+ NI
NI
(13)
than in relaxation oscillators, which make them more appropriate for timing applications (11). Multivibrators have
usually poor timing accuracy for frequencies above a few
hundred kHzs because of the random uctuations of the
threshold levels and charging and discharging currents
and their increasing importance with respect to shorter
clock periods (1), but they provide the best linearity and
temperature and supply stability at medium-low frequencies, as well as wide control and signal ranges (8). Since
the latter are features required for voltage-to-frequency
conversion, most VFCs on the market are based on multivibrators, thus we will focus on this approach in the following.
Implementations of VFCs are usually based on capacitors as energy storage elements, and the nonlinear resistor in Figure 7(b) is commonly implemented as depicted
in Figure 8(a). The current switches in Figure 8(a) are set
to the Ch or Dh position by the reversal block, corresponding to the charging phase and the discharging phase respectively. During the charging phase, the current iC enters the capacitor input node (+) and the voltage vC grows,
while it decreases during the discharging phase, wherein
the current iD leaves the capacitor input node. The circuit
operation consists in the succession of these two phases
in a cyclic behavior. The period of the cycle is determined
on the one hand by the value of iC and iD and the capacitance C associated to the timing capacitor, which set the
rate for the voltage vC to change inside each phase, and on
the other hand by the reversal block, which senses vC and
turns the switches to the Ch or Dh positions. Assuming
constant values for the currents iC and iD , the time intervals corresponding to the charging and discharging phases
are (see Figure 8(c)) ,
tC =
C(vCH vCL )
iC
tD =
C(vCH vCL )
iD
(14)
where vCH and vCL are the top and bottom limits respectively of the voltage swing in the capacitor terminals. In
the specic case of having i = ic = iD , the time invested in a
cycle is
Tc = tC + tD =
2C(vCH vCL )
i
(15)
(16)
vCH vCL
2iR
(17)
Voltage-to-Frequency Converters
Figure 7. Simplied rst-order abstractions for a relaxation oscillator. The conditions leading to the relaxation oscillator case on the LC
circuits of Figure 2(b) and Figure 2(c) are met by making C -> 0 in the former case (a) and L 0 in the latter (b).
Figure 8. Realization of relaxation oscillators based on Figure 7(b). The reversal block controls the switches to alternatively charge and
discharge the capacitor (a). The associated cyclic diagram corresponding to the general case in Figure 6 is shown in (b), while (c) depicts
the resultant waveform.
lies in the role that the reversal block plays. In the former, the reversal block works like a Schmitt trigger, which
senses vC (t) and turns the current switch to the position Ch
or Dh any time it reaches the low vCL and high vCH threshold values respectively. Thus, the reversal block imposes
the thresholds vCH and vCL , and sets directly the voltage
swing (vCH vCL ) in the CLacitor during the circuit operation. Eq. (14) applies here and the frequency of the waveform associated to vC is controlled by means of voltage controlled current sources, thus making iC = KC vin and iD =
KD vin obtain
f =
1
KC KD vin
=
t C + tD
(KC + KD)C(vCH vCL )
(18)
(i0 iD )T0
iC T0
=
C
C
(19)
CvC
i0
= ( 1)T0
iD
iD
(20)
iD
KD vin
1
=
=
tC + tD
i0 T0
i0 T0
(21)
Voltage-to-Frequency Converters
Figure 9. (a) Direct implementation of a Schmitt trigger type VFC converter, where the reversal
block in Figure 8 is implemented by a Schmitt trigger circuit.(b) Two realizations of the Schmitt
trigger block.
Direct Implementation
Emitter-coupled multivibrator
viH = VCC
R2 + R3
R1 + R2 + R3
(22)
R2
R1 + R2
(23)
Note that Eq. (22) and Eq. (23) constitute the high and low
threshold voltages, respectively, of the single-comparator
Schmitt trigger. The main drawback of this circuit is that
the parasitic capacitors at the nodes between the resistors have to be charged and discharged every time the
switch is closed or opened, which is a time-consuming task
and causes errors that limit the high frequency capability of the circuit. A better behavior is obtained by using
two comparators and xed voltage references to dene the
threshold values. The right part of Figure 9(b) shows the
dual-comparator Schmitt trigger which is based on this approach and whose threshold voltages are determined by the
resistor ladder as
R3
R2 + R3
viL = VCC
viH = VCC
(24)
R1 + R2 + R3
R1 + R2 + R3
Besides other general limitations that will be seen later,
the basic disadvantage of this strategy is the poor supply voltage stability, due to the dependence of the threshold values on VCC . Table 1 shows the typical performance
parameters associated to the NE5664 , which follows the
direct approach in Figure 9(a) and whose maximum drift
Voltage-to-Frequency Converters
1MHz
10:1
0.2 %/V
600 ppm/ C
0.4%
0kHz to 150kHz
0.15%
50 ppm/ C
0.01%/V
0kHz to 500kHz
0.05%
75 ppm/ C
0.015% of FSR%
25ppm/ C
25ppm/ C
0.002%
0.02%
0.001 %/V
Figure 10. Emitter-coupled multivibrator. The clamping block in (a) causes a voltage drop of value VBEon for a current ID VBEon /R, that
makes the basis voltage of Q2 fall, thus it is cut-off and the capacitor is charged through . The situation reiCmains until the voltage drop
at the emitter of Q2 is low enough to turn on Q2 , thus the clamping block at the top-left corner cuts-off Q3 and the situation is reversed.
The process is repeated cyclically.
VH = VCC VBEon
(26)
(27)
Since the problem is equivalent to that treated in the previous section, equation Eq. (18) is also valid here and the
10
Voltage-to-Frequency Converters
Kvin
4CVBEon
(28)
(29)
(30)
(31)
Now from Eqs. (29), (30) and (31) vC = vCH vCL = (VH
VM ) (VL VM ) = 2VR and Eq. (18) reduces to
f =
K Vin
4CVR
(32)
|vC |
| dvdtC |
= tos (
i0 R
1)
vin
(34)
1
vin
=
tos + tD
tos i0 R
(35)
Note that this coincides with Eq. (21) for KD = 1/R and T0 =
tos . The circuit in Figure 12 has two main drawbacks. First,
the change of vC causes an error in the charging current
due to the nite output impedance of the current source
that generates it, as well as in the discharging current iD ,
which is not constant as supposed above. Second, the passive RC integrator is very slow, especially for large input
voltage changes, which require a large amount of charge
to be injected in or extracted from the capacitor. Both aspects can be improved by using an active integrator, as
Figure 13 depicts (327). The negative feedback provides
low impedance at the inverting input terminal of the integrator, whose voltage remains nearly constant, thus the
error in the charging and discharging currents is reduced.
Equations Eq. (33)-Eq. (35) are valid in Figure 13(b), but
now iD = vin /R exactly. With respect to Figure 13(a), vC
Voltage-to-Frequency Converters
11
Figure 11. Improved emitter-coupled multivibrator. The clamping circuitry is redesigned to make the voltage variations at the capacitor
depend on a voltage (VR ) very stable against temperature changes, thus the frequency drift with temperature is improved.
Figure 12. Basic diagram of a charge-balancing converter based on a one-shot timer. The charging phase has xed length in time, which
is the width of the pulse generated by the one-shot timer (tos ) when it is triggered by the comparator. For vC vC , the average voltage at
the comparator inverting input is vin , thus iD vin /R. The average current injected in the bottom integrator must equal iD and depends
on the rate the one-shot is triggered with, hence this rate depends on vin .
Figure 13. Improved converters based on one-shot timers. The use of active integrators instead
of the passive one in Figure 12 achieves smaller errors and improves the dynamic response.
response because this minimizes the transients. In addition, Figure 13(a) provides high impedance for the input
terminal, which can be useful for applications with low
load requirements. Finally, note that the capacitance of the
integrating capacitor does not appear in Eq. (35), thus it
12
Voltage-to-Frequency Converters
Rb
)
Ra
(36)
The voltage divider formed by Ra and Rb is usually implemented on-chip and the accuracy obtained for the ratio
Rb /Ra is within 1%, and does not depe1nd on the temperature. Thus, the precision, as well as the temperature
stability in the expected tos are mainly determined by the
resistor R and the capacitor. However, for small values of
tos , the time delays in the input comparator, the ip-op and
the switch limit the accuracy. In addition, for low frequencies the leakage currents in the capacitor and the switch
introduce an error in the timing prediction.
Table 3 shows some typical performance parameters of
the monolithic converter ADVFC32KN (4).
Synchronized charge balancing type converters
Further improvements can be achieved by replacing the
one-shot timer by circuitry clocked by a precise oscillator,
usually a crystal one, which determines the charging phase
duration (5). Figure 15 shows the circuit which results from
substituting the one-shot timer with a block composed of
one bistable and one AND logic gate, which works as follows. If the integrator output crosses the threshold vref , the
comparator places a high value at the bistable input (suppose that Q is high). However, the output does not follow
it immediately as in Figure 13(b), but it is synchronized
with the clock by the ip-op and changes at the next clock
rising edge. At the same time, the output complement (low
value) is put at the ip-op input by the feed-back loop,
thus the output will go back down at the next rising edge
of the clock and the duration of the output pulse is just one
clock cycle. Note that the circuit operates in the charging
phase as long as the output is high, thus TCLK replaces tos in
the equations Eq. (33), Eq. (34) and Eq. (35), which remain
valid. However, since the output is now synchronized, its
frequency must be a multiple of the clock frequency ,
f =
1
1
=
tD
TCLK + TCLK iS ( TCLK
)
TCLK [1 + iS ( iv0inR 1)]
(37)
Voltage-to-Frequency Converters
13
Figure 15. Synchronized charge balancing converter. The one-shot timer is replaced by externally
clocked circuitry. The charging phase length is now xed by the clock period, which is supposed very
accurate.
output to track the input voltage). In fact, the transfer function of the AD7740 (6) is f = 0.1 f CLK + 0.8(vin /vre f ) f CLK .
Since sigma-delta modulators can be built currently
with cheap CMOS technology, the main advantage of this
approach is the cost. The AD7741 (7) with 0.012% Nonlinearity at 2.75MHz of FoutMax costs ve times less than
the synchronous AD652 with 0.02% Nonlinearity at 2MHz
of FoutMax. In addition, they need fewer external components.
vin vBEon
(38)
iB
R
which acts as charging or discharging current. However,
note that this circuit provides a current that is not proportional to vin , but contains an offset that equals (vBEon /R +
ire f =
14
Voltage-to-Frequency Converters
iB ). Besides modifying the intended proportional relationship, the added term has a strong temperature coefcient,
since the base to emitter voltage varies with the temperature as 2mV/ C.
A common alternative for eliminating the dependence
of iref on VBEon places the transistor in a negative feedback loop as Figure 17(b) illustrates. If we neglect the base
current, the output reference current for this circuit is
ire f =
A
vBEon
vin
(A + 1)R
(A + 1)R
(39)
vin
R
(40)
(VBEQ1 VBEQ2 )
VT
nI1
=
ln(
)
R2
R2
I2
(41)
R3 VT
nR3
ln(
)
R2
R1
(42)
R3 VT
ln(n)
R2
(43)
Voltage-to-Frequency Converters
15
Figure 18. Current source nite output resistance: timing error (a) and common cascode strategies to increase the output resistance (b).
Figure 19. Band-gap reference voltage. (a) The voltages VBE and VT have temperature VTcoefcients of opposite sign, thus they compensate the global dependence of VR .(b) Two realizations of this approachVR.
input stage, bias currents have to be provided from the input to the sensing nodes. Such currents are usually in the
range of a few hundred nano-amperes and add to or subtract from charging and discharging currents. MOS transistors can be used instead, to implement almost innite
input impedance thanks to the isolation provided by the
gate oxide. In addition, junction leakage or dielectric absorption currents can also be added to or subtracted from
charging currents, thus also affecting the output frequency.
However, these currents are commonly at a low nanoampere range. Hence, errors due to these base and leakage
currents can be usually tolerated as long as the charging
and discharging currents are in the range of 10A (14).
Finite Switching Times
The switching times in Figure 8 have been neglected in the
computation of the output frequency in Eq. (18). However,
they are not zero in real implementations and limit the
frequency capability of the converter. The elements that
contribute to the delay in switching are mainly the length
of the feedback path and the delays of the blocks along
it, while the gain of the active devices, the parasitic capacitances and the internal current levels determine such
delays.
Output Interfacing
The output stage of the commercial VFCs is designed to
allow easy interfacing to all digital logic families. Hence,
BJT open collector or both uncommitted collector and emit-
16
Voltage-to-Frequency Converters
(44)
io tos
C
(45)
It is possible to reduce the ripple by increasing the capacitance C in Eq. (45), but it also will enlarge the time constant
Voltage-to-Frequency Converters
17
Figure 20. Frequency-to-Voltage converters. (a) Pulse-integrating converters, where the charge packets are injected in the integrator
at a rate determined by the input frequency. (b) Phase-locked-loop type converters, where the low-pass lter output voltage in a PLL is
proportional to the deviations of the input signal frequency.
Figure 21. Telemetry with VFC. The voltage signal from a sensor is translated into a frequency signal to be transmitted through twisted
pair, ber optic or radio links. On the other side, the signal can be converted back into a voltage to drive an analog display or an actuator,
or can be processed digitally.
Figure 22. Isolation with VFCs. The frequency signal from a VFC can be transmitted through magnetic (a) or optic (b) couplers to
implement galvanic isolation, thus preventing damage due to high bias voltages, or unreasonable current leakage.
Digital Processing
The output frequency signal from a VFC can be easily
translated into a digital code in order to be processed,
stored or displayed by conventional digital circuits and systems (928).
Actually, this procedure implements an A/D converter
while transforming the dc input of the VFC into a digital word. Figure 23(a) shows the simplest way to carry out
this A/D conversion, where a counter, which is only enabled
for a given interval TG named gate time, counts the transitions of the input signal. Thus, after this known period,
the counter stores the decimal value N = f TG , which
encodes the input voltage in a binary number. Another ap-
(46)
18
Voltage-to-Frequency Converters
Figure 23. Digital processing of the output frequency signal of a VFC. The transitions in a xed time interval TG (TG ) are counted and
the average frequency is computed.
ACKNOWLEDGMENT
This work was supported in part by the spanish government under project TEC2006-12376-C02-01/MIC.
BIBLIOGRAPHY
1. A. A. Abidiand R. G. Meyer, Noise in relaxation oscillators,
IEEE Journal of Solid-State Circuits, Vol. CS-18, pp. 794802,
Dec. 1983.
2. AD537 Integrated circuits voltage-to-frequency converter,
Voltage-to-frequency converter in Analog to Digital Converters
in Product Catalogue, www.analog.com/en/
3. AD650 Voltage-to-frequency and frequency-to-voltage converter, Voltage-to-frequency converter in Analog to Digital Converters in Product Catalogue, www. analog. com/en/
4. ADVF32 Monolithic non-synchronous Voltage-to-frequency
and frequency-to-voltage converter, Voltage-to-frequency converter in Analog to Digital Converters in Product Catalogue,
http:www. analog. com/en/
5. AD652 Monolithic synchronous voltage-to-frequency converter, Voltage-to-frequency converter in Analog to Digital Converters in Product Catalogue, www. analog. com/en/
6. AD7740 3V/5V Low Power, Synchronous Voltage-to-Frequency
Converter, Voltage-to-frequency converter in Analog to Digital
Converters in Product Catalogue, www. analog. com/en/
7. AD7741/AD7742 Single and Multichannel, Synchronuos
Voltage-to Frequency Converters. Voltage-to-frequency converter in Analog to Digital Converters in Product Cata- logue,
www.analog.com/en/
8. M. Banu, MOS oscillators with multi-decade tuning range
and gigahertz maximum speed, IEEE Journal of Solid-State
Circuits, Vol. CS-23, pp. 474479, April 1988.
9. L. N. Beard, Use of a voltage-to-frequency converter to measure the integral JEdt, American Journal of Physics, Vol. 57,
Iss. 5, pp. 457476, 1989.
10. R. Carrand J. Yang, Current-to-frequency converter for electron multipliers, Review of Scientic Instruments, Vol. 60, Iss.
7, pp. 23212323, 1989.
11. L. M. DeVito, A versatile clock recovery architecture and
monolithic implementation, Monolithic phase-locked loops
and clock recovery circuits, Editor B. Razavi, IEEE Press, New
York 1996.
iv This device is obsolete, but its data are still a good example of
the performance obtained with this implementation.
Voltage-to-Frequency Converters
12. S. W. Glowacki, A current-to-frequency converter, Nuclear
Instruments and Methods, Iss. 186, pp. 549551, NorthHolland Publishing Company 1981.
13. P. R. Grayand R. G. Meyer, Analysis and Design of Analog
Integrated Circuits, 2nd Edition. Ed. John Wiley & Sons. Singapore 1984.
14. A. B. Grebene, Bipolar and MOS analog integrated circuit design, John Wiley & Sons, New York 1984.
15. K. W. Kobayashi et al., A novel self-oscillating HEMT-HBT
cascode VCO-Mixer using an active tunable inductor, IEEE
Journal of Solid-State Circuits, Vol.33, No. 6, June 1998.
16. LM231A/LM231, LM331A/LM331 Precision voltage-tofrequency converters, National Semiconductor Other Analog
Product Catalogue www.national.com/catalog/
17. J-F Loude, Current-to-frequency isolation converter, Journal of Physics E-Scientic Instruments, Vol.18,Iss. 10, pp.
824826, 1985.
18. T. D. Moore, S. Turner, C.C. Speake, A capacitance displacement transducer with large dynamic-range, good linearity and
frequency read-out, Measurement Science & Technology, Vol.
4,Iss. 12,pp. 15191521, 1993.
19. R. Pallas-Arenyand
`
J. G. Webster, Sensors and Signal Conditioning, John Wiley and Sons, New York 1991.
20. R. A. Pease, Application Note C: V/F converter ICs handle frequency-to-voltage needs,National Semiconductor Application Notes, http://www.national.com/apnotes/Voltage-toFrequencyConverters v1. html
21. R. A. Pease, Application Note 210: New Phase-LockedLoops Have Advantages as Frequency to Voltage Converters (and more),National Semiconductor Application Notes, http://www. national. com/apnotes/Voltage-toFrequencyConverters v1. html
22. B. Razavi, Analysis, Modeling and Simulation of Phase Noise
in Monolithic Voltage-Controlled Oscillators, Proc. CICC, pp.
323326, May 1995.
23. B. Razavi, Design of Monolithic phase-locked loops and clock
recovery circuits- a tutorial, Monolithic phase-locked loops
and clock recovery circuits, Editor B. Razavi, IEEE Press, New
York 1996.
19
FERNANDO VIDAL-VERDU 1
1
RAFAEL NAVAS-GONZALEZ
MANUEL DELGADO-RESTITUTO2
2,3
ANGEL RODRIGUEZVAZQUEZ
1 Dpto.
de Electrnica,
Universidad de Mlaga,
Campus de Teatinos,
Complejo Tecnologico,
Malaga,
Spain
2 Institute of Microelectronics of
Seville, CNM-CSIC, Edicio
CICA-CNM, C/ Tara sn,
Seville, Spain
3 School of Engineering,
University of Seville, Seville,
Spain