You are on page 1of 688

Linear System Theory

and Design
-~-'
Li near ~
and De~
HRW
Series in
¡
:.

Electrical and
Chi-Tsong Chen
Computer Engineering
Professor, Department
State University of Nev
M. E. Van Valkenburg, Series Editor

L. S. Bobrow ELEMENTARY LINEAR CIRCUIT ANALYSIS


C. T. Chen LINEAR SYSTEM THEORY AND DESIGN
D. J. Comer. DIGITAL LOGIC AND STATE MACHINE DESIGN
C. H. Durney, L. D. Harris, C. L. Alley ELECTRIC CIRCUITS: THEORY AND
ENGINEERING APPLlCATIONS
G. H. Hostetter, C. J. Savant, Jr., R. T. Stefani DESIGN OF FEEDBACK
CONTROL SYSTEMS
S. Karni and W. J. Byatt MATHEMATICAL METHODS IN CONTINUOUS
AND DISCRETE SYSTEMS
B. C. Kuo DIGITAL CONTROL SYSTEMS
B. P. Lathi MODERN DIGITAL AND ANALOG COMMUNICATION
SYSTEMS
C. D. McGillem and G. R. Cooper CONTINUOUS AND DISCRETE
SIGNAL AND SYSTEM ANALYSIS Second Edition
A. Papoulis CIRCUITS AND SYSTEMS: A MODERN APPROACH
S. E. Schwarz and W. G. Oldham ELECTRICAL ENGINEERING:
AN INTRODUCTION _ __ ,.__ _ . _ .~_; __
A. S. Sedra and K. C. Smith MICROELECTRONIC CIRCUITS
M. E. Van Valkenburg ANALOG FILTER DESIGN

HOLT, RINEHART /J

New York Chicagc


Montreal Toronto
Linear System Theory
and Design

Chi-Tsong Chen
g
Professor, Department of Electrical Engineering
State University ofNew York at Stony Brook

ANALYSIS
=:SIGN
;HINE OESIGN
~IRCUITS: THEORY AND

DESIGN OF FEEDBACK

HODS IN CONTINUOUS

~OMMUNICATION

US ANO DISCRETE
Id Edition
)ERN APPROACH
L ENGINEERING:

NIC CIRCUITS
;N

HOLT, RINEHART AND WINSTON, INe.

New York Chicago San Francisco Philadelphia


Montreal Toronto London Sydney Tokyo
This text is a major revision of IntroduClion lO Linear
System Theory by Chi-Tsong Chen, originally published
in 1970. © 1970 by Holt, Rinehart and Winston
Copyright © 1984 Holt, Rinehart and Winston, Inc.
AII rights reserved. No part of lhis publication may be reproduced or
transmiued in any form or by any means, electronic or mechanical, ineluding
photocopy, recording or any information storage and retrieval system, without
permission in writíng from lhe publisher.

Requests for permission to make copies of any part of the work should be
mailed to: Permissions, Holl, Rinehart and Winston, Inc., Orlando,
Florida 32887

Library of Congress Cataloging in Publication Data

Chen, Chj-Tsong.
Linear system theory and designo

Bibliography: p.

Ineludes indexo

1. System analysis. 2. System designo 1. Title.


QA402.C442 1984 003 83-12891
ISBN 0-03-060289-0

Printed in lhe United States of America


9 038 987
Holt, Rinehart and Winston
The Dryden Press
Saunders College PUblishing
To
Beatrice
and
Janet, Pauline, Stanley

cluding
without

d be
Preface XIII

Glossary of Symbol:

Chapter 1 Introe
1-1 ".
j

1-2 l

Chapter 2 linea
2-1 1

2-2 1

2-3 !

2-4 1

2-5 S

2-6 I

2-7 F

*May be omitted without 10ss (


Contents

Preface Xlll

Glossary of Symbols xvii

Chapter 1 Introduction 1

1-1 The Study of Systems 1

1-2 The Scope of the Book 2

Chapter 2 Linear Spaces and Linear Operators 6

2-1 Introduction 6

2-2 Linear Spaces over a Field 7

2-3 Linear Independence, Bases, and Representations 12

Change of Basis 17

2-4 Linear Operators and Their Representations 19

Matrix Representations of a Linear Operator 21

2-5 Systems ofLinear Algebraic Equations 26

2-6 Eigenvectors, Generalized Eigenvectors, and Jordan-Form

Representations of a Linear Operator 33

*Derivation of a Jordan-Form Representation 38

2-7 Functions of a Square Matrix 45

Polynomials of a Square Matrix 45

Functions of a Square Matrix 51

Functions of a Matrix Defined by Means of Power

Series 54

*May be omitted without loss of continuity.

vii

viii CONTENTS

*2-8 Nonns and Inner Product 57


4-4

2-9 Concluding Remarks 60

Problems 62

4-5

Chapter 3 Mathematical Descriptions of Systems 70

3-1 Introduction 70

3-2 The Input-Output Description 72 Chapter 5 Conti

Linearity 73 Equa

Causality 76

Relaxedness 77
5-1 ]
Time Invariance 80
5-2 ]
Transfer-Function Matrix 81
5-3 e
3-3 The State-Variable Description 83

The Concept of State 83

Dynamical Equations 86

Linearity 87

Time Invariance 89

Transfer-Function Matrix 90
5-4 (
Analog and Digital Computer Simulations ofLinear Dyna­
mical Equations 91

3-4 Examples 94

*Dynamical Equations for RLC Networks 101

3-5 Comparisons of the Input-Output Description and the State­


Variable Description 106
5-5 (

3-6 Mathematical Descriptions of Composite Systems 108


r
Time-Varying Case 108

Time-Invariant Case 111


*5-6 (

Well-Posedness Problem 114


I

*3-7 Discrete-Time Systems 121


*5-7 (
(
3-8 Concluding Remarks 124

Problems 125
*5-8 (
*5-9 (
F
Chapter 4 linear Dynamical Equations and Impulse-Response

Matrices 133

Chapter 6 Irredu
4-1 Introduction 133
Identi1
4-2 Solutions of a Dynamical Equation 134

Time-Varying Case 134


6-1 It
Solutions of x = A(t)x 134
6-2 T
Solutions of the Dynamical Equation E 139
R
Time-Invariant Case· 141
6-3 Ir
4-3 Equivalent Dynamical Eqúations 146
F
Time-Invariant Case 146

*Time-Varying Case 151

Linear Time-Varying Dynamical Equation with

Periodic A(o) 153

CONTENTS ix

4-4 Impulse-Response Matrices and Dynamical


Equations 154
*Time-Varying Case 154
Time-Invariant Case 157
4-5 Concluding Remarks 161
ms 70 Problems 162

2 Chapter 5 Controllability and Observability 01 linear Dynamical


Equations 168
5-1 Introduction 168
5-2 Linear Independence of Time Functions 170
5~3 Controllability of Linear Dynamical Equations 175
Time-Varying Case 175
83 *Differential Controllability, Instantaneous Controllabil­
ity, and Uniform Controllability 180
Time-Invariant Case 183
*Controllability Indices 187
5-4 Observability of Linear Dynamical Equations 192
90 Time-Varying Case 192
IUlations of Linear Dyna­
*Differential Observability, Instantaneous Observabil­
ity, and Uniform Observability 196
Linear Time-Invariant Dynamical Equations 197
letworks 101 *Observability Indices 198
lescription and the State­
5-5 Canonical Decomposition of a Linear Time-Invariant Dyna­
mical Equation 199
lposite Systems 108
Irreducible Dynamical Equations 206
*5-6 Controllability and Observability of lordan-Form Dynamical
Equations 209
114
*5-7 Output Controllability and Output Function
Controllability 214
*5-8 Computational Problems 217
*5-9 Concluding Remarks 226
Problems 227
¡I Impulse-Response

Chapter 6 Irreducible Realizations, Strict System Equivalence, and


Identification 232
n 134
6-1 Introduction 232
6-2 The Characteristic Pplynomial and the Degree of a Proper
'34
Rational Matrix 234
~quation E 139
6-3 Irreducible Realizations of Proper Rational
Functions 237
146
Irreducible Realization of ~1D(s) 237 .
Irreducible Realizations of g(s) = N(s)lD(s) 240
Observable Canonical-Fonn Realization 240
lical Equation with
Controllable Canonical-Fonn Realization 243
X CONTENTS

Realization from the Hankel Matrix 245 7-4


*Jordan-Canonical-Form Realization 249

*Realization of Linear Time-Varying Differential

Equations 252

6-4 Realizations of Vector Proper Rational Transfer

Functions 253

Realization from the Hankel Matrix 257

*6-5 Irreducible Realizations of Proper Rational Matrices: Hankel


7-5
Methods 265

Method 1. Singular Value Decomposition 268


*7-6
Method II. Row Searching Method 272 7-7
*6-6 Irreducible Realizations of G(s): Coprime Fraction

Method 276

Controllable-Form Realization 276

Realization ofN(s)D-1(s), Where D(s) and N(s) Are Not Chapter 8 Stab
Right Coprime 282 8-1
Column Degrees and Controllability Indices 284 8-2
Observable-Form Realization 285

*6-7 Polynomial Matrix Description 287

*6-8 Strict System Equivalence 292

*6-9 Identification of Discrete-Time Systems from Noise-Free


8-3
Data 300 8-4
Persistently Exciting Input Sequences 307

Nonzero Initial Conditions 309

6-10 Concluding Remarks 313


*8-5
Problems 317
*8-6
Chapter 7 State Feedback and State Estimators 324 8-7

7 -1 Introduction 324

7-2 Canonical-Form Dynamical Equations 325

Single-Variable Case 325 Chapter 9 Linea


*Multivariable Case 325 tion, ~
7-3 State Feedback 334 9-1 I
Single-Variable Case 334 9-2 (
Stabilization 339

Effect on the Numerator of g(s) 339


9-3 (
Asymptotic Tracking Problem-Nonzero Set
S
Point 340

*Multivariable Case 341

Method I 341

Method II 345
9-4 S
Method III 347
Nonuniqueness of Feedback Gain Matrix 348
Assignment of Eigenvalues and Eigenvectors 351 9-5 D
Effect on theNumerator Matrix of G(s) 352
Computational Problems 353
CONTENTS xi

atrix 245
7-4 State Estimators 354

tion 249
Full-Dímensional State Estimator 355

19 Differential
Method I 357

Method II 358

lal Transfer Reduced-Dimensíonal State Estimator 361

Method 1 361

ix 257
Method II 363

tional Matrices: Hankel


7-5 Connection of State Feedback and State Estimator 365

Functional Estimators 369

IPosition 268
*7-6 Decoupling by State Feedback 371

lId 272
7-7 Concluding Remarks 377

¡prime Fraction
Problems 378

\ 276
Chapter 8
Stability of Linear Systems 384

I~e D(s) and N(s) Are Not


8-1 lntroduction 384

\lbility Indices 284


8-2 Stability Criteria in Terms of the Input-Output

\285 Description 385

1287
Time-Varying Case 385

Time-Invariant Case 388

I,tems from Noise-Free 8-3 Routh-Hurwitz Criterion 395

8-4 Stability of Linear Dynamical Equations 400

linces 307
Time-Varying Case 400

IJ9 Time-Invariant Case 407

*8-5 Lyapunov Theorem 412

A Proof of the Routh-Hurwitz Criterion 417

*8-6 Discrete-Tíme Systems 419

8-7 Concluding Remarks 425

324
Problems 425

325
Chapter 9
Linear Time-Invariant Composite Systems: Characteriza­
tion, Stability, and Designs 432

9-1 lntroductíon 432

9-2 Complete Characterizatíon of Single-Variable Composite

Systems 434

s) 339
9-3 Controllability and Observability of Composíte

I-Nonzero Set Systems 439

Parallel Connection 440

Tandem Connection 441

Feedback Connection 444

9-4 Stability of.Feedback Systems 448

Single-Variable Feedback System 449

rain Matrix 348


Multivariable FeedIJack System 451

~d Eig~nvectors 351
9-5 Design of Compensators: Unity Feedback Systems 458

ix of G(s) 352
Single-Variable Case 458

53
Single-Input or Single-Output Case 464

xii CONTENTS

Multivariable Case-Arbitrary Pole Assignment 468


Appendix G Poly
Multivariable Case-Arbitrary Denominator­ 0-1

Matrix Assignment 478 0-2

Decoupling 486 0-3

9-6 Asymptotic Tracking and Disturbance Rejection 488


0-4

Single-Variable Case 488


0-5

Multivariable Case 495


0-6

Static Decoupling-Robust and Nonrobust

Designs 501

State-Variable Approach 504


Appendix H Pole
9-7 Design of Compensators: Input-Output Feedback
ProbJ

Sytems 506

Single-Variable Case 506


References 636

Multivariable Case 511


Index 657

Implementations of Open-Loop Compensators 517

Implementation 1 517

Implementation 11 519

Applications 523

Decoupling 523

Asymptotic Tracking, Disturbance Rejection, and

Decoupling 526

Model Matching 528

9-8 Concluding Remarks 534

Problems 536

Appendix A Elementary Transformations 542

A-l Oaussian Elimination 543

*A-2 Householder Transformation 544

A-3 Row Searching Algorithm 546

*A-4 Hessenberg Form 551

Problems 553

Appendix B Analytic Functions of a Real Variable 554

Appendix C Minimum Energy Control 556

Appendix O Controllability alter the Introduction of Sampling 559

Problems 564

Appendix E Hermitian Forms and Singular Value

Decomposition 565

Problems 570

Appendix F On the Matrix Equation AM + MB = N 572.


Problems 576

CONTENTS xiii

,Je Assignment 468 Appendix G Polynomials and Polynomial Ma~rices 577

~nominator- G-l Coprimeness of Polynomials 577

G-2 Reduction of Reducible Rational Functions 584

G-3 Polynomial Matrices 587

ce Rejection 488
G-4 Coprimeness of Polynomial Matrices 592

G-S Column- and Row-Reduced Polynomial Matrices 599

G-6 Coprime Fractions of Proper Rational Matrices 605

ronrobust Problems 618

Appendix H Poles and Zeros 623

Jut Feedback
Problems 635

References 636

Index 657

ompensators 517

¡rbance Rejection, and

e 554

of Sampling 559

N 572

This text is intended fo


linear systems and mu
pendent study and ref
mathematical backgrc
matrix manipulation ,
The unstarred section:
first graduate course (
York at Stony Brook.
during the last three ~
multivariable systems,
number of universities
Vlith the advancen
designing systems tha
Consequently, it is im:
one might unknowingl
investigation of aH the
procedures have evo IVI
study and the design ] .
control text per se, b<
optimization, and sens
This text is a revise
Theory which discusse
in 1970. Since then; se
system theory. Aman!
matrices in fractional
Preface

This text is intended for use at the senior-graduate level in university courses on
linear systemS and multivariable system designo It may also be used for inde­
pendent study and reference by engineers and applied mathematicians. The
mathematical background assumed for this book is a working knowledge of
matrix manipulation and an e1ementary knowledge of differential equations.
The unstarred sections of this book have been used, for over a decade, in the
first graduate course on linear system theory at the State University of New
York at Stony Brook. The majority of the starred sections were developed
during the last three years for a second course on linear systems, mainly on
multivariable systems, at Stony Brook and have been c1assroom tested at a
number of universities.
With the advancement of technology, engineers have become interested in
designing systems that are not merely workable but also the best possible.
Consequently, it is important to study the Iimitations of a system; otherwise,
one might unknowingly try to design an impossible system. Thus, a thorough
investigation of all the properties of a system is essential. In fact, many design
procedures have evolved from such investigations. This text is devoted to this
study and the design procedures developed thereof. This is, however, not a
control text per se, because performance criteria, physical constraints, cost,
optimization, and sensitivity problems are notconsidered.
This text is a revised and expanded edition of Introduction to Linear System
Theory which discussed mostly the state variable approach and was published
in 1970. Since then, several important developments have been made in linear
system theory. Among them, the geometric approach and the transfer-function
matrices in fractional forms, called the matrix-fraction description, are most
xv
xvi PREFACE

pertinent to the original text. The geometric approach is well covered in duced in Chapter 4.
W. M. Wonham's Linear Multivariable Control: A Geometric Approach, 2d ed., realizations (Chapter
Springer-Verlag, New York, 1979 and is outside the scope ofthis text. Hence (Appendix G). Mono
the new material of this edition is mainly in the transfer-function matrix in Hence it was decided
fractional formo Because of this addition, we are able to redevelop, probably polynomials and poi
more simply in concepts and computations, the results of the state variable This, however, will d
approach and establish a fairly complete link between the state-variable the topic was grouped
approach and the transfer-function approach. The logical sequen
We aim to achieve two objectives in the presentation. The first one is to
develop major results and design procedures using simple and efficient methods.
Thus the presentation is not exhaustive; only those concepts which are essential
in the development are introduced. For example, the Smith-McMillan form is Chapter 2
not used in the text and is not discussed. The second objective is to enable the Secs. 2-]
to 2-5
reader to employ the results developed in the text. Consequently, most results
are developed in a manner suitable for numerical computation and for digital
computer programming. We believe that solving one or two problems of each
topic by hand will enhance the understanding ofthe topic and give confidence
in the use of digital computers. With the introduction of the row searching
algorithm (Appendix A), which has been classroom tested, this is possible even Appendix A

for multivariable systems, as long as their degrees are sufficiently smal!.


The level of mathematics used in this edition is about the same as that of the
original edition. If concepts and results in modern algebra more extensive than
those in Chapter 2 are introduced, sorne results in the text can be developed
more elegantly and extended to more general settings. For example, the
Jordan form can be established concisely but abstractly by using the concepts Secs. G-I Appendix G
of invariance subspaces and direct sumo Its discussion can be found in a large
number of mathematical texts and will not be repeated here. In view of om
objectives, we discuss the computation of the required basis and then develop
the Jordan formo By using sorne concepts in abstract algebra, such as ring,
("'
principal ideal domain, and module, the realization problem (Chapter 6) can be Single-variable Secs. 9-5
cases of 9-6 ~
developed more naturally and sorne results in this text can be extended to delay Secs. 9-5 to 9- 7 9-7
differential equations, linear distributed systems, and multidimensional systems.
These are extensively studíed in AIgebraic System Theory, which was initiated The material connecte,
by R. E. Kalman in the late 1960s and has extended in recent years most of the The logical dependenc
results in this text to linear systems over rings. The concepts used in algebraic various combinations (
system theory are less familiar to engineering students and require more 1 teach a one-semestel
mathematical sophistication and will not be discussed. All the results and following chapters are I

design procedures in this text are developed by using only elementary concepts
and results in linear algebra. Chapter 1
The results in this text may eventually be implemented on digital computers. Chapter 2
Because of the finite word lengih, the sensitivity of problems and the stability Chapter 3 (Skip. Th
of algorithms become important on computer computations. These problems Chapter 4 (Ski.p Th
are complex and extensively discussed in texts on numerical analysis. In our Chapter 5 (Empha~
development, we will take note of these problems and remark briefly wherever 5~5, and
appropriate. Chapter6
The arrangementof the topics in this text was not reached without any Chapter 7
difficulty. For example, the concepts of poles and zeros seem to be best intro­ Charter 8
PREFACE xvii
eh is well covered in duced in Chapter 4. However, their complete treatments require irreducible
etric Approach, 2d ed., realizations (Chapter 6) and coprime fractions of transfer-function matrices
pe of this text. Hence (Appendix G). Moreover, the concept of zeros is used only in Section 9-6.
fer-function matrix in Hence it was decided to create an appendix for the topic. The coprimeness of
o redevelop, probably polynomials and polynomial matrices might be inserted in the main text.
, of the state variable This, however, will digress too much from the state-variable approach; thus
~en the state-variable the topic was grouped in an appendix.
The logical sequences of various chapters and appendixes are as follows:
Jll. The first one is to
: and efficient methods. Chapler 1
:pts which are essential
nith-McMillan form is Chapter 2
!
-====== Chapler 3 (These orders can be inlerchanged.)
ljective is to enable the Secs.2-1
lo 2-5
,equently, most results
utation and for digital 1

Chapter 4
: two problems of each
lic and give confidence
.!
Chapler 5 ---- Appendix B
1 of the row searching
:d, this is possible even Appendix A - Seco 5-8 ~ Appendixes e and D
lfficiently small.
the same as that of the
ra more extensive than
1
Chapler 6
text can be developed Seco 6-·2 Chapler 8 Charler 7
~s. For example, the 1 - - - - Hankel melhod
Sec.6-5 -Appendix E ....... Secs. 8-5-Appendix F ----+- Seco 7-4

by using the concepts Secs. G-l Appendix G _ Secs. 6-6 8-6

~an be found in a large G-2


lo 6-9
l here. In view of our
Jasis and then develop
algebra, such as ring,
i
Chapler 9
Secs. 9-1 / Appendix H
tlem (Chapter 6) can be Single-variable Secs. 9-5 9-2
cases of 9-6 --------9-3 --------Sec.9-4
m be extended to delay Secs. 9-5 lo 9-7 9-7
ltidimensional systems.
ry, which was initiated The material connected by a broken line is not essential in the development.
ecent years most of the The logical dependencies among Chapters 6, 7, 8, ana 9 are loose, ana their
cepts used in algebraic various combinations can b~ adopted in one- or two-semester courses. When
nts and require more 1 teach a one-semester course at Stony Brook, the unstarred sections of the
l. AII the results and following chapters are covered:
Iy elementary concepts
Chapter 1
don digital computers. Chapter 2
blems and the stability Chapter 3 (Skip Theorem 3-1 and its corollary.)
tions. These problems Chapter 4 (Skip Theorem 4-11.)
;:rical analysis. In our Chapter 5 (Emphasize the time-invariant part by skipping Theorems 5-2,
;:mark briefl.y wherever 5-5, and 5-11.)
Chapter 6
t reached without any Chapter 7
, seem to be best intro- Charter 8
xviii PREFACE

We emphasize the exact meanings of theorems and their implications; hence the and discussions has
proofs of a number of theorems are skipped. For example, we prove only matter. 1 am gratefl
Theorems 2-1 and 2-2 in Chapter 2. We skip the proofs ofTheorems 4-1, 4-2, F. Trace for typing va
and others. In the second course, we cover the following: and Winston and the ~
and to Professors S. ]
Appendix A
from the People's R
Section 5-8, controllability and observability indices
thanks go to m y wife
Hankel method (Section 6-4 and method II of Section 6-5)
their support during t
Appendix E
Singular value decomposition method (Method l of Section 6-5)
Appendix G
Sections 6-6 to 6-9
Starred sections of Chapter 7
Appendix H
Chapter 9
Those who are interested in quick access to the design methods using the
transfer-function matrix in fractional form may proceed from Sections 2-1 to
2-5, Appendixes A and G, and then to Sections 9-5 to 9-7, or only their single­
variable cases.
The problem sets form an integral part of the book. They are designed to
help the reader understand and utilize the concepts and results covered. In
arder to retain the continuity of the main text, sorne important results are
stated in the problem sets. A solutions manual is available from the pub­
Iisher.
The Iiterature on linear system theory is very extensive. The length of this
text, however, is limited. Hence the omission of sorne significant results is
inevitable and l would Iike to apologize for il. l am indebted to many people in
writing this book. Kalman's work and Zadeh and Desoer's book Linear
System Theory form the foundation ofthe original edition ofthis book. Rosen­
brock's and Wolovich's works are essential in developing the present edition.
l have benefited immensely in my learning from Professor C. A. Desoer. Even
to this date, 1 can always go to him whenever l have questions. For this, I can
never express enough of my gratitude. To Professors B. J. Leon, E. J. Craig,
I. B. Rhodes, P. E. Barry (first edition) and to Professors M. E. Van Valkenburg,
W. R. Perkins, D. Z. Zheng (present edition), l wish to express my appreciation
for their reviews and valuable suggestions. I would like to thank President
F. Zhang and Professor K. W. You of Chengdu University of Science and
Technology, Professor S. B. Park of Korea Advanced lnstitute of Science and
Technology, Professor T. S. Kuo of National Taiwan University, and Professor
S. K. Chow of National Sun Yat-Sen University, Taiwan, for providing oppor­
tunities for me to lecture on an earlier draft of Chapter 9 and Appendix G.
I especially appreciate the opportunity at Chengdu University to interact witb
several faculty members, especially Professor L. S. Zhang, from various uni­
versities in China; their suggestions have improved considerably the presenta­
tion of the text. I am grateful to many of "my graduate" students, speciaUy
C. Waters, C. H. Hsu(the first edition), I. S. Krishnarao, Y. S. Lai, C. C. Tsui and
S. Y. Zhang (the present edition) whose assistance in the form of dissertations
--~ ----~~-----~_.-.-

PREFACE xix

and discussions has clarified considerably my understanding of the subject


mplications; hence the
matter. 1 am grateful to Mrs. V. Donahue, C. LaGuardia, T. Marasco, and
ample, we prove only
F. Trace for typing variolls drafts of this text, to Mr. P. Becker of Holt, Rinehart
; of Theorems 4-1, 4-2,
and Winston and the staffofCobbjDunlop for their assistance in the production
g: and to Professors S. H. Wang, K. W. You, and D. Z. Zheng, visiting scholars
from the People's Republic of China, for their proofreading. My special
thanks go to my wife Beatrice and my children Janet, Pauline, and Stanley for
m 6-5) their support during the writing of this text.

Section 6-5) Chi-Tsong Chen

gn methods using the

d from Sections 2-1 to

·7, or only their single-

They are designed to


ld results covered. In
important results are
'ailable from the pub­

ive. The length of this


le significant results is
bted to many people in
Desoer's book Linear
n of this book.Rosen­
ng the present edition.
or C. A. Desoer. Even
~stions, For this, 1 can
B. J. Leon, E. J. Craig,
M. E. Van Valkenburg,
x:press my appreciation
ike to thank President
versity of Science and
nstitute of Science and
liversity, and Professor
n, for providing oppor­
er 9 and Appendix G.
versity to interact with
ang, from various uni­
.siderably the presenta­
late students, specially
'i. S. Lai, C. C. Tsuiand
le form of dissertations
Q.E.D.
I

A,B,P, ...
u, y, ex, •..
u, y, ex, ...

!l'
u(s), y(s), G(s), C(s)

v(A), ...

b(G(S)), deg G(s)

A',x', .

A*, x*, .

det A, ...

iC

~
~(s)
~[s]
p(A), rank A
Glossary of Symbols

Q.E.D. End of the proof of a theorem.


I This symbol denotes the end of a statement or an
example.
A,B,P, . Capital boldface letters denote matrices.
u, y, 0:, . Lowercase boldface letters denote vectors.
u, y, (l., ••• Lowercase italic and Greek type denote scalar-valued
functions or scalars. Capital italic letters are also used
in Chapter 9 and Appendix G to denote scalars.
fe
Laplace transformo
u(s), y(s), G(s), C(s)
1f a letter is used in both time and frequency domains,
circumf1ex wilI be used to denote the Laplace transform
such as u(s) = fe[lJ{t)] and (;(.1') = ft[G(r)J 1r él letter is
used in only one domain, no circumf1ex wi\l be used,
for example, C(s).
v(A), ... The nuIlity of the constant matrix A.
Ó(G(s)), deg G(s)
The degree of the rational matrix G(s).
A',x', .
The transpose of the matrix A and the vector x.
A*, x*, .
The complex-conjugate transpose of the matrix A and
the vector x.
det A, ...
The determinant of A.
iC
The field of complex numbers.
~
The field of real numbers.
~(s)
The field of ratianal functions of s withcoefficie'nts in IR.
~[sJ
The set of polynomials of s with coefficients in ~.
p(A), rank A The rank of A. lf A is a constant matrix, the rank is

xxi
xxii GLOSSAR y OF SYMBOLS

defined over iC or IR. If A is a rational or polynomial


matrix, the rank is defined over lR(s).
deg det A(s) The degree of the determinant of A(s).
diag {A, B, C} A diagonal matrix with A, B, and e as block diagonal
elements as

Linear Sys
and Desig
where A, B, and e are matrices, not necessarily square
and of the same order.
~ Equals by definition.

~A = (~a.)
dt te, dt 1) ,

ff! [A] ~ (ff! [aij]), . " When an operator is applied to a matrix or a vector,
it means that the operator is applied to every entry of
the matrix or the vector.
ational or polynomial
~(s).
. A(s).
d e as block diagonal

Linear System Theory


and Design
not necessarily square

a matrix or a vector,
Jlied to every entry of
1-1 The Study 01

The study and design


methods. We apply v.
responses. If the perfo!
eters or connect to it :
design is guided by pas
approach has undoubte
The empirical methe
performance become ve
if physical systems beco
to be experimented. Ir:
Analytical study of phy
development of matherr
distinction between ph)
fact, circuits or control:
systems. A resistor witl
of the resistor does not
inductance is again a 1
amount of current fiowi
for the success of the I

properlymodeled.
A physical system rr
asked and the different
amplifier may be modell
1
Introduction

1 -1 The Study of Systems

The study and design of a physical system can be carried oui by "empirical
methods. We apply various signals to the physical system and measure its
responses. If the performance is not satisfactory, we adjust sorne of its param­
eters or connect to it sorne compensator to improve its performance. This
design is guided by past experience, if any, and proceeds by cut and try. This
approach has undoubtedly succeeded in designing many physical systems.
The empirical method may become unsatisfactory if the specifications on the
performance become very precise and stringent. lt may also become inadequate
if physical systems become very complicated or too expensive or too dangerous
to be experimented. In these cases, analytica! methods become indispensable.
Analytical study of physical systems roughly consists of four parts: modeling,
development of mathematical-equation description, analysis, and designo The
distinction between physical systems and models are basic in engineering. In
fact, circuits or control systems studied in any textbook are models of physical
systems. A resistor with a constant resistance is a model; the power limitation
of the resistor does not appear in the resistance. An inductor with a constant
inductance is again a model; in reality, the inductance may varv with the
amount of current ftowing through it. Modeling is a very important problem,
for the success of the design dependsupon whether the physical system is
properly modeled.
A physicalsystem may have different models depending on the questions
asked and the different operational ranges used. For example, an electronic
amplifier may be modeled differently at high and low frequencies. A spaceship
2 INTRODUCTION

may be modeled as a particle in the study of its trajectory; however, it must be systems are mainly lil
modeled as a rigid body in the maneuvering. In order to develop a suitable
model of a physical system, a thorough understanding of the physical system Y(s)=
and its operatiúnal range is essential. In this book, we shall refer to models of and
physical systems as systems. Hence a physical system is a device or a collection
of devices existing in the real world; a system is a model of a physical system.
Once a system (a model) is found for a physical system, the next step in the
studyis to develop, by applying various physicallaws, mathematical equations
to describe the system. For example, we apply Kirchhoffs voltage and current and their extensions
laws to electrical.systems and Newton's laws to mechanical systems. The relationship between
equations that describe systems may assume many forms; they may be linear and is called the inp
equations, nonlinear equations, integral equations, difference equations, differ­ frequency domain. T
ential equations, or others. Depending on the question asked, one form of its elements are limitec
equation may be preferable to another in describing the same system. In con­ as ratios of two polyn
clusion, a system may have many different mathematical-equation descriptions, said to be in the poly
just as a physical system may have many different models. is called a dynamical (
Once a mathematical description of a system is obtained, the next step in set of first-order diffe:
the study involves analyses-quantitative and/or qualitative. In the quanti­ it is called the dynan
tative analysis, we are interested in exact responses of systems due to the applica­ or the internal descri¡
tion of certain input signals. This part of the analysis can be easily carried out from the eoncepts of
by using a digital or an analog computer. In the qualitative analysis, we are examples, how they ca
interested in the general properties of the system, such as stability, control­ The major portion
lability, and observability. This part of analysis is very important, because around Equations (1­
desig~ techniques may often evolve from this study. and qualitative. The
lf the response of a system is found to be unsatisfactory, the system has to puters; hence we empl
be improved or optimized. In some cases, the responses of systems can be will be thoroughly in­
improved by adjusting certain parameters of the systems; in other cases, com­ number of design proc
pensators have to be introduced. Note that the design is earried out on the The design problerr
model of a physical system. However, if the model is properly chosen, the control systems. In (
performance of the physical system should be correspondingly improved by without any constraint
introducing the required adjustments or compensators. design objectives, such
mum degrees which ce
are also interested in (
1-2 Ttae Scope of the Bool~ túe designo In the desi;
system to meet some
The study of systems may be divided into four parts: modeling, setting up steady-state error, and
mathematical equations, analysis, and designo The development of models for compensators. A des
physieal systems req uires knowledge ofeach particular field and sorne measuring from the one without
devices. For example, to develop models for transistors requires the knowledge systems is not consideff
of quantum physics and sorne laboratory setup. Developing models for auto­ of control systems, alth
mobile suspension systeins iequires actual testing and measurements; it cannot in its designo For a dis(
be achieved by the use of pencil and paper alone. Thus, the modeling problem 546, and 596. In this se
should be studied in connection with each specific field and cannot be properly chapter.
covered in this text. Hence, we shall assume that models of physical systems, We review in Chapt
or systems, are available to us in this text. The objective of thisc
The mathematical equations which will be used in this text to describe transformations, to sob
THE SCOPE OF THE BOOK 3

ry; however, it must be systems are mainly limited to


r to develop a suitable
y(s) = G(s)u(s) = N(s)D- 1(s)u(s) = F- 1 (s)H(s)u(s) (1-1 )
of the physical system
shall refer to models of and
a device or a collection x(t) = Ax(t) + Bu(t) (1-2a)
el of a physical system.
em, the next step in the y(t) = Cx(t) + Eu(t) (1-2b)
nathematical equations
,ffs voltage and current and their extensions to the time-varying case. Equation (1-1) describes the
:hanical systems. The relationship between input u and output y in the Laplace-transform domain
ms; they may be linear and is called the input-output dfscription or the external description in the
:rence equations, diITer­ frequency domain. The matrix G(s) is called the trans!erl!1nction matrix, and
on asked, one form of its elements are limited to rational functions. In the design, G(s) will be factored
: same system. In con­ as ratios of two polynomial matrices, such as N(s)D- 1 (s) or F- 1 (s)H(s), and is
[-equation descriptions, said to be in the polynomial fraction formo The set of two equations in (1-2)
~ls.
is called a dynamical equation or state-variable equation. Equation (1-2a) is a
tained, the next step in set of first-order differential equations. Ir (1-2) is used to describe a systern,
itative. In the quanti­ it is called the dynamical-equation description, the state-variable description,
tems dueto theapplica­ or the internal description. These two types of equations will be developed
an be easily carried out from the concepts of linearity and time invariance. We shall then show, by
itative analysis, we are examples, how they can be used to describe systems.
:h as stability, control­ The major portion of this text is devoted to the analysis and design centered
~ry important, because
around Equations (1-1) and (1-2). Analysis can be divided into quantitative
and qualitative. The former .can now be delegated to analog or digital com­
:tory, the system has to puters; hence we emphasize the latter. Various properties of these equations
nses of systems can be will be thoroughly investigated. Their relationships will be established. A
1S; in other cases, com­ number of design procedures will then be developed from the study.
n is carried out on the The design problem studied in this book is not exactly the design of feedback
IS properly chosen, the
control systems. In our design, we are interested in the exact conditions,
londingly improved by without any constraints on the complexity of compensators, to achieve certain
design objectives, such as stabilization or poIe placement: What are the mini­
mum degrees which compensators must possess to achieve such design? We
are also interested in developing simple and efficient procedures to carry out
the designo In the design of control systems, the ultimate purpose is to design a
system to meet sorne performance criteria, such as the rise time, overshoot,
;: modeling, setting up steady-state error, and others, with or without constraints on the degrees of
~lopment of models for compensators. A design problem with constraints is significantly different
~ld and sorne measuring from the one without constraints. In this text, the performance of control
requires the knowledge systems is not considered. Hence our study is not a complete study ofthe design
oping models for auto­ of control systems, although most of the results in this text are basic and useful
leasurements; it cannot In its designo For a discussion ofthe design of c~ntr~l systems, see References 3,
, the modeling problem S46, and S96. In this section we give a brief descripfion of the contents of each
and cannot be properly chapter.
ds of physical systems, We review in Chapter 2 a number ofconcepts and results in linear algebra.
The objective of this chapter is to enable the reader to carry out similarity
n this text to describe transformations, to solve linear algebraic equations, and to compute functions
4 INTRODUCTION

of a matrix. These techniques are very important, if not indispensable, in are studied in Chaptel
analysis and design of linear systems. the eigenvalues of the
In Chapter 3 we develop systematically the input-output description and feedback with a cons·
the state-variable description of linear systems. These descriptions are devel­ its state can be gener
oped from the concepts of linearity, relaxedness, causality, and time invariance. val ues. Various desi.
We also show, by examples, how these descriptions can be set up for systems. is also established.
Mathematical descriptions of composite systems and discrete-time equations We study in Chapl
are also introduced. We also discuss the well-posedness problem in the feedback under the heading of ~
systems. the design of a system
In Chapter 4 we study the solutions of linear dynamical equations. We output stability, stabi
also show that different analysis often leads to different dynamical-equation total stability. Their
descriptions of the same system. The relation between the input-output discuss the Lyapunov
description and the state-variable description is also established. criterion. Their cour
We introduce in Chapter 5 the concepts of controllability and observability. In the last chapter
The importance of introducing these concepts can be seen from the networks invariant composite s~
shown in Figure 1-1. Their transfer functions are both equal to 1. There is no zero cancellation of t
doubt about the transfer function of the network shown in Figure 1-I(b); with the transfer functi
however, we may ask why the capacitor in Figure 1-I(a) does not play any role ways, as shown in Fig
in the transfer function. In order to answer this question, the concepts of be studied from its con
controllability and observability are needed. These two concepts are also and (c) cannot. We a
essential in optimal control theory, stability studies, and the prediction or feedback systems. We
filtering of signals. Various necessary and sufficient conditions for a dynamical function matrices in t
equation to be controllable and observable are derived. We also discuss the tions: unity feedback
canonical decomposition of a dynamical equation and introduce an efficient design compensators '
and numerically stable method of reducing a dynamical equation to an irreduci­ ment and to achieve
ble one. reestablish in the tran
In Chapter 6 we study irreducible realizations of rational transfer-function in the state-variable
matrices. The problem is to find a controllable and observable linear time­ approaches.
invariant dynamical equation that has a prescribed rational matrix. lts solution A total of eight ;
is indispensable in analog and digital computer simulations. lt also offers a various chapters are el
method of synthesizing a rational matrix by using operational-amplifier circuits.
This result is also needed in establishing the link between the state-variable
approach and the transfer-function approach in the design of linear time­
invariant systems.
The practical implications ofthe concepts of controllability and observability (a)

11
t 1: In
,.
+

11
t In
-r
,.
+
------o:<

J
(urrent
I
SOUfCC 151 In
___ L­
lo) lb)
Figure 1-1 Two different networks withthe same transfer function, 1. Figure 1-2 Thrée differ
THE SCOPE Of THE BOOK 5
;

not indispensable, in are studied in Chapter 7. We show that if a dynamical equation is controllable,
the eigenvalues of the equation can be arbitrarily assigned by introducing state
Itput description and feedback with a constant gain matrix. lf a dynamical equation is observable,
escriptions are devel­ its state can be generated by designing a state estimator with arbitrary eigen­
, and time invariance. values. Various design procedures are introduced. The separation property
be set up for systems. is also established.
screte-time equations We study in Chapter 8 a qualitative property oflinear systems. This comes
'oblem in the feedback under the heading of stability, which is always the first requirement to be met in
the design of a system. We introduce the concepts of bounded-input bounded­
nical equations. We output stability, stability in the sense of Lyapunov, asymptotic stability, and
t dynamical-equation total stability. Their characterizations and relationships are studied. We also
~en the input-output discuss the Lyapunov theorem and then use it to establish the Routh-Hurwitz
blished. criterion. Their counterparts in the discrete-time case are also studied.
lity and observability. In the last chapter we study various problems associated with linear, time­
m from the networks invariant composite systems. One of them is to study the implication of pole­
lual to 1. There is no zero cancellation of transfer functions. For example, consider two systems
wn in Figure l-l(b); with the transfer functions l/(s -1) and (s - 1)/(s + 1) connected in three different
loes not play any role ways, as shown in Figure 1-2. We show why the system in Figure 1-2(b) can
tion, the concepts of be studied from its composite transfer function, but the systems in Figure 1-2(a)
vo concepts are also and (c) cannot. We also study stabilities of single-variable and multivariable
nd the prediction or feedback systems. We then study the design of compensators by using transfer­
itions for a dynamical function matrices in the fractional formo We study two feedback configura­
We also discuss the tions: unity feedback and plant input-output feedback connections. We
mtroduce an efficient design compensators with minimum degrees to achieve arbitrary pole place­
luation to an irreduci­ ment and to achieve asymptotic tracking and disturbance rejection. We
reestablish in the transfer-function approach essentially the results developed
)nal transfer-function in the state-variable approach and complete the link between the two
lservable linear time­ approaches.
Ll matrix. lts solution A total of eight appendixes are introduced. Their relationships with
ions. lt also offers a various chapters are covered in the Preface.
Illal-amplifier circuits.
~en the state-variable
esign of linear time­
(a)
·ility and observability

(b)

GJ-r
~J s - \

(c)

;tion, 1. Figure 1-2 Three dilTerent connections 01" 1/(s -1) and (s - 1)/(s + 1).
Section 2-7, we study
and the Cayley-Ham
concepts of inner pro<
This chapter is in
have some basic knm
multiplication, and im

2
be used. Let A, B, (
respectively. Let 3i be
we have

Linear Spaces and


Linear Operators
CA=C[a
b1

rt
BD= b

bm
These identities can b
2-1 Introduction the product of an n x
The material presel
In this chapter we shall review a number of concepts and results in linear 38, 39,43 to 45, 77, 86
algebra that are essential in the study of this text. The topics are carefully emphasize the differen
selected, and only those which will be subsequently used are introduced. The (2-12) and Definition
purpose of this chapter is to enable the reader to understand the mechanism of matrix representation
similarity transformation, to solve linear algebraic equations, to find Jordan­ follow naturally.
form representations of square matrices, and to compute functions of a matrix,
in particular, exponential functions of a matrix (see Section 2-9, Concluding
Remarks 1 ). 2-2 linear Spaces
In Section 2-2 we introduce the concepts of field and linear space over a
field. The fields we shall encounter in this book are the field of real numbers, In the study oí mathe
the field of complex numbers, and the field of rational functions. In order to forms the center of stu
have a representation of a vector in a linear space, we introduce, in Section 2-3, For example, in arith
the concept of basis. The relationship between different representations of the algebra, we study the
same vector is established. In Section 2-4, we study linear operators and their examples of sets includ
representations. The concept of similarity transformation is embedded here. the set of aH polynomi
In Section 2-5, the solutions of a set of linear algebraic equations are studied. matrices. In this sech
The concepts of rank and nullity are essential here. ' In Section 2-6, we show that one of those just menti
every square matrix has a Jordan-form representation; this is achieved by· Consider the set 01
introducing eigenvectors and generalized eigenvectors as basis vectors: . in plication with the corr
set. The sum and pr<
set has elements O and
1 It is recommended that the reader keeps the concludin~ remarks in mind, for they provide the
reader with' motivations for studying the mathematical theorems introduced in this chapter.
2 Numbers correspond to th
6
LINEAR SPACES OVER A FIELD 7

Section 2-7, we study functions of a square matrix. The minimal polynomial


and the Cayley-Hamilton theorem are introduced. In the last section, the
concepts of inner product and norm are introduced.
This chapter is intended to be self-contained. The reader is assumed to
have some basic knowledge of matrix theory (determinants, matrix addition,
multiplication, and inversion). The matrix identities introduced below will also
be used. Let A, B, C, D be n x m, m x r, 1 x n, and r x p constant matrices,
respectively. Let a¡ be the ¡th column of A, and let bj be the jth row of B. Then
we have

AB~ [a, a,··· a.{~J ~ a,b, +a,b, +... +a.b. (2-1 )

CA = C[al a z ' - . a m ] = [Cal Caz'" Cam ] (2-2)

BD= (blJ
bt D= (blDJ
b~D (2-3)

bm bmD
These identities can be easily checked. Note that a¡b i is an n x r matrix, it is
the product of an n x 1 matrix a¡ and a 1 x r matrix b¡_
The material presented here is well known and can be found in References 5,
; and results in linear 38,39,43 t045, 77, 86, and 116. z However our presentation is different. We
he topics are carefully emphasize the difference between a vector and its representations [see Equation
i are introduced. The (2-12) and Definition 2-7]. After stressing this distinction, the concepts of
land the mechanism of matrix representation of an operator and of the similarity transformation
ltions, to find Jordan­ follow naturally.
: functions of a matrix,
:ction 2-9, Coneluding
2-2 Linear Spaces over a Field
Id linear space over a
: field of real numbers, In the study of mathematics we must first specify a collection of objects that
"unctions. In order to forms the center of study. This coHection of objects or eíements is called a seto
:roduce, in Section 2-3, For example, in arithmetic, we study the set of real numbers. In boolean
representations of the algebra, we study the set {O, 1}, which consists of only two elements. Other
:ar operators and their examples of sets inelude the set of complex numbers, the set of positive integers,
ion is embedded here. the set of aH polynomials of degree less than 5, the set of aH 2 x 2 real constant
equations are studied. matrices. In this section when we discuss a set of objects, the set could be any
ction 2-6, we show that one of thosejust mentioned or any other the reader wishes to specify~
1; this is achieved by Consider the set of real numbers. The operations of addition and multi­
as basis vectors. In plication with the commutative and associative properties are defined Tor the
set. The sum and product of any two real numbers are real numbers.. The
set has elements O and 1. Any real number ex has an additive inverse ( - ex) and
l mind, for they Pfovide the

:oduced in this chapter.


2 Numbers correspond to the References at the end of the book.
8 LINEAR SPACES AND LINEAR OPERATORS

a multiplicative inverse (l/a, except a = O) in the set. Any set with these proper­ field are, respectively,
ties is caBed afield. We give a formal definition of a field in the foBowing.

Definition 2-1
A field consists uf a set, denoted by :F, of elements caBed sealars and two Note that the set of all
operations caBed addition "+" and multiplication "'''; the two operations From the foregoin:
are defined over :F such that they satisfy the following conditions: field could be anythin~
objects. The fields W{
1. To every pair of elements a and 13 in :F, there correspond an element a + [3 familiar ones: the field
in :F called the sum of a and [3, and an element a . [3 or af3 in :F, called the field of rational functic
produet of a and [3. tions of these fields ar
2. Addition and multiplication are respective1y commutative: F or any a, [3 in :F, show that they satisfy ;
a+f3=[3+a to denote the field of rea
and use !R(s) to denote
3. Addition and multiplication are respectively associative: For any a, 13, y in :F, with indeterminate s. 1
(a + fJ) + y = a + (fJ + y) (a' f3)' y = a' (/3' y) field because it has no :
nomials do not forro a
4. Multiplication is distributive with respect to addition: For any a, 13, y in :F, Before introducing
a . (13 + y) = (a . 13) + (a . y) two-dimensional geom
the plane can be consi<
5. :F contains an element, denoted by O, and an element, denoted by 1, such A vector can be shrunl
that a + O= a, 1 . a = a for every a in :F. product of two points
6. To every a in :F, there Is an element 13 in :F s\.1ch that a + 13 = 0. The element matical terminology, is
13 is caBed the additive inverse. spaee.
7. To every a in :F which is not the element O, there is an element y in :F such
that a· y = 1. The element y is caBed the multiplieative inverse. I Definition 2-2
A linear space over a fie
We give some examples to illustrate this concept.
of elements caBed vecte
and sealar multiplieatioj
Example 1 that they satisfy all the
Consider the set of numbers that consists of O and 1. The set {O, 1} does not 1. To every pair of ve<
form a fie1d ifwe use the usual definition of addition and multiplication, because in :!f, called the SUD:
the e1ement 1 + 1 = 2 is not in the set {0,1}. However, if we define 0+ O = 2. Addition is commu
1 + 1 =0, 1 + O = 1 and 0·1 =0,0 =0,1'1 = 1, then it can be verified that {O, 1} 3. Addition is associa
with the defined addition and multiplication satisfies aB the conditions listed (x 2 +x 3 )·
for a field. Hence the set {O, 1} with the defined operations forms a field. It is 4. fj( contains a vecto
called the field of binary numbers. I The vector Ois caB(
5. To every x in fI, th<
Example 2 6. To every a in :F, an(
the sealar product o
Consider th~ set of all 2 x 2 matrices of the form
7. Scalar multiplicatic
X -YJ
.[ · y · x .
a(f3x) = (af3)x .

3 A sel wilh all properlies of


where x and y are arbitrary real numbers. The set with the usual definitions of precisely, a cornrnulalive rir
matrix addition and multiplication forms a fie1d. The elements O and 1 of the does lhe sel of polynornials .
LINEAR SPACES OVER AFIELO 9

set with these proper­ field are, respectively,


l in the fo11owing.
and

Note that the set of a11 2 x 2 matrices does not form a field. I
alled scalars and two
'; the two operations From the foregoing examples, we see that the set of objects that forms a
:onditions: field could be anything so long as the two operations can be defined for these
objects. The fields we sha11 encounter in this book are fortunately the most
lond an element ex + 13 familiar ones: the field of real numbers, the field of complex numbers, and the
:>r rxf3 in ?F, called the field of rational functions with real coefficients. The additions and multiplica­
tions of these fields are defined in the usual ways. The reader is advised to
:ive: For any ex, 13 in ?F, show that they satisfy aH the conditions required for a field. We use IR and iC
to denote the field of real numbers and the field of complex numbers, respectively,
and use lR(s) to denote the field of rational functions with real coefficients and
e: For any a, /3, '}' in ?F, with indeterminate s. Note that the set ofpositive real numbers does not form a
field because it has no additive inverse. The set of integers and the set of poly­
(13 . y)
nomials do not form a field because they have no multiplicative inverse. 3
For any ex, 13, y in ?F, Before introducing the concept of vector space, let us consider the ordinary
two-dimensional geometric planeo lf the origin is chosen, then every point in
the plane can be considered as a vector: it has direction as we11 as magnitude.
11, denoted by 1, such A vector can be shrunk or extended. Any two vectors can be added, but the
product of two points or vectors is not defined. Such a plane, in the mathe­
+13=0. Theelement matical terminology, is ca11ed a linear space, or a vector space, or alinear vector
space.
n element y in ?F such
Je ínverse. I Defínition 2-2
A linear space over a field ?F, denoted by (f![, ?F), consists of a set, denoted by f![,
of elements ca11ed vectors, a field fF, and two operations called vector addition
and scalar multiplication. The two operations are defined over f![ and ff such
that they satisfy a11 the fo11owing conditions:
~he set {O, 1} does not f![, there corresponds a vector Xl +X z
1. To every pair of vectors Xl and x 2 in
nultiplication, because in 9(, ca11ed the sum ofx l and X;.
r, if we define 0+0= 2. Addition is commutative: For any Xl' X2 in ~ Xl + X2 =X 2 +x l .
. be verified that {O, 1} 3. Addition is associative: For any x¡, x 2 , and x 3 in f![, (Xl +X 2 )+X 3 =X l +
J the conditions listed (X 2 +X 3 )·
ms forms a fie\d. It is 4. f![ contains a vector, denoted by O, such that O +x = X for every x in f![.

• The vector O is ca11ed the zero vector or the origino


5. To every x in ~ there is a vector x in f![, such that x +x =0.
6. To every rx in ff, and every J!: in f![, there corresponds a vector r:xx in f![ ca11ed
the scalar product of rx and x.
7. Scalar multiplication is associ:Üive: For any rx, f3 in fF ap<:l any x in f![,
a(f3x) = (rxf3)x.

3 A set with aB properties of a field except property 7 in Definilion 2-[ is caBed a ril1g or, more
the usual definitioris of precisely, a commutative ring with (multiplicative) identity. The set of integers forms a ring, as
lements Oand 1 of the does the set of polynomials with real" codficients.
-- - - - --_..
-- - ----- -------._~-----_._-_._--- --~-----
-~- -~-----------
_--~-

10 LINEAR SPACES AND LINEAR OPERATORS

8. Scalar multiplication is distributive with respect to vector addition: For Let the vector additioj
anya in ff and any x¡, x 2 in f!l', a(x l +X 2 ) = cxx l +cxx 2 .
9. Scalar multiplication is distributive with respect to scalar addition: For
any a, {3 in ff and anyx in f!l', (a + {3)x = cxx + {3x.
10. For any x in !!l, Ix =x, where 1 is the element 1 in ff. I

Example 1
A field forms a vector space over itself with the vector addition and scalar lt is easy to verify tha
multiplication defined as the corresponding operations in the field. For not a linear space.
example, (~,~) and (iC, iC) are vector spaces. Note that (iC,~) is a vector
space but not (~, ic). (Why?) We note that (~(s), ~(s)) and (~(s), ~) are also Example 5
vector spaces, but not (~, ~(s)). I
Let !!l denote the set e
Example 2 x+2x + 3x =0. Then
scalar multiplication dI
The set of all real-valued piecewise continuous functions defined over ( - 00, 00) homogeneous, then ( Pi
forms a linear space over the field of real numbers. The addition and scalar
multiplication are defined in the usual way. lt is called a function space. I
We introduce one 1
Example 3
Given a field ff, let ff" be all n-tuples of scalars written as columns Definition 2-3
Let (f!l', ~) be a linear ~

X¡= xt' Xlil (2-4)


be a subspace of (!!l, ff:
space over ff.
[ x.,
where the first subscript denotes various components of X¡ and the second We remark on the
subscript denotes different vectors in ff". lf the vector addition and the scalar the vector addition an
multiplication are defined in the following way: space (!!l, ff), they sati:
2-2. Hence we need t(
aXli

l
whether a set iJ}j is a su
Xli + X1Jl
x¡ +Xj= X2i ~ X2] cxx.= aX2i (2-5) is in iJ}j for any Yl, Y2 in
[ XIIi +Xllj~ !
ra~lli-,
then (ff", ff) is a vector space. lf ff =~, (~", ~) is called the n-dimensional
6 are satisfied. Henc~

real vector space; if ff = ic, (e, ic) is caHed the n-dimensional complex vector Example 6
space; if ff = ~(s), (W(s), ~(s)) is called the n-dimensional rational vector
In the two-dimensiona
~~ I
through the origin is a
Example 4
Consider the set !R"[s] of aH polynomials of degree less than n with real co­
effiCients 4
n- 1
ror any fixed real a is a
L a¡si
¡~O

Example 7
4 Note that IR (s), with parenlheses, denotes the field of rational functions with real coefficients;
whereas IR [s J, with brackets, denotes the set of polyti.o~ials with real coefficients. The real vector space 1
LINEAR SPACES OVER A FIELD 11

, vector addition: For Let the vector addition and the scalar multiplication be defined as
~2' n-l n-l "-1
) scalar addition: For ¿ lX¡si + ¿ f3iSi = ¿ (IX¡ + f3¡)S¡
i=O i=O ;=0
¡;
• IX Ct~ lXiS¡) = ~t~ (IXIX¡)S¡

Ir addition and scalar 1t is easy to verify that (rrln[S], rrl) is a linear space. Note that (rrln[s], rrl(s)) is
ns in the fie1d. For not a linear space. I
hat (iC, IR) is a vector
) and (rrl (s), rrl) are also Example 5
• Let fL denote the set of aH solutions of the homogeneous differential equation
x + 2x + 3x = O. Then (fL, rrl) is a linear space with the vector addition and the
scalar multiplication defined in the usual way. lfthe differential equation is not
defined over (- 00, 00) homogeneous, then (gr, rrl) is not a linear space. (Why?) 11
le addition and scalar
a function space. I
We introduce one more concept to conclude this section.

as columns Definition 2-3


Let (fL, .?") be a linear space and let óJI be a subset of fL. Then (óJI, .?") is said to
be a subspace of (gr, .?") if under the operations of (fL, .?), óJI itself forms a vector
(2-4)
space over !/F •

We remark on the conditions for a subset of fL to form a subspace. Since


of Xi and the second
the vector addition and scalar multiplication have been defined for the linear
Lddition and the scalar
space (fL, .?), they satisfy conditions 2, 3, and 7 through 10 listed in Definition
2-2. Hence we need to check only conditions 1 and 4 through 6 to determine
whether a set óJI is a subspace of (fL,.?). 1t is easy to verify that if IX lY 1 + 1X 2 Y2
(2-5) is in óJI for any y i> Y2 in óJI and any 1Xi> 1X 2 in .?", then conditions 1 and 4 through
6 are satisfied. Hence we conclude that a set óJI is a subspace of (gr, .?") if
Cf. l Y1 + Cf. 2 Y2 is in óJI, IfJl" an y y j , y 2 in óJI and any o: 1, 0: 2 in .'!fr.

lled the n-dimensional


lsional complex vector Example 6
sional rat ional vector
In the two-dimensional real vector space (1R 2 , IR), every straight line passing
• through the origin is a subspace of(rrl 2 , rrl). That is, the set

s than n with real co­

for any fixed real IX is a 's~bspace of (rrl 2 , rrl). •


Example 7
tions wíth real coefficients;
.1 coefficients. The real vector space (rrl n, ~) is a subspace of the vector space (iC", rrl).

12 LINEAR SPACES AND LINEAROPERATORS

then the linear indep


2-3 Linear Independence, Bases, and Representations
lowing definition.
Every geometric plane has two coordinate axes, which are mutually perpendic­
ular and of the same scale. The reason for having a coordinate system is to Definition 2-4'
have sorne reference or standard to specify a point or vector in the planeo In
A set of vectors Xl, Xz
this section, we will extend this concept of coordinate to general linear spaces.
only if the equation
In linear spaces a coordinate system is called a basis. The basis vectors are
generally not perpendicular to each other and have different scales. Before
proceeding, we need the concept of linear independence of vectors.
implies or; = O, where (
Definition 2-4
ingly, Q[ can be consid

A set of vectors Xl' Xz, ... ,XII in a linear space over a field ff, (Er, ff), is said to
Observe that line
be linearly dependent if and only if there exist scalars (Xl' (Xz, . .. , (XII in ff, not
but also on the field.
all zero, such that
(2-6)

X
lf the only set of (Xi for which (2-6) holds is (Xl = 0, (Xz = O, ... , (XII = 0, then the
set of vectors xl> Xz, ... ,XII is said to be linearly independent. I

Given any set of vectors, Equation (2-6) always holds for (Xl = 0, (Xz = 0, ... ,
is linearly dependent
(XII = O. Therefore,in order to show the linear independence of the set, we have
°
to show that (Xl = 0, (Xz = 0, ... , (XII = is the only set of (Xi for which (2-6) holds;
that is, if any one of the (X;'s is different fmm zero, then the right-hand side of
Indeed, if we choose

(2-6) cannot be a zero vector. lf a set of vectors is linearly dependent, there are
generaHy infinitely many sets of (x¡, not aH zero, that satisfy Equation (2-6).
However, it is sufficient to find one set of (x¡, not aH zero, to conclude the linear then Cl.1X l +(Xzx z =0.
dependence of the set of vectors. the field of real numbe
zero, such that (X 1X1 +
Example 1 pendent in (IRZ(s), IR), 1

ConsÍder the set of vectors Xl' Xz, ... , XII in which Xl =0. This set of vectors
lt is clear from tI
is always linearly dependent, because we may choose (Xl = 1, (Xz = 0, (Xz = 0, ... ,
Xl? x2 ·> ", Xi¡ are HrJea
Cl. = O, and Equation (2-6) holds.
H id!
as a linear combinatio
every one of them can
Example 2
Consider the set of vector Xl which consists of only one vector. The set of
Definition 2-5
vector Xl is linearly independent if and only if Xl #= O. lf Xl #=0, the only way
tohaveCl.1X l =OiS(Xl =0. Ifxl=O,wemaychooseCl.l=l. I The maximal number
is called the dimension
lf we introduce the notation
In the previous sec
(IR", IR). The meaning
thereare, at most, n lin
dimensional real vect(
(2-7)
pendent veCtors. (Try
LINEAR INDEPENDENCE, BASES, AND REPRESENTATIONS 13

then the linear independence of a set of vectors can also be stated in the fol­
'esentations
lowing definition.
e mutually perpendic­
.ordinate system is to Definition 2-4'
:ctor in the planeo In
A set of vectors x ¡, x 2, ... , x" in (fll', 9') is said to be linearly independent if and
general linear spaces.
only if the equation
The basis vectors are
ferent scales. Before [Xl X2 X,,]ex=ü
)f vectors.
implies ex = O, where cvery component of ex is an element of ff or, correspond­
ingly, ex can be considered as a vector in ff". •

1ff, (fll', ff), is said to


Observe that linear dependence depends not only on the set of vectors
, 1X2' ... ,IX" in ff, not but also on the field. For example, the set of vectors {Xl' X2}, where

), ... , IX" = 0, then the


x¡ ~
s~ 1] X2~
s+2
(S+1:S+3)
]

[ 1 [
'nt. I
s+2 s+3
for IX¡ =0, 1X 2 = O, ... ,
is linearly dependent in the field of rational functions with real coefficients.
ce of the set, we have
lndeed, if we choose
for which (2-6) holds;
:he right-hand side of s+3
( dependent, there are IX¡ = -1 and Ci. 2 =-­
s+2
atisfy Equation (2-6).
lo conclude the linear then IX¡X¡ +Ci. 2X2 =0. However, this set of vectors is linearly independent in
the field of real numbers, for there exist no Ci.¡ and Ci. 2 in IR that are different from
zero, such that Ci.¡x¡ +Ci.2X2 =0. In other words, x¡ and X2 are linearly inde­
pendent in (1R 2(s), IR), but are linearly dependent in (1R 2(s), IR (s)). •

. This set of vectors


lt is clear from the definition of linear dependence that if the vectors
= 1, Ci. 2 = 0, Ci. 2 = 0, ... ,
X¡,"2, ... ,X" are linearly dependent, then at least one of them can be written
Iill
as a linear combination of the others. However, it is not necessarily true that
every one of them can be expressed as a linear combination of the others.

e vector. The set of


Definition 2-5
fx¡1=-O, the only way
d. • The maximal number of linearly independent vectors in a linear space (fl, ff)
is called the dimension of the linear space (fil, ff). •

In the previous section we introduced the n-dimensional real vector space


(IR", IR). The meaning of n-dimensional is now clear. lt means that in (lR'i, iR)
there are, at most, n linearly independent vectors (over the field IR). In the two­
dimensional real vector space (IR 2; IR), one cannot find three linearly inde­
pendent vectors. (Try 1)
~"Ja (2-7)
14 LINEAR SPACES AND LINEAR OPERATORS

Example 3 of el> e z, ... , en' No,"


Consider the function space that consists of aH real-valued piecewise continuous is another linear com
functions defined over ( - 00, 00). The zero vector in this space is the one which
is identicaHy zero on ( - 00, 00). The foHowing functions, with - 00 < t < 00,
Then by subtracting (
O=(p
are clearly elements ofthe function space. This set of functions {t n , n = 1,2, ...}
is linearly independent, because there exist no real constants, a/s, not aH zero, which, together with 1
such that
L ai=O
i= 1
This completes the pI

There are infinitely many of these functions; therefore, the dimension of this This theorem ha:
• space is infinity. • vector space (f!{, g-), i
represented by a set (
We assume that aH the linear spaces we shaH encounter are of finite dimen­ (2-7), we may write (2­
sions unless stated otherwise.

Definition 2-6 where P = [{Jl> {Jz,···


A set of linearly independent vectors of a linear space (f!{, g-) is said to be a ~ctor P can be cons
basis of q; if every vector in fE can be expressed as a unique linear combination one-to-one correspond
of these vectors. • the same dimensional,

Theorem 2-1 Definition 2-7

In an n-dimensional vector space, any set of n linearly independent vectors In an n-dimensional


qualifies as a basis. then every vector x iJ
called the re presentatl
Proof
Example 4
Let el> e z, ... , en be any n linearly independent vectors in f!{, and let x be an
arbitrary vector in!!( Then the set of n + 1 vectors x, el> e z, ... ,en is linearly The geometric plane s1
dependent (since, by the definition of dimension, n is the maximum number of real vector space. An
linearly independent vectors we can have in the space). Consequently. there ~-
exist a o, al, ... , an in ff, not all zero, such that
aoX+alel +azez + ... +anen=O (2-8)

We claim that a o =/=-0. Ir a o =0, Equation (2-8) reduces to


(2-9)

which, together with the linear independence assumption of el' ez, ... , en,
implies that al = O, az == O, ... , a = O. This contradicts the assumption that
ll

not aH ao, al>' .. , a n are zero. Ir we define {Ji ~ - aJa o, for i = 1, 2, ... , n, then
(2-8) becomes
(2-10)

This shows that every vector x in fE can be expressed as a linear combination Figure 2-1 A two-dim<
LINEAR INDEPENDENCE, BASES, AND REPRESENTATlONS 15

of el> ez, . .. , en· Now we show that this combination is unique. Suppose there
piecewise continuous is another linear combination, say
:pace is the one which x= /3l e ¡ + /3z e z + ... + /3nen (2-11 )
, with - CIJ < t < CIJ,
Then by subtracting (2-11) from (2-10), we obtain
0= (/31 -/3¡)e l + (/3z -/3z)ez + ... + (/3n -/3n)en
:tions {t n, n= 1,2, ... }
mts, lX/S, not all zero, which, together with the linear independence of {eJ, implies that
i = 1,2, ... , n
This completes the proof of this theorem. Q.E.D.

:he dimension of this This theorem has a very important implication. In an n-dimensional
• vector space (f![, .?"), if a basis is chosen, then every vector in f![ can be uniquely
represented by a set of n scalars /31, /3z, ... , /3n in.?". lf we use the notation of
er are of finite dimen­ (2-7), we may write (2-10) as
x= [el e z ... en]p (2-12)

where p = [/31' /3z, ... , /3nJ' and the prime denotes the transpose. The n x 1
vector p can be considered as a vector in (.?"n, .?"). Consequently, there is a
f![,.?") is said to be a
one-to-one correspondence between any n-dimensional vector space (f![, .?") and
le linear combination
I
the same dimensional linear space (.?"n, .?") if a basis is chosen for (f![, .?").

Definition 2-7

independent vectors In an n-dimensional vector space (f![, .?"), if a basis {el> ez, . .. ,en} is chosen,
then every vector x in f![ can be uniquely written in the fonn of (2-12). P is
called the representation of x with respect to the basis {el, ez' ... , en}. •

Example 4
n and let x be an
f![,
L, ez, ... , el! is linearly
The geometric plane shown in Figure 2-1 can be considered as a two-dimensional
maximum number of real vector space. Any point in the plane is a vector. Theorem 2-1 states that
Consequently, there

(2-8)

(2-9)

üon of el' ez, ... , e,,,


the assumption that
or i = 1, 2, ... ,n, then

(2-10)

i a linear combination Figure 2-1 A two-dimensional real vector space.


16 LINEAR SPACES AND LINEAR OPERATORS

Table 2-1 Different Representations ofVectors real vector space (IR


space (IRn(s), lR(s)); a
'"
Bases~s b
functions, written as

[~J [~J [~J [~J [~J

This array of n numl


such; that is, it is a ve·
of a vector with res¡
numbers, unless it is
any set of two linearly independent vectors forms a basis. Observe that we vector. However we :
have not only the freedom in choosing the directions of the basis vectors (as vectors 5 :
long as they do not lie in the same line) but also the magnitude (scale) of these
1
vectors. Therefore, given a vector in (IR 2, IR), for different bases we have different
O
representations of the same vector. For example, the representations of the
o¡ =
O
vector b in Figure 2-1 with respect to the basis {e¡,e 2} and the basis {e¡,e 2}
are, respectively, [1 3]' and [-1 2]' (where the "prime" symbol denotes the
O
transpose). We summarize the representations ofthe vectors b, el, e 2, el, and e 2
O
with respect to the bases {el, e 2}, {el, e2} in Table 2-1.
as the basis of(lRn, IR),
Example 5 can be interpreted as
Consider the linear space (1R 4 [s], IR), where 1R 4 [s] is the set oC all real poly­ °
the basis {O¡, 2 " •. , JI
nomials of degree less than 4 and with indeterminate s. Let e ¡ = S3, e 2 = S2, representation and th,
e 3 =s, and e 4 = 1. Clearly, the vectors e¡, i = 1,2,3,4, are linearly independent
and qualify as basis vectors. With this set of basis vectors, the vector x =
3s 3 + 2s 2 - 2s + 10 can be written as

Changa of basis. ,
representations with ]
hence [3 2 - 2 10]' (where the "prime" denotes the transpose) is the relationships are bet\
representation of x with respect to {e ¡, e 2, e 3, e 4}. Ir we choose e ¡ = S3 - S2, In this subsection, thi:
e2 = S2 - s, e3 = s -1, and e4 = 1 as the basis vectors, then Let the representa1

x: ::: +::' ~,~ :,;OW" -,') + 5(,' -,)+ 3(, -1) + 131
and {e¡, e 2 , .•• ,en} be
x=[
In order to derivf
infOI:mation of the rel

Hence the representationofx with respectto {e¡, e 2, e 3, e 4} is [3 5 3 13]'. I


5 This set of vectors is c~lle
In this example, there is a sharp distinction between vectors and representa­ 6 One migot be. tempte,
tions. However, this is not always the case. Let us consider the n-dimensional [e 1 e z ... en] - 1 may
LINEAR INOEPENOENCE BASES, ANO REPRESENTATIONS 17

real vector space (IR n, IR), complex vector space (en, C), or rational vector
space (W(s), lR(s)); a vector is an n-tuple of real, complex, or real rational
functions, written as

This array of n numbers can be interpreted in two ways: (1) lt is defined as


such; that is, it is a vector and is independent of basis. (2) It is a representation
of a vector with respect to sorne fixed unknown basis. Given an array of
numbers, unless it is tied up with sorne basis, we shall always consider it as a
:is. Observe that we vector. However we shall also introduce, unless stated otherwise, the following
. the basis vectors (as vectors 5 :
nitude (scale) of these 1 O O O
Jases we have different O 1 O O
epresentations of the O O , O O
and the basis {e 1> ez } 01 = , Oz= nn-l = , o n­
­ (2-13)
~" symbol denotes the
O O 1 O
ors b, el, ez' el' and e z O O O 1

as the basis of(lR n, IR), (en, C), and (IRn(s), lR(s)). In this case, an array ofnumbers
can be interpreted as a vector or the representation of a vector with respect to
~ set of aH real poly­ the basis {o¡, 0z, ... , on}, because with respect fo this particular set of bases, the
Let el =s3,e z =sz, representation and the vector itself are identical; that is,
: linearly independent
ctors, the vector x =
(2-14 )

Change of basis. We have shown that a vector x in (~, ff) has different
representations with respect to different bases. it is naturalto ask Vvhái: the
:he transpose) is the relationships are between these different representations of the same vector.
e choose el = S3 - sZ, In this subsection, this problem will be studied.
L Let the representations of a vector x in (,qr, ff) with respect to {e¡, e z, ... , en}
and {el' ez, ... , en} be ~ and p, respectively; that iS,6
s-1)+13-1
x=[e¡ e z ._. en]~=[e¡ ez en]p (2-15)

In order to derive the relationship between ~ and p, we need either the


information of the representations of ej, for i = 1, 2, ... , n, with respect to the

Hs [3 5 3 13]'. I
5 This set or vectors is called an orthonormal set.
:ctors and representa­ 60ne might be tempted to write p=[e 1 ez '" en]-l[e¡ ez ... cn]jl. However,
der the n-dimensional [el e z .. . en] - ¡ may not be defined as can be seen rrom Example 5.
18 LINEAR SPACES AND LINEAR OPERATORS

basis {e b ez, ... ,en}, or the inforrnation of the representations of el' for i = Example 6
1,2, ... , n, with respect to the basis {e b ez, ... , en}. Let the representation of
e i with respect to {el, e z,···, en} be [PI¡ PZi P3¡ ... Pni]'; that is, Consider the two set
readily verified that

i= 1,2, ... , n (2-16)

where E ~ [el ez en], Pi ~ [Pli PZi ... Pn¡J'· Using rnatrix nota­ and
tion, we write
[el e z ... en] = [EpI Epz ... EPn] (2-17)

which, by using (2-2), can be written as


[el e z ... en] = E[PI pz ... P/I]
Clearly, we have PQ .

l
PII PIZ
en] Pfl pzz PI]
PZn

pnl P/lZ ;/ln


~ [el ez en]P (2-18)

Substituting (2-18) into (2-15), we obtain


2-4 linear Opel
x = [el e z ... en]Pp = [el (2-19)

Since the representation of x with respect to the basis {e b ez, ... ,en} is uniqtie, The concept of a func1
(2-19) irnplies úJI, ifwe assign to each
p=Pp (2-20)
of assignments is can
Figure 2-2(a) is a fur
ith colurnn: the
P = representation of
J usually denoted by th(
to the element x of f![
where (2-21 )
e with respect to defined is caBed the d
[ i
{el' ez, ... , e/l} to sorne elernent of ~
This establishes the relationship between p and p. In (2-16), if the repre­
sentation of e¡ with respect to [e j , e z, . .. , e,J is used, then we shall obtain
(2-22 )

ith colurnn: the ]


where Q = ~epr.esentation of (2-23)
ei w¡th respect to
f
{e b e z,···, en} o
Different representations of a vector are related by (2-20) or (2-22). There­ y =Lx
fore, given two sets of bases, if the representation of a vector with respect to -1
one set of bases is known, the representation of the sarne vector with respect
to the other set ofbases can be cornputed by using either (2-20) or (2-22). Since
(a)
p = Pp and p = QP, we have p = PQp,for aH p; hencewe conclude that
Figure 2-2 Examples i
PQ = 1 or P = Q - 1 (2-24 ) not representa function.
LINEAR OPERATORS AND THEIR REPRESENTATlONS 19

rations of ei' for i = Example 6


the representation of
Consider the two sets of basis vectors of (~4[SJ, IR) in Exarnple 5. lt can be
Oni]'; that is,
readily verified that

2, ... , n (2-16)

Using rnatrix nota­ and

(2-17)
~
le. e, " '4] l'. " " '4] l- ~ -1
o
1
O
O

O -1
1 ~J~l" "" '4]Q

Clearly, we have PQ = 1 and ~ = P~, or

PI]

P2n

Pnn
(2-18)
rn =Pl ~lJ::
ll~J ~~ U 1
PJ~ rl~~J~l

2-4 linear Operators and Their Representations


(2-19)

, e2 , . . . , en} is unique, The concept of a function is basic to all parts of analysis. Given two sets f![ and
11JJ, ifwe assign to each elernent of f![ one and only one elernent of 11JJ, then the rule
(2-20)
of assignrnents is called a function. For exarnple, the rule of assignrnents in
Figure 2-2(a) is a function, but not the one in Figure 2-2(b). A function is
usually denoted by the notation f : f![ ---> 11JJ, and the elernent of l1JJ that is assigned
(2-21 ) to the elernent x of f![ is denoted by y = f(x). The set f![ on which a function is
defined is called the domain o! the function. The subset of l1JJ that is assigned
to sorne elernent of f![ is called the range of the function. For exarnple, the
:n (2-16), if the repre­
1 we shall obtain 1{

(2-22)

(2-23)

--"-0f----\---r----+--------'L----.X
·20) or (2-22). There­ y =Lx
'ector with respect to -1
.e vector with respect
2-20) or (2-22). Since
(a) (b)
:conclude that
Figure 2-2 Examples in which (a) the curve represents a function and.(b) the curve does
(2-24 ) not represent a functlon.
20 LINEAR SPACES AND LINEAR OPERATORS

domain ofthe function shown in Figure 2-2(a) is the positive realline, the range is a linear transforma1
of the function is the set [ -1, 1], which is a subset of the entire real Hne úJ/. all equal to (O/t, IR).
The functions we shall study in this section belong to a restricted class of
functions, called linear functions, or more often called linear operators, linear Matrix representa
mappings, or linear transformations. The sets associated with linear operators
two examples that the
are required to be linear spaces over the same field, say (gr, ff) and (úJ/, ff). A defined may be of fini
linear operator is denoted by L:( gr, ff) ---> (úJ/, ff). In words, L maps (gr, ff)
every linear operatc
into (úJ/, ff). dimensional (úJ/, ff) h:
Ir (gr, ff) and (úJ/, ff)
Definition 2-8 operator can still be I
A function L that maps (gr, ff) into (úJ/, ff) is said to be a linear operator if infinite order or a fon
and only if and will not be discus
L((;(¡x¡ + (;(zx z ) = (;(¡Lx¡ + (;(zLx z
Theorem 2-2
for any vectors Xl' X z in grand any scalars (;(¡, (;(z in ff. I
Let (gr, ff) and (úJ/, ¿
Note that the vectors Lx¡ and Lx z are elements of 0Jj. The reason for over the same field. ]
requiring that úJ/ be defined over the same field as gr is to ensure that (;(¡Lx¡ in gr. Then the line:
and (;(zLx z be defined. by the n pairs of m,
respect to the basis {x
Example 1 be represented by an
column af A is the rep]
Consider the transformation that rotates a point in a geometric plane counter­
clockwise 90° with respect to the origin as shown in Figure 2-3. Given any Proof
two vectors in the plane, it is easy to verify that the vector that is the sum of the
Let x be an arbitrary \
two vectors after rotation is equal to the rotation of the vector that is the sum
the set af vectors qua
ofthe two vectors before rotation. Hence the transformation is a linear trans­
pressed uniquely as (;(
formation. The spaces (gr, ff) and (úJ/, ff) of this example are all equal to
of L, we have
(IR Z , ~). I

Example 2
which implies that fOj
Let O/t be the set of all real-valued piecewise continuous functions defined over i = 1, 2, ... ,n. This p
[O, T] for sorne finite T > O. lt is clear that (O/t, IR) is a linear space whose
Let the represeI
dimension is infinity (see Example 3, Section 2-3). Let 9 be a continuous
[au aZi ... ami]';
function defined over [0, T]. Then the transformation

y(t)= r g(l-r)u(r)dr (2-25)

where the a¡Js are ele!


2
L[x¡ Xz

------------
---
.-------- -­
~s -3 Y2 -0.5 xI 1.5
Figure 2-3 The transformation that rotates a vector counterclockwise 90°.
LINEAR OPERATORS AND THEIR REPRESENTA:rIONS 21
ive realline, the range
is a linear transformation. The spaces (gr, ff) and (qy, ff) of this example are
e entire realline qy.
aH equal to (Ii/l, IR). I
o a restricted class of
inear operators, linear
Matrix representations of a linear operator. We see fram the aboye
with linear operators
two examples that the spaces (gr, ff) and (qy, ff) on which a linear operator is
gr, ff) and (qy, ff). A
defined may be of finite or infinite dimensiono We show in the foHowing that
'ords, L maps (gr, ff)
every linear operator that maps finite-dimensional (gr, ff) into finite­
dimensional (qy, ff) has matrix representations with coefficients in the field ff.
lf (gr, ff) and (qy, ff) are of infinite dimension, a representation of a linear
operator can still be found. However, the representation will be a matrix of
Je a linear operator if infinite order or a form similar to (2-25). This is outside the scope of this text
and will not be discussed.

Theorem 2-2
I
Let (gr, ff) and (qy, ff) be n- and m-dimensional vector spaces, respectively,
f O.!/. The reason for over the same field. Let Xl' X2, ... ,XII be a set of linearly independent vectors
to ensure that cxlLx l in gr. Then the linear operator L:( gr, ff) ~ (qy, ff) is uniquely determined
by the n pairs of mappings Y¡ = Lx¡, for i = 1,2, ... ,n. Furthermore, with
respect to the basis {Xl' X2, ... , XII} of gr and a basis {u l , U2, ... , um} of qy, L can
be represented by an m x n matrix A with coefficients in the field ff. The ith
column of A is the representation OfYi with respect to the basis {Ul' U2,.··, u m }.
metric plane counter­
gure 2-3. Given any Proof
. that is the sum úf the
vector that is the sum Let X be an arbitrary vector in gr. Since Xl' X2, ... , X" are linearly independent,
ltion is a linear trans­ the set of vectors qualifies as a basis. Consequently, the vector X can be ex­
nple are aH equal to pressed uniquely as CXIX l +CX2X2 + ... +CX"x1l (Theorem 2-1). By the linearity
of L, we have
I
Lx = cxlLx l +cx2Lx2 + ... + cx"Lx"
=CX1Yl +CX2Y2 + ... +CX"YII
which implies that for any X in fI, Lx is uniquely determined by Y¡ = LXi' for
'unctions defined over
.i = 1, 2, ... ,n. This praves the first part of the theorem.
a linear space whose
Let the representation of y¡ with respect to {1Ul 1" lffi2" ' , . ,íl¡¡ml be
~t g be a continuous
[al ¡ a2; ... am¡J'; that is,

l
l

(2-25) n.] a a,: '] i= 1,2, ... , n (2-26)

amI
22 LINEAR SPACES AND LINEAR OPERATORS

Note that the elements of A are in the field 9' and the ith column of A is the between IX and eX and 1
representation of y i with respect to the basis of 1fJ/. With respect to the basis related by eX = Pez and
{x¡,x 2 , ... ,x n} of (gr, 9') and the basis {u¡,u 2 , ... ,u m} of (1fJ/, 9'), the linear in the field 9' and the
operator y = Lx can be written as the basis {el' e2, ... ,e
[u l U2 ... umJJI= L[x l X2 xnJex (2-28)
where JI ~ [,81 ,82 . .. ,8mJ' and ex ~ [al a2 anJ' are the representa- Hence, by the uniqué
tions ofy and x, respectively. After the bases are chosen, there are no differences we have Aa=PAP-l
between specifying x, y, and ex, JI; hence in studying y = Lx, we may just study
the relationship between JI and ex. By substituting (2-27) into (2-28), we obtain
or
[u l u2 ... umJJI= [u l U2 ... umJAex (2-29)

where Q ,6,. p-l.

which, together with the uniqueness of a representation, implies that


Two matrices A a
JI=Aex (2-30) matrix P satisfying (:
similarity transformat.
Hence we conclude that if the bases of (gr, 9') and (1fJ/, 9') are chosen, the to different bases) oftJ
operator can be represented by a matrix with coefficients in 9'. Q.E.D.
Example 3
We see from (2-30) that the matrix A gives the relation between the repre­
sentations ex and p, not the vectors x and y. We also see that A depends on the Consider the linear o¡:
basis chosen. Hence, for different bases, we have different representations of {x¡, x 2 } as a basis, the
the same operator.
We study in the following an important subclass of linear operators that
maps a linear space (g[, 9') into itself; that is, L:(fI, 9')-+(gr, 9').. In this
case, the same basis is always used for these two linear spaces. If a basis of g[, Hence the representat
say {el e2, ... , en}, is chosen, then a matrix representation A of the linear
operator L can be obtained by using Theorem 2-2. For a different basis
{e1, e2, ... , en}, we shall obtain a different representation Á of the same operator
L. We shall now establish the relationship between A and Á. Consider The representation of
Figure 2-4; x is an arbitrary vector in g[; ex and eX are the representations of x
with respect to the basis {e¡, el>' .. ,en} and the basis {e¡, e2' ... ,en}, respec­
tively. Since the vector y = Lx is in the same space, its representations with
respect to the bases chosen, say JI and p, can also be found. The matrix repre­ 1t is easy tQ verify that
sentations A and Á can be computed by using Theorem 2-2. The relationships

Independent
of basis

L
x ----+y(=Lx) ith column:
Representation of
]
ith
f
column:
Representation of
J
A = ~¡ wit? respect to ,P = e~ Wi!h respect to or

Basis [ the basls {el' e 2, ... , en}


{el' e b · .. , en}
[el If, instead of {Xl'

1
~
'th
column:
Representation of
] llh coiumn:
Representation of Yl = LXI = [Xl
Basis Á = Le¡ with respect to ,Q = e¡ with respect to
the basis {el' e 2, ... , en} Hence the representat
[el
{el' ez,· .. ,en}
Figure 2-4 Relationships between dilTerent representations of the same operator.
LINEAR OPERATORS AND THElR REPRESENTATlONS 23

ith column of A is the between IX and (j and between p and ~ have been established in (2-20); they are
:h respect to the basis related by (j = Prx and ~ = pp, where Pis a nonsingular matrix with coefficients
of (óJI, ~), the linear in the field ff and the ith column of P is the representation of e i with respect to
the basis {el, ez, ... , en}. From Figure 2-4, we have
(2-28) ~= A(j and ~= Pp= PAIX= PAP-¡(j
nJ' are the representa­ Hence, by the uniqueness of a representation with respect to a specific basis,
here are no differences we have A(j = PAP-¡(j. Since the relation holds for any (j, we conclude that
Lx, we may just study A = P AP - ¡ = Q - ¡ AQ (2-31a)
into (2-28), we obtain
or A = P - ¡AP = QAQ - ¡ (2-31b)
lm]AIX (2-29)
where Q ~ P- . 1
implies that
Two matrices A and A are said to be similar if there exists a nonsingular
(2-30) matrix P satisfying (2-31). The transformation defined in (2-31) is caBed a
similarity transformation. Clearly, al! the matrix representations (with respeet
~, ff) are chosen, the
to different bases) of the same operator are similar.
; in ff. Q.E.D.
Example 3
Jn between the repre­
that A depends on the Consider the linear operator L of Example 1 shown in Figure 2-3. If we choose
ent representations of {Xl' Xz} as a basis, then

linear operators that Y¡=Lx¡=[x 1 xz][~J and Yz=Lxz=[x¡ X Z][ -~J


ff) -4 (El", ff). In this
laces. lf a basis of El", Hence the representation of L with respect to the basis {x¡, Xz} is
liion A of the linear
For a different basis
\. of the same operator
A and A. Consider The representation of x 3 is
e representations of x
e¡, ez, ... , en}, respec­ 1.5J
[ 0.5
; representations with
ld. The matrix repre­ It is easy to verify that the representation af)' ~ with respect te (~( 1 -)"(2] is equal to
.-2. The relationships
O -IJ_[1.5J=[-0.5J
[1
ith column: ~ O 0.5 1.5
Representation of

,P =[ e¡ with respect to or Y3=[X¡ x z] [ -0.5J

{el' e2,···, en} 1.5


If, instead of {x¡, XZ}, we choose {Xl' X3} as a basis, then from Figure 2-3,

"th column:
Representation of
~ Y¡=LX¡~[~~. X3] [-~J and Y3=Lx3 =[x¡ X3J[-~J
r
, Q = e, with respect to

{el. e z,··., en} Bence the representation of L· with respect t6 the basis {x¡, X3} is

he same operator.
24 LINEAR SPACES AND LINEAR OPERATORS

The reader is advised to find the P matrix for this example and verify A =
PAP- 1 . • Ab=

In matrix theory, a matrix is introduced as an array ofnumbers. With the It can be shown that tI
concepts of linear operator and representation, we shall now give a new inter­
pretation of a matrix. Given an n x n matrix A with coefficients in a field ff,
if it is not specified to be a representation of sorne operator, we shall consider it Since the set of vectors
as a linear operator that maps (ff n , ff) into itself. 7 The matrix A is independent basis. We compute TI
of the basis chosen for (ff n , ff). However, if the set of the vectors 01' 02, ... ,on It is clear that
in Equation (2-13) is chosen as a basis of (ff n , ff), then the representation of
the linear operator A is identical to the linear operator A (a matrix) itself. This
can be checked by using the fact that the ith column of the representation is
equal to the representation of Ao¡ with respect to the basis {o 1o 02, ... , 0n}. lf
a¡ is the ith column of A, then AO i = a i . Now the representation of a¡ with
respect to the basis (2-13) is identical to itself. Therefore we conclude that the
representation of a matrix (a linear operator) with respect to the basis (2-13)
is identical to itself. For a matrix (an operator), Figure 2-4 can be modified as in
Figure 2-5. The equation Q = [q 1 q2 ... qnJ follows from the fact that the and
ith column ofQ is the representation of q¡ with respect to the basis {0 1o "2,"" Oll}'
If a basis {ql' q2, ... , qn} is chosen for (!F", ff), a matrix A has a representa­ The last equation is ob
tion A. From Figure 2-5, we see that the matrix representation A may be respect to the basis {b,
computed either from Theorem 2-2 or from a similarity transformation. In
most of the problems encountered in this book, it is always much easier to
compute A from Theorem 2-2 than from using a similarity transformation.

Example 4 The matrix A can al:


of a matrix and n 3 mI
Consider the following matrix with coefficients in IR :
more easily, QA = AQ,
3 2
L=A= [ -~ ~ -~ -~l
1 -iJ
Let b~ m
;::xample 5
We extend Example 4
with real coefficients.
7 This interpretation can be extended to nonsquare matrices. Ab, ... , A"- 1b are Hne;
1X1An-lb(see Section 2­
{b,Ab, ... ,An- 1b} is
lndependent of basis ~.
L~A
,JI rro¡umn J
Á = Representation of Aq¡
with respeet to
A
Basis [o¡ °2 '"
°n] ~ IJI {q¡, ih, ... , qn}

Qr IQ Á=Q-¡AQ
q2 .. ,
Á
Basis [q¡ qn] ~
'Ji Q=[q¡ q2 qn] =p-¡

Figure 2-5 Oifferent representations ofa matrix (an operator).


LINEAR OPERATORS AND THEIR REPRESENTATIONS 25

mple and verify Á =

• Then 8

,f numbers. With the


It can be shown that the following relation holds (check!):

now give a new inter­


~fficients in a field ff,
A 3 b = 5A 2b -15Ab + 17b (2-32)

Ir, we shall consider it Since the set of vectors b, Ab, and A2b are linearly independent, it qualifies as a
atrix A is independent basis. We compute now the representation of A with respect to this basis.
~ vectors "1' "2' ... ,"n It is clear that

A(b)~ [b
the representation of
la matrix) itself. This
the representation is
Ab A'bJ [!J
lis {" l' n2 , . . . , "n}. lf
~esentation of a¡ with
we conclude that the
:ct to the basis (2-13)
A(Ab)~ [b Ab A'bJ m
A(A'b)~ [b [-:~J
f can be modified as in
from the fact that the and Ab A'bJ
le basis {n¡, "2,···, Un}.
X A has a representa­
The last equation is obtained from (2-32). Hence the representation of A with
esentation Á may be respect to the basis {b, Ab, A2b} is
I transforrnation. In
ways much easier to 0:0: 17]
ity transformation. A= 1 ;0: -15
[o 1 5
I I
I •

The matrix A can also be obtained from Q - 1 AQ, but it requires an inversion
of a matrix and n 3 multiplications. However, we may use Á = Q -1 AQ, or
more easily, QA = AQ, to check our resulto The reader is asked to verify

[~ -~l[~ ~ _~~l ~
-1
-~]
-1
O = [_; O
1 -iJ O 1 ~J 4 3 1 -3
Example 5
We extend Example 4 to the general case. Let A be an n x n square matrix
with real coefficients. lf there exists a real vector b such that the vectors b,
Ab, ... ,An-Ib are linearly independent and if Anb= -(;(nb-(;(n_IAb- ... ­
(;( 1 An - 1 b (see Section 2-7), then the representation of A with respect to the basis
{b, Ab, ... , A n - 1 b} is

~~ Aql
of
O O
1 O
O -(;(n

O -(;(n-I

~} J A=
O 1 O -(;("-2
(2-33)

O O O -(;(2

O O 1 -(;(1


8 A 2 ~ AA, A 3 ~ AAA.
26 LINEAR SPACES AND LINEAR OPERATORS

A matrix of the form shown in (2-33) or its transpose is said to be in the operator which maps
companionform. See Problem 2-26. This form will constantly arise in this text. n
(ff ,ff) that undergo
As an aid in memorizing Figure 2-5, we write A = Q - 1AQ as
D&finition 2-9
QA=AQ
The range of a linear (
Since Q= [ql q2 qnJ, it can be further written as
flI (A)= {all the elem(
[ql q2 qnJ A = [Aql Aq2 AqnJ (2-34)
vector x ir
From (2-34), we see that the ith column of A is indeed the representation of
Aqi with respect to the basis {ql' q2, ... , qn}. Theorem 2-3
We pose the following question to conclude this section: Since a linear
operator has many representations, is it possible to choose one set of basis The range of a linear (
vectors such that the representation is nice and simple? The answer is affirma­
tive. In order to give a solution, we must first study linear algebraic equations. Proof

lf Y1 and y 2 are elemel


x 2 in (ff n , ff) such th;
2-5 Systems of Linear Algebraic Equations in ff, the vector (X1Y 1 +
of A, it is easy to shO\
Consider the set of linear equations: (X1X 1 + (X2 X 2 is an elem
( ffm, ff) (see the rema
a11x 1 +a12 x 2+ +alnxn=Yl
a21 x l + a22x 2 + +a2nXn = Y2 (2-35)
Let the ith column
then the matrix equati­

where the given aJs and y/s are assumed to be elements of a field ff and the
unknown x/s are also required to be in the same field ff. This set of equations where Xi, for i = 1, 2, ..
can be written in matrix form as range space fAlCA) is, b
Ax=y (2-36) (ff n , ff). lt is the sam
(2-37) ranging througl

[
~~: ~~~ :::
that flI(A) is the set of
Since fAlCA) is a linear s¡:
where A~ : : number of linearly in
9j'(Aj ¿s the maximurn n.
a m1 a m2
Clearly, A is an m x n matrix, x is an n x 1 vector, and y is an m x 1 vector. Definition 2-10
No restriction is made on the integer m; it may be larger than, equal to, or The rank of a matrix j
smaller thanthe integer n. Two questions can be raised in regard to this set of independent columns
equations: first, the existence of a solution and, second, the number of solutions. of A.
More specifically, suppose the matrix A and the vector y in Equation (2-36) are
given; the first question is concerned with the condition on A and y under which Example 1
at leastone vectorx exists such that Ax = y. If solutions exist, then the second
question is concenled with the number of linearly independent vectors x such Consider the matrix
that Ax == y. Inorder to answer these questions, the rankand the nullity of
the matrix A have to be introduced.
We have agreed in the previous section to consider the matrix A as a linear
SYSTEMS OF LINEAR ALGEBRAIC EQUATlONS 27

se is said to be in the operator which maps (ff", ff) into (ff m , ff). RecaB that the linear space
tant1Yarise in this tex t. (ff", ff) that undergoes transforrnation is caBed the domain of A.
-IAQ as
D&finition 2-9
The range of a linear operator A is the set Yf.(A) defined by
fJJI (A)= {aH the elements y of (ff m , ff) for which there exists at least one
(2-34 )
vector x in (ff", ff) such that y = Ax} !II
the representation of
Theorem 2-3
ection: Since a linear
lOO se one set of basis The range of a linear operator A is a subspace of (ff m , ff).
The answer is affirma­
ar algebraic equations. Proof
lf YI and y z are elements of Yf.(A), then by definition there exist vectors XI and
Xz in (ff", ff) such that YI =Ax l , yz=Ax z. We claim that for any (XI and (X2
in ff, the vector (X IY I + (XzY z is also an element of fJJI(A). lndeed, by the linearity
of A, it is easy to show that (XIYI + (XzYz = A((Xlx l + (Xzx 2), and thus the vector
(X IX I + (X2XZ is an element of (ff", ff). Hence the range .<?l(A) is a subspace of
(ff m , ff) (see the remark following Definition 2-3). Q.E.D.

(2-35 ) Let the ith column of A be denoted by a¡; that is, A = [al az
then the matrix equati.on (2-36) can be written as
(2-37)
:s of a field ff and the
This set of equations where Xi, for i = 1,2, ... , n, are components of x and are elements of ff. The
range space fJJI(A) is, by definition, the set of y such that y = Ax for some x in
(2-36) (ff", ff). lt is the same as sayingthat fJJI(A) is the set ofy with XI' X2"'" X" in
(2-37) ranging through aB the possible values of ff. Therefore we conclude
that fJJI(A) is the set of all the possible linear combinations of the columns of A.
Since fJJI(A) is a linear space, its dimension is defined and is equal to the maximum
number of linearly independent vectors in fJJI(A). Hence, the dimension of
fJJI(A) is the maximum number of linearly inde pendent columns in A.

i Y is an m x 1 vector. Definition 2-10


~ger than, equal to, or The rank of a matrix A, denoted by peA), is the maximum number of linearly
in regard to this set of independent columns in A, or equivalently, the dimension of the range space
le n umber of sol utions. clA •
. in Equation (2-36) are
In A and y under which Example 1
s exist, then the second
Jendent vectors x such Consider the matrix
ank and the nullity of

he matrix A as a linear
28 LINEAR SPACES AND LINEAR OPERATORS

1. Given A and giver


The range space of A is aH the possible linear combinations of aH the columns
Ax = Y if and only
of A, or correspondingly, all the possible linear combinations of the first two
columns of A, because the third and the fourth columns of A are linearIy depen­
dent on the first two columns. Hence the rank of A is 2. •
2. Given A, for ever)
if and only if ~(A)
The rank of a matrix can be computed by using a sequence of elementary
transformations (see Appendix A). This is based on the property that the rank
Proof
of a matrix remains unchanged after pre- or postmultiplications of elementary
matrices (Theorem 2-7). Once a matrix is transformed into the upper triangular lo It follows immedia
form as shown in (A-6), then the rank is equal to the number of nonzero rows in vector y is not an e
(A-6). From the form, it is also easy to verify that the number of linear inde­ is said to be incons,
pendent columns of a matrix is equal to the number of independent rows. 2. The rank of A, p(J
Consequently, if A is an n x m matrix, then is a subspace of (
(!F m , ff), then for a
rank A = no. of linear independent columns
= no. of linear independent rows
lf p(A) < m, there e;
~(A), for which the
':::;;min (n, m) (2-38)

The computation of the rank of a matrix on digital computers, however, is The design of con
not a simple problem. Because of limited accuracies on digital computers, reduced to the solutio
rounding errors always arise on numerical computations. Suppose a matrix, very important in OUl
after transformation, becomes then a11 rows of A are 1
lf A has a full row ranl
[A:y] cannot increase
spanned by the colun
where e is a very small number, say, 10 -1 0. This e may arise from the given combination of the co
data (assuming no rounding errors) or from rounding errors. lf e arises from there exists a x such ti
rounding errors, we should consider e as zero, and the matrix has rank 1. lf said to have a full coh
e is due to the given data, we cannot consider it as a zero, and the matrix has vector a will lie in the
rank 2. To determine what value is small enough to be considered as a zero is a be written as a linear c(
complicated problem in computer computations. For problems encountered a 1 x m vector b such t
on matrix computation, the reader is referred to References S181, S182, S200, After we find out th
and S212. to ask how many solUl
In matrix theory, the ranle of a matrix is defined as the largest order of aH 'Ne discuss only the ha
nonvanishing minors of A. In other words, the matrix A has rank k if and
only if there is at least one minor of order k in A that does not vanish and every Definition 2-11
minor of order higher than k vanishes. This definition and Definition 2-10 The null space of a line
are, in fact, equivalent; the proof can be found, for example, in Reference 43.
A consequence is that a square matrix has full rank ifand only ifthe determinant X(A)= {aH
ofthe matrix is different from zero; or correspondingly, a matrix is nonsingular The dimension of X(A
if and only if all the rows and columns of the matrix are linearly independent.
With the concepts of range space and rank, we are ready to study °the
In other words, tht
o

existence problem of the solutions ofAx = y.


It is easy to show tha
o

.;V(A), veA), is 0, then


Theorem 2-4
Considerthe matrix equationAx = y, where the m x n matrix A maps (!F", !F) 9 lt is also called the right nt
nu 11 space of A. See Probl
into (!F m , ,'1».
SYSTEMS OF LINEAR ALGEBRAIC EQUATlONS 29

1. Given A and given a vector y in (ff m , ff), there exists a vector x such that
ms of aH the columns
Ax = Y if and only if the vector y is an element of P4(A), or equivalently,
ltions óf the first two
i A are linearly depen­ p(A) = p([A:y])
• m
2. Given A, for every y in (ff , ff), there exists a vector x such that Ax = Y
if and only if ~(A) = (ff m , ff), or equivalently, p(A) = m.
quence of elementary
lroperty that the rank
Proof
cations of elementary
o the upper triangular 1. lt follows immediately from the definition of the range space of A. lf the
ler of nonzero rows in vector y is not an element of ~(A), the equation Ax = Y has no solution and
umber of linear inde­ is said to be inconsistent.
of independent rows. 2. The rank of A, p(A), is by definition the dimension of ~(A). Since ~(A)
is a subspace of (ff m , ff), if p(A) = m, then ~(A)=(ffm, ff). lf ~(A)=
(ff m , ff), then for any y in (ff m , ff), there exists a vector x such that Ax = y.
mns
lf p(A) < m, there exists at least one nonzero vector y in (ff m , ff), but not in
~(A), for which there exists no x such that Ax = y. Q.E. D.
(2-38)

tmputers, however, is The design of compensators to achieve various design objectives can be
)n digital computers, reduced to the solution of linear algebraic equations. Hence this theorem is
s. Suppose a matrix, very important in our application. lf A is an m x n matrix and if p(A) = m,
then all rows of A are linearly independent and A is said to have afull row rank.
lf A has a full row rank, no matter what column y is appended to A, the rank of
[A:y] cannot increase and is equal to p(A). In other words, y lies in the space
spanned by the columns of A and, consequently, can be written as a linear
I arise from the given combination of the columns of A. Hence if A has a full row rank, for any y,
Tors. lf e arises from there exists a x such that Ax = y. Similarly, if p(A) = n, the m x n matrix A is
natrix has rank 1. lf said to have a full column rank. lf A has a full column rank, then every 1 x n
o, and the matrix has vector a willlie in the space spanned by the rows of A and, consequently, can
msidered as a zero is a be written as a linear combination of the rows of A. In other words, there exists
lroblems encountered a 1 x m vector b such that bA = a.
Ices S181, S182, S200, After we find out that a linear equation has at least one solution, it is natural
to ask how many solutions it may have. Instead of studying the general case,
he largest order of all we discuss only the homogeneous linear equation Ax: = O.
. A has rank k if and
; not vanish and every Definition 2-11
I and Definition 2-10 The null space of a linear operator A is the set .IV(A) defined by
nple, in Reference 43.
mly ifthe determinant .IV(A) = {aH the elements x of (ff", ff)for which Ax = O}
1 matrix is nonsingular
The dimension of .IV(A) is called the nullity of A and is denoted by v(A). •
inearly independent.
e ready to study the Inother words, the null space fi(A) is the set of all solutions ofAx =0. 9
lt iseasy to show that .A!(A) is indeed a linear space. lf the dimension of
fi(A), v(A), is O, then fi(A) consists of only the zero vector, and the only

9 It
is also called the right null space or A. The set of all y satisfying y A = O will be called the lert
atrix A maps (ff", ff)
null space of A. See Problem 2-51.
30 LINEAR SPACES AND LINEAR OPERATORS

solution ofAx = Ois x = O. Ir v(A) = k, then the equation Ax = Ohas k linearly be arbitrarily chosen.
independent vector solutions. vectors in Y(A). Hen
Note that the null space is a subspace of the domain (ff n , ff), whereas the theorem; its formal pn
range space is a subspace of (ff m , ff).
Theorem 2-5
Example 2
Let A be an m x n mat
Consider the matrix

h[! i i ~ =iJ
Corollary 2-5
The number of linear
which maps (~5,~) into (~3, ~). lt is easy to check that the last three columns n - p(A), where n is th
of A are linearly dependent on the first two col umns of A. Hence the rank of linearly independent c(
A, p(A), is equal to 2. Let x = [Xl Xz X3 X4 xs]'. Then
This corollary follc
null space of A. 1t is
is x = O, which is calle
find a nonzero vector x
then Ax = O has a nor
(2-39) det (A) = O, where det s
We introduce three

Since the vectors [O 1 2J and [1 2 O]' are linearly independent, we Theorem 2-6 (Sylves
conclude from (2-39) that a vector x satisfies Ax = Oif and only if Let A, B be q x n and 1

XI +X3 +xs =0 p(A)


xz+X3+ 2x 4- XS=0
Proof
Notethat the number of equations is equal to the rank of A, p(A). The solution
x ofAx = Ohas five components but is governed by only two equations; hence The composite matrix
three of the five components can be arbitrarily assigned. Let X3 = 1, X 4 = O, applied successively to
X5 = O; then XI = - 1 and xz = - 1. Let X 3 = O, X 4 = 1, X s = O; then Xl = Oand of AB is ~(B) and the r
Xz = -2. Let X3 = O, X 4 = O, Xs = 1; then XI = -1 and Xz = 1. lt is clear that p(AB) S min (p(A), p(B)
the three vectors

are linearly independent, and that ~very solution ofAx = O must be a linear
combination of these three vectors. Therefore the set of vectors 'forro a: basis of
.A'~(A) and v(A)= 3. I

We see fram this example that thenumber of equations that the vectors of
()?, J)­
JIf(A) should obey is equal to p(A) and that there are n components in every
vector of X(A). Therefore n - p(A) components of the vectors of X(A) can Figure 2"6 A composite
SYSTEMS OF LINEAR ALGEBRAIC EQUATlONS 31

n Ax = Ohas k linearly be arbitrarily chosen. Consequently, there are n - peA) linearly independent
vectors in .;V(A). Hence we conclude that n - peA) = veA). We state this as a
1 (ff n , ff), whereas the theorem; its formal proof can be found, for example, in References 43 and 86.

Theorem 2-5
Let A be an m x n matrix. Then
peA) + veA) = n

Corollary 2-5
The number of linearly independent vector solutions ofAx = O is equal to
the last three columns n - peA), where n is the number of columns in A, and peA) is the number of
A. Hence the rank of linearly independent columns in A. •
Then
This corollary follows directly from Theorem 2-5 and the definition of the
null space of A. lt is clear that if peA) = n, then the only solution ofAx = O
x, [=iJ is x = O, which is called the trivial solution. lf peA) < n, then we can always
find a nonzero vector x such that Ax = O. In particular, if A is a square matrix,
then Ax = O has a nontrivial solution if and only if peA) < n, or equivalently,
X,)[i] (2-39)
det (A) = 0, where det stands for the determinant.
We introduce three useful theorems to conclude this section.

Theorem 2-6 (Sylvester's inequality)


:arly independent, we
Id only if Let A, B be q x n and n x p matrices with coefficients in the same field. Then
peA) + p(B) - n.:5: p(AB).:5: min (p(A), p(B))

Proof
fA, peA). The solution
. two equations; hence The composite matrix AB can be considered as two linear transformations
ed. Let X3 = 1, X4=0, applied successively to (ffP, ff) as shown in Figure 2-6. Since the domain
~ 5 = O; then Xl = °
and
"2 = 1. lt is clear that
of AB is ~(B) and the range of AB is a subspace of ~(A), we have immediately
p(AB) .:5:min (p(A), p(B)) by using (2-38). From Figure 2-6 we have p(AB) =

-T---,t-- I ------~--
r I.L --r- --- .
--T-- --­ IR(AB'L R{A)

,iB) R{B)

_í_ L-_'J7(_B)---.J--l-j IL--'J7(_A)~ 1


= O must be a linear
1(

vectors form a basis of



ons that the vectors of
(5P,5)--B-----..- (]",5)
! components in every
e vectors of .;V(A) can Figure 2-6 A composite transformation.
32 l.1NEAR SPACES AND LINEAR OPERATORS

p(B) - d, where d is the dimension of the intersection of ~(B) and .K(A).10 The assumption of p(A
dimension oL/nA) is n - p(A); hence, d S n - p(A). Consequently, p(AB) 2: ACl=[!31 f32
p(B)-n+p(A). Q.E.D.

lf B is an n x n matrix and nonsingular, then and

p(A) + p(B) - n = p(A) S p(AB) smin (p(A), n) = p(A) Hence Cl*A*ACl = O


which, in turn, imp
Hence we have the following important theorem:
conclude that p(A *
2. This part can be Sil
Theorem 2-7
using the fact p(A)
Let A be an m x n matrix. Then
p(AC) = p(A) and p(DA) = p(A)
2-6 Eigenvectol
for any n x n and m x m nonsingular matrices e and D. I Form Represent.

In words, the rank of a matrix will not change after the pre- or postmulti­ With the background
plication of a nonsingular matrix. Because of this property, gaussian elimina­ posed at the end of Sec
tion or the row-searching algorithm discussed in Appendix A can be used to that map (en, iC) into i1
compute the rank of a matrix. to any operator that r
The reason for restrict
Theorem 2-8 immediately.
Let A be an n x n n
Let A be an m x n matrix with coefficients in a field, and let A * be the complex
consider A as a linear
conjugate transpose of A. Then
1. p(A) = n if and only if p(A *A) = n, or equivalently, Definition 2-12
det (A*A) 1=0 Let A be a linear oper
2. p(A) = m if and only if p(AA *) = m, or equivalently, is ca11ed an eigenvalue
Ax = ílx. Anynonzerc
det (AA*) 1=0 associated with the eig

Note that A*A is an n x n matrix and AA* is an m x m matrix. In order to


In order to find an
have p(A) = n, it is necessary to have n s m. This theorem will be proved by
using Definition 2-4'. More specifically, we use the fact that if AFiJ. = ~ implies
Cl = O, then a11 the columns of A are linearly independent, and p(A) = n, where n
where 1 is the unit mat]
is the number of the columns of A.
(2-40) is a set of hom
n x n square matrix. f
Proof
nontrivial solution if a
1. Sufficiency: p(A*A)=n implies p(A)=n. We show that ACl=O implies an eigenvalue oJA if ar
Cl=O under the assumption of p(A*A)=n, where Cl is an n x 1 vector. lf is a polynomial ofdegr
ACl = O, then A*ACl = O; which, with the assumption p(A *A) = n, implies Cl = O. Since ~(íl) is of degree
Hence we conclude that p(A) = n. N ecessity: p(A) = n implies p(A *A) = n. a11 distinct).
Let Cl be an n x 1 vector.·· We show that A*ACl = O implies Cl = O under the·

I lIt is also called a right eige


10 The intersection of two linear spaces is a linear space. y is called a left eigellvecto
REPRESENTATIONS OF A LINEAR OPERATOR 33

~(B)and .;V(A).10 The assumption of p(A) = n. The equality A*Aa. = O implies a.*A*Aa. = O. Let
onsequently, p(AB) ~ Aor:=[.81 .82 ... .8mJ'. Then
Q.E.D. a.*A* = [,B! .8!.,. .8::'t]
and a.*A*Aor: = 1.81\2 +1.821 + ... +1.8mI 2
2

n)= p(A) Hence a.*A*Aa. = Oimplies .8 i = 0, for i = 1, 2, ... , m; or, equivalently, Aa. = O,
which, in turn, implies a. = Ofrom the assumption of p(A) = n. Therefore we
conclude that p(A*A) = n.
2. This part can be similarly proved or directIy deduced from the foregoing by
using the fact p(A) = p(A *). Q.E. D.

<\)
2-6 Eigenvectors, Generalized Eigenvectors, and Jordan­
I Form Representations of a Linear Operator

the pre- or postmulti­ With the background of Section 2-5, we are now ready to study the problem
erty, gaussian elimina­ posed at the end of Section 2-4. We discuss in this section only linear operators
ndix A can be used to that map (en, C) into itself with the understanding that the results are applicable
to any operator that maps a finite-dimensional linear space over Cinto itself.
The reason for restricting the field to the field of comp1ex numbers will be seen
immediately.
Let A be an n x n matrix with coefficients in the field C. We have agreed to
let A * be the complex
consider A as a linear operator that maps (en, C) into (en, C).

Definition 2-12
Let A be a linear operator that maps (en, C) into itself. Then a scalar A in e
is called an eigenva/ue of A if there exists a nonzero vector x in en such that
Ax = AX. Any nonzero vector x satisfying Ax = AX is called an eigenveetor of A
I associated with the eigenvalue A. 11 I

m matrix. In order to In order to find an eigenvalue of A, we write Ax = A.x as


rem wil\ be proved by
t that ir Aa. = (} implies
, and peA) = n, where n where I is the unit matrix of order n. We see that for any fixed), in e, Equation
(2-40) is a set of homogeneous linear equations. The matrix (A - Al) is an
n x n square matrix. From Corollary 2-5, we know that Equation (2-40) has a
nontrivial solution if and only if det (A - Al) = O. lt follows that a sea/ar A is
N that Aa. = O implies an eigenva/ue of A if and on/y if it is a so/ution of ~(A) ~ det (Al - A) = O. ~(A)
is an n x 1 vector. lf is a polynomial of degn~e n in Aand is called the eharaeteristie po/ynomia/ of A.
<\ *A) = n, implies (X = O. Since ~(A) is of degree n, the n x n matrix A has n eigenvalues (not necessarily
=n implies p(A*A)=n. all distinct).
mplies (X = O under the
11 It is also caBed a right eigenveáor of A.. Ifa 1 x n nonzero vector y exists'such thiÚ yA = J.y, then
y is caBed a left eigenveclOr of A associated with J•.
34 LINEAR SPACES AND LINEAR OPERATORS

Example 1 Equation (2-42) implie:


Consider the matrix
(A-.
-1J
-1
(2-41 )
Since
which maps (lR z, IR) into itself. We like to check whether Definition 2-12 can and
be modified and applied to a linear operator that maps (IR", IR) into (IR", IR).
A modified version of Definition 2-12 reads as a scalar }, in IR is an eigenvalue of the left-hand side of (2­
A if there exists a nonzero vector x such that Ax = Ax. Clearly A is an eigenvalue CI.
of A if and only if it is a solution of det (Al - A) = O. Now
By assumption the },¡'s
det(AI-A)=det [ ..1.-1 1 ]
=A z +1
-2 ..1.+1
which has no real-valued solution. Consequently, the matrix A has no eigen­
value in IR. implies Cl. 1 = O. This is
Since the set of real numbers is a part of the field of complex numbers, there is linearly independent
is no reason that we cannot consider the matrix A in (2-41) as a linear operator
that maps (iC Z , iC) into itself. In so doing, then the matrix A has eigenvalues Let A be the repre:
+i and -i where ¡Z ~ -1. • RecaH from Figure 2-5
A¡v ¡ with respect to {v l'
The constant matrices we shaH encounter in this book are aH real-valued. A¡ is located at the ith
{v¡, vz, ... , v,,} is
However in order to ensure the existence of eigenvalues, we shaH consider them
as linear operators under the field of complex numbers.
With these preliminaries, we are ready to introduce a set of basis vectors
such that a linear operator has a diagonal or almost diagonal representation.
We study first the case in which aH the eigenvalues of A are distinct; the case
where A has repeated eigenvalues will then be studied.
This can also be (
Case 1: AII the eigenvalues of A are distinct
Figure 2-5, we have
Let Al' }'z, ... , A" be the eigenvalues of A, and let Vi be an eigenvector of A
associated with A¡, for i = 1,2, ... , n; that is, Av¡ = A¡v i' We shaH use the set of
vectors {v ¡, V z, ... , v,,} as a basis of (e", iC). In order to do so, we have to show Since
that the set is linearly independent and qualifies as a basis.

Theorem 2-9
we have
Let Al' Az, ... , A.. be the distinct eigenvalues of A, and let Vi be an eigenvector
of A associated with A¡, for i = 1, 2, ... , n. Then the set {v l' Vz, ... , v,,} is linearly We conclude that if
independent (over ic). into itself are aH distin
the operator A has a d
Proof the diagonal.
We prove the theorem by contradiction. Suppose VI' V z, ... , v.. are linearly
Example 2
dependent; then there exist CI.¡, Cl.z, ••• ,!X" (not aH zero) in ic such that
Cl.IV I +azv z ... +a"v" =0 (2-42) Consider

We assume Cl. 1 fO. If Cl. 1 =0, we may reorder Ai in such a way that al fO.
REPRESENTATIONS OF A LINEAR OPERA1OR 35

Equation (2-42) implies that

(2-41 )
(A - AzI)(A - A3 1) ... (A - AJ) Ct <XiV) = O (2-43)

Since (A-Ajl)v¡=(A¡-A)V¡ ifj+i


ter Definition 2-12 can and (A - A;I)v¡ = O
)S (~", IR) into (IR", IR).
the left-hand side of (2-43) can be reduced to
in IR is an eigenvalue of
early Ais an eigenvalue <Xl V'l - Az)(A l - A3 ) .•• (Al - AII)V 1= O
)w
By assumption the A;'S, for i = 1,2, ... , n, are all distinct; hence the equation
.z + 1 <Xl n (Al -A¡)V l =0
11

i= Z
natrix A has no eigen­
implies <Xl = O. This is a contradiction. Thus, the set of vectors {v¡, Vz, ... , VII}
omplex numbers, there is linearly independent and qualifies as a basis. Q.E.D.
H) as a linear operator
trix A has eigenvalues Let A be the representation of A with respect to the basis {v¡, Vz, ... , VII}.
I Recall from Figure 2-5 that the ith column of A is the representation of Av¡ =
A¡viwithrespectto{vl,vZ, ... ,vlI}-thatis,[O O Ai O O]',where
oo. o ••

A¡ is located at the ith entry. Hence the representation of A with respect to


.ok are aH real-valued.
{v¡,vz, ... ,v lI } is
we shal1 consider them

: a set of basis vectors O


Az O
O ..
. o·
0 O.]
O
agonal representation. O A3 · o· O (2-44)
o
o.. ..
\. are distinct; the case • • o

O O . o o A II

This can also be checked by using a similarity transformation. From


Figure 2-5, we have
e an eigen vector of A
We shall use the set of Q = [v 1 VZ .. . VII]
jo so, we have to show Since
siso
AQ=A[v l "'z ... VII] = [Av l Av z ... Av n]
= [A1V l AzVz ... AllvlI]=QA
we have A=Q-1AQ
et Vi be an eigenvector
'1' vz,···, VII} is linearly We conclude that if the eigenvalues of a linear operator A that maps (C", C)
into itself are all distinct, then by choosing the set of eigenvectors as a basis,
the operator A has a diagonal matrix representation with the eigenvalues on
the diagonal:
Vz, ... ,V" are linearly
Example 2
C such that
(2-42) Consider

lch a way that <Xl 1-0.


-IJ
-1
36 LINEAR SPACES AND LINEAR OPERATORS

The characteristic polynomial of A is A? + 1. Hence the eigenvalues of A are


+ i and - i. The eigenvector associated with )'1 = i can be obtained by solving Example 4
the following homogeneous equation: Consider

(A-AII)VI=[l~i -~~J[~::J=o

Clearly the vector VI = [1 1 - i]' is a solution. Similarly, the vector v2 =


[1 1 + i]' can be shown to be an eigenvector of A2 = - i. Hence the repre­
The eigenvalues of A al
sentation of Awith respect to {v 1> v2} is
with Al = 1 can be fom
A=[iO O.J -1

The reader is advised to verify this by a similarity transformation. I

Case 2: The eigenvalues of A are not al! distinct Since the matrix (A - A
Unlike the previous case, if an operator A has repeated eigenvalues, it is not Consequently, we can
always possible to find a diagonal matrix representation. We shall use examples VI = [1 O O]', associ
to illustrate the difficulty that may arise for matrices with repeated eigenvalues. A3 = 2 can be found as
sufficient to form abas
Example 3
From this example
Consider
values, it is not alway
Consequently, the A c:
it is possible to find a s
is almost a diagonal fa
eigenvalues of A on thl
The eigenvalues of A are Al = 1, A2 = 1, and A3 = 2. The eigenvectors associated
example, if A has an ~
with Al can be obtained by solving the following homogeneous equations:
with multiplicity 1, ther

(A-A'I)v~[~ ~ -nv~o
forms.
(2-45)

Note that the matrix (A - AII) has rank 1; therefore, two linearly independent
vector solutions can be found for (2-45) (see Corollary 2-5). Clearly, VI =
[1 O O]' and V2 = [O 1 O]' are two linearly independent eigenvectors
associated with Al = A2 = 1. An eigenvector associated with A3 = 2 can be
found as v 3 =[ -1 O 1]'. Since the set of vectors {V I,V 2 ,V 3 } is linearly
independent, it qualifies as a basis. The representation of A with respect to
{v 1> V2, V3} is Al O: O O
O Al 1 :O O
A= [O1 O0J
1 O I
O O Al: O
o---6--o-~-il-:
O
O
O O 2 1 L __

O O O O: A2
In this example, although A has repeated eigenvalues, it can still be diag­
onalized. However, this is not always the case, as can be seen from the following
Which form it will as~
example.
discussed in the next su
¡
'"

REPRESENTATIONS OF A LINEAR OPERATOR 37


le eigenvalues of A are
be obtained by solving Example 4
Consider

J=o
(2-46)

nilarly, the vector V 2 =


, - i. Hence the repre­
The eigenvalues of A are Al = 1, }o2 = 1, and }'3 = 2. The eigenvectors associated
with Al = 1 can be found by solving

O1 2J
(A-A 1 I)v= O O 3 v=O
formation. I [O O 1
Sínce the matrix (A - ..1. 11) has rank 2, the null space of (A - ,111) has dimension 1.
:d eigenvalues, it is not
ConsequentIy, we can find only one linearly independent eigenvector, say
We shall use exampies
VI = [1 O O]', associated with Al =..1. 2 = 1. An eigenvector associated with
h repeated eigenvalues.
. 1. 3 = 2 can be found as V 3 = [5 3 1]'. Clearly the two eigenvectors are not
sufficient to form a basis of (C 3 , C). I

From this example, we see that if an n x n matrix A has repeated eigen­


values, it is not always possible to find n linearly independent eigenvectors.
Consequently, the A cannot be transformed into a diagonal formo However,
it is possible to find a special set of basis vectors so that the new representation
is almost a diagonal form, called aJordan canonical formo The form has the
eigenvectors associated eigenvalues of A on the diagonal and either O or 1 on the superdiagonal. For
example, if A has an eigenvalue Al with multiplicity 4 and an eigenvalue . 1. 2
~eneousequations:
with multiplicity 1, then the new representation will assume one of the following
forms.

{O
(2-45)

linearly independent
Al : O
o--:-iL
-:
1 .1.
__
O
O
¡
O
O
O
O
f
A1 1 : O ,O O
O )'1:L O
______ O , O
l
ry 2-5). Clearly, '"'1= O O:!e l O O' , O O: l¡ 1 ,O ,

l~ ~ :_~---~'-:-tJ
- - - ..! - - -.
lependent eigenvectors O O O: Al' O _ _ _ 1---:­

ed with ,13 = 2 can be O O O O: A2


; {v¡, v2 , v3 } is linearly
m of A with respect to
Al 1 O :0 O )'1 1 O O O

O Al 1 , :0 O
O Al 1 O O
O O Al : O O O O Al 1 ,
O (2-47)
I
O---O--O-~-il-: O
O O O Al : O
L __

O O O
1_ _ _
,
O , ,12 0- - -O -- 0- - -O -:-l;
les, it can still be diag­
seen from the following
Which form it will assunie depends on the characteristics of A and will be
discussed in the next subsection. The matrices in (2-47) are all ofblock-diagonal
38 LINEAR SPACES AND LINEAR OPERATORS

formo The blocks on the diagonal are of the form A. Define

A 1 O O O
O A 1 O O
(2-48)
O O O 1 O
O O O A 1 and
O O O O A Th is set of vectors ['i
of length k.
with the same eigenvalue on the main diagonal and l's on the diagonal just
Let JV¡ denote tb!
aboye the main diagonal. A matrix of this form is caBed al ordan block associ­
that (A - AI)ix = o. lt
ated with A. A matrix is said to be in the lordan canonicalform, or the lordan
is a subspace of .;v¡+ 1
form, if its principal diagonal consists of Jordan blocks and the remaining
tion 2-13 is in .;Vk but
elements are zeros. The fourth matrix in (2-47) has two Jordan blocks associated
defined in (2-49) is in
with Al (one with order 3, the other with order 1) and one Jordan block associ­
ated with A2' A diagonal matrix is clearly a special case ofthe Jordan form: all (A - AI)i1
of its Jordan blocks are of order 1. and (A-AI)i-l,
Every matrix which maps (C n , C) into itself has a Jordan-form representa­
tion. The use of Jordan form is very convenient in developing a number of hence Vi is in .;Vi but r
concepts and results; hence it will be extensively used in the remainder of this Let A be an n x n
chapter. discuss in the followin
vectors of A associatet
alized eigenvectors of
Derivation of a Jordan-form representation. 12 In this subsection, i =0,1,2, ... , until rar
we discuss how to find a set of basis vectors so that the representation of A with notations, weassume n
respect to this set of basis vectors is in aJordan formo The basis vectors to be are as shown in Table :
used are called the generalized eigenvectors. and is equal to, f
.;VO e ';vI e ';v2 e ....
Defihition 2-13 Now we are read)
associated with A. Bet
A vector v is said to be a generalized eigenvector of grade k of A associated with only one linearly inde¡
A if and only if 13
(A -AI)kv = O
and (A -AIt-lv +0

Note that if k = 1, Definition 2-13 reduces to (A - AI)v = Oand v O, which + Table 2-2 Chains of Ge
is the definition of an eigenvector. Hence the term "generalized eigenvector"
is well-justified. p(A - AI)o = 10 Vo =0
Let v be a generalized eigenvector of grade k associated with the eigenvalue p(A -AI)=7 VI =3
p(A-AI)2=4· v2 =6
p(A-AI)3=3 v3 =7
PlA-AI)4",i v4 =8

12 This section may be skipped withollt loss of continllity. However, it is sllggested that the reader
glances throllgh it to gain a better feeling aboLlt the Jordan-form representation.
13 (A - AI)k ~ (A- AI)(A - Al)' .. (A - AI)(k terms), (A _ .U)O ~ 1.
REPRESENTATlONS OF A LINEAR OPERATOR 39

A.. Define
Vk ~ V
Vk- l ~ (A-AI)v=(A-AI)v k
Vk- 2 ~ (A-AI)2v =(A-AI)vk_' (2-49)
(2-48)
and
This set of vectors {V¡, v2, ... , vk} is called a chain of generalized eigenvectors
of length k.
on the diagonal just Let .;Vi denote the null space of (A - AI)i, that is, .;Vi consists of all x such
1 Jordan block associ­
that (A-AI)ix=O. It is clear that ifx is in .;Vi' then it is in ';vi+l' Hence.;Vi
11 form, or the J ordan isasubspaceof .;Vi+"denotedas ';viC .;Vi +l' Cleariy,thevdefinedin Defini­
s and the remaining tion 2-13 is in ';vk but nbt in .;v k-l' In fact, for i = 1, 2, ... , k, Vi = (A - AI)k-i v
dan blocks associated
defined in (2-49) is in .;Vi but not in ';vi-l' Indeed, we have
Jordan block associ­
fthe Jordan form: all (A - AI)ivi = (A -AlhA -AI)k-i v = (A - AI)kv =0
and (A -AI)i-' vi = (A - AI)i-'(A -AV-iV = (A - AI)k-' v 1=0
jan-form representa­
e10ping a number of hence Vi is in .;Vi but not in ';vi-l'
the remainder of this Let A be an n x n matrix and have eigenvalue A with multiplicity m. We
discuss in the following how to find m linearly independent generalized eigen­

vectors of A associated with A.. This is achieved by searching chains of gener­

alized eigenvectors of various lengths. First we compute ranks of (A - AI)i,

In this subsection, i = O, 1,2, ... , until rank (A - AI)k = n-m. In order not to be overwhelmed by

,resentation of A with . notations, weassume n = 10, m = 8, k =4,and the ranks of(A -AW, i =0,1,2,3,4,

le basis vectors to be are as shown in Table 2-2. The nullity Vi is the dimension ofthe null space ';vi'

and is equal to, following Theorem 2-5, n - peA -AI)i. Because of

fi o C .;v 1 C .;v 2 C •.• , we have O = Vo :-:;; V1 :>: V2 :-:;; . •• :-:;; Vk = 111.


Now we are ready to find m = 8 linearly independent eigenvectors of A
associated with A. Because of .;v 3 C .;v4 and V4 - V~ = 1, we can find one and
, of A associated with only one linearly independent vector u in .;v 4 but not in .;v 3 such that
B 4 u =0 and B 3 u 1=0
where El ~ A -,ll From this Illi, we can generate a chHin al faur generalized

= O and v 1= O, which Table 2-2 Chains of Generalized Eigenvectors, where B ~ A - Al
eralized eigenvector"
No. of independent
p(A ­ ,1.1)0 = 10 1'0 =0
vectors in Xi but
j with the eigenvalue p(A -,1,1)=7 VI =3
not in Ñ¡_I
p(A _,1,1)2 =4 1'2=6
. w2 =w. WI =Bw TWO chains with
p(A _,1,1)' =3 V, =7 }
v2 =v, v, =Bv length 2
p(A _,1,1)4 =2 1'4=8
One chain with

I )(2--~-
length 4

s suggested that the reader


esentaÚon.

L
40 LINEAR SPACES AND LINEAR OPERATORS

eigenvectors as which implies, because (


(2-50) FinaHy, we have C1 =Cs
This completes the prol
Because of X z e X 3 and V3 - Vz = 1, there is only one linearly independent
vector in X 3 but not in X z. The u 3 in (2-50) is such a vector; therefore we Theorem 2-11
cannot find any other linearly independent vector in X 3 but not in X z. Con­
sider now the vectors in X z but not in Xl' Because Vz -V1 =3, there are The generalized eigem
three such linearly independcnt vectors. Since Uz in (2-50) is one of them, we linearly independent.
can find two vectors v and w such that {u z, v, w} are linearly independent and
This theorem can t
BvjoO (A-A.il)k(A-A.jlt Th
and BwjoO Now we discuss thf
u4 v 1 V 2 W 1 W Z ­
From v and w, we can generate two chains of generalized eigenvectors oflength associated with other ei,
2 as shown in Table 2-2. As can be seen from Table 2-2, the number of vectors tion A are the represent
in Xl is equal to v 1 - Vo = 3, hence there is no need to search other vector in v1 , V z, w1 , wz,x,x}. Bec
X l ' This completes the search of eight generalized eigenvectors of A associated (A - A.I)u 4 = u 3 , we ha"
with ..1.. Q[l A. O O ... O
AU 4 =U 3 +A.U 4 = Q[O
transpose. Proceeding
Theorem 2-10
The generalized eigenvectors of A associated with A. generated as in Table 2-2
are linearly independent.

Proof
First we show that if {u z, v, w} is linearly independent, then {u¡, v1. W1} is linearly
independent. Suppose {u¡, v1 , w¡} is not linearly independent, then there
exist ci,i=1,2,3, not aH zero, such that C1U1 +C Zv1 +C3W1 =0. However,
we have

This is a Jordan-form m;
Since y is a vector in X z, the only way ta have Ey = {D is that y = Gi. Siúce {;nz, '1, w}
is linearly independent by assumption, y = O implies Ci = O, i = 1,2,3. This
is a contradiction. Hence if {u z, v, w} is linearly independent, so is {u¡, Vl' w¡}. Example 4 (Continued
Now we show that the generalized eigenvectors {u i , i = 1, 2,3,4; vj' Wj,j = 1, Consider
2} are linearly independent. Consider

(2-51 )

The application of B 3 = (A - A.lf to (2- 51) yields

which implies; because of B3 U4 jo O, C4 = O. Similarly, we can show c 3 =0 by 14 rhis number is called the 9
multlplicity o¡os the number
applying B2 to (2-51). With C3 = C4 = O, the application of B to (2-51) yields
orders of aH lordan blocks
REPRESENTATIONS OF A LINEAR OPERATOR 41

which implies, because ofthe linear independenceof {uzo Vz, Wz}, Cz = C6 = Cs =0.
(2-50) Finally, we have C1 =C s =C7 =0 following the linear independence of {U 1, '1 1, w 1}.
This completes the proof of this theorem. Q.E. D.
;: linearly independent
a vector; therefore we Theorem 2 -11
; but not in j{ z. Con­
: Vz - v1 = 3, there are The generalized eigenvectors of A associated with different eigenvalues are
·50) is one of them, we linearly independent. I
early independent and
This theorem can be proved as in Theorem 2-10 by applying repetitively
(A - A¡I)k(A - Ajlf The proof is left as an exercise.
Now we discuss the representation of A with respect to Q ~ [u 1 Uz u 3
u 4 V1 VZ W1 WZ x x]. The last two vectors are the eigenvectors of A
1eigenvectors oflengfh assoqated with other eigenvalues. The first four columns ofthe new representa­
, the number of vectors tion A are the representations of Au¡, i = 1, 2, 3, 4, with respect to {u¡, u z, u 3 , u 4 ,
search other vector in Vlo Vz, W1, Wz, X, x}. Because (A - AI)u1 = O, (A -AI)1z = u 1, (A - AI)u 3 =u z, and
.vectors of A associated (A-AI)u 4 =u 3, we have AU1=AU1=Q[A O O ... 0J',Au Z =u 1 +Au z =
Q[l A O O ... O]', AU 3 =U z +Au 3 =Q[O 1 A O ... OJ', and
AU 4 =u 3 +Au 4 =Q[O O 1 A O ... OJ', where the" prime denotes the
transpose. Proceeding similarly, the new representation A can be obtained as

A 1 O 0:0 O O O

lerated as in Table 2-2


O A 1 0:0 O
O
O O A 1:O
O O O ,,1.:0
O O O
O O O

°
A= O--Ó--Ó-O-:-~--(: O

, , O (2-52)
O O O 0:0 A: O
fu
m lo v lo W¡} is linearly
dependent, then there O O O O - 0- -6: ),- - i -:,
1
°
+ C3W1 =0. However, O O O O O 0:0 , A: _ _ _ _ _ _1

This is a Jordan-form matrix. Notethat the number of Jordan blocks associated


at y =O.Since {uz, v, w} with ). is equal to v 1 = 3, the dimension af the null space Ol (A. - U).14.
c¡=O,i=1,2,3. This
ldent, so is {U l , '1 1, w¡}. Example 4 (Continued)
= 1, 2, 3,4; Vj, Wj,j = 1,
Consider

L +cswz=O (2-51 )

14 This number is caBed the geometric multiplicity of ), in Reference 86. In oLher words, geometric
VoIe can show C3 = O by
multiplicity is the number of lordan block S, and the (algebraic) multiplicity is the sum of the
of B to (2-51) yields orders of aH lordan blocks associated with .l..

l.
42 LINEAR SPACES AND LINEAR OPERATORS

Example 5
Its eigenvalues are Al = 1, A2 = 1, A3 = 2. An eigenvector associated with
),3 = 2 is v3 = [5 3 1J'. The rank of (A - Al1) is 2; hence we can find only one Transform the followi
eigenvector associated with Al' Consequently, we must use generalized eigen­
vectors. We compute

B@(A-A'l)~[~
1 A
O
O :]

(A~A,Il'~[~ m~
:] ~[~ ¡]
1 1 O
O O O 1. Compute the eigen'
and
O O O det (A -AI)=

Since pB 2 = 1 = n-m, we stop here. We search a v such that B 2 v = O and Hence A has eigen'
Bv =/= O. Clearly, v = [O 1 O]' is such a vector. It is a generalized eigenvector plicity 1.
of grade 2. Let 2. Compute (A - 21Y,

B f:::,. (A -21) =

Theorems 2-10 and 2-11 imply that v1 , v2 , and v 3 are linearly independent.
This can also be checked by computing the determinant of [v 1 V2 v 3]. If
we use the set of vectors {v 1, V2' V3} as a basis, then the ith col umn of the new
representation A is the representation of Av; with respect to the basis {v 1 , v 2 , v3 }.
Since AV 1 =A 1V¡,Av 2 =v 1 +A 1V2 , and AV 3 =A 3 V3 , the representations of (A-21)2 =
Av l' Av 2 , and AV 3 with respect to the basis {v 1 , v2 , v 3 } are, respectively,

r:
where Xl = 1, }'3 =2. Hence we have
(A-21)' ~l

A=[~O 0:2
~-~-~J
__
(2-53)
Since peA - 21)3 = n
a generalized eigen'
This can also be obtained by usihg the similarity transformation

A=Q- 1 AQ
15 We use the ract that

~[i ~]
O
where Q =[v 1 v2 v,] 1
where A and e are square
O I
REPRESENTATlONS OF A LINEAR OPERATOR 43

vector associated with Example 5


ce we can find only one Transform the following matrix into the lordan form:
t use generalized eigen­
3 -1 1 1 O O
1 1 -1 -1 O O
O O 2 O 1 1
A= (2-54 )
O O O 2 -1 -1
O O O O 1 1
O O O O 1 1

, O5]
O 3 1. Compute the eigenvalues of AY
O 1 det (A -Al) = [(3 -..1.)(1-..1.) + 1](..1. -2)2[(1 _..1.)2 -1] =(..1. _2)5 A
such that B 2 v =0 and Hence A has eigenvalue 2 with multiplicity 5 and eigenvalue O with multi­
~eneralized eigenvector plicity 1.
2. Compute (A - 21Y, for i = 1, 2, ... , as follows:
1 -1 1 1 O O
1 -1 -1 -1 O O
O O O O 1 1 p(A-21)=4
B ~ (A-21)=
O O O O -1 -1 v1 =6-4=2
O O O O -1 1
• linearly independent. O O O O 1 -1
at of [v 1 V 2 v3 ]. Ir
ith column of the new O O 2 2 O O
to the basis {v 1 , v2 , v3 }. O O 2 2 O O
le representations of O O O O O O p(A-2If =2
(A -21)2 =
re, respectively, O O O O O O V2 =4
O O O O 2 -2
O O O O -2 2

r~ ~l

O O O O

O O O O

~J
n

~ l~
v O O O p(A -21iY = 1
(A-2Ij' O O O O v3 =5
O O O -4
(2-53) O O O 4 -4
Since p(A - 21)3 = n-m = 1, we stop here. Because v3 - V2 = 1, we can find
a generalized eigenvector u of grade 3 such that B3u ;= O and B 2 u 1= O. It is
rmation

[5 We use the ract that

de{: ~J = det A det e


where A and e are square matrices, not necessarily or the same order.

44 LINEAR SPACES AND LINEAR OPERATORS

easy to verify that u = [O O 1 O O O] is such a vector. Define where

2 1 O
2 -1 O
~
O O 1
UI BZu= U z ~ Bu = u 3 ~ u=
O O O
O O O
O O O
This is a chain of generalized eigenvectors oflength 3. Because of Vz - VI = 2, In this example,
there are two linearIy independent vectors in .K z but not in .K l. The vector {w, V z, V¡, U 3, U z, u¡};
U z is one of them. We search a vector v which is independent of U z and has
the property BZv = O and Bv i= O. It can be readily verified that v =
[O O 1 -1 1 1]' is such a vector. Define

O O
O O
2 1
VI ~ Bv=
-2 -1
O 1 This is also called a J,
O 1 (2-55), we see that thl
below the main diago
Now we have found five generalized eigenvectors of A associated with A. = 2. In this book, we use
3. Compute an eigenvector associated with A. z =0. Let w be an eigenvector everything discussed
of A associated with A. z =0; then given in Equation (2-:
A Jordan-form re
3 -1 1 1 O O into itself is unique u¡:
1 1 -1 -1 O O Jordan blocks and ti
O O 2 O 1 1 w=O by A. However, bec~
(A ­ A.zI)w = different Jordan-form
O O O 2 -1 -1
O O O O 1 1
O O O O 1 1
2-7 Functions (
Clearly, W = [O O O O -1]' is a solution
4. With respect to the basis {u I , u z, u 3 , VI, vz, w}, A has the following Jordan­ In this section we shall
form representation: tion that maps (en, 1[:)
extensively, because il
2 1 0:0 O O function of a matrix c:
021:0 O O
matrix, and then defil
O O 2:0 O O
matrix.
A= 0- -6 - -6 ~ 2- -¡ -: O (2-55)
O O 0:0 2:0
O O O - O--6 - 0­ Polynomials of a s
(C", C) into (1[:",1[:). If
5. This may be checked by using

A=Q-¡AQ or and
FUNCTlONS OF A SQUARE MATRIX 45

a vector. Define where Q=[U 1 Uz u3 v1 Vz w]


, ,
2 , 1 , O' O: O: O
O ,
2 : -1 , O , O', O: O
O , ,
O , O 1 , 2: 1 O
~ u= 1 O
,
O , O : -2 :, -1 O
O
O : O O O, 1 1
O
O O O O 1 -1 I
O

Becauseofvz -V 1 =2, In this example, if we reorder the basis {Ul> u z, u3, v1, V Z, w} and use
10t in j( l' The vector {w, v z, Vl' U 3 , U z, U 1 } as a new basis, then the representation will be
:pendent of U z and has
lily verified that v =
efine

O A= (2-56)
O
1
1
1 This is also called a lordan-form representation. Comparing it with Equation
1 (2-55), we see that the new lordan block in (2-56) has l's on the diagonal just
below the main diagonal as a result ofthe different ordering ofthe basis vectors.
l associated with A. = 2. In this book, we use mostly the lordan block of the forro in (2-55). Certainly
:t w be an eigenvector everything discussed for this form can be modified and be applied to the form
given in Equation (2-56).
A lordan-forro representation of any linear operator A that maps (en, C)
o into itselfis unique up to the ordering of lordan blocks. That is, the number of
O lordan blocks and the order of each lordan block are uniquely determined
1 by A. However, because of different orderings of basis vectors, we may have
w=o different lordan-form representations of the same matrix.
-1
1
1
2-7 Functions of a Square Matrix
ution.

the following lordan­ In this section we shall study functions ofa square matrix or a linear transforma­
tion that maps (en, C) into itself. We shall use the lordan-form representation
extensively, because in terms of this representation almost all properties of a
function of a matrix can be visualized. We study first polynomials of a square
matrix, and then define functions of a matrix in terms of polynomials of the
matrix.
(2-55 )

Polynomials of a square matrix. Let A be.a square matrix that maps


(en, C) into (C n , C). Ifk is a positive integer, we define
Ak.~AA···A (k terms) .' (2:57a)
and AO ~ 1 (2-57b)
46 LINEAR SPACES AND LINEAR OPERATORS

where 1 is a unit matrix. Let f(A) be a polynomial in A of finite degree; then polynomial of A is
f(A) can be defined in terms of(2-57). For example, if f(A) = A3 +2A z +6, then
f(A) ~ A 3 +2A z +61
We have shown in the preceding section that every square matrix A that Assume that aJorda:
maps (1[", C) into itself has a Jordan-form representation, or equivalently, there
exists a nonsingular constant matrix Q such that A = QAQ - I with A in a
Jordan canonical formo Since

where the n¡ x n¡ mat


we have

f(A) = Qf(A)Q-I or f(A) = Q-l f(A)Q (2-58) Definition 2-15


The largest order of
for any polynomial f(A).

index of A¡ in A.
One of the reasons to use the Jordan-form matrix is that if

The multiplicity (
(2-59) the matrix in (2-52),
nz = 1; for the matr
where Al and A z are square matrices, then n¡'::;; ni.

f(A) = [f(OAd O ] (2-60) Theorem 2-12


f(A z )
The minimal polynol
This can be easily vetified by observing that

A=
k [A1O
where ñ¡ is the index
Definition 2-14
Proof
The minimal polynomial of a matrix A is the monic polynomial 16 if¡(A) of least
degree such that if¡(A) =0. I
Since the matrices A
as showing that tf¡(A)
'"f/e -6.rst shovv that tb
Note that the O in if¡(A) =0 is an n x n square matrix whose entries are all
A¡ consists of r Jorda
zero. A direct consequence of(2-58) is that f(A)=O ifand only if f(A)=O.
Consequently, the matrices A and A have the same minimal polynomial, or A¡=di
more general1y, similar matrices have the same minimal polynomial. Computing
the minimal polynomial of a matrix is general1y not a simple job; however, if
the Jordan-form representation of the matrix is available, its minimal poly­
nomialcan be readily foundo
Let Al, Az, . oo,A m be the distirict eigenvalues of A with multiplicities
nI> nz, ooo,n m , respectivelyo It is .the same as saying that the characteristic

. 16.\ monic polynomial is. a polynomial the coefficient of whose highest power is 1. For example,
3x +1 aTid -x 2 +2x +4 are not monicpolynomials, but x 2 -4x +7 is.
FUNCTIONS OF A SQUARE MATRIX 47

of finite degree; then polynomial of A is


A)=A 3 +2A z +6,then m

~(A) ~ det (Al - A) =


¡= 1
n (A - AJ"¡ (2-61 )

square matrix A that Assume that a Jordan-form represemation of A is


or equivalently, there
QAQ-I with A in a
.(2-62 )

where the n¡ x n¡ matrix A¡ denotes aB the Jordan blocks associated with A¡.

f(A)Q (2-58) Definition 2-15


The largest order of the Jordan blocks associated with A¡ in A· is caBed the
index of A¡ in A. 11
hat if
The multiplicity of A¡ is denoted by ni, the index of Ai is denoted by ni' For
(2-59) the matrix in (2-52), nl=8,nl=4; for the matrix in (2-53), nl=n l =2,nz=
n z = 1; for the matrix in (2-55), ni = 5, nI = 3, nz = nz = 1. lt is clear that
n¡sn¡.

(2-60) Theorem 2-12


The minimal.polynomial of A is

n (A - A¡)ñ¡
m
lj¡(A) =
¡= 1

where ni is the index of A¡ in A.

Proof
nomiaP 6 lj¡(A) of least
Since the matrices A and A have the same minimal polynomial, it is the same
• as showing that lj¡(A) is the polynomial with least degree such that lj¡(A) = o.
We first show that the minimal polynomial of A¡ is ¡J¡¡()c) = (.>1-- AS'. Suppos\'::
. whose entries are aH
and only if feA) = O.
A¡ consists of r Jordan blocks associated with A¡. Then
nimal polynomial, or A¡ = diag (A il , A¡z, ... ,A¡r)
lynomial. Computing
imple job; however, if
:>le, its minimal poly- and

A with multiplícities
hat the characteristic
(Ail - A¡lt' A O _
.= .. ?. (Aiz ~ A;I)"'
.. (2-63)
t power is 1. For example, .
, is.
f. O O
48 LINEAR SPACES ANO LINEAR OPERATORS

Ifthe matrix (Aij - A;I) has dimension n¡j, then we have they have, respecth

=[~ ¿ ~
... O]
... O
minimal polynomiah

(Aij - A¡I) (2-64a) Because the chafé


(nij x ni) O O O ... 1 by the minimal polyr
O O O O Theorem 2-12.
O O 1 O O Corollary2-12 (C,
O O O 1 O
Let i1(A) ~ det (J,l - .
(Aij-A;I?= O O O O (2-64b)
1 polynomial of A. Th(
O O O O O
~(A
O O O O O

[~ ~ ~ ~
The Cayley-Ham
Theorem 2-12 (see Pr
(A¡j - A¡I)n¡j-¡ = °oll (2-64c)
The reason for in1
O O O O J in the following theoJ
and (Aij - AJ)k = O for any integer k ;::: nij (2-64d)
Theorem 2-13
By definition, ñ¡ is the largest order of the Jordan blocks in A¡, or equivalently,
Let A¡, Az, . .. , Am be
ñ¡=max(nij,j=1,2, ... ,r). Hence (Aij-A¡I)ñ,=O for j=1,2, ... ,r. Con­
Let f and 9 be two pe
sequently, 1jJ¡(A;) =0. It is easy to see from (2-63) and (2-64) that ifljJ¡(A)=(A-o:)k
with either o: I=A¡ or k < ñ¡, then ljJ ¡(A;) 1=0. Hence we conclude that 1jJ¡ = (A - A¡)ñ¡ 1. feA) = g(A).
is the minimal polynomial o'f Aj . Now we claim that feA;) = Oif and only if f 2. Either f =h¡1jJ +f
is divisible without remainder by ljJ ¡, denoted as ljJ d.r Inde:ed, if ljJ ;\f; then A, and h¡ and h 2 a
f can be written as f = ljJ eh, where h is the quotient polynomial, and feA;) = 3.
1jJ¡(A;)h(A;) = O' h(A;) = O. If f is not divisible without remainder by ljJ ¡, then pe)(A¡) = g('
lcan be written as f = ljJ¡h +g where gis a polynomial of degree less than ñ¡.
Now f(A¡) = O implies g(A¡) = O. This contradicts the assumption that ljJ ¡ is
where j(l)(A;) ~ d~
{
the minimal polynomial of Á¡, for 9 is a polynomial of degree less than that of Proof
ljJ i and g( A¡) = O. With these preliminaries, the theorem can be readily proved.
From (2-62), we have IjJ(Á) = diag (IjJ(Á¡) IjJ(Á 2 ) ••• IjJ(Árn )). Since IjJ(Á¡) = O The equivalence of sta
i[ and only ifljJ eontains the [actor (il- ),¡)ñ;, we conclude that the mínima1 poly­ Statements 2 and 3 ar
nomial of Á and, correspondingly, of A is . .
rn
TI (A _A¡)ñ¡ Q.E.D.
i= 1 In order to apply 1
Example 1 The minimal polynol
form or by direct com
The matrices Therefore it is desirab

r~ ~ i ~J r~ ~ i ~J r~ ~ I~J

polynomial can be av

Corollary 2-13
Let the charaeteristic
al! have the same characteristic polynomial i1(A) = (A - 3)3(A -1); however,
-_._--_....::_..__ .:..~.~,----------"---~----_.:..._=--.=--.....:....-~=--_:...._-_:...-::~_.,---...:_-,--,~_:._-----

FUNCTlONS OF A SQUARE MATRIX 49

they have, respectively, (A - 3)(A - 1), (A - 3f(A -1), and (A - W(A -1) as
minimal polynomials. I

(2-64a) Because the characteristic polynomial is always divisible without remainder


by the minimal polynomial, we have the following very important corollary of
Theorem 2-120

Corollary 2-12 (Cayley- Hamilton theorem)


Let Ll(A)~det(AI-A)~An+txl;,n-l +'0' +txn-IA+tx ll be the characteristic
(2-64b) polynomial of A. Then
Ll(A)=An+txIAn-l +00. +txll_IA+txJ=O I

The Cayley-Hamilton theorem can also be proved directly without using


Theorem 2-12 (see Problems 2-39 and 2-40).
(2-64c) The reason for introducing the concept of minimal polynomial will be seen
:J in the following theoremo

k 2: nij (2-64d) Theorem 2-13


in A¡, or equivalently, Let Al, A2"'" Am be the distinct eigenvalues of A with indices nI' ii 2, ... , iimo
j=1,2, ... ,r. Con­ Let f and 9 be two polynomialso Then the following statements are equivalent.
.) that ifi/J¡(A)=(A-txt
lude that lj; ¡ = (A - A¡)n¡ 1. feA) = g(A).
~¡) = O if and only if f 2. Either f =hllj; +g or 9 =hi.lj; + f, where lj; is the minimal polynomial of
Indced, if lj;dI, then
3.
A, and h 1 and h 2 are sorne polynomials.
ynomial, and f(A¡) =
f([)(A¡) = gO)(A;) for 1= 0, 1,2, ooo, n¡ - 1; i = 1,2, o.. , m (2-65 )
:emainder by lj; ¡, then
of degree less than ni'
where f
(1)
(Ad
6.
=
d1(A)i
~ A=A¡ and
1
g( )(A¡) is similarly defined.
o

issumption that 0lj;¡ is


¡:gree less than that of Proof
~n be readily proved.
Theequivalence ofstatements 1 and 2 follows directly from the fact that lj;(A) = O.
(Am))o Since lj;(A¡) = O
Statements 2 and 3 are equivalent following
that the minimal poly-
lj;(A) = n (A - A¡)ñ¡
m

i= 1
Q.EoD.
Q.E.D.
In_order to apply this theorem, we must know the minimal polynomial of A.
The minimal polynomial can be obtained by transforming A into aJordan
form or by direct computation (Problem 2-42)0 Both methods are complicated.

;~ ~l
Therefore it is desirable to modify Theorem 2-13 so that the use of the minimal
polynomial can be avoided.

) 3
) O 1
° Corollary 2-13
Let the characteristic polynomial of A be

n (A - A¡)n
m
- W(A -1); however, Ll(A) ~ det (Al - A) = i

i= 1
SO LINEAR SPACES AND LINEAR OPERATORS

Let f and 9 be two arbitrary polynomials. If Now, froro Corollary


On the spectrum of A
¡<l)(A¡) = g(l)(A;) for l = 0, 1,2, , ni - 1
i= 1, 2, ,m (2-66)

then f(A)=g(A). I
Solving these two eqt
This follows immediately from Theorem 2-13 by observing that the condition
(2-66) implies (2-65). The set of numbers ¡<L)(A;), for i = 1,2, ... ,m and l = A 100 =g(A) =c
0, 1,2, ... , ni -1 (there are totally n = I7= 1 n¡) are called the values of f on
the spectrum of A. Corollary 2-13 implies that any two polynomials that have
Obviously A 100 can a
the same values on the spectrum of A define the same matrix function. To
different g(A) such as 1
state it in a different way: Given n numbers, if we can construct a polynomial
which gives these numbers on the spectrum of A, then this polynomial defines
uniquely a matrix-valued function of A. It is well known that given any n Functions of a Sql
numbers, it is possible to find a polynomial g(A) of degree n - 1 that gives these
n numbers at sorne preassigned A. Hence if A is of order n, for any polynomial Definition 2-16
f(A), we can construct a polynomial of degree n - 1,
Let f(A) be a functie
g(A)=a o +a 1A + oo. +a n_ 1An- 1 (2-67) spectrum of A. If g(A
spectrum of A, then th,
such that g(A) = f(A) on the spectrum of A. Hence any polynomial of A can
be expressed as
This definition is :
f(A) =g(A) =aol +a 1 A + ... +a n_ 1An-1 be precise, functions e
This fact can also be deduced directly from Corollary 2-12 (Problem 2-38). (2-65). The condition
Corollary 2-13 is useful in computing any polynomial and, as will be dis­ is easier to obtain the
cussed, any function of A. lf A is of order n, the polynomial g(A) can be chosen willlead to the same J
as in (2-67) or as any polynomial of degree n - 1 with n independent parameters. If A is an n x n me
For example, if all eigenvalues, Ai' i = 1,2, ... , n, of A are distinct, then g(A) can find a polynomial
can be chosen as
n -1 n

g(A)= I
i=O
Pi n (A-Aj)
j= 1
which is equal to f(A
know that every funct
Ni
n-l i
or g(A) = I Pi f1 (A - Aj ) We summarize th,
i=O j= 1
an n x n matrix A al
In conclusion, the form of g(A) can be chosen to facilitate the computation.
polynomial of A, say
Example 2
Compute A 100, where
Let
A=[~ ~J

l
In other words, given f(A) = A 00, compute f(A). The characteristic poly­ wheré a o, al'" ., ex n - 1
nomial bf A is ~(A) = det (Al - A) = (A - 1)Z. Let g(A) be a polynomial of degree to compute these ex/s
n -1 = 1, say we have f(A)=g(A).
n independent paramé·
FUNCTIONS OF A SQUARE MATRIX 51

Now, from Corollary 2-13, if f(A.) = g(A.) on the spectrum of A, then f(A) = g(A).
On the spectrum of A, we have
,-1
(2-66) f(l) = g(l) (1)100 =exO +ex 1
100'(1)99=ex 1
• 1'(1) = g'(l)
Solving these two equations, we obtain ex 1 = 100 and exo = -99. Hence
ting thatthe condition
= 1, 2, ... , m and 1=
ed the values of f on
1 0J
AI00=g(A)=exoI+exlA=-99 [ O 1 +100 O 1 = O 200J
1 [1 2J [1
Jolynomials that have
Obviously A 100 can also be obtained by multiplying A 100 times or by using a
matrix function. To
different g(A.) such as g(A.) = exo +ex 1(A. -1) (Problem 2-33). •
mstruct a polynomial
ris polynomial defines
)wn that given any n Functions of a square matrix
: n - 1 that gives these
n, for any polynomial Definition 2-16
Let f(A.) be a function (not necessarily a polynomial) that is defined on the
(2-67) spectrum of A. If g(A.) is a polynomial that has the same values as f(A.) on the
spectrum ofA, then the matrix-valued function f(A) is definedas f(A)~g(A). •
polynomial of A can

This definition is an extension of Corollary 2-13 to include functions. To


n-l
be precise, functions of a matrix should be defined by using the conditions in
.2 (Problem 2-38). (2-65). The conditions in (2-66) are used because the characteristic polynomial
al and, as will be dis­ is easier Oto ohtain than the minimal polynomial. Of course, both conditions
rial g(A.) can be chosen will lead to the same resulto
dependent parameters. If A is an n x n matrix, given the n values of f(A.) on the spectrum of A, we
He distinct, then g(A.) can find a polynomial of degree n - 1,
g(A.) = exo +ex 1A. + ... +ex n_12n-l
which is equal to f(2) on the spectrum of A. Hence from this definition we
know that every function of A can be expressed as
f(A) =exol +ex 1A + ... +ex n _¡An-l
We summarize the procedure of computing a function of a matrix: Given
an n x n matrix A and a function f(A.), we first compute the characteristic
:e the computation.
polynomial of A, say

n (A. - A.Jn¡
m
.1\(2) =
i= 1

Let
g(2)=exo +ex 1A. + ... +ex n_ 12 n- 1
e characteristic poly­ where exo, ex 1, ... ,ex n-1 are n unknowns. Next we use the n equations in (2-66)
polynomial of degree to compute these ex/s in terms of the values of f on the spectruID of A. Then
we have f(A) = g(A). We note that other polynomial g(A.) of degree n-1 with
n independent parameters cari also be used.
52 LINEAR SPACES AND LINEAR OPERATORS

Example 3 ofthe form


Let g(A) = (;(0 +

A, ~ [~ ! -~] Then the conditions i

Compute eA". Or equivalently, if feA) = eM, what is f(A¡}? Hence,


The characteristic polynomial of Al is (A -lf(A - 2). Let g(A) = Cio +CilA +
2
1X2/" Then
f(l) = g(l) é =(,(0 +(,(1 +(,(2
f'(I)=g'(I) tel =Ci l +2(,(2 (note that the derivative is with respect
to A, not t)
f(2) = g(2) e21 = (,(0 + 2(,( 1 +4(,(2

Solving these equations, we obtain (,(0 = -2te' +e21'(,(1 =3tel +2el -2e 21 ,
and (,(2 = e21 - el - tel. Hence, we have Here in the last step v
If feA) = eJ..I, then
eA" = g(A ¡) = ( - 2te' +e 21 )I +(3tel + 2e' - 2e 21 )A 1 +(e21 - e' - tel)A ~

=
2e' - e
O
21
O
el
2é - 2e
O

2
']

e AI = re~'1 t;
[ _el +e21 O 2e 21 _el I
lo (
Example 4 Note that the derivati
Let A function of a n
therefore, the relations

A'~[~ j -~]
of a matrix. F or exan

lts characteristic polynomial is ~(A) = (A - 1)2(A - 2), which is the same as the and if
one of Al in Example 3. Hence we have the same g(A) as in Example 3. Con­
sequently, we have
2e' -e2' then
e A21 = g(A 2 ) = O
[ 21 for any function f tt
e -é I
(2-71), any function o
mediately.
Example 5
Example 6
Given
Consider
1 O

~ ~~
[ A, Al 1
(2-68)
(n n) O
O
O
O
1.]
The characteristic polynomial of Á is (A - Alr. Let the polynomial g(A) be
FUNcnoNs OF A SQUARE MATRIX 53

of the form
g(A) = Cto +Ct 1(A - A¡} +Ctz(A - Ad + ... +Ct n-1(A - A¡)"-l
Then the conditions in (2-66) give immediately
, f(n-1)(A¡}
Ct O =f(A 1), Ct 1=f(A¡), ... , Ctn- 1= (n-l)!
<\¡)?
Hence,
. Leí g(A) = Cto +Ct 1A +
f '(A ) f(lI- 1)(' )
f(A) =g(A) = f(AI)1 +---yf- (A - A11) + ... + (n -1;t (A -A 11)n-1

f(A1) f'(A 1)/1! f"(A 1)/2! ¡<n-1)(A 1)/(n _1)']


erivative is with respect O f(A 1) f'(A 1)/1! f(n-Z)(A¡)/(n-2)!
=?. .
? f(¿1) f(n-3)(A~)/(n-3)! (2-69)
[ . . .
. ..
O O O f(A¡)
Here in the last step we have used (2-64).
If f(A) = e AI , then
e AIl teA,l t Ze AIl/2! ... t n- 1e A,I/(n-l)!l


Al _

l
e - ..
.
O
O eA,1

O
..
.
..
O
.
...
te Al ... t n-ZeA,l/(n - 2)!

eA,1
..
.

Note that the derivatives in (2-69) are taken with respect to Al' not to t.
(2-70)

11

A function of a matrix is defined through a polynomial of the matrix;


therefore, the relations that hold for p'olynomials can also be applied to functions
of a matrix. F or example, if A = QAQ -1, then
f(A) = Qf(A)Q-1

¡lich is the same as the


15 in Example 3. Con­
and if A [AlO= O]
A2

then [(A) = [f(:¡) f(~2)J (2~1'í1 ~


for any function f that is defined on the spectrum of A. Using (2-69) and
• (2-71), any function of a Jordan-canonical-form matrix can be obtained im­
mediately.

Example 6
Consider

Al 1 O: O O
(2-68) OA 1 '1 : O O
A=O O A1 :O O (2-72)
-------~-~----- ­
O O O :A z 1

O O O: O A2

he polynomial g(A) be
54 LINEAR SPACES AND LINEAR OPERATORS

lf feA) = eAt , then If the absolute values


be shown that the in
Instead of provin:
matrix [unction, we ~
f(A)=e At = (2-73) (2-69).

Example 7
lf feA) = (s - Af 1, where s is a complex variable, then
Consider the Jordan-
feA) = (si - A)-l
1 1 1
O O
S-Al (S-A l )2 (S-A l )3:
then
1 1 ,
O O O
S -Al (S-A l )2:
,
1 ,
O O , O O (2-74 )
,
s -Al ,
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ .l. _ _ _ _ _ _ _ _ _ _ _ _ _
Since (A -AlI)i is of
immediately to (2-69).
1
O O O
: s -A 2 (s -A2)2
Example 8
1
O O O O The exponential funct
s-A2 I

Functions of a matrix defined by means of power series. We


have used a polynomial of finite degree to define a function of a matrix. We
shall now use an infinite series to give an alternative expression of a function converges for all finité
of a matrix.
Definition 2-17
Let the power series representation of a function f be
ce
f(A) = ¿ a)i A remarl< is in ord
i=O by using Definition 2­
with the radius of convergence p. Then the function f of a square matrix A requires the computat
is defined as the infinite series (2-78
ce the resulting matrix m:
f(A)~ L a¡A i
(2-76) converges very fast, th(
i=O
We derive some irr
if the absolute values of all the eigenvalues of A are smaller than p, the radius to close this section. \
of convergence; or the mafrix A has the property Ak.= O for some positive
integer k. •

This definition is meaningfulonly if the infinite series in (2"76) converges.


lf A k = O for some positive integer k, then (2-76) reduces to· .
k-l
In (2 e 79), if we choose s
f(A) = L ai A ¡
i=O
FUNCTlONS OF A SQUARE MATRIX 55

If the absolute values of aH the eigenvalues of A are smaller than p, it can also
be shown that the infinite series converges. For a proof, see Reference 77.
Instead of proving that Definitions 2-16 and 2-17 lead to exactIy the same
matrix function, we shall demonstrate this by using Definition 2-17 to derive
(2-73) (2-69).

Example 7
Consider the lordan-form matrix A given in (2-68). Let

f(A) = f(A l ) + f'(Al)(A - Al) +f'i~l) (A - AIf + ...


o
then
o
(2-77)

o (2-74)
Since (A-A1I)i is of the form of (2-64), the matrix function (2-77) reduces
immediately to (2-69). •
1

Example 8

I The exponential function

ih 2 Antn
ower series. We e).! = 1 + At +-- + ... + - + ...
2! n!
:tion of a matrix. We
pression of a function converges for all finite Aand t. Hence for any A, we have
1
eA! = L
<Xl
_tkAk (2-78)
k=O k!
I

(2-75 ) A remark is in order concerning the computation of eAl . If e Al í.s computed


by using Definition 2-16, a closed-form matrix can be obtained. However, it
. of a square matrix A requires the cornputation of the eigenvalues of A. This step can be avoided if
the infinite series (2-78) is used. ClearIy, the disadvantage of using (2-78) is that
the resulting matrix may not be in a clbsed formo However, since the series (2-78)
(2-76) converges very fast, the series is often used to compute eA! on a digital computer.
We derive sorne important properties of expone;ntial functions of matrices
lller -ihan p, the radius to close this section. Using (2-78), it can be shown that .
;= O for sorne positive
eO =1
I oA(!+s) =eA!eAs (2-79)
es in (2-76) converges. e(A+B)1 =eAie Ht if and only if AB =BA (2-80)
to O
In (2-79), if we choose s = - t, then frorn the fact that e = 1, we have
[eA!J - 1 = e - Al (2-81 )
56 LINEAR SPACES AND LINEAR OPERATORS

By differentiation, term by term, of (2-78), we have 2-8 Norms and

~eAt= I _1_tk-1Ak=A(I ~tkAk) All the concepts intre


dt k=l{k-l)! k=ok! over the field of comp
=AeAl=eAtA (2-82) for convenience in th
space (C", C).
Here we have used the fact that functions of the same matrix commute (see The concept of th!
Problem 2-36). idea of length. Any r
The Laplace transform of a function f defined on [O, (0) is defined as as a norm if it has the
1. Ilxll?:ü and Ilxll =1
(2-83)
2. IlctX11 = Ictlllxll.
It is easy to show that 3·ll x l+ x zll::;llx 1 11+

ff' [tJ = S-(k+ 1)


The last inequality is e
Letx =[Xl Xz .

By taking the Laplace transform of (2-78), we have


00 00

ff'{e At ) = L S-(k+l)A k =s-1 L {s-IA)k (2-84) or


k=O k=O
It is well known that the infinite series or

00
lt is easy to verify that
f{Je)={l-Je)-1 = 1 +Je +Je z + ... = L Jek
norm 11'llz is called th
k=O
used mainly in the sta
converges for 1,1.1 < 1. Now if s is chosen sufficiently large, the absolute values if al! the eomponents 01
of all the eigenvalues of S-1 A are smaller than 1. Hence from Definition 2-17, The concept of nor
we have into itself, or equivalel
00
norm of a matrix A is •
{1-s- 1A)-1 = L {s-IA)k (2-85)
k=O
Hence from (2-84) we have
where "sup" stands for
(2-86) least upper bound of
IIAIIis, for any x in (C"
In this derivation, Equation (2-86) holds only for sufficiently large s. However,
it can be shown by analytic continuation that Equation (2-86) does hold for all s
except at the eigenvalues of A. Equation (2-86) can also be established from The norm of A is de
(2-82). Because of ff'[dh{t)/dt] = sff'[h{t)] - h{O), the application of the Laplace norm. For different 11"
transform to (2-82) yield~ then

sff'{e At ) - eO = Aff'{eA1 )
or ..(sl - A)ff'{e A1 ) = I
which yields imrrüidiate\y (2-86).For the matrices in{2-73) and (2-74), we can 17 May be skipped without Ir
also readily establish (2-86). and its study may be coup
-_ .. - .. - -- _..
- - - -----------_. __....._--,------_._----_._--­
...
- _-..

NORMS ANO INNER PROOUCT 57

2-8 Norms and Inner Product 17

All the concepts introduced in this section are applicable to any linear space
over the field of complex numbers or over the field of real numbers. However,
(2-82) for convenience in the discussion, we restrict ourself to the complex vector
space (C", C).
natrix commute (see The concept of the norm of a vector x in (C", C) is a generalization of the
idea of length. Any real-valued function of x, denoted by Ilxll, can be defined
X)) is defined as as a norm if it has the properties that for any x in (C", C) and any a in e
1. Ilxll ~O and Ilxll =0 if and only ifx =0.
(2-83)
2·ll ax ll=l a lllxll·
3. Ilx I + xzll ::;llxlll + Ilxzll.
The last inequality is called the triangular inequality.
Let x = [Xl Xz ..• X n ]'. Then the norm ofx can be chosen as
11

Ilxlld~ ¿
i; 1
IXil (2-87)

f (2-84) or Ilxllz ~ i~l


( 11

Ixd
}IZ
z (2-88)

or Ilxll oo ~ m~x IXil (2-89)


I

It is easy to verify that each of them satisfies all the properties of a norm. The
norm 11'11z is called the euelidean norm. In this book, the concept of norm is
used mainly in the stability study; we use the fact that Ilxll is finite if and only
~, the absolute values if al! the eomponents ofx are finite.
froro Definition 2-17, The concept of norm can be extended to linear operators that map (C", C)
into itself, or equivalently, to square matrices with complex coefficients. The
norm of a matrix A is defined as
(2-85)
IIAII ~ ~~~ I\~III = 1I~~el IIAxII
where "sup" stands for supremum, the largest possible number of IIAxII or the
(2-86) least upper bound of IIAxII. An immediate consequence of the definition of
IIAII is, for any x in (C", C),
tly large s. However, (2-90)
IIAxII ::;IIAllllxll
·86) does hold for all s
) be established from The norm of A is defined through the norm ofx; hence it is called an indueed
lcation ofthe Laplace norm. For different Ilxll, we have different IIAII. For example, if Ilxlll is used,
then

73) and (2-74), we can 17 May be ski~ped without los~ (lC continuity The ~~terial in tbi's secti~-n is used only in Chapter 8,
and its study may be coupled with that chapter. .
58 LINEAR SPACES AND LINEAR OPERATORS

where aij is the ijth eleroent of A. lfllXl12 is used, then and IIABxII

IIAI12 = (Amax(A*A))1/2 for any x.


The norro is a fun
where A* is the coroplex conjugate transpose of A and AmaxCA*A) denotes the two vectors, caBed th
largest eigenvalue of A*A (see Appendix E). Ir Ilxll", is used, then two vectors x and y iJ

=ro~x (tl laijl)


the properties that fo
IIAII",
(x,
These norms are aH different, as can be seen froro Figure 2-7. (ex
The norro of a roatrix has the foHowing properties (x,:
IIA +BII ~IIAII +IIBII (2-91 )
where the "overbar"
IIABII ~IIAIIIIBII (2-92)
property iroplies that
These inequalities can be readily verified by observing that (x, ay) = ex (x, y).
li(A +B)xll =IIAx +Bxll~IIAxII +IIBx\I~(IIAII +IIBII)llx\\ In the coroplex ve<

where x* is the coro¡


roatrix A, we have

and

Consequently we have
The sum of these two This magni tude gives
magnitudes gives the the norm of A.
norm of A. The inner product
(a) (b) In fact, this is the norn

Theorem 2 -14 (Sch'

3 Ilx 11 ~ 1
If we define Ilxll = ( (x, ,
2 !.~

----­
.....:;:-----­

--l--'¡=~/--/--':::::""k-l--l--+--
-4 -3 -2 ~Q.. 1 ------;.::",j
2 r-4_ 5
Xl
Proof
Ax
-2 The inequality is obvio

This magnitude gives O~(x +exy,


the norm of A.
for any ex. Let ex = - <:
(el

Figure 2'-7 Solid lines denote x.; broken lines denote Ax, where A = [_ ~ ~J
(a) IIAIll =4. (b) IIAllz =3.7. (e) tlAII", = 5. which gives the Schwat
. _ - . - 0
-
- - ­
----

FUNCTIONS OF A SQUARE MATRIX 59

and IIABxII :$IIAIIIIBxII :$IIAIIIIBllllxll


for any x.
The norm is a function of a vector. Now we shall introduce a function of
Amax(A *A) denotes the two vectors, called the scalar product or inner product. The inner product of
used, then two vectors x and y in (C", C) is a complex number, denoted by (x, y), having
the properties that for any x, y in (C", C) and any al' a 2 in C,

(x, y) = (y,x)
: 2-7. (alx l +a 2x 2, y) =a l (Xl' y) +a 2(x 2, y)
(x,x»O for all xi=O
(2-91 )

(2-92) where the "overbar" denotes the complex conjugate of a number. The first
property implies that (x, x) is a real number. The first two properties imply
that (x, ay) =a(x, y).
In the complex vector space (C n , C), the inner product is always taken to be
\ +IIBII)\\xll
n

(X, y) =x*y = LX¡Y¡ (2-93)


i=l

where x* is the complex conjugate transpose of x. Hence, for any square


matrix A, we have

(x, Ay) =x*Ay


and (A*x, y) = (A*x)*y =x*Ay

Consequently we have

This magnitude gives (x, Ay) = (A*x, y) (2-94)


the norm of A.
The inner product provides a natural norm for a vector x: Ilxll = (x, x) )1 /2
(b)
In fact, this is the norm defined in Equation (2-88).

Theorem 2-14 (Schwarz inequality)


If we define Ilxll =(x,x»)1/2, then
I(x, y)1 :$llxllllyll
Proof
The inequality is obviously true if y = O. Assume now y i= O. Clearly we have

0:$ (x +ay,x +ay) = (x,x) +a(y,x) +a(x, y) +aa(y, y) (2-95)

for imy CJ. • . Let a = - (y, x)j (y, y); then (2-95) becomes

(x x) > (x, y)(y, x) J(x, y)i2


A= [ 3. 2J.
, . . (y, y) (y, y)
(, where
-1 O
which gives the Schwarz inequality. Q.E.D.
60 LINEARSPACES'AND LINEAR OPERATORS

2-9 Concluding Remarks computation. The g


unstable and should b
In this chapter we have reviewed a number of concepts and results in linear is a numerically stabh
algebra which are useful in this book. The following three main topics were by using complete 1
covered: However, according 1
formance, which inel
1. Similarity transformation. The basic idea of similarity transformation
of use, the gaussian t
(change of basis vectors) and the means of carrying out the transformation
general matrices.
are summarized in Figures 2-4 and 2-5. Similarity transformations can be
The gaussian elim
carried out (a) by computing A =PAP- 1 = Q-l AQ, 2r (b) more easily, by
matrix into a triangul
using the concept of representation: the ith column of A is the representation
then be determined.
of Aq¡ with respect to the basis {q¡, q2,' .. , qn}' The second method will be
small a number shoul
constantly employed in the remainder of this book.
numerical property e
2. Jordanform representation of a matrix. For each eigenvalue of A with
transformations. It (
multiplicity m, there are m linearly independent generalized eigenvectors.
are employed togeth,
Using these vectors as a basis, the new representation of A is in a lordan
method of computin~
canonical formo A systematic procedure for searching these generalized
composition (see App
eigenvectors can be found in Reference S43. If aH eigenvalues are distinct,
a small number is a
then a lordan form reduces to a diagonal formo
"distance" to a matri
lordan-form representation is usually developed by introducing invariant
value decomposition
subspaces and their direct sumo We bypass these concepts and concentrate
position however is q
on the search of the required basis vectors. The interested reader may
orten oversimplified; 1
correlate our derivation with the concepts of invariant subspaces and their
comparable results al
direct sumo
transformations with
3. Functions of a square matrix. Three methods of computing a function of a
The eigenvalues o
matrix f(A), where A is an n x n constant matrix, were introduced. (a) lJse
Once the characteristi
Definition 2-16: First compute the eigenvalues of A, and then find a poly­
subroutines for solviJ
nomial g(A.) of degree n -1 that is equal to f(A.) on the spectrum of A, then
may be ill-conditione
f(A)=g(A). (b) Use the lordan canonical form of A: Let A=QAQ-l.
cause large changes in
Then f(A) = Qf(A)Q -1, where A is in the lordan form and f(A) is computed
the eigenvalues may el
in (2-69) and (2-71). (c) Use Definition 2-17.
problem and should t
Remarks are in order regarding the computer computation of the topics ejgenvalues is the SO-(
covered in this chapter. A problem is said to be ill-conditioned ifsmall changes Hessenberg form and
in data willlead to large changes in solutions. In engineering terminology, if a Problems A-S and A­
problem is very sensitive to the variations of data, then the problem is ill­ compute the charactel
conditioned; otherwise, it is weH-conditioned. A procedure for solving a the Leverrier algorith
problem, or an algorithm, is said to be numerically stable if numerical errors S181 (see also Referen
inside the procedure will not be amplified; otherwise, the algorithm is said to be The lordan canon
numerically unstable. The condition of a problem and the numerical stability and results (see, e.g., I
of an algorithm are two independent concepts. Clearly, whenever possible, solving Riccati equatic
numerically stable methods should be used to solve a problem. There are a canonical form, howe'
large number of texts on the subject of computer computations. The reader is this problem, see Refer
referred to, e.g., References S138, S181, S182, S200, and S212. Several weH­
tested computer programs such as LINPACK and EISPACK are discussed in 18 Roughly speaking, balan,
References S82, S103, and S182. or (he norrns of sorne col
The solution of linear algebraic equations is a basic topic in computer can often be irnproved. :
._-_._-~_-~._---_. .. ---

CONCLUDlNG REMARKS 61

computation. The gaussian elimination without any pivoting is numerically


unstable and should be avoided. The gaussian elimination with partial pivoting
s and results in linear is a numerically stable method. Its numerical stability can be further improved
hree main topics were by using complete pivoting or by using the Householder transformation.
However, according to Reference S138, from the point of view of overall per­
formance, which inc1udes efficiency, accuracy, reliability, generality, and ease
ilarity transformation
of use, the gaussian elimination with partial pivoting is satisfactory for most
mt the transformation general matrices.
ransformations can be
The gaussian elimination with partial pivoting can be used to transform a
or (b) more easily, by
matrix into a triangular form (see Appendix A), and the rank of the matrix can
Ais the representation then be determined. However, because of the difficulty in determining how
second method will be
small a number should be considered as a zero, ambiguity always occurs. The
numerical property of this process can be improved by using Householder
eigenvalue of A with
transformations. It can be further improved if Householder transformations
leralized eigenvectors.
are employed together with pivoting (see Appendix A). The most reliable
)n of A is in aJordan
method of computing the rank of a matrix is by using the singular value de­
ling these generalized
composition (see Appendix E). Although the problem of determining whether
igenvalues are distinct,
a small number is a zero or not remains, the decomposition does provide a
"distance" to a matrix of lower rank. A subroutine to achieve the singular
y introducing invariant
value decomposition is available in LINPACK. The singular value decom­
lcepts and concentrate
position however is quite expensive, and according to Reference S82, its use is
interested reader may
often oversimplified; the results depend highly on balancing. 18 Furthermore,
nt subspaces and their
comparable results are often obtainable with less cost by using Householder
transformations with pivoting.
Iputing a function of a The eigenvalues of A are the roots of the characteristic polynomial of A.
,e introduced. (a) lJse
Once the characteristic polynomial is known, the roots can be solved by using
and then find a poly­ subroutines for solving the roots of polynomials. However a polynomial
le spectrum of A, then may be iII-conditioned in the sense that small changes. in coefficients may
'A: Let A=QÁQ-l.
cause large changes in one or more roots. Hence this procedure of computing
1 and feA) is computed
the eigenvalues may change a well-conditioned prob1em into an iII-conditioned
problem and should be avoided. The most reliable method of computing the
putation of the topics ejgenvalues is the so-called QR method. The method transforms A into the
tioned if small changes Hessenberg form and then carries out a sequence of QR factorizations (see
:ering terminology, if a Problems A-S and A-6 and References S181 and S2ÜÜ). lf it is necessary to
en the problem is ill­ compute the characteristic polynomial of a matrix, several methods, inc1uding
;)cedure for solving a the Leverrier algorithm (Problem 2-39), are discussed in References S82 and
ble if numerical errors S181 (see also Reference S2Ü8).
algorithm is said to be The Jordan canonical form is useful in developing a number of concepts
the numerical stability and results (see, e.g., Problems 2-22, 2-23, 2-37, and 2-4S). It is also useful in
'Iy, whenever possible, solving Riccati equations (Problem 2-46). Computer computation of aJordan
problem. There are a canonical form, however, is an iII-conditioned problem. For a discussion of
.tations. The reader is this problem, see Reference S108. For a computer program,see Reference S124.
d S212. Several .well­
lACK are discussed in 18 Roughly speaking, balancing or scaling is lo rnake al! enlries of a rnalrlx be ofcorriparable size
or Ihe notrns of sorne colurnn and row lo be equal. Afler balancing, Ihe. resulls of cornpulalion
sic topic in computer can oflen be irnproved. See References S82, S84, SI 03 and S202.
62 LINEAR SPACES AND LINEAR OPERATORS

Problems 2-12 Consider Table


respect to the basis {el' e
2-1 With the usual definition of addition and multiplication, which of the following sets b, el' ez, el' and e 2 with r
forms a field?
2-13 Show that simil¡
a. The set of integers quently, the same set of <
b. The set of rational numbers
c. The set of all 2 x 2 real matrices 2-14 Find the P matri
d. The set of polynomial of degree less than n with real coefficients
2-15 Given
2-2 Is it possible to define rules of addition and multiplication such that the set {O, 1, 2}
forms a field?
A
2-3 Given the set {a, b} with a f= b. Define rules of addition and multiplication such that
°
{a, b} forros a field. What are the and 1 elements in this field?
what are the representati
2-4 Why is (e, IR) a linear space but not (IR, C)? {h, Ah, A2 b, A3 h}, respee

2-5 Let lR(s) denote the set of all rational functions with real coefficients. Show that 2-16 What are the ranl
(IR (s), lR(s)) and (IR (s), IR) are linear spaces.

2-6 Which of the following sets of vectors are linearly independent? A, ~[: i
. Hl [:!l HJ ;n (R'. R)
2-17 Find the bases oftt
2-16.

b. [21++31iJ. , [104 -+2iJ1. ' [-iJ3 in (C 2 , IR) 2-18 Are the matrices

c. e- l , te -1, e - 21 in (a¡¡, IR), where a¡¡ denotes the set of all piecewise continuous functions
defined on [0, (0).
d. 3s 2 +s-10, -2s+3,s-5 in(1R 3 [s],IR) nonsingular in the field o
the matrices become nur
3s 2 -124s 5 +s 3 -2s-1 1
e. , - - - - - - , ---::---- in (IR (s), IR) matrices nonsingular in t
3
2s +4s - 1 S2 +s-l

2-19 Does there exist a :


2-7 Is the set in Problem 2-6b linearly independent in (C 2 , C)? Are the sets in Problem
2-6d and e linearly independent in (IR (s), lR(s))?

2-8 What are the dimensions of the following linear spaces?


a. (IR, IR) b. (C, C) c. (e, IR) d. (IR (s), lR(s)) e. (lR(s), IR)
If so, find one.
2-9 Show that the vectors Xl' X 2, ... , X k are linearly dependent in (IR "(s), lR(s)) if and only
2-20 Consider the set o
if there exist polynomials c¡(s), i = 1, 2, ... , k, not all zera, such that
CI(S)X I +C2(S)X 2 + ... +Ck(S)X k =0 x(n) = A"x(O) -t
where A is an n x n cons!;
See Equation (G-31) in Appendix G. x(O), under what conditi(
the equation? Hint: Wri
2-10 .Show that the set of a1l2 x 2 matrices with real coefficients forms a linear space over
IR with dimension 4.
x(n) .
2-11 In an n-dimensional vector space (!!l", $'), given the basis el' e 2, ... , en, what is the
representation of e¡ with respect to the basis?
PROBLEMS 63

2-12 Consider Table 2-1. Suppose the representations of b,e¡,el,e l , and el with
respect to the basis {el' el} are known, use Equation (2-20) to derive the representations of
b, e¡, el, el' and el with respect to the basis {el' el}.
hich of the following sets

2-13 Show that similar matrices have the same characteristic polynomial, and conse­
quently, the same set of eigenvalues. [H int: det (AB) = det A det B.J

2-14 Find the P matrix in Example 3, Section 2-4, and verify A = PAP -1 .
nts

2-15 Given
such that the set {O, 1, 2}

1 multiplication such that

what are the representations of A with respect to the basis {b, Ab, Alb, A3 b} and the basis
{b, Ab, Alb, A3 b}, respectively? (Note that the representations are the same!)

2-16 What are the ranks and nullities of the fol\owing matrices?
coefficients. Show that

lent?
Al = [~ ~ =~]
130
Al =[~ ~~]
001
A3 =[~ ~ ~;
34500
;]

2-17 Find the bases ofthe range spacesand the null spaces ofthe matrices given in Problem
2-16.

2-18 Are the matrices

vise continuous functions


~J
nonsingular in the field of rational functions with real coefficients .1R(s)? For every s in e,
the matrices become numerical matrices with elements in C. For every s in e, are the
matrices nonsingular in the field of complex numbers e?

2-19 Does there exist a solution for the fol\owing linear equations"!
Are the sets in Problem

IR (s), IR) If so, find one.

in (IRn(s), lR(s)) if and only


2-20 Consider the set of linear equations
at
x(n)=A"x(O) +An-lbu(O) +An-lbu(l) + ... + Abu(n -2) +bu(n -1)
where A is an n x n constant matrixand b is an n x 1 column vector. Given any x(n) and
x(O), under what conditions on A and b will there exist u(O), u(l), ... , u(n -1) satisfying
the equation? Hint: Write the equation in the form
forms a linear space over

x(n)-A"x(O)=[b Ah ...
u(n
An-lbJ u(n:-2) .
-1)1
el' el, ... , e", what is the [
u(O)
64 LINEAR SPACES AND LINEAR OPERATORS

2-21 Find the Jordan-canonical-fonn representations of the following matrices: 2_27


19
Consider the m

A, ~[¡ l~] A, ~[ : -!J

4
2
O -2
1
O
-4
of the matrix with multiI
following k vectors,

A'~[: -I~] A'~[: 3]

4 4
-150 20 16
200 160 -25 -20

A,{t 1'] A'~[~


1 1 1
ll. O 1 1

¡]
_J.
2
2
-"-
2
-2
-2
O
O
O
O

O
O
1
O
O
where

are generalized eigenvect

A, ~r!
1 O

!l
O 1 2-28 Show that the ma
O O
that its inverse is given b
-4 -3

2-22 Let A¡ for i = 1,2, ... ,n be the eigenvalues of an n x n matrix A. Show that

detA= TI }'i
j=l

2-23 Prove that a square matrix is nonsingular if and only if there is no zero eigenvalue. 2-29 Show that the dete

2-24 'Under what condition will AH = AC imply B' = C? (A is assumed to be a square O


matrix.) O
2-25 Show that the Vandermonde determinant
O
f3m(s
is equal to
s'
where n=k l +k 2 + ... -t

is eq ua] to TI (J'r-q·
1 s.i<jS."

O 1
2-26 Consider the matrix O O
1 O
1
O O

tJ
O
-al1"1 -a 11 (nl-l)
O O
O O
O O
Show that the characteristic polynomial of A is
6(A) ~ det(AI-A)=A n +CX1An-l +CX2An-2 + ... +CXn-1A +cx n O O

lf Al is an eigenvalue of A [that is, 6(A¡) =0], show that [1 Al AI .,., A~-lris an


eigenvector associated with Al' [The matrix A is called the companion matrix of the poly­ 19 See Reference 6.

nomial 6(A), lt is said to be in the Frobenius form in the numericaL analysis literature.]
20 See Reference SI7.
PROBLEMS 65

lowing matrices: 19
2-27 Consider the matrix shown in Problem 2-26. Suppose that }'l is an eigenvalue
ofthe matrix with multiplicity k; that is, ~(A) contains (le _XI)k as a factor. Verify that the

-!] fol1owing k vectors,


O o O

I~]
O O O
1 O O
-20
(n-1) X'¡-k
~ ~]
'.k-1
1 1
O 1 where (
11- 1) ~ (11 - 1)(11- 2)' .. (n - i)
i ;?: 1
i 1·2·3···i
O O
are generalized eigenvectors of A associated with XI'
2-28 Show that the matrix A in Problem 2-26 is nonsingular if and only ir a" f O. Veriry
that its inverse is given by

_a"i l/a"
:ix A. Show that A- I = O
[
O O

¡:re is no zero eigenvalue. 2-29 Show that the determinant of the m x m matrix
Sk m -1 O O O
assumed to be a square O Skm-l -1 O O
O O Skm-Z O O

O O O Sk2 -1
~m(s) ~m-l(S) ~m - zls) ~2(S) SkI +~I(S)

is equal to
Sn +~I(S)S"-k, +~2(S)sn-kl-k, + ... +~",(s)

where 11 = k l +k 2 + ... +km and ~i(S) are arbitrary polynomials.


20
2_J0 Show that [he characterislic polynornial al lhe rnalrix
O 1 O O O O O O

O O O O O O O O

]
O O O O O O O

-a 122 -a121
---------------------------------~----------------~---
, -------------
O O O O O 1 O O

O O O O O O O O

O O O O O O O
a"_I A +a n
-a21n! -a21(",-I) -a212 -a2!! : - a22n 2 -a22(", -1) -a222 -a221
[( AI .. , A~-IJ' is an
Janion matrix of the poly­ 19 See Reference 6.
:rical analysis literature.J 20 See Reference S17.
66 LINEAR SPACES AND LINEAR OPERATORS

is given by not exist. Let

where ¿\¡¡(s) = sn i +aiils(ll¡-l) + ... +aji(n¡-1)S +aiill¡


D.i}s) =aijls("i- I ) + ... + aij("rI)S + aij"i
Find a matrix B such th¡
Note that the submatrices on the diagonal are of the companion form (see Problem 2-26);
for any nonsingular mat]
the submatrices not on the diagonal are aH zeros except the last row.
2-38 Let A be an /1 x I~
2-31 Find the characteristic polynomials and the minimal polynomials of the foHowing
A with k ~n can be writ
k
matrices:
the minimal polynomial

l~' ~J l~' ~J l~' U l~'


1 O 1 O 1 O 1 O
11 1 1 11 O 11 O 2-39 Define

1]

}'I
O 11 O 11 O 11 O 11
O O O O O O O O (sI-Ar l

What are the multiplicities and indices? What are their geometric multiplicities?
where D.(s) ~ det (sI -A)
stant matrices. This defir
2-32 Show that if 1 is an eigenvalue of A with eigenvector x, then f(l) is an eigenvalue
most /1-1. Verify that
of f(A) with the same eigenvector x.

tr ARo
al =- - ­
2-33 Repeat the problems in Examples 2 and 3 of Section 2-7 by choosing, respectively,
1

g(l)=a ol +a l (1-1) and g(1)=a o(1-1) +al(1-1f(1-2) +a z(1-2). tr AR I

az=- - ­
2
2-34 Given
tr AR z
a, = - ~~-
- 3

tr AR"_I
2-35 Compute eA! for the matrices (X'l=- ~--
n

[~ O O
~ 1~l~J [~O -15~
200
-12~1
160
where tr stands for the tr
matrix. This procedure (
The right-hand-side equa
by using Definition 2-16 and by using the Jordan-form representation. Rn-zs+R,,_d. Forade
2-36 Show that functions of the same matrix commute; that is, 2-40 Prove the Cayley­
f(A)g(A) = g(A)f(A) R"_I' R,,-z, ... , from 0=,

Consequently, we have Ae A' =eA!A. 2-41 Show, by using Pn

2-37 Let (sI-A)-1 =­


L

2-42 Let
B
Find a matrix B such that e = C. Show that if 1, = O for sorne i then the matrix B does
-, - .. ~. - _.. - _.
-------------- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - .------------­

PROBLEMS 67

not exist. Let

c=
A.
O }, O
1 O]
[
¡ni O O A.

Find a matrix B such that eH = C. [Hint: Let I(J.) = log}, and use (2-69).] Is it true that
fonn (see Problem 2-26);
for any nonsingular matrix C, there exists a matrix B such that f!l =C?
·ow.
2-38 Let A be an n x n matrix. Show by using the Cayley-Hamilton theorem that any
'nomials of the following
Ak with k ~ n can be written as a linear combination of {I, A, ... , A" -I}. Ir the degree of
the minimal polynomial of A is known, what modification can you make?

2-39 Define

Tic multiplicities? where ~(s) ~ det(sl-A) ~ s" +als"-I +a2s"-2 + ... +anand Ro, R I, ... , R n- I are con­
stant matrices. This definition is valid because the degree in s ofthe adjoint of (sI - A) is at
len j(A) is an eigenvalue most n - 1. Verify that
tr ARo
a l =- - - ­ Ro=I

JY choosing, respectively, 1

.-2). tr AR I
a2=- - 2 ­

tr AR 2

a,=- - - ­
. 3
......................................................................................

trAR"_2

ct.'l-l ==­
n-l
tr AR"_J
(/.,,=---­ O=AR n_ 1 +anl
n

where tr stands for the trace and is defined a~ the sum of all the diagonal elements of a
matrix. This procedure of computing ai and R¡ is called the Leverrier algorithm. [H int:
The right-hand-side equations can be verH-ied from L\(s)Jl = (sI - A)(iRos"·· J + iR IS" - 2 + ... +
:ation. R,,_ 2S + R,,_ d. For a derivation of the left-hand-side equations, see Reference SI85.]

2-40 Prove the Cayley-Hamilton theorem. (Hint: Use Problem 2-39 and eliminate
R n- I, R n- 2, ... , from 0= AR n_ 1 +anI.)

2-41 Show, by using Problem 2-39,


1 .
(si - A)-I = ~(s) [An-I +(s +adA"- 2 +(S2 +als +(2)An- 3 + ...

+(s"-I +a l s n- 2 + ... +an_¡)I]


2-42 Let
1 .
i then the matrix B does (si -Al-! = - Adjoint (si -A)
~(s)
68 LINEAR SPACES AND LINEAR OPERATORS

and let m(s) be the monic greatest common divisor of all elements of Adjoint (sI - A). q ualifies as an inner prol
Show that the minimal polynomial of A is equal to /'.i(s)/m(s). space. What is the forrr

2-43 Let all eigenvalues of A be distinct and let q¡ be a (right) eigenvector of A associated 2-50 Show that an n >
with A¡, that is, Aq¡ = A¡q¡. Define Q ~ [q1 q2 ... q"J and define has eigenvalue O with m
called a nilpotent matrix.

p~ ~ ~~ Q'
2-51 Let A beanm xn
iorms a linear space, can

where p¡ is the ith row of P. Show that p¡ is a left eigenvector of A associated with A¡,
that is, p¡A = A¡p¡.

2-44 Show that if all eigenvalues of A are distinct, then (sI - A)-l can be expressed as
1
(sI-Ar 1 =¿-q¡p¡
s -A¡
where q¡ and p¡ are right and left eigenvectors of A associated with A¡.

2-45 A matrix A is defined to be cyclic if its characteristic polynomial is equal to its


minimal polynomial. Show that A is cydic if and only if there is only one Jordan block
associated with each distinct eigenvalue.

2_46 21 Consider the matrix equation


PEP +DP +PF +G=O
where all matrices are./) x n constant matrices. lt is called an algebraic Riccati equation.
Define
-F
M= [
G
Let
Q=[Q1 Q2J
Q3 Q4
consist of all generalized eigenvectors of M so that Q -1 MQ = J is in aJordan canonical
formo We write

Show that if Q 1 is nonsingular, then P = Q3Q 1 1 is a solution of the Riccati equation.

2-47 Give three different norms of the vector x = [1 -4 3J.


2-48 Verify the three norms of A in Figure 2-7.

2-49 Show that the· set of all piecewise continuous complex-valued functions defined
over [O, co) forms·a linear space over C. Show that

.. (g, h) ~ IX> g*(t)h(t) dt

21 See Reference S4.


PROBLEMS 69

nents of Adjoint (sI - A). q ualifies as an inner product of the space, where 9 and h are two arbitrary functions of the
space. What is the form of the Schwarz inequality in this space?

1genvector of A associated 2-50 Show that an n x n matrix A has the property A k = O for k> m if and only if A
iefine has eigenvalue O with multiplicity n and index m. A matrix with the property A k =0 is
caBed a ni/potent matrix. [Hint: Use Equation (2-64) and Jordan canonical fonu.]

2- 51 Let A be an m x n matrix. Show that the set of aB! x m vectors y satisfying yA = O


forms a linear space, caBed the left null space of A, of dimension In - peA).

r of A associated with Ai,

A)-l can be expressed as

¡th Á¡.

lolynomial is equal to its


is only one Jordan block

..Igebraic Riccati equation.

r is in a Jardan canonical

the Riccati equation.

·valued functions defined


input u and the outp
be related by an equ

If the input and out

3 can be reduced to

Mathematical Descriptions which can be further

of Systems
if the system is relaxe
input and output of
external description (
cepts of time invarim
of state is introduced

3-1 Introduction
that relates the input
The very first step in the analytical study of a systerri is tó set up mathematical set of two equations
equations to describe the system. Because of different analytical methods used, used to describe a s:
or because of different questions asked, we may often set up different mathe­ state-variable descript
matical equations to describe the same system. For example, in network to illustrate the proce
analysis, if we are interested in only the terminal properties, we may use the Comparisons betweer
'impedance or transfer fundion to describe the network; if we want to know description are given.
the current and voltage of each branch of the network, then loop analysis or descriptions of paral
node analysis has to be used to find a set of differential equations to describe Finally, the discrete-t
the network. The transfer function that describes only the terminal property in the last section.
of a system may be called the external or input-output description of the system, We are concernec
The set of differential equations that describes the internal as well as the terminal physical systems; hel
behavior of a system may be called the internal or state-variable description assumed to be real-va
of the system. variable or a multivaJ
In this chapter we shall introduce the input-output description and the
state variable description of systems from a very general setting. They will be Definition 3-1
developed from the concepts of linearity, relaxedness, time invariance, and
causality. Therefore they will be applicable to any system, be it an electrical, A system is said to be.
a mechanical, or a chemical system, provided the system has the aforementioned terminal and only om
properties. system if and only if
The class of systems studied in this book is assumed to have someinput output terminal.
terminals and output terminals. The inputs, or the causes, or the excitatioris u
are applied at the input terriúnals; the outputs, or the effects, or the responses y The references for
are measurable at theoutput terminals. In Section 3-2 we show that if the and116. The main ob
of linearity, causality,
70
._-_... _ ... .. _.. -_._.- _.... -
_-_.~

.---~-------_._-.- ,

lNTRODUCTlON 71

input u and the output y of a system satisfy the linearity property, then they can
be related by an equation of the form

y(t) = J~eo G(t, r)u(r) dr (3-1 a)

If the input and output have, in addition, the causality property, then (3-1a)
can be reduced to

y(t) = feo G(t,r)u(r)dr (3-1 b)

which can be further reduced to


t

y(t) = r G(t, r)u(r) dr (3-1 e)


Jto
if the system is relaxed at too Equation (3-1) describes the relation between the
input and output of a system and is caBed the input-output description or the
external description of the system. We also introduce in Section 3-2 the con­
cepts of time invariance and the transfer function. In Section 3-3 the concept
of state is introduced. The set of equations
x(t) = A(t)x(t) + B(t)u(i) (3-2a)

y(t) = C(t)x(t) + E(t)u(t) (3-2b)

that relates the input u, the output y, and the state x is then .introduced. The
o set up mathematical set of two equations of the form (3-2) is called a dynamical equation. If it is
alytical methods used, used to describe a system, it is caBed the dynamical-equation description or
et up different mathe­ state-variable description ofthe system. We give in Section 3-4 many examples
example, in network to illustrate the procedure of setting up these two mathematical descriptions.
~rties, we may use the Comparisons between the input-output description and the dynamical-equation
:; if we want to know description are given in Section 3-5. We study in Section 3-6 the mathematical
then loop analysis or descriptions of parallel, tandem, and feedback connections of two systems.
equations todescribe FinaBy, the discrete-time versions of Equations (3-1) and (3-2) are introduced
the terminal property in the last section.
,cription of the system. We are concerned with descriptions of systems that are models of actual
¡as well as the terminal physical systems; hence all the variables and functions in this chapter are
te-variable description assumed to be real-valued. Before proceeding, we classify a system as a single­
variable or a multivariable system according to the following definition.
t description and the
setting. They will be Definition 3-1
time invariance,. and
A system is said to be a single-variable system if and only if it has only one input
em, be it an electrical,
. terminal and only one output terminal. A system is said to be a multivariable
Las the aforementioned
system if and only if it has more than one input terminal or more than one
output terminal. •
d to have sorne input
es, or the excitations u
The ·references for this chapterare 24, 27, 31, 53, 60, 68, 70, 73, 92, 97, 109,

~cts, or the responses y


and 116. The main objective of Sections 3-2 and 3-3 is to introduce the concepts

2 we show that if the


of linearity, causality, time invariance, and the state, and to illustrate their

72 MA'rHEMATlCAL DfSCRIPTlONS OF SYSTEMS

importance in developing the linear equations. They are not introduced very will be excited solely al
rigorously. For a more rigorous exposition, see References 60, 68, 109, and tion, it is Iegitimate to
116.

3-2 The Input-Output Description where H is Some oper


terms of the input u of
The input-output description of a system gives a mathematical relation between
initially relaxed at - (
the input and output of the system. In developing this description, the know­
short. Note that Equ
ledge of the internal structure of a system may be assumed to be unavailable to
this section, whenever
us; the only access to the system is by means of the input terminals and the
mean only those input
output terminals. Under this assumption, a system may be considered as a
"black box," as shown in Figure 3-1. Clearly what we can do to a black box is
Linearity. We intro
to apply all kinds of inputs and measure their corresponding outputs, and then
exactly the same as thl
try to abstract key properties of the system from these input-output pairs.
We digress at this point to introduce sorne notations. The system shown
Definition 3-2
in Figure 3-1 is assumed to have p input terminals and q output terminals.
The inputs are denoted by Ul' U2' ... ,up or by a p x 1 column vector A relaxed system is saie
u = [Ul U2 ... upJ'. The outputs or responses are denoted by Y10 Y2"'" Y q
or by a q x 1 column vector y = [Yl Y2 ... yqJ'. The time interval in which
the inputs and outputs will be defined is from - 00 to +00. We use u or u(·) for any inputs U 1 and ti
to denote a vector function defined over ( - 00, 00); u(t) is used to denote the relaxed system is said 1<
value ofu at time t. Ifthe function u is defined only over [to, tI)' we write u[to,tt!.
If the output at time tI of a system depends only on the input applied at time In engineering litera
t 10 the system is called an instantaneous or zero-memory system. A network
.that consists of orily resistors is sueh a system. Most systems of interest,
however, have memory; that is, the output at time tI depends not only on the
input applied at tI' but also on the input applied before and/or after tI' Hence,
for any u 1 , U 2 and any re
if an input u[tl>oo) is applied to a system, unless we know the input applied
in (3-4) and the set of COi
before tI' the output Y[tl.oo) is generally not uniquely determinable. In fact,
ship in (3-5) is called th(
for different inputs applied before t lo we will obtain different output Y[tl> 00)'
caBed the property of ha¡
·although the same input U[II,OO) is applied. It is clear that such an input-output
the system is said to satis
pair, which lacks a unique relation, is ofno use in determining the key properties
of the system. Hence, in developing the input-output description, before an whether or not there is 1
of homogeneity does no
input is applied, the system must be assumed to be relaxed or at rest, and that
che íollowing exampíe.
the output is excited solely and uniquely by the input applied thereafter. If the
concept of energy is applicable to a system, the system is said to be relaxed at
time tI if no energy is stored in the system at that instant. As in the engineering Example 1
literature, we shall assume that every system is relaxed at time - oo. Con­
Consider a single-variab
sequently if an input u( _ 00, 00) is applied at t = - 00, the corresponding output

for all l. It is easy to vel


homogeneity but not the

The property of addi


Figure 3-1 A system with p input terminals and q output terminals. geneity. To be precise, 1
_ _ • __ •• 0 •• _ . •
--~,--,_ .... .. _._."­
• ~ o . -_._.-...-_ .. _~-----.-"---~ ._-_. ,_._-_.. __ .
_'_"._._ .._.-,-'-"-
..'---""--- -~~_._--- _._'._. __
__ ._ .. ... _---_ ..­

THE INPUT-OUTPUT DESCRIPTION 73

:e not introduced very will be excited solely and uniquely by u. Hence, under the relaxedness assump­
ences 60, 68, 109, and tion, it is legitimate to write
y=Hu (3-3)

where H is sorne operator or function that specifies uniquely the output y in


terms of the input u of the system. In this book, we shall call a system that is
atical relation between initially relaxed at - 00 an initially relaxed system-or a relaxed system, for
description, the know­ short. Note that Equation (3-3) is applicable only to a relaxed system. In
~d to be unavailable to this section, whenever we talk about the input-output pairs of a system, we
Lput terminals and the mean only those input-output pairs that can be related by Equation (3-3).
lay be considered as a
an do to a black box is Linearity. We introduce next the concept of linearity. This concept is
ding outputs, and then exactly the same as the linear operators introduced in the preceding chapter.
nput-output pairs.
1S. The system shown Definition 3-2
Jd q output terminals.
p x 1 column vector A relaxed system is said to be linear if and only if
enoted by Yl,Y2,··· ,Yq H(a l Ul + a2 u 2) = alHu l + a 2 Hu 2 (3-4)
;: time interval in which
+00. We use u or u(·) for any inputs u l and u 2 and for any real numbers al and a 2 . Otherwise the
) is used to denote the relaxed system is said to be nonlinear. I
. [to, tI)' we write U[lo,I,)·
he input applied at time In engineering literature, the condition of Equation (3-4) is often written as
'ry system. A network H(u l + U2) = HUI + HU 2 (3-5)
)st systems of interest,
epends not only on the H(au l ) = aHu l (3-6)
andJor after tI. Hence, for any u l , u 2 and any real number a. It is easy to verify that the condition given
:now the input applied in (3-4) and the set of conditions in (3-5) and (3-6) are equivalen1. The relation­
determinable. In fact, ship in (3-5) is caBed the property of additivity, and the relationship in (3-6) is
different output Y[I¡,OO)' caBed the property of homogeneity. If a relaxed system has these two properties,
at such an input-output the system is said to satisfy the principIe ofsuperposition. The reader may wonder
lining the key properties whether or not there is redundancy in (3-5) and (3-6). GeneraBy, the property
t description, before an of homogeneity does not imply the property of additivity, as can be seen from
lxed or at rest, and that the following example.
'Jplied thereafter. lf the
is said to be relaxed at
Example 1
1. As in the engineering
ed at time - oo. Con­ Consider a single-variable system whose input and output are related by
e corresponding output
if u(t -1) 1=0
ifu(t-1)=0
for aB t. It is easy to verify that the input-output pair satisfies the property of
homogeneity but not the property of additivity. •
>
The property of additivity, however, almost implies the property of homo­
terminals.
geneity. To be precise, the condition H(u l + u 2) = HUI + Hu 2 for any u l and
74 MATHEMATICAL DESCRIPTlONS OF SYSTEMS

sider a relaxed single-

I
1
/i
As shown in Figure 3-.
by a series of pulse fUJ
u(t¡)(jt,(t - tJ~, we can

Figure 3-2 A pulse functían (l/l(t-t l ). If the input-output pe


then we have 2
U 2 implies that H(IXU 1 ) =IXHu¡, for any rational number IX (see Problem 3-9).
Since any real number can be approximated as closely as desired by a rational
number, if a relaxed system has the continuity property that Un -4 U implies Now as t,. tends to z(
H Un -4 Hu, then the property of additivit)' implies the property of homogeneity. summation becomes
We shall develop in the following a mathematical description for a linear toward a 6-function.
relaxed system. Before proceeding we need the concept of the delta function
or impulse function. We proceed intuitively because a detailed exposition
would lead us too far astray.1 First let (j t,(t - t 1) be the pulse function defined
in Figure 3-2; that is, Now if H 6(t - r) is km
for t < t 1 puted from (3-9). The
the relaxed system due
for t 1.::;; t < t 1 + ~

Note that g is a functi


time at which the 6-fun
Note that (jt,(t - t 1 ) has unit area for all~. As ~ approaches zero, the limiting
at which the output is
"function"
function, it is called th
write the output at tim

is called the impulse function or the Dirac delta function or simply (jlunction.
Thus the delta function (j(t - t 1) has the properties that
oo 111 +E 3(t -
J -ce
a(t - t 1) dt =
ti-E;
tí) dt = i 2 Th;,::; condiUon fer inlerchai
cussion of this problem, see
for any positive c and that
U(¡¡)O¡:,(I-I¡).

(3-7) /
for any function f that is continuous at t l' T~

u (1) I
1
I
I
I 11
I 11
With the concept of impulse functión, we are ready to develop a mathe­ 1
I
1 I I 11
=U(I,)/i0¡:' 1 I I~.
rilatical description for relaxed linear systems. We discuss first single-variable I 1 I :1
systems; the result can then be easily extended to multivariable systems. Con­
J I
I
I
:
I
:
I
: I

1 For a rigorous development of the subsequent material, the theory of distributions is needed;
see Reference 96. Figure 3-3 Pulse-func
~._------_.-_.,-_._._-,_ .. __ .- _. -"'~'- _-
.. __ ._ ..•. .. _-_ ... "---'- _
.. _•.. .•..._. __ ._---_. __.. __ ..... _._- .-

THE INPUT-OUTPUT DESCRIPTION 75

sider a relaxed single-variable system whose input and output are related by
y=Hu
As shown in Figure 3-3, every piecewise continuous input can be approximated
by a series of pulse functions. Since every pulse function can be described by
u(tJb,,(t - tJ~, we can write the input function as
u =i= ¿ u(t¡}b,,(t - t¡}~
i

If the input-output pairs of the relaxed system satisfy the linearity property,
then we have 2
r ex (see Problem 3-9). (3-8)
; desired by a rational
ty that U n - u implies Now as ~ tends to zero, the approximation tends to an exact equality, the
perty ofhomogeneity. summation becomes an integration and the pulse function b,,(t - t¡} tends
escription for a linear toward a b-function. Consequentiy, as ~-+O, Equation (3-8) becomes
: of the delta function
a detailed exposition y= J~oo (Hb(t -'l:))u('l:)d'l: (3-9)
pulse function defined
Now if Hb(t - r) is known for aH r, then for any input, the output can be com­
puted from (3-9). The physical meaning of Hb(t - r) is that it is the output of
the relaxed system due to an impulse function input applied at time r. Define
Hb(t - r) = g(', r) (3-10)

Note that g is a function of two variables-the second variable denoting the


time at which the b-function is applied and the first variable denoting the time
ches zero, the limiting at which the output is observed. Since g(', r) is the response of an impulse
function, it is called the impulse response of the system. Using (3-10), we can
write the output at time t as

or simply b{unction. y(t) = roo g(t, r)u(r) dr (3-11 )

2 The condilion lor inlerchanging the order of H and lhe Slil11mation is disregarded. For a dis­
cussion of this problem, see Reference S125, pp. 2-6.

u(¡¡ J6{',.(t - ti) {',.

(3-7) /
T~u (ti)
I
I
I
I
I 1I
I 1 I

to develop a mathe~ 1
= u (t) -.{',. I
I
I
I
I 11

I
.ss first single-variable ,{',. I I \
I I I
ariable systems. Con- I I I

of distributions is needed;
Figure 3-3 Pulse-furiction approximation of an input function.
76 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

Hence if g(', r) for aH r is known then for any input u, the output can be com­ Consequently, the in
puted fram (3-11). In other words, a linear relaxed system is completely becomes
described by the superposition integral (3-11), where g(', r) is the impulse
response of the system, and theoreticaHy it can be obtained by direct measure­
ments at the input and the output terminals of the system.
Relaxedness. Rec
If a system has p input terminals and q output terminals, and if the system
is initially relaxed at - 00, the input-output description (3-11) can be extended to the system is--relaxed
solely and uniquely b
y(t) = roo G(t, r)u(r) dr (3-12)
Definition 3-3

gll(t, r) gdt, r) ... glP(t, r)l A system is said to be


where G(t,r)= g21~t,r) g22~t,r) .,. g2P~t,r) and uniquely excited

r gql(t,r) gqz(t,r) gqp(t,r)


and gij(t, r) is the response at time t at the ith output terminal due to an impulse
This concept can
If all the voltages ac
are zero at time to, the
function applied at time r at the jth input terminal, the inputs at other terminals at t o and if an input u
being identically zero. Equivalently, gij is the impulse response between the by the initial conditio
jth input terminal and the ith output terminal. Hence G is called the impulse­ will be excited by the
response matrix of the system. at to, then its input-ol
A1though the input and output of a relaxed linear system can be related by
an equation ofthe form in Equation (3-12), the equation is not readily applicable
because the integration required is from - 00 to 00 and because there is no It is clear that if ~
way of checking whether or not a system is initially relaxed at - oo. These system is still relaxed ;
difficulties will be removed in the subseqtient development. are linearly related, it
for a system to be re.
words, if the net efTec1
Causality. A system is said to be causal or nonanticipatory if the output of system is relaxed at t o
the system at time t does not depend on the input applied after time t; it depends
only on the input applied before and at time t. In short, the past affects the Example 2
future, but not conversely. Hence, if a relaxed system is causal, its input and
output relation can be written as A unit-time-delay syst.
one unit of time; that
y(t) = Hu( - (3-13)
oo,!]
uf'" - ! . te) == O, although
for all t in ( - 00, 00). If a system is not causal, it is said to be noncausal or
anticipatory. The output of a noncausal system depends not only on the past Now if a system v
input but also on the future value of the input. This implies that a noncausal to be relaxed at to, the
system is able to predict the input that will be applied in the future. For real
physical systems, this is impossible. Hence causality is an intrinsic property of
every physical system. Since in this book systems are models ofphysical systems,
Hence, if a system is dI
all the systems we study are assumed to be causal.
If a relaxed system is linear and causal, what can we say about the impulse
response matrix, G(t, r), ofthe system? Every element of G(t, r) is, by definition,
theoutput due to a D-function input applied at time r. Now if a relaxed system
we know that the inpul
is causal, the output is identically zera before any inputis applied.. Hence a
linear system is causal if and only if . and that the system is 1
A legitimate q uesti.
G(t,r)=O for aH r and all t < r (3-14 ) to, how do we know th¡
THE INPUT-OUTPUT DESCRIPTlON 77

he output can be com­ Consequently, the input-output description of a linear, causal, relaxed system
system is completely becomes
g(', -r) is the impulse
y(t) = Loo G(t, ,)u(,) d, (3-15 )
Iled by direct measure­

:m.
Relaxedness. RecaH that the equation y = Hu of a system holds only when
,nals, and if the system

the system is-relaxed at - 00, or equivalently, only when the output y is excited
-11) can be extended to

solely and uniquely by u(-oo,oo)' We may apply this concept to an arbitrary to.
(3-12 )
Definition 3-3
A system is said to be relaxed at time t o if and only if the output is solely
~: :~j
Y[to,oo)
and uniquely excited by u[to,oo)' •

This concept can be easily understood if we consider an RLC network.


~, ,)
If aH the voltages across capacitors and aH the currents through inductors
lÍnal due to an impulse are zero at time to, then the network is relaxed at to. If a network is not relaxed
puts at other terminals at t o and if an input u[to,oo) is applied, then part of the response will be excited
? response between the by the initial conditions; and for different initial conditions, different responses
; is caHed the impulse- will be excited by the same input u[to,oo)' If a system is known to be relaxed
at to, then its input-output relation can be written as
stem can be related by
(3-16)
s not readily applicable
ld because there is no It is clear that if a system is relaxed at - 00, and if u( _ oo,to) == 0, then the
:laxed at -~. These system is still relaxed at to. For the class of systems whose inputs. and outputs
lent. are linearly related, it is easy to show that the necessary and sufficient condition
for a system to be relaxed at t o is that y(t) = HU(-oo,to)=0 for all t~to' In
words, if the net effect of u( _ oo,to) on the output after to is identicaHy zero, the
'patory if the output of system is relaxed at to.
after time t; it depends
>rt, the past affects the Example 2
is causal, its input and
A unit-time-delay system is a device whose output is equal to the input delayed
one unit of time; that is, y(t) = u(t - 1) for aH t. The system is relaxed at t o if
(3-13) u[ro- I.to) == 0, although u( - 00,10 - 1) 1= O. ~
aid to be noncausal or
Now if a system whose inputs and outputs are linearly related is known
1s not only on the past
to be relaxed at to, then the input-output description reduces to
lplies that a noncausal
n the future. For real y(t) = roo G(t, ,)u(,) d,
an intrinsic property of
1els ofphysical systems,
t
Hence, if a system is described by
say about the impulse y(t) = r' G(t, ,)u(,) d, (3-17)
. G(t, ,) is, by definition,
"¡ow if a relaxed system
t
we know that the input-output pairs of the system satisfy the linearity property
It is applied. Hence a
and that the system is causal and is relaxed at to.
A legitimate question lllay beraised at this point: Given a system at time
(3-14 ) to, how do we know thatthe system is relaxed? For a system whose inputs and
78 MATHEMATlCAL DESCRIPTIONS OF SYSTEMS

outputs are linearly related, this can be determined without knowing the Proof
previous history of the system, by the use of the foHowing theorem.
If u[to,oo) == O, theout¡:

Theorem 3-1
lo
r
y(t) = Loo 4
A system that is describable by Since

y(t) = J~oo G(t, ,)u(,) d,


is relaxed at t o if and only if U[lo, 00) == O implies Y[lo, 00) =o. I is a constant vector, t
on [to, (0). Conseq
Proof (see Appendix B), anl

N ecessity: If a system is relaxed at to, the output y(t) for t ? t o is given by This is an import¡
Corollary 3-1, its rela

1 00
lo
G(t, ,)u(,) d, over any nonzero intl
interval, then the sys
Hence, if U[lo,oo) = o, then Y[lo, 00) = O. Sufficiency: We show that if U[lo,oo) =O next chapter that the
implies Y['o,oo) =0, then the system is relaxed at to. Since function matrices 01
satisfies the condition
y(t) = roo G(t, ,)u(,) d, = roo G(t, ,)u(,) d, + 1~ G(t, ,)u(,) d, able.
We give an examp
= =
the assumptions U[lo. 00) O, Y[lo,oo) O imply that whose inputs and out

r
lo
J-oo
G(t, ,)u(,)d, =0 for aH t ? to
Example 3
Consider the system
In words, the net effect ofu(-oo,lo) on the output y(t) for t ?t o is zero, and hence the output is the volt.
the system is relaxed at to. Q.E. D. stored in the capacitor
zero if no voltage is a¡:
An implication of Theorem 3-1 is that given a system at time to, if the system relaxed at to, because
is known to be describable by depending on which i
roo
Loo G(t,,)u(,)dT
the relaxedness of the system can be determined from the behavior of the 1--------­
I R = I 11
system after t o without knowing the previous history of the system. Certainly
I
it isimpractical or impossible to observe the output from time to to infinity; I
I + +
fortunately, for a large class of systems, it is not necessary to do so. I
I
e
I
I
Corollary 3-1 IL _
Ir the impulse-response matrix G(t, ,) of a systerncan be decomposed into
G(t, ,) = M(t)N(,), and ir every element of M is analytic (see Appendix B) .on (a)
( - 00, (0), then ihe system is relaxed at t o if and only if for sorne fixed positive Figure 3-4 A nonline¡
e, U[lo,lo+<) = O implies Y[lo,lo+<) = O. linear capacitor C.
._..~... ._-_.._.._._--_.__..__.."--_ .....__..... _._--:~:::::~::-. . ::::..-==~~-:::::=:.=.=-:~-=.:-~=-...:.:=.:.-=:..::..:..====::.~---------;------:----------
- • __ .~ •• _,_. _ _. , _ . _ •• ' _ 0 • • . '•••

I
i

THE INPUT-OUTPUT DESCRIPTlON 79

without knowing the Proof


g theorem. Ir u[co,oo) == O, the output y(t) of the system is given by

y(t) = roo G(t,t)u(t)dt=M(t) roo N(t)u(t)dt for t 2: t o

Since

roo N(t)u(t) dt

• is a constant vector, the analyticity assumption of M implies that y(t) is analytíc


on [to,oo). Consequently, if Y[to,lo+e) ==0 and if u[Co.oo) ==0, then Y[lo,oo)==O
(see Appendix B), and the corollary follows from Theorem 3-1. Q.E.D.

t 2: t o is given by This is an important result. For any system that satisfies the conditions of
Corollary 3-1, its relaxedness can be easily determined by observing the output
over any nonzero interval of time, say 10 seconds. Ir the output is zero in this
interval, then the system is relaxed at that moment. It will be shown in the
how that if U[lo, 00) == O next chapter that the class of systems that are describable by rational transfer­
function matrices or linear time-invariant ordinary differential equations
satisfies the conditions of Corollary 3-1. Hence Corollary 3-1 is widely applic­
G(t, t)u(t) dt able.
We give an example to illustrate that Theorem 3-1 does not hold for systems
whose inputs and outputs are not linearly related.

Example 3
Consider the system shown in Figure 3-4; the input is a voltage source, and
2:t o is zero, and hence the output is the voltage across the nonlinear capacitor. Ir the electric charge
Q.E.D. stored in the capacitor at time t ois either O, q b or qz, the output wil\ be identically
zero if no voltage is applied at the input. However, the system is not necessarily
at time to, if the system relaxed at to, because if an input is applied, we may obtain different outputs
depending on which initial charge the capacitor has. •
.\ ~}

1-------------1

n the behavior of the I R=I.l1 I

the system. Certainly +


Jm time t o to infinity; + +q
'y to do so. e y
q

. be decomposed into
:; (see Appendix B) on (a) (b)

'or some fixed positive Figure 3-4 A nonlinear network. (a) The network. (b) The characteristic o(the non­
linear capacitor C.
80 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

Time invariance. Ifthe characteristics ofa system do not change with time, this to the multivaria
then the system is said to be time invariant, fixed, or stationary. In order to
define it precisely, we need the concept of a shifting operator Q". The effect of
the shifting operator Q" is illustrated in Figure 3-5. The output of Q" is equal for all t and -r. Heno
to the input delayed by \1. seconds. Mathematically it is defined as ii ~ Q"u if
and only if ii(t) = u(t - \1.) or ii(t +\1.) = u(t) for all t.

Definition 3-4 we know that its inp


A relaxed system is said to be time invariant (or stationary or fixed) if and only if invariance properties
time-invariant case, tI:
HQ"u=Q"Hu (3-18) to be Oand the time ir
for any input u and any real number \1.. Otherwise the relaxed system is said we start to consider
to be time varying. I (3-19) becomes

The relation HQ"u = Q"Hu can also be written as HQ"u = Q"y, which implies y(t)
that if an input is shifted by \1. seconds, the waveform of the output remains the
same except for a shift by \1. seconds. In other words, no matter at what time an The second equality (
input is applied to a relaxed time-invariant system, the waveform of the output The integration in (3
is always the same. G(t, -r) represents the
If a relaxed linear system is known to be time invariant, what condition time -r. G(t) represent~
does that impose on its impulse response? The impulse response g(', -r) is the at -r =0. Following (
output due to a J-function input applied at time -r; that is, g(', -r) = HJ(t - -r). if G(t) = O for all t < O
Ifthe system is time invariant, then we have
Transfer-function
Q~(', -r) = Q"HJ(t - -r) = HQ"J(t --r) describable by convol
= HJ(t - (-r +\1.)) = g(', -r +\1.) transform, because it \
Now by the definition of Q", the equation Q~(', -r) = g(', -r +\1.) implies g(t, -r) = an algebraic equation :
g(t +\1., -r +\1.), which holds for any t, -r, and \1.. By choosing \1. = - -r, we have of y; that iS,4
g(t, -r) = g(t - -r, O) for all t, -r. Hence the impulse response g(t, -r) of a relaxed,
linear, time-invariant system depends only on the difference of t and -r. Extending

Since G(t - -r) = O for


u set at 00 ; herree, from

y(s)

3 Note thilt G(e, r) and G(t


symbol G is used.
4 If Y contains delta functic
Figure 3-5 The effect of a shifting operator on a signa\.
inc1ude the delta function:
_._ ...
.----
_-",,_
._.~-----_ .. _.. ­ ..­
..... _-----_ --' .. _.. -_ ..... - __ . _ - ; • • ._. --:-_ ••
••• ".

_ • • _ ..

_~ •• _-_. - - _ .... _
••~ __ •
•• _ . . ._ - _ . . .
L __ •• _ . . . •
• • • • o._• • • • _.
~~ • ~~
_ ••• _ ••
~._
_~

..__ •
_ _ .. _ - ._,_
._. _ •• - . - -_ _ . _ . • ~ ~

THE INPUT-OUTPUT DESCRIPTION 81

10t change with time, this to the multivariable case, we obtain 3


tionary. In order to
tor Q(/.. The elfect of G(t, ,) = G(t -', O) = G(t -,)
output of Q(/. is equal for all t and ,. Hence, if a system is described by
defined as ü ~ Q(/.u if
y(t) = rr G(t -,)u(,) d, (3-19 )
t
we know that its input-output pairs satisfy the linearity, causality, and time­
)r fixed) if and only if invariance properties; furthermore, the system is relaxed at time to. In the
time-invariant case, the initial time t o is always chosen, without loss of generality,
(3-18) to be Oand the time interval of interest is [O, (0). Note that t o =0 is the instant
elaxed system is said we start to consider the system or to apply the input u. Ir t o =0, Equation
I (3-19) becomes

= Q(/.y, which implies


le output remains the
y(t) = t G(t-,)u(,)d,= J~ G(,)u(t-,)d, (3-20)

latter at what time an The second equality of (3-20) can be easily verified by changing the variables.
tveform of the output The integration in (3-20) is called the convolution integral. Since G(t - ,) =
G(t, ,) represents the responses at time t due to Ci-function inputs applied at
'iant, what condition time " G(t) represents the responses at time t due to Ci-function inputs applied
response g(', ,) is the at, =0. Following (3-14), a linear time-invariant system is causal if and only
is, g(·,,)=HCi(t-,). if G(l) =0 for all l < O.

Transfer-function matrix. In the -study of the class of systems that are


describable by convolution integrals, it is of great advantage to use the Laplace
transform, because it will change a convolution integral in the time domain into
; +ex) implies g(t, ,) = an algebraic equation in the frequency domain. Let y(s) be the Laplace transform
¡ing ex = - " we have of y; that iS,4
e gel, ,) 01 a relaxed,
01 l and ,. Extending y(s) ~ 5t'(y) = too y(t)e- SI dt
Since G(t - ,) =0 for ,> t, the upper limit of the integration in (3-20) can be
set at 00; hence, from (3-20), we have

y(s) = tOO (tOO G(t-,)U(,)d,)e-S1dt

= too (tOO G(t _,)e-S(I-t) dt) u(,)e- St d,

= tOO G(v)e- SV dv tOO u(,)e- St d,


~ G(s)ñ(s) (3-21 )

3 Note thatG(t,"t) and G(t -"t) are two <;lirrerent runctions, However. ror convenience, the same
symbol G is used. - ­

4 Ir y contilins delta runctions at t =0, the lower ¡imil or the integration should start rrom O-to
inelude the delta runctions in the transrorm.
82 MATHEMATlCAL DESCRlPTIONS OF SYSTEMS

Here we have changed the order of integration, changed the variables, and used rational functions al
the fact that G(l) = Ofor l < O. As defined in (3-21), C(s) is the Laplace transform proper rational funcl
of the impulse-response matrix; that is,
A rro
G(s) = Jo G(t)e - SI dt 3-3 The State·

and is called the transfer-function matrix of the system. For single-variable The concept of 51

systems, C(s) reduces to a scalar and is cal1ed the transfer function. Hence the able only when the S)
transfer function is the Laplace transform of the impulse response; it can also be say at time t o, then tb
defined, following (3-21), as output y [lo, ro) depend
ditions at too Henc
.. ;t'[y(t)] I y(s) I (3-22) addition to the input
g( s) = ;t'[ ( )] the system is = -;;;--() relaxed
U (relaxedatl=O U S atl=O initial conditions is ca
together with the in
where the circumf[ex (.. ) over a variable denotes the Laplace transform of the
example, in newtoni
salIle variable; for example,
a particle (system) al
not uniquely determi
known. How the p~
immaterial in determ
We see that the familiar transfer functions are the input-output descriptions two numbers, the po~
of systems. It is important to note that this input-output description is obtained state of the system at t
under the relaxedness assumption of a system; hence, if the system is not relaxed the momentum at lo,
at t = O, the transfer function cannot be directly applied. Thus whenever a
transfer function is used, the system is always implicitly assumed to be relaxed Definition 3-6
at t =0.
A transfer function is not necessarily a r¡ltional function of s. For example, The state of a system ;
the impulse response g(t) ofthe unit-time-delay system introduced in Example 2 with U[IO, ro)' determinl
is (j(t -1), and its transfer function is e-s, which is not a rational function of s.
However, the transfer functions we shall study in this book are exclusively By the behavior o
rational functions ofs. In fact, we study only a special class ofrational functions. the system. If the sysl
branch of the networ
be computed. If the
Definition ::S-5
U[IO'''')' in computing 1

A. rational function g(s) is said to be proper if g(oo) is a finite (zero al" nonzero) [1 summarizes the ess

constant. [¡(s) is said to be strictly proper if [¡( (0) = O.. A rational matrix C(s) needed in determining
is said to be proper if C(oo) is a finite (zero or nonzero) constant matrix. C(s) input ui- 00,11)' i = 1, 2,
is said to be strictly proper if C( (0) = O. • though ui - 00,11)' i = 1,
the system are identic;
We give sorne exal
For example, [¡(s) = s2/(s -1) is not proper; [¡(s) = s2/(S2 - s + 2) is proper
and [¡(s) = s2/(S3 - s) is strictly proper. It is clear that if [¡(s) = N(s)/D(s), [¡(s) is
Example 1
proper if and only if degN(s).:s;degD(s); [¡(s) is strictly proper if and only if
deg N(s) < deg D(s), where deg stands for the degree of a polynomial. A rational Consider the network
matrix is proper if andortly if al1 of its elements are proper. A rational matrix current through the i
.is strictly proper if and only if all of its elements are strictly proper.
Ifatransfer function is not proper, high-frequency noises will be greatly 5 The exceptions are transl
amplified and will overwhelm information-bearing signals. Hence improper metelS. See Reference 54
-_.. __ ._ .. - -_._---- .. "---._.... _-_._~--,---,~."._----- - _.-._-~_.~_.-~-----_.-
" ..._.. _---,. __ .. ~---_ .. _.- ._--_.. ,._- .. _----_._~---~--_._._- --_... ~,._-- .. _- -_ ... _._._----~._. __ - _. __ ._._ --_._--------.
.. _. __ _----. -----
.._.'--'-" ... __ .. ..

THE STATE-VARIABLE DESCRIPTION 83

he variables, and used rational functions are hardly used in practice. 5 In this book, we study only
the Laplace transform proper rational functions and proper rational matrices.

3-3 The State-Variable Description

. F or single-variable The concept of state. The input-output description of a system is applic­


r function. Hence the able only when the system is initially relaxed. If a system is not initially relaxed,
?sponse; it can also be say at time t o, then the equation Y[to.,.,) = HU[to.,.,) does not hold. In this case the
output Y[to.,.,) depends not only on the input u[to.,.,) but also on the initial con­
ditions at too Hence, in order to determine the output Y[to.,.,) uniquely, in
addition to the input u[to.,.,) we need a set of initial conditions at too This set of
(3-22)
initial conditions is called the state. Hence the state at t o is the information that,
together with the input u[to.,.,)' determines uniquely the output Y[to.,.,)' For
place transform of the example, in newtonian mechanics, if an external force (input) is applied to
a particle (system) at time t o, the motion (output) of the particle for t '2:. t o is
not uniquely determinable unless the position and velocity at time t o are also
known. How the particle actually attained the position and velocity at t o is
immaterial in determining the motion of the particle after too Hence the set of
lt-output descriptions two numbers, the position and velocity at time t o, is qualified to be called the
lescription is obtained state ofthe system at time.t o. Note that the set oftwo numbers, the position and
;.: system is not relaxed the momentum al t o, is also qualified as the state.
:d. Thus whenever a
1ssumed to be relaxed [)efinition 3-6
The state of a system at time t o is the amount of information at to that, together
)n of S. For example,
with u[to. ,.,), determines uniquely the behavior of the system for all t '2:. too •
roduced in Example 2
rational function of S.
By the behavior of a system, we mean aH responses, including the state, of
book are exclusively
the system. If the system is a network, we mean the voltage and current of every
s of rational functions.
branch of the network. Hence from the state at t o, the state at all t> t o can
be computed. If the state at ti> to is known, we need u[l" ,.,), rather than
u[to.ool' in computing the behavior of the system for t '2:. ti' Hence the state at
tI summarizes the essentia! information about the past input Ql( - ,.,.'¡) which is
nite (zero or nonzero)
\. rational matrix G(s) needed in determining the future behavior ofthe system. We note that different
;onstant matrix. G(s) input ui_«>.t¡),i=1,2, ,may yield the same state at ti' In this case, even
though ui _ ,.,.t¡), i = 1, 2, , are different, their effects on the future behavior of
• the system are identical.
We give sorne examples to illustrate the concept of state.
/(s2- s +2) is proper
9(5) = N(s)jD(s), 9(s) is Example 1
proper if and only if
llynomial. A rational Consider the network shown in Figure 3-6.. It is well known that if the initial
er. A rational matrix , current through the inductor and the initialvóltage across the capacitor are
tly proper.
noises wifl be greatly ,5 The exceptions are transfer functions ,of sorne transducers such as tachorneters and accelera­
als. Hence improper rneters. See Reference s46. '
~'-'--'._' _.. _._.-.~-- ~-_._._._--~ ~._ _._. __ _.. _. _._._._- _'
.~~~~~~~~~~~~~~--~~~~
••• _ •••••••• •• ~. __ ••• _ ••• - _ . _ •••.••• - -_._.~_ •••• __ •• _ •••• _.~-- - _ • • • • • • _. -.-.'-~'--"-"-'-'--"" - - -••• _ ••• _ . _ •• , __ o . _ . _•• •• __ •• ~'._~

84 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

VOltag:
source

Figure 3-6
E3i
~ ~
A network.
Y
l
~_-_
output after t ~ t o ex(
minable. From (3-24

Taking the derivative

y(t) = -2e

known, then for any driving voltage the behavior of the network can be deter­ which, together with f
mined uniquely. Benee the inductor current and the capacitor voltage qualify
as the state of the network. •
Solving for Cl and C2 I
Example 2
We consider again the network in Example 1. The transfer function from
u to y of the network can be easily found as
Bence if the network
A 222
g(5)=(5+1)(5+2)=5+1- 5+2 y(t) = (2

Bence the impulse response of the network is -j

g(t) =2e- 1 -2e- 21 (3-23)


We see that if y(to) an,
which is the inverse Laplace transform of the transfer function g(5). Now we determined even if thE
apply an input U[lo,oo) to the network. If the network is relaxed at to, the output y(to) and y(to) qualifie~
is given by' . also qualifies as the st

r g(t-L)U(L)dL
l

y(t)= for t ~to In this example, w


Jl
o
(- is summari2
00, t o)
If the network is not relaxed at to, the output must be computed from concept of state is ver:

y(t)= Loo g(t-L)U(L)dL= roo g(t-L)U(L)dL + L g(t-L)U(L)dL Example 3


Consider the network
for t ~ to (3-24)
voltages are known, tl
because the input that had been applied before t o may stHI have sorne effect on for any applied input.
the output after t o through the energy stored in the capacitor and inductor. as the state of the net...
We consider now the effect on Y[lo, OO) due to the unknown input u(_ 00,10)' From
(3-23) we have
6 d
-d J' g(t - r)u(r) dr = g(t ­
t 21 2t t 'o
roo g(t-L)U(L)dL=2e- roo e u(L)dL-2e-
l
roo e u(L)dL

(3-25)
fx:
r----'\AII~------JE
R
for t ~ to, where
----=t
XI

Cl ~ Looo etu(L) dL and C2 ~ LOoo e 2t u(L) dL


~

Note that Cl and C2 are independent of t. Bence if C l and C2 are known, the Figure 3-7 A netwo
••. - • • • • ~ .. _. __ o • • _

THE STATE-VARIABLE DESCRIPTION 85

output after t"2: t o excited by the unknown input u{ _ 00.10) is completely deter­
minable. From (3-24) and (3-25) we have
y(t o)=2e- IO c l -2e- Ztoc 2 (3-26)

Taking the derivative of (3-24) with respect to t yields6


1
0

letwork can be deter­


y(t) = -2e-lcl +4e- Zlc 2 +g(O)u(t) +

which, together with g(O) = o, implies that


1lO
-o g(t-"l:)u("l:)d"l:
t

acitor voltage qualify


• Solving for Cl and Cz
y(to) = -2e- IO c l +4e- 2IO cz
from (3-26) and (3-27), we obtain
(3-27)

Cl = 0.5e'O (2y(to) + y(to))

ansfer function from Cz = 0.5e 210 (y(to) + y(to))

Hence if the network is not relaxed at t o, the output y(t) is given by


y(t) = (2y(t o) + y(to))e-(t-t o) - (y(t o) + y(to))e- 2 (t-to)

r g(t -"l:)u("l:)d"l:
t
+ for t "2:t o
(3-23)
t
We see that if y(to) and y(to) are known, the output after t "2:t o can be uniquely
nction 9(S). Now we determined even if the network is not relaxed at too Hence the set of numbers
laxed at to, the output y(to) and y(to) qualifies as the state of the network at too Clearly, the set {Cl' C2}
also qualifies as the state of the network. •

In this example, we see that the effect of the input over the infinite interval
(- co, t o) is summarized into two numbers {y(to), y(to)} or {Cl> C2}; hence the
omputed froro concept of state is very efficient and powerful.

g(t - "l: )U("l:) d"l: Example 3


Consider the network shown in Figure 3-7. It is clear that if aH the capacitor
for t "2: t o (3-24)
voltages are known, then the behavior of the network is uniquely determinable
Hhave sorne effect on for any applied input. Hence the set oI capacitor voltages Xl' X2, and X3 qualifies
pacitor and inductor. as the state of the network. Let us examine the network more carefully. If we
inputu(_oo"o)' From

'10
6 d
d). r g(t-T)u(r)dT=g(t-T)U(T) I
t

+ J' -g(t-T)U(T)dT
a
t lo t=t roat

(3-25 )

Lnd C2 are known, the Figure 3 c 7 A network with a loop that consists of capacitors only.
• _. _ •• _ _ .• ' •• _ •••• _ •• •.•••• •• •• _ _ _... • •• _ _ •••• _ .•• •• _._. _ _ ••• •••..• _._ ••• _ . _ . _ _. _ • • • • •.•• _ . ••• ._0 __ - '0_" • __ ._._•. , __ .... • "'._. __ •. • __ •• ... • . __ . • •••• _

~
_
.
~
~
~
.
_
~
_
~
~
_
,
~
86 MATHEMATICAL DESCR1PT10NS OF SYSTEMS

apply the Kirchhoff voltage law to the loop that consists of three capacitors, dynamical equation.
we have Xl(t) +xz(t) +X3(t)=Ü for all t. lt implies that if any two of Xl,X Z, the form
and X 3 are known, the third one is also known. Consequently, if any two of the
capacitor voltages are known, the behavior of the network is uniquely deter­ X(l
minable for any input applied thereafter. In other words, any two of the three y(t
capacitor voltages qualify as the state. If all the three capacitor voltages are
chosen as the state, then there is a redundancy. In choosing the state of a or, more explicitly,
system, it is desirable to choose a state that consists of the least number of xl(t) =J
variables. How to pick the state with the least number of variables for general xz(t) = J
RLC networks will be studied in the next section. •
xnCt)=J
Example 4
Yl(t) =6
A unit-time-delay system is a device whose output y(t) is equal to u(t -1) for
yzCt) = 6
all t. For this system, in order to determine Y[lo,CO) uniquely from U[lo,CO)' we
need the information U[lo-l,lo)' Hence the information U[lo-l,lo) is qualified to
yq(t) =9
be called the state of the system at time too I
where x = [Xl Xz ..
From these examples, we may have the following observations concerning and u = [Ul U z
the state of a system. First, the choice of the state is not unique. For the state x are real-valued
network shown in Figure 3-6, the state may be chosen as the inductor current for (3-28) to qualify a
and the capacitor voltage, or chosen as y(to) and y(to) or Cl and Cz. For the initial state x(to) and :
network shown in Figure 3-7, any two of the three capacitor voltages can be sufficient condition for
chosen as the state. Different analyses often lead to different choices of state. initial state is that h¡ an
Second, the state chosen in Example 1 is associated with physical quantities, see References 24,77, al
whereas in Example 2 the state is introduced for mathematical necessity. that the solution can 1
Hence, the state of a system is an auxiliary quantity that may or may not be the state at to, as expe<
easily interpretable in physical terms. Finally the state at each instant may and is caBed a state eq
consist of only a finite set of numbers, as in Examples 1,2, and 3, or consist of an an output equation. N(
infinite set of numbers, as in Example 4. Note that there is an infinite number of equation in the form 01
points between [to -1, to); hence the state of this example consists of an infinite ledge ofx(t) and u(t) su
set of numbers. The state space of
In this book we study only the class of systems whose states may be chosen set of equation (3-28) ¡¡
to consist of a finite number of variables. The state of a system can then be
represented by a finite-dimensional column vector x, called the state vector. linearity. We use tI:
The components of x are called state variables. The linear space in which the
state vector ranges is denoted by L. Since state variables are usually real­
valued, and since we study only systems with a finite number of state variables, to denote that the statt:
the state spaces we encounter in this book are the familiar finite-dimensional state x(t), for t '2: to, al
real vector space(IW, IR). output pair of a system

Dynamical equations. In addition to the input and output of a system we Definition 3-7
have now the state of the system. The state at time t o is, by definition, the
required information at t o that, togetherwith input U[lo,CO)' determines uniquely A system is said to bell .
the behavior (output and state) ofthe system for all t '2: too The set 01 equations
that describes the unique relations between the input, output, and state is caUed a
----,._._---~ _.
... __.__ .. _-_._.-._--- -- ._ .•.. __ ...._---- ---- _._'-.-'."- ----.--_.•.•.. -.- ....._-_.-. __.-.-_ .. ~.,-_ ..-----_.- --.. ,-----_ .. _-_.-._­

THE ST ATE-VARIABLE DESCR¡PTION 87

.ts of three capacitors, dynamical equation. In this book, we study only the dynamical equations of
t if any two of Xl> Xz, the form
lentiy, ifany two ofthe
i(t) = h(x(t), u(t), t) (state equation) (3-28a)
ork is uniquely deter­
s, any two of the three y(t) = g(x(t), u(t), t) (output equation) (3-28b)
capacitor voltages are
or, more explicitly,
100sing the state of a
,f the least number of x¡(t) =h¡(x¡(t), xz(t), ... , xn(t), u¡(t), uz(t),.·., up(t), t)
)f variables for general xz(t)=hz(x¡(t), xz(t),···, xn(t), u¡(t), U2(t),···, up(t), t)
I (3-29a)
Xn(t) = hn(x¡ (t), xz(t), ... , xn(t), u¡(t), uz(t), ... , up(t), t)

y¡(t) =g¡(x¡(t), xz(t),···, xn(t), u¡(t), uz(t),···, up(t), t)


is equal to u(t -1) for yz(t) =gz(x¡(t), xz(t), ... , xn(t), u¡(t), uz(t), ... , up(t), t)
iquely from u[to,oo)' we (3-29b)
U[lo-¡,lo) is qualified to yq(t) =gq(x¡(t), xz(t), ... , xn(t), u¡(t), uz(t), ... , up(t), t)
I
where x = [x¡ Xz ... xnJ' is the state, y = [y¡ Yz '" yqJ' is the output,
lservations concerning and u = [u¡ Uz .. . upJ' is the input. The input u, the output y, and the
not unique. For the state x are real-valued vector functions of t defined over ( - ro, (0). In order
IS the inductor current for (3-28) to qualify as a dynamical equation, we must assume that for any
)r c¡ and Cz. For the initial state x(t o) and any given u, Equation (3-28) has a unique solution. A
acitor voltages can be sufficient condition for (3-28) to have a unique solution for a given u and a given
ferent choices of state. initial state is thath¡ and ahJaxjare continuous functions oft for i,j = 1, 2, ... , n;
th physical quantities, see References 24, 77, and 92. lf a uniquesolution exists in (3 c 28), it can be shown
athematical necessity. that the solution can be solved in terms of x(to) and U[lo,l)' Hence x serves as
at may or may not be the state at to, as expected. Equation (3-28a) governs the behavior of the state
e at each instant may and is called a state equation. Equation (3-28b) gives the output and is called
, and 3, or consist of an an output equation. Note that there is no loss of generality in writing the output
is an infinite number of equation in the form of (3-28b), because by the definition ofthe state, the know­
e consists of an infinite ledge ofx(t) and u(t) suffices to determine y(t).
The state space of (3-28) is an n-dimensional real vector space; hence the
e states may be chosen set of equation (3-28) is called an n-dimensional dynamical equation.
r a system can then be
~alled the state vector. linearity. We use the notation
Lear space in which the
(U[lo,oo), x(to)} -+ {x[to, 00)' Y[to, oo)}
ables are usually real­
nber of state variables, to denote that the state x(t o) and the input U[lo,oo) excite the output y(t) and the
iliar finite-dimensional state x(t), for t ~to, and call it an input-state-output pair. An input-state­
output pair of a system is called admissible ifthe system can generate such a pairo

l output of a system we Definition 3-7


~o is, by definition; the
A system is said to be linear if and only if for any two admissible pairs
o), determines uniquely

o. The set of equations {x¡(t o), utto,oo)} -+ {xtto,oo), yttó,oo)}


Jut, and state is caUed a {XZ(t o), ufro,oo)} -+ {xfro,oo), Yfto,oo)}
88 MATHEMATlCAL DESCRIPTlONS OF SYSTEMS

and any real numbers al and a2 , the following pair The responses due to {x
2 exclusively by the non¡
{alxl(tO) +a 2x (t o), alutto,ro) +a 2 u fro,ro)}
are caBed zero-state re~
----> {alxtto,ro) +a 2x fro,ro), alytto,ro) +a 2 Yfto,ro)} (3-30)
Equation (3-31) follows
is also admissible. Otherwise, the system is said to be nonlinear. • xl(to) =x(t o
Similar to Equations (3-4) to (3-6), if al = a 2 = 1, the relationship in (3-30) is
Hence for linear syster
caBed the property of additivity; if a 2 = 0, it is caBed the property of homogeneity.
zero-state responses in
The combination of these two properties is caBed the principie of superposition.
in the previous section I
In this definition, the superposition property must hold not only at the output
If a system is linear,
but also at aB state variables; it must hold for zero initial state as weB as non­
zero initial state. Hence this definition is much more stringent than the one in h(x(t), u(t), t) = j.
Definition 3-2. Consequently, a system may not be linear according to Defini­
where A, B, e, and E :
tion 3-7 but may be linear according to Definition 3-2. For example, the system
(Problem 3-36). Heno
in Figure 3-S (a) has a nonlinear capacitor C. Ifthe voltage across the capacitor
form
is zero at t o = 0, it will remain to be zero for aB t;::: t o no matter what input
waveform is applied. Hence, as far as the behavior at the input and output E: x(
terminals is concerned, the nonlinear capacitor can be disregarded. Hence the
yt
system is linear according to Definition 3-2 but not linear according to Definition
3-7. Consider the network in Figure 3-S(b) with a nonlinear capacitor C and a A sufficient condition I
nonlinear inductor L. Because the L-C loop is in a series connection with the of A(') be a continuous
current source, its behavior will not transmit to the output y. Hence the system the entries of B(' ), C(' ), ~
in Figure 3-S(b) is linear according to Definition 3-2 but not linear according to Since the values of A(
Definition 3-7. equation E in (3-32) is 1
equation.
We discuss now the implication of Definition 3-7. If al =a 2 = 1 and if
xl(t O)= -x 2 (t O) and utto,ro) = -ufro,ro) Time invariance. 1
then the system is said
then the linearity implies that {O,O}---->{O[to,ro),O[to,ro)}' Hence a necessary
°
condition for a system to be linear is that if x(to) = and u[to, ro) == 0, then the
responses of the system are identically zero. A very important property of any
lowing. Let Q" be the

Definition 3-8
linear system is that the responses of the system can be decomposed into two
parts, as A system is said to be t
Responses due to {x(t o), IDI[to, ro)}
= responses due to {x(t o), O} +responses due to {O, u[to, ro)} (3-31 )
and any real number a,
e {Q,,~
u In
is also admissible. Ott

Ll t
L .

.2 H 3F
T Y
In words, for time-i
the waveforms of excita
will always be the same

~) ~)
1 For linear time-invariaJ
are independent of time

Figure 3-8 Two nonlinear systems which are linear according to Definition 3-2.
•••• _ _ • ••• .•• _ _ ,_u__ , ,_"" ,_ _ ,,
----,---_..- ..•- .. __

.
~
~
~
_
~
.
_
~
~
,
~
.
_
,
~
_-----~_. ._~------_._---_._--_.~--~.,~---~-----.-

THE STATE-VARIABLE DESCRIPTION 89

The responses due to {x(to), O} are caBed zero-input responses; they are generated
exclusively by the nonzero initial state x(to). The responses due to {O, u¡to.co)}
are called zero-state responses; they are excited exclusively by the input u[to.oo)'
x2 yfro. col} (3-30)
Equation (3-31) foBows directly from (3-30) if we choose IX 1 = IX 2 = 1 and
IOnlinear. • xl(to) =x(to) u 1 =0 x 2 (t O) =0 ufro.CO) =U[lo.oo)
relationship in (3-30) is
Hence for linear systems, we may consider the zero-input responses and the
roperty of homogeneity.
zero-state responses independently. The input-output description discussed
'inciple of superposition.
in the previous section describes only the zero-state responses of linear systems.
not only at the output
If a system is linear, the h and g in (3-29) become linear functions ofx and u as
ial state as well as non­
:ringent than the one in h(x(t), u(t), t) = A(t)x(t) + B(t)u(t), g(x(t), u(t), t) =C(t)x(t) + E(t)u(t)
lar according to Defini­
where A, B, e, and E are, respectively, n x n, n x p, q x n, and q x p matrices
~or example, the system
(Problem 3-36). Hence an n-dimensional linear dynamical equation is of the
1ge across the capacitor
form
, no matter what input
t the input and output E: x(t) = A(t)x(t) + B(t)u(t) (state equation) (3-32a)
lisregarded. Hence the
y(t) =C(t)x(t) + E(t)u(t) (output equation) (3-32b)
according to Definition
inear capacitor e and a A sufficient condition for (3-32) to have a unique solution is that every entry
ies connection with the of A(') be a continuous function of t defined over (- 00, 00). For convenience,
ut y. Hence the system the entries of B('), C('), and E(') are also assumed to be continuous in ( - 00, 00).
not linear according to Since the values of A('), B('),C('), and E(') change with time, the dynamical
equation E in (3-32) is more suggestively caBed a linear time-varying dynamical
equation.
If IX I =0: 2 = 1 and if
- ufro. col Time invariance. Ifthe characteristics ofa system do not change with time,
then the system is said to be time invariant. We define it formally in the fol­
" Hence a necessary
lowing. Let Q" be the shifting operator defined in Figure 3-5.
md U[to. 00) == O, then ihe
í'ortant property of any
Definition 3-8
~ decomposed into two
A system is said to be time invariant if and only if for any admissible pair

(3-31 )
and any real number IX, the pair
{Q"x(t o), Q"u[to. col} -+ {Q"x¡to. co)' Q"y[co. co)}
u In
is also admissible. Otherwise, the system is said to be time varying. •

2H
T
1 In words, for time-invariant systems, if the initial states are the same and
the waveforms of excitations are the same, then the waveform of the responseS
will always be the same no matter at what instant the excitations are applied.

-------
3F
T For linear time-invariant systems, the matrices A('), B('), C('), and E(') in (3-32)
are independent of time, and the equation reduces to
FE: x(t) = Ax(t) +Bu(t) (3-33a)
(b)
ling lo Definilion 3-2. y(t) = Cx(t) + Eu(t) (3-33b)
90 MATHEMATlCAL DESCRIPTlONS OF SYSTEMS

where A, B, e and E are, respectively, n x n, n x p, q x n, and q x p real constant (sI - A) always exists.
matrices. This set of equations is called a linear time-invariant n-dimensional at t =O-then (3-35b)
dynamical equation and is denoted by FE (fixed equation). For linear time­
invariant systems, the responses are independent of the initial time; hence it is
always assumed without loss of generality that t o =0. The time interval of A comparison of this
interest then becomes [0, ro).
The state space L of E or FE is an n-dimensional real vector space (IR", IR),
Hence we can think of the n x n matrix A as a linear operator which maps L Hence if a linear tirr.
into L. As mentioned in the preceding chapter, it is very convenient to intro­ (;(s) and the dynamil
°
duce the set of the orthonormal vectors {01' 2 " " , 0n}, where 0i is an n x 1 related by (3-36). We
column vector with 1 at its ith component, and zero elsewhere, as the basis of
the state space. In doing so, we may also think of the matrix A as representing a
linear operator with respect to this orthonormal basis. Hence, unless other­
wise stated, the basis of the state space of E nr FE is assumed to be the set of the Every entry ofthe adj(
orthonormal vectors the degree of the deter
rational matrix. If E
rational matrix. Note
The dynamical equations in (3-29), (3-32), and (3-33) can be solved, given
x(to) and u('), in the direction of positive time or in the direction of negative
time. Clearly we are interested in only the direction of positive time. In the Analog and digité
positive-time direction, the input u(t 1) affects only the future responses, the equations. As willl
responses for t ~ t 1; it does not affect the past responses. Hence the dynamical number of state variab:
equations are aH causal. ical equations. We sh(
dyriamical equation ca .
Transfer-function matrix. In the stuoy oflinear time-invariantdynamical
connecting integrators
equations, we may also apply the Laplace transformo Taking the Laplace grator, 8 summer, and ¡
transform of FE and assuming x(O) =x o, we obtain puter, and their functio
si(s) -x o = AX(s) +BU(s) (3-34a) block diagram of analo
time-invariant dynamil
y(s) =CX(s) +Eii(s) (3-34b)

where the circumflex over a variable denotes the Laplace transform of the
same variable; for example,

From (3-34), we have


B In practice, pure dilTerentia
x(s) = (sI - A)-1 XO + (sI - A)-1Bu(s) (3-35a) noises. On the other hand.

y(s) =C(sI -A)-1 xo +C(sI -A)-lBu(s) + Eu(s) (3-35b)

They are algebraic equations. Ifx o and u are known, x(s) and y(s) can be com­
puted from (3-35). Note that the determinant of (sI - A) is different from zero
(the zero ofthe field of rational functions of S)7; htmce, the inverse ofthe matrix

<al
7 That is, the determinant of (sI - A) is not identically equal to zero. Note that det (si - Al = O Figure 3-9 Analog com
for sorne s is permitted. or attenuator.
~-'------'------'''-----,-------------- ----­ -------_._-_.

THE STATE-VAR1ABLE DESCR1PTlON 91

nd q x p real constant (si - A) always exists. If the initial state X o is O-that is, the system is relaxed
variant n-dimensional at 1 =O-then (3-35b) reduces to
)n). For linear time­ Y(s) = [C(sl - A)-l B + E]u(s)
nitial time; hence it is
The time interval of A comparison of this equation with (3-21) yields
G(s)=C(sI-A)-lB +E (3-36)
l vector space (IR", Iln
lerator which maps L Hence if a linear time-invariant system is described by the transfer matrix
y convenient to intro­ G(s) and the dynamical equation {A, B, C, E}, the two descriptions must be
, where ni is an n x 1 related by (3-36). We write (3-36) as
:where, as the basis of A 1
rix A as representing a G(s) = d (1 ) C[,Adj (sI - A)]B + E
et s -A
Hence, unless other­
/'led to be the set 01 the Every entry of the adjoint of (si - A) is a polynomial of degree strictly less than
the degree ofthe determinant of(sI - A); hence C(sI -A)-lB is a strictly proper
rational matrix. If E is a nonzero matrix, then C(sl - A)-l B + E is a proper
rational matrix. Note that we have
I can be solved, given G(oo) =E (3-37)
direction of negative
positive time. In the Analog and digital computer simulations of linear dynamical
future responses, the equations. As will be illustrated in the next section, systems that have a finite
Hence the dynamical number of state variables can always be described by finite-dimensional dynam­
ical equations. We show in this subsection that every finite-dimensionallinear
dynamical equation can be readily simulated on an analog computer by ihter­
e-invariant dynamical connecting integrators, summers, and amplifiers (or attenuators). The inte­
Taking the Laplace grator,8 summer, and amplifier are three basic components of an analog com­
puter, and their functions are illustrated in Figure 3-9. We give in Figure 3-10 a
(3-34a) block diagram of analog computer connections ofthefollowing two-dimensional
time-invariant dynamical equation
(3-34b)

lace transform of the


aa22_-I[xX2(t)(t)] +[bb
12 1 ll b12][U1(t)]
b 22 U2(t)
21

~::][::~~G +[:::

8 In practice, pure difTerentiators are not used for the reason that they wil\ amplify high-frequency
(3-35a) noises. On the other hand, integrators wil\ smooth or suppress noises.

tEu(s) (3-35b) Xl

;) and y(s) can be com­


) is different from zero
e inverse of the matrix

(a) (b) (e)

Note that det (sI - A) = O Figure 3-9 Analog eomputer components. (a) Integrator. (b) Summer. (e) Amplifier
or attenuator.
92 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

can be easily preparel


directly from a dynam
We give a simple e
sider the dynamical eq

U¡ Xl

Find the output y and


condition Xl(O)= 1, X2('
The CSMP input s

'-----------~ e21I-----------J
DYNAMIC
PARAMETER
'-------------+1 e22 1--- ------'
X1DT=
Figure 3-10 Block diagram oftwo-dimensional dynamical equation.
X2DT=
X1 =IN
X2=IN
Note that for a two-dimensional dynamical equation, we need two integrators. Y=X1­
The output 01 every integrator ean be assigned as a state variable. We see that TIMER DELT,
even for a two-dimensional dynamical equation, the wiring of the block diagram PRTPLT X1,Y
is complicated; hence, for the general case, we usually use a matrix block STOP
diagram. The matrix block diagram of the dynamical equation E is shown in END
Figure 3-11. If E is n-dimensional, the integration block in Figure 3-11 consists
of n integrators. The matrix E represents the direet transmission part from the The first part of the
input u to the output y. Ifthe matrix A is a zero matrix, then there is no "feed­ step size. "FINTIM" i
back" in the block diagram. For a discussion of analog computer simulations, interval in which the n
see, e.g., Reference S46. result, DELT is usually
Dynamical equations can be readily simulated on a digital computer. print out every comput
There are many specialized subroutines for solving dynamical equations, such larger than DELT. In
as MIDAS (Modified Integration Digital Analog Simulator), MIMIC (an DELT and OUTDEL.
improved version of MIDAS), CSMP (Continuous System M odeling Program), as well as plot the outp
TELSIM (Teletype Simulator), and others; see Reference 26. These programs Comparisons of ana
tions are in order. On
are limited to a range,
range, sorne componen
of the sim ulation will t
equations must be pro]
u y carried out by cut and t
+ of numbers which a dig
magnitude scaling genel
is often limited to 0.1 1
precision of eight or m
computer simulation is
Figure 3-11 Matrix block diagram of the dynamical equation E. putero The generation
THE STATE-VARIABLE DESCRIPTION 93

can be easily prepared from the block diagram of a dynamical equation or


directly from a dynamical equation.
We give a simple example of the application of System/360 CSMP. Con­
sider the dynamical equation

[Xl]
X2 =
[2 -lJ[XIJ +[O ]
1 5_ X2 1.5 u

y = [ 1 0.6J [ :J
Find the output y and state variable Xl from O to 20 seconds due to the initial
condition Xl (O) = 1, X2(0) = O, and a unit-step-function input.
The CSMP input statements for this problem are listed in the following:

DYNAMIC
PARAMETER U = 1.0
X1 DT=2.0*X1-X2
X2DT=X1 +5.0*X2+1.5 *U
~quation.
X1 = INTGR(1.0,X1 DT)
X2 = INTG R(0.0,X2DT)
e need two integrators. Y=X1 +0.6 ·X2
. variable. We see that TIMER DELT =0.001 ,FINTIM =20.0,OUTDEL=0.10
Ilg ofthe blockdiagram PRTPLT X1,Y
ly use a matrix block STOP
;:quation E is shown in END
in Figure 3-11 consists
nsmission part from the The first part ofthe program is self-explanatory. "DELT' is the integration
, then there is no "feed­ step size. "FINTIM" is the final time of computation. "OUTDEL" is the
computer simulations, interval in which the responses will be printed. In order to have an accurate
result, DELT is usually chosen to be very smal!. It is, however, unnecessary to
n a digital computer. print out every computed result; therefore the printout interval is chosen much
amical equations, such larger than DELT. In employing CSMP the user has to decide the sizes of
mulator), MIMIC (an DELT and OUTDEL. For this program we have asked the computer to print
:m Modeling Program), as well as plot the output y and the state variable X l'
~e 26. These programs
Comparisons of analog computer simulations and digital computer simula­
tions are in order. On analog computer simulations, the magnitudes of signals
are limited to a range, typically ± 10 volts. Ir the magnitudes go over the
range, sorne components of the analog computer will saturate and the result
of the simulation will be erroneous. Hence on analog computer simulations,
equations must be properly scaled. This is a difficult problem and usually is
carried out by cut and try. On digital computer simulations, because the range
of numbers which a digital computer can manage is very large, the problem of
magnitude scaling generally does not arise. The accuracy of an analog computer
is often limited to 0.1 percent of its full scale; a digital computer may have a
predsion of eíght or more decimal digits. Therefore the result from a digital
computer simulation is much moreaccurate than íhat from an analog conb
ion E. puter. The generation of nonlinear functions is easier on a digital computer
94 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

than on an analog computer. However, the interaction between an analog y =0


computer and the user is very good. By this we mean that the parameters of a
simulation can be easily adjusted, and the consequence of the adjustments can
~
be immediately observed. The interaction between a large-frame general­
purpose digital computer and the user is generally not very satisfactory, but
this interaction has been improving in recent years with the introduction of
time sharing and remote terminals. With the increasing speed of computation (a)
and graphic display, the interaction between the user and a special-purpose
digital computer or a general-purpose mid-frame digital computer is now very Figure 3-12 A mechó
good. Furthermore, digital computers are much more versatile; they can be Force
used to carry out the designo Hence digital computer simulations are gaining
more popularity in recent years. t . Stallc

---+o·-t--- v,
3-4 Examples
-----.::------. Coulol11 b
In this section we shal1 give sorne examples to illustrate how the input-output
descriptions and state-variable descriptions of linear systems are developed.
(a)
A system generally consists ofmany subsystems or components. For example,
the network in Figure 3-6 consists of three components: one resistor, one Figure 3-13 (a) Stati(
inductor and one capacitor. Ir any component of the system is nonlinear or
time varying, then the overall system is nonlinear or time varying. Hence in • Force
order to obtain a linear time-invariant model for a physical system, every com­
ponent of the system must be modeled as a linear time-invariant element.
Stricdy speaking, no physical system is linear and time invariant. A tele­
vision set, an automobile, or a communication satel1ite cannot function forever;
its performance will deteriorate with time because of aging or other factors. Y2
However, if the changes of characteristics are very smal1 in the time interval of
interest-say, one year-then these physical systems can be considered as time
invariant. Hence over finite time intervals, a great number of physical systems
can be modeled by time-invariant systems.
A necessary condition for a system to be linear is that for any admissible Figure 3-14 Charact
pair {O, u¡o.ro)} ---+ {x¡o.ro), Y[O,ro)}, the pair {O, au[o.ro)} ---+ {ax¡o.OO)' ay[o.GO)} for any
a, even very very large, is also admissible. For any physical system, ifthe applied
signal is larger than certain limit, the system will burn out or saturate. Hence fricUon is clearly not a l
no physical system is linear according to Definition 3-2 or 3-7. However, we neglect the static a
linear models are often used in practice to represent physical systems. This is friction. Let k¡ be thl
possible because most physical systems are designed to operate only in a certain f isgiven by f=k¡dy/,
operational ranges. Limited to these ranges, physical systems can often be 3-14. It is a nonlinear
approximated by linear models. This is accomplished by.linearlization or (y¡, yz) as shown, then I
simplification, as will be discussed in the following examples. constant. Hence the lin
and simplification.
Example 1 Now we shal1 de vele
external force u (input)
Consider the mechanical system shown in Figure 3-12. The friction force Newton's law yields
between the floorand the mass generaUy consists of three distinct parts: static
friction, Coillomb friction, and viscous friction as shown in Figure 3-13. The
EXAMPLES 95

>n between an analog y=o


Lat the parameters of a ~
)f the adjustments can ti (s) y (s)
l large-frame general-
very satisfactory, but
th the introduction of
speed of computation (a) (b)
and a special-purpose
Figure 3-12 A mechanical system.
computer is now very
versatile; they can be Force Force
imulations are gaining
Viscous friction
Sta tic

----+-o-+----- Velocity ---~iL------ Velocity

Coulomb
how the input-output
ystems are developed. (a) (b)
lonents. For example,
Figure 3-13 (a) Static and Coulomb friction. (b) Viscous friction.
:nts: one resistor, one
system is nonlinear or
me varying. Hence in Force

ical system, every com­ Break


invariant element. roo
ime invariant. A tele­
annot function forever; ---+-----,.¡.<::------1e---_ Displacement
aging or other factors.
1 in the time interval of
n. be considered as time
ber of physical systems

hat for any admissible Figure 3-14 Characteristic of the spring.


xx[O, 00)' IXY[o. oo)} for any
:al system, ifthe applied
)lit or saturate. Hence friction is clearly not a linear function of the velocity. To simplify the analysis,
3-2 or 3-7. However, we neglect the static and Coulomb frictions and consider only the viscous
ysical systems. This is friction. Let k¡ be the viscous friction coefficient. Then the friction force
Iperate only in a certain f is given by f = k1dy/dt. The characteristic of the spring is shown in Figure
1 systems can often be 3-14. lt is a nonlinear element. However if the displacement is limited to
~d by linearlization or (Yl, Y2) as shown, then the spring force is equal to k 2y, where k 2 is the spring
nples. constant. Hence the linear model in Figure3-12(a) is obtained by linearization
and simplification.
Now we shall develop the input-output description of the system from the
external force u (input) to the displacement y (output). The application of
12. The fdction force Newton's law yields
ree distinct parts: static
vn in Figure 3-13. The
96 MATHEMAT1CAL DESCRIPTIONS OF SYSTEMS

Taking the Laplace transform and assuming the zero initial condition, we obtain For example, if the iJ
cart is moved to the
u(s) = (ms 2 +k 1 s +k 2 )y(s)
pendulum back to thf
Hence the input-output description in the frequency domain of the system is as a model of a space
Let H and V be, rl
A 1 A
y(s) = 2 k k u(s) the cart to the pendul
ms + lS + 2
linear movements yiel,
If m = 1, k 1 = 3, k 2 = 2, then the impulse response of the system is

g(t) = ~ [S2 +~s +2] = ~ [(S ~ 1) - ~2)] = e-t - e- 2t


-1 -1 (S

and the input and output are related by H=

y(t) = J>(t - r)u(r) dr and mg- V=

We next derive the dynamical-equation description of the system. Let the


The application of Ne'
displacement and the velocity of the mass m be the state variables; that is, yields
Xl = y, X2 = y. Then we have

or [Xl]
. =
X2
[0-;;¡ - 1][X1] +[0]
-
k2 k1
m X2
1
;;; u
These are nonlinear eql
the pendulum at the Vf
small. U nder this ass
cos e= 1. By retainin¡
y=[l O] [:J terms with e2 , ip, ee, a;

This is a state-variabIe description of the system. By choosing a different set


of state variables, we may obtain a different state-variable description of the
system. I
which imply
Example 2

Consider a cart with an inverted pendulum hinged on top of it as shown in and


Figure 3-15. For simplicity, the cart and the pendulum are étssumed to me'!'"
in only one plane, and the friction, the mass of the stick, and the gusts of wind With these linearized eq
are neglected. The problem is to maintain the pendulum at the vertical position. variable descriptions of
to (3-38) and (3-39) yiel<
e m

and

From these two equatioJ


function 90..(s) from u to

~Y-----1
Figure 3-15 A cart with an inverted pendulum. and A

gOll
EXAMPLES 97

11 condition, we obtain For example, if the inverted pendulum is falling in the direction shown, the
cart is moved to the right and exerts a force, through the hinge, to push the
pendulum back to the vertical position. This simple mechanism can be used
main of the system is as a model of a space booster on takeoff.
Let H and V be, respectively, the horizontal and vertical forces exerted by
the cart to the pendulum as shown. The application of Newton's law to the
linear movements yields
ystem is d2
M---l=u-H

dt 2

J
- =e -1 -e -2/
) d
2
J i . 2
H =m dt 2 (y +1 sin e) =mji +ml cos e!J -mi sin e (e)

d2 . •
and mg - V = m dt 2 (l cos e) = mi [ - sin e (j - cos e (efJ

)f the system. Let the The application of Newton's law to the rotational movement of the pendulum
tate variables; that is, yields

m1 2(j =mgl sin e + VI sin e- HI cos e


These are nonlinear equations. Since the purpose ofthe problem is to maintain
the pendulum at the vertical position, it is reasonable to assume e and é to be
u small. Under this assumption, we may use the approximation sin e= e and
cos e= 1. By retaining only the linear terms in e and é, that is, dropping the
terms with e 2, é2, ee, and ee', we obtain V = mg and

Mji = u -mji -mle'


hoosing a different set
ml 2lf =mgle + mgle - (mji + mllJ)1
1ble description of the
• which imply

(M+m)ji+mllJ=u (3-38)

top of it as shown in and 2¡{J - 2ge + ); = O (3-39)


1 are assumed to move
, and the gusts of wind With these linearized equations, we can now derive the input-output and state­
at the vertical position. variable descriptions of the system. The application of the Laplace transform
to (3-38) and (3-39) yields, assuming zero initial conditions,
y
(M + m)s2 (s) + mls 28(s) = D(s)
and (21s 2 - 2g)8(s) + S2 y(S) = O
From'these two equations, the transfer function gyu(s) from u to y and the transfer
function gouCs) from u to e can be readily obtained as
A 21s 2 - 2g
(3-40)
gy,,(s) s2[(2M +m)/s 2 - 2g(M +m)]

-1
and (3-41 )
gou(s) = (2M +m)/s 2 - 2g(M +m)
98 MATHEMATlCAL DESCRIPTlONS OF SYSTEMS

To develop a dynamical equation, we define the state variables as Xl = y, and


X2 e,
= y, X3 = and X4 = e.
From (3-38) and (3-39), we can solve ji and (J as
The term kVa is the f(
.. 2gm 2 maintain their velocil
Y= - 2M +m e +2M +m u
that Yi,i+ ¡(t) and v¡(t)
we have, for i = 1, 2,3,
(J=2g(M+m)e_ 1 u
and
(2M +m)l (2M +m)l
From these two equations and the definition of Xi' the state-variable equation From (3-46) and (3-44
description of the system can be readily obtained as

Xl O O O Xl O
-2mg 2
X2 O O O X2 These equations can b
2M+m 2M+m u (3-42)
O O
+ O -k
X3 O X3 VI (t) O O
2g(M +m) -1 mi
X4 O O O X4 y 12(t) 1 O -1
(2M +m)l (2M +m)l
-k
V2(t) O O
y =[1 O O O]x m2
ydt) O O 1
Note that Equations (3-39) to (3-42) are obtained l;lnder simplification and
linearization and are applicable only for small e and e. I v3(t) O O O

Example 3 (Reference 512) Y34(t) O O O

Consider four vehicles moving"in a single lane as shown in Figure 3-16. Let V4(t) O O O
Yi, Vi' mi' and Ui be, respectively, the position, velocity, mass and the applied
force of the ith vehicle. Let k be the viscous friction coefficient and be the same This is the state equati
for a11 four vehicles. Then we have, for i = 1,2,3,4, interested in the diste
therefore the absolute
V¡=Yi (3-43) on what will be consic
(3-44 ) developed.
Ui =kv¡ +m¡vi
The purpose of this problem is to maintain the distance between adjacent Example 4 (ReferencE
vehicles at a predetermined value ha and to maintain the velocity of each Consider a satel!ite of
vehicle as clase as possibie to a desired velocity Va. Define altilude of lúe sateilite
Yi,i+l(t)=y¡(t)-Yi+l(t)-h a i=1,2,3 (3-45)

v¡(t) = Vi(t) - Va i = 1, 2, 3, 4 (3-46)

4-c:J U3-c:J " u2-2 UI_c:J

U
'l"/./##$////#//./#//#//.//#/##//./$#####/###//#///////#/#///##//////////////#/#/////////////////.
-Y4~ Y3~ ~
.
- - - - - - - - - - - - - - - - - . : . . . - - - . - - - - Y2 YI-J
Figure 3-16 Four vehicles moving in a single lane. Figure 3-17 Satellite il
-" --- ~--"--_=:...::-

EXAMPLES 99

lte variables as Xl = y, and u¡(t) = ult) - kvo (3-47)


m solve ji and (j as
The term kvo is the force needed to overcome the friction for the vehicles to
maintain their velocities at va. Now the problem reduces to find ü¡(t) such
that Y¡,¡+l(t) and viCt) are as close as possible to zero for all t. From (3-45),
we have, for i = 1, 2,3,
Yi,¡+ 1(t) = y¡(t) - y¡+ I (l) = v¡(t) - Vi + I (t)
;tate-variable equation From (3-46) and (3-44), we have, for i = 1,2,3,4
'. -k 1 -k 1
V¡\t)=-Vi +-u¡=-v¡(t) +-ü¡(t)
o m¡ m¡ m¡ m¡
2
These equations can be arranged in matrix form as
2M +m
u (3-42)
1
O
-k
VI(t)
O O O O O O VI (t) O O O
-1 mi mI
YdO
1 0-1 O O O O jidt) O O O O ül(t)
(2M +m)l
-k 1
V2(t)
O O O O O O V2(t) O O O Ü2(t)
m2 m2
Ydt)
O O 1 O -1 O O ydt) +
O O O O ü}(t)
der simplification and -k 1

V3(t) O O O O
m}
O O v}(c) O O
m}
O Ü4(t)

Y34(t) O O O O 1 O -1 Y34(t) O O O O
-k o, 1
V4(t) O O O O O O v 4(t) O O (3-48)
Tn in Figure 3-16. Lef m4 m4
mass and the applied
This is the state equation description of the system. In this problem, we are
fficient and be the same
interested in the distances between adjacent vehicles and their velocities;
therefore the absolute distances do not appear in the equations. Depending
(3-43) on what will be considered as outputs, an output equation can be similarly
developed.
(3-44)
Example 4 (Reference 596)
lnce between adjacent
n the velocity of each Consider a satellite of mass m in earth orbit as shown in Figure 3-17. The
fine aititude of the satellite is specHled by r(e),O(tj, and cPU) as shown. The orblt

2,3 (3-45 )

(3-46)

U&

-
] UI-G

VI

y -
I Salellite (mass m)
I
1",
.
.
& --
.....................
1""

'"
7//////'///'//$///'//&//////////,I/. --- ---- - .""v

Figure 3-17 Sateilite in earth orbit.


---- ------~------~-- --- ----- ~---- ---~- -_:::.._~.

100 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

can be controlled by three orthogonal thrusts U,.(t), ua(t), and u.p(t). The state, lf the perturbation is '
input, and output of the system are chosen as can be linearized as
r(t) O 1 O
I\t)
8(t) U,.(t)] r
u)] 3w 2 O O
x(t) = u(t)= ua(t) y(t)= 8(t)
e(t) [ uq,(t) [ O O O
cP(t) cP(t)
-2w
x(t) = O O
~(t) ro
Then the system can be shown to be described by ------------
O O O
r
re 2COS 2cP +r~2 - k/r 2 +u,./m O O O
x= h(x u) =
,
. .. e
-2r8/r +284> sin <PIcos 4> +ua/mr cos cP
(3-49a)
l O O O: O
. ~. y-( t ) = O O 1 O:, O
-:- ­
_8 2 COS cP sin cP -2rcP/r +u.p/mr

1 O O O O O]
f
-- - -- -- - -
O O O O: 1
where A =oh(x, u)/ox,
and Y= ex = O O 1 O O o x (3-49b)
[ and u o. This is a sixth
o o o o 1 o
lt can be used to descri
One solution which corresponds to a circular, equatorial orbit is specified by the circular orbit rema
ro(t) ro The dashed lines in
ro(t) . O two uncoupled parts, (
~o(t) wt two parts independentl:
xo(t) = uo(t) =0 (3-50)
8 0 (t) W
<ro(t) O °Dynamical equatio
cPo(t) O section a systematic prc
cal equations for genera
where ro and w are related by r6w2 =k =a known physical constant. Once pendent voltage and Cl
the satellite reaches this orbit, it will remain in the orbit as long as there are currents and the capac
no disturbances. lf the satellite deviates from the orbit, thrusts must be applied behavior of the networ
to push the satellite back to the orbit. Define is not necessary to cho
'ro + cX l(t) -, state variables. This ca
t:X2(t) lf we assign all the cap,
as shown, then we see tI
x(t) =xo(t) +t:X:(t) = wt +t: X3(t)
w +6: 4 (t)
cXs(t)
R
cX6(t)
+

u(t) =uo(t) +t:ü(t) = [:~~~~~]


t: U3(t)

ro +cXl(t)]
and y(t)=yo(t) +t:y(t) = wt +t:x~U) Figure 3-18 Circuits wit:
[ consists ofinductors only.
cXs(t)
- - - - - - - -- ------- -

EXAMPLES 101

" and u,¡,(t). The state, If the perturbation is very small, or equivaIentIy, é: is very small, h(x, u) in (3-49)
can be linearized as
,
o 1 O O ,, O O O O O
, 1
,
[~l)
8(t)
,] 3w 2

O
O
O
O 2wro , O
O
,
,
1 ,, O
O
O
m
O
O
,
O ,, O
O

cP(t) -2w , 1 ,
x(t) = O O O , O O x(t) + O O üU)
, ~,
(3-51a)
ro , mro:
, ,
------------------,------­
, ----------­
, I

O O O O , O
I 1 O O ,, O
, 1
O O O O O
l,/m , mro

I
(3-49a)
/mrcoscjl I O O O: O 0J
y(t)= ~ __~ __1__ ~_~-~--~ x(t) (3-51b)
u<f¡/mr
0000:10

(3-49b)
where A =oh(x, u)/ox, B =oh(x, u)/ou, and computed at the circular orbit at Xo
and uo. This is a sixth-dimensionallinear time-invariant dynamical equation.
lt can be used to describe and control the satellite so long as the deviation from
rial orbit is specified by the circular orbit remains smal!.
The dashed lines in (3-51) show that the equation can be decomposed into
two uncoupled parts, one involving r and 8, the other <p. By studying these
two parts independently, the analysis and design can be considerably simplified.
(3-50)
*Dynamical equations for RlC networks. We introduce in this sub­
section a systematic procedure for assigning state variables and writing dynami­
cal equations for generallumped linear RLC networks which may contain inde­
ysical constant. Once pendent voltage and current sources. lt is well known that if all the inductor
)it as long as there are currents and the capacitor voltages of an RLC networks are known, then the
thrusts must be applied behavior of the network is uniquely determinable for any input. However, it
is not necessary to choose aH the inductor currents and capacitar voltages as
state variables. This can be seen from the simple circuits shown in Figure 3-18.
lf we assign all the capacitor voltages and inductor currents as state variables
as shown, then we see that Xl (t) = xz(t) for aH t. Clearly there is a redundancy

R LI
+
¡-
X2 R
XI

C
c l C2
d L2
Figure 3-18 Circuits witha: loop which consists of capacitors only or a cutset which
consists ofinductors only.
102 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

here. Hence the state nf an RLC network can be chosen to consist of only inde­
pendent capacitor voltages and independent inductor currents.
Before proceeding, we review briefly the concepts of tree, link, and cutset
of a network. We consider only connected networks. A tree of a network is
defined as any connected graph (connection of branches) containing aB the
nodes of the network and not containing any loop. Every branch in a given
tree is caBed a tree branch. Every branch not in the tree is caBed a link. A
cutset of a connected network is any minimal set of branches such that the (voltage ­
source)
removal of aB the branches in this set causes the remaining network to be un­
connected. With respect to any fixed tree, every link and sorne tree branches
form a unique loop caBed a fundamental loop; every tree branch with sorne
links forro a unique cutset caBed a fundamental cutset. Hence every funda­
Figure 3-19 A networ
mental loop ineludes only one link, and every fundamental cutset ineludes only
one tree branch. With these concepts, we are ready to give a systematic pro­
By applying the Kirchl
cedure for developing a dynamical-equation description of any RLC network we have immediately Ü
that may contain independent voltage sources or current sources 9 : the voltage across resis
1. Choose a tree caBed a normal tree. The branches of the normal tree are are expressed in terms
chosen in the order of voltage sources, capacitors, resistors, inductors, and to the fundamental cuts
current sources. Hence, a normal tree consists of aB the voltage sources,
the maximal number of permissible capacitors (those that do not form a
loop), the resistors, and finaBy the minimal number of inductors. UsuaBy The application of the I<
it does not contain any current source.
2. Assign the charges or voltages of the capacitors in the normal tree and the
flux or current of the inductors in the linksas state variables. The voltages This application of the }
or charges ofthe capacitors in the links and the flux or current ofthe inductors
in the normal tree need not be chosen as state variables.
3. Express the branch variables (branch voltage and current) ofaB the resistors, These three equations ca
the capacitors in the links, and the inductors in the normal tree in terms ofthe
state variables and the inputs by applying the Kirchhoff voltage or current
law to the fundamentalloops or cutsets of these branches.
4. Apply the Kirchhoff voltage or current law to the fundamental loop or
cutset of every branch that is assigned as a state variable. The output equation can

Example 5
Consider the linear network shown in Figure 3-19. The normal tree is chosen Example 6
as shown (heavy lines); it consists of the voltage source, two capacitors, and Find the input-output d{
one resistor. The voltages of the capacitors in the normal tree and the current equivalently, the transfer­
of the inductor in the link are chosen as state variables. Next we express the
i~puts, one output, the tn
variables of resistors CD and (2) in terms of the state variables and inputs. By
[g¡¡(~) 9ds)], where 9¡
applying the Kirchhoff voltage law (KVL) to the fundamental loop of branch and gds) is the transfer J
CD, the voltagé across CD is found as '(u¡ ~ Xl); hence its current is (u¡ - xJ. network reduces to the Oll<

9 A network with a loop that consists of only vo'ltage sources and ~apacitors or with a cutset that
consists or onlycurrent sources and inductors is excluded, because in this case its dynamical­
L.
equation description cannot be or the form in (3-33). - -1
s
EXAMPLES 103

to consist oIonly inde­ IFundamental cutset of (3)


I
I
nts. I J F~ndamental cutset of 0)
(U¡ -xl)

: tree, link, and cutset / /

/ I
A tree of a network is
les) containing aH the
rery branch in a given lF
~ee is caHed a link. A (voltage ­
ranches such that the source) G)'
ling network to be un­
nd some tree branches
.ree branch with some
. Hence every funda­ Figure 3-19 A network with voltage and current sources.
tal cutset includes only
give a systematic pro­
By applying the Kirchhoff current law (KCL) to the fundamental cutset of (Z),
1 of any RLC network
we have immediately that the current through resistor (2) is X3' Consequently
11 sources9 :
the voltage across resistor (2) is X3' Now the characteristics of every branch
are expressed in terms of the state variables as shown. If we apply the KCL
of the normal tree are
to the fundamental cutset of branch Q), we have
esistors, inductors, and

aH the voltage sources,


(u¡-x¡)-x¡ +U2 -X3=0

)se that do not form a


The application of the KCL to the fundamental cutset of ® yields
of inductors. Usually

the normal tree and the


This application of the KVL to the fundamental loop of (J¡ yields
rariables. The voltages

.current ofthe inductors


X3 +X3 -Xl +X2 =0
)les.
These three equations can be rearranged in matrix form as
[reut) of aH the resistors,

)rmal tree in terms of the

h.hoff voltage or current

o
O
-1
-1] [1
-1
1 x + O
O
.nches.

le fundamental loop or
The output equation can be easily found as
·iable.

-- .1'"
. . /.,

Example 6
he normal tree is chosen

rce, two capacitors, and


Find the input-output description of the network shown in Figure 3-19, or
'mal tree and the current
equivalently, the transfer-function matrix of the network. Since there are two
es. Next we express the
inputs, one output, the transfer-function matrix is a 1 x 2 matrix. Let G(s) =
ariables and inputs. By
[§ ¡¡(s) gds)], where g11 is the tninsfer function from U¡ to y with U2 = O,
iamental loop of branch
and gds) is the transfer function from U2 to y with U¡ =0. With U2 ""'0, the
e its current is (u¡ - Xl)'
network reduces to the one in Figure 3-20(a). By loop analysis, we have

(1 +D ¡ ¡(s) ~.~ ¡ 2(S) ~ u¡(s)


~apacitors or with a cutset that
.use in this case its dynamical- L
-~11ls)+ 2)A
( l+s+~ 1 2 (s)=0
104 MATHEMATlCAL DESCR¡PTlONS OF SYSTEMS

1.11
1 R

T,--~[_':1_
S

= E c
Ca) Cb)

Figure 3-20 Two reduced networks.

Solving ror I 2(s), we obtain R

1s
1+
_ (l/s)u¡(s)
2( )-(1 +l/s)(l +s +2/s)-1/s 2 L LCT
r
Hence
A y(s)
g¡¡(S)=-A- u,~o
sI 2(s)
=-A-= ~ 32
S2
1
El__ ~
u¡(s) inilially u¡(s) s- +2s + s +
relaxed Cb)

Ir U¡ = 0, the network in Figure 3-19 reduces to the one in Figure 3-20(b). Figure 3-21 Network w
By node analysis,
1 )
These can be rearranged
( S +s
A A
V¡(s) -sV2(s) =0

-sV¡(s) +(1 +2s)V2(s) -(1 +s)V3(s) =U2(S)


-(1 +s)V2(s) +(2 +s)V;,(s) = -uz{s)
Solving ror V¡ (s), we obtain
This nonlinear time-invé
the general case. N ow il
then h(x¡(t» can be appr
A V¡(s) work in Figure 3-21(a) Cl
Hence gds)=~) u,~o
u z{s initially scribed by the linear timé
re1axed

Consequently, the input-output description in the rrequency domain or the


network is given by
S2

y(s)= [ S3 +2s 2 +3s +1



Now ir x¡(t) is known
Example 7 duce. the variables x¡ (t) ==
'Consider the network shown in Figure 3-21(a), where T is a tunneI diode with as h(x¡ (t)) = i o - x¡ (t)/R 2.
the characteristic shown. Let x¡ be the voltage across the capacitor and X2
thecurrent through the inductor. Then we have
..
X2(t) = Cx¡(t) +i(t) = Cx¡(t) +h(x¡(t)) [x~¡(t)J
2 (t)
=
LX2(t) E - RX2(t) - x ¡ (t)
EXAMPLES 105

-llR l
L i = h(u)

E=.

(h)
T
(a)
L L Xl
xl

+ +
R R
1+ J+ C
Xl e u Rl
Xl ¡¡ -R 2
i=­
-L
+2s z +3s +1 (h) (e)

Figure 3-21 Network with a tunnel diode.


he one in Figure 3-20(b).

These can be rearranged as


=0 . ) - h(x 1(t)) Xz(t)
x 1 (t = e +c
0= uz(s) (3-52)
uz(s) . () -X1(t)-Rxz(t) E
0= ­
Xz t = L +I
This nonlinear time-invariant dynamical equation describes the network for
~(s) the general case. Now ir x 1 (t) is known to operate only inside the range (a, b),
then h(x 1(t)) can be approximated as h(x 1(t))=x 1(t)jR 1. In this case, the net­
work in Figure 3-21(a) can be reduced to the one in Figure 3-21(b) and is de­
scribed by the linear time-invariant state equation
3s +1

freq uency domain of the


Now ir x 1(t) is known to operate only inside the range (e, d), we may intro­
duce the variables Xl(t)=Xl(t) -vo, xz(t) =xz(t) - io and approximate h(Xl(t))
ash(x l(t)) = i o - Xl(t)/ R z. The substitution or these into (3-52) yields .
~ T is a tunnel diode with

~J[~l(t)J +l~JE
~oss the capacitor and Xz

1· .. R xz(t) ­
__ L·
L L
106 MATHEMATlCAL DFSCRIPTIONS OF SYSTEMS

where E = E - Vo - Rio. This linear equation is applicable only if v(t) is limited


inside the range (e, d). An equivalent network of this linear equation is shown
in Figure 3-21(c). I In In
-IF
u t
3-5 Comparisons of the Input-Output Description In
and the State Variable Description

We compare in this section the input-output description and the state-variable Figure 3-23 Active:
description of systems.

1. The input-output description of a system describes only the relationship


between the input and the output under the assumption that the system is
initial1y relaxed. Hence, if the system is not initial1y relaxed, the description A dynamical equ~
is not applicable. A more serious problem is that it does not reveal what input and output 1
will happen ifthe system is not initiaHy relaxed nor does it reveal the behavior condition; hence a
inside the system. For example, consider the networks shown in Figures 3-22 2. For ~xtremely com
and 3-23. The networks are assumed to have a capacitor with -1 farad. Ir the dynamical-equ;
the initial voltage of the capacitor is zero, the current will distribute equally find the input-out¡:
among the two branches. Hence the transfer function of the network in each input termim
Figure 3-22 is 0.5; the transfer function of the network in Figure 3-23 is 1. output terminals g
Because of the negative capacitance, if the initial voltage across the capacitor system. In practic<
is difIerent from zero, then the voltage Yl in Figure 3-22 will increase with narrow pulse, or ide
time, whereas the voltage Y2 in Figure 3-23 remains equal to u(t) for aH t. unit step function.
It is clear that the network in Figure 3-22 is not satisfactory, because the

output will increase without bound if the initial condition is different from

zero. Although the output of the network in Figure 3-23 behaves well,

the network is still not satisfactory, because the voltages in branches 1 and 2
The response of a I
will increase with time (in different polarity), and the network will eventuaJJy given by
burn out. Hence the networks in Figures 3-22 and 3-23 can never function

properly. Ir the internal structure of the network is not known, this fact

cannot be detected from its transfer function. Consequently, the input-output

description sometimes does not characterize a system completely.

The dynamical-equation descriptions of the networks in Figures 3-22 at to). DifIerentiatiJ


and 3-23 (see Problem 3-18) can be found, respectively, as

u i¡

In In
r Y¡
For the time-invaria

u
t
iz -IF
+"
~
X Thus the impulse re~
In In (3-55) or (3-56). See
For linear time
functions or impuls<
Figure 3-22 A network with linear, time-varying elements. frequency s = jw can
COMPAR1SONS OF THE lNPUT-OUTPUT DESCR1PTlON 107

lle only if v(t) is limited


lear equation is shown
• 1n 1n
-1 F
CD1
t
scription
u

In ;n
+

1 Y2

and the state-variab1e Figure 3-23 Active networks.

only the rel¡¡tionship


x=x x=x
lion that the system is
{Y1 =O.5x + O.5u {Yz,=u
elaxed, the description A dynamical equation describes not only the relationship between the
: does not reveal what input and output but also the behavior inside a system under any initial
:s it reveal the behavior condition; hence a dynamical equation characterizes a system completely.
; shown in Figures 3-22 2. For ~xtremely complicated linear systems, it might be very involved to find
:itor with - 1 farad. Ir the dynamical-equation descriptions. In these cases, it might be easier to
will distribute equally find the input-output descriptions by direct measurements. We apply at
ion of the network in each input terminal a very sharp and narrow pulse; the responses of the
,rk in Figure 3-23 is 1. output terminals give us immediately the impulse-response matrix of the
ge across the capacitor system. In practice, we may have difficulty in generating a very sharp and
3-22 will increase with narrow pulse, or ideally a b-function; however, this can be avoided by using a
equal to u(t) for all t. unit step function. A unit step function b 1 (t -' t o) is defined as .
tisfactory, because the
dition is different from ~
U1(t-t O)= O
{l for t ¿to
fort<t
o
( 5 )
3- 3
lre 3-23 behaves well,
ges in branches 1 and 2 The response of a linear causal system due to a unit-step-function input is
letwork wil\ eventually given by
0-23 can never function
s not known, this fact
lently, the input-output
9 1(t, to) = r' g(t, r) dr
J,o
(3-54 )

n completely. where ~lí(t, to) is called the step response (dlle to a unit step function applied
works in Figures 3-22 at to). Differentiating (3-54) with respect to to, we obtain
Iy, as
(3-65 )

For the time-invariant case, (3-55) reduces to


d
g(t)= dtg1(t) (3-56)

Thus the impulse response can be obtained from the step response by using
(3-55) or (3~56). See Problem 3-25.
For linear time-invariant systems, we may measure either transfer
functions or impulse responses. The transfer function g(s) of a system at
frequency s = jw can be measured easily and accurately by employing fre­
108 MATHEMATlCAL DESCRIPTIONS OF SYSTEMS

quency response analyzers. After measuring gUw) at a number of fre­ two multivariable s:
quencies, we can then find a g(s) to match the measured gUw). This method
of determining transfer functions is often used in practice.
3. Prior to 1960, the design of control systems had been mostiy carried out by
using transfer functions. However, the design had been limited to the single­ where u i and y¡ are 1
variable case; its extension to the multivariable case had not been successful. of the system Si' 1
The state-variable approach was developed in the 1960s. In this approach, the impulse-respon:
the formulations in the single- and multivariable cases are the same, and a 3-24 that in the pa
number of results were established. These results were not available in the tandem connection,
transfer-function approach at that time; consequentiy, interest in this we have u 1 =u - Y2'
approach was renewed in the 1970s. Now the results in the state-variable SI and S 2 have com¡:
approach can also be obtained in the transfer-function approach. It also cannot be properly
appears that the latter approach is simpler in concepts and computations. effect in the connect
4. The dynamical equation can be extended to the time-varying case; the unchanged after COI
extension of the transfer function to the time-varying case has not been matrix of the paralle
successful. In the optimal design, dynamical equations can be used to study
finite terminal time problems; this is not possible by using transfer functions.
5. In rhe study of nonlinear systems, depending on the approach taken, either For the tandem e
description can be used. For example, in the study of the stability problem,
we use the input-output description in the functional analysis and operator
approaches. See References 28, 95, 117, and S79. Ir Lyapunov's second
method is employed, then we must use the dynamical-equation description. We prove (3-59) for ti
6. Ir the dynamical-equation description of a system is available, the system definition, the respon
.can be readily simulated on an analog or a digital computer. response at Y1 due to 1
gl(t,r)is
From the foregoing discussion, we see that the input-output and the state­
variable descriptions have their own merits. In order to carry out a design
efficientiy, a designer should make himself familiar with these two mathematical
descriptions. In this book, these two descriptions will be developed equally
and their relationships will be explored.

3-6 Mathematical Descriptions of Composite Systems 10

Time-varying case. In engineering, a system is often built by interconnect­


ing a number ofsubsystems. Such a system is called a composite system. Whether
or not a system is a composite system depends on what we consider as building (al
blocks. For example, an analog computer can be looked upon as a system,
or as a composite system that consists of operational amplifiers, function u
generators, potentiometers, and others as subsystems. There are many forms
of composite systems; however, mostly they are built from the following three
basic connections: the parallel, the tandem, and the feedback connections, as
shown in Figure 3~24. In this section we shall study the input-output and
state-variable descriptions of these three basic composite systems.
We study first the input-output description of composite systems. Consider
Figure 3.24 Composit.

10 The material in this section is not used until Chapter 9; thus its study may be postponed. Tandem connection. (e)

MATHEMATICAL DESCRIPTIONS Of COMPOSITE SYSTEMS 109

at a number of fre­ two multivariable systems Si, which are described by


:d gUw). This method
tice. Yi(t) = roo G¡(t,r)u;("r)dr i= 1,2 (3-57)
mostly carried out by
n limited to the single­ where Ui and Yi are the input and the output, G i is the impulse-response matrix
l.d not been successful. of the system Si. Let u, y, and G be, respectively, the input, the output, and
,Os. In this approach, the impulse-response matrix of a composite system. We see from Figure
~s are the same, and a
3-24 that in. the paral1el connection, we have u 1 =u z =u, Y=Yl +Yz; in the
re not available in the tandem connection, we have u =u 1 , Y1 = u z, Yz = y; in the feedback connection,
:ntly, interest in this we have U1 = u - Yz, Y= Yl' Here we have implicitly assumed that the systems
s in the state-variable SI and Sz have compatible numbers of input and output; otherwise, the systems
on approach. lt also cannot be properly connected. lt is also assumed that there is no loading
>ts and computations. effect in the connection; that is, the impulse response matrices G¡, G z remain
lme-varying case; the unchanged after connection. 1t is easy to show that the impulse response
ng case has not been matrix ofthe parallel connection of SI and Sz shown in Figure 3-24(a) is
lS can be used to study G(t, r) = G 1(t, r) + Gz(t, r) (3-58)
;ing transfer functions.
l.pproach taken, either For the tandem connection shown in Figure 3-24(b), we have
. the stability problem,
analysis and operator
1f L yapunov's second
G(t, r) = f Gz(t, v)G 1(v, r) dv (3-69)

-equation description. We prove (3-59) for the single-variable case. The impulse response g(t, r) is, by
available, the system definition, the response at yz due to a b-function applied at time r at Ul' The
nputer. response at Y1 due to this o-function is g 1(t, !). The output of Sz due to the input
gl(t,r)is
-output and the state­
to carry out a design
lese two mathematical
J: gz(t, v)g I(V, r) dv
be developed equal1y

y =
u Y¡ + Y2
site Systems 10 U=U;{~
O-~/
+

built by interconnect­
Josite system. Whether
'e consider as building (a) (b)

:ed upon as a system,


I amplifiers, function u + U¡ y

rhere are many forms


,m the fol1owing three
jback connections, as
. the input-output and
e systems.
ite systems. Consider (e)

Figure 3.24 Composite eonneetions of two systems. (a) Parallel eonneetion. (b)
Tandem eonneetion. (e) Feedback eonneetion.
I may be postponed.
110 MATHEMATICAL DESCRIPTlONS OF SYSTEMS

Hence, g(t, r) = f gz(t, v)g[(v, r) dv description is

The same procedure can be used to prove the multivariable case.


For the feedback connection shown in Figure 3-24(c), the impulse-response

r
matrix is the solution of the integral equation

G(t, r) = G[(t, r) - f G[(t, v) Gz(v, s)G(s, r) ds dv (3-60)

where G[ and G z are known and G is unknown. This equation can be easily y = [Y [(t)C[(t)
verified from the definition of G(t, r) (Problem 3-26). There is a general iterative
method for solving Equation (3-60), but it is very involved.
where y [(t)=(I+E l ,
Now we study the state-variable descriptions of composite systems. Let
order for (3-64) to be
the systems S [ and S z in Figure 3-24 be described by
and (1 + E z(t)E l (t)) eJ
X¡ = A¡(t)x¡ + B¡(t)u¡ i= 1, 2 (3-61 a) verified by observing
y¡ =C¡(t)x¡ +E¡(t)u¡ (3-61b)
Time-invariant ca
where x¡ is the state, o¡ is the input, and y¡ is the output; A¡, B¡, C¡ and E¡ are
applied to the time-;
matrices of compatible order whose entries are continuous function of t defined
discuss the transfer-J
over l - 00, (0). The state space of S ¡ is denoted by L¡.
Gz(s) be the proper
Let us introduce the concept of the direct sum of two linear spaces. The
tively; then the transl
linear space L is the direct sum of two linear spaces L[ and L2> written as
is G l (s) +Gz(s). Th
L = L[ EB LZ' if every vector in L is of the form [x{ xí]', where x[ is a vector
SI followed by Sz is
in L[ and X z is a vector in LZ' The dimension of Lis the sum of those of L[
reversed. It is clear
and L z.
G Z(s)G l (s). It is im
It is clear that the composite vector
compatible in both c;
In order to discus
shown in Figure 3-24
qualifies as the state of any composite connection of S [ and S z; its state space
Theorem 3-2
is the direct sum of the state spaces of S [ and S z, L [ EB LZ' F or the parallel
connection, we have o[ = U z = o, y = y [ +y z; hence its dynamical equation is Let G l (s) and Gz(s) b
(not necessarily prope
[::] == [A~(t) A~(t)J[:J +[::~:~J u (3-62a)

y = [C[ (t) Cz(t)] [:J +(E[ (t) + Ez(t))u (3-62b)


Observe that the [
The dynamical equation of the tandem connection of S [ and Sz is given by while the matrix on tl
of order m. The eler
(3-63a) since the rational fUn<
be applied.

(3-63b)
Proof of "Theprem 3-:

1t is well knowti that (


which can be easily obtainedby observing o[ = U, Y1 = U z, y = y z.
For the feedback connection shown in Figure 3-24(c), its dynamical-equation det (J
MATHEMATICAL DESCRIPTIONS OF COMPOSITE SYSTEMS 111

description is

.le case.
the impulse-response

(3-64a)
r) ds dv (3-60)

:¡uation can be easily


~ is a general iterative
y = [Y 1 (t)C1(t) - Y 1 (t)E 1(t)C2(t)] [::J + Y1(t)E 1(t)u (3-64b)

l. where Y 1 (t) = (1 + E 1(t)E 2(t»- 1 and Y 2(t) = (1 + E 2 (t)E 1 (t))-1. It is clear that in
lposite systems. Let order for (3-64) to be defined, we must assume that the inverses of (1 + El (t)E 2 (t»
and (1 + E 2 (t)E 1 (t» exist for all t. The dynamical equation (3-64) can be easily
(3-61 a) verified by observing U1 = U-Y2' Y1 =U2' Y=y [. (See Problem 3-28.)
(3-61 b)
Time-invariant case. All the results in the preceding subsection can be
. A¡, Di, Ci and E¡ are applied to the time-invariant case without any modification. We shall now
; function of t defined discuss the transfer-function matrices of composite systems. Let G1 (s) and
G2(s) be the proper rational transfer-function matrices of SI and S 2, respec­
o linear spaces. The
tively; then the transfer-function matrix of the parallel connection of SI and S 2
[ and ~2' written as
is G[(s) + Gz(s). The transfer-function matrix of the tandem connection of
, where x[ is a vector
SI followed by S2 is G 2(5)G 1 (5). Note that the order of G 2(S)G[(5) cannot be
le sum of those of ~ [
reversed. It is clear that if Gi (5), i = 1, 2, are proper, so are G1(5) + G2(s) and
G 2 (5)(;[(5). It is implicitly assumed that the orders of G[(5) and G 2 (5) are
compatible in both connections.
In order to discuss the transfer-function matrix of the feedback connection
shown in Figure 3-24(c), we need sorne preliminary results.

,nd S 2; its state space Theorem 3-2


~2' For the parallel
lamical equation is Let G[(s) and G 2 (s) be, respectively, q x p and p x q rational function matrices
(not necessarily proper). Then we have

liI

¡(t)Ju (3-62b)
Observe that the matrix on the right-hand side oí (3-65) is a q x q matrix,
while the matrix on the left-hand side is a p x p matrix. 1m is the unit matrix
S [ and S2 is given by
of order m. The elements of these matrices are rational functions of s, and
[(t)
)E[(t) u
J (3-63 a )
since the rational functions form a field, standard results in matrix theory can
be applied.

(3-63b) Proof of Theorem 3-2


It is well known that det (NQ) = det N det Q. Hence we have
,Y=Y2'
:s dynamical-equation det (NQP) = det N det Q det P= det (PQN) (3-66)
112 MATIlEMATlCAL DESCR¡PTlONS OF SYSTEMS

where N, Q, and Pare any square matrices of the same order. Let us choose the feedback system
N - [Ip 0J Q _ [I p - (;2(S)J P = [I p (;2(S)J G(s) = G I,
- - (;¡(s) I q - (;¡(s) Iq O Iq
They are square matrices of order (q + p). lt is easy to verify that Proof
From Figure 3-24(c),
NQP=[I p AO A ]

O I q +G¡(s)G 2 (s)

and pQN=ep+(;2~)(;¡(S) :J which, together with

Note that the COI


hence det (NQP) =det (I q + 91(S)(;~(s)) system to be defined.
det(PQN) =det(I p + (;2(s)(;¡(s)) meaningless in the se
and the theorem follows immediately from (3-66). For a different proof, see Equation (3-69).
Problem 3-32. Q.E. D.
Example 1
Theorem 3-3 Consider a feedback:
lf det (Iq + (;1 (S)(;2(S)) + 0, then
GI(s)(I p +G 2(S)G I(s))-1 =(I q + G I(S)G 2(sW IGI(S)

Proof
Note that the zero in det(I q +G I(S)G 2 (S))+0 is the zero element in the field
of rational functions. Hence, it can be written more suggestively as
lt is easy to verify tha
det(Iq +G I(s)G 2 (s))+O for sorne s. The ~ondltion det(Iq +G¡(s)G 2 (s))+O
implies that the inverse of the matrix (I q +G I(S)G 2 (s)) exists. From Theorem
3-2, we have
det(I q +G I(s)G 2 (s))=det(I p +G 2 (S)G 1(S)) +0
Hence, both (Iq +G 1(S)G 2 (s))-1 and (I p +G 2 (S)G 1(s))-1 exist. Consider the
identity

which can be written as


(I q +G 1(s)G 2 (S))G 1(s)(I p +G 2 (S)G 1(S))-1 = G1(s) (3-67)

Premultiplying (I q + G 1(s)G 2 (S))-1 on both sides of (3-67), we obtain the desired


equality. Q.E. D.
l;
Knowing Theorem 3-3, we are ready to investigate the transfer-function Obviously there is no
matrix of the feedback connection of SI and S2' is said to be inconsi~
det (I q + G 1(S)G 2 (s)) +
Corollary 3-3
Recall from (3-64)
Consider the feedback system shown in Figure 3-24(c).. Let G1(s) and ·G 2 (s) equation description o .
be q x pand p x q proper rational transfer-function matrices of SI and S2' where E¡ is the direc
respectively_ lf det (I q + G 1(s)G 2 (s)) + 0, then the transfer-function matrixof Because of G¡(oo) = E¡
MATHEMATICAL DESCRIPTIONS Of COMPOSLTE SYSTEMS 113

order. Let us choose the feedback system is given by


(;(s) = (;l(s)(I p +(;iS)(;l(S))-l =(I q +(;l(s)(;is))-l (;l(S) (3-68)

Proof
erify that
From Figure 3-24(c), we have (;l(s)(ii(s) - (;2(S)Y(S)) = y(s), or
(I q + (;1(S)(;2(S))Y(s) = (;l(s)ii(s) (3-69)

which, together with Theorem 3-3, implies this corol1ary. Q.E.D.

+
Note that the condition det (I q + (; 1(S)(;2(S)) O is essential for a feedback
system to be defined. Without this condition, a feedback system may become
meaningless in the sense that for certain inputs, there are no outputs satisfying
Equation (3-69).
a different proof, see
Q.E.D.
Example 1
Consider a feedback system with

-s
A s +1
G 1(s)= _1_
s +2
1 ]
G 2 (s)= O 1
A [1 0J
[
s +1
~s~1
o element in the field
aor( suggestively as
lt is easy to verify that det (1 2 + (;1(S)(;2(S)) =0. Let us choose
et(I q + (;1 (S)(;2(S)) +0
dsts. FroID Theorem

+O
A

=
s +21]
s))
exist. Consider the
u(s)
[~
(s 1)2

Then (3-69) becomes

(3-67)

, we obtain the desired


Q.E.D. r+s +1
(3-70)

Obviously there is no y(s) satisfying (3-70). In matrix theory, Equation (3-70)


: the transfer-function
is said to be inconsistent. Hence, in the feedback connection, we require
+
det (I q + (; 1 (s) (;2(S)) O for some s. •

Recal1 from (3-64) that the condition for the existence of the dynamical­
Let (;l(S) and (;2(S) equationdescription ofthe feedback system in Figure 3-24(c) is det (1 + E 1 E 2 ) O +
[atrices of Sl and S2, where Eiis the direct transmission part of the dynamical equation of Si'
fer-function matrix of +
Because of(;¡(co) = E¡ [see Equation (3-37)J, if det (1 + E 1 E 2 ) O, then det(1 +
I
I

ji

114 MATHEMATICAL DESCR¡PTIO~S OF SYSTEMS

G¡ (s)G 2 (s)) +-2. However, the converse is not true; that is, the condition
det(1 +G¡(s)G 2 (s))+-0 may not imply det(1 +E¡EJ+-O. Hence, a feedback
system may have the transfer-function matrix description without having the U¡
state-variable description. This discrepancy will be resolved in the following +
H
subsection.

Well-posedness problem. We showed in Corollary 3-3 that if G¡(s) and 1


s
G2 (s)are proger ratipnal fun~tion .matrices and if det (1 + G1(S)G 2 (S))+-0 for
sorne s, then G(s) = G¡(s)(1 + G 2 (s)G¡(s)r 1 is well defmed. However, nothing
has been said regarding whether or not G(s) is proper. In this subsection, this 1
problem will be studied. s+

Example 2 F
Uz + J -s .¡
Consider the feedback system shown in Figure 3-24(c) with s+

[
-1

G1(s)= _1_

s +1
-JJ
s+lJ
Figure 3-25

of the signa!. Hence,


although all of its su
Feedbal

design of feedback sy~


The rational function matrix proper transfer functi<

l
rational matrix. In th,
Before proceeding, we
s
1 Theorem 3-4
-1
s +1 Let M(s) be a square r

is clearly nonsin~ular in the field of rational functions. The overall transfer­


function matrix G(s) is given by where Mp(s) is a poI)
matrix. Then M- i(s)
r
G(s)= _1_
-1
~ lJl
s
-s-2
O
1
~ Jl-l =[-s-s+1
-1
Proof

[ We first show that if 1\

s +1 s+l s+l s +1
M-¡(s) = {Mp(s)[1
which is not a proper rational function matrix. The block diagram of this If M; ¡(s) is proper, tr
feedback system is plotted in Figure 3-25. • strictly proper, M sp( G
constant matrix. Thu:
For thefeedback system in Figure 3-25, if the input signals u(t) is corrupted From M-1(s)=[
by high-frequenS?' noises o(t), then the noises will be amplified by the improper (Mp(S)+O)-l-.[Adj l.\­
rational matrix G(s). For example, let u(t)= sin t U o and o(t) =0,01 sin 1000t 0 0
where U o and oo' are 2 x 1 constant vectors. Although the amplitude of the 11 A sufficient condition for
noise is only one hundredth ofthe amplitude ofthe signa! atthe inpútterminals, Appendíx G. For necess
the ainplitude of the noise at the output terminals is 10 times larger than that 3-4 of Reference S34.
MATHEMATICAL DESCRIPTIONS Of COMPOSITE SYSTEMS 115

:hat is, the condition


). Hence, a feedback
In without having the u,
+o-~
-
-1
)Ived in the following + +
H

'1 3-3 that ifG1(s) and ~ 1


s
(+ (;1(S)(;2(S)) #=0 for
:d. However, nothing
[n this subsection, this 1
'-­ ~ s+T
E
F
Uz + J -s + 2 + yz
s+1 +
ith -

Figure 3-25 Feedback system with an improper transfer-function matrix.


=1
of the signa\. Hence, the system in Figure 3.25 is only of limited use in practice,
although al1 of its subsystems have proper rational functions. Thus in the
design of feedback systems, we shal1 require not only that al1 subsystems have
proper transfer f.unctions but also that the resulting overal1 system has a proper
rational matrix, In the remainder ofthis subsection, we shal1 study.this problem.
Before proceeding, we need a preliminary result.

Theorem 3-4
Let M(s) be a square rational matrix and be decomposed uniquely as
M(s) = Mp(s) + Msp(s) (3-71 )
The overall transfer­
where Mis) is a polynomial matrix and Msp(s) is a strictly proper rational
matrix. Then M - i(s) is proper if and only if M; 1(S) exists and is proper. 11

s +1
Proof
-s We first show that if M; 1(S) is proper, so is M- 1(s). We write (3-71) as
M- 1(s) = {Mp(s)[1 +M; 1(s)Msp(S)]} -1 = [1 + M; 1(s)Msp(S)] - 1M; I(S)
block diagram of this If M; 1(S) is proper, then M; 1((0) is a finite constant matrix. Since Msp(s) is
• strictly proper, Msk~)=O. 'Hence we have M-1(00)=M;1(OO), a finite
constant matrix. Thus M -1(s)'is proper.
ignals u(t) is corrupted From M- 1 (s)=[M p(s)+M sj,(s)]-1, we have, as· S~OO,M-1(S)~
,¡¡tied by the improper (Mp(S)+O)-1~[Adj Mp(s)]fdet Mp(s). Hence if Mp(s) is singular, M- 1(oo) is
n(t) =0.01 sin l000t no
L the amplitude of the l' A sufficient condition for M; '(s) to be proper is Mp(s) column-reduced or row-reduced. See
at the input terminals, Appendix G. For necessaryj'and sufficiént conditions, see Problem G.l7 and Sections 2-4 and
times larger than that 3-4 of Reference 534.
116 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

not a finite constant matrix and M- 1(s) is not proper. 12 Now we assume The condition de
Mis) to be nonsingular and show that if M- 1(s) is proper, so is M; 1(s). From theorem is the same a
M(s) = Mp(s)[1 + M; 1(s)Msp(s)], we have dynamical descriptior
resolves the discrepan
Mo(s)~ [1 + M; 1(s)Msp(s)] -1 = M- 1(s)Mis) = M- 1(s)[M(s) - Msp(s)]
Before discussin
= 1- M -1(s)M sp(s)
det [1 +G 2 (00)G 1(oof
which, together with the finiteness ofM -1(00) and Msp(oo) =0, implies Mo(oo) = Figure 3-26. This sy~
1. Hence we have M- 1(00) = Mo(oo)M; 1(00) = M; 1(00), and M;I(OO) is a and will be developed
finite constant matrix. Thus M; I(S) is proper. Q.E.D. The transfer matrices
matrices of appropriat
This theorem shows that the properness of M- 1(s) depends only on the
e(s)=í
polynomial part of M(s). We give a special case in the following.
which implies
Corollary 3-4
e(s) = I
If M(s) is a square proper rational matrix, M -1(S) is proper if and only if M( (0)
Hence the transfer mal
is nonsingular.
Gf(s) = G2
Proof
Ip view of Theorem 3-:
lf M(s) is proper, the polynomial part of M(s) is Mis) = M(oo). Hence M- 1(s) G o(00 ) is nonsingular, .
is proper if and only if M( (0) is nonsingular. Q.E. D.
Go(.
Using these results, we are ready to answer the question posed at the begin­ lf G¡(s) are proper, the
ning of this subsection: the properness of G(s) in (3-68). nonsingular, Corollary
in (3-73) is proper. H{
Theorem 3-5 condition for Gf(s) to b
Consider the feedback system shown in Figure 3-24. Let G 1(s) and G 2 (s) be
q x pand p x q proper rational transfer matrices of SI and S2' Then the overall
transfer matrix
G(s)=G 1(s)[1 +G 2 (S)G 1(S)]-1
is proper if and only ifl + G 2 ( oo)G1(00) is nonsingular.
Then we have Go(co)=
Proof
lf I +G 2 (00)G 1(00) is nonsingular, [1 +G 2 (S)G 1(S)]-1 is proper. G(s) is the which is proper.
product of two proper rational matrices; hence it is proper.
lf G(s) is proper, so are G 2 (s)G(s) and 1- G 2 (s)G(s). Since
[1 + G2 (s)G 1(s)] -1 = 1- G 2 (S)G 1(s)[1 + Gz(s)G 1(s)] -1 = 1- G 2 (s)G(s)
which c~n be readily verified by postIllultiplying 1+ Gz(s)G 1{s), we ~conclude
that if G(s) is proper, so is [1+G 2 (s)G 1(S)]-I. Hence [1+G 2 (oo)G 1(00)]-1
'¡s:finite and 1+ G 2 (00)G 1(00) is nonsingular. This establishes the theorem.
Q.E.D.

12 For adifferenl proof of this sta temen l. see Problein3A4. Figure 3-26 A feedbad
MATHEMATICAL DESCRIPTIONS OF COMPOSITE SYSTEMS 117

r. 12 Now we assume The condition det [1 + (;1(00)(;2(00)J =det [1 + (;z(oo)(;I(oo)J =/=-0 in this
~r, so is M; 1 (s). From theorem is the same as the condition det (1 + E 1 E 2 ) =/=- O for the existence of the
dynamical description of the feedback system shown in Figure 3-24(c). This
resol ves the discrepancy between these two descriptions.
(s) [M(s) - Msis)] Before discussing the physical implications of the condition
det [1 + (;2(00)(;¡{00)] =0, we study first the composite system shown in
)= 0, imp\ies M 0(00 ) = Figure 3-26. This system is a generalization of the system in Figure 3-24(c)
X»), and M;I(oo) is a and will be developed in Chapter 7 and then extensively studied in Chapter 9.
Q.E.D. The transfer matrices (;¡(s) in Figure 3-26 are assumed to be proper rational
matrices of appropriate orders. Clearly, we have
1 depends only on the e(s) = o(s) - (; 3(S)(;1 (s)e(s) - (;4(S)(;z(S)(; 1 (s)e(s)
tllowing.
which implies
e(s) = [1 + (;3(S)(;I(S) + (;4(S)(;2(S)(;1 (s)] -I U(S) (3-72)
Jer if and only if M( 00 ) Hence the transfer matrix from u to y is given by
(;¡(s) = (;z(S)(;1 (s)[1 + (;3(S)(;I(S) + (;4(S)(;2(S)(;, (s)]-' (3-73)

Ip view ofTheorem 3-5, one may wonder whether (;¡(s) is proper ifand only if
M(oo). Hence M-'(s) G o( (0) is nonsingular, where
Q.E.D. (;o(s) ~ I + (;3(S)(;,(s) + (;4(S)(;2(S)(;I(S)
Ion posed at the begin- If Gj(s) are proper, the polynomial part of (;o(s) is (;0(00). Hence, if Ca((0) is
nonsingular, Corollary 3-4 implies that (;0 '(s) is prop~r. Consequently, (; ¡(s)
in (3-73) is proper. However, the nonsingularity of G o( (0) is not a necessary
condition for (;¡(s) to be proper. For example, consider
Jet (;1(S) and (;z(s) be

l
1
d 52' Then the overall ... s
Gz(s) = O

Then we have Co( (0) =! -! +0 = O, which is singular: however, we have


(;¡(s) = (;2(s)[1 - 1 + (;2(S)J -1 =I
s proper. (;(s) is the which is proper.
;:Jer.
Sin ce u y

-1 = I - (;2(S)(;(S) i-----r-r----.,/] G¡ (s) I-- ~,.______./

2(S)(; lis), we...c onclude


[1 + G 2(oo )G 1 (eú)J - 1
tab\ishes the theorem.
Q.E.D.

Figure 3-26 A feedbacksystem.


118 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

For this system. GAs) is proper; wh...e reas Go( (0) is singular. Will such a from any point to an
system be acceptable in practice? lf Go(oo) is singular, the transfer matrix establishes the first p,
from u(s) to c(s) in (3-72) is improper. In this case, if u(t) contains a part which the polynomial part of
has a high-frequency spectrum or is discontinuous, the amplitude of e(t) will be proper if and only if é
very large or infinite, and the system may saturate or bum out. Hence, in the
design of a feedback system, we shall require not only the overall transfer matrix We discuss now the
but also all transfer functions from all possible input-output pairs of the system and more generally, de
to be proper. In this case, no signal will be unduly amplified, and in sorne sense, A loop of a block diag
the smoothness of signals throughout the system will be preserved. direction of the path b
For example, the feedl
Definition 3_9 13 is indicated by the hea'
of all transfer function
Let every subsystem of a composite system be describable by a rational transfer
In our discussion, we ~
function. Then the composite system is said to be well posed if the transfer
Clear1y if a loop gain i
function of every subsystem is proper and the c1osed-loop transfer function
zero at s = oo. lf aH t
from any point chosen as an input terminal to every other point along the
has a nonzero loop ga
directed path is well defined and proper. I
loop are exactIy prope
degree of its numerato
In this definition, if a point, say E in Figure 3-26, is chosen as an input
only loops with nonze
terminal, then we must add a fictitious input as shown. Then the c1osed-loop
Consider the feedb:
transfer functions from r to e, W, and y are, respectively.
not well posed becaus
e(s)= ---, [1 + G 3 (S)G 1 (s) + G 4 (s)Gz(s)G 1 (S)]-l (S) r loop with loop gain +
w(s) = - G 3 (S)G 1 (s)[1 + G 3 (S)G 1(s) + G 4 (S)G 2 (S)G 1(s)] -l r (S) (3-74) improper transfer func
conc1ude that if a systel
and y(s) = - Gi(s)G 1 (s)[1 + G 3(s)Gi(s) + G 4 (S)G 2(S)G 1(s)] -l r (S)
+ 1 at s = 00 which is
We note that in computing these transfer functions, no branch in the system function.
should be disconnected. Hence we compute c1osed-loop transfer functions. The system in FigUJ
lf we disconnect any branch, then the system becomes a different system. Go(00)=1-1 +2#0,
from the existence of 2
Theorem 3-6 loop will offset the 100
The system in Figure 3-26, where G¡(s) are rational transfer matrices of appropri­
ate orders, is weH posed if and only if Gi(s), i = 1, 2, 3,4, are proper and the
rational matrix ~~- '---.J

(3-75)

exists and is proper, or equivalentIy, the constant matrix


Go(oo)= I +G 3 (00)G 1 (00) +G 4 (00)G 2 (00)G 1(00)
is nonsingular. (a)

Proof
The transfer matrices in (3-72) to (3-74) are c1early proper if and only if G l(S) o
o
is proper. Similarly, it can be shown that if G l(S) is proper, the transfer matrix

(e)
13 This definition is similar to the one in Reference S34 and is applicable only to linear time-invariant
lumped systems. For a more general definition and discussion, see References SS, S2Ü7 and S214. Figure 3-27 Feedback
MATHEMATlCAL DESCRIPTlONS OF COMPOSITE SYSTEMS 119

singular. Will such a from any point to any other point along the directed path is proper. This
tr, the transfer matrix establishes the first part of the theorem. If a11 Gi(s), i = 1, 2, 3, 4, are proper.
I contains a part which the polynomial part of Go(s) is Goloo). Hence (;0 I(S) is, fo11owing Corollary 3-4,
mplitude of e(t) will be proper if and only if Go((0) is nonsingular. Q.E. D.
1m out. Rence, in the
overall transfer matrix We discuss now the implications ofthe conditions det [1 + G2 ( CIJ )G 1 (CIJ)J f- O
put pairs of the system and more genera11y, det Go(oo) f- O. Before proceeding, we need sorne concepts.
fied, and in sorne sense, A loop of a block diagram is a closed path which travels from a point along the
preserved. direction of the path back to the same point and does not pass any point twice.
For example, the feedback system in Figure 3-25 has three loops; one of them
is indicated by the heavy lines. The loop gain of a loop is defined as the product
of a11 transfer functions along the loop including the signs at summing points.
e by a rational transfer
In our discussion, we are interested in only the value of the loop gain at s = oo.
II po sed if the transfer
Clearly if a loop gain is a strictly proper rational function, then its loop gain is
loop transfer function
zero at s = CIJ. If a11 transfer functions of a block diagram are prdper, a loop
other point along the
has a nonzero loop gain at s = 00 if and only if a11 transfer functions along the
I
loop are exactly proper (a rational function is said to be exactly proper if the
degree of its numerator is equal to that of its denominator). In the following,
is chosen as an input
only loops with nonzero loop gains at 5=00 will come into the discussion.
Then the closed-Ioop
Consider the feedback systems shown in Figure 3-27(a) and (b). Theyare
not well posed because det Go(oo) =0 in both systems. Each system has a
'(5)
loop with loop gain + 1 at s = 00, which is equivalent to a gain of infinity or an
)G 1(5)]-11-(5) (3-74)
improper transfer function as shown in Figure 3-27(c) and (d). Hence, we may
)G 1(5)] -11-(5)
conclude that if a system is not well posed, the system has a loop with loop gain
+ 1 at s = CIJ which is equivalent to an infinite gain or an improper transfer
o branch in the system function.
10P transfer functions. The system in Figure 3-28(a) has a loop with loop gain 1; however, we have
dilTerent system. Go(oo) = 1-1 +2 f-O,and the system is we11 posed. This discrepancy arises
from the existence of another loop with a nonzero loop gain at s = oo. This
loop will offset the loop with loop gain 1, as can be seen from Figure 3-28(c)
:r matrices of appropri­
4, are proper and the

(3-75)

(a) (b)

er if and only if G I(S)


o
per, the transrer matrix

(e) (d)
: onlyto linear time-invariant
teferences S8, S207 and S214. Figure 3-27 Feedback systems which are not well po sed.
120 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

or an improper transfe
+ }--E..-_ _---+I 2s + 1 to be well posed.
-s­

3-7 Discrete-TiI

The inputs and outpul


defined for all t in ( - e
(a) (b) invariant case. Theya
study a different class I

outputs of discrete-tirr
For example, a digital
of variables at discrete
continuous-time syster
responses are of intere:
For convenience, ti
output appear will be ;
(e) (d) time interval T is cal
{y(k) ~ y(kT)}, k = O, j
Figure 3-28 (a) A well-posed system. (b) An ill-posed system. (e), (d) Equivalent
In the following, we di
systems of (a).
the continuous-time S)
called a relaxed discre,
and (d). Henee, if there are two or more loops with nonzero loop gains at discrete-time system sa
s = 00 passing the same point, such as the point E shown in Figure 3-28(a),
these loops must be grouped into a combined loop with a net loop gain equal to
the sum of all individual loop gains. For the system in Figure 3-28(a), the
combined loop has a net loop gain 1 - 2 = -1 which is different from + 1; where g(k, m) is called t
henee the loop will not cause any problem, and the system is well posed. The due to the application (
system in Figure 3-28(b) has a combined loop with loop gain 3 - 2 = 1; hence
the system is not well posed (Problem 3-39).
From these examples, we conclude that if a system has a combined loop with
net loop gain 1 at s = 00, the combined loop is equivalent to a gain of infinity with the system relaxe
or an improper transfer function. Note that what causes the problem is the that is, the output doe
combined loop instead of individual loops. ylk? = Q [ol' k < ;Ti. "
To further verify the aforementioned statement, we consider the system relaxed at ka, then (3-7(
shown in Figure 3-25. There are two loops passing through the point H.
Along the loop indicated by the heavy lines, there is a loop between points
E and F. In computing the loop gain of the heavy-lined loop, the transfer
function from E to F should be computed. However because of the two strictly lf a linear causal relaxe
proper transfer functions, the loop gain is zero at s = oo. Hence, the combined g(k, m) = g(k - m) for al]
loop which passes through the point H has a net loop gain 1 at s = oo. This ka = Oand the set of tim
loop is equivalent to again of infinity. Similarly, the combined loop which linear time-invariant cal
passes the point J has anet loop gain 1 at s = oo. The transfer function from
Uz to yz through the loop is (s + 2)/1, which is improper. For the system in y(k) =
Figure 3-25, wehave det [1 +GkD)G¡(OO)] =0.
In conclusion; a system is well posed if and onlyif the system has nocom­
bined loop with net loop gain 1 at s = 00, which is equivalentto an infinite gain 14 This section maybe skippe<
D/SCRETE-TIME SYSTEMS 121

or an improper transfer function. In practice, every system should be designed


to be well posed.

3-7 Discrete-Time Systems 14

The inputs and outputs of the systems we studied in the previous sections are
defined for all t in ( - 00, 00) for the time-varying case or in [0, ro) for the time­
(b) invariant case. They are called continuous-time systems. In this section we shall
study a different class of systems, called discrete-time systems. The inputs and
outputs of discrete-time systems are defined only at discrete instants of time.
For example, a digital computer reads and prints out data that are the values
of variables at discrete instants of time; hence it is a discrete-time system. A
continuous-time system can also be modeled as a discrete-time system if its
responses are of interest or measurable only at certain instants of time.
For convenience, the discrete instants of time at which the input and the
output appear will be assumed to be equally spaced by an amount of T The
(d)
time interval T is called the sampling periodo We use {u(k) ~ u(kT)} and
{y(k) ~ y(kT)}, k =0, ± 1, ±2, ... , to denote the input and output sequences.
temo (e), (d) Equivalent In the following, we discuss only single-variable discrete-time systems. As in
the continuous-time system, a discrete-time system that is initially relaxed is
called a relaxed discrete-time system. lf the inputs and outputs of a relaxed
lonzero loop gains at diserete-time system satisfy the linearity property, then they can be related by
)wn in Figure 3-28(a), .00

. net loop gain equal to y(k) = L g(k, m)u(m) (3-76)


-00
in Figure 3-28(a), the
is different fram + 1; where g(k, m) is called the weighting sequence and is the response of the system
~m is wel1 posed. The
due to the application of the input

{~
) gain 3 - 2 = 1; hence i=m
u(i) =
ifm
; a combined loop with
nt to a gain of infinity with the system relaxed at time m-o lf the discrete-time system is causal­
,es the problem is the that is, the output does not depend on the future values of the input-then
g(le, m) = O for le < m. Consequentíy, if a discrete-time system is causal and is
;: consider the system relaxed at ko, then (3-76) reduces to
k
through the point H.
L loop between points
y(k) = L g(k, m)u(m) (3-77)
m~ko
ned loop, the transfer
ause of the two strictly lf a linear causal relaxed discrete-time system is time invariant, then we have
g(k, m) = g(k - m) for all k ~m. In this case, the initial time is ehosen to be
Benee, the combined
gain 1 at s = ro. This °
k o = and the set of time of interest is the set of positive integers. Bence for a
linear time-invariant causal relaxed discrete-timesystem, we have
combined loop which
k
transfer function from
;:r. F or the system in y(k) = ¿ g(k - m)u(m) k =0, 1, 2, ... (3-78)
m~O

le system has no com­


14 This section may be skipped withoutloss in continuity.
,lent to an infinite gain
122 MATHEMATICAL DESCRIPTlONS OF SYSTEMS

Comparing with continuous-time systems, we see that in discrete-time omitted. We introd­


systems, we use summation instead of integration; otherwise, all the concepts time-varying, discrete
are the same. In the continuous-time case, if we apply the Laplace transform,
a convolution integral can be transformed into an algebraic equation. We
have the same situation here, the transformation which will be used is cal1ed
the z-transform. where x is the state VI
time dynamical equa
Definition 3-10 set of first-order diffe
The z-transform of the sequence {u(k), k =0,1,2, ...} is defined as If A(k), B(k), C(k),
00

u(z) ~ Q[u(k)] ~ L U(k)Z-k


k=O
where z is a complex variable. I which is called a finé
x(z) be the z-transforr
Example 1

lf u(k) = 1 for k = O, 1,2, ... , then


That is,
A ~ k 1 z
u(z) = L., z- = =-­ I
k=O 1-z- 1 z-l

Example 2 Then
lf u(k) =e- 2k for k =0,1,2, ... , then
00 ~[x(k+
z
u(z) = L e- 2kz- k I
k=O

Now we shall apply the z-transform to (3-78). For a causal, relaxed system, Hence the application
we have g(k - m) = O for k < m; hence we may also write (3-78) as
00

y(k) = L g(k - m)u(m)


m=O

Consequently, we have
00 00 00
x(z)
y(z) = L y(k)Z-k = L L g(k - m)u(m)z-(k-lII l z -111
= (

k=O k=O 111=0 y(z) = t


00 00

= L L g(k - m)z-(k-lIIlu(m)z-m They are algebraic eql


m=Ok=O the same as (3-35) and
00 00

= L g(k)Z-k L u(m)z-III =g(z)u(z) (3-79) . y(z


k=O m=O
where G(z
Here we have changed the order of summations and used the fact that g(k - m) =
O for k < m. The function g(z) is the z-transform of the weighting sequence' is called the sampled ,
{g(k)}k'=o and is called the z-ú'ansfer function or sampled transfer function. identical to the coritin
The extension of the input~output description of single-variable discrete­ Similar to the co
time systems to the multivariable case is straightforward and its discussion is practical interest are p
- ----------,-----­

OISCRETE-TIME SYSTEMS 123

that in diserete-time omitted. We introduce now the diserete-time dynamical equation. A linear,
rwise, aH the concepts time-varying, discrete-time dynamical equation is defined as
~he Laplace transform,
DE: x(k + 1) = A(k)x(k) + B(k)u(k)
;ebraie equation. We y(k) = C(k)x(k) + E(k)u(k)
will be used is caBed
where x is the state vector, u the input, and y the output. Note that a diserete­
time dynamieal equation is a set of first-order difference equations instead of a
set of first-order differential equations, as in the continuous-time case.
~fined as If A(k), B(k), C(k), and E(k) are independent of k, then DE reduces to

DFE: x(k + 1) = Ax(k) + Bu(k)


y(k) = Cx(k) + Eu(k) (3-80)

• which is caHed a linear, time-invariant, discrete-time dynamical equation. Let


x(z) be the z-transform of

{x(k)}k'=o
That is,

• x(z) ~ .?l'[x(k)J ~
<X)

¿ X(k)Z-k
k=O
Then
<X) <X)

.?l'[x(k+ l)J = ¿ x(k +l)z-k=z¿ x(k +l)z-<k+l)


I
k=O k=O

=Z[ k=~ 1 x(k + l)z-(k+ 1) -X(O)] = z[x(z) -x(O)J


causal, relaxed system, Hence the application of the z-transform to (3-80) yields
(3-78) as
zx(z) - zx o = Ax(z) + Bu(z) (3-81 )
Hz) =Cx(z) + Eu(z)

where x(O) = Xo Equation (3-81) CHD be arranged as

x(z) = (zI - Af 1 zx o + (zI - Af 1 Bu(z) (3-82)

y(z) = C(zI - A)- 1 Zxo + C(zI - A)-l Bo(z) + Eu(z) (3-83)

They are algebraic equations. The manipulation of these equations is exaetly


the same as (3-35) and (3-36). Ifx(O) = 0, then (3-83) reduces to
(3-79) y(z) = [C(zI -A)-lB +EJu(z)~ G(z)o(z)
where G(z)=C(zI-A)-lB+E (3-84)
thefact that g(k - m) =
IS called the sampled transferlunction matrix of (3~80). We see that (3-84) IS
le weighting sequence
'd tri:msfer function.
identical to the continuous-time case in (3-36) if z is replaced by s.
ngle-variable discrete­ Similar to the continuous-timecase,. the sampled transfer functions of
'd and its discussion is practical interest are proper rational functions. Consider the improper transfer
124 MATHEMATICAL DESCRIPTIONS OF SYSTEMS

function whereP, Q, R and W


is called the polynomü
A Z2+2 1 2 1
g(z)=--=z+1+3z- +3z- +3z-'+'" Both the input-ou
z-l in this chapter are use
The inverse z-transform of 9(z) is g( -1) = 1, g(O) = 1, g(k) = 3, k = 1, 2, 3, .... on the problem, on tl
Its impulse sequence is not zero for g(k) 1= Ofor k < O; hence {g(k)} is not causal.
Consequently, an improper rational function does not describe a causal system. Problems
AH physical systems are causal; hence we are interested in only proper sampled
transfer functions. This reason is different from the continuous-time case 3-1 Consider the merr
where differentiations due to improper transfer functions are to be avoided whieh u denotes the in¡:
because of high-frequency noises. possible to introduce a n
The topics discussed in Sections 3-5 and 3-6 are directly applicable to the
discrete-time case without any modification. 3-2 We may define th
response due to a b-fun<
describing the system?
3-8 Concluding Remarks
3-3 The impulse respol
In this chapter we have developed systematically the input-output description tand T. Is this system e:
and the state-variable description of linear systems. Although the input­
output description can be obtained by analysis, it can also be developed from 3-4 The impulse respor
the measurements at the input and output terminals without knowing the
internal structure ofthe system. For linear, time-invariant systems, the transfer­
function description can also be used. Whenever transfer functions (including
where w and t o are const:
impedances and admittances in network theory) are used, the systems are ideallow-pass filter in tho
implicitly assumed to be relaxed at t =0.
The condition for a set of first-order differential equations x =f(x, u, t), 3-5 Consider a relaxed
y =g(x, u, t)-in particular, x = A(t)x +B(t)u, y =C(t)x + E(t)u-to be qualified
as a dynamical equation is that for any initial state X o and any input u[to.co)'
there is a unique solution Y[to. co) satisfying the equations. This uniqueness
condition is essential in the study of solutions of a dynamical equation. The for any u, where C( is a fix
dynamical equations studied in this book are of the form x = A(t)x + B(t)u, ehops off the input after t
y =C(t)x +E(t)u. They can be extended to include derivatives ofu in the output
3-6 Let P a be the trunca
equation, such as
relaxed system y = Hu. w(
yU) =C(t)x(t) +E(t)u +E¡U)uU) +E 2 U)üU) + ...
However, these extensions are of limited interest in practice. This faet is often used in ~

One may wonder why we study a set of first-order differential equations y


instead of studying higher-order differential equations. The reasons are as
follows: (1) every higher-order differential equation can be written as a set of
first-order differential equations, (2) the notations used to describe first-order
equations are compact and very simple, and (3) first-order differential equations ---,..¡.....':----. II

can be readily simulated on an analog or a digital computer.


In addition to transfer functions and dynamical equations, one may' also
encounter the following description . .
(a)
P(s);(s) = Q(s)íi(s)
y(s)=R(sj~(s) +W(s)u(s) Figure P3~1
PROBLEMS 125

where P, Q, R and W are polynomial matrices, in the analyses. This description


is called the polynomial matrix system description and will be studied in Chapter 6.
Both the input-output description and the state-variable description studied
in this chapter are useful in practice. Which description should be used depends
7(k) = 3, k = 1,2,3, .... on the problem, on the data available, and on the question asked.
lce {g(k)} is not causal.
scribe a causal system. Problems
1 only proper sampled
continuous-time case 3-1 Consider the memoryless systems with characteristics shown in Figure P3-1, in
IllS are to be avoided which u denotes the input and y the output. Which of them is a linear system? ls it
possible to introduce a new output so that the system in Figure P3-1(b) is linear?
~ctly applicable to the
3-2 We may define the impulse response of a relaxed nonlinear system gL r) as the
response due to a ¿¡-function input applied at time r. ls this impulse response useful in
describing the system? Why?

3-3 The impulse response of a relaxed linear system is found to be g(t, r) = e- h -d for al!
mt-output description t and r. Is this system causal? Is it time invariant?
Although the input­
Iso be developed from 3-4 The impulse response of an ideallow-pass filter is given by
without knowing the sin 2w(t - to)
g(t)=2w---­ for aH t
tt systems, the transfer­ 2w(t - to)
er functions (including where w and t o are constants. Is the ideallow-pass filter causal? ls it possible to build an
used, the systems are ideal low-pass filter in lhe real world?

equations X =f(x, u, t),


E(t)u-to be qualified U(t)
and any input u(to,oo)' y(t) = (p.u)(t) ~ { O
for t > ex
ms. This uniqueness
3.mical equation. The for any u, where ex is a fixed constant. In words, this system, called a truncation operator,
chops off the input after time ex. ls this system linear? ls it time invariant? ls it causal?
form X = A(t)x + B(t)u,
ttives ofu in the output
3-6 Let Pa be the truncation operator defined in Problem 3-5. Show thal for any causal
relaxed system y = Hu, we have
) + ... p.Y =P.Hu =PaHPau

tice. This fact is often used in stability studies of nonlinear feedback systems.

y y y
differential equations
The reasons are as
be written as a set of
to describe tirst-order _ _ _-,//c-,--­ u
----:*"------u --07"''''---+:,----~ u
r differential equations
uter.
llations,one may also

(a) (b) (e)

Figúre P3-1
126 MATHEMATICAL DF5CRIPTIONS OF SYSTEMS

3-7 In Problem 3-6, is it true that 3-14 The input u and t


(P.y)(t) = (P.Hu)(t) = (H p.u)(t)
for al! t in (- 00, oo)? Ir not, find the interval of time in which the equation holds. What is the transfer fune

3-8 Consider a linear system with input u and output y. Three experiments are per­ 3-15 Consider a multi'
formed on this system using the inputs u¡(t), uz(t), and U3(t) for t ;::::0. In each case, the the ijth element of (;(s) e
initial state at t =0, x(O), is the same. The corresponding observed outputs are y¡(t), h(t),
and Y3(t). Which of the following three predictions are true if X(O) 1= O?
a. lfu 3 =u¡ +uz, then Y3 = Y¡ +Yz· where Yi is the ith compe
b. lfU3=1(u¡ +uz), thenY3 =1(Y¡ +Yz).
c. lfu 3 =u¡ -uz, then Y3 = Y¡ - Yz· 3-16 Consider a multiv
Which are true if x(O) = O? (Answers: No, yes, no, for x(O) 1= O; all yes, if x(O) = O.)

3-9 Show that if H(u¡ + u z ) = Hu¡ + Hu z for any u¡, u z, then Hrxu =rxHu for any rational
number rx and any u. where the N;]s and D;]s:
of the system?
3-10 Show that for a fixed rx, the shifting operator Q. defined in Figure 3-5 is a linear time­
invariant system. What is its impulse response? What is its transfer function? Is this 3-17 Find the dynamic
transfer function a rational function? lf 8,8 ¡, and 8 z are very srr
functions and dynamical
3-11 The causality of a relaxed system may also be defined as follows: A relaxed system
is causal if and only if u¡(t) = uz(t) for all 15,10 implies (Hu¡)(t) = (Huz)(t) for aH 15,1 0 , 3-18 Find the dynamic
Show that this definition implies that y(t) = Hu(_ 00,1]' and vice versa. and 3-23.

3-12 Let g(t, L)=g(t +rx, L +rx) for all t, L, and rx. Define x =t +L,y=t -L, then g(t, L)= 3-19 Find the dynamic:
g«x + y)/2, (x - y)/2). Show that ag(t, L)/aX = O. [From this fact we may conclude that if network in Figure P3-19.
g(t, L) = g(t + rx, L + rx) for all t, L, and rx, then g(t, L) depends only on t - L.]

3-13 Consider a relaxed system that is described by

y(t)= Ig(t-L)U(L)dL \ 1

,,(,,~
If the impulse response 9 is given by Figure P3-13(a), what is the output due to the input
shown in Figure P3-13(b)? (Use graphical method.)

[m g

u(tJ (a)

Figure P3-17

g(tJ
E-­
c¡ Cz
o 2

bl-----t-I
R1
L

1 - - - + - 1-----,.-,' t
-1 + t II Z

11 1 rv Rz
o '2 3 4
(a) (b)

Figure P3-13 Figure P3-19


PROBLEMS 127

3-14 The input u and the output y of a system is described by

ji+2¡i+3y=2ú+u

e equation holds. What is the transfer function of the system?

'ee experiments are per­ 3-15 Consider a multivariable system that is describable by y(s) = G(s)íI(s). Show that
t 2': O. In each case, the the íjth element of G(s) can be defmed as
,d outputs are Yl(t), yz(t),
O) 1= O?
A

9 ij\S) =
If[Yi(t)]
ro-[
()J
I
iniLially relaxed
.;L Uj t and u,~Ofork-l=j
where Yi is the ith component of y and Uj is the jth component of u.

3-16 Consider a multivariable system whose inputs and outputs are described by
LIl yes, if x(O) = O.) N 11 (P)Yl(t) + N 12(p)Yz(t) = D 11 (p)u 1(t) + D ¡z(p )uz(t)

N ZI (p )Yl (t) + N zz(p)Yz(t) = D 21 (p)u 1(t) + Dzz(p)uz(t)

u = rxH u for any rational


where the No/s and D¡Js are polynomials of p ~ d/dt. What is the transfer-function matrix
of the system?
igure 3-5 is a linear time­
3-17 Find the dynamical-equation descriptions of the systems shown in Figure P3-17.
ansfer function? Is this
Ife, e e
l' and z are very smal!, can you consider the two systems as linear? Find the transfer
functions and dynamical equations to describe the Iinearized systems.
Jl!ows: A relaxed system
)= (HuzXt) for al! 1:::; too 3-18 Find the dynamical-equation descriptions of the networks shown in Figures 3-22
:sa. and 3-23.

't, Y = t -'t, then g(t, 't) = 3-19 Find the dynamical-equation description and the transfer-function matrix of the
we may conclude that if network in Figure P3-19.
)ll t - 't.]

: output due to the input


12

(a) (b)

Figure P3-17

(b)

Figure P3-19
128 MATHEMATlCAL DESCRIPTlONS OF SYSTEMS

3-20 Consider the simplified model of an aircraft shown in Figure P3-20. lt is assumed 1F
that the aircraft is dynamically equivalent at the pitched angle eo, elevator angle uo, altitude
ha, and cruising speed vo. lt is assumed that small deviations of e and u from eo and Uo
E-­
e
generate forces JI = k I and J2 = k 2u, as shown in the figure. Let m be the mass of the
-I.Q
aircraft, 1 the moment ofinertia about the center of gravity P, be the aerodynamic damping,
and h the deviation of the altitude from h o. Show that the transfer function from u to h Currenl
is, by neglecting the effect of 1, source t

3-21 The soft landing phase of a lunar module descending on the moon can be modeled Figure P3-23
as shown in Figure P3-21. lt is assumed that the thrust generated is proportional to til,
where m is the mass of the module. Then the system can be described by my = - krh - mg,
where 9 is the gravity constant on the lunar surface. Define the state variables of the
where gl(l, t o ) is the step r
system as XI = y, X2 = y, X, = m, and u = rh. Find the dynamical-equation description of
[Hint: This can be proved
lhe system.
(3-55).J

3-22 Show that the output of a linear causal relaxed system due to the input u(t)i5 I(t - to)
3-23 Find the transfer fu
is given by
in Figure P3-23. Do you
y(t) = u(to)gl(t, to) + l'
'o
U(T)gl(t, T) dT for t ~to
3-24 Show that single-va
in the sense that the order
h true for the time-varying s

3-25 The impulse and Stl


relaxed system are, by del
e function and i5 I is a step I
using the property given in
h~
:C--~----­ 3-26 Verify that the impl
given by Equation (3-60).
~12~11r­
3-27 Verify the identity
Figure P3-20
where El is a q x p matrix.

3-28 Verify (3-64) by usin¡

3-29 Show that

l
I
Th rus! = kli¡
det(I"
I

Lun"ar ·surúlce Note that the matrix in the


rank 1, show that det(I n +
Figure P3-21
elements of G. [lf G is not
PROBLEMS 129

~ure P3-20. lt is assumed


.eleva tor angle uo, altitude
::>f 8 and u from 80 and Uo +
Let m be the mass of the
-In
¡he aerodynamic damping, In
nsfer function from u to h Curren! y
source f

~l
IH

the moon can be modeled Figure P3-23


ated is proportional to rñ,
;ribed by mji = - klil- mg,
the state variables of the where 9 1(t, to) is the step response and b ¡ is the step function defined in (3-53) and (3-54).
:al-equation description of [Hin!: This can be proved either by decomposing u into a sum of step functions or by using
(3-55).]

le to the input u(t)O¡(t - to) 3-23 Find the transfer function and the dynamical-equation description of the network
in Figure P3-23. Do you think the transfer function is a good description of this system?

t"2:to 3-24 Show that single-variable, linear, time-invariant, causal, relaxed systems commute
in the sense that the order of the tandem connection of two systems is irnmaterial. ls this
true for the time-varying systerns?

3-25 The impulse and step responses of a single-variable, linear tirne-invariant, causal,
relaxed systern are, by deftnítion, given by 9 =Hb(t) and 9 ¡ = Hb¡(t), where b is a delta
d
fllnction and 15¡ is a step function. lt can be shown that b(t) =-b¡(!). Verify (3-56) by
d!
lIsing the property given in Problem 3-24.

3-26 Verify that the impulse-response rnatrix of the feedback systern in Figure 3-24(c) is
given by Equation (3-60).

3-27 Verify the identity


I p - Ez(Iq + E¡E z )- 1 El = (lp + EzE¡)- I

where E¡ is a q x p matríx and E z is a p x q matrix.

3-28 Verify (3-64) by using the identity in Problem 3-27.

3-29 Show that

Note that the rnatrix in theleft-hand side isan.n xn rnatrix~ Let'G be an 11 x 11 matrix of
rank 1, show that det(ln+G)= 1 + trace G, where trace G is thesumof aH diagonal
elements of G. [H Gis not of rank 1, seeEquation (9-19) of Chapter 9.]
130 MATHEMATlCAL DESCRIPTlONS OF SYSTEMS

3-30 Find the transfer-function matrix of the feedback system shown in Figure 3-24(c),
where the transfer-function matrices of SI and Sz are, respectively,

ss +1:2J 2.5

s +2

'8
Y-B
3-31 Find the dynamical-equation description of the feedback system in Figure 3-24(c),
where SI and Sz are, respectively, described by
0.5

[~IIJ=[-2O
XIZ
_IJ[XuJ
1 XIZ
+[_ 4
lJU I
1 2
YI =[0 lJx l +[1 -IJu I
Figure P3-35
and [~::J=[~}z
YZ=[~ _~JXZ
Draw a block diagram of a computer simulation of this feedback system.

3-32 Prove Theorem 3-2 by using det N = det P = 1 and det NQ = det PQ.

Figure P3-37
3-33 Find the overall transfer matrix ofthe feedback system shown in Figure 3-24(c)with

3-36 A function h(x(t), u((


s

l
-1

'i' J
A s +1 al h(xl(t), uI(t»+
GI(s)= O Gz(s)= _1_
[ for any real numbers a¡, a¿
s +1 linear function ofx(t) and [:

Can you find a combined loop with a net loop gain 1 at s = oo?
for sorne A and B. If h, x,
3-34 Which of the systems in Figure P3-34 have improper overall transfer functions? does the assertion hold? I

3-35 Can you find a combined loop with a net loop gain 1 in the system shown in Fig. 3-37 Consider the multiv
P3-35? What is its overall transfer matrix? following composite lransfe
e(s)J
[ u(s) =
[(1 +
C(s)(

3-38 The transfer matrix (


have a different form of Gf(
as in Equation (3-68)?

3"39 Show that the systen


transfer function in the syst,

Figure P3-34 3-40 Show that the parall


G
posed if and only if I(s) al:
PROBLEMS 131

1shown in Figure 3-24(c),


Iy, Y¡

: system in Figure 3-24(c),

Figure P3-35

+ e

:k system.

Q =detPQ.

Figure P3-37
own in Figure 3-24(c)with

3-36 A function h(x(t), uU)) is said to be a linear function ofxU) and u(t) if and only if
(;( ¡ h(x ¡ (t), U ¡ (t)) + (;(z h(xz(t), uz(t)) = h«(;( ¡ x ¡ (t) + (;(zxz(t), (;( ¡ u 1 (t) + (;( zuz(t))
for any real numbers (;(¡, (;(z, any x¡(t), xz(t), and any u¡(t), uz(t). Show that h(x(t), u(t)) is a
linear function ofx(t) and u(t) ifand only ifh is of the form
h(x(e), u(e)) = A(e)x(e) + B(e)u(e)
for sorne A and B. If h, x, and u are square matrix functions instead of vector functions,
erall transfer functions? does the assertion hold? If not, what modification do you need?

3-37 Consider the multivariable feedback system shown in Figure P3-37. Verify the
\ the system shown in Fig.
following composite transfer matrix:
e(s)] [(1 + (;(s)C(s)t ¡ - (;(s)(1 +C(S)(;(S))-¡][f(S)]
[ u(s) = C(s)(1 + (;(s)C(s))- 1 (1 +C(s)G(s))-¡ 3(s)

3-38 The transfer matrix of the system in Figure 3-26 is given in (3-73). Is it possible to
have a different form of (; ¡(s) with (;¡(s) or (;z(s) on the right-hand side of the parentheses
as in Equation (3-68)?

3-39 Show that the system in Figure 3-28(b) is not well posed by finding an improper
transfer function in the system.

3-40 Show that the parallel andtandem connections in Fi6ure 3-24(a) and (b) are wel1
posed if and only if (; ¡(s) and (;z(s) are proper.
132 MATHEMATlCAL DESCRIPTIONS OF SYSTEMS

3-41 Let (;(s) = G o +G1s- 1 +G 2s- 2 + ... and C(S) =C o +CIS- I +C 2S- 2 +.... Show
that (1 +C(s)(;(s)) -1 is proper if and only if (1 +C( <X) )(;( co)) = (1 +CoG o) is nonsingular.
Prove it directly by using the power series without using Theorem 3-5.

3-42 A rational matrix (;(s) is proper if (;(00) is a finite constant matrix and improper
if (;( (0) is not a finite constant matrix. Are the following statements valid?
1. Ir (;I(S) and (;2(S) are proper, (;2(S)(;I(S) is proper.
2. If (;I(S) and (;2(S) are improper, (;2(S)(;I(S) is impropero
Answer: Yes; no. Consider (;2(S)(;¡(S) = (;2(S)U(s)U- 1(S)(;I(S). Let (;¡(s) be 2 x 2 proper
rational matrices and let l
S" +1
U(s) = [ 1

3-43 Let (;(s) be a q x p rational matrix, not necessarily proper. Show that the rational
matrix
(;I(s) = (;(s)(I + K(;(s)r 1
is proper for almost aH p x ~ const~nt matrix K (or, in mathematical terminology, for a
generic K). [Hint: Express G(s) as G(s) = N(s)D-1(s), where N(s) and D(s) are polynomial
matrices, and use Theorem G-9. See Reference SI96.]

3-44 Let M(s) be a nonsingular rational matrix decomposed uniquely as


4-1 IntroductioDl
M(s) = Mp(s) + Msp(s)
where Mp(s) is a polynomial matrix and Msp(s) is a strictly proper rational matrix. Show Linear systems can be e
that if Mp(s) is singular, M-1(s) is not proper. [Bint: Write Msp(s)=N(s)D-1(s)= output description and
(N(s)U(s))(D(s)U(s)r 1, where N(s), D(s), and U(s) are polynomial matrices, and write are obtained, the next
of analyses: qualitative
M-1(s) = (D(s)U(s))(Mp(s)D(s)U(s) + N(s)U(S))-1
interested in the generé
If Mp(s) is singular, so is Mp(s)D(s). Consequently, there exists a nonsingular U(s) such stability, of the equati<
that at least one column, say the jth column, of Mp(s)D(s)U(s) is a zero column. Then the quantitative analysl
apply Theorem G-9.]
due to sorne excitations
can be used to carry Ol
hO\iVeVer are not in eJe,::;·
from these solutions. I
of the input-output de~
tionships between these
be discussed are applic:
descriptions are first co
If the irnpulse-respc
input, the output"y can

by direct compuíation c·
the equation Y(s) = G(s
matrix G(s) is a rationa
lS-1+C 2S- 2 + .... Show
,(1 +CoG o) is nonsingular.
rem 3-5.

stant matrix and improper


:ments valid?

j. Let G¡(s) be 2 x 2 proper


4
Linear Dynamical Equations
and Impulse-Response
,er. Show that the rational Matrices

~matical terminology, for a


i(s) and D(s) are polynomial

uniquelyas
4-1 Introduction

'per rational matrix. Show


Linear systems can be described, as shown in the previous chapter, by the input­
Trite Msp(s)=N(s)D-1(s)= output description and the state-variable description. Once these descriptions
tI matrices, and write are obtained, the next step is natural1y to analyze them. There are two types
of analyses: qualitative and quantitative. In the qualitative analyses, we are
J(s)) - 1
interested in the general properties, such as controllability, observability, and
sts a nonsingular U(s) such stability, of the equations. These wil\ be discussed in Chapters 5 and 8. In
s) is a zero column. Then the quantitative analyses, we are interested in the exact responses of equations
due to some excitations. Digital computers are now widely available, and they
can be used to carry out these analyses (see Section 3-3). Computer solutions
however are not in closed forms. and it is difficl11t to extrapolate any propert\es
from these solutions. In this chapter we shal1 study the closed-form solutions
of the input-output description and the state-variable description. The rela­
tionships between these descriptions will also be studied. We note that what will
be discussed are applicab1e to composite systems if their overal1 mathematical
descriptions are first computed.
If the impulse-response matrix G(t, r) of a system is known, then for any
input, the output y can be obtained from

y(t)= L G(t, r)u(r) dr

by direct computation or by a graphical method. In the time-invariant case,


the equation y(s) = G(s)íi(s) can also be used. Unless the transfer-function
matrix G(s) is a rational matrix, general1y it is easier to compute y directly in
133
134 LINEAR DYNAMICAL EQUATlONS AND IMPULSE-RESPONSE MATRICES

the time domain than from the frequency domain. It is elementary to compute Theorem 4-1
y from the input-output description; hence it will not be discussed further.
Solutions of linear dynamical equations are studied in Section 4-2. Solu­ The set of all solutio
tions are stated in terms of the state transition matrix <.I>(t, r), which is the over the field of real 1
unique solution of
Proof
o
ot <.I>(t, r) = A(t)<l>(t, r) <.I>(r,r)=I Let WI and W2 be tW(
solution of (4-2) for al
In the time-invariant case, we have <D(t, r) = eA(!-t). Various methods for the d
computation of eAI and (sI - A)-I are discussed. In Section 4-3 the concept of dt (ex l W¡ +ex
equivalent dynamical equations is introduced. Equivalent dynamical equa­
tions are obtained by changing the basis of the state space. We show that
every time-varying linear dynamical equation has an equivalent linear dy­
namical equation with a constant A matrix. We al~o establish the theory Hence the set of solut
space of (4-2). We n
01 Floquet. In the last section, the relation between linear dynamical equations
el' e 2, ... , en be any li
and impulse-response matrices is studied. The necessary and sufficient con­
tions of (4-2) with th
dition for an impulse-response matrix to be realizable by a linear dynamical
equation is established. We also show that every proper rational matrix has a show that W¡, for i = L
linear time-invariant dynamical-equation realization. of(4-2) can be written
The references for this chapter are 24,31,60,68, 77, 109, 114, and 116. assertion is proved. \
independent. Suppos.
definition, there exists

4-2 Solutions of a Dynamical Equation


Note that the O in the
Time-varying case. Consider the n-dimensional linear time-varying space; therefore, it is r
dynamical equation
[WI(t) W2 1
E: x(t) = A(t)x(t) + B(t)u(t) (state equation) (4-1 a)
In particular, we have
y(t) = C(t)x(t) + E(t)u(t) (output equation) (4-1 b)
[WI(t O) W;
where A('), B('), C('), and E(') are n x n, n x p, q x n, and q x p matrices whose
entries are real-valued continuous functions of t defined over ( - 00, 00). Since which implies that e j ,
A(') is assumed to be continuous, for any initial state x(to) and any Ql!, there existe: dicts the hYDothesis: h
a unique solution in the dynamical equation E. This fact will be used frequently ( - 00, 00).
in the following development. Before studying the entire dynamical equation Let Wbe any solut
E, we study first the solutions of the homogeneous part of E; namely, linearly independent "
be written as a unique:
X =A(t)x (4-2) as

Solutions of X =A(t)x
1t is clear that
The set of first-order differential equations in (4-2) has a unique solution for
every initial state X o in (W,~). Since there are infinitely many possible initiaJ
states, Equation (4~2) has infinitely many possible solutions. This set of SOlU7
tions forms a linear spaceover~. There are only n linearly independent initial is a solution 0((4-2) wi
states in(W, ~); hence the linear space is of dimension n. This fact will be
formally established in the following theorem.
\TRICES SüLUTIONS OF A DYNAMICAL EQUATION 135

~lementary to compute Theorem 4-1


e discussed further. The set of aH solutions of x(t) = A(t)x(t) forms an n-dimensional vector space
in Section 4-2. Solu­ over the field of real numbers.
x. <1>(t, o), which is the
Proof
Let '!tI and '!t2 be two arbitrary solutions of (4-2). Then al VI +a2V2 is also a
1
solution of (4-2) for any real al and IX 2 . We prove this by direct verification:
lrious methods for the d d d
- (IX IVl +a2V2) =IX 1 - VI +a 2 - V2 =a 1 A(t)Vl + IX 2A (t)V2
tion 4-3 the concept of dt dt dt
llent dynamical equa­
= A(t)(IX 1 VI + IX 2V2)
space. We show that
equivalent linear dy­ Hence the set of solutions forms a linear space over~. It is called the solution
o establish the theory space of (4-2). We next show that the solution space has dimension n. Let
lr d ynamical equations el> e 2, ... , en be any linearly independent vectors in (IR n, ~) and Vi be the solu­
lfY and sufficient con­ tions of (4-2) with the initial condition Vi(t O) = e¡, for i = 1, 2, ... ,n. If we
bY a linear dynamical show that Vi, for i = 1,2, ... , n, are linearly independent and that every solution
r rational matrix. has a of (4-2) can be written as a linear combination of V;, for i = 1, 2, ... , n, then the
assertion is proved. We prove by contradiction the fact that the Ijfs are linearly
109, 114, and 116. independent. Suppose that Vi, for i = 1,2, ... , n, are linearly dependent; then, by
definition, there exists a nonzero n x 1 real vector ot such that
(4-3)

Note that the O in the right-hand side of (4-3) is the zero vector of the solution
. linear time-varying space; therefore, it is more informative to write (4-3) as
[Vl(t) V2(t) Vn(t)]ot=O for a11 t in (-00,00)
uation) (4-1 a) In particular, we have

:quation) (4-1 b)

:l q x p matrices whose
which implies that ei, for i = 1, 2, ... , n, are linearly dependent. This contra­
over ( - 00, 00). Since
dicts the hypothesis; hence \ji" for i = 1, 2, ", n, are Ijnearly i.ndependenl ,,'!"'"r
) and any VI, there exists (-00,00).
t will be used frequently
Let V be any solution of (4-2), and let V(to) =e. Since el' e 2, ... , en are n
lre dynamical equation
linearly independent vectors in the n-dimensional vector space (~m ~), e can
of E; namely,
be wriUen as a unique linear combination of ei , for i = 1, 2, ... , n-for example,
(4-2) as n

e= ¿ IX¡e i
¡= 1

It is clear that

; a unique solution for

ly many possible initial

i = 1
:ions. This set of solu­

arly independent initial is a solution of (4-2) with the initial condition

on n. This fact will be n

¿ IXiV¡(tO) =e

i= 1

\
I
136 LINEAR OYNAMICAL EQUATIONS ANO IMPULSE-RESPONSE MATRICES

Hence, from the uniqueness of the solution, we conclude that Proof

\jiU = ¿"
Before we prove the th
1X¡\jIk)
i;:;; 1 *= A(t)x and if \jI(t o) =
that is, \jI(') == O. It is ob
This completes the proof that the solutions of (4-2) form an n-dimensional
Again, froro the unique
vector space. Q.E. D.
solution with \jI(t o) =0.
We shall now pro"
Definition 4-1 that det 'JI(to) = det [~
*
An n x n matrix function 'P is said to be a fundamental matrix of = A(t)x if
set of n constant col um¡
(IR", IR). It follows that
and only if the n columns of'P consist of n linearly independent solutions of
*= A(t)x. I

Example 1 which, together with th


Consider the dynamical equation

*=[~ ~Jx
is a solution of * = A(t)

It actually consists of two equations: Xl =0, X2 = tXI. Their solutions are


XI(t) = x ¡(to) and X2(t) = 0.5t 2XI(t O) - 0.5t6X¡(to) + X2(tO), wbich are obtained by
first solving for x¡(t) and then substituting x¡(t) into X2 = tx¡. Now two linearly This contradicts the aS5
independent solutions \jII = [O 1]' and \jI2 = [2 t 2]' can be easily obtained pendent. Hence, we ce
by setting to=O, x¡(to) =0, X2(tO) = 1 and x¡(to)=2, X2(tO) =0. Hence, the
matrix Definition 4-2
Let 'PO be any fundam<
<I>(t, to)~
is a fundamental matrix. I
is said to be the state trl
Each column of'P, by definition, satisfies the differential equation * = A(t)x;
The physical meanir
hence, it is evident that 'P satisfies the matrix equation
for all t, its inverse is .
..y = A(t)1Jf'
immediateiy the follow
with 'P(t o) = H, where H is sorne nonsingular real constant matrix. Conversely, matrix:
if a matrix M satisfies (4-4) and if M(t) is nonsingular for sorne t, then from the
proof of Theorem 4-1 we know tbat all the columns of M are linearly indepen­
dent. Hence, the matrix function M qualifies as a fundamental matrix. Thus
we conclude that a matrixfunction"qt is afundamental matrix of* = A(t)x ifand
only if'P satisfies (4-4) and 'P(t) is nOlisingular for some t. for any t, to, ti' and t 2 ir
An important property of a fundamental matrix 'PO is thatthe inverse of Note that CI>(t, to) is
'P(t) exists for each t in ( - 00, 00). This füllows from the following theorem. particular 'P chosen. Ll
*= A(t)x. Since the col
basis vectors, there exist
Theorem 4-2
.matrix P such that 'JI 2 =
Every fundamental matrix 'P is nonsingular for al! t in ( - 00, 00). of the itb column of 'P 2 1
\TRICES SOLUTlONS Of A DYNAMICAL EQUATlON 137

: that Proof
Before we prove the theorem, we need the following fact: Ir '!lO is a solution of
x = A(t)x and if '!ICto) = O for sorne to, then the solution '!lO is identicaHy zero;
that is, '!lO =:= O. It is obvious that '!lO =:= Ois a solution ofx = A(t)x with '!ICto) = O.
Jrm an n-dimensional Again, from the uniqueness ofthe solution, we conclude that '!lO =:= O is the only
Q.E.D. solution with '!I(to) =0.
We shaH now prove the theorem; we prove it by contradiction. Suppose
that det 'PCto) = det ['!IICtO) '!IzCto) .. , '!InCtO)] =0 for sorne too Then the
set ofn constant column vectors '!I1(t O), '!IlCtO)"'" '!InCtO) is linearly dependent in
! matrix of x = A(t)x if
(~n, ~). It follows that there exist real IX¡, for i = 1,2, ... , n, not aH zero, such that
iependent solutions of
n
I
¿ IXi'!l¡CtO) = O
", "
i= 1

which, together with the fact that


n
¿ IX¡'!Ik)
i= 1

is a solution ofx =A(t)x, implies that


n
. Their solutions are ¿ IX¡'!I¡(')=:=O
which are obtained by i= 1

tXI' Now two linearly This contradicts the assumption that '!Ik), for i = 1, 2, ... , n, are linearly inde­
:an be easily obtained pendent. Hence, we conclude that det 'P(t) 1=0 for aH t in (- 00, (0). Q.ED.
"zCto) =0. Hence, the
Definition 4-2
Let 'PO be any fundamental matrix of x = A(t)x. Then
for all t, t o in (-00, (0)
I
is said to be the state transition matrix of x = A(t)x. I
.tial equation x = ACt)X;
The physical meaning of<l>Ct, to) will be seen latero Since 'P(t) is nonsingular
for aH t, its inverse ís well defined for each t. From the definition we have
(4-Ll. )
ímmediateíy the following very important properties ol the state transition
ant matrix. Conversely, matrix:
r sorne t, then from the <l>(t, t) = 1 (4-5)
VI are linearly indepen­ <l> - l(t, to) = 'PCto)'P - 1 (t) = <l>Cto, t) (4-6)
lamental matrix. Thus <l>(t z, to) = <l>Ctz, t ¡}<l>(t 1> to) (4-7)
latrix ofx = A(t)x if and
for any t, to, ti' and t z in ( - 00,(0).
~.) is that the inverse of Note that <l>(t, to) is uniquely determined by A(t) and is índependent of the
following theorem. particular 'P chosen. Let 'P 1 and 'P z be two different fundamental matrices of
x = A(t)x. Sínce the columns of 'P l' as well as the columns of 'P 2' qualify as
basis vectors, there exists, as shown in Section 2-3, a nonsingtilar real constant
matrix P such that 'P z = 'P 1 P. In fact, the ith column of P is thereprese~tation
(- 00,00). of the ith column of 'P 1 with respect to the basis that consists of the columns of
... .... ... ..... ==..-'....::c.....-­

138 LINEAR DYNAMICAL EQUATIONS AND IMPULSE-RESPONSE MATRICES

'P 1('). By definition, we have Solutions of the dy¡


<I>(t, to) = q¡ z(t)(q¡ z(to)) - 1 = q¡ 1(t)PP -1 (q¡ 1 (t O)) - 1 We use cj)(l; lo, x o, u)
= q¡ 1(t)(q¡ l(t O))-1 X(t o) =x o and the apl

which shows the uniqueness of <I>(t, to). From Equation (4-4), it is evident that Theorem 4-3
<I>(t, to) is the unique solution of the matrix equation
The solution of the Sl

(4-8)
is given by
with the initial condition <I>(t o, to) = l.
Remarks are in order concerning the solutions of (4-4) and (4-8). Ir A(t)
is a continuous function of t, then <I>(t, to) and q¡(t) are continuously differenti­
able 1 in t. More generally, if A(t) is n times continuously differentiable in t,
then <I>(t, to) and q¡(t) are n + 1 times continuously differential in t; see References
24 and 77.
The computation of the solution of (4-8) in a closed form is generally very where <I>(t, '[) is the s
difficult, if not impossible, except for sorne special cases. Ir A(t) is a triangular unique solution of
matrix, then its solution can be reduced to solving a set of scalar differential
equations, and its closed-form solution can be readily obtained (Problem 4-1).
If A(t) has the following commutative property
Proof
Equation (4-11) is ob
first show that (4-10) ~
for all t and lo, then the unique solution of (4-8) is given by
d a
dl X(l) = al <1>(
<I>(l, to)=exp [ ( A('[)d'[] (4-9)

=A(l)q
(Problem 4-31). Ir A(t) is a diagonal matrix or a constant matrix, then it meets
the commutative property, and its transition matrix is given by (4-9). For other
special cases, see Problems 4-14 and 4-15. See also References S229 to S231.
From the concept of state transition matrix, the solution of Ji = A(l)x
=A(l) I
follows immediately. To be more informative, we use ,p(t; t o, 1<0' u) lo denote
the solution of x = A(l)x at time l due to the initial condition x(lo) =x o. The At l = lo, we have
fourth argument of cP denotes the fact that u == O. The solution of Ji = A(l)x
with x(lo) =x o is given by
X(l)~ cP(t; lo, x o, O) = <I>(l, lo)xo
In other words, (4-10)
which can be verified by direct substitution. The physical meaning ofthe state

. trarisition matrix <D(l, lo) isnow clear. It governs the motion of the state vector

in the time interval in which the input is identically zero. <I>(l, lo) is a linear
We consider again
transformation that maps the· state X o at lo ¡nto the state x at time l. (4-10) reduces to .

lA function is said to be continuous/y differentlab/e if its first derivative exists and is conUnuous. 2
ata J''o f(l, e) de = f(l, e)I,=,
TRICES SOLUTlONS OF A DYNAMICAL EQUATION 139

Solutions of the dynamical equation E

We use cP(t; to, x o, u) to denote the state resulted at time t due to the initial state
I(,on- ' x(to) =x o and the appUcation of the input u.

(4-4), it is evident that Theorem 4-3


The s01ution of the state equation

(4-8) x =A(t)x +B(t)u


is given by

(4) and (4-8). If A(t)


,mtinuous1y differenti­
x(t)~cP(t; to, xo, u)=<I>(t, to)xo + l'
to
<I>(t, r)B(r)u(r) dr (4-10)

Is1y differentiable in t,
\:ial in t; see References
= <I>(t, to{x o + L <I>(to, r)B(r)u(r) dr] (4-11 )

form is generaUy very


where <I>(t, r) is the state transition matnx of x= A(t)x; or, equiva1ent1y, the
unique solution of
[U A(t) is a triangular
t of scalar differential a <I>(r,r)=1 I
at <I>(t, r) = A(t)<I>(t, r)
I,tained (Problem 4-1).

Proof
I
1) Equation (4-11) is obtained from (4-10) by using <I>(t, r) = <I>(t, t o)<1>(to, r). We
first shüw that (4-10) satisfies the state equation by direct substitution 2 :
[bY
el
-x(t) a <I>(t, to)xo +­a
=- l' <I>(t, r)B(r)u(r) dr
(4-9) dt at at 'o

,a
=A(t)<1>(t, to)xo +<I>(t, t)B(t)u(t) +
1 -<I>(t, r)B(r)u(r) d!
'o at
t
J matrix, then it meets
ven by (4-9). For other
= A(t) [ <I>(t,to)x o + <I>(t, r)B(r)u(r) dr] +B(t)u(t)
ences S229 to S231.
solution of x = A(t)x
I(t; to, X o, O) to denote
ldition x(to) =x o. The At t = t o, we have
s01ution of x= A(t)x '0
x(to) = <I>(t o, to)xo -r
1
'o
<I>(to, r)B(r)u(r) dr = Ix o +0 =X o

In other words, (4-10) a1so meets the initia1 condition. Hence it is the s01ution.
11 meaning of the state Q.E.D.
ltion of the state vector
We consider again the so1ution given by Eqtiation (4-10). U u == O, then
I~o. <I>(t, to) is a linear (4-10) reduces to .
x at time t.
(4-12)

': exists and is continuous. 2 -o


oC 'o
'1' ¡(e, rl dr = f(t, rl \ + l' -a ¡(e, rl dr
,~, 10 0C
140 LINEAR DYNAMICAL EQUATIONS AND IMPULSE-RESPONSE MATRICES

Ifx o =0, Equation (4-10) reduces to By substituting(4­


t 4-3. The output y ca

tr <I>(t, r)B('r)u(r) dr
cjl(t; to, O, u) = (4-13) zero-input response.
Equatian (4-16) beco
For obvious reasons, cjl(t; to, x o, O) is called the zero-input response, and
cjl(t; to, O, u) is called the zero-state response of the state equation. It is clear that y(t)
cjl(t; to, x o, O) and cjl(t; to, O, u) are linear functions of X o and u, respective1y.
Using (4-12) and (4-13), the solution given by Equation (4-10) can be written as
cjl(t; to, xo, u) = cjl(t; to, x o, O) +cjl(t; to, O, u) (4-14)
The matrix function
This is a very important property; it says that the response 01a linear state equa­
tion can always be decomposed into the zero-state response and the zero-input
response. This is consistent with Equation (3-31).
is called the impulse-¡
Note that Equation (4-13) can be derived directly from the fact that it is a
the input-output re1a1
linear function of u. The procedure is exactly the same as the one in deriving
We see from (4-H
t
r G(t, r)u(r) dr
Jto
the solution of the d:
is the unique solution
in Section 3-2. The response cjl(t; to, O, u) is, by definition, the solution of
x= A(t)x +B(t)u with O as the initial state. If we cut the input u into small
pulses, say
Unfortunately, there
and except for very :
u= L u[t¡,t¡+Ó) easily faund. Theref
theoretical study af lil
then we have
of a dynamical equati
(4-15) <I>(t, r). The solution (
digital camputer.
where we have used the fact that if ~ is very sma11, the solution of x= A(t)x
+ B(t)u due to the input u[t¡,t¡H) with Oas the initial state is approximately equal
Time-invariant ca:
to B(t¡)u(t¡)~. The input u[t¡,t¡H) outside the time interval [ti, ti +~) is identica11y
(fixed) dynamical equ.
zero; hence, the response between t¡ +~ and t is governed by <I>(t, ti +~).
Summing up (4-15) for a11 i and taking the limit ~-+O, we immediately obtain
the equation
t

tr <I>(t, r)B(r)u(r) dr
cjl(t; to, O, u) = where A, B, e, and E
respectively. Since th
We give now the solution of the entire dynamical equation E. dynamical equatian E
applied here. We ha,
Corollary 4-3
The solution of the dynamical equation E in (4-1) is given by
t
y(t) = C(t)<D(t, to)xo +C(t) r $(t, r)B(r)u(r) dr + E(t)u(t)
t I and e At is nonsingula
In fact, e At is nonsingl

~'C(t)<D{t, to{x o + L <I>(to, r)B(r)U(r)dr] +E(t)u(t) (4-16)


tion matrix of x = Ax
rRICES SOLUTIONS OF A DYNAMICAL EQUATlON 1411

By substituting (4-10) and (4-11) into (4-1 b), we immediately obtain Corollary
4-3. The output y can also be decomposed into the zero-state response and the
(4-13) zero-input response. If the dynamical eguation is initially in the zero state,
Eguation (4-16) becomes
¡-input response, and t

lation. It is clear that y(t) = r [C(t)<II(t, r)B(r) + E(t)b(t - r)]u(r) dr


, and u, respectively. t
-10) can be written as
, u) (4-14 )
~ I'
'o
G(t, r)u(r) dr (4-17)

The matrix function


,ofa linear state equa­
lse and the zero-input G(t, r)~ C(t)<II(t, r)B(r) + E(t)b(t - r) (4-18 )

is called the impulse-response matrix of the dynamical eguation E. It governs


m the fact that it is a
the input-output relation of E if E is initially in the zero state.
lS the one in deriving
We see from (4-16) that if the state transition matrix <II(t, r) is known, then
the solution of the dynamical eguation can be computed. Recall that <II(t, r)
is the unique solution of
ition, the solution of
1e input u into small
ata <II(t, r) = A(t)<1l(t, r), <II( r, r) = 1
Unfortunately, there is in general no simple relation between <II(t, r) and A(t),
and except for very simple cases, state transition matrices <II(t, r) cannot be
easily found. Therefore, Eguations (4-10) and (4-16) are used mainly in, the
theoretical study of linear system theory. If we are required to find the solution
of a dynamical equation due to a given Xo and u, it is unnecessary to compute
(4-15 ) <II(t, r). The solution can be easily computed by using existing subroutines on a
digital computer.
solution of x = A(t)x
; approximately egual
Time-invariant case. In this subsection, we study the linear time-invariant
-t¡,t i +8)isidentically
(fixed) dynamical eguation
erned by <1>( t, ti + 8).
'e immediately obtain FE: x =Ax +Bu (4-19a)
y =Cx +lElUl (4-19b)

where A, B, e, and E are n x n, n x p, q x n, and q x p real constant matrices,


respectively. Since the equation FE is a special case of the linear time-varying
ation E. dynamical eguation E, all the results derived in the preceding subsection can be
applied here. We have shown in (2-82) that

1 by

+ E(t)u(t) and eAI is nónsingular at t = O; hence eA' is a fundamental n1atr~x of x = Ax.


In fact, eA' is non'singular for all t and (eA') - 1 = e- A'; therefore, the state transi­
tion matri", .af X =Áx. is, by the use of (2-81) and (2-79),
+ E(t)u(t) (4-16)

• <II(t, to) = eA'(eA'O) -1 = eA(t-' 'o) = <II(t - to)


142 LINEAR DYNAMICAL EQUATlONS AND IMPULSE-RESPONSE MATRICES

It follows from (4-10) that the solution of (4-19a) is We may apply them
t
cP(t;to,xo,u)=eA(t-to)xo+ r eA(t-<JBu(,)d, (4-20)
1. Using Definition
t polynomial g(A) o
then e At = g(A).
If t o = O, as is usually assumed in the time-invariant equation, then we have the
fol1owing theorem. 2. Using the Jordan (
where Á is of the J,
(2-70).
Theorem 4-4
3. Using the infinite
The solution ofthe linear time-invariant dynamical equation FE given in (4-19) give a closed-form
is tation.
We introduce one me
(4-21 )
we have
and
Hence, to compute l
(4-22) inverse Laplace tran
inverse of a matrix i¡
I
angular 4 or of order l
the inverse of a trian~
The impulse response matrix of FE is
Note that(sl-A)
G(t, ,) = G(t -e) =CeA(t-<JB + El5(t -e) many methods to con
1. Taking the inverse
or, as more commonly written,
2. Using Definition 2
G(t) = CeAtB + El5(t) (4-23) 3. Using (si - A)-l =
4. Using Definition 2
The solution of a linear time-invariant dynamical equation can also be 5. Taking the Laplacl
computed in the frequency domain. Taking the Laplace transform of (4-21)
In addition, there is a
and (4-22), and using g' [e Al ] = (si - At 1 [see Eq uation (2-86)], we obtain
x(s) = (si - A)-lX(O) +(sl - A)-l Bú(s) (4-24)
Example 2
We use methods 1 anl
and y(s) = C(sl - A)-lX(O) +C(sl - A)-l Bn(s) + En(s) (4-25)

where the circumflex denotes the Laplace transform 01 a variable. 1 nese


equations have been derived in Section 3-3 directly from the dynamical equa­
tion. As is defined there, the rational-function matrix

(;(s) = C(sl - A)-l B + E (4-26)


1. (sl-A)-l =[ -1
s

S +
is cal1ed the transfer-function matrix of the dynamical equation FE. It is the
Laplace transform ofthe impulse-response matrix given in (4-23). The transfer­
function matrix governs the zero-state response of the equation FE.
= (s ;J
We give now sorne remarks concerriing the computation of e Al . 3 We [
(s +J
introduced in Section 2-7 three methods of computing functions of a matrix.

3See also Reference S159. 4 A squarematrix is said lo be


-----------------,------~.-.-~~~_ ..

[RICES SOLUTIONS OF A DYNAMICAL EQUATION 143

We may apply them to compute eA!:

1. Using Definition 2-16: First, compute the eigenvalues of A; next, find a


r) dr (4-20)
polynomial g(A) of degree n - 1 that is equal to eA! on the spectrum of A;
then eA! = g(A).
:ion, then we have the
2. Using the Jordan canonical form of A: Let A = QÁQ-l; then e At = QitQ-l,
where Á is of the Jordan formo i t can be obtained immediately by the use of
(2-70).
3. Using the infinite series eA! = L:'=o tkAk/k!: This series will not, generally,
ion FE given in (4-19) give a closed-form solution and is mainly used on digital computer compu­
tation.

(4-21 )
We introduce one more method of computing eA'. Since se [e
At
] = (sI - A)-1 ,
r) dr we have
eAt = -1(sI - At 1 se (4-27)

Hence, to compute eA!, we first invert the matrix (sI - A) and then take the
u(t) (4-22) inverse Laplace transform of each element of (sI - A)-I. Computing the
inverse of a matrix is generally not an easy jobo However, if a matrix is tri­
I
angular 4 or of order less than 4, its inverse can be easily computed. Note that
the inverse of a triangular matrix is again a triangular matrix.
Note that (sl- A)-1 is a function of the matrix A; therefore, again we have
r) many methods to compute it:
1. Taking the inverse of (sI - A).
2. Using Definition 2-16.
(4-23) 3. Using (sI - A)-1 = Q(sI - Á)-IQ -1 and (2-74).
4. Using Definition 2-17.
~quation can also be 5. Taking the Laplace transform of eA!.
;e transform of (4-21)
In addition, there is an iterative scheme to compute (sI - A) - 1 (Problem 2-39).
(2-86)], we obtain
Example 2
(4-24 )
We use methods 1 and 2 to compute (sI - A)-I, where
+Eu(s) (4-25)
. 10 -1l
of a variabie. These Fl=Ll -2J
the dynamical equa­

(4-26)
1. (sI-A)-1 =[ - 1 s
s
+2
1 J-l = + S2
1
2s +1
[s+2
1
-sI]
luation FE. lt is the

1 (4-23).The transfer­

uation FE.

=[(; ~~' (, ::)'1


utation of e At . 3 W'e.
(s + 1)2 (s + 1)2J
:unctions of a matrix.

4 A square matrix is said to be triangular if al! the elements below or aboye the main diagonal are zero.
144 LINEAR DYNAMICAL EQUATIONS AND IMPULSE-RESPONSE MATRICES

2. The eigenvalues of A are -1, -1. Let g(A) =0:0 +0: 1,1. Iff(A)~(S-A)-l
= g(A) on the spectrum of A, then
8
f( -1) =g( -1): (s +1)-1 =0:0 -0: 1
j'(-1)=g'(-l): (s +1)-2=0: 1
Hence g(A) = [(s + 1)- 1 +(S + 1)- 2J +(S + 1)- 2A
and (si - A)-I = g(A) = [(s + 1)- 1 +(s + 1)-2JI +(s + 1)- 2A

J(: ~~' (, ::)~ o


l(S + 1)2 (s + 1)2J
Example 3
Consider the state equation Figure 4-1 TrajectoI

know that every elen


2, ... ,m. Hence ev(
The solution is given by We caH tke Ai! a mode

x(t) = é!x(O) + I eA(t-tlBu(r) dr (4-28)


The responses of
mainly by its modes,
a negative real part,
The matrix function eA! can be obtained by taking the inverse Laplace transform approaches zero eitt
of (si - A)- l. which is computed in Example 2. Hence its imaginary part is
mode will approach
S +2
zero real part and ha
(s + 1)2
its index ni is 2 or h

(s~w
tñ; -1.
[
If aH eigenval ues
can be written as, by
and

)=L1(1 +t)e-
t

x(t te -/
where q¡ and Pi are, I
The sol utions with u == O are plotted in Figure 4-1 for the initial states with A¡. In the time dI
x(O) = [8 8]' and x(O) = [8 5]'. Note that the velocity at each point of the
trajectories in Figure 4-1 is equal to Ax. •

If the matrix A has m distinct eigenvalues A¡ with index ni, for i = 1, 2, ... , m If Xo ís chosen so tha
(see Definition 2"15),5 we claim that every element of eA! is a linear combination
ofthe factors tkeA¡t, for k = 0,1, ... ,ni - 1; i = 1,2, ... , m. Let Á be a Jordan-form
representation of A and A = QÁQ - 1. Then eA! = QeA!Q - l. From (2-69) we F or this il1itial state, '
direction of the eigen
formula: similar to (4
5If aH eigenvalues are distinct, then the index of every eigenvalue is 1. If an eigenvalue has mullí-o eigenvectors of A. 1
plicily k, then ilsindex could be 1,2, ... , k -1, or k. See Seclíon 2-6 and Definition 2-15. discussed.
'RICES SOLUTIONS Of A DYNAMICAL EQUATlON 145

A. If f(A)6 (5 - A)-I

-----!L--+---+--+--+-+--+--1--1--- xl
o 4 8

Figure 4-1 Trajectories.

know that every element of e Al is ofthe form tkeA¡t, for k =0,1, ... , ñ¡ - 1; i = 1,
2, ... ,m. Hence every element of e At is a linear combination of these factors.
We call tkeA¡t a mode of the dynamical equation FE in (4-19).
The responses of a linear time-invariant dynamical equation are dictated
(4-28) mainly by its modes, or equivalently, the eigenvalues of A. If an eigenvalue has
a negative real part, its mode will approach zero exponentially as t --> 00; it
rse Laplace transform approaches zero ·either monotonically or oscillatorily depending on whether
its imaginary part is zero or not. If an eigenvalue has a positive real part, its
mode will approach infinity exponentially as t --> oo. If an eigenvalue has a
zero real part and has index 1,5 its mode is a constant or a pure sinusoidal; if
-te-
(l-t)e- t
t
J its index ñ¡ is 2 or higher, then its mode will approach infinity at the rate of
tñ¡-I.
If all eigenvalues of A are distinct, the response of x = Ax due to x(O) =x o
can be written as, by using Problem 2-44,

'e
-(t-r) l x(s) = (sI - A)
A -1"
o = L,
X --1
1 Qi1ll1¡X
o (4-29)
')Je -(t - rl J U(T) dT i s -- /"'i
where q¡ and p¡ are, respectively, a right and a left eigenvector of A associated
for the initial states with Aj • In the time domain, (4-29) becomes
I at each point of the
• (4-30)

{ ñ¡, for i = 1,2, ... , m lf Xo is chosen so that p¡XO =0 for all iexcept i = j, then (4-30) reduces to
; a linear combination
,et Á. be a lordan-form x(t) = (p }'o)e Ajt qj
r l. From (2-69) we For this initial state, only the mode e Ajt is excited and x(t) will travel along the
direction of the eigenvector qj. lf A has eigenvalues with indices 2 or higher,a
formula similar to (4-29) can be derived by using (2-74) and the genera:Jized
Ir an eigenvalue has multi­ eigenvectors of A. The situationis much more complicated and will not be
and Definition 2-15. discussed.
146 LINEAR DYNAMICAL EQUATIONS AND IMPULSE-RESPONSE MATRICES

4-3 Equivalent Dynamical Equations matrix E is the dire


Since it has nothing 1
We introduce in this section the concept of equivalent dynamical equations. lence transformation
This concept in the time-invariant case is identical to the one of change of basis We explore noW
introduced in Figure 2-5. Hence, we study first the time-invariant case and then Recall that the state
the time-varying case. unique relation betv
initiaUy relaxed. n
Time-invariant case. Consider the linear time-invariant (fixed) dynamica\ analysis often lead te
equation
Example 1
FE: X =Ax +Bu (4-31a)
y=Cx +Eu (4-31b) Consider the networl
inductor X¡ and the v
where A, B, C, and E are, respectively, n x n, n x p, q x n, and q x p real constant
ables, then the dynan
matrices; u is the p x 1 input vector, y is the q xl output vector, and x is the
n x 1 state vector. The state spaCe L of the dynamical equation is an n-dimen­
sional real vector space and the matrix A maps L into itself. We have agreed in
Section 3-3 to choose the orthonormal vectors {n¡, n z, ... , n,.} as the basis
vectors ofthe state space L, where ni is an n x 1 vector with I at its ith component
and zeros elsewhere. We study now the effect of changing the basis of the state
space. The dynamical equations that result from changing the basis of the state lf, instead, the loo
space are called equívalent dynamical equations. In order to allow a broader the dynamical eq uati
class of equivalent dynamical equations, we shall in this section extend the field
of real numbers to the field of complex numbers and consider the state space as
an n-dimensional complex vector space. This generalization is needed in
order for the dynamical equation FEto have an equivalent lordan-form dy­
namical equation.
The dynamical eq
Definition 4-3 describe the same sys
Let P be an n x n nonsingular matrix with coefficients in the field of complex formation between ti
numbers e, and let x = Px. Then the dynamical equation clear that X¡ =x¡. S
we have xz=(x¡-x
FE; x=Áx +Bu (4-32a)
y=Cx+Eu (4-32b)

where A =PAP-¡ B =PB e =CP-¡ (4-33)

is said to be equivalent to the dynamical equation FE in (4-31), and P is said to or


be an equivalence transformation. •

The dynamical equation F E in (4-32) is obtained from (4-31) by the substitu­


tion of x = Px. In this substitution we have changed the basis vectorsof the
state space from the orthonormal vectors to the columns ofP-¡ (see Figure 2-5).
Observe that the matrices A and Á are similar; they are different representations
of the same operator. If we let Q = P - ¡ = [q¡ qz ... qn], then the i.th
column of A is therepresentation of Aqi with respect to the basis {q¡, qz,·· ., q,.}.
From the equation B == PB or B = P - ¡ B = [q¡ qz ... qn]B, we see that
the ith column of B is the representation of the ¡th column of B with respect to
e
the basis {q¡, qz, ... , qn}' The matrix is to be computed from CP ~ ¡. The Figure 4-2 A networl
"RiCES EQUlVALENT DYNAMICAL EQUATIONS 147

matrix E is the direct transmission part between the input and the output.
Since it has nothing to do with the state space, it is not affected by any equiva­
iynamical equations. lence transformation.
,ne of change ofbasis We explore now the physicaL meaning of equivalent dynamical equations.
variant case and then RecaIl that the state of a system is an auxiliary quantity illtroduced to give a
unique relation between the input and the output when the system is not
initially relaxed. The choice of the state is not unique; different methods of
ant (fixed) dynamical analysis often lead to different choices of the state.

Example 1
(4-31 a)
(4-31b) Consider the network shown in Figure 4-2. If the current passing through the
inductor Xl and the voltage across the capacitor Xl are chosen as the state vari­
Id q x p real constant
ables, then the dynamical equation description of the network is
t vector, and x is the
lation is an n-dimen­
(4-34a)
r. We have agreed in
... , "ti} as the basis
Lat its ith component (4-34b)
; the basis of the state
~ the basis of the state If, instead, the loop currents Xl and Xl are chosen as the state variables, then
r to aIlow a broader the dynamical equation is
ction extend the field
der the state space as (4-35a)
ization is needed in
ent Jordan-form dy­ (435b)

The dynamical equations in (4-34) and (4-35) have the same dimension and
describe the same system. Hence they are equivalent. The equivalence trans­
the field of complex formation between these two equations can be found from Figure 4-2. 1t is
I clear that Xl =Xl' Since Xl is equal to the voltage across the l-ohm resistor,
we have Xl =(Xl -Xl)' Thus
(4-32a)
(4-32b)

(4-33)
(4-36)
f-31), and P is said to or

4-31) by the substitu­
~ basis vectors of the
'p-I (see Figure 2-5). Xl

erent representations
.. qtl], then the ¡th
basis {ql, qz, ... , qtl}.
qtl]B, we see that
1 of B with respect to
ed from Cp-l. The Figure 4-2 A network with two different choices of state variables.
148 LINEAR DYNAMICAL EQUATIONS AND IMPULSE-RESPONSE MATRICES

It is easy to verify that the dynamical equations in (4-34) and (4-35) are indeed Example 2
related by the equivalence transformation (4-36). 11 Consider the two net
capacitor is zero, the
Definition 4-4 impulse responses are
Two linear dynamical equations are said to be zero-state equivalent if and only if are zero-state equivale
they have the same impulse-response matrix or the same transfer-function tions are zero-state eq
matrix. Two linear dynamical equations are said to be zero-input equivalent any initial voltage in 1
if and only if for any initial state in one equation, there exists a state in the other u = O are all identicall~
equation, and vice versa, such that the outputs of the two equations due to A necessary condit
zero input are identical. 11 they have the same d
network in Figure 4-31
Note that this definition is applicable to linear time-invariant as well as ofthe network in FigUl
linear time-varying dynamical equations.

Theorem 4-5 Example 3

Two equivalent linear time-invariant dynamical equations are zero-state equi­ The two networks in ]
valent and zero-input equivalent. descriptions are zero-¡
if there is a nonzero ir
Proof of Figure 4-4(b) is no]
is identically zero.
The impulse-response matrix of FE is
G(t) = CeA1B + Eb(t) lfthe matrix A in a
tion is said to be a lord
The impulse-response matrix of FE is
G(t) = CeA1Ü + Eb(t)
If the equations FE and FE are equivalent, we have Á = PAP- 1 ,
C = CP-l, and E =E. Consequently we have

and
eA' = PeA'P- 1
CeA'B + Eb(t) = Cp- 1 PeA'p- 1 PB + Eb(t) = CeAtB + Eb(t)
B = PB,

u t 1n
1 y u(

Hence, two equivalent dynamical equations are zero-state equivalent. -_Y-,


The zero-input response of FE is (a)

y(t) = CeA(t - 'Olx(t o) Figure 4-3 Two netw<


zero-input equivalent wit
The zero-input response of FE is
y(t) = CeA(t -IO)X(tO) = Ce A(' -Iolp - 1x(to)
Hence, for any x(to), if we choose x(t o) = Px(t o), then FE and FE have the same 0.5 n
r y

zero-input response. Q.E. D. u t ~ u(


We note that although equivalence implies zero-state equival~nce and zero­ 0.5 n
input equivalence, the conve¡;seis. not true. That is, two .linear dynamical
equations can be zero-state equivalimt and ze~o-input equivdlent witho~t being (a)
equivalent. Furthermore, two linear equations can be zero-st~Úe equivalent
Figure 4-4 Two netwo
without being zero-input equivalent.
·TRICES EQUlVALENT DYNAMICAL EQUATIONS 149

and (4-35) are indeed Example 2


I
Consider the two networks shown in Figure 4-3. If the initial state in the
capacitor is zero, the two networks have the same input-output pairs; their
impulse responses are al! equal to b(t), the Dirac b-function. Therefore, they
'quivalent if and only if are zero-state equivalent, or more precisely, their dynamical-equation descrip­
ame transfer-function tions are zero-state equivalent. Because of the symmetry of the network, for
: zera-input equivalent any initial voltage in the capacitor, the outputs of these two networks due to
ists a state in the other u =0 are al! identical!y zero. Hence, they are zero-input equivalent.
two equations due to A necessary condition for two dynamical equations to be equivalent is that
I they have the same dimensiono The dynamical-equation description of the
network in Figure 4-3(a) has dimension O; the dynamical-equation description
e-invariant as wel! as ofthe network in Figure 4-3(b) has dimension 1. Hence, they are not equivalent.

Example 3

ns are zero-state equi- The two networks in Figure 4-4 or, more precisely, their dynamical-equation
descriptions are zero-state equivalent but not zero-input equivalent. Indeed,
if there is a nonzero initial condition in the capacitor, the zero-input response
of Figure 4-4(b) is nonzero, whereas the zero-input response of Figure 4-4(a)
is identically zero. •

Ifthe matrix A in a dynamical equation is in the Jordan form, then the equa­
tion is said to be a Jordan-formdynamicaJ equation. We have shown in Section

+ Eb(t)
u
t In
1ty u
In
+
x

e
In)'
y

~
AIB In In
e equivalent. -'-- ---"'~-
-
---
(a) (b)

Figure 4-3 Two networks whose dynamical equations are zero-state equivalent and
zero-input equivalent without being equivalent.

o)

mdFE have the same


0.5 n
Q.E.D.
u t I F
equivalence and zero­ 0.5 n In In
twa linear dynamical
úvalent withaut being
(a) (b)
zero-state equivalent
Figure 4-4 Two networks whose dynamical equations are zero-state equivalent.
150 LINEAR OYNAMICAL EQUATIONS ANO IMPULSE-RESPONSE MATRICES

2-6 that every operator that maps (Cm C) into itself has a lordan-form matrix This can be written as
representation. Hence every linear, time-invariant dynamical equation has an
E +CBS-l +CABs- 2
equivalent Jordan1orm-dynamical equation.

Example 4 This equality holds fOl


Consider the dynamical equation
FE: This establishes the th,
(4-37a)
*Time-varying caSE
varying dynamical eqt
AH the discussion and
(4-37b) well. The only conCf
invariant case are fixe<
time-varying case may
lt is easy to verify that if we choose Consider the linear

5]-1 = [1
3 O
O
1
-5]
-3
1 O O 1 where A, B, C, and E al
continuous functions o
then the new A matrix will be in the lordan formo The method for finding the
matrix P was discussed in Section 2-6. If we substitute x = Px into (4-37), we Definition 4-5 6
obtain immediately the following equivalent lordan-form dynamical equation:
Let P(·) be an n x n rr

[;:]= [~--U-.~][~:] [~ 1~][~:J


and P(t) are nonsingul
dynamical equation
+
X3 O 0, 2 X3 -1-2

[;}[: ~ 4:j I where

We give the following important theorem to concJude this subsection..


is said to be equivalent
Theorem 4-6 be an equivalence tran~
Two linear time-invariant dynamical equations {A,B, L, E} and {A, B, C,E}, not
The dynamical eql
necessarily of the same dimension, are zero-state equivalent or have the same
by the substitution of ji¡
transfer-function matrix if and only if E = E and
matrix of E. Then we
i =0,1,2, ....

Proof is a fundamental matri:

Theyare zero-state equivalent if and .only if


E + C(sI - A)-1 B = E + C(sI---'A)-IB 6This definition reduces to O
~ATRICES EQUlVALENT DYNAMICAL EQUATIONS 151

, a J ordan-form matrix This can be written as, by using (2-85),


amical equation has an E +CBs- 1 +CABs- 2 +CA 2 Bs- 3 + ...
E+ CBs - 1 + CABs - 2 + CA 2Bs - 3 +
= ...
This equality holds for every s if and only if E = E and
i =0,1,2, ...
This establishes the theorem. Q.E.D.
(4-37a)
*Time-varying case. In this subsection we study equivalent linear time­
varying dynamical equations. This is an extension of the time-invariant case.
AH the discussion and interpretation in the preceding subsection applies here as
(4-37b) well. The only conceptual difference is that the basis vectors in the time­
invariant case are fixed (independent of time), whereas the basis vectors in the
time-varying case may change with time.
Consider the linear time-varying dynamical equation
x= A(t)x + B(t)u
-5]
-3

E:
y = C(t)x + E(t)u
(4-38a)
(4-38b)

1
where A, B, C, and E are n x n, n x p, q x n, and q x p matrices whose entries are
continuous functions of t.
: method for finding the
e x = Px into (4-37), we Definition 4.5 6
·m dynamical equation:
Let PO be an n x n matrix defined over (- ro, ro). It is assumed that P(t)
and P(t) are nonsingular and continuous for aH t. Let x = P(t)x. Then the
dynamical equation
E: x=Á(t)x+"8(t)u (4-39a)
y = qt)x + lt(t)u (4-39b)

I where A(t) = (P(t)A(t) + P(t))P -l(t) (4-40a)


B(t) = P(t)B(t) (4-40b)
~(t) = C(t)P-l(t) (4.-40c)
ie this subsection. E(t) = E(t} (4-4001)

is said to be equivalent to the dynamical equation E in (4-38), and PO is said to


be an equivalence transformation. •

,E} and {A, B, C, E}, not The dynamical equation E in Equation (4-39) is obtained from (4-38)
'alent or have the same
x
by the substitution of x = P(t)x and = P(t)x + P(t)x. Let'JI be a fundamental
matrix of E. Then we claim that
q¡(t)~ P(t)'JI(t) (4-41 )

is a fundamental matiix of E. The matrix 'JI is, i:Jy assumption, a fundamental

6This definition reduces to Definition 4-3 ir P is independent of time.


152 LINEAR DYNAMICAL EQUATlONS AND IMPULSE-RESPONSE MATRICES

matrix of E; hence 'iJ(t) = A(t)'P(t), and 'P(t) is nonsingular for all t. Conse­ Definition 4-6
quently, the matrix P(t)'P(t) is nonsingular for aH t (see Theorem 2-7). Now
x
we show that P(t)'P(t) satisfies the matrix equation =A(t)x. Indeed, A matrix P(·) is calle,
tinuous and bounded
d • •
di (P(t)'P(t)) =P(t)'P(t) + P(t)'P(t)
= CP(t) + P(t)A(t))P -l(t)P(t)'P(t)
= A(t)(P(t)'P(t))
Because of (4-43) a
Hence, P(t)'P(t) is a fundamental matrix of x= Ax. of P- 1(t) = - P-l(t)PI
Consequently, we car
Theorem 4-7 7 P- 1 (.). Clearly a nor
Let Ao be an arbitrary constant matrix. Then the dynamical equation in (4-38) In Definition 4-5, if P(
is equivalent to the one in (4-39) with A(t) = A o. are said to be equivale
preserves, as will be (
Proof namical equation, bu
required to be a Lya
Let 'P(t) be an fundamental matrix of x = A(t)x. That is, 'P(t) is nonsingular for
general. In other wo
~1I t and satisfies 'iJ(t) = A(t)'P(t). The differen~iation of '11 - 1 (t)'PJt) = 1 yields equivalent in the sense
'P- 1(t)'P(t) + '11 - 1 (t)'P(t) = O, which implies '11 -l(t) = - 'P- 1(t)'P(t)'P- 1(t) =
This is possible for the
- 'P- 1(t)A(t). We define, in view of (4-41),
be discussed in the foL
P(t) = eAo1'P - 1(t)
Linear time-varying e
Clearly P(t) is nonsingular and continuously differentiable for aH t and qualifies
as an equivalence transformation. We compute Consider the linear tir
A(t) = (P(t)A(t) + P(t))P- 1(t)
= (e A01'P- 1(t)A(t) + AoeAo1'P - l(t) +eAol'iJ-1(t))'P(t)e - Aot
for aH t and for sorne
=A o A(-) is a periodic funct
This establishes the theorem. Q.E.D. matrix of x = A(t)x. 1
Indeed we have
In this theorem, '11(.) and consequently PO are generaHy not known, there­
fore nothing is really gained in this transformation. If Ao is chosen as zero, 'P(t .
then P(t) = 'P- 1 (t) and (4-40) becomes The matrix function '1
A(i) = ir» B(t) = q;-l (t)B(i) CU) = qt)\f(t) E(t) = lE(i) (4-42) Colum_l1s of 'f'(f) Etild e
solution space; hence, 1
Its block diagram is plotted in Figure 4-5. Unlike the one in Figure 3-11, there 2-3) such that
is no feedback in Figure 4-5.

7 This theorem was pointed out to the author by Professor T. S. Kuo in 1972. I::or the nonsingular J
AT
e = Q (Problem 2-37

ECt) ~=========;l
Define
+ y

+ We show that PO is a 1
Figure 4-5 Matrix block diagraril 6f the dynamical equation E in (4-42). P(t +T) =e
rRICES EQUlVALENT DYNAMICAL EQUATIONS 153

lar for a11 t. Conse­ Definition 4-6


Theorem 2-7). Now
A matrix PO is called a Lyapunov transformation if (1) PCt) and P(t) are con­
:t)x. Indeed,
tinuous and bounded on [to, 00) and (2) there exists a constant m such that
O<m<\detP(t)1 forallt2t o (4-43)

'P(t) •
.Because of (4-43).and the boundedness of P(t), P- 1(t) is b.ounded. Because
of P - l(t) = - P -1(t)P(t)P- 1(t) and the boundedness of P(t), P -1(t) is bounded.
Consequently, we can show that if PO is a Lyapunov transformation, so is
P -lO. Clear1y a nonsingular constant matrix is a Lyapunov transformation.
ical equation in (4-38) In Definition 4-5, if PO is a Lyapunov transformation, the dynamical equations
are said to be equivalent in the sense of Lyapunov. A Lyapunov transformation
preserves, as will be discussed in Chapter 8, the stability properties of a dy­
namical equation, but an equivalence transformation does not. lf PO is
required to be a Lyapunov transformation, Theorem 4-7 does not hold in
P(t) is nonsingular for general. In other words, not every time-varying dynamical equation can be
. 'P- 1(t)'I'(t) =1 yields equivalent in the sense of Lyapunov to a dynamical equation with a constant A.
- 'P- 1(t)'f(t)'I'-1(t) = This is possible for the class of time-varying equations with periodic ACt) as will
be discussed in the following.

Linear time-varying dynamical equations with periodic A(')


e for aH t and qualifies
Consider the linear time-varying dynamical equation in (4-38). We assume
A(t + T) = A(t)
for all t and for sorne positive constant T. This means that every element of
AO is a periodic function with the same period T. Let 'PCt) be a fundamental
Q.E.D. matrix of x = A(t)x. Then 'P(t + T) is also a fundamental matrix of x= A(t)x.
Indeed we have
llly not known, there­
'fCt +T)=A(t +T)'PCt +T)=A(t)'P(t +T)
Aa is chosen as zero,
The matrix function 'P(t) is nonsingular for al! t; consequent\y, so is 'P(t + T).
Co\umns of ~(t) and columns of qJ(t + T) forrn two sets of basis vectors in the
E(O = JE(t) (4-42)
solution space; hence, there exists a nonsingu\ar constant matrix Q (see Section
le in Figure 3-11, there 2-3) such that
'P(t + T) = 'PCt)Q (4-44)

~or the nonsingular matrix Q there exists a constant matrix Á such that
n 1972.
eAT = Q (Prob\em 2-37). Hence, (4-44) can be written as
'P(t + T) = 'P(t)e AT (4-45)

Define
P(t)~eAl'P-l(t) (4-46)

We show that PO is a periodic function with period T:


: in (4-42). PCt +T) = eA{I+ T)'P -1(t +T) = eA1i T e- AT'P-1(t) = P(t)
_........
"-~---~-
,'- ., .. _~
---

154 LINEAR DYNAMICAL EQUATlONS AND IMPULSE-RESPONSE MATRICES

Theorem 4-8 where <Il(t, T) is the sta1


Assume that the matrix A in the dynamical equation E in (4-38) is periodic with (4-49), we immediatel:
period T. Let P be defined as in (4-46). Then the dynamical equation E in
(4-38) and the dynamical equation
G(t, T) =
E: x(t) = Áx(t) +P(t)B(t)u(t)
y"(t) = C(t)P -l(t)X(t) + E(t)u(t)
That G(t, T) = O for t <
where Á is a constant matrix, are equivalent in the sense of Lyapunov. • itly embedded in writi
Ir the state-variab
The matrix PO in (4-46) is periodic and nonsingular; hence it is bounded.
description of the sys
Its derivative is clearly continuous and bounded. Hence, PO is a Lyapunov
problem-to find the
transformation. The rest of the theorem follows directly fram Theorem 4-7.
tion of a system-is n
The homogeneous part of this theorem is the so-called theory of Floquet.
two prablems: (1) Is j
It states that if x = A(t)x and)f A(t + T) = A(t) for all t, then its fundamental
fram the impulse-resp
matrix is ofthe form P- 1 (t)e At , where P- 1 (t) is a periodic function. Further­
state-variable descript
more, x =A(t)x is equivalent in the sense of Lyapunov to =Áx. x the first problem in th
studied in Chapter 6.
4-4 Impulse-Response Matrices and Dynamical Equations Consider a system
a linear finite-dimensil
Time-varying case. In this section we shall study the relation between the response matrix, then
impulse-response matrix and the dynamical equation. Let the input-output equation E, or more
description of a system with p input terminals and q output terminals be G(t, T). The terminol!
dynamical equation, '
y(t) = rtG(t, T)U(T) dT (4-47) generate G(t, T). Noté
Jo impulse-response matl
where y is the q x 1 output vector, u is the p x 1 input v(¡ctor, and G is the not have any physical
q x p impulse-response matrix of the system. We have implicitly assumed in tion gives only the Sal
(4-47) that the system is initially relaxed at too The ijth (ith row, jth column) equation is not in the
element of G(', T) is the response at the ith output terminal due to a b-function system.
input applied at time T at the jth input terminal. Suppose now that the internal Ir the realization (
structure of the same system is accessible, and that analysis of this system dimensional linear dyr
leads to a dynamical equation of the form not every G(t, T) is rea
form (4-48) that will gel
E: x(t) = A(t)x(t) + J3(t)..(t) (4-,f¡J.a~)
¡he impUlse response 1
y(t) = C(t)x(t) + E(t)u(t) (4-48b) ficient condition for G
where x is the n x 1 state vector of the system, and A, B, e, and E are n x n,
n x p, q x n, and q x p matrices whose entries are continuous functions of t Theorem 4-9
defined over (- 00, (0). Since Equations (4-47) and (4-48) are two different
descriptions of the same system, they should give the same input-output pairs A q x p impulse-respo
ifthe system is initially relaxed. The solution ofthe dynamical equation E with linear dynamical equa
x(t o) = O is given by composed into
t
y(t) = C(t) r <I>(t, T)B(T)U(T) dr + E(t)u(t) G(t, T)
Jo
t where E is a q x p mat
= r [C(t)<I>(t, T)B(T) +E(t)b(t-T)]U(T) dT (4-49) tinuous matrices of t.
Jo
TRICES IMPULSE-RESPONSE MATRICES AND DYNAMICAL EQUATlONS 155

where <I>(t, T) is the state transition matrix ofi = A(t)x. By comparing (4-47) and
(4-49), we immediately obtain
(4-38) is periodic with
amical equation E in
G(t, T) = {~(t)<I>(t, T)B(T) + E(t)(j(t - T) for t ¿ T

for t < T (4-50)

That G(t, T) = O for t < T fol\ows from the causality assumption which is implic­
of Lyapunov. • itly embedded in writing (4-47); that is, the integration is stopped at t.
If the state-variable description of a system is available, the input-output
; hence it is bounded. description of the system can be easily obtained from (4-50). The converse
e, PO is a Lyapunov problem-to find the state-variable description from the input-output descrip­
from Theorem 4-7. tion of a system-is much more complicated, however. It actual\y consists of
led theory of Floquet. two problems: (1) Is it possible at al\ to obtain the state-variable description
then its fundamental from the impulse-response matrix of a system? (2) If yes, how do we obtain the
ic function. Further­ state-variable description from the impulse-response matrix? We shal\ study
x
to =Ái. the first problem in the remainder of this section. The second problem will be
studied in Chapter 6.
¡mical Equations Consider a system with the impulse-response matrix G(t, T). If there exists
a linear finite-dimensional dynamical equation E that has G(t, T) as its impulse­
e relation between the response matrix, then G(t, T) is said to be realizable. We cal\ the dynamical
Let the input-output equation E, or more specifical\y, the matrices {A, B, e, E}, a realization of
G(t, T). The terminology "realization" is justified by the fact that by using the
out terminals be
dynamical equation, we can build an operational amplifier circuit that will
(4-47). generate G(t, T). Note that the state of a dynamical7equation realization of the
impulse-response matrix of a system is purely an auxiliary variable and it may
: vector, and G is the not have any physical meaning. Note also that the dynamical-equation realiza­
implicitly assumed in tion gives only the same zero-state response of the system. If the dynamical
1 (ith row, jth column)
equation is not in the zero-state, its response may not have any relation to the
lal due to a b-function system.
;e now that the internal If the realization of an impulse response G(t, T) is restricted to a finite­
lllalysis of this system dimensional linear dynamical equation of the form (4-48), it is conceivable that
not every G(t, T) is realizable. For example, there is no linear equation of the
form (4-48) that will generate the impulse response of a unit-time-delay system or
(4-4881 ) the impulse response l/U - ,). We give in the [oliowing the necessary ana suf­
(4-48b) ficient condition for G(t, T) to be realizable.
B, e, and E are n x n,
tinuous functions of t Theorem 4-9
~-48) are two different
.me input-output pairs A q x p impulse-response matrix G(t, T) is realizable by a finite-dimensional
lmical equation E with linear dynamical equation of the form (4-48) if and only if G(t, T) can be de­
composed into . ..

G(t, T) = E(t)<5(t - T) + M(t)N(T) for al\ t ¿T (4-51 )

where E is a q x p matrix and M and N are, respectively, q x n and n x p con­


T) dT (4-49) tinuous matrices of t.
156 LINEAR DYNAMICAL EQUATIONS AND IMPULSE-RESPONSE MATRICES

Proof response matrix may


Necessity: Suppose the dynamical equation networks in Figure 4·
o(t - r). For further r
E: X = A(t)x. + B(t)u
y = C(t)x. +E(t)u Time-invariant ca~
is a realization of G(t, r); then case and see what cal
G(t,r)=G(t-r). Ce
G(t, r) = E(t)O(t - r) + C(t)<ll(t, r)B(r) response G(t - r) is re~
= E(t)o(t - r) +C(t)'P(t)'P- 1 (r)B(r) equation if and only if
where'P is a fundamental matrix ofx = A(t)x.. The proof is completed by identi­ G(t -r
fying
There are two objectio
M(t) = C(t)'P(t) and N(t) = 'P- 1 (t)B(t) terms of G(t - r) inste
Sujficiency: Let to be realizable by a l
desirable is to have Cl
G(t, r) =E(t)o(t -r) +M(t)N(r) invariant dynamical-e<
where M and N are q x n and n x p continuous matrices, respectively. Then the we may also study as'
following n-dimensional dynamical equation condition of realizatio~
terms of G(t).
E: x(t) = N(t)u(t) (4-52a)
Consider a system '
y(t)=M(t)x.(t) +E(t)u(t) (4-52b)
is a realization of G(t, r). Indeed, the state transition matrix of E is an n x n
identity matrix; hence,
G(t, r) = M(t)IN(r) + E(t)O(t - r) Q.E.D. or, in the frequency do

We note that the dynamical equation in (4-52) can be simulated without


using feedback as shown in Figure 4-5 with B(t) = N(t) and C(t) = M(t). where G(o) is the imp'
matrix of the system.
Example 1 is found to be

Consider g(t, r) = g(t - r) = (t - r )e W - <). It is easy to verify that

g(t-r)=(t-r)eA(t-<)=[e At teAtll
r -re-A<l
.L'

~ Le'" J
Hence, the dynamical equation
E: Since (4-53) and (4-55) <

As discussed in (3-}
matrix, and C(sI - At 1
is a realization of g(t, r). I function matrix of the d

Theorem 4-10
All the equivalent dynamical equations have the same impulse-response
matrix; hence, ifwe find a realization of G(t, r), we may obtain different realiza­ A transfer-funciion ma
tions of G(t, r) by applying equivalence transformations. Note that an impulse- time-invariant dynamic~
[RICES IMPULSE-RESPONSE MATRICES AND DYNAMICAL EQUATIONS 157

response matrix may have different dimensional realizations; for example, the
networks in Figure 4- ~ are two different dimensional realizations of g(t, r) =
b(t - r). For further results in realization, see Reference S128.

Time-invariant case. We shall first apply Theorem 4-9 to the time-invariant


case and see what can be established. For the time-invariant case, we have
G(t, r) = G(t - r). Consequently, Theorem 4-9 can be read as: An impulse
response G(t - r) is realizable by a finite-dimensional (time-varying) dynamical
T)
equation if and only if there exist continuous matrices M and N such that
; completed by identi­ G(t - r) = M(t)N(r) + E(t)ó(t - r) forallt;:::r
There are two objections to using this theorem. First, the condition is stated in
)B(t) terms of G(t - r) instead of G(t). Second, the condition is given for G(t - r)
to be realizable by a linear time-varying dynamical equation. What is more
desirable is to have conditions on G(t) under which G(t) has a linear time­
invariant dynamical-equation realization. Since, in the time-invariant case,
we may also study a system in the frequency domain, we shall first derive the
~spectively. Then the
condition of realization in terms of transfer-function matrix, and then state it in
terms of G(t).
(4-52a) Consider a system with the input-output description
(4-52b)

latrix of E is an n x n y(t) = t G(t - r)u(r) dr

or, in thefreqúency domain,


Q.E.D.
(4-53)
be simulated without
where GO is the impulse-response matrix and G(s) is the transfer-function
ld tU) =M(t).
matrix of the system. Suppose now a state-variable description of the system
is found to be
FE: x=Ax +Bu (4-54a)
lfy that
y=Cx +Eu (4-54b)
,-A<l By taking the Laplace transform and assuming the zera initial state, we obtain
Át J
y(s) = [C(sI - At 1 B + E]u(s) (4-55)

Since (4-53) and (4-55) describe the same system, we have


G(s) = C(sI - A)-l B + E (4-56)

As discussed in (3-36) and (3-37), C(sI - At 1 B is a strictly proper rational


matrix, and C(sI - A) -1 B +E is a ptoper rational matrix. Hence, the transfer
function matrix ofthe dynamical equation in (4-54) is a proper rational matrix.

Theorem 4-10
ame impulse-response
,btain difIerent realiza­ A transfer-function matrix G(s) is realizable by a finite-dimensional linear
Note that an impulse- time-invariant dynamical equation ifand only if G(s) is a proper rational matrix.
158 LINEAR DYNAMICAL EQUATlONS AND IMPULSE-RESPONSE MATRICES

Proof forward paths and loo


If (;(s) is realizable by a finite-dimensional linear time-invariant dynamical
equation, then from (4-56), we know that (;(s) is a proper rational matrix.
Before proving that every proper rational matrix (;(s) is realizable, we first and Li¡ = 1, for i = 1,2,
prove that every scalar (1 x 1) proper rational function is realizable. The most input u to the output y
general form of a proper rational function is
A() /3I S/l-I+"'+/3/1-IS+/3/1 e +~s !/3z/s
gS = e +---,,-''--.:------,-,------;---'---''-----=-------''--''---- (4-57) 1 +a¡fs +(;(2.
s/l +O:IS/l 1 + ... +O:II-IS +a/l
This proves the assertic
We claim that the n-dimensional linear time-invariant dynamical equation
We are now ready
XI O 1 O O O XI O In order to avoid cumb
X2 O O 1 O O X2 O Let

+ u
x/I_ I O O O O X/I_ I O
and let
X/I -CJ." -ct. n - 1 -0:/1-2 -0: 2 -(;(1 X/I 1 (4-58a)

y=[/3/1 /3/1-1 /311-2 /32 /3IJX +eu (4-58b) be a realization of 9¡j, f(


is a realization of g(s). What we have to show is that the transfer function of that the bJs are colum
(4-58) is g(s). We shall demonstrate this by using Mason's formula for a signal­ posite dynamical equati
f10w graph. 8 ,9 Let us choose XI, XI, X2' X2"'" X/I' X/I as nodes; then the signal­
f10w graph of (4-58) is of the form shown in Figure 4-6. There are n loops with
loop gain ·-a¡fs, -a 2/s 2, ... , -aJs/l; and there are, except the direct trans­
mission path e, n forward paths with gains /3ds, /32/S2, ... , /3Js/l. Since all the

8 Mason's gain formula for signal-flow graph is as follows: The transfer fllnction of a signal-flow
graph is

where Ó = l-(I: all individual loop gains) +(I: aH possible gain prodllcts of two nontouching
loops)-' .. ; g¡ = gain of the ith forward path, and ó; = the part of Ó nol louching lhe ith forward is a realization of (;(s).
path. Two loops or two parts of a signal-flow graph are said to be nontollching if they do '10l have
any poinl in common. See, e.g., Reference S46.
9This can also be pro ved by computing algebraicaHy the transfer fllnction of (4-58). This is done in
Chapter 6.

e
{JI
{J2'-_­__

X2

Figure 4-6 Signal-flow graph of the dynamical equation in (4-58).


.TRICES
IMPULSE-RESPONSE MATRICES AND DYNAMICAL EQUATlONS 159

forward paths and loops have common nodes, we have

~-invariant al a2 an
dynamical /!,. = 1 +- +- + ... +- (4-59)
·oper rational matrix. s S2 sn
is realizable, we first and /!,.i = 1, for i = 1, 2, ... , n. Hence, the transfer function of (4-58) from the
realizable. The most input u to the output y is
{J IIs + {J 21S2 + ... + {J niSn {JI Sn - 1 + ... + (Jn A

e+ 2 n=e+ =g(s)
(4-57) 1 +aIls +a2/s + ... +anls s" +a 1s" 1 + ... +a n

This proves the assertion that every scalar proper rational function is realizable.
namical equation We are now ready to show that every proper rational matrix is realizable.
Xl o In order to avoid cumbersome notations, we assume that (;(s) is a 2 x 2 matrix.
Let
X2 O
+ u (4-60)
O
and let
1 (4-58a)

(4-58b) be a realization of g¡j, for i,j = 1,2; that is, gJs) =cJsI -A¡F 1bu +eu' Note
he transfer function of that the bds are column vectors and the cL/s are row vectors. Then the com­
posite dynamical equation

r'J ~b" b¡'j[::J

's formula for a signal­


lodes; then the signal­

[~"j C'
O O
There are n loops with
~cept the direct trans:
.. , (Jjs". Since all the
X12 _
X21 -
O
O
A 12
O
O
A 21 O
O
O
X 12
X 21
+ O
b 21
b 22
(4-61 a)

i 22 O O O A 22 X 22 O

C
["]
er runction or a signal-flow

[~J=[C~l
O O el!
12
O C21 C22 J X 12
X 21 +[e 21 e
12 1
J[u J
e 22 U2
(4-61 b)

X 22
oducts of two nontouching
not touching the ith forward is a realization of (;(s). Indeed, the transfer-function matrix of (4-61) is
touching ir they do not have

Dn of (4-58). This is done in


[C~l C~2 C~l c~J
O
(sI-Ad- 1
O

~
O

~Y !ln

(4-62)
4-58).
---- --------~--- ------- -- =--=---=--=-=---=----=--==-=---,.--,,-=--=--=--=-----=--=- =--~-,-===========-~-- -=-=-=--=-'==--=.. _=--='=---=---=--=' -----
~-~--

160 LINEAR DYNAMICAL EQUATlONS AND IMPULSE-RESPONSE MATRICES

Thus every proper rational matrix is realizable by a finite-dimensional linear Since a realizable
time-invariant dynamical equation, Q.E.Do since entries of 'G(l)
i = 1, 2, ... , the entri<
The realization procedure discussed in (4-60) and (4-61) is simple and straight­ ConsequentIy, the mal
forwardo We realize every element of a transfer-function matrix independently real lineo As a consec
and then connect them from the input and the output as shown in Figure 4-70
The resulting realization, however, is generalIy not satisfactory for the folIowing Theorem 4-11
two reasons: First, the realization is internalIy not coupled as can be seen from
Figure 4-7. The coupling or interacting of alI variables, however, is a feature of A system that has a pr,
most physical multivariable systemso Second, the dimension of the realization only if U[lo,lo+EJ =0 in
is generalIy unnecessarily large. These two problems will be resolved in Chapter
6. This is a restateme
The condition ofTheorem 4-10 is stated in terms of transfer-function matri­ have rational transfer
ces. We may translate it into the time domain as folIows: linear time-invariant e
an interval (no matter
Corollary 4-10 interval.

An impulse-response matrix G(t) is realizable by a finite-dimensional linear


time-invariant dynamical equation if and only if every entry of G(t) is a linear
combination of terms of the form tke Ail (for k = 0, 1, 2, o.. , and i = 1, 2, o. o) and 4-5 Concluding
possibly contains a b-function at t =0. I
The solutions of !ine;
The impulse-response matrix G(t) is the inverse Laplace transform of the The solution hinges on
transfer-function matrix (;(s). If a proper rationaI function is of the form ties <1>(1, t) = 1, <1>- l(t,
varying case, <1>(t, -r) is
A N(s) <1>(t, -r) is equal to eA(l­
g .(s) = d + .,.---:..---,----,-­
o

IJ (s - Ad'(s - A2 t 2••• duced in Section 2-7.


then its inverse Laplace transform %(t) is a linear combination of the terms may bypass <1>(t, -r) an<
integration.
Oifferent analyses l
Hence, Corollary 4-10 follows directly from Theorem 4-100 of a system. Mathema
basis chosen for the st
nothing to do with the
,-------------, to the same input-outPl
I [

description over the stc


L :
"':

iT~l
lA

:""' I ~r :
Y,
Every proper ratior
dimensional!inear time

-~.~
synthesis problem in ne
: i was also constructedo
generally it is possible t
: I
, 1---
I I
A

g,2(S)
I
f-
l'
I
I
I
be discussed in Chapter
bility,
; \ I

~
"2 • I r~
~ :Y2
+rc--'--­
To conclude this el
equationso The state
: 1 . :

A(k)A(k -lJx(k -1) = .


. I

I
...J A(m) (Problem 4-27). 1
putable. In terms of <1>(,
Figure 4-7 lnternally uncoupled realization of (4-60), can be obtained as in Pr,
TRICES CONCLUDING REMARKS 161

te-dimensional linear Since a realizable G(t) can be decomposed into G(t - r) = M(t)N(r), and
Q.E.O. since entries of "G(t) are linear combinations of tkeÁ¡t, for k =0,1, ... , and
i = 1,2, ... , the entries of M(t) and N{t) must be linear combinations of tke Áit .
is simpleand straight­ Consequently, the matrices M(t) and N(t) are analytic functions of t on the entire
matrix independently realline. As a consequence of this fact, we have the following theorem.
: shown in Figure 4-7.
~tory for the following Theorem 4-11
:d as can be seen from A system that has a proper rational transfer function matrix is relaxed at to if and
owever, is a feature of
only if u[lo.lo+'l = O implies Y[to.to+'l = O for sorne positive real s. I
sion of the realization
be resolved in Chapter
This is a restatement of Corollary 3-1. Hence, for the class of systems that
have rational transfer-function matrices-or equivalently, are describable by
ansfer-function matri­
linear time-invariant dynamical equations-if the output is identically zero in
s: an interval (no matter how smal1), then the system is relaxed at the end of that
interval.

itecdimensional linear
ntry of G(t) is a linear 4-5 Concluding Remarks
., and i = 1, 2, ...) and
I
The solutions of linear dynamical equations were studied in this chapter.
The solution hinges on the state transition matrix <I>(t, r), which has the proper­
)lace transform of the ties <I>(t, t)=I, <I>-l(t, r)=<I>(r, t), and <I>(t, r)<l>(r, to)=<I>(t, tolo For the time­
ion is of the form varying case, <I>(t, !) is very difficult to compute; for the time-invariant case,
<I>(t, !) is equal to eA(t-t l , which can be computed by using the methods intro­
duced in Section 2-7. In both cases, if only a specific solution is of interest, we
may bypass <I>(t, !) and compute the solution on a digital computer by direct
lbination of the terms integration.
Oifferent analyses often lead to different dynamical-equation descriptions
of a system. Mathematically, it means that dynamical equations depend on the
10. basis chosen for the state space. However, the input-output description has
nothing to do with the basis; no matter what analysis is used, it always leads
to the same input-output description. This is an advantage of the input-output
description over the state-variable description.
Every proper rational function has been shown to be realizable by a finite­
dimensional linear time-invariant dynamical equation. This corresponds to the
synthesis problem in network theory. A realization of a proper rational matrix
was also constructed. However, the realization is not satisfactory, because
general1y it is possible to construct a lesser-dimensional realization. This will
be discussed in Chapter 6 after the introduction of controllability and observa­
bility.
To conclude this chapter, we remark briefly the solution of discrete-time
equations. The state transition matrix <I>(k, m) of .x(k + 1) = A(k)x(k) =
A(k)A(k -l),x(k - 1) = ... can be readily computed as <I>(k, m) = A(k -1)'"
A(m) (Problem 4-27). Unlike the continuous-tim~ case, <I>(k, m) is easily com­
putable. [n terms of <I>(k, m), the solution ofa discrete-time dynamical equation
can be obtained as in Problem 4-29. For the time-invariimt case, the equivalent
162 LINEAR DYNAMICAL EQUATlONS AND IMPULSE-RESPONSE MATRICES

dynamical equations and the realization problem are identical to the con­ 4-7 Given
tinuous-time case. For example, a sampled transfer-function matrix (;(z) has a
finite-dimensional realization of the form in (3-80) if and only if G(z) is a proper
rational matrix in z. The realization procedure in Theorem 4-10 is also directly
applicable to the discrete-time case. show that

Problems where o<1l(t, tol/Ot = A(t)lIJ

4-1 Find the fundamental matrices and the state transition matrices of the following
o
homogeneous equations: o,

4-8 Givenx(t)=A(t)x.
where A* is the complex '
transition matrioes of x =
and

4-2 Show that ol1>(t, ')/0' = -11>(t, ,)A(,).


4-9 Consider

4-3 Find the solution of


and its adjoint equation

x=[~ ~ ~]x+[~ ~]u


-2 -4 -3 -1 1 shown in Figure P4-9. N,
. Ga(t, ,) be their impulse re:
1
2
with Show that if A, R, and e are
matrices, then

uU) = [~J for t 2':0

4-4 Let

o 1

1 O

O 1

O O

Find eA' by using the formula 2'[é'] =(sI-A)-l. w

4-5 lf T- l(t) exists ~nd is <:Íifferentiable for aH t, show that

d
-"[T- l(t)] = - r l(t) [el
- TU) ] r l(t)
de . elt ..
4-6 From l1>(t, ¡), show how to compute AU). Figure P4-9
PROBLEMS 163
.RICES

identical to the con­ 4-7 Given


ion matrix (;(z) has a
mly if G(z) is a proper
m 4-10 is also directly
show that

det <I>(t, to)=exp [ [ (at¡(r) +adr))dr]

where 8<1>(t, to)lat = A(t)<I>(t, to) and <I>(to, to) = I. Hint: Show that
a
latrices of the following - det <I>(t, to) = (a¡¡(t) +a 22 (t)) det <I>(t, to)
ot
4-8 Given x(t) =A(t)x. The equation z= -A*z is called the adjoint equation ofx =A(t)x,
where A* is the complex conjugate transpose of A. Let <I>(t, to) and <l>a(t, to) be the state
transition matrices of x= A(t)x and z= - A*z, respectively. Verify that

4-9 Consider

:X = A(t)x + B(t)u y =C(t)x

and its adjoint equation


z= -A*(t)z +C*(t)v w =B*(t)z
shown in Figure P4-9. Note the reversal of the flow direction of signals. Let G(t, r) and
Gu(t, r) be their impulse response matrices. Show that

G(t, r) = G:(r, t)
Show that ir A, B, and C are constant matrices and G(s) and Gis) are their transfer-function
matrices, then
G(s) = - G:( -s)

Figure P4-9
164 LINEAR OYNAMICAL EQUATIONS ANO IMPULSE-RESPONSE MATRICES

4-10 Every e1ement of <1>(t, to) can be interpreted as the impulse response of sorne input­

output pairo What is the input and the output of the·ijth element of <1>(t, to)?

4-11 Let
In

Figure P4-17
be the state transition matrix of
and the initial capacitor ,

X=[A~¡ :::}
immediately of the form ,

4-18 Find an equivalen


Show that <1>2¡(t, to) =0 for all t, t o and that (a/at)<J>;;(t, to) = A¡;<1>¡;(t, to) for i = 1, 2.

4-12 Verify that B(t) = <1>(t, t o)Bo<1>*(t, t o) is the solution of


ti 4-19 Find an equivalen"
-B(t)=A(t)B(t) +B(t)A*(t) B(to)=B o

[
tlt
where <1>(t, to) is the state transition matrix of x = A(t)x.

4-13 Verify that X(t) = eA'CeBl is the solution of


ti
-X=AX +XB X(O)=C
tlt
4-14 Show that if Á(t) =A¡A(t) -A(t)A¡, then
4-20 Find an equivalen!
A(t) = eA,lA(O)e- A"
X i (11 -t
Show also that the eigenvalues of A(t) are independent of t.
x 2(11 -t
[
4-15 Show that if Á(t) = A¡A(t) - A(t)A¡, then a fundamental matrix of x = A(t)x is given X 3(11 -t
by

where A2~A(O)-A¡.

4-16 Find the impedance (the transfer function from u to i) of the network in Figure
4-21 Can you transform
P4-16. lf the initial conditions of the inductor and the capacitor are zero and if an input
equivalence transformatioJ
voltage u(t)=e- c is applied, what are ;(t), ;¡(t) and iz(t)? Note that ;j(t) and ;2lt) contain
sorne exponential functions that do not appear in ¡(t). How do you explain this? 4-22 Find a time-varying
realization of the impulse r
4-17 Consider the network shown in Figure P4-17. Find the initial inductor current
4-23 Find a dynamical-e(
find a linear time-invarianl

4-24 Use a signal-ftow g


+ 2n variable, linear time-invari
u 'V

4H

Figuré P4-16

y=[l
.TRICES PROBLEMS 165

: response of some input­

tt of <1>(t, to)?

Figu re P4-17

and the initial capacitor voltage such that for the input u(t) = e- 41 the output y(t) wil1 be
immediately of the form e- 41 without containing any transient.

4-18 Find an equivalent time-invariant dynamical equation of


(t, to) for ; = 1,2.
x= (cos t sin t)x
4-19 Find an equivalent Jordan-canonical-form dynamical equation of

1~][::]+[-~]
4
20 u
-25 -20 X3 O

4-20 Find an equivalent discrete-time Jordan-canonicai-form dynamical equation of

XI(n +
x¡(n+l) =
1)] [O
O 20
4
3][X1(n)] + [-1]
16 x¡(n) 3 u(n)
[
natrix of x= A(t)x is given x3(n + 1) O -25 -20 x3(n) O

XI(nl]
y(n)=[ -1 3 OJ x¡(n) +4u(n)
[x (n)
3

of the network in Figure 4-21 Can you transform a time invariant {A, B, C} into {O, B, C} by a time-varying
Jr are zero and if an input equivalence transformation?
that ;1(1) ~.nd ;1(t) contain
you explain this? 4-22 Find a time-varying dynamical equation realization and a time-invariant dynamical
realization of the impulse response g(t) = t 2 e!·I.
he initial inductor current
4-23 Find a dynamical-equation realization of g(t, e) = sin te - (r -<) cos e. ls it possible to
find a linear time-invariant dynamical-equation realization for it?

4-24 Use a signal-f1ow graph to show that the transfer function of the following single-
variable, linear time-invariant dynamical equation

X~ [¡
O O
O O
1 O
O

O -"o
-ct,,_l j
13U-l .

O -~n-¡ x + P";-2 u
[~.]
O O -!XI PI

y=[O O O O lJx+eu
166 LINEAR OYNAMICAL EQUATIONS ANO IMPULSE-RESPONSE MATRICES

is Verify tha t
A {31 S ,,-1 + ... +{3"
g(s)=e + S" +O:IS"-1 + ... +CI."_IS +0:"

4-25 Realize the proper rational matrices Verify also that

1
S+2 2+2

ls +1
5
s +1
s +3
Ss
1

s +2
+1
1
l
into continuous-time i:\nd discrete-time dynamical equations.
2

2
+1
5
+1
2

52
2
+3
1
+1
+2
if and only if A(t) and A((

4-31 Show that if f


lO
A(e

(
4-26 Consider the equivalent dynamical equations

{*y=Cx
Ax + Bu
= {x ~x + Bu =
é

y=Cx is
where x = Px. Their adjoint equations are, respectively,
[Hint: Use Problem 4-30 a
:i= -A*z+C*u (la)
{ y=B*z (lb)
z= ::-A*z + C*u (2a)
{y=B*z (2b)

where A* and Á* are the complex conjugate transposes of A and Á, respectively. Show
that Equations (1) and (2) are equivalent and that they are related by z= (p-l )*z.

4-27 Consider x(k + 1) = A(k)x(k). Define


<1>(k, m)~ A(k - 1)A(k - 2)A(k - 3) ... A(m) for k>m

<1>(m, m)~ 1

Show that, given the initial state x(m) = xo, the state at time k is given by x(k) = <1>(k, m)x o·
If A is independent of k, what is <1>(k, m)?

4-28 For continuous-time dynamical equations, the state transition matrix <1>(t, e) is
defined for aH t, T. However, in discrete-time dynamical equation, <1>(k, m) is defined only
for k 2: m. What condition do we need on A(k) in order for <1>(k, m) to be defined for !< < m?

4-29 Show that the solution of x(k + 1) =A(k)x(k) + B(k)u(k) is given by


k-l
x(k) = <1>(k, m)x(m) +I <1>(k, 1 + 1)B(l)u(l)
l=m

[This can be easily verified by considering B(l)u(l) as an initial state at time (l + 1).J Show
that if A(k) and B(k) are independent of k, then the solution becomes
k-l
x(k) = Akx(O) +I Ak - 1
-
III
Bu(m)
m=O

4-30 Let A(t) ~ (aij(t)). Then, by definition,

d (d
- A(t) ~ -'--- adt)
dt dt J
)
ArRICES PROBLEMS 167

Verify that

d . •
~ (A(t)B(t)) = A(t)B(t) + A(t)B(t)
dt
Verify also that
d d .
- [A(tW ~- (A(t)A(t)) = 2A(t)A(t)
dt dt

;j
if and only if A(t) and A(t) commute; that is, Á(t)A(t) = A(t)Á(t).

4-31 Show that if r A(r) dr and A(t) commute for all t, then the unique solution of
'o o
- <1>(t, to) = A(t)<l>(t, to) <1>(to, to) = 1
oc
is <1>(t, to)=exp flO
A(r) dr

[Hint: Use Problem 4-30 and Equation (2-78).J


(la)
(lb)
(2a)
(2b)

d f,., respectively. Show


an -1 *
edbyZ==(P )Z.

for k>m

.IS g'¡ven by x(k) = <1>(k, m)xo·

trans ition matrix <1>(t, r) is


. .....(k, m) is defined only?
Ilion,....
<:.111) to be defined for k < 111.

lis given by

)11(1)

state at time (1 + l).J Show


ecorn es
observable from the ou
in II can be detected fre
This iIIustration, thouÉ
cepts of controllability
The concepts of cor
study of control and fi\¡
systern shown in Figun
5 of the platform are sUPI
The rnass of the platfon
Controllabi Iity and rnents of the two spring
both ends of the platfe
vibrate. If no force is ¡
Observability of Linear come back to rest. No~
to apply a force to bring
Dynamical Equations this question, the conce]
This chapter is orga
rnatical background is
for linear independence
results in controllabilit,
three theorerns. The c:
Necessary and sufficien
5-1 Introduction tions and linear time-il
derived. The concept o
System analyses generally consist of two parts: quantitative and qualitative. to the concept of contrc
In the quantitative study we are interested in the exact response of the system theorern is also devel0l
to certain input and initial conditions, as we studied in the preceding chapter. which are uncontrolIabh
In the qualitative study we are interested in the general properties of a systern. theorern is developed.
In this chapter we shall introduce two qualitative properties oflinear dynamical function rnatrix of a dy
equations: controllability and observability. We shall first give the reader sorne equation that is control
rough ideas of these two concepts by using the network shown in Figure 5-1. controllability and obser
The input u ofthe network is a current so urce. It is c1ear that ifthe initial voltage cal equation. Their COI
in the capacitor e 2 in loop II is zero,no rnatter what input u is applied, the inspection. In Section :
mode e- t in II can never be excited. Hence the mode e- t in H is said te be y:: function controllabHitv :
controllable by the input u. On the other hand, the rnode e - t in loop 1 can be of the input-output d~sc
excited by the application of the input u; hence the mode in I is controllable by sorne cornputational proi
the input u. Although the mode e -1 in 1 can be excited by the input, its presence Although elements 01
can never be detected frorn the output terminal y. Hence it is said to be not mathernatical convenienc

Xl ¡ er'
u
Dampi.ng
coefficien!

,11 Spring

constan!

(curren!
2
1
service)

Figure5-1 A simple network.

Figure 5-2 A platform sys


168
INTRODUCTION 169

observable from the output y. On the other hand, the presence ofthe mode e- l
in II can be detected from the output y; hence the mode is said to be observable.
This illustration, though not very accurate, may convey the ideas of the con­
cepts of controllability and observability.
The concepts of contrQllability and observability are very important in the
study of control and filtering problems. As an example, consider the platform
system shown in Figure 5-2. The system consists of one platform; both ends
of the platform are supported on the ground by means of springs and dashpots.
The mass ofthe platform is, for simplicity, assumed to be zero; hence the move­
ments ofthe two spring systems are independent. Ifthe initial displacements of
both ends of the platform are different from zero, the platform will start to
vibrate. If no force is applied, it will take an infinite time for the platform to
come back to rest. Now we may ask: For any initial displacements, is it possible
to apply a force to bring the platform to rest in afinite time? In order to answer
this question, the concept of controllability is needed.
This chapter is organized as follows. In Section 5-2 the required mathe­
matical background is introduced. Three theorems that give the conditions
for linear independence of a set of vector functions are presented. All the
results in controllability and observability follow almost directly from these
three theorems. The concept of controllability is introduced in Section 5-3.
Necessary and sufficient conditions for linear time-varying dynamical equa­
tions and linear time-invariant dynamical equations to be controllable are
derived. The concept of observability is introduced in Section 5-4. It is dual
to the concept of controllability; hence its discussion is rather brief.· Duality
ltitative and qualitative.
theorem is also developed. In Section 5-5, we study dynamical equations
:t response of the system
which are uncontrollable and/or unobservable. The canonical decomposition
n the preceding chapter.
theorem is developed. A consequence of this theorem is that the transfer­
II properties of a system.
function matrix of a. dynamical equation depends solely on the part of the
~rties oflinear dynamical
equation that is controllable and observable. In Section 5-6, we study the
first give the reader sorne
controllability and observability of linear time-invariant Jordan-form dynami­
)rk shown in Figure 5-l.
cal equation. Their conditions are very simple and can be checked almost by
.r that ifthe initial voltage
inspection. In Section 5-7, the concepts of output controllability and output
lt input u is applied, the
function controllability are introduced. It is shown that they are properties
e- t in 11 is said to be not
of the input-output description of a system. In the last section, we discuss
node e -1 in loop 1 can be
sorne computational problems encountered in this chapter.
)de in 1 is controllable by
Although elements of the matrices A, B, e, and E are all real-valued, for
by the input, its presence
mathematical convenience they are considered as elements of the field of com­
-lence it is said to be not

¡2U
Xl
I .'":';'
¡ x2
Damping Spring Damping Spring
coefficient constant coefficient constant
2 1 J 1

Figure 5-2 A platform system.


170 CONTROLLABILlTY ANO OBSERVABILITY OF LINEAR OYNAMICAL EQUATlONS

plex numbers. Consequently, the state space of an n-dimensional dynamical The concept of lin
equation will be taken as an n-dimensional complex vector space (C", iC). tions. Let f j , for i = 1,
The references for this chapter are 2, 8,11,13,14,20,21,48,55,56,60,61, 1 x p complex-valued
69, 71, 98, 103, and 105 to 107. if there exist complex
The reader who is interested in only the time-invariant case may skip
alfl(t) +a
Theorems 5-2, 5-5, and 5-6.
Otherwise, the f¡'s are
in Equation (5-2) is a
5-2 Linear Independence of Time Functions we may also state tha
only if
The concept of linear independence of a set of vectors of a linear space was
introduced in Section 2-3. We shall now apply this concept to a set offunctions
ofa real variable. A set of complex-valued functions Jl,f2> ... , Jn is said to be
linearly dependent on the interval 1 [t 1> t 2] over the field of complex numbers
if there exist complex numbers al' a2' ... , a m not all zero, such that
aI!I(t)+al!2(t)+··· +allfn(t) =0 foral/tin [tI' t 2] (5-1) implies IX = 0, where
Otherwise, the set of functions is said to be linearly independent on [t 1> t 2 ] over
the field of complex numbers. In this definition, the specification of time
interval is crucial.

Example 1
Clearly, Ot is a constan
Consider the two continuous functions 11 and 12, defined by The linear indepen
J~(t)=t for / in [ -1, 1] an interval; hence in t
for t in [0, 1] entire interval. Let F'
12(t) = { t
-t for t in [ - 1, O]
Theorem 5-1
lt is clear that the functions(1 and.f2 are linearly dependent on [O, 1], since if Let f j , for i = 1, 2, ... , t
we choose al = 1, a2 = -1, then a¡j"l(t) +ad1.(t) =0 for aH I in [0, 1]. The on [/,. I J. Let F be t
functions JI and j~ are also linearly dependent on [ -1, O]. However, JI and
.f2 are linearly independent on [ -1, 1]. I

From this example, we see that although a set of functions is linearly inde­ Then f l , f 2 , .. o. f are
ll

pendent on an interval, it is not necessary that they are linearly independent on constant matrix W(t l ' I

any subinterval. However, it is true that there exists a subinterval on which


they are linearly independent. For example, in Example 1 the functionsJl and Proof
J2 are linearly independent on the subinterval [ -e, e] for any positive eo On The proof of this theore
the other hand, if a set of functions is linearly independent on an interval [t l ' t 2],
is nonsingular ir and (
then the set of functions is linearly independent on any interval that contains
We prove first the ne
[1 l ' t 2 ]. Assume that the f/s are
Then there exists a nOI

1 TÍJe functions we study are mostly continuous functions: hence there ¡s·no sUDstantia\ difreorence
between using the open intervallt ,. e2) and the closedinterval [e l' e2] Every [e ,. e2 ] ís assumed to 2 The matrix W(t ,. e2) is in ri
be a nonzero interval. in this chapter.
.__ ._~.:=~---.--_.-':::_ .. _-~---~

ICAL EQUATIONS LINEAR INDEPENDENCE OF TIME FUNCTlONS 171

limensional dynamical The concept of linear independence can be extended to vector-valued func­
:tor space (Cn, C). tions. Let f¡, for i = 1, 2, ... , n, be 1 x p complex-valued functions of t; then the
),21, 48,55, 56, 60, 61, 1 x p complex-valued functions f l , f l , ... , f n are linearly dependent on [t l' tlJ
if there exist complex numbers al' al" .. ,a", not all zero, such that
variant case may skip (5-2)

Otherwise, the f;'s are linearly independent on [t 1> t l ]. Note that the zero vector
in Equation (5-2) is a 1 x p row vector [O ° ...
O]. As in Definition 2-4',
we may also state that f¡, f l , ... , f n are linearly independent on [ti' tlJ if and
'Os
only if
; of a linear space was
:ept to a set offunctions
,fl' ... , fn is said to be for aH t in [tI' tlJ (5-3)
Id of complex numbers
J, such that

'tin [ll,tlJ (5-1) implies t:x = O, where

F"r~:l
)endent on [t l' t lJ over
~ specification of time
... J
an

lf..J
Clearly, t:x is a constant 1 x n row vector and F is an n x p matrix function.
ed by
The linear independence of a set of functions is a property associated with
an ¡nterval; hence in testing for linear independence, we have to consider the
entire interva\. Let F*(t) be the complex conjugate transpose of F(t).

Theorem 5-1
ldent on [0, lJ, since ir Let f j , for i = 1, 2, ... ,n, be 1 x p complex-valued continuous functions defined
)r all I in [0, 1]. The on [1 l' I J. Let F be the 11 x (! matrix with f; as its ith row. Define
,0]. However, fl and
I W(t]. 12)~ r" F(t)F*(t) dI
Jt l

rlctions is linearly inde­ Then f l • f 2 , ...• f n are linearly independent on [1 \. I J ir and only ir the 11 x 11

linearly independent on constant matrix W(t l' ( 2 ) is nonsingular. 2


a subinterval on which
e 1 the functions fl and Proof
for any positive 6. On
The proof ofthis theorem is similar to that ofTheorem 2-8. RecaH that a matrix
ltonan interval [tI' tlJ, is nonsingular if and only if all the columns (rows) are linearly independent.
V interval that contains
We prove first the necessity of the theorem; we prove it by contradiction.
Assume that the f/s are linearly independent on [t l' t lJ, but W(t l ' t 2) is singular.
Then there exists a nonzero 1 x n row vector t:x such that t:xW(t¡, tl)=O. This

re is no substantia\ difference
Every [t,. t,] is assumed to , The matrix W(t ,. t ,) is in faet positive definite(see Definition 8-6). This property is not needed
in this ehapter. .
=. . .=-.--=..
·--'="--==.C'-... =----C-.~._."" .• ,_•.• -._.-~ •• -" ..••. _.-.-.
,_U . , , _ . _ _. _ •• _ - • • •'_ •• _-_ •• _ - , - _ ••••• _ ••• "~ ••

172 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAM¡CAL EQUATIONS

implies exW(t¡, t 2 )ex* =0, or and the C;s are linear


'2 nonzero 1 x n row v'
exW(t¡, t 2 )ex* =
1'1 (exF(t))(exF(t))* dt =0 (5-4)

This implies that


Since the integrand (exF(t))(exF(t))* is a continuous function and is nonnegative
for all t in [t ¡, t 2 J, Equation (5-4) implies that exF(k)( t) = I

exF(t) =0 for all t in [tI' t 2 J Hence we have


This contradicts the linear independence assumption of the set of f;, i = 1, 2, ex[F(t) : F
... , n. Hence ifthef¡'s are linearly independent on [tI' t 2 J, then det W(t l , t 2 ) f O.
We next prove the sufficiency of the theorem. Suppose that W(t l , t 2 ) is in particular,
nonsingular, but the lis are linearly dependent on [t l' t 2]. Then, by definition,
there exists a nonzero constant 1 x n row vector ex such that exF(t) = O for a11
t in [tI' t 2 ]. Consequently, we have which implies that al

r exF(t)F*(t)dt=O
t2

exW(t l ,t 2 )= (5-5)
J" are linearly depender
which contradicts the assumption that W(t l , t 2 ) is nonsingular. Hence, if
W(t l' t 2 ) is nonsingular, then the f¡'s are linearly independent on [t ¡, t 2 ].
Q.E.D. has rank n. Hence ti

The determinant of W(t l' t 2) is called the Gram determinant of the f¡'s. In The condition of
applying Theorem 5-1, the functions f i , for i = 1,2, ... , n, are required to be functions to be linea
example.
continuous. lf the functions f;, for i = 1, 2, ... , n, have coIitinuous derivatives
up to order (n -1), then we may use the fo11owing theorem.
Example 2
Thearem 5-2 Consider the two fun(
Assume that the 1 x p complex-valued functions f l , f 2 , . . . , f" have continuous
derivatives up to order (n - 1) on the interval [t ¡, t 2]. Let F be the n x p matrix
with f i as its ith row, and let F(k) be the kth derivative of F. lf there exists some
defined ayer l -1, 1].
t o in [tI' t 2J such that the n xnp matrix

has rank n, then the f¡'s are linearly independent on [t J, t 2J over the field of
complex numbers. 3

Proof
and
We prove the theorem by contradiction.' Suppose that there exists some t o in
[tI' t 2 J such that
To check the linea
continuous, we can em
interval. If the functit
, Ir lo is al eilher 1, or 12 , lhe end points of an interval, then FÚ<l(to) IS denned as· ·F1k)(I). with I then Theorem 5-2 can 1
approaching lo from inside the interval. lheorem 5-1; however
CAL EQUATIONS LINEAR INDEPENDENCE OF TIME FUNCTIONS 173

and the f;s are linearly dependent on [t b t2]. Then by definition, there exists a
nonzero 1 x n row vector ex such that
=0 (5-4 )
exF(t) =0
This irnplies that
on and is nonnegative
exF(kl(t) = O for all t in [t l> t 2] and k = 1, 2, ... , n - 1
Hence we have
f the set of f j , i = 1, 2, ex[F(t) : F(l)(t) : ... : F(ll- ll(t)] = O
], then det W(t l , t 2) 1= O.
in particular,
)pose that W(t l' t 2) is
l. Then, by definition, ex[F(to): F(ll(t O) : ... : F(ll-l)(t O)] =0
1 that exF(t) =0 for aH
which irnplies that aH the n rows of
[F(to) : F(l)(t o) : ... : Fln-l)(to)]
(5-5 )
are linearly dependent. This contradicts the hypothesis that
)nsingular. Hence, if [F(t o) : F(l)(t o) : ... : F(ll-l)(t O)]
lependent on [ti' t 2 J. Q.E.D.
has rank n. Hence the f;s are linearly independent on [ti' t 2 ].
Q.E.D.
The condition of Theorern 5-2 is sufficient but not necessary for a set of
rminant of the f¡'s. In functions to be linearly independent. This can be seen from the following
, n, are required to be exarnple.
;;ontinuous derivatives
~rn.
Example 2
Consider the two functions
.. , f n have continuous fl(t) = t3
:t F be the n x p matrix f2(t) = It 3 1
F. If there exists sorne defined over [ -1, 1]. They are linearly independent on [ -1, 1]; however,

to)] (S-S) IJIU) f"?l(tll ,rt 3


3t 2 1
1J
p Lf2(t) J?)(t)J = ¡; Lt 3 J
3t 2 = 1

rOí" aH L in lO,
l' t 2 ] over the field of 3 2

fl(t) f?)(t)J= [ t 3t J=1 for all t in [-1, O)


p [ f2(t) fPl(t) p _t 3 -3t 2

fl(t) j¡<l)(t)J

there exists sorne t o in


and p [ f2(t) fP)(t) =0 at t =0 •
To check the linear independence of a set of functions, if the functions are
continuous, we can employ Theorem 5-1, which requires an integration over ario
interval. If the functions are continuously differentiable up to certain order,
) IS defined as F(k)(t), with t then Theorern 5-2 can be used. lt is clear that Theorern 5~2 IS easier to use than
Theorem 5-1; however, it gives only sufficient conditions. lf the functions are
._ •... ._ .. -.- -_ - ._._.
" ..
---~------"'---------------

174 CONTROLLABILlTY AND OBSERVABILlTY OF LINEAR DYNAMICAL EQUATlONS

analytic, then we can use Theorem 5-3, which is based on the fact that if a every subinterval of
function is analytic on [tI' t 2 J, then the function is completely determinable essential. The staten
from a point in [t 1 , t 2 J if aH the derivatives of the function at that point are Note that Theore
known. (See Appendix B.)

Theorem 5-3 This can be seen fron


Assume that for each i, f i is analytic on [t l' t 2]. Let F be the n x p matrix with
f i as its ith row, and let F(k) be the kth derivative of F. Let t o be any fixed point Example 3
in [t l' t 2]. Then the f/s are linearly independent on [t l' t 2J if and only if Let
(5-7)

Proof
Then
The sufficiency of the theorem can be proved as in Theorem 5-2. Now we
prove by contradiction the necessity of the theorem. Suppose that [F(t) : 1
p[FU o) : F(l)UO) : ... : F(n-l)(to) : .. 'J < n
It is easy to verify th
Then the rows of the infinite matrix
p[F(t) : flll(t) : F(2 l (t) : F(
[F(to): F(I)(t o): ... : F(II-l)U O): "'J sin lO't '
=p [
are linearly dependent. Consequently, there exists a nonzero l x n row vector sin 2 x lO't : 2 x 1
a. such that for al! t.
(5-8)
The matrix in (5-~
The Í¡'s are analytic on [t 1 , t 2 J by assumption; hence there exists an 6>0 such many cases, it is not no
that, for aH t in [to -6, t o +6J, F(t) can be represented as a Taylor series about Example 3, we check
the point t o:
~o)" F(II)(t O)
n.
F(t) = ffor aH t in [to - 6, to +6J
11=0
(t - (5-9)
then we have the follo
Premultiplying a. on both sides of (5-9) and using (5-8), we obtain
a.F(t) =0 for al! t in [to -6, t o +6J (5-10) Corollary 5-3
Since the sum of analytic functions is an analytic function, the analyticity Assume that, for each
assumption of the f/s implies that a.F(t) as a row vector function is analytic over independent on [t b t 2
[t 1 , t 2 ]. Consequently, Equation (5-10) implies that
a.F(t) =0 for al! t in [t 1, t 2 J
for almost all t in [t 1> j
or, equivalently, the f/s are linearly dependent on [t1> t 2 ]. This is a contra­ This corollary wil
diction. Q.E.D. omitted.
A direct conseque'nce of this theorem is that if a set of anal ytic functions is
linearly independent on [t1> t 2 J, then
5-3 Controllabil
p[F(t): f'(I)(t) : ... : F(n-1)(t)=--'J =n
Time-varying case
for al! t in [t l' t 2]. Itfollows that if asid of analytic functions is linearly inde­ trollability of linear d
pendent on [t l' t 2 J, then the set of analytic functions is linearly independent on state controllability o
o\L EQUATIONS CONTROLLABILlTY OF LINEAR DYNAMICAL EQUATlONS 175

on the fact that if a every subinterval of [tI' t 2 ]. In this statement, the analyticity assumption is
pletely determinable essential. The statement does not hold without it, as we have seen in Example 1.
ion at that point are Note that Theorem 5-3 is not true if the infinite matrix in (5-7) is replaced by
p[F(ro) : F(l)(ro) : ... : Fin - l)(tO)J = n

This can be seen from the following example.


the n x p matrix with
Example 3
t o be any fixed point
Jif and only if Let
=n (5-7) _ [sin 1000tJ
F(t) - sin 2000t

Then
~orem 5-2. Now we
sin lOOOt: 10 3 cos lOOOt J
'pose that [F(r) : F(l)U)] =.
[ sm 2000t :' 2 x 10 3 COS 20001
<n It is easy to verify that p[F(t): F(l)(t)] < 2 at t = O, ± 10 - 3n;, . . .. However,
p[F(t) : Fl)(t) : F(2)(t) : F( 3 )(t)]

=p [
sin 10 3 r 10 3 cos 10 3 r -10·sinI0 3 r
3
' -10·cosI0 3 r J
zero 1 x n row vector sin 2 x 10 3 { : 2 x 10 3 cos 2 X 10 3 r -4xI0.sin2xl0 r :-8xI0 9 cos2xI0 3 r =2
for all t. I

=O (5-8)
The matrix in (5-7) has n rows but infinitely many columns. However, in
:e exists an e> O such many cases, it is not necessary to check aH the derivatives of F. For instance, in
a Taylor series about Example 3, we check only up to F(3 l . lf we use the matrix
[F(t) : F(l)(t) : ... : F(n -l)(t)]
e, t o +eJ (5-9)
then we have the following corollary.
e obtain
] (5-10) Corollary 5-3
ction, the analyticity Assume that, for each i, f¡ is analytic on [t l' t 2]. Then f l' f 2 , ... , f n are linearly
lction is analytic over independent on [t 1, t 2 J if and only if
p[F(t) : F(l)(t) : ... : F(n -l)(t)] = n
for almos! all t in [tI' t 2 ].
2]. This is a contra­ This corollary will not be used in this book, and therefore its proof is
Q.E.D. omitted.
.f analytic functions is

5-3 Controllability of Linear Dyn~mical Equations


=n
Time-var\'¡ng case. In this secti~n, we sh~ll int~oduce the conceptof con­
ctions is linearly inde­ trollability of linear dynamical equations. Tobe more precise, we study the
nearly independent on state controllability of linear state equations. As will be seen immediately,

1
176 CONTROLLABILlTY AND OBSERVABILlTY OF LINEAR DYNAMICAL EQUATIONS

the state controllability is a property of state equations only; output equations precisely, the dynami
do not play any role here. at any to.
Consider the n-dimensional, linear state equation
Example 2
E: x= A(t)x(t) + B(t)u(t) (5-11 )
Consider the system ~
where x is the n x 1 state vector, u is the p x 1 input vector, and A and B are,
X2 in the system, T
respectively, n x n and n x p matrices whose entries are continuous functions
cannot transfer Xl ay.
of t defined over ( - 00,00). The state space of the equation is an n-dimensional
then no matter what
complex vector space and is denoted by ~.
Hence the equation t
Definition 5-1
The solution of tn
The state equation E is said to be (state) controllablé at time to, if there exists a
finite ti > t o such that for any x(to) in the state space ~ and any Xl in ~, there x(t) = «I»(t; i
exists an input u[to,ld that will tIansfer the state x(t o) to the state Xl at time ti'
Otherwise, the state equation is said to be uncontrollable at time to.

This definition requires only that the input u be capable ofmoving any state
in the state space to any other state in afinite time; what trajectory the state where <I>(t, t o) = 'P(t)\
should take is not specified. Furthermore, there is no constraint imposed on the nonsingular for aH t.
input. Its magnitude can be as large as desired. We give some examples to
illustrate this concept. Theorem 5-4
The state equation E
Example 1 ti> t o such tqat the 11
Consider the network shown in Figure 5-3. The state variabk x ofthe system is independent on [t o, ti
the voltage across the capacitor. Ir x(to) = O, then x(t) = Ofor all t '2:. t o no matter
what input is applied. This is due to the symmetry ofthe network, and the input Proof
has no effect on the voltage across the capacitor. Hence the system-or, more Sufficiency: If the row
Theorem 5-1 the n x y.

W(t
4 In lhe lileralure, if a slale can be lransferred lo lhe zero state O, the slate is said to be controllable.
If a state can be reached frarn O, lhe state is said to be reachable. Our definition does not rnake
this distinction to simplify the subsequent presentatiol1. Furthermorc. lhe cquation E is said lo
be, in the literature, completely controllable. For conciseness, the adverb "cornpletely" is dropped
in this book. u(t) = ­

will transfer X o to the

In In
1 Y
+ ~.
11 rv IF;::

In In 1.11:
.
Figure 5-3 An uncontrollable network. Figure 5-4 An uncont
:AL EQUATlONS CONTROLLABILITY OF LINEAR DYNAMICAL EQUATIONS 177

[lly; output equations precisely, the dynamical equation that describes the system-is not controllable
at any too

Example 2
(5-11 )
Consider the system shown in Figure 5-4. There are two state variables X¡ and
tor, and A and B are, Xz in the system. The input can transfer X¡ al' Xz to any value; however, it
continuous functions
cannot transfer x¡ and Xz to any values. For example, if X¡(t o) 0:=0, xz(to)=O,
III is an n-dimensional
then no matter what input is applied, x¡(t) is always equal to xz(t), for all t> too
Benee the equation that describes the system is not controllable at any too iI

The solution of the state equation E with x(to) = Xo is given by


ime to' if there exists a
md any x ¡ in ~, there
:he state x¡ at time ti.
x(t) = cP(t; t o, xo, u) = 1I>(t, to)x o + l'
[o
<1>(t, L)B(-r)U(L) dL

at time too
= 1I>(t, t o{ Xo + L lI>(t o, L)B( L)u(L) dLJ (5-12)

le of moving any state


where <1>(t, t o) = 'I'(t)'I' -¡(t o); '1' is a fundamental matrix of i = A(t)x and is
lt trajectory the state
nonsingular for all 1.
;traint imposed on the
¡ve sorne examples to
Theorem 5-4
The state equation E is controllable at time t o if and only if there exists a fmite
ti> t o such that the n rows of the n x p matrix function <1>(t o,')B(') are linearly
independent on [to, ti].
abk x ofthe system is·
lor aH t?: t o no matter·
Proof
letwork, and the input
the system-or, more Sufflciency: Ir the rows of <1>(t o, ·)B(-) are linearly independent on [t o, ti], from
Theorem 5-1 the n x n constant matrix

(5-13)
.te is said to be co/ltrolfable.
r definition does not make
js nonsingular. Given any x.(t o) = ~~o and any x j", '.ve claii1J that tb.e input
c. tl1e equation E is said lo
:rb "completely" is dropped
(5-14 )

will transfer X o to the state Xl at time ti. Indeed, by substituting (5-14) into

1F
+tzX
-=.t

Figure 5-4 An uncontrollable network.


178 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

(5-12), we obtain In the following we sl


A and B without solv
X(t¡) = <1>(t¡, tO){X O- [ <1>lt o• r)B(r)B*(r)<1>*(to, r) elr W-¡(t o, I¡) need sorne additional ;
times continuously di
. [x o -<I>(t o, t¡)X¡J} M¡(·), ... , by the equ:

= <1>(t¡, (0){Xo - WU o• I¡)W-1U o. (¡)[x o -<1>(to, (¡)x¡J}


= <1>(1 ¡, 10)<1>(t O· (I)x I
=x¡ with Mo(t) = I

Thus we condude that the equation E is controllable. Necessity: The proof is Observe that
by contradiction. Suppose E is controllable at to, but the rows of <I>(to,' )B(')
$(t o, t)B
are Iinearly dependent on [1 (l. (IJ for aH (1 > (o. Then there exists a nonzero,
constant IXIl row vector a. such that o
ot $(t o, t)B
a.<1>(t (l. t)B(t) = O for all 1 in [lo. IIJ (5-15 )

Let us choose x(t(l)~x(l = a.* Then Equation (5-12) beco mes

<1>U (l. tl)xU 1 ) = a.* + r" <1>(1 (l. r )B( r )u( r) dr (5-16 )
10 and, in general,
Premultiplying both sides of (5-16) by a.. we obtain Ok
ot k $(t o, c:
a.<1>(t(l.t l )x(tI) =a.a.* + r" a.<1>(to. r)B(r)u(r) dr (5-17 )
10 Theorem 5-5
By hypothesis, E is controHable at t o ; hence for any state-in particular,
Assume that the matri(
XI =O-there exists ullO,ld such that X(I¡)=O. Since a.<1>(10, t)B(t) =0, for allt
n - 1 times continuom
in [to, t¡J, Equation (5-17) reduces to
lable at t o if there exist
a.a.* =0
p[r
which, in tum, implies that ex = O. This is a contradiction. Q.E.D.
Proof
In the proof of this theorem, we also give in (5-14) an input u(t) that transfers
Define
x(t o) to x¡ at time tI' Because of the continuity assumption of A and B, the
input u in (5-14) is a continuous function oí t in [t o, tI]. ¿
If a linear dynamical equation is controllable, there are generally many dif­
ferent inputs u that can transfer x(t o) to x¡ at time t ¡, for the trajectory between Then, from (5-19) and 1

x(t o) and Xl is not specified. Among these possible inputs that achieve the
same mission, we may ask which input is optimal according to sorne criterion. [$(to, t¡)B(t¡) : O~l $(t
If the total energy

r Ilu(1)111 dt
ll

Since $(t o, t ¡) is nonsir


Jea
is used as a criterion, the input given in (5-14) wiII use the minimal energy in p[~
transferring x(t o) to Xl at time tI' Thisisestablished in Appendix C. implies
In order to apply Theorem 5-4, a fundamental matrix 'P or the state transi­
tion matrix <l>{t, r) of:x: = A(t)x has to be computed. As we mentioned earlier,
p [$(t o, tl)B(t¡) :
this is generally a difficult task, Hence, Theorem 5-4 is not readily applicable.
CAL EQUATIONS
CONTROLLABILITY OF LINEAR DYNAMICAL EQUATIONS 179

In the following we shall give a controllability criterion based on the matrices


A and B without solving the state equation. However, in order to do so, we
¡(t o, t ¡) need sorne additional assumptions on A and B. Assume that A and B are (n - 1)
times continuously differentiable. Define a sequence of n x D matrices M o('),
M¡(·), ... , by the equation
. [X O- C1>(tO, t ¡}X¡J}
d
M k+ ¡(t) = -A(t)Mk(t) + dt Mk(t) k = O, 1,2, ... , n - 1 (S-18a)

with Mo(t)=B(t) (S-18b)

lecessity: The proof is Observe that


:he rows of C1>(to,' )B(') <I>(to, t)B(t) = <I>(to, t)Mo(t) (S-19a)
here exists a nonzero,
:t <I>(to, t)B(t) = 'P(to) {[:t '11 - ¡(t)J B(t) +'11 - ¡(t) :t B(t)}
(S-1S)

;omes = 'P(to)'P-¡(t) [-A(t)B(t) + :t B(t)J

= <I>(to, t)M¡ (t) (S-19b)

) dr (S-16)
and, in general,
ak
k=O,I,2, .. ,n-l
at k <I>(to, t)B(t) = <I>(to, t)Mk(t) (S-19c)

(r) dr (5-17 )
Theorem 5-5
state-in particular,
Assume that the matrices A(') and B(-) in the n-dimensional state equation E are
I(tO, t)B(t) = O, for aH t
n -1 times continuously differentiable. Then the state equation E is control­
lable at t o if there exists a finite t¡ > t o such that
(S-20)
n. Q.E.D.
Proof
nput u(t) that transfers
Define
lption of A and B, the
a ""\lo, ¡',)1lJl(
- ,-,,1. "'\
j;]) LJ
L a t"', (e.0' " )1R1-
~'a'--- 0' ,¡ M\L¡i
'
at 1=1, t¡
lre generally many dif­
the trajectory between Then, from (5-19) and using (2-2), we have:
lputS that achieve the
jing to sorne criterion.
a an -¡
[<I>(t o, t¡)B(t¡) : 'Jt;<I>(t o, t¡)B(t¡) : ... : at1-¡ <I>(t o, t¡)B(t¡)
J
= <I>(t o, t¡)[Mo(t¡} : M¡(t¡): ., . : Mn-1(t¡)] (S-21 )
Since <I>(to, ti) is nonsingular, the assumption

the minimal :energy in p[Mo(t¡} : M¡(t¡) : ... : Mn-¡(t 1 )] =n


Appendix c.' implies
( '11 orthe state ttansi~
we mentioned earlier,
not readily applicable.
180 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

It follows from Theorem 5-2 that the rows of $(t o, .)B(-) are linearly independent controllability. The (
on [t o, t 1 ] for any t 1 > tá. Thus, from Theorem 5-4, we conclude that the state obtained by slight mo
equation E is controllable. Q.E.D. matrices A and B in the
the following theorem.
As in Theorem 5-2, the condition of Theorem 5-5 is sufficient but not neces­
sary for the controllability of a state equation.
Theorem 5-6
Example 3 If the matrices A and E
Consider equation E is different
for any fixed to in ( - ex

Xl]
~z
[X3 =
10J° [1] +[0]
[t° Xz 1 u (5-22) p[Mol

°° t
Z
X3 1 If the matrix A is an
From (5-18), we have tion matrix $(t o, .) of x
of two analytic functior
implies that $(t o, -)B(-)
is that a set of analytic I
if the set of analytic fm
matter how small) of( ­
from Theorems 5-3 an
A and B is controllable
t in ( - 00, 00).

Definítion 5-3
The linear dynamical
in ( - 00, 00) if and onl)
Since the matrix [Mo(t) : M 1 (t) : Mz(t)] has rank 3 for all t -/=0, the dynamical
equation is controllable at every t. I

Differential eontrollability. instantaneous eontrollability, where the M/s are as d


and uniform eontrollability
The remainder of this subsection will be devoted to three distinct types of Gon­ Ir a dynamical equa
trollability. This material may be omitted with no loss of continuity. the states can be achie'
consists of b-functions :
Definition 5-2 that instantaneous co
most important implic:
A state equation E is said to be differentially (completely) controllable at time t o of a single input, the
if, for any state x(to) in the state space L and any state Xl in L, there exists an equivalence transforma
input u that will transfer x(t o) to the state Xl in an arbitrarily small interval of form equivalent dynaI
~~ I controllable dynamical

If every state in L can be transferred to any other state in a finite time (no
matter how long), the state eqúation is said to be controllable. If this can be
5 In lhe engineering lileraluro
achieved in an arbitrarily small interval of time, then the state equation is said
firsl used in Reference S6 lo
to be differentially controllable. Clearly, differential controllability implies we adopt lhe terminology "
MICAL EQUATIONS
CONTROLLABILJTY OF LINEAR DYNAMICAL EQUATlONS 181

are linearly independent controllability. The condition for differential controllability can be easily
e conclude that the state obtained by slight modifications of Theorems 5-4 and 5-5. However, if the
Q.E.D. matrices A and B in the state equation E are analytic on ( - 00,00), then we have
the following theorem.
sufficient but not neces­

Theorem 5-6
Ifthe matrices A and B are analytic on ( - 00, CJ:)), then the n-dimensional state
equation E is differentially controllable at every t in ( - 00, 00) if and only if,
for any fixed t o in ( - CJ:), (0),

(5-22) 11

If the matrix A is analytic on (- 00, 00), it can be shown that the state transi­
tion matrix <I>(t o, .) of x= A(t)x is also analytic on (- 00, 00). Since the product
oftwo analytic functions is an analytic fundion, the assumption ofTheorem 5-6
implies that <I>(t o, ')BO is an analytic function. An implication of Theorem 5-3
is that a set of analytic functions is linearly independent on ( - 00,(0) if and only
if the set of analytic functions is linearly independent on every subinterval (no
matter how smaH) of ( - 00, (0). With this fact, Theorem 5-6 foHows immediately
from Theorems 5-3 and 5-4. Consequently, if a state equation with analytic
A and Bis controllable at any point at all, it is differentially controllable at every
tin(-oo,oo).

Definition 5-3
The linear dynamical equation E is said to be instantaneously controllable 5
in (- 00, (0) if and only if
- aH t 1=0, the dynamical
I for aH t in (-00, (0)

ilitv, where the M¡'s are as defined in Equation (5-18). 11

~ee distinct types of caD­ U a dynamical equation is instantaneously controHable, then the transÍer oí
s of continuity. the states can be achieved instantaneously at any time by using an input that
consists of b-functions and their derivatives up to an order of n - 1. It is clear
that instantaneous controllability implies differential controllability. The
most important implication of instantaneous controllability is that in the case
y) controllable at time t o of a single input, the matrix [Mo(t) MI(t) ... Mn-l(t)] qualifies as an
: Xl in L, there exists an equivalence transformation (Definition 4-5). Consequently, many canonical­
itrarily small interval of form equivalent dynamical equations can be obtained for instantaneously
I
controHable dynamical equations. See References 10,99, 103, and 110.
state in a finite time (no
itrollable. If this can be
he state equation is said s In the engineering literature, it is called uni(orm controllability. However, this terminology was
first used in Reference 56 to define a difTerent kind of eontrollability (see Definition 5-4); hence
IcontroHability implies we adopt the terminology"instantaneous eontrollability."
182 CONTROLLABILlTY AND OBSERVABILlTY OF LINEAR DYNAMICAL EQUATIONS

Definition 5-4
The dynamical equation E is said to be uniformly controllable if and only if there
exist a positive (J e and positive CXi that depend on (J e such that
0< cx¡((Je)I ,:::;;W(t, t +(Je) ~cxz(aJI
and 0< cx 3 ((Je)I ~<l>(t +(J" t)W(t, t +(Je)<l>*(t +(J" t)':::;;cx 4 ((Je)I -3

for all t in ( - 00, (0), where <l> is the state transition matrix and W is as defined
in Equation (5-13). I

Figure 5-5 The function


By A> B, we mean that the matrix (A - B) is a positive definite matrix (see
Section 8-5). Uniform controllability ensures that the transfer of the states
can be achieved in the time interval (J e' The concept of uniform controllability
achieved in the time intt
is needed in the stability study of optimal control systems. See References 56
will not be arbitrarily lar
and 102.
W - ¡]. In optimal cont
Instantaneous controllability and uniform controllability both imply con­
sometimes required to e
trollability. However, instantaneous controllability neither implies nor is
implied by uniform controllability.
Time-invariant case.
n-dimensional linear tim
Example 4
Consider the one-dimensionallinear dynamical equation
where x is the n x 1 statt:
x=e-1rlu and n x p real constant .
equations. the time inte
Since p(Mo(t)) = p(e- Itl ) = 1 for al1 t, the dyilamical equation is instantaneously
that is, [O, 00). .
controllable at every t. However, the dynamical equation is not uniformly
The condition for a t
control1able because there exists no CX¡ that depends on (Je but not on t such that
that there exists a fini te ti
,+ U
e
in t. on [t o, tI]. In the
W(t,t+(Je)= 1 e-2Idt=0.5e-21(I-e-2Ue»cx¡((Je)
J. As discussed in Chapter
terms of form t k (!/.'; hencl
for all t > O. I
sequently, if its rows ar
Example 5 independent on [too t lJ f
time-invariant state eqm
Consider the one-dimensional dynamicai equation and the transfer of any st
x= b(t)u time intervaL Hence th,
trol1ability study of linea
with b(t)defined as in Figure 5-5. The dynamical equation is not instantaneous­
Iy control1able in the interval (-1, 1). However, it is uniformly control1able Theorem 5-7
in ( - 00, oo). This can be easily verified by choosing (J e = 5. I The n-dimensional linear
if and only if any of the fe
A remark is in order concerning controllability, differential control1ability,
and uniform controllability. lf a dynamical equation is differentially con­ 1. Al1 rows of e-AIB (an
trollable, a state can be transferred to any other state in an arbitrarily smal1 [0, (0) over e, the fiel,
interval of time. However, the magnitude of the input may become very large; 1'. All rows of (sI - A)-I
in the extreme case, a delta-function input is required. lfadynamical equation
is merely controllable, the transfer of the states may take a very long interval of 6 Although aH the entries of A al
time. However, if it is uniformly controllable, the transfer of the states can be of the field of complex numbe:
nCAL EQUATlONS CONTROLLABILITY OF LINEAR DYNAMICAL EQUATIONS 183

b (1)

'lable if and only if there


that

t)saiaJI -3 -2 -1 o 2 3

trix and W is as defined


I

Figure 5-5 The functian b(t).


.tive definite matrix (see
le transfer of the states
.uniform controllability achieved in the time interval a e ; moreover, the magnitude of the control input
:ms. See References 56 will not be arbitrarily large [see (5-14) and note that the input is proportional to
W-l]. In optimal control theory, the condition of uniform controllability is
ability both imply con­ sometimes required to ensure the stability of an optimal control system.
neither implies nor is
Time-invariant case. In this subsection, we study the controllability oC the
n-dimensional linear time-invariant state e<tuation
FE: x= Ax + Bu (5-23)
)n
where x is the n x 1 state vector, u is the p x 1 input vector; A and B are n x n
and n x p real constant matrices, respectively. For time-invariant dynamical
equations, the time interval of interest is from the present time to infinity;
ation is instantaneously that is, [O, (0).
lation is not uniformly The condition for a time-varying state equation to be controllable at t o is
re but not on t such that
that there exists a finite t 1 such that aH rows of cl>(t o' t)B(t) are linearly independent,
in t, on [to, ti]. In the time-invariant case, we have eJ>lt o , t)B(t)=eA(to-l)B.
As discussed in Chapter 4, al! elements of eA(to-tlB are linear combinations of
terms of form tke Á1 ; hence they are analytic on [O, 00) (see Appendix B). Con­
I
sequently, if its rows are linearly independent on [O, 00), they are Iinearly
independent on [t o' ti] for any t o and any ti> t o· 1n other words, if a linear
time-invariant state equation is controllable, it is controllable at every t o 2:0
and the transfer of any state to any other state can be acÍlieved in any nonzero
time interval. Hence the reference of t o and ti is often dropped in the con­
trollability study of linear time-invariant state equations.
ion is not instantaneous­
; uniformly controllable Theorem 5-7
re = 5. I The n-dimensional linear time-invariant state equation in (5-23) is control1able
if and only if any of the fol1owing equivalent conditions is satisfied:
Jerential controllability,
m is differentially con­
1. All rows of e-AtB (and consequently of eA/B) are linearly independent on
[O, (0) over ic, the field of complex numbers. 6
: in an arbitrarily small
may become very large; 1'. All rows of (sI - At 1 B are Iinearly independent over iC.
Ir a dynamical equation
~e a very long interval of 6 Althollgh all the enfríes of A and B are real ímmbers. we have agreed to consider them as elements
lsfer of the states can be of the field of complex nllmbers.
184 CONTROLLABILITY ANO OBSERVABlLITY Of LINEAR OYNAMICAL EQUATlONS

2. The controllability grammian Example 6

W el ~ t eA'BB*é *, dr
Consider the inverted r
equation is developed il
2g(M +m)/(2M +m)l=
is nonsingular for any t> 0. 7
(3-42) becomes
3. The n X (np) controllability matrix
1
U~[B: AB: A 2 B:"': A"-lBJ (5-24 )
O
has rank n. O
4. For every eigenvalue A of A (and consequently for every ), in C), the n x (n + p)
O
complex matrix [Al -A : BJ has rank n. s
We compute
Proof
The equivalence of statements 1 and 2 follows directly from Theorems 5-1 and
5-4. Since the entries of e- AIB are analytic functions, Theorem 5-3 implies that u= [B Al
the rows of e - AIB are linearly independent on [0, (0) if and only if
p[e-AIB: -e-AIAB:"': (-l)"-le-AIA"-lB: "'J=n
This matrix can be read
for any t in [0, (0). Let t = O; then the equation reduces to Thus, if X3 = eis differer
to bring it back to zero.
p[B: - AB: . ..; ( - 1)" - lA" - 1 B: ( -1)" A"B : ...] = n
derivatives back to zer
From the Cayley-Hamilton theorem, we know that Am with m ¿ n can be written balancing a broom on (
as a linear combination of 1, A, ... , A"-l; hence the columns of AmB with
m "2: n are linearly dependent on the columns of B, Aa, ... ,A"-1B. Conse­
quently, Example 7
p[B: -AB: ... : (-l)"-IA"-IB:' "J=p[B: -AB:"': (-I)"-1A"-lBJ Consider the platform s
of the platform is zero ;
Since changing the sign will not change the linear independence, we conclude
systems. The spring c(
that the rows of e-AIB are linearly independent if and only if p[B : AB : ...
coefficients are assumed
: A n -1 BJ = n. This proves the equivalence of statements 1 and 3. In the
X2 +X2 =u, or
foregoing argument we also proved that the rows of e - AIB are linearly inde­
pendent if and only if the rows of eAIB are linearly independent on [0, (0) over
the field of complex numbers. Next we show the equivalence of statem.eEt:::
and 1'. Taking the Laplace transform of eAIB, we have
This is the state-variabh
5t'[e AI BJ = (si - A)-1 B Now if the initial di~
platform will oscillate,
Since the Laplace transform is a one-to-one linear operator, if the rows of
platform to come to re~
eAIB are linearly independent on [0, (0) over the field of complex numbers, so
are the rows of (si - A) -1 B, and vice versa. and X2(O) = -1, is it po~
.seconds? The answer d,
The proof of statement 4 will be postponed to Section 5-5 (page 206). Q.E. O.
applied to thetwo sprin
For the state equatic

p
7 The matrix is in raet positive definite.· See Problem E-II.
s This eondition implies that (sI - Al and B are left eoprime. See Appendix G. hence the state equation
~AL EQUATlONS CONTROLLABILlTY OF LINEAR DYNAMICAL EQUATIONS 185

Example 6
Consider the inverted pendulum system studied in Figure 3-15. Its dynamical
equation is developed in (3-42). For convenience, we assume 2mg/(2M +m) = 1,
2g(M +m)/(2M +m)l = 5, 2/(2M +m) = 1, and 1/(2M +m)l = 2. Then Equation
(3-42) becomes

x~l~
1 O

!}+[j}
(5-24 )
O -1
y=[1 O O OJx (5-25)
O O
O 5
We compute

am Theorems 5-1 and


~orem 5-3 implies that
nd only if
U=[B AB A2B A'B] ~ [ !
-2
1
O
-2
O
2
O
O -10 -l~l
l:-"J=n This matrix can be readily shown to have rank 4. Hence (5-25) is controllable.
to Thus, if X3 = e is different from zero by a small amount, a control u can be found
to bring it back to zero. In fact, a control exists to bring XI = y, X3 = e, and their
:'-'J=n derivatives back to zero. This is certainly consistent with our experience of
:h m ¿ n can be written balancing a broom on our hand. Ii
::olumns of AmB with
, ... ,An-IB. Conse­
Example 7
Consider the platform system shown in Figure 5-2. lt is assumed that the mass
of the platform is zero and the force is equally divided among the two spring
endence, we conclude
systems. The spring constants are assumed to be 1, and the viscous friction
anly if p[B : AB. : ...
coefficients are assumed to be 2 and 1 as shown. Then we have XI +2xI = u and
nts 1 and 3. In the
X2 +X2 =u, or
AtB are linearly inde­
~ndent on [0, (0) over
(5-2~)
llence of statements 1
This is the state-variable description of the system.
Now if the initial displacements XI (O) and X2(0) are different from zero, the
platform will oscillate, and it will take, theoretically, an infinite time for the
erator, if the rows of
platform to come to rest. Now we pose the following problem: If Xl (O) = 10
complex numbers, so
and X2(0) = - 1, is it possible to apply a force to bring the platform to rest in 2
seconds? The answer does not seem to be obvious because the same force is
-5 (page206). Q.E.D.
applied to the two spring systems.
For the state equation in (5-26), we compute

p[B : ABJ = p [~.5=~.25J = 2


:ndix G. hence the state equation is controllable. Consequently, the displacements can
186 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

be brought to zero in 2 seconds by applying a proper input u. Using Equations

r
(5-13) and (5-14), we have and u 1 (t)=-[0.5

W(O, 2) = [(l0~5' °J[0.5J [0.5


(1' 1
e
1] [ O
O5
. ' 0]
(1' dr =
[1.6
6.33
6.33J
27
for t in [0, 2]. If a for
at t = 2. The behaviol
Hnes in Figure 5-6.
u (1)

In Figure 5-6 we a]
40
xl(O)=lO and X2(O) =
in transferringx(O) to 2
of the input. If no re:
x(O) to zero in an arbi
20
the input may become
u 2/ imposed, we might nc
3 ..... /4 interval of time. For
o
might not be able to tr
--- 1, seconds

Example 8
-20
Consider again the plal
the viscous friction coe
-40 a11 equal to 1. Then t

xl (1)

10 Clearly we have
............

........

............

1 ........... 2
...------- ....
3 4

o
2.5 1, seconds
and the state equation
an input to transfer x(C
X (1) xz(O), no input can tfar
2

* Controllabilíty índicE
10 Let A and B be n x

2 3 4
ol---l--+--I----+-+----t--+-----:,f--------.­
-5 -- ,.,ti' 1, seconds It consists of k + 1 bloc
The matrix U ~ Un - 1
thenU n_ 1 has a rank 01
Note that there are a to
ways to choose these ni
ing the most natural anc
Figure 5-6 Behavior of x¡Ü) and X2(t) and the waveformof u.. Let b , ; = 1, 2, ... , p, be
;AL EQUATlONS CONTROLLABILlTY Of LINEAR DYNAMICAL EQUATlONS 187

Jt u. Using Equations
and ul(t) =- [0.5 l][e~'St ~] W-I(O, 2) [ ~~J = -44.1eo. St +20.7e t


1.6
- [ 6,33 27
6.33J for t in [0,2]. If a force of the form U I is applied, the platform will come to res t
at t = 2. The behavior of Xl' X2 and of the input u I are plotted by using solid
lines in Figure 5-6. •

In Figure 5-6 we also plot by using dotted lines the input U2(t) that transfers
XI(O) = 10 and X2(0) = -1 to zero in 4 seconds. We see from Figure 5-6 that,
in transferringx(O) to zero, the smaller the time interval the larger the magnitude
of the input. If no restriction is imposed on the input u, then we can transfer
x(O) to zero in an arbitrarily small interval of time~ however, the magnitude of
the input may become very large. If sorne restriction on the magnitude of u is
imposed, we might not be able to transfer x(O) to zero in an arbitrarily small
interval of time. For example, if we require iu(t)\.::; 5 in Example 5, then we
might not be able to transfer x(O) to zero in less than 4 seconds.

Example 8
Consider again the platform system shown in Figure 5-2. Now it is assumed that
the viscous friction coefficient and the spring constant ofboth spring systems are
all equal to 1. Then the state-variable description of the platform system is

Clearly we have

-lJ
-1
=1<2

and the state equation is not controllable. If XI(O) = X2(0), it is possible to find
an input to transfer x(O) to the zero state in a finite time. However, if X1 (O) 1=
X2(0), no input can transfer both Xl (O) and X2(O) to zeíO in a finite time. íill

*Controllability índices
Let A and B be n x n and n x p constant matrices. Define

U k = [B :. AB :• ... :• AkB] k =0,1,2, ... (5-27)

It consists .of k + 1 block columns of the form AiB and is of order n x (k + l)p.
The niatrix U~ U _¡ is the controllability matrix. If {A, B} is controllable,
II

then Un_¡has a rank of n and consequently has n linearly independent columns.


Note that there are a total of np colurrins.in U ¡; hence there are many possible
II
_

ways to choose these n linearly ind.ependeiltcolumns. We discuss in the follow­


ing the most natur~1 and also the most important way of choosingthese columns.
Let b . i = 1, 2, ... ,p, be the ith column ofB. Then the matrix Uk can be written
188 CONTROLLABlLITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

explicitly as Hence, aH columns


columns. Consequer
U k = [b¡ bz
Uk must increase by
,

ro increase. Since the r


to add at most n - p
(5-28)
fJ. -l'::;n - p or ¡J.'::;/
yields the upper boun
/"k (no. of dependent columns) In order to have
rows. Hence, we neec
Now we search linearly independent columns of U k in order from left to right; (5-31).
that is, if a column can be written as a linear combination of its left-hand-side The degree ñ of th
columns, the column is linearly dependent; otherwise, it is lintlarly independent. min (ñ, n - p + 1) in (
This process of searching can be carried out by using, for example, the column­ Theorem 5-7, we havt
searching algorithm discussed in Appendix A. Let ri be the number of linearly
dependent columns in AiB, for i = O, 1, ... ,k. If B has a full column rank, then
ro = O. We note that if a column, say Ab z, is linearly dependenton its left-hand­ Corollary 5- 7
side columns, then all Aib z, with j = 2, 3, ... , will be linearly dependent on its The state equation FJ
left-hand side columns. lndeed, if
Ab z = IX¡ b¡ +lXzb z + ... +lXpb p +lX p+¡Ab¡
then AZb z =1X¡Ab¡ +lXzAb z + ... +lXpAb p +lXp+¡Azb¡ where p is the rank (
singular.
We see that A zb z is a linear combination of its left-hand-side columns. Pro­
ceeding similarly, we can show that Aib z, for j = 3, 4, ... ,are linearly dependent
on its left-hand-side columns. Because of this property, we have Example 9
O'::;ro'::;r¡'::;rz'::;'" .::;p Consider the sateHite
equation is developed
Since there are at most n linearly independent columns in U oo , there exists an
the control of Xi' i = 1
integer fJ. such that are entirely independel
O.::;ro'::;r¡'::;··· '::;r¡L_¡ <P (5-29 a)
(5-29b)
of (3-51):
and r¡L=r¡L+¡="'=P

Equivalently, fJ. is the integer such that


pU o < pUl < ... < pU/._¡ = pUl' = pUl' +1 = ... (5-30)

In other words, the rank of Uk increases monotonically until k reaches fJ. - 1;


thereafter, all P col umns of AiB will be linearly dependent of their left-hand-side
columns. Hence the controllability of {A, B} can be checked from U¡L-¡ and fJ. y=
is called the controllability indexo We claim that the controllability index
satisfies the inequalities where we have assum<
n
- '::;fJ. .::;min(ñ, n - p+ 1) (5-31 ) bars. The rank of B is
P the matrix
where ñ is the degree of the minimal polynomial of A and Ji is the rank of B.
From the definition of minimal polynomial, we have
A ñ =1X¡A ñ-¡ +1X2Añ-Z + ... +lXñ l [B: AB

for sorne lX i , which implies


AñB =IX¡A ñ-¡B +IX Z A ñ - zB + ... +IX ñB has rank 4. This is thc
LINEAR INDEPENDENCE OF TIME FUNCfIONS 189
ICAL EQUATIONS

Hence, all columns of AñB are linearly dependent on their left-hand-side


columns. Consequently, we have fl ~ ñ. From (5-30), we see that the rank of
Uk must increase by at least 1 as k inereased by 1; otherwise, it will cease to
inerease. Sinee the rank of U", is at most n, if the rank of B is p, it is suffieient
to add at most n - p number of AjB. Hence, we eonclude from (5-30) that
fl- 1 ~ n - p or fl::; n - p + 1. The eombination of fl ~ ñ and fl ~ n - p + 1
(5-28)
yields the upper bound in (5-31).
In order to have rank n, the matrix U 1"-1 must ha ve more columns than
rows. Hence, we need flP ~ n or fl ~ nI p. This establishes the lower bound in
Irder from left to right; (5-31).
on of its left-hand-side The degree ñ of the minimal polynomial of A is not easy to compute; hence
is lin6arly independent. min (ñ, n - p + 1) in (5-31) is often replaeed by n - p + 1. From (5-30) and
r example, the column­ Theorem 5-7, we have immediately the following eorollary.
the number of linearly
full column rank, then
endenton its left-hand­ Corollary 5- 7
learly dependent on its The state equation FE is controllable if and only if the n x (n - p + l)p matrix
Un - p = [B AB .. , An-PB]
\.b 1
xp + ¡A 2 b¡ where p is the rank of B, has rank n, or the n x n matrix lJ n _ pU~_ p is non­
singular. •
ld-side columns. Pro­
, are linearl y dependent
we have Example 9
Consider the satellite system studied in Figure 3-17. Its linearized dynamical
equation is developed in (3-51). As can be seen from the dotted lines in (3-51),
; in U"" there exists an the control of Xi> i = 1,2, 3, 4, by ü¡ and Ü2 and the control of X5 and X6 by Ü3
are entirely independent; hence, we may consider only the following subequation
(5-29a)
of(3-51):
(S-29b)
o
~l I~o ~lo
1
. 3 ro o o
'1"+1 ='" (5-30) x= o o o 1 x+ u (S-32a)

y until k reaches fl- 1; LO -2 o OJ LO 1j


lt of their
left-hand-side o
y=[~ ~J x
O
ecked froro U 1 and fl (S-32b)
¡J. -

e controllability index
o 1

where we have assumed, for simplicity, w = m = ro = 1 and dropped the over­


(5-31 ) bars. The rank of B is 2; hence, the state equation is controllable if and only if
the matrix
and p is the rank of B. O,

~l
O :, O

~l!
1
O O 2 : -1
1
[B' AB: A'B] O,, O 1 : -2
. O 1 : -2 O' O -4
has rank 4. This is the case; hence, the state eqilation in (5-32) is eontrollable.
190 CONTROLLABILlTY AND OBSERVABILlTY OF LINEAR DYNAMICAL EQUATIONS

For this problem, we can readily show the controllability inc


pU o =2<pU¡=4=pU 2 =pU 3 = ... 1, ... , /1}.

hence, the controllability index of (5-32a) is 2. I Theorem 5-8

We study more the controllability matrix U. We assume that the linearly The set of the controll~
independent columns of U in order from left to right have been found. We transformation and ar
now rearrange these independent columns as
Proof
b¡, Ab¡, ... , Al'l -¡ b¡, b 2, Ab 2, ... , AI'2-¡ b 2, ... , b p, Ab p, ... , Al'p-¡ b p
Define
The integer /1¡ is the number of linearly independent columns associated with b i
in the set, or is the length of the chain associated with b¡. Clearly we have
where A=PAP-¡, Ü.
/1 = max{/1¡, /12, ... , /1 p}
easily verified that
and /1¡ +/12 + ... + /1p::;; n. The equality holds if {A, B} is controllable. The
set {/1¡, /12, ... ,/1 p} will be called the controllability indices of {A, B}.
Now we shall establish the relationship between the controllability indices which implies
and the r¡'s defined in (5-28). In order to visualize the relationship, we use an
example. We assume p = 4, /1¡ = 3, /12 = 1, /13 = 5, and /14 = 3. These indepen­
dent columns are arranged in a crate diagram as shown in Figure 5-7. The Hence, the r¡ defined
(i, j)th cell represents the column Ai - ¡bj- A column which is linearly indepen­ indices are invariant Ul
dent of its left-hand-side columns in (5-28) is denoted by "x"; otherwise denoted The rearrangement
by "O". The search of linearly independent columns in (5-28) from left to right
is equivalent to the search from left to right in each row and then to the next
row in Figure 5-7. Hence, the number of zeros in the ith row of Figure 5-7 is where M is a p x p eler
equal to ri-¡ as shown. From the crate diagram, we can deduce that r¡ is equal forward to verify that
to the number of {bk , k = 1, 2, ... ,p} with controllability indices equal to or - f:, - .
smaller than i. Hence, we conclude that Uk=[B: 1

r¡ - r¡_¡ = no. of {bk , k = 1, 2, ... , p} with controllability index i (5-33) where diag {M, M, ...
singular. Hence, we h¡
with l' _ d~O. For example, r¡ - ro = 1, and b 2 has controllability index 1;
1'2 - r¡ =0 and no b¡ has controllability index 2; 1'3 - 1'2 = 2, and b¡ and b4 have
controllability index 3; 1'5 -1'4 = 1, and b 3 has controllability index 5. Hence, Consequently. we cane
the ordering ol" the eoll
b¡ b2 b3 b4
Now we discuss a (
1 x x x x ro =0 umns in U = [B AB
A x O x x r¡ = 1
A2 X O X X 1'2 = 1
A3 O O X O 1'3 =3
A4 O O X O 1'4= 3 and then search its linl
AS O O O O In terms af the crate di
1'5 =4

are searched in order fr,


/1i 3 1 5 3

column and so forth. 1


Figure 5-7 Crate diagram of Ai-1bj .
:AL EQUATIONS LINEAR INDEPENDENCE OF TIME FUNCTIONS 191

the controllability indices of {A, B} are uniquely determinable from tri, i =0,
1, ... , Ji}.

I Theorem 5-8
The set ofthe controllability indices of {A, B} is invariant under any equiva1ence
ume that the linearly
transformation and any ordering of the columns of B.
ave been faund. We
Proof
Ab p, ... , AJ.Lp-l b p
Define
nns associated with b¡
]early we have
Ük=[B: AB:"': AkBJ
where A= PAP- , l
B= PB and P is any nonsingular matrix. Then it can be
easily verified that
is controllable. The
for k = O, 1, 2, ...
es of {A, B}.
controllability indices which implies
~lationship, we use an
for k =0,1,2, ...
. = 3. These indepen­
11 in Figure 5-7. The Hence, the r¡ defined in (5-28) and, consequently, the set of controllability
;;h is linearly indepen­ indices are invariant under any equivalence transformation.
x"; otherwise denoted The rearrangement of the columns of B can be represented by
;-28) from left to right
B=BM
, and then to the next
h row of Figure 5-7 is where M is a p x p elementary matrix and is nonsingular. Again it is straight­
deduce that ri is equal forward to verify that
ty indices equal to or
iJd~~ [B
: AB : ... : AkBJ = U k diag{M, M, ... , M}
where diag {M, M, ... , M} consists of k + 1 number of M, and is clearly non­
ability index i (5-33)
singular. Hence, we have
mtrollability index 1;
for k =0, 1,2, ...
2, and b¡ and b 4 have
)i1ity index 5. Hence, Consequently, we conclude that the controllability indices are independent of
the ordering of the columns of B. Q.E.D.

Now we discuss a different method of searching linearly independent col­


umns in U = [B AB ... An-lBJ. We first rearrange the columns of U as
b l , Ab l , AZb¡, ... , An-¡b¡; b z, Ab z, ... , An-lb z ; ... ; b p, Ab p, ... , An-lb p
(5-34)

and then search its linearly independent columns in order from left to right.
In terms of the erate diagram in Figure 5-7, the linearly independent columns
are searched in order from top to bottom in the firstcolumn, then in the second
column and so forth. Let .
b l' Ab b ... , Aii,-lb l'. b 2' Ab 2, ... , Aii2- l b"
2, ... , b P' Ab P' ... , Aiip-lb p
...:.=---_._-_.•_­----
-----~ -----=-------=-:..:::...-.::--:..:'--- - - - - - - - - - - ' -.." ' - - - , - - - - - -- - - ----- _. -------------------,~=~

192 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

be the resulting linearly independent columns. If {A, B} is controllable, we have Consider the n-dirr
ji ¡ + ilz + . .. + il p = n. The lengths of these chains are {il¡, ilz,···, il p }·
Unlike the controllability indices, these lengths of chains depend highly on the
ordering of {b¡, i = 1,2, ... , p}.
where A, R, e, and E a
Example 10 continuous functions (
Consider the state equation in (5-32). If we search linearly independent columns
from left to right in Definition 5-5
b¡, Ab¡, AZb¡, A 3 b¡; b z, Ab z, AZbz, A3 b z The dynamical equati
if there exists a finite tI
the resulting linearly independent columns are of the input U[lo,l,] and
b¡, Ab¡, AZb¡, b z to determine the state
unobservable at too
Its lengths are {3, 1}. If we search from left to right in
Example 1
b z, Ab z, AZb z, A 3 b z ; b¡, Ab¡, AZb¡, A3 b¡
Consider the network ~
the resulting linearly independent columns are the initial voltage acros
b z, Ab z, AZb z, A3 b z the output is identicall:
cally zero), but we are n
Its lengths are {4, O}. The lengths are indeed different for different ordering of hence the system, or m
b¡ and b z. The controllability indices of (5-32) can be computed as {2, 2} and system, is not observab
are independent of the ordering of b¡ and b z. I
Example 2
Let il = max{il¡, ilz, ... ,ilp }. It is clear that il can never be smaller than /1, Consider the network sl
the controllability indexo Since /1 = max{/1¡, /1z, ... , /1p}, we may conclude reduces to the one sho'
that /1 is the smallest possible maximum length of chains obtainable in any
search of linearly independent columns of U. This controllability index will
play an important role in the design of state feedback in Chapter 7 and the design
of compensators in Chapter 9. In In
lF
+
urv +

5-4 Observability of linear Dynamicai EqiUia-á:imu$

Time-varying case. The concept of observability is dual to that of control­


lability. Roughly speaking, controllability studies the possibility of steering Figure 5-8 An unobserv
the state from the input; observability studies the possibility of estimating the
lH
state from the output. If a dynamical equation is controllable, all the modes of
the equation can be excited from the input; if a dynamical equation is observ­
able, all the modes of the equation can be observed at the output. These two
concepts are defined under the assumption that we have the complete knowledge
of a dynamical equation; that is, the matrices A, B, e, and E are known before­
hand. Hence, the problem of observability is different from the problem of
realization or identification. The problem of identification is a problem of
estimating the matrices A, B, e, and E from the information collected at the (a)

input and output terminals. Figure 5-9 An unobserv


ICAL EQUATIONS OBSERVABILlTY OF LINEAR DYNAMICAL EQUATIONS 193

is controllable, we have Consider the n-dimensional linear dynamical equation


are {P1' ¡¡z, ... , ¡lp}.
IS depend highly on the
E: x= A(t)x(t) + B(t)u(t) (S-3Sa)
y = C(t)x(t) + E(t)u(t) (S-3Sb)

where A, B, e, and E are n x n, n x p, q x n, and q x p matrices whose entries are


continuous functions of t defined over (- 00, 00).
ly independentcolumns
Definition 5-5
The dynamical equation E is said to be (completely state) observable at t o
if there exists a finite t 1 > t o such that for any state X o at time to, the knowledge
of the input ulto,tl! and the output Ylto,tI! over the time interval [to, t 1J suffices
to determine the state xo. Otherwise, the dynamical equation E is said to be
unobservable at to. 111

Example 1
Consider the network shown in Figure 5-8. lfthe input is zero, no matter what
the initial voltage across the capacitor is, in view ofthe symmetry ofthe network,
the output is identically zero. We know the input and output (both are identi­
cally zero), but we are not able to determine the initial condition ofthe capacitor;
for different ordering of hence the system, or more precisely, the dynamical equation that describes the
computed as {2, 2} and system, is not observable at any to.
111
Example 2
never be smaller than /1, Consider the network shown in Figure 5-9(a). Ir no input is applied, the network
, /1p}, we may conclude reduces to the one shown in Figure 5-9(b). Clearly the response to the initial
lains obtainable in any
ontrollability index will
::::hapter 7 and the design

:luations

, dual to that of control­


Figure S-S An unobservable network.
e possibility of steering
;ibility of estimating the IH IH
'ollable, all the modes of
lical equation is observ­
the output. These two
the complete knowledge In.
I y In.
I

y
md E are known before­
nt from the problem of
ication is a problem of
' -_
F
_---4-----_
1
1 L ­_ _ ~------JIF 1
(a)
~mation collected at the (b)

Figure S-9 An unobservable network.


-

-~' ~~--~-~. __ - _,o

.~--~_.~- --~~-~--~--~~-_ .. __._-------­


• __ •

194 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

current in the inductor can never appear at the output terminal. Therefore, Prove by contradictio
there is no way of determining the initial current in the inductor from the input ti> t o such that the co
and the output terminals. Hence the system or its dynamical equation is not Then there exists an n
observable at any too I

The response of the dynamical equation (5-35) is given by Let us choose x(to) = cx.

y(t) = C(t)<1>(t. to)x(to) +C(t) L l1>(t, r)B(r)u(r) dr + E(t)u(t) (5-36)


Hence the initial statex(
y(t

where l1>(t. r) is the state transition matrix of X = A(t)x. In the study of observ­ tion that E is observa!
ability. the output y and the input u are assumed to be known, the initial state ti> t o such that the col
x(t o) is the only unknown: hence (5-36) can be written as
y(t) = C(t)<1>{t. t o)x(t o) (5-37) We see from this th~
tion depends only on e
where y(t)~ y{t) - C(t) Ir l1>{t. r)B(r)u(r) dr - E(t)u(t) (5-38) can also be deduced [re
lO
servability study, it is s
is a known function. Consequently. the observability problem is a problem of x = A(t)x, Y = C(t)x.
determining x(to) in (5-37) with the knowledge of y, C, and l1>(t. to). Note that The controllability
the estimated state x(t o) is the state not at time t, but at time too However, if independence of the rol1
x(to) is known, the state after t o can be computed from by the linear independ
between these two conc
x(t) =l1>(t, to)x(to) + Irlo l1>(t, r)B(r)u(r) dr (5-39)
Theorem 5-10 (Theort
Theorem 5-9 Consider the dynamiea
The dynamical equation E is observable at t o if and only if there exists a finite defined by
ti> t o such that the n columns of the q x n matrix function C(' )<1>(', to) are linearly
independent on [t o, ti].

Proof where A*, B*, C*, and E


E in E. The equation
Sufficiency: Multiplying l1>*(t, to)C*(t) on both sides of (5-37) and integrating equation E* is observab
from t o to tI' we obtain
Proof
[ l1>*(t, to)C*(t)y(t) dt =[[ l1>*(t, to)C*(t)C(t)l1>(t, to) dt] X o ~ Veto, t l)xO
From Theorem 5-4, the
(5-40) rows of l1>(t o, t)B(t) are 1
1,
5-9, the dynamical equ.
Veto, t ¡}~ B*(t)l1>a(t, to) are linearly
where
1 l1>*(t, to)C*(t)C(t)l1>(t,to) dt
lo

(5-41 )

[B*(t)<I>a(t, to)J* = <I>;:'(t, t


From Theorem 5-1 and the assumption that all the eolumns of C(')l1>(', to) are <l>a is the state transition r
linearly independent on [to, t 1J, we concIude that Veto, tI) is nonsingular. l1>Uo, t) (see Problem 4-8)
Hence, from (5-40) we have

Xo = V- 1 (t o, tI) 1
1

'
lo
l1>*(t, to)C*(t)y(t) dt (5-42) We list in thefollowir
tions 5~6 to 5-8, which a
Thus, ifthe function y[lo, Id is known,x o can be computed from (5-42). Necessity: 5-4 for controllability. ~
OBSERVABILITY 01' LINEAR DYNAM1CAL EQUATIONS 195
CAL EQUATIONS

terminal. Therefore, Prove by contradiction. Suppose E is observable at to, but there exists no
Iductor from the input ti> t o such that the columns of COP(', to) are linearly independent on [to, ti].
lmical equation is not Then there exists an n x 1 nonzero constant vector a such that

• C(t)<.l>(t, tola = O for al1 t> t o

:n by
Let us choose x(to) = a; then
Y(t) = C(t)<.l>(t, to)a = O for al1 t> t o
11; + E(t)u(t) (S-36)

Hence the initial statex(t o) = a cannot be detected. This contradicts the assump­
[n the study of observ­
tion that E is observable. Therefore, if E is observable, there exists a finite
;nown, the initial state ti> t o such that the columns of C(· )<.1>(', t o) are linearly independent on [t o' ti].
Q.E.D.

(S-37 ) We see from this theorem that the observability of a linear dynamical equa­
tion depends only on C(tl and <I>(t, to) or, equivalently, only on C and A. This
(S-38) =
can also be deduced from Definition 5-5 by choosing u O. Hence in the ob­
servability study, it is sometimes convenient to assume u =0 and study only
'oblem is a problem of X= A(t)x, Y= C(t)x.
nd <I>(t, to). Note that The controllability of a dynamical equation is determined by the linear
t time to. However, if independence of the rows of <I>(to, .)B('), whereas the observability is determined
by the linear independence of the columns of q. )<1>(', t o). The relationship
between these two concepts is established in the following theorem.
d, (5-39 )
Theorem 5-10 (Theorem of duality)
Consider the dynamical equation E in (5-35) and the dynamical equation E*
defined by
y if there exists a finite
IC(' )<1>(', t o) are linearly E*: z= - A*(t)z +C*(t)v (S-43a)
1 =B*(t)z +E*(t)v (S-43b)

where A*, B*, C*, and E* are the complex conjugate transposes of A, B, C, and
E in E. The equation E is control1able (observable) at t o if and only if the
(5-37) and integrating equation E* is observable (controllable) at too

Proof
1 dtJxo~ Veto, t1)xo From Theorem 5-4, the dynamical equation E is controllable if and only if the
(S40) rows of <I>(to, t)B(t) are linearly independent, in t, on [to, ti]. From Theorem
5-9, the dynamical equation E* is observable if and only if the columns of
) dt (541) B*(t)<I>a(t, to) are linearly independent, in t, on [to, t 1J, or equivalently, the rows of
[B*(t)<I>a(t, to)]* = <I>:(t, to)B(t) are linearlY independent, in t, on [to, tlJ, where
lmns of q.)<I>(', t o) are <l>a is the state transition matrix ofi = - A*(t)z. lt is easy to show that <I>:(t, to) =
<I>(t o, t) (see Problem 4-8); hence E is control1able if and only if E* is observable.
lo, t d is nonsingular.
Q.E.D.

(S-42)
We list in the following,for observability, Theorems 5-11 to 5-14 and Defini­
dt
tions 5-6 lo 5-8, which are dual to Theorems5-5 to 5-8 and Definitions 5-2 to
·rom(5-42). Necessity: 5-4 for control1ability. Theorems 5-11 to 5-14 can be proved either directly or
i
I
L
196 CONTROLLABILlTY ANO OBSERVABILlTY OF LINEAR OYNAMICAL EQUATIONS

by applying Theorem 5-10 to Theorems 5-5 to 5-8. The interpretations in the Definition 5-7
controlIability part also apply to the observability part.
The linear dynamical
Theorem 5-11 ( - 00, 00) if and only i

Assume that the matrices A(') and C(-) in the n-dimensional dynamical equation
E are n - 1 times continuously differentiable. Then the dynamical equation E
is observable at t o if there exists a finite ti> t o such that

(544) where the N;'s are as d

Definition 5-8

where k = 0, 1, 2, ... , n-1 (5-45a)


The linear dynamical e
if and only if there exist
WIth (5-45b)
I 0< {31(O"o)I:
O < {33(O"o)I:
*Differential observability, instantaneous observability,
for alI t, where el> is the
and uniform observability
Differential, instantaneous, and uniform observabilities can be defined by using
Time-invariance ca
the theorem of duality; for example, we may define {A, C} to be differentialIy
equation
observable if and only if { -A*, C*} is differentialIy controlIable. However,
°for ease of reference we shalI define them explicitIy in the folIowing.

Definition 5-6
The dynamical equation E is said to be difJerentially observable at time t o if, where A, B, C, and E al
for any state x(t o) in the state space L, the knowledge of the input and the output time interval of interest
over an arbitrarily smalI interval of time suffices to determine x(t o). I time-invariant dynamic
t o ~ 0, and the determino
Theorem 5-12 time interval. Hence, tI:
bility study of linear tirr.
Ir the matrices A. and iC are analytic on (-oo. o')), then the i1-o.imensioi1cJ ay·
namical equation E is differentialIy observable at every t in ( - 00,00) if and only
if, for any fixed t o in ( - 00, 00), Theorem 5 -1 3
No(to) The n-dimensional linea:
NI (to) able if and only if any 01
1. AH columns of CeA!
p =n complex numbers.
N n - 1(to) 1'. AlI columns of C(sI­
2. The observability gra
I

*Sections noted with an asterisk may be skipped withollt loss or continllity.


is nonsingular for an;
'-.-.e-_­
-~----~ .. _---~~, .. ~

.
~
_
(lCAL EQUAUONS OBSERVABILlTY OF LINEAR DYNAMICAL EQUATIONS 197

le interpretations in the Definition 5-7

The linear dynamical equation E is said to be instantaneously observable in

l~;~:;
( - 00, ro) if and only if

nal dynamical equation


~dynamical equation E p l=n for all t in ( - 00, (0)
.t
Nn~l(t)J
where the N/s are as defined in (5-45). 11
(5-44)

Definition 5-8
The linear dynamical equation E is said to be uniformly observable in ( - ro, (0)
~, ... ,n-1 (5-45a)
if and only if there exist a positive (Jo and positive f3i that depends on (Jo such that
(5-45b) 0< f31((Jo)I.:::::;V(t, t +(Jo).:::::;f3z((Jo)1
I
0< f33((Jo)I ':::::;$*(t, t +(Jo)V(t, 1 +(Jo)$(t, t + (Jo) .:::::;f34((Jo)I
Ity, for all t, where $ is the state transition matrix and V is as defined in (5-41). 11

: can be defined by using Time-invariance case. Consider the linear time-invariant dynamical
~, C} to be differentially equation
controllable. However,
he following. FE: x=Ax +Bu (5-46a)
y=Cx +Eu (5-46b)

where A, B, C, and E are n x n, n x p, q x n, and q x p constant matrices. The


observable at time to if,
time interval of interest is [O, (0). Similar to the controllability part, if a linear
the input and the output
time-invariant dynamical equation is observable, it is observable at every
termine x(to). I

t o ::::: O, and the determination of the initial state can be achieved in any nonzero
time interval. Hence, the reference of t o and t 1 is often dropped in the observa­
bility study of linear time-invariant dynamical equations.
:n the n-dimensional dy·­
t in (-ro, (0) ifand only
Theorem 5-13- _
The n-dimensional linear time-invariant dynamical equation in (5-46) is observ­
able if and only if any of the foHowing eq uivalent conditions is satisfied:
1. AH columns of CeA! are linearly independent on [O, ro) over e, the field of
complex numbers.

1'. AH columns of C(sI - A)-l are lineady independent over C.

2. The observability grammian


• Wot~ t eA''C*Ce A < dI

_ontínuity. is nonsingular for any 1 > O.


------------------------- ------~_...:-_~~~~-=--.---------- ..

198 CONTROLLABILlTY ANO OBSERVABILlTY Of LINEAR OYNAMICAL EQUATlONS


CANONICAL OECOMPO~

3. The nq x n observability matrix and

Equivalently, v is the

(5-47) pV o <
The integer v is called
have

has rank n.
4. For every eigenvalue Aof A (and consequently for every A in C), the (n +q) x n
complex matrix where ñ i5 the degree (

Corollary 5-13

has rank n, or equivalently, (sI - A) and e are right coprime. The dynamical equatif
V n _ q , where q is the n
·Observability Indices V:_qVn_q is nonsingul¡

Let A and e be n x n and q x n constant matrices. Define Consider the matri)

1
in order from top to bo
l
independen t rows asso
C2
C: ro (no. of dependent rows) the observability indice!J

Cq

Theorem 5 -14
(5-48)
The set of observability
transformation and any

5-5 CalT1lonical Dre


DYl1léllmicai Equatic

In the remainder of thi


dynamical equations. e
lt consists of k + 1 block rows of the form CAí and is of order (k + l)q x n. The
matrix V = V n _ 1 is the observability matrix. The q rows of e are denoted by
ej, i = 1, 2, ... ,q. Let us search linearly independent rows of Vk in order from
top to bottom. Let rí be the number of linearly dependent rows in CAí,
i = O; 1, ... , k. Similar to the controllability part, we have where A,B, e, and E ar
Weintroduced in the p
ro '::;/"1 '::;/"2'::;" . '::;q
observability. The cond
Since there are at most n linearly iridependent rows in V ro' there exists an integer able are also derived. A ql
v such that said if the eq ua tion is un<
shall study this problem.
(5-49a)
transformation. Let x =
CAL EQUATIONS CANONICAL DECOMPOSITION OF A LINEAR TIME-INVARIANT OYNAMICAL EQUATION 199

and rv=r v +¡ = ... =q (5-49b)

Equivalently, v is the integer such that


pVo<pV¡< ... <pVV_¡=pVv=pV v +¡='" (5-50)
(5-47)
The integer v is called the observability index of {A, C}. Similar to (5-31), we
have

~svsmin(ñ,n-q+l) (5-51)
q
:y .A. in (:), the (n + q) x n
where ñ is the degree of the minimal polynomial of A and q is the rank of C.

Corollary 5-13
The dynamical equation FE in (5-46) is observable if and only if the matrix
coprime. Vn _ q , where q is the rank of C, is of rank n, or equivalently, the n.X n matrix
V:_qVn_q is nonsingular. I

ine Consider the matrix V n _ ¡. lt is assumed that its linearly independent rows
in order from top to bottom have been found. Let Vi be the number of linearly
independent rows associated with e i. The set of {Vi' i = 1,2, .. , , q} is cal1ed
the observability indices of {A, C}. Clearly we have
l rows)
v = max [Vi, i = 1, 2, ... , q l

and VI +V 2 + ... +\'q Sil. The equality holds ir [A, C) is observable.

Theorem 5-14
(5-48) The set of observability indices of [A, el is invariant under any equivalence
transformalion and any ordering or the rows of C. I

S-5 Canonical Decomposition of a Linear Time-llI1Ivariall1lt


Dyn<llmical IEquatioU1J

In the remainder of this chapter, we study exclusively linear time-invariant


dynamical equations. Consider the dynamical equation
I order(k + l)q x n. The
)WS of C are denoted by FE: x=Ax +Bu (5-52a)
,ows of Vk in order from y=Cx+Eu (5-52b)
i
iependent rowS in CA ,
where A, B, C. and E are n x n, n x p, q x n, and q x p real constant matrices.
have
We introduced in the previoús sections the concepts of controllability and
observability. The conditions for' the equation to be controllable and observ­
roo, there exists an integer able are also .d~rived. A questi()ll that m;w be raised at this point is: What can be
saidif the equation is uncontrol1able and/or unobservable? In this section we
shall study this problem. Before proceeding, we review briefly the equivalence
(5-49a) transformation. Let x = Px, where P is a constant nonsingular matrix. The
200 CONTROLLABILlTY ANO OBSERVABILlTY Of LINEAR OYNAMICAL EQUATIONS CANONICAL OECOMPOSI1

substitution of X = Px into (5-52) yields Proof


FE: X=Ax +Bu (5-53a) If the dynamical equati
+Eu
y = Cx (5-53b) have
where A = PAP-¡, B = PB, C = CP-¡, and E = E. The dynamical equations pU~
FE and FE are said to be equivalent, and the matrix P is called an equivalence
Let q¡, qz, ... , qn, be an
transformation. Clearly we have
each i = 1, 2, . . . , n ¡ , .
V~[B: AB: ... : An-¡B] =[PB: PAB:····: PAn-lB] {q¡, qz, ... ,qnJ (Why
=P[B: AB: ... : An-¡B]~PU p-¡

Since the rank of a matrix does not change after multiplication of a nonsingular where the last n - n¡ coi
matrix (Theorem 2-7), we have rank U = rank Ü. Consequently FE is con­ Q is nonsingular. We c
trollable if and only if FE is controllable. A similar statement holds for the into the form of (5-54).
observability part. x = Px we are actually u:
state space. The ith coll
Theorem 5-15 Aq¡ with respect to {q¡, ~
linearly dependent on th
The controllability and observability of a linear time-invariant dynamical
given in (5-54a). The c(
equation are invariant under any equivalence transformation. I
B with respect to {q¡, (
{q¡, qz,· .. , q!!.l}; hence
This theorem is in fact a special case of Theorems 5-8 and 5-14. For easy
Let U and U be the co
reference, we have restated it as a theorem.
we have pU = pU = n¡ (s
In the following, e will be used to stand for controllable, efor uncontrollable,
o for observable, and (5 for unobservable. - [i
U=
Theorem 5-16
Consider the n-dimensional linear time-invariant dynamical equation FE.
If the controllability matrix of FE has rank n ¡ (where n¡ < n); then there exists
=[
an equivalence transformation X = Px, where P is a constant nonsingular where
-k-
Ve
represents the
.
matrix, which transforms FE into AcB wlth k ~n¡ are line
pU =n¡ implies pUc = nI'
FE:
[;:J = [:' t J[::J
z
+ [~'J u (5-54a)
We show now that the
FE and FiL. hav~ the ss­

y = [Ce Ce] [::J + Eu (5-S4b) -A~ZJ [(S]


sI -Ae
and the nI-dimensional subequation of FE Thus the transfer-functiOl

Xc = Acx c + Bcu (5-55a) -


[ Ce - _][sI - Ac
Ce O
y=Ccx c +Eu (5-55b) sI·

is controllable 9 and has the same transfer function matrix as FE. = [Ce Ce] I

=CcCsI -Ac
9 It is easy to show that ¡f¡he equation FE ís observable. then its subequation FE, is also observable.
(Try) which is the transfer-func
IlCAL EQUATlONS CANONICAL DECOMPOSITlON OF A LINEAR T1ME-INVARIANT DYNAMICAL EQUATION 201

Proof

(5-53a) If the dynamical equation FE is not controllable, then from Theorem 5-7 we
(5-53b) have
¡e dynamical equations pU~p[B : AB : ... : An-¡BJ =n¡ < n
is caBed an equivalence Let q¡, q2' ... , qn, be any n¡ linearly independent columns of U. Note that for
each i = 1, 2, ... , n¡, Aq¡ can be written as a linear combination of
: PAn-lB] {q¡, Q2,"" qnJ (Why?) Define a nonsingular matrix
~n-¡B]~ PU (5-56)

¡cation of a nonsingular where the last n - n¡ columns of Q are entirely arbitrary so long as the matrix
msequently FE is con­ Q is nonsingular. We claim that the transformation ji: = Px will transform FE
;tatement holds for the into the form of (5-54). Recall from Figure 2-5 that in the transformation
ji: = Px we are actuaBy using the columns of Q ~ p-¡ as new basis vectors of the
state space. The ith column of the new representation A is the representation of
Aq¡ with respect to {q¡, q2' ... , qn}' Now the vectors Aq¡, for i = 1,2, ... , n 1> are
linearly dependent on the set {q¡, q2' ... , qn,}; hence the matrix A has the form
le-invariant dynamical given in (5-54a). The columns of B are the representations of the columns of
¡ation. • B with respect to {q¡, q2"'" qn}' Now the columns of B depend only on
{q¡, q2' ... , qn,}; hence B is of the form shown in (5-54a).
;-8 and 5-14. For easy Let U and Ü be thecontrollability matrices of FE and FE, respectively. Then
we have pU = pÜ = n¡ (see Theorem 5-15). It is easy to verify that
lle, efor uncontrollable,
- .[BeJ~ AeBe !I • . • !I A~-¡BeJ
U=
O: O: 1 O
=[Üe! A~'B:¡ ... ·¡A~-¡BeJ}n¡ rows
namical equation FE. O¡ O : : O }(n-n¡)rows
~¡ < n), then there exists
where Üe represents the controBability matrices of F Eco Since columns of
constant nonsingular
A~B with k ?:::n¡ are linearly dependent on the columns of Ü e, the condition
pU = n¡ implies pÚe = ni' Hence the dynamical equation F Ee is controllable.
We show now that the dynamical equations FE and FEe , or correspondingly,
FE and FEc • have the same transfer-functíon matrix. It is easy to verify 1:h3.t

(5-54b) (5-57)

Thus the transfer-function matrix of FE is

(5-55a) [Ce CcJ[sl ~Ae s~~~cTT~eJ +E


(5-55b)

'ix as FE.
= [Ce C-J [(Si - Aer 1 (si -AJ-l A:?lSI - Ac)-¡J [Be] +E
e O (si - Acr ¡ O
=Ce(sl -AJ-¡Be +E
ualion FE, is <j.lso observable.
which is the transfer-function matrix of FEe . Q.E.D.
202 CONTROLLABILlTY AND OBSERVABILlTY OF LINEAR DYNAMICAL EQUATIONS CANONICAL DECOMPO;

In the equivalence transformation X = Px, the state space L of FE is divided


into two subspaces. One is the nl-dimensional subspace of L, denoted by Lb
which consists of all the vectors [~c J the other is the (n - n 1 )-dimensional
and

subspace, which consists of all the vectors [:J Since F Ec is controllable, all
Hence, the reduced ce

Xc =
the vectors Xc in Ll are controllable. Equation (5-54a) shows that the state
variables in Xc are not affected directly by the input u or indirectly through the
state vector xc; therefore, the state vector Xc is not controllable and is dropped in
the reduced equation (5-55). Thus, if a linear time-invariant dynamical equation Dual to Theorem
is not controllable, by a proper choice of a basis, the state vector can be decom­ dynamical equations.
posed into two groups: one controllable, the other uncontrollable. By
dropping the uncontrollable state vectors, we may obtain a controllable dy­ Theorem 5-17
namical equation of lesser dimension that is zero-state equivalent to the
original equation. See Problems 5-22 and 5-23. Consider the n-dime
If the observabilitv m:
valence transform~tio:
Example 1
Consider the three-dimensional dynamical equation FE
1

1
y=[1 1 l]x (5-58)
1
and the nrdimensiom
The rank of B is 2; therefore, we need to check U 1 = [B : AB] in determining
the controllability of the equation. Since
1
is observable and has 1
1
1 This theorem can l
the state equation is not controllable. The first n2 rows of P i
Let us choose, as in (5-56), the observability matr
arbitrary so long as P
ro 1 I 1l Xo does not appear eh
p-l = Q = 11 o¡oJ vector Xi; is not observ:
Lo 1 lo Combining Theore
theorem.
The first two columns of Q are the first two linearly independent columns of
U 1; the last column of Q is chosen arbitrarily to make Q nonsingular. Let
Theorem 5-18 (Cane
x=Px. We compute

m~
Consider the linear tirr

A~PAP-' ~[~ J[~


1
o
o
By equivalence transf<

B~Ph[~ ~ ~][~o ~] r~lb __~]o


-1 1
=
10 This is á simplified versior
References 57, 60. and 116
ICAL EQUATIONS CANONICAL DECOMPOSITION OF A LINEAR TIME-INVARIANT DYNAMICAL EQUATION 203

pace ~ of FE is divided O 1
;e of~, denoted by L. l , and 1 1] 1 O
[
le (n - nl)-dimensional O 1
Hence, the reduced controllable equation is
FE c is controllable, all
1) shows that the state
. indirectly through the
Xc = G ~] Xc + [~ ~] u y = [1 2Jx c (5-59 )

llable and is dropped in


:mt dynamical equation

Dual to Theorem 5-16, we have the following theorem for unobservable
:e vector can be decom­ dynamical equations.
r uncontrollable. By
tain a controllable dy­ Theorem 5-17
tate equivalent to the
Consider the n-dimensional linear time-invariant dynamical equation FE.
If the observability matrix of FE has rank nZ(n2 < n), then there exists an equi­
valence transformation X = Px that transforms FE into

FE: (5-60a)

1Jx (5-58) y=[C o O][::J+EU (5-60b)

and the n2-dimensional subequation of FE


I : AB] in determining
Xo= Aox o+Bou (5-61 a)
y =Cox o +Eu (5-61 b)

is observable and has the same transfer-function matrix as FE. I


:3
This theorem can be readily established by using Theorems 5-16 and 5-10.
The first n2 rows of P in Theorem 5-17 are any n2 linearly independent rows of
the observability matrix of {A, C}; the remaining n - n2 rows of Pare entirely
arbitrary so long as P is nonsingular. Equation (5-60) shows that the vector
x¡¡ does not appear directly in the output y al' indirectly through %0' T-lence the
vector Xi; is not observable and is dropped in the reduced equation.
Combining Theorems 5-16 and 5-17, we have the following very important
theorem.
ldependent columns of
~e Q nonsingular. Let Theorem 5-18 (Canonical decomposition theorem) 10

Consider the linear time-invariant dynamical equation


FE: x=Ax +Bu
y=Cx +Eu
By equivalence transformations, FE can be transformed into the following

lO This is a simplified version of the canonical decomposilion theorem. For the general form, see
References 57, 60, and 116. See also Reference S127.
---~._---------

204 CONTROLLABILlTY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS CANONICAL DECOMPC

canonical form This theorem cal


which the uncontror
observable parts. V
(5-62 a ) response matrix of a
and observable part
y = [O Cco : Ce]x +Eu (5-62b) matrix (the input-out¡
controllable and obse
ca
where the vector x is controllable but not observable, x is controllable co input-output descripl
and observable, and Xc is not controllable. Furthermore, the transfer function us why the input-ou
of FE is system, for the uncor
Cco(sI - Aco )-) Bco +E appear in the transfel

which depends solely on the controllable and observable part of the equation Example 2
FE.
Consider the networl
source, the behavior
Proof
detected from the ou
If the dynamical equation FE is not controllable, it can be transformed into the L 1 are not observabl{
form of (5-54). Consider now the dynamical equation FE c which is the con­ the state variable aSSt
trollable part of FE. If FE c is not observable, then FE c can be transformed into metry, the state varial
the form of (5-60), which can also be written as By dropping the state
the network in Figun
[~caJ [AOco
X co
= ~12J[~caJ
A co X
+[~caJ
co B
u
co
the transfer function

-y = [O Cco]i + Eu Before moving to


of Theorem 5-7.
Combining these two transformations, we immediately obtain Equation (5-62).
Following directly from Theorems 5-16 and 5-17, we conclude that the transfer
function of FE is given by A CcoCsI - co)-1 co +
B E. Q.E. O.

LI In

!l
J/
u
ca
y
I
I ni
1------1
I I

I I

I 1
I I

I I

:e :
10
1
I
I
I
CO
1
I
1
I

L -.J

Figure 5-10 Canonical decomposition of a dynamical equation. (e stands for control­


lable, e for uncontrollable, o for observable, and o for unobservable.) Figure 5-11 An uncon
CAL EQUATIONS CANONICAL DECOMPOSITION Of A LINEAR T1ME-INVARIANT DYNAMICAL EQUATION 205

This theorem can be illustrated symbolically as shown in Figure 5-10, in


which the uncontrollable part is further decomposed into observable and un­
observable parts. We see that the transfer-function matrix or the impulse­
(5-62a) response matrix of a dynamical equation depends solely on the controllable
and observable part of the equation. In other words, the impulse-response
matrix (the input-output description) describes only the part of a system that is
(5-62b)
controllable and observable. This is the most important relation between the
>le, xco is controllable input-output description and the state-variable description. This theorem tells
le, the transfer function us why the input-output description is sometimes insufficient to describe a
system, for the uncontrollable and/or unobservable parts of the system do not
appear in the transfer-function matrix description.

le part of the equation Example 2


Consider the network shown in Figure 5-11(a). Because the input is a current
so urce, the behavior due to the initial conditions in C I and L I can never be
detected from the output. Hence the state variables associated with el and
e transformed into the L I are not observable (they may be controllable, but we don't care). Similarly
FE c which is the con­ the state variable associated with L 2 is not controllable. Because of the sym­
ru be '<ausfmmed tuto metry, the state variable associated with C2 is uncontrollable and unobservable.
By dropping the state variables that are either uncontrollable or unobservable,
the network in Figure 5-11(a) is reduced to the form in Figure 5-11 (b). Hence
the transfer function of the network in Figure 5-11(a) is g(s) = 1. 11

Before moving to the next topic, we use Theorem 5-16 to prove statement 4
of Theorem 5-7.
lbta;,., Equattou (5-62)
Ilclude that the transfer Cl
R
Q.E.D.

+
Li Lz
In In

" f c2
y

In In

(a)

In In
y

In In

(b)

Figure 5-11 An uncontrollable and unobservablc systcm with transfer function 1.


206 CONTROLLABILlTY ANO OBSERVABILlTY OF LINEAR OYNAMICAL EQUATlONS CANONICAL OECOMPO

Proof of statement 4 of Theorem 5-7 equation still has the


The matrix (si - A) is nonsingular at every s in iC except at the eigenvalues of A; ing definition.
hence the matrix [si - A : B] has rank n at every s in iC except possibly at the
eigenvalues of A. Now we show that if {A, B} is controllable, then Definition 5-9
p[sl - A: B] =n at every eigenvalue of A. lfnot, then there exist an eigenvalue A linear time-invaria
}, and a 1 x n vector ex f. O such that only if there exists a l
ex[AI -A : B] =0 sion that has the sam
or exA = exA and exB = O state equivalent to FE

which imply Theorem 5-19


A linear time-invariar
and, in general, i = 1,2, ... is control1able and ob

Hence we have Proof


Ifthe dynamical egual
This contradicts the controllability assumption of {A, B}. Hence, if {A, B} is is reducible (Theorem
controllable, then p[ si - A : B] = n at every eigenvalue of A and, conse­ FE is control1able an,
quently, at every s in iC. contradiction. Supp<
Now we show that if {A, B} is not controllable, then p[AI -A : B] < n and observable and th
for sorne eigenvalue A of A. If {A, B} is not controllable, there exists an equi­ FE,
valence transformation P that transforms {A, B} into {A, R} with

A=PAP- 1 =[~c ~c2] B=PB=[~C] of lesser dimension, sa


Let A be an eigenvalue of Ac. We choose pf.O such that PAc=AIJ. Now we Theorem 4-6, we have
form a 1 x n vector ex = [O JI]. Then we have

ex[AI-A; 8]=[0 p][AI~Ac -A 12 '


oc] =0
R Consider now the pro<
Al -Ac '
which implies
o=et[P(Ali - A)p-l : PBJ =~'p[(},~ - A)p-l : RJ

Since ex f.O, we have ii~exP f.O. Because p- 1 is nonsingular, ii(AI - A)P- 1 =0


implies ii(AI - A) = O. Hence we have
a[AI -A : B] =0
In other words, if {A, B} is not controllable, then [si-A: B] does not have a
rank of n for sorne eigenvalue of A. This completes the proof. Q.E. O.
l,
By (5-64), we may replc
Irreducible dynamical equation: We have seen from Theorems 5-16 and
5-17 that if a linear time-invariant dynamical equation is either uncontrollable
orunobservable, then there exists a' dynamical equation of lesser dimension where \'''_1 and Un -¡
that has the same transfer-functioh rilatrix as the original dynamical equation. lable and observable"
In other words, if a linear time-invariant dynamical equation is either uncontrol­ 2-6 that p(VU) = n. N<
lable orunobservable,its dimension caribe reduced such that the reduced matrices; hence the ID
--------------- .~,~"~~-~.

:AL EQUATIONS CANONICAL DECOMPOSITION OF A LINEAR TIME-INVARIANT DYNAMICAL EQUATION 207

equation still has the same zero-state response, This fact motivates the follow­
ing definition.
, the eigenvalues of A;
~xcept possibly at the
Definition 5-9
,s controllable, then
:re exist an eigenvalue A linear time-invariant dynamical equation FE is said to be reducible if and
only if there exists a linear time-invariant dynamical equation of lesser dimen­
sion that has the same transfer-function matrix as FE or, equivalently, is zero­
state equivalent to FE. Otherwise, the equation is said to be irreducible.

Theorem 5-19
A linear time-invariant dynamical equation FE is irreducible if and only if FE
is controllable and observable.

Proof
"-la.BJ=O If the dynamical equation FE is either uncontrollable or unobservable, then FE
is reducible (Theorems 5-16 and 5-17). Hence what we have to prove is that if
f. Hence, if {A, B} is FE is controllable and observable, then FE is irreducible. We prove this by
lue of A and, conse­
contradiction. Suppose that the n-dimensional equation FE is controllable
and observable and that there exists a linear time-invariant dynamical equation
t1 p[ Al - A : BJ < n
FE,
:, there exists an equi­
S} with FE: x=Ax +Su (5-63a)
y =Cx +Eu (5-63b)

of lesser dimension, say ni < n, that is zero-state equivalent to FE. Then, from
Theorem 4-6, we have E = E and
tí IJA c = AfI. Now we
CAkB=CAkS k=O,I,2,.,. (5-64)

Consider now the product

l
-1: BJ
ular, a,(AI - A)P-l =0
CB CAn-lB
CAB CAnB
(5-65)
[
~ : BJ does not have a = CAn-lB CAnB CA2(~-1)B
proof. Q.E.D.
By (5-64), we may replace CAkB in (5-65) by CAkB; consequently, we have
1m Theorems 5-16 and (5-66)
'VU=Vn-1Ü n- l
¡ either uncontrollable
n of lesser dimension where V"-l and Ü"-l are defined as in (5-48) and (5-27)., Since FE is control­
II dynamical equation. lable and observable, we have p U = n and pV = n, lt follows 'from Theorem
Ion is either uncontrol­ 2-6 that p(VU) = n. Now Vn - l and Ü n - 1 are, respectively, qn xn¡ and ni x np
:uch that the reduced matrices; hence the matrix V"-lÜ"-l has a rank of at most ni' However,
208 CONTROLLABILlTY AND OBSERVABILlTY OF LINEAR DYNAMICAL EQUATIONS CONTROLLABILlTY A1

(5-66) implies that PeY n _ 1 Ü n _ ¡) = n > ni' This is a contradiction. Hence, if invariant dynamical e
FE is controllable and observable, then FE is irreducible. Q.E. O. is referred to Referen<

Recal1 from Section 4-4 that if a dynamical equation {A, B, C, E} has a pre­
scribed transfer-function matrix G(s), then the dynamical equation {A, B, C, E} *5-6 Controllabi
is cal1ed a realizatian of G(s). Now if {A, B, C, E} is control1able and observable, Dynamical Equa
then {A, B, C, E} is cal1ed an irreducible realizatian of G(s), In the following we
shall show that all the irreducible realizations ofG(s) are eql..livalent. The controllability
equation are invarial
Theorem 5-20 ceivable that we ma~
by transforming the
Let the dynamical equation {A, B, C, E} be an irreducible realization of a
equation is in a lord:;
q x p proper rational matrix G(s). Then {A, B, C, E} is also an irreducible
almost by inspection
realization of G(s) if and only if {A, B, C, E} and {A, B, C, E} are equivalent;
Consider the n-d
that is, there exists a nonsingular constant matrix P such that A = PAP-1,
equation
B = PB, C = CP -1, and E = E.

Proof
The sufficiency fol1ows directly from Theorems 4-6 and 5-15. We show now the where the matrices A
necessity of the theorem. Let U, V be the controllability and the observability The n x n matrix A is
matrices of {A, B, C, E}, and let Ü, V be similarly defined for {A, B, C, E}. If A.m' A¡ denotes aH tb
{A, B, C, E} and {A, B, C, E} are realizations of the same G(s), then from (5-64) the number of lordan
and (5-65) we have E = E,
A¡ = diag (Ái!, A
VU=VÜ (5":67)
and VAU=VAÜ (5-68) Table 5-1 Jordan-Fu
The irreducibility assumption implies that pV = n; hence the matrix (V*V) is
nonsingular (Theorem 2-8). Consequently, from (5-67), we have
Ü = eY*V) - 1 V*VU ~ PU (5-69) A
(n x 11
where P ~ eY*V)-l V*V. From (5-69) we have p Ü .=s; min (pP, pU), which,
together with pÜ = n, implies that pP = n. Hence P qualifies as an equivalence (

transformation. The first p columns of (5-69) give B = PB. Since pU = n,


Equation (5-69) implies that

P = (ÜU*)(UU*)-l
(n¡ x 11;)
With P=eY*V)-lV*V=(ÜU*)(UU*)-l, it is easy to derive from (5-67) and
(5-68) that V = VP and PA = AP, which imply that C = CP and A = PAP - 1.

(q xn¡
Q.E.O.

This theorem implies that al1 the irreducible realizations of G(s) have the
"same dimensiono Physically, the dimension of an irreducible dynamical equa­ Aij

tion is the minimal number of integrators (if we simulate the equation in an (l1¡j x nij

" analog computer) or the minimal number of energy-storage elements (if the
system" Is an RLC n"etwork) required to generate the given transfer-function
matrix.
We studied in this section only the canonical decomposition of linear time-
~AL EQUATIONS CONTROLLABILITY AND OBSERVABILITY OF JORDAN-FORM DYNAMICAL EQUATIONS 209

tradiction. Hence, if invariant dynamical equations. For the time-varying case, the interested reader
Q.E.D. is referred to References 106 and 108.

A, B, e, E} has a pre­
equation {A, B, e, E} *5-6 Controllability and Observability of Jordan-Form
llable and observable, Dynamical Equations
1, In the following we

quivalent. The controllability and observability of a linear time-invariant dynamical


equation are invariant under any equivalence transformation; hence it is con­
ceivable that we may obtain simpler controllability and observability criteria
by transforming the equation into a special formo Indeed, if a dynamical
cible realization of a
equation is in a Jordan forro, the conditions are very simple and can be checked
is also an irreducible
almost by inspection. In this section, we shall derive these conditions.
C, E} are equivalent; Consider the n-dimensional linear time-invariant Jordan-form dynamical
lch that A = P AP - 1,
equation
JFE: X =Ax +Bu (5-70a)
y=Cx +Eu (5-70b)

15. We show now the where the matrices A, B, and e are assumed of the forms shown in Table 5-1.
and the observability The n x n matrix A is in the Jordan form, with m distinct eigenvalues A¡, ,,1,2' ..• ,
d for {A, D, e, E}. If Am • Ai denotes all the Jordan blocks associated with the eigenvalue Ai; r(i) is
(;(5), then from (5-64) the number of Jordan blocks in A¡; and A¡j is thejth Jordan block in A¡. Clearly,
A i = diag (A¡l' A i2 , ... , Air(i») and A = diag (Al' Az, ... , A m)
(5-67)
(5-68) Table 5-1 Jordan-Form Oynamical Equation
e the matrix W*V) is
ve have

nin (pP, pU), which,


(5-69)

ifies as an equivalence
(n
A ­
x n)

lA'
C=[C I C 2
A2

J (n~p) ~[J
Cm]
-,
= PE. Since pU = n,


(ni x n¡)
{U A'2
A,j
I
B. _ Bi2
(n, "p) -
lB" 1J
,
.

BHj )
erive from (5-67) and
Cj =[C i1 C'2 Cir(il]

CP and A=PAP- l .
(q x n¡)

Q.E.D.

ions of (;(5) have the


~ible dynamical equa­
te the equation in an
. A¡J =[A'~; [ ]
(nij x ni) Al
Hage elements (if the
iven transfer-function Ai

osition of linear time-


210 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS CONTROLLABILITY Al'

Let n¡ and nij be the order of A¡ and A ij , respectively; then observable are that tI
m m r{i) linearly independent.
n= I
i=I
ni = I I
¡=lj=l
nij dent, the set {C IZ1 }'
Hence J F E is not ob:
Corresponding to A¡ and A¡j, the matrices B and C are partitioned as shown.
The first row and the last row of Bij are denoted by bl¡j and b lij , respectively. The conditions 1
The first column and the last column of Cij are denoted by Clij and c lij . required that each of
pendence. The linee:
Theorem 5-21 Furthermore, the rov
determining the cont]
The n-dimensional linear time-invariant Jordan-form dynamical equation J FE The physical mea
is controllable if and only if for each i = 1, 2, ... , m, the rows of the r(i) x p the block diagram o:
matrix studying the general
Jordan-form dynamic
an integrator and a fe.
(5-71 a) block, or more preci:
variable. Each chain
Consider the last cha
are linearly independent (over the field of complex numbers). J FE is observable variables in that chail
if and only if for each i = 1, 2, ... , m, the columns of the q x r(i) matrix that chain can be obs(
same eigenvalue, the
(5-71b) vectors of these chair
are linearly independent (over the field of complex numbers). 11 be studied separately.

Example 1
Consider the Jordan-form dynamical equation
Al 1: O O O O O O O O
O Al: O O O O O 100 ~bl11
------­
-0- - ­ 6- :-l~: O O O O O 1 O ~bl1z
------­
JFE: x= O O -ü-:-il -:
___ .L
O O O x
_
+ O O 1 U~bl13 (5-72a)
------­
O O O O: Az 1 O 1 1 2

lo O °O °O O: O
O: O
Az
O
1 J'
Az l~ ~ ~J
11;2:0;020]
y= [ 110:1:2:02121 x (5-72b)
0:2:3:0
i i i i

The matrix A has two distinct eIgenvalues Al and Az- There are three Jordan
blocks associated with Al; hence r(l) = 3. There isonly one Jordan block associ­
ated with Az ; hence r(2) = 1. The conditions for J FEto be controllable are that
the set {blll , b l1z , b ll3 } and the set {bIZl } be, individually, linearly independent.
This is the case; hence JFE is controllable. The conditions for JFE to be Figure 5-12 Block di¡
AL EQUATIONS CONtROLLABILITY AND OBSERVABILITY OF JORDAN-FORM DYNAMICAL EQUATIONS 211

observable are that the set {C III , C l1 2> CI 13 } and the set {C 12 ¡} be, individually,
linearly independent. Although the set {C III , C 112 , C I13 } is linearly indepen­
dent, the set {C 12 ¡}, which consists of a zero vector, is linearly dependent.
Hence J FE is not observable. E
mrtitioned as shown.
and b 1ij, respectively. The conditions for controllability and observability in Theorem 5-21
/ C I ij and Clij' required that each of the m set of vectors be individually tested for linear inde­
pendence. The linear dependence of one. set on the other set is immaterial.
Furthermore, the row vectors of B excluding the bli/s do not play any role in
determining the controllability of the equation.
amical equation JFE The physical meaning of the conditions of Theorem 5-21 can be seen from
~ rows of the r(i) x p
the block diagram of the Jordan-form dynamical equation JFE. Instead of
studying the general case, we draw in Figure 5-12 a block diagram for the
Jordan-form dynamical equation in (5-72). Observe that each block consists of
an integrator and a feedback path, as shown in Figure 5-13. The output of each
(5-71 a) block, or more precisely the output of e.ach integrator, is assigned as a state
variable. Each chain of blocks corresponds to a Jordan block in the equation.
Consider the last chain of Figure 5-12. We see that if b l21 #=0, then aH state
;). J FE is observable variables in that chain can be controHed; if C l21 #=0, then al! state variables in
>( r(i) matrix
that chain can be observed. Ir there are two or more chains associated with the
same eigenvalue, then we require the linear. independence of the first gain
(5-71 b) vectors of these chains. The chains associated with different eigenvalues can
:rs). 11 be studied separately.

)
y
) ~b111
) ~b112
-
1 U~b!13 (5-72a) u
y
2
);

1J ~b121 u
y

(5-72b)

u
y
tere are three Jordan
Jordan block associ­
controlÍable are that
linearly independent.
tions for J FEto be Figure 5-12 Block diagram of the Jordan-form equation (5-72).
212 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS
CONTROLLABIl.ITY A~

linearly independent.
theorem can be estab

Observe that in <


independent, it is ne<
that is, p = 1-it is n,
Figure 5-13 Analog computer simulation of 1/(5 -).¡).
linearly independent.
form dynamical eqm
Proof of Theorem 5-21 block associated with
to a vector. Thus we
We use statement 4 of Theorem 5-7 to prove the theorem. In order not to be
overwhelmed by notation, we assume [sI - A : B] to be of the form Corollary 5-21
S-Al -1 O b ll1 A single-input linear
O S - Al -1: b Zll trollable if and only
O O S-Al: b111 distinct eigenvalue an
----------------------:- -. .: X; ------= i -- -,
~ b 11Z spond to the last row
A single-output li
: s -Al , : bl1z
- - - - - - - - - - - - -: -s- ~ Jc~ -----~-( --:- -»l-Z-l observable if and onl
distinct eigenvalue an
s - Az : blZ1 to the first column of

(5-73) Example 2
The matrix A has two distinct eigenvalues Al and Az. There are two Jordan Consider the single-v~
blocks associated with Al' one associated with A z. If s =A b (5-73) becomes
O -1 O b ll1 o :
O O -1:
. b Zl1
x=
. o (
O O O:
- - - - - - - - - - - - - - - : - Ó- - - - =---( - -:
,
: O O:
,
b'12 I
b 111
b 11Z (5-74)
r~- -~
There are two distinc
-----------T----------------T-----­
, Al -A z I
- 1 , b 1Z1 I
responds to the last
A1 -A z : b lZ1 _i is zero; therefore, the
I
corresponding to the fl
By a sequence of elementary column operations, the matrix in (5-74) can be therefore, the equatior
transformed into
, Example 3
O -1 O: O
O O -1: O Consider the followin!
I

O O O: b l11
- ------------- -: - ó-- -- =---( - -: O (5-75)
~ _? 0 \ :__ ~l!! _
I Al -A z O, O and
I •

A1 -A z : O
That the state eqmi.ti,
Note that Al -A z is different from zero. The matrixin (5-75), or equivalently, Equation (5-77) is a t
the matrix in (5- 73) at s = Al' has a full row rank if and only if b 1ll aild b 11Z are matrix is in the Jordar
ICAL EQUATIONS CONTROLLABlUTY AND OBSERVABILlTY OF JORDAN-FORM DYNAMICAL EQUATlONS 213

linearly independent. By proceeding similarly for each distinct eigenvalue, the


theorem can be established. Q.E.D.

Observe that in order for the rows of an r(i) x p matrix B¡ to be linearly


independent, it is necessary that r(i) s p. Hence in the case of single input­
that is, p = l-it is necessary to have rU) = 1 in order for the rows of B¡ to be
linearly independent. In words, a necessary condition for a single-input Jordan­
form dynamical equation to be controllable is that there is only one Jordan
block associated with each distinct eigenvalue. For p = 1, the matrix B¡ reduces
to a vector. Thus we have the following corollary.
m. In order not to be
be of the form Corollary 5-21
, b lll A single-input linear time-invariant Jordan-form dynamical equation is con­
b 211 trollable if and only if there is only one Jordan block associated with eaeh
b L11 distinct eigenvalue and all the components of the column vector B that corre­
b l12 spond to the last row of each Jordan block are different from zero.
A single-output linear time-invariant Jordan-form dynamical equation is
: b L12
:t~- --- --=--( - -:- -b l-2­ 1 observable if and only if there is only one Jordan block associated with eaeh
distinct eigenvalue and all the components ofthe row vector e that correspond
s- "'-2 : b l21 to the first column of each Jordan block are different from zero. •

(5-73) Example 2
There are two Jordan Consider the single-variable Jordan-form dynamical equation
becomes
=,1.1' (5-73)

~r--H\~l ~1}
I

b lll
O 1 O: O 10
1
b 211

b l11

x=
. O O 1 0
I x+ 9
y = [1 O O: lJx

b l12
(5-74)
, bL12
There are two distinct eigenvalues O and 1. The component of B which cor­
-T-----­
: b l21 responds to the last row of the Jordan block associated with eigenvalue O
: : b 121 _
is zero; therefore, the equation is not controllable. The two components of e
corresponding to the first column of both Jordan blocks are different from zero;
aatrix in (5-74) can be therefore, the equation is observable. 11

Example 3
o
O Consider the following two Jordan-form state equations:

., O (5-75) [~J=[-~ _~Jx +GJ u (5-76)

., b
o

,
--------
, O
o

l12
and [XIJ
X2 =
[-1
O 0J x + [e- J
-2 e-21
I

u (5-77)


o

o
O
That the state equation (5-76) is controllable follows from Corollary 5-21.
(5-75), or equivaleritly, Equation (5-77) is a time-varying dynamical equation; however, since its A
mly if blll and bl12 are matrix is in the Jordan form and since the components of B are different fram
,=.c:::..:c..:==-C=",.c.'="'='_'="='--=----==='-''''=--=''------ ,_, _--'

214 CONTROLLABILlTY AND OBSERVABILlTY Of LINEAR DYNAMICAL EQUATlONS OUTPUT CON

zero for all l, one might be tempted to conclude that (5-77) is controllable. where A, R, and e are
Let us check this by using Theorem 5-4. For any fixed to, we have response matrix of the

- ('o -
et>(to-t)B(t)= [ e O
t) O J[:-2'­
e- 2('0-'l_
t ]
= :-21
[ - 'o ]
0

The transfer-function n
It is clear that the rows of et>(to - t)B(l) are linearly dependent in t. Hence the
state equation (5-77) is not controllable at any to. I

From this example we see that, in applying a theorem, all the conditions It is clear that G(s) is a
should be carefully checked; otherwise, we might obtain an erroneous con­
clusion. Corollary 5-22

*5-7 Output Controllability and Output Function A system whose transfe


Controllability output controllable if a
over the field of comple
Similar to the (state) controllability of a dynamical equation, we may define
controllability for the output vector of a system. Although these two concepts
are the same except that one is defined for the state and the other for the output, has rank q.
the state controllability is a property of a dynamical equation, whereas the
output controllability is a property of the impulse-response matrix of a system.
Consider a system with the iilput-output description The proof of Theore
trivial consequence of tl
y(t) = roo G(t, r)u(r) dr controllable. We see
can also be stated in te
where u is the p x 1 input vector, y is the q x 1 output vector, G(t, r) is the q x p indeperidence of G(s), tl
impulse-response matrix of the system. We assume for simplicity that G(t, r) The state controlla
does not contain o-functions and is continuous in t and r for t> r. the output controllabili
these two concepts are
Definition 5-10
A system with a continuous impulse-response matrix G(t, r) is said to be output Example 1
conlrollable at time lo if, for any y l ' there exist a finite l I > lo and an input Consider the network sI
U[Io,',] that transfers the output from y(to) = Oto Y(l 1) = y l ' I observable, but it is out

Theorem 5-22
A system with a continuous G(t, r) is output controllable at t o if and only if there
exists a finite ti > t o such that all rows of G(t l , r) are linearly independent in r
on [to, ti] over the field of complex number. I
ID. ID.
+ IF
The proof of this theorem is exactly the same as the one of Theorem -5-4
and is therefore omitted.
u
-
- Ir
1\

We study in the following the class of systems that also have linear time­ ID. ID.
invariant dynamical-equation descriptions. Consider the system that is
describable by
FE: x=Ax +Bu Figure 5-14 A network ,
y=Cx observable.
ICAL EQUATIONS OUTPUT CONTROLLABILITY AND OUTPUT FUNCTION CONTROLLABILITY 215

(5-77) is controHable. where A, B, and C are n x n, n x p, q x n real constant matrices. The impulse­
t o, we have response matrix of the system is
-fo ]

= [ :- 2 10
The transfer-function matrix of the system is
mdent in t. Hence the
I (5-78 )

rem, aH the conditions It is clear that (;(s) is a strictly proper rational function matrix.
ain an erroneous con­
Corollary 5-22

runction A system whose transfer function is a strictly proper rational-function matrix is


output controllable if and only if aH the rows of (;(s) are linearly independent
over the field of complex numbers or if and only if the q x np matrix
luation, we may define [CB: CAB: ... : CAn-lB] (5-79)
.ugh these two concepts
the other for the output, has rank q. I
equation, whereas the
mse matrix of a system.
1
The proof of Theorem 5-7 can be applied here with slight modification. A
trivial consequence of this corollary is that every single-output system is output
controllable. We see that although the condition of output controllability
can also be stated in terms of A, B, and C, compared with checking the linear
ector, G(t, r) is the q x p independence of (;(s), the condition (5-79) seems more complicated.
Ir simplicity that G(t, r) The state controllability is defined for the dynamical equation, whereas
lrfort>r. the output controllability is defined for the input-output description; therefore,
these two concepts are not necessarily related.

,(t, r) is said to be output Example 1


ite tI> t o and an input Consider the network shown in Figure 5-14. It is neither state controllable nor
y\. I observable, but it is output controllable.

~at to if and only ifthere

1
nearly independent in r
I In
In

:he one of Theorem 5-4


+
- '1;
1\
Y
I

-
t also have linear time­
er the system that is
In In
J
Figure 5-14 A network which is output controllable but neither (state) controllable nor
observable.
- - -_ ..•_.. __ ._--_. __._._._._. __ ._--_._._-~--_._._- .... " ..- ...._.. _._- .._-"._--

216 CONTROLLABILITY ANO OBSERVABILITY Of LINEAR OYNAMICAL EQUAll0NS

1-----------------,
range of G(s), for wh
I I
I I YI
I I
u I I
I I
If the input is rest
: l-----+:--.-Y2
I I then the given outpu
L ~

computed from (5-81


Figure 5-15 A system which is controllable and observable but not output controllable. given output has som
needed to generate th
The condition fOl
Example 2
terms of the matrices,
Consider the system shown in Figure 5-15. The transfer-function matrix of interested reader is re
the system is Dual to the output
These problems are Í1

[,;1]
q x p proper rational
exists a p x q rational

s +1
the rows of which are linearly dependent. Hence the system is not output A system is said to ha"
controllable. The dynamical equation of the system is A necessary and suffi
pG(s) = q in IR (s). Tl
x= -x +u controllability; Man~
unique? Is it a prop
which is controllable and observable. I stable? What are its
problems will not be ;
If a system is output controllable, its output can be transferred to any References SI72, S185
desired value at certain instant of time. A related problem is whether it is
possible to steer the output following a preassigned curve over any interval of
time. A system whose output can be steered over any interval of time is said
to be output function control/able or functional reproducible. "5-8 Computatic
In this section, we dis
Theorem 5-23 chapter. As discussed
A system with a q x p proper rational-function matrix (;(s) is output function conditioned; a comput
controllable if and only if p(;(s) = q in lR(s), the field of rational functions with unstable. If we use a
real coefficients. problem, the result wil
if we use a numericall)
Proof
result will be correcto
If the system is initially relaxed, then we have problem, well or ill cor
a problem, if we must
y(s) 7' (;(s)o(s) (5-80)
stable method, the un
If p(;(s) = q-that is, all the rows of (;(s) are linearly indepe~dent over the possible, in the compu
field of rational functions-then theq x q matrix (;(s)(;*(s) is nonsingular As discussed in TI
(Theorem 2-8). Consequently, for any y(s), if we choose trollability of astate el
more suitable for comI
o(s) = (;*(s)(;(s)G*(S))-ly(S) . (~-81)
computational probler
then Equation (5-80) is satisfied. Consequently, if pG(s) = q, then the system is The computation (
output function controllable. If pG(s)< q, we can always find a y(s), not in the is sttaightforward. L(
CAL EQUATIONS COMPUTATIONAL PROBLEMS 217

range of G(s), for which there exists no solution u(s) in (5-80) (Theorem 2-4).
Q.E.D.

Ir the input is restricted to the class of piecewise continuous functions of t,


then the given output function should be very smooth; otherwise, the input
computed from (5-81) will not be piecewise continuous. For example, if the
1t not output controllable. given output has sorne discontinuity, an input containing b-functions may be
needed to generate the discontinuity.
The condition for output function controllability can also be stated in
terms of the m:ttrices A, B, and C. However, it is much more complicated. The
:fer-function matrix of interested reader is referred to Reference 8.
Dual to the output function controllability is the input function observability.
These problems are intimately related to the inverse problem. A system with a
q x p proper rational matrix (;(s) is said to have a right (left) inverse if there
exists a p x q rational matrix GR1(S) [Gu(s)] such that

system is not output A system is said to have an inverse if it has both a right inverse and a left inverse.
A necessary and sufficient condition for (;(s) to have a right inverse is that
p(;(s) = q in rR(s). This condition is identical to that of the output function
controllability; Many questions may be raised regarding a right inverse. Is it
unique? Is it a proper rational matrix? What is its minimal degree? Is it
I stable? What are its equivalent conditions in dynamical equations? These
problems will not be stu<;lied in this text. The interested reader is referred to
be transferred to any References S172, S185, S218, and S239.
oblem is whether it is
ve over any intervalof
interval of time is said
ble. *5-8 Computational Problems
In this section, we discuss sorne computational problems encountered in this
chapter. As discussed in Section 2-9, a problem may be well conditioned or ill
(;(s) is output function conditioned; a computational method may be numerical1y stable or numerical1y
rational functions with unstable. Ir we use a numerically stable method to solve a weíl-conditioned
problem, the result wil\ generally be good. Ir a problem is i1l conditioned, even
if we use a numerical1y stable method to solve it, there is no guarantee that the
result will be correct. Ir we use a numerical1y unstable method to solve a
problem, well or in conditioned, the result must be carefully scrutinized. Given
a problem, if we must use an unstable method because of nonexistence of any
(5-80) stable method, the unstable method should be applied at a stage, as late as
independent· over the possible, in the computation.
5)(;*(S) is nonsingular As discussed in Theorem 5-7, there are several ways of checking the con­
trollability of a state equation. Among them, statements 3 and 4 appear to be
moresuitable for computer computation. However, they may encounter sorne
(5-81 ) computationalproblems, as will be discussed in the fol1owing.
= q, then the system is The computation of the control1ability matrix U = [B AB ... A"- 1 B]
IS find a y(s), not in the is straightforward. Let Ko~ B. We compute K¡ = AK¡_l; i = 1, 2, ... , n-1.
- ... '
---~~.
._... _-~

218 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATlONS

At the end, we have U = [K o K 1 K"-l]. The rank of U can then be For example, consider
computed by using the singular value decomposition (Appendix E) which is a
numerically stable method. If the dimension n of the equation is large, this
process requires the computation of AkB for large k and may transform the
problem into a less well conditioned problem. For convenience in discussion,
we assume that all eigenvalues Ai of A are distinct and B is an n x 1 vector.
We also arrange Ai so that IA 11 ¿IA21 ¿ ... ¿IA"I. Clearly, we can write B as

where v¡ is an eigenvector associated with eigenvalue Ai; that is, Av¡ = AiVi. It is (see Reference S212). "
straightforward to verify on the superdiagonal a
except the 1:: at the (20, 1
computed as

If IA 11is much larger than all other eigenvalues, then we have

for k large
lf e =0, the eigenvalue:
In other words, AkB tends to approach the same vector, vl' as k increases. the migration of the eil
Hence, it will be difficult to check the rank of U if n is large. 5-16, the root locus of i

The same conclusion can also be reached by using a different argument. The indicate the migration (
condition number of a matrix A may be defined as cond A~ IIAI1211A -1112 = a¡JaS' eigenvalues except Al a
where al and as are the largest and smallest singular values of A. It can be shown
that asslA,,1 < IA 11 sal. Hence, if IA 11 ~ IA"I, Cond A is a very large number.
In computer computation, the multiplication of a matrix with a large condition 11 This ¡lnalysis was suggest(
number will introduce large computational error and should be avoided.
Hence, the use of [B AB ... A"-lBJ to check the controllability of {A, B}
is not necessarily a good method.
As an example consider the 10-dimensional state equation (see Reference
S169)

We compute U = [B AB '.. A9 BJ and then compute its singular values.


The three smallest singular values are 0.712 x 10- 7 , 0.364 X 10- 9 , and
0.613 x 10- 12 . If we use a digital computer with a relative precision of 10-10
(or a number will be considered as a zero if its absolute value is smaller than
10-10), then the rank of U is smaller than 10, and we will conclude that the
equation is not controllable, although the equation is clearly controllable
following Corollary 5-21.
If we use the condition rank [sI - A BJ = n, for all eigenvalues of A, to
check the controllability of {A, B}, we must compute first the eigenvalues of A.
The QR method is a stable and reliable method of contputing the eigenvalues of
a matrix. However, the eigenvalues ofsome matrices can be very illconditioned. Figure 5-16 The root
MICAL EQUATIONS COMPUTAT10NAL PROBLEMS 219

~ rank of U can then be For example, consider the matrix


(Appendix. E) which is a
e equation is large, this 20 20
and may transform the 19 20 O
mvenience in discussion, 18 20
nd B is an n x 1 vector.
Lrly, we can write B as o 2 20
6 1

; that is, Av¡ = A¡V¡. lt is (see Reference S212). The diagonal elements range from 1 to 20; the elements
on the superdiagonal are aH equal to 20. The rest of the matrix are aH zeros
except the 6 at the (20, l)th position. The characteristic polynomial of A can be
computed as

:have Ll(s) = n
20

i= 1
(s -i) -(20)19 6

If 6 =0, the eigenvalues of A are clearly equal to A¡ = i, i = 1, 2, ... ,20. To see


:ctor, Vi> as k increases.
the migration of the eigenvalues of A as 6 increases from O, we plot, in Figure
large.
5-16, the root locus of Ll(s) (see Reference S46),u The heavy hnes with arrows
different argument. The
indicate the migration of the eigenvalues as 6 increases from O. We see that aH
~~ IIAlbllA -lllz = (J¡/(J" eigenvalues except Al and AzO become complex if 6 is sufficiently large (actually
Les of A. It can be shown
is a very large number.
11 This analysis was suggested by Proressor D. Z. Zheng.
ix with a large condition
md shouldbe avoided.
controHability of {A, B}

equation (see Reference

El
lpute its singular values.
)-7, 0.364 x 10- 9 , and 10 11
lative precision of 10-10
lte value is smaller than
re will conclude that the
1 is clearly controllable

. all eigenvalues of A, to
lrst the eigenvalues of A.
puting the eigenvalues of
m be very ill conditioned. Figure 5-16 The root locus of ~(s).
220 CONTROLLABILITY ANO OBSERVABILITY Of LINEAR OYNAMICAL EQUATIONS

if e is larger than 7.8 x 10- 14 ). We list in the following Alo and Al! for sorne e:
el = 7.8 x 10- 14 10.5 ±jO
10
82=10- 10.5±j2.73
83 = 10- s 10.5 ±j8.05 where AW is a nz x n
e4 = 1 10.5 ±j16.26
They are computed directly from A by usingthe QR method. [Thecomputation
of the eigenvalues from fl(s) is not advisable because the characteristic poly­
nomial may be more sensitive to 8 than the eigenvalues and the roots of a poly­ (
and
nomial are very sensitive to the variations of the coefficient.J We see that the p
eigenvalues are very sensitive to e. Thus the eigenvalues of A are very ill
conditioned. For this reason, the use of the criterion rank [sI - A BJ = n where P2AWP~ has bl
to check the controllability of {A, B} may yield an erroneous result. if A~2d =0, the control
It turns out that the use of Theorem 5-16 is the best way of checking the
controllability of a state equation. Furthermore, if the equation is not control­
lable, the computation also yields a reduced controllable equation. The proof
of Theorem 5-16 provided a procedure of computing the required equivalence
transformation. The procedure, however, uses the controllability matrix and is has rank nI + n 2 (whel
not satisfactory from the computational point of view. In the following, we lable and can be red!
shall introduce an efficient and numerically stable method to transform a state is transformed, by a se
equation into the form in (5-54). For the single-input case, the method is
essentially the procedure of transforming the matrix A into the Hessenberg A 11
form (see Appendix A). We discuss in the following the general case. A 21
Let PIbe an orthogonal matrix,12 that is, P l 1 = P'I' such that Á~PAP'= O

x x x x
o x x o
P 1 B= O O x : =[B~I] (5-82) C=CP'=[C 1
where pB\I) = nI' pA 2 ]
O O O O smallest integer such
A k ,k-l has dimension ¡
where B\I) is an nI x P upper triangular matrix and nI = rank B. This step can
be accomplished by a sequence of Householder transformations,u We then
compute UnI +n 2 + ... +nk=
A(I) : A(l)] By this process, the co:
- J :_'- - : ~
PI AP'I = A(l) (5-83) The Á in (5-89) is in
[ 2I!I A(l) 22 formation of {A, B} ir
where AW is an nI x nI matrix, AW is an (n -nI) x nI matrix, and so forth. follows (see Reference
If AW = O, the controllability matrix
Step 1. P =Im Ao =A,
U=[B AB:'''; An-IBJ=Pl[~Y}--~W~~~_!~!D~~~~_~~~J
O O O ... (5-84 )
Step 2. Find an ortho~

has rank nI < n and {A, B} is not controllable and can be reduced to controllable
where rank Zj
12If {A, S} has complex elements. we use a unitary matrix, that is. P,' = P¡'
13The P [ can also be obtained by using gaussian elimination with partial pivoting. In this case.
P [ is not orthogonal. However. the method is stiB numericaBy stable. 14 The form of Á is identical t
COMPUTATlONAL PROBLEMS 221
ICAL EQUATIONS

{AW, B~l)}. Ir AW f. O, we find an orthogonal matrix P 2 such that


tIa and ,1,11 for sorne E:
P2AW=[A~2lJ (5-85)

where AW is a n2 x nI matrix with n2 = rank AW and n2 snl' We compute

(5-86)
~d. [The computation
he characteristic poly­
md the roots of a poly­
(5-87)
ient.] We see that the
Llues of A are very ill
rank [sI -A B] =n Now
leous result.
,t way of checking the
:quation is not control­
e equation. The proof
.e required equivalence
:J PI
[
B(l) A(1)B(l)
- _-_-
O O
x x

O O ...
"']
~_- ~}j~y~- _- _~_- _- ~_-_-~-~-~- (5-88)

'ollability matrix and is has rank nI + n2 (where x denotes nonzero matrices), and {A, B} is not control­
In the following, we lable and can be reduced. Ir AW f.O, we continue the process until {A, B}
od to transform a state is transformed, by a sequence of orthogonal transformations, into 14
lt case, the method is
A 11 A 12 A 13 A 1,k-l AIk B\l)
\. into the Hessenberg
: general case. A 2l An A23 A 2,k-l A2k O
o such that Á~ PAP' = O A 32 A 33 A 3,k-l A 3k , B~ PB = O (5-89)

O O O Ak,k-l O

] (5-82) C=CP'=[C l C3 •.• Ck - 1 C k]

where pB~I)=n¡, pA 21 =n2' pA 32 =n 3, ... , pA k,k-l =nk' The integer k is the


smallest integer such that either nk=O or nI +n2 +- .. +nk=n. Note that
A k,k-l has dimension nk x nk-l and has a full row rank. Clearly, we have
.rank B. This step can
p ¿nI ¿n2 ¿n3 ... ¿nk ¿O
~rmations.13 We then
If nI +n2 + ... +nk = n, {A, B} is controílabie; otherwise, it is noí controllable.
By this process, the controllability of {A, B} can be determined.
(5-83) The Á in (5-89) is in the block Hessenberg form (see Appendix A). The trans­
formation of {A, B} into the form in (5-89) can be carried out recursively as
1 matrix, and so forth. follows (see Reference S203):
Step 1. P = 1m Aa = A, B a = B, ñ = 0, j = 1.
Step 2. Find an orthogonal transformation P j such that
!~~l~--:~~J (5-84)
P.B. = [Zj]}n j
) )-1 O
reduced to controllable
where.rankZ j = nj' Ir nj =0, go to step 7.
1 = Pf.
1

'artial pivating. In this case.· 14 The farin of Á is identical to the one in (A-17).
lble.
222 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

Step 3. Compute controllability indices,


computatian. The se,
P A ¡ pi. = Xo Yo] dependent on its left­
o

))- ) [ _!
B A
J..
j j dependent; hence}J.2 =
where X j is a nj x nj matrix. not appear in the subst
Step 4. Update
p=p
[1- O]
n
01' Á, both columns an
Á; hence A 2b¡ and .~21
O Pj
Á, the first column is 1
Step5. ñ=ñ+nj. Ifñ=n,gotostep8.
Á 3 b¡ is linearly depenc
Step 6. j = j + 1 and go to step 2.
column is linearly indo
Step 7. {A, B} is not controllable and can be reduced to an ñ-dimensional
linearly independent o
controllable equation. these indices are equal
Step 8. {A, B} is controllableo triangles as shown in (:
The {A, B} in (5-90)
This is a numerically stable method 01' checking the controllability 01' a state
tion into the following
equation. This algorithm was first proposed by Rosenbrock [S185] and then
discussed in References S6 and S155. These papers used the gaussian elimina­ 1 x
tion with partial pivoting in finding Pj' Van Dooren [S203] and Patel [S170] O
suggested to use orthogonal transformations to improve the numerical stability
O O
01' the algorithm. The singular value decomposition was also suggested in
Reference S170 to determine the rank 01' B i . O O
The algorithm actually reveals more than the controllability 01' {A, B}; it O O
also reveals its controllability indices. A comparison 01' (5-84) and (5-88) with O O
(5-28) yields immediately O O
i =0,1,2, ... O O
where r; is the number 01' linearly dependent columns in AiB and ni+ ¡ is the The matrix Á is said to
number 01' linearly independent columns in AiB. Since the set of controllability formo This is achieved
indices is uniquely determinable from {r;, i=O, 1,2, ... } [see Figure 5-7 or
Equation (5-33)], it is also uniquely determinable from {ni, i = 1, 2, ... }.
Actually we can saya little more about the controllability indices. In order 1
not to be overwhelmed by notation, we assume that {A, B}, where A and B
are respectively 8 x 8 and 8 x 3 matrices, has been transformed into
, ,
~:(~ x
í 1
v x A X ,
, X X
.,, X X
,X
,

, ,
O : 1: x x x x ,x x ,x x ,x
A , ,
O O /!\
------- :x.... x x'x
0J x'x x'x L 1 _

O O O
A= :..t: x x: x x: x x: x
(5-90)
O O O O O ,ti\.: x x: x x: x Then we have
-------
0- -ó- -Ü Xü-~-: ~- -;;-:-; ­
- - I I 1

O O O

O O O
O O O: Oo '1' :b1 e1: d 1 I L -)"t I
-------
O O O 0--0- -Ü-:-6-0-:0-:'E:f¡­
~ '--v--' "-v--'
1st 2nd 3rd
block co\umn

Let B = fb¡ b2 b31; that is, b; is the ith calumn ofB. Thcn from the structure
01' Á and B in (5-90), we can readily veril'y that 111=3,jJ.2 = 1, and}J.3 =4. These
CAL EQUATlONS COMPUTATlONAL PROBLEMS 223

controllability indices can actually be read out directly from {A, B} without any
computation. The second column of the first block column of A is linearly
dependent on its left-hand-side columns. This implies that Ab 2 is linearly
dependent; hence /12 = 1. Once a column becomes linearly dependent, it will
not appear in the subsequent block columns of A. In the second block column
of A, both columns are linearly independent of their left-hand-side columns of
A; hence A2 b1 and A2 b3 are linearly independent. In the third block column of
A, the first column is linearly dependent on its left-hand-side columns; hence
A3"b 1 is linearly dependent, and /11 = 3. The second column of the third block
column is linearly independent, and there is no other column of A which is
to an ñ-dimensional linearly independent of its left-hand-side columns; hence /13 = 4. We see that
these indices are equal to the numbers of 1 enclosed by circles, squares, and
triangles as shown in (5-90).
The {A, B} in (5-90) can be further transformed by equivalence transforma­
ontrollability of a state tion into the following form:
brock [S185J and then , ,
i the gaussian elimina­ 1 x x x x x. x x,, x x,x , ·
;203J and Patel [S170J
the numerical stability
O 1 x
O O 1
x x x: x x', x x:x
x x x. x x. x x' x
_______
,
·
~ _____ L
. ____ ~ __
------­
was also suggested in O O O 1 O O ,, O O; O 0:0
B= O O O Á= O O 1 :0 0:0 O: O
(5-91 )
ollability of {A, B}; it ------­
_______ ~ _____ L ____ ~ __

f (5-84) and (5-88) with O O O O O O 1 0;0 0;0


, ··
O O O O O O :0
_______ _____1 , ____
O O:J_._
O
Ó- -Ó --6

~
~
,
O O O :0 o' O 1 , O .
n AiB and l1i+ 1 is the The matrix Á is said to be in the block companionform or the block Frobenius
le set of controllability formo This is achieved by choosing
. } [see Figure 5-7 or
/'ji, i = 1,2, ... }.

LJ
Jility indices. In order 1
A, B}, where A and B 1
ormed into 1 1
1
l ~.j
( x', x
x: x: x 1
l( x', x
"------­
l( x: x
1
O -\.J
~ - - ~ _:->: ­
(5-90) Then we have
x x x:x x:x
,
x.x
.
l( x:x x x x:x x.x . x:x
\ _':.1_: d 1_ x
- .. -
x
- .. -
X'x
- ..' - - -
X'X
- - ~ - - -
x:x
- _1 .. _ _
) /! :.1'1'.1
1 x x:x x:x x:x
--v--' (5-92)
3rd O O 1:, x x:, 'x x:, x
0- -Ó- -6 : -( - :-x- : :-x- - ~ -:-; ­
~hcn rrom the structure O O O: O 1 : b2 C2 :d 2
= 1, and f13 = 4. These 0- -Ó- -O : -0- - b- :- 0- - i -:"0 ­
224 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQU.\ 11\ ' ' ' '

Proceeding upward, {A, B} in (5-90) can be transformed into the form in (5-91). are left coprime. Sirr
This process can be easily programmed on a digital computer (see Reference SG).
The process of transforming {A, B} into the block Hessenberg form in (5-89) -A B: O
is numerically stable. The process of transforming the block Hessenberg form 1 O:-A
into the block companion form in (5-91), however, is not numerically stable. T)l= O O: I
The matrix P¡I in (5-92) carries out gaussian elimination on columns of A
• • I •
[ : :
,
I :

without any pivoting. lf pivoting is used, the form of A will be altered, and we O O : O
can never obtain the form in (5-91). Hence, P¡I must be chosen without any '-v

pivoting and the process is numerically unstable. 1"0 1"1

The {A, B} in (5-90) or, equivalently, the {A, B} in (5-91) has the con_troJ­
lability indices /ll = 3, /lz = 1, and /l3 = 4. Based on these indices, the {A, B} There are /l + 1 block
can be transformed into A and I and p columns
, B columns. Now we:
1 x x x X x,x'x
, , x x X left to right. Because
O O O 1 O O ,' O ,' O O O O independent of their 1<
O O O O 1 O : O L: __________
O O O O dependent B columns i
_ _ _ _ _ _ _ ..I _ _
------- of the structure of T )l' .
O 1 x
B=PB= - - - - - - -
X x x:x:x x x x
A=PAP- I = -------'--r---------­ (5-93)
O O 1 x x x,x'x
, , x x x
O O O O O 0:0: 1 O O O
O O O O O 0:0:0 1 O O Let /l be the least inte~
, ,
O O O O O 0:0:0 O 1 O
and
by the equivalence transformation 15
It can be shown directl
{):~ O O O O O O O in (5-94) are the same 2
O O O (O O O O O
O O O O O (Ü O O rank U = rank [H A
= total numb<
P= O ~! ~ O
O O O O O
O O ,¡. O OL _ , O O O Consequently, we coo(
A

O O O O L '1'
_ , O O O number of linearly inde
O O O O O O ,¡.l. _ .1
O independent columns e
O O O O O O O -1- J
L _ >
position of [he rows oi
apply Householder tr
The positions of 1 in this matrix are determined from the positions of 1 in pivoting on the rows o
(5-90) with the same encirclements. The transformation of {A, B} into (A, B} Appendix A). Once in
can also be easily programmed without first determining the controllability readily determined.
indices (see Reference SG). The form in (5-93) is said to be in the controllable There are two disa
fotm and is very useful in the design of state feedback as will be seen in Chapter 7. {A, B}. First, the size (
We mentian one more method of checking controllability to conclude this larger than the size of
section. Frómstatement 4 ofTheorem 5-7 and Theorem 0-8', we may conclude controlla1:?le eguation c
that {A, B} iscontrollable if and only ir the polynomial matrices si - A and B of {A, B} into a block H
the controllability of {A
The discussion· of t
15 This performs only permutations of columns and rows. part and will not be rep
:reAL EQU.\ 11' '''' COMPUTATIONAL PROBLEMS 225

into theform in (5-91). are left coprime. Similar to Theorem G-14, we form the matrix
lUter (see Reference S6). -A B: O O:

:ssenberg form in (5-89)


I O:, -A B:
,
block Hessenberg form
not numerically stable. T=
l' 0 0 '' 1 0 ' ," , (5-94 )
• • I • • I

Ltion on columns of A [ :
,
: I :
, :'

will be altered, and we O O : O O:


'-.--' '----v--'
be chosen without any
rl (no. of dependent columns)

(5-91) has the control­


lese indices, the {A, B} There are ¡.¡. + 1 block columns in TI'; each consists of n col umns formed from
A and I and p columns formed from B. We call the former A columns, the latter
B columns. Now we search linearly independent columns of TI' in order from
x x x x left to right. Because of the unit matrix 1, all A columns in TI' are linearly
O O O O independent of their left-hand-side columns. Let ri be the number of linearly
dependent B columns in the (i + 1)th block column as shown in (5-94). Because
O O O O of the structure of TI" we have
----------
x x x x
---------- (5-93)
x x x x
1 O O O
Let /l be the least integer such that
O 1 O O
O O 1 O O~rO~rl ~ ... ~rl'-l <p
and rl'=rl'+l ='" =p

It can be shown directly (Problem 5-35) or deduced from Chapter 6 that the ri
O in (5-94) are the same as the ri in (5-28). Hence we have
O
rankU=rank[B AB ... AI'-IB]=(p-ro)+(p-r¡)+'" +(p-rl')
O =total number oflinearly independent B columns in (5-94)
O
O Consequently, we conclude that {A, B} is controllable if and only if the total
O number of linearly independent B columns in (5-94) is n. Note that the linearly
O independent columns of TI' are to be searched in order from left to right. The
position of the rows of TI" however, can be arbitrary altered. Bence, we may
'1-', apply Householder transformations or gaussian eliminations with partial
TI the positions of 1 in pivoting on the rows of TI' to transform TI' into an upper triangular form (see
on of {A, B} into (A, B} Appendix A). Once in this form, the linearly independent B columns can be
ning the controllability readily determined.
o be in the controllable There are two disadvantages in using TI' to check the controllability of
will be seen in Chapter 7. {A, B}. First, the size ofTI' is (2 +/l)n x (n +p)(¡.¡. +1), which is generally much
lability to conclude this larger than the size of A. Second, ir{ A, B} is not controllable, its reduced
n G-8', we may conclude controllable equation cannotbe readily obtained. Bence, the transformation
tI matrices si - A and B of {A, B} into a block Hessenberg formseems to be a better method of checking
the controllability of {A, B}.
The discussion of the observability part is similar to the controllability
part and will not be repeated.
226 CONTROLLABILlTY AND OBSERVABILlTY Of LINEAR DYNAMICAL EQUATlONS

5-9 Concluding Remarks Similar to Definition :


given any X o and any)
In this chapter we have introduced the concepts of controllability and observa­
transfer X o to Xl' 16 T
bility. Various theorems for linear dynamical equations to be controllable and
A"-lB] =n (see Prob
observable were derived. We shall discuss briefly the relations among some of
Theorem 5-7. Hence,
these theorems. We list first those t'.leorems which are dual to each other:
case are applicable to

TYT T sr r I~::,;em
Controllability: Theorems the time-varying case,
discussed.

Observab ility: Theorems 5-9 5-11 5-12 5-13 5-14 5-17


The theorems in the observability part can be easily derived from the con­ Problems
trollability part by applying Theorem 5-10 (theorem of duality), and vice versa.
Theorems 5-1 and 5-4 (or 5-9) are two fundamental results of this chapter. 5-1 Which of the followil
They are derived with the least assumption (continuity), and hence they are
a. {t, t 2 ,e', e 2 ',te'}
most widely applicable. If additional assumptions (continuous differenti­
b. {e', te', t 2 e', te", te 3 ']
ability) are introduced, then we have Theorems 5-2 and 5-5 (or 5-11), which give
c. {sin t, cos t, sin 2t}
only sufficient conditions but are easier to apply. If we have the analyticity
assumption (the strongest possible assumption) on time-varying dynamical 5-2 Check the controllal
equations, then we have Theorems 5-3 and 5-6 (or 5-12). Theorem 5-7 (or
5-13), which follows directly from Theorems 5-1, 5-3, and 5-4, gives the necessary
and sufficient conditions for a linear time-invariant dynamical equation to be
controllable.
The relationship between the transfer-function matrix and the linear time­
invariant dynamical equation was established in this chapter. This was achieved
by decomposing a dynamical equation into four parts: (1) controllable and
observable, (2) controllable but unobservable, (3) uncontrollable and unobser­
b. x=[~
-2
vable, and (4) uncontrollable but observable. The transfer-function matrix
depends only on the controllable and observable part of the dynamical equation.
If a linear time-invariant dynamical equation is not controllable and not
observable, it can be reduced to a controllable and observable one.
The concepts of controllability and observability are essential in the study of
Chapters 6 to 8. They will be used in the realization of a rational matrix
2~ l~l
-25 -20-,
x

(Chapter 6) and the stability study oflinear systems (Chapter 8). Sorne pr:,tctic:;¡l
implications of these concepts will be given in Chapter 7. ) = L -- J. 3 üJ Ji.
The computational problems of the various controllability and observability
5-3 Show that a linear dyr
conditions are also discussed. A1though the conditions can be stated nicely in finite ti> to such that for an
terms of the ranks of [B AB '" An- 1 B] and [51 -A BJ in the con­ Hint: Use the nonsingularil
trollability case, they are not suitable for computer computations. An efficient
and numerically stablemethod is introduced to transform a dynamical equation 5-4 Show that ir a linear,
into the form in (5-54) or (5-60), and its controllability or observability can then at any t < too ls it true tha
be determined. The algorithm can also be used to reduce a reducible dyilamical controllable at any t> t o?
equation to an irreducible one.
Before conc1uding this chapter, we remark on the controllability of the "6(1) the Ií!erature, ifx o =0, it
n-dimensional linear time-invariant descrete-time equation encompasses both and doe:
abilit)i and die condition o
x(k + 1) = Ax(k) + Bu(k) identical.
CAL EQUATlONS
PROBLEMS 227

Similar to Definition 5-1, we may define {A, B} to be controllable ifand only if,
given any X o and any Xl' there exists an input sequence {u(k)} of finite length to
ollability and observa­ transfer Xo to x1.16 The condition of controllability is that rank [B AB
to be controllable and A"-1B] =n (see Problem 2-20). This condition is identical to statement 3 of
lations among sorne of Theorem 5-7. Hence, most of the results in the time-invariant continuous-time
ual to each other: case are applicable to the discrete-time case without any modification. For
the time-varying case, the situation is different but is simpler. This will not be
-8 51-16 lrheorem discussed.
1

,-14 5-17
5-10

ierived from the con­ Problems


.uality), and vice versa.
results of this chapter. 5-1 Which of the following sets are linearly independent over ( - ro, ro)?
r), and hence they are a. {t, e2 , el, e 2l, te l }
continuous difTerenti­ b. {e', te', t 2 e', te 2 ', te 3t }
-5 (or 5-11), which give c. {sin t, cos t, sin 2t}
le have the analyticity
ne-varying dynamical 5-2 Check the controllability of the following dynamical equations:
12). Theorem 5-7 (or
5-4, gives the necessary
lamical equation to be
a. [~J=[~ :I::l+[~lu
ix and the linear time­
y=[O lJ [:J
x=[ ~ ~ ~Jx+[ ~ ~]u
ter. This was achievéd
;: (1) controllable and b.
trollable and unobser­ -2 -4 -3 -1 1
ansfer-function matrix
h.e dynamical equation.
controllable and not
y=[~ ~ -~lx
=[~O 2~ I~J x + [- ~l u
rvable one.
;:ssential in the study of c. x
1 of a rational matrix -25 -20 ~J
)ter 8). Sorne practica! y = [-1 3 OJ %
7
bility and observability 5-3 Show that a linear dynamical equation is controllable at to if and only if there exists a
can be stated nicely in finite ti> ro such that for any x o, there exists a u that transfers X o to the zeTo state at time ti'
-A B] in the con­ R/nt: Use the nonsingularity of the state transition matrix.
Jutations. An efficient
il a dynamical equation 5-4 Show that if a linear dynamical equalion is controllable al to, lhen ít is controllable
observability can then at any t < to, ls it true that if a linear dynamícal equation is controllable at to, then it is
1::ontrollable at any t> to? Why?
~ a reducible dynamical

: controllability of the 161n the líterature, ifxa =0, it is caBed reachable:irx,=O, il is called contro[[able. Ollr definilion
encompasses both and does not make this distinclion. Ir A is singular, the condítion of reach­
tion ability and the condition of controllability are slightlY dirferenl. If A is nonsinglllar, they are
idenlical.

L
228 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

5-5 Is it true that p[B: AB:···: An- 1BJ =p[AB: A2B:··· : AnBJ? If not, under
Is it possible to choose an
what condition wi1l it be true? is of the form y(t) = te-' f
5-6 Show that if a linear time-invariant dynamical equation is control1able, then it is
5-16 Consider the dyna
uniformly control1able. state ofthe equation is no
yiste-'fort~l?
5-7 Check the observability of the dynamical equations given in Problem 5-2.
5-17 Show that the state
5-8 State (without proof) the necessary and sufficient condition for a linear dynamical
equation can be determÍl
equation E to be differential1y control1able and differentia11y observable at too
11 -1 order. [Hin!: Com

5-9 Check the contro11ability of the fo11owing state equations: 5-18 Show that controll;

a. x=[~ ~Jx +CJu tions are invariant under a


for a11 t and continuously

b. x= [~ ~J x +L~'] u 5-19 Reduce the dynami,

c. x =[~ ~J x +L~2,] u
to a controllable one.
5-10 Check the control1ability and observability of
5-20 Reduce the equatior

~ ~J +[~ ~J
2 -1
x
O O
u
5-21 Reduce the fol1owin.

y = [1 O 1J x
by using statement 3 ofTheorems 5-7 and 5-13.

5-11 What are the contro11ability indices and the observability index of the equation in
Problem 5-1O?
J.
5-12 Compute the contro11ability indices and the contro11ability index of the state equa­ to a contro11able and obser
tion in (3-48).
5-22 Consider the n-dimer
5-13 Given a COl1trollable linear time-invariant single-input dynamical equation
x=Ax +bu
where A is an 11 x 11 matrix and b is an 11 x 1 column vector. What is the equivalent dy­ The rank of its controllabili
namical equation if the basis {b, Ab, ... , An-1 b} is chosen for the state space? Or, equi­
valently, if i = Px and if p-1 = [b Ab ... A"-l bJ, what is the new state equation?
is assumed to be n1 « n). 1
5-14 Find the dynamical equations for the systems shown in Figures 5-3, 5-4, and 5-1\ independent columns ofU.
and check the contro11ability and observability of these equations. n1 x n1 unit matrix. Show

5-15 Consider the dynamical equation

is control1able and is zero-st


Hint: Use Theorem 5-16.

5-23 In Problem 5-22, the


y=[l 1J x Find a method to solve P 1 ir
.fICAL EQUATIONS PROBLEMS 229

... : A"B]? lf not, under Is it possibleto choose an initial state at t =0 such that the output ofthe dynamical equation
is ofthe form y(t)=te- r for t>O?

1 is controllable, then it is 5-16 Consider the dynamical equation in Problem 5-15. 1t is assumed that the initial
state ofthe equation is not known. Is it possible to find an input u¡O,,,,) such that the output
y is te -, for t :?1 ?
n in Prob\em 5-2.
5-17 Show that the state of an observable, n-dimensional linear time-invariant dynamical
ion for a linear dynamica\ equation can be determined instantaneous]y from the output and its derivatives up to
.bservable at too n - 1 order. [Hint: Compute y(t), j;(t), . .. , yI"-¡)(t).J

5-18 Show that controllability and observability of linear time-varying dynamical equa­
tions are invariant under any equivalence transformation x = P(t)x., where Pis nonsingular
for all t and continuously differentiable in 1.

5-19 Reduce the dynamical equation

y=[1 lJx

to a controllable one.

5-20 Reduce the equation in Problem 5-19 to an observable one.

5-21 Reduce the following dynamical equation

lity index of the equation in


.~ ¡l' }'l
O A¡
O O
O O
O
1
O
O
O
..1. 2

O
},[~}
y=[O O lJ x

i1ity index of the state equa­ to a controllable and observable equation.

5-22 Consider the n-dimensional linear time-invariant dynamical equation


iynamical equation F ¡; . y. = Av" + 131.1
y=Cx+Eu
What is the equivalent dy­ The rank of its controllability matrix,
.r the state space? Or, equi­ U=[B: AB:"': A"-¡BJ
is the new state equation? is assumed to be n¡ «n). Let Q¡ be an n x n¡ matrix whose columns are any ni linearly
independent columns ofU. Let Pibe an ni x n matrix such that P ¡Q¡ = 1,,\, where 1"1 is the
in Figures 5-3,5-4, and 5-11
n¡ x n¡ unit matrix. Show that the following ni-dimensional dynamical equation
ions.
FE: x¡ =P¡AQ¡x¡ + P¡Bu
y=CQ¡x¡ +Eu
is controllable and is zero-state equivalent to FE. In other words, FE is reducible to FE.
Hint: Use Theorem 5-16

5-23 In Problem 5-22, the reduction procedure reduces to solving for p¡ in p¡Q¡ = I,,¡.
Find a method to solve P ¡ in P ¡Q¡ = 1"1
230 CONTROLLABILITY AND OBSERVABILITY OF LINEAR DYNAMICAL EQUATIONS

5-24 Develop a similar statement as Problem 5-22 for an unobservable linear time-

5-32 Consider two SYSl


invariant dynamical equation.

5-25 Is the following Jordan-form dynamical equation controllable and observable?

O
2 1
2
O
O
O
O
O
O
O
O
O
O
2 1 1
2 1 1
e, ~r
O O 2 O O O O 1 1 1 are they output controlla
x= O O O 2 O O O x+ 3 2 1 u
O O O O 1 1 O -1 O O 5-33 AA q X P ra tional rr
m!-trix G l f(S) such that 1
O O O O O 1 O 1 O 1 pG(s) = p in lR(s).
O O O O O O 1 1 O O
5-34 Let P be a nonsing

y~[:
2 1 3 -1 1
1
1
1
1
2
1
O
O
O
1 i}
where B I is an nI x p matl
PE

5-26 Is it possible to find a set of bij and a set of cij such that the following Jordan-form sions (n - n¡) x nI and (n­
equation ifand only if {A 22 , A 21 } is

.{j
1 O O
5-35 Show that
1 O O O b 21 b
O
O
1
O
1
1
0] ["
O x + b 31
O b41
b'j
b
22

32

b42
u pU =p[B AB '" A"­

O O O 1 b 51 b5 wh\?re Iinearly independent


Hint: Premultiply T~ by

y=
['" C2l

C31
CI2

C22

C32
Cl3

C23

C33
CI4

C24

C34
";]
C25

C35
X

is controllabIe? Observable?

5-27 Show that A is cyclic (see Problem 2-45) if and only ifthere exists a vector b such that
5-36 Show that {A, C} is o
{A, b} is controllable. (Hint: Use Corollary 5-21.)
are, respectively, n x n and q
pose ofe.
5-28 Show that if {A, !3} is controllable and A is cyclic. then there exists 8 fJ x 1 solumn
vector r such that {A, Br} is controllable. (H int: Use Theorem 5-21.)
5-37 Showthat{A, B}isc
that is orthogonal to aH coll
5-29 Show that a necessary condition for {A, B} where Bis an n x p matrix, to be control­
lable is p ?m, where m is the largest number of Jordan blocks associated with the same left eigenvector O( of A such ti
eigenvalue of A. (See Problem 5-26 and use Theorem 5-21.)
This is called the Popov-Befe
5-30 In Corollary 5-22 we have that a dynamical equation with E = O is output of Theorem 5-7.)
controllable if and only if p[CB: CAB: . --: CA" - lB] = q. Show that if E F O, then the
dynamical equation is output controllable ir and only ir
p[eB : CAB: ... : CA" - lB: E] = q

5-3i Consider a linear time-inváriant dynamical equation with E =0. Under ·what
condition on e will the (state) controllability imply output cóntrollability?
.... _..--:..:=.

PROBLEMS 231
CAL EQUATIONS

Inobservable linear time- 5-32 Consider two systems with the transfer-function matrices

able and observable?


G[--rS~11 (S~S2;s3~I)J G-l(S~2)J
(s +2) 2- (s+1)
211 (s +1) (s +3)(s +1) (s +2)
211 are they output controllable? üutput function controllable?
11 1
321 u 5-33 A q x p rational matrix G(s) is said to have a left inverse if there exists a rational
matrix Gu(s) such that GLI(s)G(s) =11'. Show that G(s) has a left inverse if and only if
·11 OO O1 pG(s) = p in lR(s).

1O O 5-34 Let P be a nonsingular matrix such that

PB=[~IJ I
PAP- =[ ~~:
where B I is an ni x P matrix and pB = pB I = ni' The matrices A21 and A 22 have dimen­
sions (n - n ¡) x ni and (n - n¡) x (n - n d, respectively. Show that {A, B} is controHable
the foHowing Jardan-form
if and only if {A n , A2 ¡} is controllable.

5-35 Show that


'21
12J pU = p[B AB ... A"-IB] = total number of linearly independent B columns ofT"
'32 U in (5-94)
'42
where linearly independent columns of TI' are to be searched in order from left to right.
)5 Hint: Premultiply TI' by

re exists a vector b such that 5-36 Show that {A, c} is observable if and only if {A, C*C} is observable, where A and C
are, respectively, 11 x n and q x /l constant matrices and C* is the complex conjugate trans­
pose of C.
1 there exists a p x 1 col uml1
m 5-21.) 5-37 Show that {A, B} is controllable if and only if there exists no left eigenvector of A
that is orthogonal to aH columns of B, that is, there exist no eigenvalue A. and nonzero
n n x p matrix, to be control­ left eigenvector a of A such that
ks associated with the same ).a=aA and IXB=O
This is called the Popov-Belevitch-Hautus test in Reference S125. (Hint: See statement 4
of Theorem 5-7.)
ation with E =0 is output
Show that if E ~O, then the

.n with E =. O U nder what


:ontrollability?
_do •. _ _ • _ _ •••• _."-" .-.- '" -----

_
.
~
_
.
_
~
.
~
function is the only a\
tion is obtained, by a~
system can be readily
operational amplifier (
link between the state
For every realizat

6 number of linear timt


a major problem in n
that a dynamical-eqm
Irreducible Realizations, good realization. We
dimension must be a c(
Strict System Equivalence, if a linear time-invaria
if the equation is uncor
and Identification 5-19 it is possible to rec
stil1 has G(s) as its trar
if the equation is contr
realization of (;(s) with
vable dynamical equati
Such a realization is ,
In this chapter we study
(1) A rational transfer
observable pan of a dy!
6-1 Introduction an irreducible one. (2)
network, the number of.
Network synthesis is one of the important disciplines in electrical engineering. reasons of economy an
lt is mainly concerned with determining a passive or an active network that found, any other irredUl
has a prescribed impedance or transfer function. The subject matter we shall lence transformation (T
introduce in this chapter is along the same line-that is, to determine a linear This chapter is orga
time-invariant dynamical equation that has a prescribed rational transfer cept of the degree for prt
matrix. Hence this chapter might be viewed as a modern version of network in Theorem 6-2. In Se(
synthesis. rational functions. H
A lillear time-invariant dynamical Aequation that has a prescribed transfer methods are then exten
matrix G(s) is called a realization of G(s). The term "realization" is justified different irreducible rea
by the fact that, b)' using the dynamical equation, the system with the transfer­ the other on coprime f1
function matrix G(s) can be built in the real world by using an operational Sections 6-5 and 6-6. Ir
amplifier circuit. Although we have proved in Theorem 4-10 that every proper tion, cal1ed polynomial rr.
rational matrix has a finite-dimensional linear time-invariant dynamical­ relationships with trans
equation realization, there are still many unanswered questions. In this chapter lished. In Section 6-8 th
we shal1 study this and other related problems. to strict system equivaler
We study the realization problem for the following reasons: First, there are under the coprimeness ,
many design techniques and many' computational algorithms developed ex­ have the same transfer rr
clusively for dynamical equations. In order to apply these techniques and we study the identificatic
algorithms, transfer-function matrices must be realized into dynamical equa­ pairs. The concept of P'
tions. Second, in the design of complex system it is always desirable to simulate This chapter is baseG
the system on an analog or a digital computer to check its performance before 98, 115, S27, S48, SS2, S
the system is built. A systemcannbt be simulated efficiently if the transfer realization of impulse~re~
to References 32, 100, 10
232
=C.C--:_~~I

!
!
INTRODUCTlON 233
i
function is the only available description. After a dynamical-equation realiza­
tion is obtained, by assigning the outputs of integrators as state variables, the
system can be readily simulated. The realization can also be built by using
operational amplifier circuits. Final1y, the results can be used to establish the
link between the state-variable approach and the transfer-function approach.
For every realizable transfer-function matrix (;(s), there is an unlimited
number of linear time-invariant dynamical-equation realizations. Therefore
a major problem in the realization is to find a "good" realization. lt is c1ear
that a dynamical-equation realization with the least possible dimension is a
good realization. We c1aim that a realization of (;(s) with the least possible
dimension must be a control1able and observable dynamical equation. lndeed,
if a linear time-invariant dynamical-equation realization of (;(s) is found, and
,I , if the equation is uncontrol1able or unobservable, then fol1owing from Theorem
5-19 it is \?ossible to reduce the realization to a lesser-dimensional equation that
still has G(s) as its transfer-function matrix. This reduction is impossible only
if the equation is control1able and observable. Therefore, we conclude that a
realization of (;(s) with the least possible dimensioñ is a control1able and obser­
vable dynamical equation, or equivalently, an irreducible dynamical equation.
Such a realization is cal1ed a minimal-dimensional or irreducible realization.
In this chapter we study mainly irreducible realizations for the following reasons:
(1) A rational transfer-function matrix describes only the control1able and
observable part of a dynamical equation; hence a faithful realization should be
an irreducible one. (2) When an irreducible realization is used to synthesize a
network, the number of integrators needed will be minima\. This is desirable for
reasons of economy and sensitivity. Note that if an irreducible realization is
; in electrical engineering.
found, any other irreducible realization can be obtained by applying an equiva­
,r an active network that
lence transformation (Theorem 5-20).
le subject matter we shall
This chapter is organized as fol1ows. In Section 6-2, we introduce the con­
t is, to determine a linear
cept of the degree for proper rational matrices. 1ts significance is demonstrated
scribed rational transfer
in Theorem 6-2. In Section 6-3, various realizations are introduced for scalar
Ddern version of network
rational functions. Hankel theorem is also introduced. The realization
methods are then extended to vector rational functions in Section 6-4. Two
has a prescribed transfer
different irreducible realization methods, one based on Hankel matrices and
1 "realization" is justified
the other on coprime fractions, for proper rational matrices are discussed in
: system with the transfer-
Sections 6-5 and 6-6. In Section 6-7 we introduce a new mathematical descrip­
by using an operational
tion, cal1ed polynomial matrix description, for linear time-invariant systems. lts
-em 4-10 that every proper
relationships with transfer functions and dynamical equations are also estab­
ime-invariant dynamical­
lished. In Section 6-8 the concept ofequivalent dynamical equation is exteDded
questions. In this chapter
to strict system equivalence for polynomial matrix description. lt is shown that,
under the coprimeness assumption, al1 polynomial matrix descriptions which
_g reasons: First, there are
have the same transfer matrix are strictly system equivalent. In the last section,
algorithms developed ex­
we study the identification of discrete-time systems from arbitrary input-output
'ply these techniques and
pairs. The:concept of persistent exciting is introduced.
zed into dynamical equa­
This chapter is based mainly o.n references 15,42,47,60,62,67,68,83,89,
ways desirable to simulate
98, 115, S27: S48, S52, S126, S158, SI61,SI85, S187, S209; and S218. For the
~ck its performance before
realization ofimpulse-response matrices G(t, T), the interested reader is referred
i efficiently if the transfer
to References 32, 100, 101, 114, and S128.
234 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICAnON THE CHARACTERlSTIC F

6-2 The Characteristic Polynomial and the Degree Ir FE 1 is a realiz,


of a Proper Rational Matrix

In this section we shall introduce the concepts of degree and characteristic


polynomial for proper rational matrices. These concepts are the extension of It is easy to show th;
the denominator of a proper rational function and its degree to the matrix + N Ils) are coprime.
case. Consider a proper rational function g(s) = N(s)/D(s). It is assumed that if and only if
N(s) and D(s) are coprime (have no nontrivial common factor). Then the [
denominator of g(s) is defined as D(s), and the degree of g(s) is defined as the or
degree of D(s). Without the assumption of coprimeness, the denominator and
the degree of g(s) are not well defined. In the following, we use det, deg, and dim Ir the denominator 01
to stand, respectively, for the determinant, degree, and dimensiono then we need a nonze

Theorem 6-1 With this theorem


dynamical equation (
Let the single-variable linear time-invariant dynamical equation
function. It is desira
FE 1: X = Ax + bu case. The answer is a
y =cx +eu matrix.

be a realization of the proper rational function g(s). Then FE 1 is irreducible Definition 6-1
(controllable and observable) if and only if
The characteristic poZ
det(sI - A) = k[ denominator of g(s)] (6-1 ) the least common de
denoted by D(;(S), is (
or dim A = deg g(s) of (;(S).1
where k is a nonzero constant.
Example 1
Consider the rational-
Proof
Let Ll(s)~ det (sI - A) and let

c(sI _ A)-I b~ N I(S) (6-2)


- Ll(s)

First we show that {A, b, c} is irreducible if and only if Ll(s) and N I(S) are co­ The minors of order
prime. Indeed, if {A, b, c} is not irreducible, then Theorems 5-16 and 5-17 The minor of order 2
imply the existence of a {A, b, e} such that dim A < dim A and (;1(S) is s +1 and b(;
1/(s + 1), 1/(s + 1), and
Hence the characteristi
~1(S) ~c(sI _A)-1 b =c(sI -A)-lb = N I(S)
Ll(s) Ll(s)
From this example
where J,(s) = det(sI - A). Since deg J,(s) = dim Á < dim A = deg Ll(s), we con­ general different from
elude that Ll(s) and NI (s) have common factors. Reversing the aboye argument,
we can show that if Ll(s) and N I(S) are not coprime, then {A, b, c} is not irredu­ lItis also caBed lhe McMilL
cible. Hence we have established that {A, b, c} is controllable and observable only lo proper ralional mal
if and only if Ll(s) and NI (s) are coprime. thepoles .al s = oo. See Re
:, AND IDENTIFICAnON
THE CHARACTERISTIC POLNOMIAL AND THE DEGREE Of A PROPER RATlONAL MATRIX 235

le Degree lf FE 1 is a realization of g(s), then we have


I(S) eL'1(s) + N I(S)
A( )
g s -e
-
+ NL'1(s) L'1(s)
~ree and characteristic
)ts are the extension of lt is easy to show that L'1(s) and N 1 (s) are coprime if and only if L'1(s) and eL'1(s)
; degree to the matrix + N Ils) are coprime. Consequently, we have established that FE¡ is irreducible
I(S). lt is assumed that if and only if
lon factor). Then the Denominator of g(s) =L'1(s) = det(sI - A)
)f g(s) is defined as the or deg g(s) = deg L'1(s) = dim A
>, the denominator and
'Ale use det, deg, and dim If the denominator of g(s) is not monic (its leading coefficient is not equal to 1),
limension. then we need a nonzero constant k in (6-1). Q.E.D.

With this theorem, the irreducibility of a single-variable linear time-invariant


dynamical equation can be easily determined from the degree of its transfer
~quation function. It is desirable to see whether this is possible for the multivariable
case. The answer is affirmative if a "denominator" can be defined for a ratianal
matrix.

'hen F El is irreducible Definition 6-1


The characteristic polynomial of a proper rational matrix (;(s) is defined to be
(6-1 ) the least common denominator of aH minors of (;(s). The degree of (;(s),
:)]
denoted by 15(;(5), is defined to be the degree of th.e characteristic polynomial
of (;(S).1 . '

Example 1
Consider the rational-function matrices

A
--
5+1
1
--
s+1
1 1 A
~-2-s+1

Gl(S)=.s~l s~l G2(S)=,_s~_1


(6-2)
r
f L'1(s) and N I(S) are co­ The minors of order 1 of (;1 (s) are 1/(s + 1), 1/(s + 1), 1/(s + 1), and 1/(s + 1).
heorems 5-16 and 5-17 The minor of order 2 of (;1(S) is O. Hence the characteristic polynomial of
1 A and
(;1(S) is s +1 and ¿¡(;I(S)=1. The minors of order 1 of (;2(S) are 2/(s +1),
l/(s +1), 1/(s +1), and 1/(s +1). The minar of order 2 of (;2(S) is 1/(s +1)2.
Hence the characteristic polynomial of (;2(S) is (s + 1)2 and 15 (;2(S) = 2. •
N I(S)
- L'1(s) From this example, we see that the characteristic polynomial Of(;(5) is in
general dilTerent from the denominator of the determinant of (;(s) [if (;(s) is a
LlA = deg L'1(s), we cQn­
mg the above argument,
1 {A, b, e} is not irred.u~ . 1 11 is also called lhe McMillan degree or lhe Smilh McMillan degree. The definition is applicable
rollable and observable' on.ly lo proper ralional malr,ices. If G(s)'íS nol proper, lhe d'efinition musl be modified lo include
lhe poles al s = oo. See References 34.62.83. and S185.
236 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTIFICAnON

square matrix] and different from the least common denominator of al1 the We remark that 1
entries of (;(S). Ir (;(S) is scalar (a Ixl matrix), the characteristic polynomial of nonzero constant [se,
(;(s) reduces to the denominator of (;(s). the characteristic po:
6-1' can be establisr
Example 2 reader is referred to 1
Consider the 2 x 3 rational-function matrix
*Theorem 6-2
S
Let the multivariable
A S +l (s +1)(s +2)
G(s) =
[
-1 1
s +1 (s + 1)(s +2)
be a realization of ti
The minors of order 1 are the entries of (;(s). There are three minors of order 2. (controllable and obs
Theyare det (sI ­
s 1 s+1 1 or dir
(s+lf(s+2)+(s+lf(s+2)=(s+lf(s+2)=(s+l)(s+2) (6-3)
where k is a nonzero ,
sil s +4
--'-+ =----­
The irreducibility
s +1 s (s +1)(s +3) (s +1)(s +3)

case will be establishe


1 1 3
bility without relying
(s+ 1)(s+2)s (s+ 1)(s+2)(s+3) s(s+ 1)(s+2)(s+3) here. In fact, this tl
realization discussed '
Hence the characteristic polynomial of (;(s) is s(s + 1)(s + 2)(s + 3) and b(;(s) =4. analogy to Theorem I
I
of the characteristic p,

Note that in computing the characteristic polynomial of a rational matrix,


every minor of the matrix must be reduced to an irreducible one as we did in
(6-3); otherwise we will obtain an erroneous result. 6-3 Irreducible I
In the following, we introduce a different but equivalent definition of the
characteristic polynomial and degree of a proper rational matrix. This defini­ Irreducible real iza1
tion is similar to the scalar case and requires the concepts developed in Appendix study first the transfer
G.
9(S) =
*Definition 6-1'
Consider a proper rational matrix (;(s) factored as (;(s) = N r (s)D r- 1 (s) = where P and (Xi, for i'
D¡l(S)N,(s). lt is assumed that Dr(s) and Nr(s) are right coprime, and D¡{s) and coefficient of D(s) is 1.
N¡(s) are left coprime. Then the characteristic polynomial of (;(s) is defined as
det Dr(s) or det D¡(s)
or, in the time dornain
and thedegree of (;(s) is defined as
deg (;(s) = deg det D,.(s) = deg detD¡(s)
where pi stands for di/ ú
where deg det stands for the degree of the determinant. • - known that -in an nth­
solution for any u, we
* Maybe skipped without loss of continuity_ vector will consist of n (
, AND IDENTIFICATlON IRREDUCIBLE REALIZATION OF PROPER RATIONAL FUNCTIONS 237

lenominator of all the We remark that the polynomials det D,.(s) and det D/s) differ at most by a
Lcteristic polynomial of nonzero constant [see Equation (6-189a)]; hence either one can be used to define
the characteristic polynomial of G(s). The equivalence of Definitions 6-1 and
6-1' can be established by using the Smith-McMillan formo The interested
reader is referred to References 15, S125, and S185.

*Theorem 6-2
Let the multivariable linear time-invariant dynamical equation
FE: x=Ax +Bu
y=Cx +Eu
be a realization of the proper rational matrix (;(s). Then FE is irreducible
:hree minors of order 2. (controllable and observable) if and only ir
det (sI - A) = k [characteristic polynomial of G(s)]
1 or dimA=degG(s)
(6-3)
: + 1)(s +2) where k is a nonzero constant. 11

The irreducibility of the realizations in this chapter for the multivariable


case will be established by using the conditions of controllability and observa­
bility without relying on this theorem. Hence this theorem will not be proved
3) here. In fact, this theorem will be a direct consequence of the irreducible
realization discussed in Section 6-6. 1t is stated in this section because of its
t- 2)(s + 3) and I>G(s) =4. analogy to Theorem6-1 and its bringing out the importance of the concepts
• of the characteristic polynomial and degree of G(s).

ial of a rational matrix,

ucible one as we did in

6-3 Irreducible Realizations of Proper Rational Functions


valent definition of the
Irreducible realization of (3/D(s). Before considering the general case, we
lal matrix. This defini­
study first the transfer function
;developed in Appendix
R ¡,
1
~_.
g(s) =
A /)
Lé> _ _ (6-4)
s" +a¡s"-¡ + ... +all_¡s +a ll = D(s)
where (3 and ai, for i = 1, 2, .. _,n, are real constants. Note that the leading
lS G(s) = N.(s)D.- ' (s) = coefficient of D(s) is 1. Let u and y be the input and output; then we have
. coprime, and D¡(s) and
(lial of (;(s) is defined as D(s)y(s) = (3u(s) (6-5a)

or, in the time domain,


(p" +aIP"- I + ... +all)y(t) = (3u(t) (6-5b)

where pi stands for i/deí. This is an nth-order differential equation. It is well


• known that in an nth-order differential equation, in order to have a unique
solution for any u, we needn number of initial conditions. Hence the state
vector wi\l consist of n components. In this case the output y and its derivatives
i
t
238 IRREDUCIBLE REALlZATIONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTlFICATION IF

up to the (n - l)th arder qualify as state variables. Define (6-8). To show that (
to Figure 6-1 to ShOVi

r
r
A()D. 2(s) D.
X S=:
XI(S~
x
=:
y(s)
sy(s) = J [1 S
:
] A
y(s) (6-6a)
equal to g(s). A diffe

. . .
Xn(s) Sn-¡y(S) S"-¡ where
or, in the time domain,
X¡(t)~ y(t)

X2(t)~y(t) = py(t) = x¡(t)

X3(t)~ y(t) = p2y(t) =X2(t) (6-6b)

Xn(t)~ y<n- I)(t) = pn - 1 y(t) = Xn_ ¡ (t) It is c1ear that (sI - A


valently, to the cofact
Differentiating xn(t) once and using (6-5) we obtain
xn(t) =pny(t) = -(XnX¡ -(Xn-¡X2 - ... -(X¡X n +f3u ¿l(s
(6-7)

These equations can be arranged in matrix form as (see Problem 2-26). 1


Of(f3/¿l(s))(sI-A) can
o 1 O O O
O O 1 O O
O O O O O
x= x+ u (6-8a)
Hence,
O O O 1 O
-(Xn -(Xn -1 -ct n - Z -(X¡ f3
y= [ 1 O O O ]x (6-8b)

The first n - 1 equations of (6-8a) are obtained directly from (6-6b). Theyare
the consequence of the definition of Xi, i = 1, 2, ... , n, and are independent of
the given transfer function. The transfer function g(s) comes into (6-8a) only
at its last equation through (6-7). We draw in Figure 6-12 a block diagram of

2 Nole Ihal ir f3 is moved lo Ihe OUlpUI, Ihen b and e in (6-8) beco me b' = [O O ... O 1] and
e = [f3O O ... O). and g(s) = c(s]

f3
¿l(s

This verifies that (6-8)

dim A, (6-8) is an irredl

, by showing that (6-8)

obserVable. Note tha

Figure 6-1 The blockdiagram of Equation (6-8). coefficients ofg(s) in (6

~-_ _-~-_ _-_ ....


~. ... ... .
--_.... ,_._
.;;.:...:::...~..::.::.:.:.:.::::......,,_.- - '' : : " : : : " - ~ _ . _ - - - - ~ - , - - - - - _ .

ANO IDENTIFICATION IRREDUCIBLE REALIZATIONS OF PROPER RATIONAL FUNCTIONS 239

ine (6-8). To show that (6-8) is a realization of g(s), we may apply Mason's formula
to Figure 6-1 to show, as in Figure 4-6, that the transfer function from u to y is
equal to g(s). A different way is to verify
(s) (6-6a) g(s) =c(sI - A)-l b
where
-1
s
o
O
O - I 0l
O]
··· O
...

~~,
(6-9)
O s
(6-6b)
s O J
It is clear that (sI - A)-l b is equal to the last column of (sI - A)-l 13 or, equi­
valently, to the cofactors of the last row of (131 ~(s))(sI - A), where
~(s)~ det (sI - A) = sn +CtlSn-1 + ... +Ct n
Xn + f3u (6-7)
(see Problem 2-26). In view of the form of si - A, the cofactors of the last row
of (f3/~(s))(sI - A) can be easily computed as
O
13 S S2 ...
O
~(s) [1
O
+ u (6-Sa)
Hence,

In
O -1 O

(SI_A)-'b~[l
13 s O
~ ~ 13 s.
(6-Sb)
O:
-1
: = ~(s)
O
[1 : ]
sn-2
rom (6-6b). They are O s
ld are independent of Ct n Ct n_ 1 Ct 2 S +Ct 1 13 Sn- I
:omes into (6-8a) only (6-10)
12 a block diagram of

11'=[0 o" O I]and


and g(s)=c(sl-A)-lb=[l O .. , O

13 13 (6-11 )
~(s) sn +Ct1s n- 1 + ... +Ct n

This verifies that (6-8) is indeed a realization of g(s) in (6-4). Since deg g(5) =
dim A, (6-8) is an irreducible realization (Theorem 6-1). This can also be verified
by showing that (6-8) is controllable (except for the trivial case 13 = O) and
observable. Note that the realization (6-8) can be obtained directly from the
coefficients of g(s) in (6-4).

J
.. _~_ •• ~ .~~ .. ~ _ _ •• ,~~ • • " , ' _ _ •• ~~_,._,_'~'_'._ _ ••· _ •• 0_ ••' _ . ' ,_~ •• "~

-------~-----~------------

i
-1
I
240 IRREDUCIBLE REALlZATIONS, STRICT SYSTEM EQUIVALENCE AND lDENTIFICATION 11 !
!

IrreduciblR realization of g(s)=N(s)/D(s). Consider the following scalar The term


proper transfer function

(6-12)
in the right-hand sic
remainder gives the r
where Ci i and Pi,
for i =0, 1, 2,. _., n, are real constants. It is assumed that coefficients associatet
Ci o i= O. By long division, 91 (s) can be written as u a unique y can be d
A f31 Sn - 1 +f3 zs n-Z + ... +f3n 6A

as
gl(S)= sn + a sn-l + . . . + a - s + a +e=g(s) +e (6-13)

1 n 1 n Xn(t)~y(t)
X n -1(t)~y< 1)(t)
where e = 91( 00) = Po/Ci o. Since the constant e gives immediately the direct X n _ Z(t)~ y<Z)(t)
transmission part of a realization, we need to consider in the following only the
strictly proper rational function
X 1(t)~ y<n -1)(
N(s) 13 sn - 1 + 13 sn - Z + ... + 13
A( )
g s-6
_ -
)6 _ 1 Z n
(6-14 ) then x = [Xl X z
D(s - sn +a 1sn 1 + ... +a n in (6-17) yields
Let u and y be the input and output of g(s) in (6-14). Then we have

D(s)y(s) = N(s)u(s) (6-15a)

or, in the time domain,

D(p)y(t) = N(p)u(t) (6-15b)


Differentiating Xl in (1

where D(p) = pn + a1p n-l + .. , -i-a mN(p) = f3 1pn-l + f3 zpn- Z + ... + 13m and pi
stands for di/dt i . Clearly, the transfer function of(6-15) is g(s) in (6-14). In the
following we introduce several different realizations of (6-15). The foregoing equatio

Observable canonical-form realization

Consider the nth-order differential equation D(p)y(t) = N(p)u(t). It is well knowh


that if we have n initial conditions-for exam pie, y(to), y(l)(t o), ... , y<n - ll(t 0)­
then for any input UUO.l,j' the output Y[CO,I,j is completely determinabk. in tbi:
case, however, if we choose y(t), y(l)(t), . .. , y<n-l)(t) as state variables as we did
in (6-6), then we cannot obtain a dynamical equation of the form x =Ax+ bu,
y = ex. Instead, we will obtain an equation ofthe form x = Ax + bu, y = ex +elU
+ezu(l) +e3ulZ) + .. '. Hence in order to realize N(s)/ D(s) in the form
x = Ax + bu, y = ex, a different set of state variables has to be chosen. A dynamical equal
Taking the Laplace transform of(6-15) and grouping the terms associated canonical formo The 1
with the same power of s, we finally obtain (6-18) is derived from
This can be verified by
y(s) = N(S)) u(s) +~) {y(O)sn- 1 + [y< 1l(0) +a ¡ y(O) _ t11 U(O)]Sn - z + ... f10w graph or computi
D(s D(s . .
n 1 n 3 e(sI - A
+ [i - l(0) +a¡in~Z)(o) - f3¡ul - )(0) +adn - ){0)
2

-f31uln-3)(0)+'" +C(n-ly(O)-f3n~IU(O)]] (6-16) withthe aid of (6-10), \


~~=~ .=.-=... =.
__ .=~= =.... = = = = = . =......-.... =-=.:..-=.=..
.. =;=="..=..=~_.=.-.=.-=

"ND IDENTIFlCATlON
IRREDUCIBLE REALIZA nONs OF PROPER RA T10NAL FUNcnONs 241

. the following scalar The term


N(s) A A A

--) u(s) =g(s)u(s)


D(s
(6-12 )
in the right-hand side of (6-16) gives the response due to the input u(s); the
remainder gives the response due to the initial conditions. Therefore, if all the
s. It is assumed that coefficients associated with Sil - 1, Sil - z, ... , SO in (6-16) are known. then for any
u a unique y can be determined. Consequently. if we choose the state variables
as
~g(s) +e (6-13 )
XII(t)~y(t)
XII_¡(t)~y<I)(t) +Cl. l y(t)-f3¡u(t)
nmediately the direct XII_Z(t)~y<Z)(t) +CI.¡y(\)(t) - f3¡dl)(t) +:Xzy(t) - f3zu(t) (6-17)
the following only the

(6-14 )
then x = [Xl Xz XII]' qualifies as the state vector. The set of equations
in (6-17) yields

Then we have y=X"


X" +CI.¡x,,- f3¡u
X,,_I =
(6-15a) x,,_z=x,,_¡ +<X zx -f3 zu lI

Differentiating x I in (6-17) once and using (6-15), we obtain


)II-Z + ... +/3",and pi X¡ = -o:"X" +/3I1U
.s 9(S) in (6-14). In the
;-15). The foregoing equations can be arranged in matrix form as

x~ -..-~"-z~ 1~3 I1 ~"-z1


O O O O
Xz 1 O O O -Cl. 1I _
1 Xz
x, /311-¡
P.
'J )u(t). lt is well knowh X3 O 1 O O
, 1+ ,u (6-18a)
(\)(t
o), ... , yI"-¡)(to)­
determinable. ln this
He variables as we did
x,J
O O O
O O O
O
1
-o:z
-o:¡
JlX"-¡J
XII
lf3z
/3¡
J
. the form x =Ax+ bu,
= Ax + bu, y = ex + e I u y= [O O O O ]x (6-18b)
J(s)/D(s) in the form
lo be chosen. A dynamical equation in the form of (6-18) is said to be in the observable
canonical formo The block diagram of (6-18) is shown in Figure 6-2. Since
19 the terms associated
(6-18) is derived from (6-15), the transfcr function of (6-18) is 9(S) in (6-14).
This can be verified by either applying to Figure 6-2 Mason's formula for signal­
flow graph or computing
6¡U(0)]S"-Z + ...
e(sI - A)- 1 b = [e(sI - A)- lb]' = b'(sI -::- A')- le'
- 3)(0)
(6-16)
with the aid of (6·10), where the "prime" symbol denotes the transpose.
242 IRREDUCIBLE REALlZATIONS, STRICT SYSTEM EQUIVALENCE, AND lDENTIFlCATION IR.

Controllable canonie
We shall now introdu(
form realization, of g(.
a new variable v(t) del

and
Equation (6-21) has t
variables as, similar te

Figure 6-2 Block diagram of the observable canonica\-form dynamica\ equation (6-18).
x(,)J:~i:~~
The observability matrix of (6-18) is lxn(s}J

e o O O O O This definition implies


cA O O O O 1 -!Xl sxn(s) = snv(s)
V~ cA 2 O O O 1 -!Xl -!X 2 +!Xi
which becomes, in the
O x x x x
1 x x x x x
where x denotes possible nonzero elements. The matrix V is nonsingular for Using (6-23), Equation
any (Xi and f3i; hence (6-18) is observable no matter D(s) and N(s) in (6-14) are
coprime or not, and is thus called the observable canonicalform realization.
The dynamical equation in (6-18) is controllable as well if D(s) and N(s) are
coprime. lndeed, if D(s) and N(s) are coprime, then deg 9(S) = deg D(s) = dim A
and (6-18) is controllable and observable following Theorem 6-1. If D(s) and
N(s) are not coprime, the observable dynamical equation (6-18) cannot be con­ or, in the time domain,
trollable as well; otherwise, it would violate Theorem 6-1. In the following we
shall show directly that if D(s) and N(s) are not corprime, then (6-18) is not
controllable. Let s - A be a common factor of N(s) and D(s); then we have
These equations can be
D(A)=An+!XIA n- I + ... +!Xn-IA+!Xn=O (6-19)
and N(A)=f3iA n- 1 +f32 An - 2 + ... +f3n=O (6-20)

Define the 1 x n constant vector lX as lX~ [1 A A2 . . . An-lJ. Then (6-20) where


can be writtenas lXb=O, where b isdefined in (6-18a). Using(6-19), it is easy to O
verify that lXA =AlX, lXA 2 =A 2 lX, , lXAn- 1 =An-1lX. Hence the identity lXb =0
O
implies that lXAib =0, for i = 0,1, , n - 1, which can be written as
O
A=

o
which, together with lX =1=0, implies that the controllability matrix of (6-18) has a
-el"
rank less than n. . Hence if D(s)and N(s) are not coprime, the realization in
(6-18) is not controllable. e = [f3n
AND IDENTIFICATION IRREDUCIBLE REALlZATlONS OF PROPER RATIONAL FUNCTlONS 243

Controllable canonical-form realization


We shall now introduce a different realization, called the control1able canonical­
form realization, of g(s) = N(s)D- 1 (s) 01' .v(s) = N(s)D- 1(s)U(s). . Let us introduce
a new variable v(t) defined by ves) = D- 1(s)ú(s). Then we have
D(s)v(s) =ú(s) (6-21 )
and .v(s) = N(s)v(s) (6-22)

Equation (6-21) has the same form as (6-5a); henee we may define the state
variables as, similar to (6-6a),

ynamical equation (6-18).


01'

This definition implies Xl =X l , Xl =X3'" . 'X n- I =Xn- From (6-21), we have


1
SXn(s)=snv(s) = -a1Sn-1v(s)-alsn-líJ(s)_'" -a 1v(s) +ú(s)
=[-a n -a n- 1 -al -a1]X(S)+U(S)
which becomes, in the time domain,
x (6-24)
x
x V· is nonsingular for Using (6-23), Equation (6-22) can be written as
and N{s) in (6-14) are
'a/form rea/izatíon.
lel1 if D{s) and N{s) are
~
y(s) = [.8n .8n-1 ... .81]
ves)
sv(s)
:
J =[.8n
1{s) = deg D{s) = dim A
orem 6-1. lf D{s) and
1 (6-18) cannot be con­
r
S·-l V(s)
01', in the time domain,
\. In the fol1owing we
¡me, then (6-18) is not y(t) = [.8" .8" - 1 .. , .8 l]X(t) (6-25)
D(s); then we have
These equations can be arranged in matnx form as
=0 (6-19)
(6-20) X=Ax +bu y=cx (6-26)
=0
where
. .1.,,-1]. Then (6-20)
sing (6-19), it is easy to O 1 O O O
nce the identity cxb = O O O 1 O O
: written as O O O O O
A= b=

O O O .. 1 O
ymatrix of (6-18) has a
-a" -(Xn - I -a.- l -!Xi 1
ime, the realization in
e = [.8n .8n-1 .8n - 2 ... .81]

.J
244 IRREDUCIBLE REALIZATlONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTlFICATION IRREO"

This is a realizalion of 9(s) in (6-14). Unlike (6-17), there are no simple relation­ Realization from the Ha
ships between Xi, u, and y. The dynamical equation in (6-26) is always control­ Consider the proper rati(
lable no matter whether D(s) and N(s) are coprime or not and is said to be in the
controllab/e canonica/fonn. Ir D(s) and N(s) are coprime, then (6-26) is observ­
able as weJl; otherwise, it is not observable. This assertion is dual to that of the
observable canonical form, and its proof is left as an exercise. A block diagram We expand it into an inf
of (6-26) is drawn in Figure 6-3.
9(S
Example 1 The coefficients {hU), i =
Consider the proper irreducible transfer function These parameters can be
3 2 h(O) = {Jo
A(S) = 4s + 25s +45s + 34
g 2s 3 + 12s 2 + 20s + 16 h(l) = -(1.
h(2) = -(1.
By long division, 9(S) can be written as
A) 0.5s 2 +2.5s + 1 h(n) = -(1.
g(s = S3 +6s2 +10s +8 +2
h(n +0 = -(1.
Hence its controllable canonical-form realization is
i= 1, 2

l:}U -1~ -l}l~}

These equations are obt


(f3 os" + f3 1s" - 1 + . .. + f3

y=[ 2.5 0.5J x +2u and then equating the c


matrix
Its observable canonical-form realization is
h(l)

X~[! ~ -~~]x +(~.5]U

h(2)
H(cx, f3)~ h\3)
1 -6 0.5 [
y=[O O lJ x + 2u h(cx)
lt is caJled a Hanke/ me
{hU), i = 1, 2, 3, ... }.
invo/ved in H«(1., f3).

Theorem 6-3
The proper transfer funl
pH(n, n)=pH
where pdenotes the rar

Proof

Figure6-3 Block diagram of the controllable canonical-form dynamical equation We show first that if del
(6-26). lfdegg(s)=n, then(6-3C
'CE, AND IDENTIFICATION IRREDUCIBLE REALIZATIONS OF PROPER RATIONAL FUNCTIONS 245

ore are no simple relation­ Realization from the Hankel matrix


1 (6-26) is always control­ Consider the proper rational function
tot and is said to be in the
me, then (6-26) is observ­ A() /3 os· + /31 s· - 1 + . .. + /3.
9s = (6-27)
rtion is dual to that of the s· +al s• 1 + ... +a.
(ercise. A block diagram
We expand it into an infinite power series of descending power of s as
g(s)=h(O) +h(l)s-1 +h(2)s-2 + ... (6-28)

The coefficients {hU), i=O, 1,2, ... } will be called the Markov parameters.
These parameters can be obtained recursively from a i and /3i as
h(O) = /30
h(l) = -a¡h(O) +/31
h(2) = -a 1h(1)-a2h(O) +/32 (6-29)

h(n)= -a 1h(n-l)-a 2h(n-2)- ... -a.h(O) +/3.


h(n + i) = -a 1h(n +i -1) - a2h(n + i - 2) _ ... - a.hU) (6-30)
i = 1,2, ...
These equations are obtained by equating (6-27) and (6-28) as
(/3os· +/31 S·-1 + ... +/3.)
=(s· +als·-1 + ... +a.)(h(O) +h(l)s-1 +h(2)s-2 + ...)
and then "equating the coefficients of the same power of s. We form the a x {3
matrix
h(1) h(2) h(3)
h(/3) ]
h(2) h(3) h(4) h(/3 + 1)
u
H(a, /3)~ h~3) h(4) h(5) h(/3 + 2) (6-31 )
[
h(a) h(a + 1) h(a + 2) h(a +/3 -1)
lt is called a Hankel matrix of order a x {3. It is formed from the coefficients
{hU), i = 1, 2, 3, ... }. lt is important to note that the coefficient h(O) is not
involved in H(a, /3).

Theorem 6-3
The proper transfer function g(s) in (6-27) has degree n if and only if
pH(n, n)=pH(n +k, n +l)=n for every k,l = 1,2,3, . . . (6-32)

where pdenotes the rank.

Proof

\-rorm dynamical equation We showfirst that if deg g(s) = n, then pH(n, n) = pH(n + 1, IX)) = pH(oo, (0) = n.

If degg(s) = n, then (6-30) holds, alld n is the smallest integer having this property.

246 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIV ALENCE, AND IDENTIFICATION IR

Because of (6-30) the (n+ l)th row of H(n+ 1, (0) can be written as a linear Note that there is on
combination of the n rows of H(n, CX)). Hence we have pH(n, (0) = up to h(2n) are used ir
pH(n+ 1, (0). Furthermore, we have pH(n, oo)=n; otherwise, there would be algorithm 3 discussed i
an integer ñ smal!er than n with the property (6-30). Because oflhe structure of Hin (6-35) in order fn
H. the matrix H(n + 2,(0) without the first row reduces to the matrix H(n + 1, ro) pendent and the (O" + 1
without the first column. Hence the (n +2)th row of H(n +2, (0) is linearly Then Theorem 6-3 iml
dependent on its previous n rows and, consequently, on the first n row of dependent on their pre
H(n + 2, (0). Proceeding in this manner, we can establish pH(n, CX) ) = pH( cx), (0) = Iinearly dependent ro\\
n. Again using (6-30), the (n+1)th column of H(n, (0) is linearly dependent cal! the (O" + l)th row
on the columns of H(n, n). Proceeding similarly, we have pH(n, n) = (a + k)th row, k = 2, 3,
p H(n + k, n + 1) = n for every k, 1= 1, 2 .... D(s) and N(s) are coprir
Now we show that if (6-32) holds, then g(s) =h(O) +h(l)s-I + ... can be ing algorithm wilI also
reduced to a proper rational function of degree n. The condition pH(n, n) =
pH( 00, CX)) = n implies the existence of {CXi' i = 1,2, ... , n} to meet (6-30). Using [al l
(6-29) we can compute {/3¡, i =0,1,2, ... , n}. Hence we have This equation expresse
/3os· +/3I S·-1 + ... +/3. combination of its pre'
g(s)=h(O) +h(l)s-1 +h(2)s-2 + ... dependent row. Note
- s· +cxls·- 1 + ... +cx.
we do not have a¡ = (J.._
Since the n is the smal!est integer having this property, we have deg g(s) = n.
This completes the proof of this theorem. Q.E. D.
with
Consider the dynamical equation
O 1 O
FE: x=Ax +bu O O 1

lts transfer function is clearly equal to


y =cx +eu
h[ O O O
9(s) = e +c(sI - A)-I b =e +s-lc(I -S-I A)-I b -al -a 2 -a,

which can be expanded as, by using (2-85), c=[ O O


9(S) =e +cbs- l +cAbs- 2 2
+cA bs- 3
+... (6-33) is a controllable and ,
Theorem 6-3, we have
From (6-28) and (6-33), we conclude that {A, b, c, e} is a realization of g(s) in
(6-27) if and only if e =h(O) and h(tJ-i- i)= -'Qlh(a

h(i)=cAi-lb i=1,2,3,... (6-34) Using this, we can read

With this background, we are ready to introduce a different realization.


Consider a proper transfer function 9(s) = N(s)/D(s) with deg D(s) = n. Here
we do not assume that D(s) and N(s) are coprime; hence the degree of g(s) may
AJ ::~: J, Ab, 2

be less than n. We expand 9(s) as in (6-28) by using the recursive equations in lh(O" + 1)
(6-29) and (6-30). We then form the Hankel m'atrix The effect of the multip
h(l) h(2) h(n) 1, or equivalently, shifts
fl(2) h(3) h(n + 1) c, cAkb just picks up thé
H(n +1, n)= (6-35) cb=,
h(n) h(n + 1) h(2n -1)
f¡(n + 1) 'h(n +2) h(2n) 3 For eomputer eOmpul'l.!ion.
:E, AND IDENTIFICATION IRREDUCIBLE REALIZATION OF PROPER RATIONAL FUNCTIONS 247

L be written as a linear Note that there is one more row than column, and the Markov parameters
we have pH(n, 00) = up to h(2n) are used in forming H(n + 1, n). Now we apply the row-searching
herwise, there would be algorithm 3 discussed in Appendix A to search the linearly independent rows of
:cause ofthe structure of Hin (6-35) in order from top to bottom. Let the first (J rows be linearly inde­
o the matrix H(n + 1, oc!) pendent and the ((J + l)th row of H be linearly dependent on its previous rows.
. H(n + 2, oc!) is linearly Then Theorem 6-3 implies that the ((J +k)th rows, k = 1, 2, 3, ... , are alllinearly
, on the first n row of dependent on their previous rows and the rank ofH(n + 1, n) is (J. Hence once a
pH(n, 00) = pH(oo , 00) = linearly dependent row appears in H(n + 1, n), we may stop the search. We shall
)) is linearly dependent call the ((J + l)th row of H(n + 1, n) the primary linearly dependent row; the
, we have pH(n, n) = ((J + k)th row, k = 2, 3, ... , nonprimary linearly dependent rows. Note that if
D(s) and N(s) are coprime, then (J = n; otherwise, we have (J < n. The row scarch­
) +h(l)s- 1 + ... can be ing algorithm will also yield {a¡, i = 1, 2, , } such that
le condition pH(n, n) =
[al az .,. a" (Il O O]H(n+l,n)=O (6-36)
!} to meet (6-30). Using
lave This equation expresses the primary linearly dependent row as a unique linear
sn- 1 + +f3n combination of its previous rows. The element ({l corresponds to the primary
,-1 + +lXn dependent row. Note that if (J = n, then a¡ = IX n _;, i = 1, 2, ... ,n. If (J < n, then
we do not have a¡ = IX n -;. We claim that the (J-dimensional dynamical equation
y, we have deg g(s) = n.
Q.E.D. x=Ax +bu y=cx +eu (6- 37)

with

A~[ O
O 1 O O
[h(ll ¡
n
O O 1 O h(2)
b- : (6-38)
O O O - h((J -1)
-al -az -a3 -a,,_ I -a" h((J)

e=[ O O O O] e= h(O)

+ ... (6-33) is a controllable and observable realization of g(s). Because of (6-36) and
Theorem 6-3, we have
s a realization of g(s) in
h((J + i)= -alh((J + i -l)-azh((J+i - 2)- ... - a"h(i) i= 1,2,3, ...

(6-34) Using this, we can readily show

: a different realization.
¡ith deg D(s) = n. Here
;e the degree of g(s) may
Ab =[~i~~ j, =[~;~: l. . , ~[~~~ :~n, . .
AZb A'b (6-39)

le recursive equations in h((J + 1) h((J +2)J h(k +(JJ

The effect of the multiplication of A simply increases the atgumént ¡in h(i) by
h(n) 1, or equivalently, shifts the elements up by one position. Because ofthe form of
e, cAkb just picks up the first element of Akb as
h(n + 1)
(6-35 ) eb =h(I), eAb =h(2), eAzb =h(3), (6-40)
h(2n -1)
h(2n) 3 For computer compullttion, numerically stable methods should be used. See Appendix A.
248 IRREDUCIBLE REALlZATlONS, STRlcr SYSTEM EQUlVALENCE, AND IDENTIFlCATlON IR

This shows that (6-37) is indeed a realization of g(s). The controllability matrix We note that this
of (6-37) is nominator of g(s) are
smaller than the degrel
[b Ab ... A"-lbJ=H(u,u)
Dual to the introd
The Hankel matrix H(u, u) has rank u; hence {A, b} in (6-37) is controHable. pendent columns of H(
The observability matrix of (6-37) is irreducible realization.
e 100 O
•Jordan-canonical-fo
cA O 1 O O
cAz O O 1 O We use an example to i
a lordan-form dynami
general case. Assume
eA"-1 o O O and assume that D(s) c~
Clearly {A, e} is observable. Hence (6-37) is an irreducible realization of g(s), assume that g(s) can be

A( ) = -el
gs ­
Example 2
(s -;
Consider
The block diagrams of
ficients el!, e lZ , e 13 , ez,

We form the Hankel matrix


:- -6-; _1- J. 1­
I , Z 4 Z U
_1- _J.
. 1­ ~.

" Z ,
I 4
I
Z
1
H(S, 4) = ,--
3 I
I
1- 1 _1­
,__ 1-1 Z 8
1- 1 _1- _2
Z 8 4
1 _1- _2 _J.
8 4 8

Using the row searching algorithm discussed in Appendix A, we can readily


show that the rank of H(S, 4) is 3. Hence an irreducible realization of g(s) has
dimension 3. The row searching algorithm also yields

kH(S, 4)~[~ :I~ 0JH(S,4)=0

Hence an irreducible realization of g(s) is

X~U5
1
O
-1
y=[ O OJ x+O.Su

The last row ofthe companion-form matrix A consists ofthe first three elements
of k with the signs reversed. The b vector consists of the first three Markov
parameters of g(s) [excluding h(O)]. The form of e is fixed and is independent
of g(s). • Figure 6-4 Two block di;
, AND IDENTIFICATION IRREDUCIBLE REALlZATION OF PROPER RATIONAL FUNCTIONS 249

: controllability matrix We note that this procedure also reveals whether the numerator and de­
nominator of 9(S) are coprime. If the rank of the Hankel matrix of 9(S) is
smal1er than the degree of its denominator, then 9(S) is not irreducible.
Dual to the introduced procedure, we may also search the linearly inde­
I (6-37) is controHable. pendent columns of H(n, ri + 1) in order from left to right and obtain a different
irreducible realization. The procedure wil1 not be repeated.

•Jordan-canonical-form realization
We use an example to illustrate the procedure to realize a transfer function into
a Jordan-form dynamical equation. The idea can be easily extended to the
general case. Assume that D(s) consists of three distinct roots Al, ..1. 2 , and ..1. 3 ,
and assume that D(s) can be factored as D(s) = (s - ..1. 1)3(S - . 1. 2 )(s - ..1. 3), We also
;ible realization of 9(S). assume that 9(S) can be expanded by partial fraction expansion into

9(S)= e ll 3 + e12 2 + e l3 + e2 +_e_3_ (6-41)


(s-..1.¡) (s-..1.¡) (S-Al) (s-..1. 2 ) (s-..1. 3 )
The block diagrams of 9(s) are given in Figure 6-4. In Figure 6-4(a), the coef­
ficients e 11' e 12' e 13' e2, and e3 are associated with the output. In Figure 6-4(b),

:ndix A, we can readily (a)


le realization of 9(s) has

1=0 e13 \ - - - - - - - - - - - - - - - ,

XlI=A¡XI\+el\u x12=A¡x¡i+e12u x13=A¡x13+ e 13 u


Xl!

fthe first three elements


. the first three Markov
ixed and is independent (b)

I Figure 6-4 Two block diagniffis of g(5) in Equalion (6-41).


~~"~._.~.-=.=.-~._=.-_.=-_.=--=_ ..._=._=-. -'='-=-'-=-=--='--=-'====-.=--=.--=--=-=---== ... --
. _=._=-._="=---=-= --·C'=~="-,"=c--·--c·-·.=c<=="--·:·=--==_=··-··-----·-='=='C~'O' '-"--'-'-'-,,-~--'" .._..... ---- --- ..--_._..... _......-_ .. :' ...... - _ .....

250 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTlFICATION 11

they are associated with the input We note that every block in Figure 6-4 can numbers. In this ca
be viewed as consisting of an integrator, as shown in Figure 6-5. Hence the puter, for complex m
output of each block qualifies as a state variable. By assigning the output of this can be taken car
each block as a state variable, and referring to Figure 6-5, we can readily obtain will be demonstratec
the dynamical equation for each block of Figure 6-4 as shown. By grouping assumed to be reaL if
the equations in Figure 6-4la), we obtain /:, is also a pole of 9(5).
following subequatioi
1.[ O : O O O
O ;'1 1: O O O
O O 1'1: O O x + 1 u (6-42a)
-o'---6---6-:--~;-: O 1
O O O· ""0"" : ;.~ 1
(6-42b) where Al is the Jordar
Equation (6-42) is in the Jordan canonical farm. There is one Jardan block (no transpose) of Al'
associated with each eigenvalue. The eq uation is clearly controllable (Carollary introduce the equivale
(5-21); it is also observable, except for the trivial cases e ll = O, ez = O, or e 3 = O.
Therefore the dynamical equation (6-42) is an irreducible realization of the
g(5) in (6-41).
If the block diagram in Figure 6-4(b) is used and if the state variables are and
chosen as shown, then the dynamical equation is
XII
.1, O O: O 01 ell Then it can be easily

_~_ _~l ;~ 1.. ~ _, 6 x +


XlZ e lZ transformed into
-X 1 3 .. __ el3 u
Xz O O o' Az ' O
'- __ 1 __ ­
ez
,X3 O O O O: ..1. 3_1 e3
y=[O O 1] x y=[
which is another irreducible Jordan-form realization of 9(5). Note the dif­
where Re A and 1m A e
ferences in the assignment of the state variables in Figures 6-4(a) and 6-4(b).
tively. Since aH the co,
There are two difficulties in realizing a transfer function into aJordan canoni­
cal formo First the denominator of the transfer funcUon must be factored. o. analog computer simul,
correspondingly, the poles of the transfer function g(5) must be computed. This iorm dynamical equati,
is a difficult task if the degree of g(5) is larger than 3. Second, if the transfer transformation introdu
function has complex poles, the matrices A, b, and e will consist of complex
Example 3

Consider the Jordan-fo


3.. : ;: ax + bu
,--------------------1 1 +2i
I I
1


~xu
s-a
---+--­
11 I

I
I
b
x l\'

x=
O
O
- - -
1 +2i :
- -

O ,'
- - - - l.

I
I
O O ,:
I
L ~
O '0 .

Figure 6-5 Internal structure of block b/(s - a). y=[ -i


IRREDUCIBLE REALlZATION I Of PROPER RATIONAL FUNCTlONS 251
, ANO IDENTIFlCATlON

numbers. In this case. the equation cannot be simulated on an analog com­


)Iock in Figure 6-4 can
puter, for complex numbers cannot be generated in the real world. However,
~igure 6-5. Hence the
this can be taken care of by introducing sorne equivalence transforrnation, as
lssigning the output of
will be demonstrated in the fol!owing. Since all the coefficients of g(s) are
i, we can readily obtain
assumed to be real, if a complex number A is a pole of g(s). its complex conjugate
lS shown. By grouping
l., is also a pole of g(s). Hence in the Jordan-form realization of g(s), we have the
following subequa tion:
o
O
(6-42a)
[ ~lJz =[AIO A¡~ J[X1J +[~¡J
X b
XZ
u
l
(6-43a)

1 u
1
1
y=[c¡ cIJ[::J
(6-43b)

(6-42b) where Al is the Jordan block associated with A and Al is the complex conjugate
(no transpose) of A¡. Clearly, Al is the Jordan block associated with 1 Let us
re is one Jordan block introduce the equivalence transformation x = Px, where
controllable (Corol!ary
11 =0, ez =0, or e3 =0.

:ible realization of the


p= [ 1
il
1
-il
J
. the state variables are and P
_¡ _~
-2 1
[1 -;IJ il

Then it can be easily verified that the dynamical equation in (6-43) can be
transformed into

l ~ll=[
x2
ReAl
-1m A¡
ImA1J[~¡J+[
Re Al x2
2Reb l
-21m b¡
Ju (6-44a)

y = [Re e¡ 1m clJ [::] (6-44b)

of g(s). Note the dif­ where Re A and 1m A denote the real part and the imaginary part of A, respec­
Hes 6-4(a) and6-4(b). tivel)'. Since al! the coefficients in (6-44) are reaL this equation can be used on
on intoaJordancanoni­ analog computer simulations. Another convenient way to transform a Jordan­
on must be factored. 01' form dynamical equation into an equation with real coefficients is to use the
lUSt be computed. This transformation introduced in Problems 6-14 and 6-15.
Second. if the transfer
will consist of complex Example 3
Consider the Jordan-form equation with complex eigenvalues.
" 1 +2i 1 O D O 2-3i
I
Ix
O 1 +2i () O
- - - - - - - - - - - - - - '- - - - - - ... - - - - - - - ... -.
e 1
----¡-----­
I
x= O o 1 -2; 1 O x + 2 +3i u (6-45)
l
O O O 1-2;: O 1
I

I
O o '----0-------0---:--2 2
I

.~
y=[ -; i 2J x
252 IRREDU('IBLE REALlZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION REALIZA

Let X = Px. where by choosing y(t), y(t),


exactly the same form
1 O 1 O , O (6-46) consists of the
O 1 O 1 O dynamical equation, t:
- - - - - - _, - - - - - - - - -1- - ­

p= O -i O ,, O general formula, we gi
,
O
_____
i O -i ,
__ , _ _ _ _ _ _ _ _ _ 1___
O
O O O O ,, 1 Example 4

Then t6-45) can be transformed into Consider the followin¡


[p2 + o:¡((
Tbe procedure of f<
We first assume that (E

y
whose coefficients are all real. I and then verify this by
terms of the coefficien
A remark is in order regarding these realizations of y(s). Clearly, the con­ we obtain
trollable canonical form and observable canonical form realizations are the
easiest to obtain. These realizations. however. are not necessarily irreducible y=PY=X2+'
.. 2
unless the given 9(s) = N(s)/D(s) is known to be irreducible [N(s) and D(s) are y=py=-a;
coprime]. The realization obtained from the Markov parameters and the Substituting these into
Jordanform realization are always controllable and observable no matter the obtain
given 9(S) is irreducible or not. Because of the requirement of computing the
poles of 9(S). the Jordan-form realization is generally more difficult to compute. e(t) = 130(1)
The Jordan-form realization however is. as discllssed in References 23, 81 b¡(t) = fi¡(t)­
and 87, least sensitive to parameter varialions among all realizations. The b 2(t) = fi2(t)­
sensilivity is defined as the shifting of the eigenvallles dlle to parameter varia­ Since the time functions
lions. In practice, in order to reduce sensHivity. a transfer function 9(S) is the differential equatior
ractored as ofthe form in (6-48). \\
9(s) =g¡(S)92(.':)··· the relations between b,

or 9(S)=91(S)+92(S)+'"

where 9¡(s) and y¡(s) are transfer functions of degree I or 2. We then realize
6-4 Realizations,
each Ms) and y¡(s), and then connect them together. Thc first one is called a
Transfer Function:
tandem realization; the second one, a parallel realizatioll. This typc of realiza­

tion is often used in the design of digital filters (see Refercncc S.t7).
In this section realizatio
studied. By a vector rati
function matrix. Consi<
·Realization of linear time-varying differehtial equations. Before

conc1uding this section, we shall briefly discuss the se"tup of dynam ical eqllations

for linear, time-varying dirrerential eq uations. Ir an ~th~ordcr. linear. time­

varying dirrerential equatiori is or.the form

(pn +a¡(t)pn-l + ... +an(t))y(t)=fi(t)u(t) (6-46)


:, AND IDENTIFICATION
REALlZATIONS OF VECTOR PROPER RATIONAL TRANSFER FUNCTIONS 253

by choosing y(t), y(t), ... , yln -¡)(t) as state variables, a dynamical equation of
exactly the same form as (6-8) can be set up. However, if the right-hand side of
(6-46) consists of the derivatives of u, although it can still be realized into a
dynamical equation, the situation becomes very involved. lnstead of giving a
general formula, we give an example to iI1ustrate the procedure.

Example 4
Consider the following second-order time-varying differential equation
[p2 + IX¡ (t)p + IX2(t)]y(t) = [,80(t)p2 + ,8¡ (t)p + ,8 2(t)]U(t) (6-47)

The procedure of formulating a dynamical equation for (6-47) is as follows:


We first assume that (6-47) can be set into the form

1
-IX¡(t)
J[Xl] +
X2
[b¡(t)]
b 2(t) u
(6-48a)

y=f O ] x + e(t)u (6-48b)

I and then verify this by computing the unknown time functions b¡, b 2 , and e in
terms of the coefficients of (6-47). Differentiating (6-48b) and using (6-48a),
. gIs). Clearly, the con­ we obtain
rm realizations are the y= py = X2 +b¡(t)u(t) + é(t)u(t) + e(t)u(t)
t necessarily irreducible
ji = p2 y = -1Y.2X¡ -1Y.¡X2 +b 2u +b¡u +b¡u + eu + 2éú + eii
cible [N(s) and D(s) are
ov parameters and the Substituting these into (6-47) and equating the coefficients of u, u, and Ü, we
bservable no malter the obtain
:ment of computing the
e(t) = f3 o(t)
ore difficult to compute.
b ¡ (t) = ,8 ¡ (t) - IY.¡ (t),8 o(t) - 2~ o(t) (6-49)
;ed in References 23, 81
b 2(t) = ,8 it) - b ¡(t) - IX¡ (t )b¡ (t) - IX ¡(t)~o(t) ~ 1Y.2(t ),8o(t) - fio
g all realizations. The
due to pararneter varia­ Since the time functions b¡, b 2 and d can be solved from (6-49), we conclude that
transfer function g(s) is the differential equation (6-47) can be transformed into a dynamical equation
of the form in (6-48). We see that even for a second-order differential equation,
the relations bet ween !J, e. and lhe 1:J.¡'s and f3¡'s are verv complic'olled. tjj

or 2. We then realize 6-4 Realizations of Vector Proper Rational


Thc lirst one is called a Transfer Functions
111. This typc of realiza­
In this section realizations of vector proper ralional transfer functions will be
Crcncc S'+7). studied. Bya vector rational function we mean úther a 1 x por a q x 1 rational­
function matrix. Consider the q x 1 proper rational-function matrix
JI equations. Befo re
) of dynamical equations
nth-ordcr. \lnear, time­ (6-50)

J)u(t) (6-46)
=========",-=',='-=,."., .

REALIZI
254 IRREDUCIBLE REALIZAT10NS, STRicr SYSTEM EQUIVALENCE, AND IDENTIFICAT10N

Example 1
It is assumed that every g;(S) is irreducible. We first expand Ginto
Consider
l
(;(s) =
e ] rgl(S)]

e
t
eq
+ ~z:(S)
gq(s)
(6-51 )
A

G(s) =
(8 +s1)(s
+3
+2)j =
[ 5 +4
where e¡ =g;(oo), and gi(S)~g;(s) -e¡ is a strictly proper rational function. 5 +3
We compute the least common denominator of g¡, for i = 1,2, ... , q, say sn +
al s" - 1 + ... + a", and then express G(5) as

(6-52 )
Hence a minimal-dime

lt is c1aimed that the dynamical equation

o 1 O O O
O O 1 O O
x= x+ u (6-53a)
O O O 1 O
-Cf.
n -Cf. n- l -Cf. n- Z -Cf.
l 1 We study now the
Since its development i
the result. Consider th
YI]
~2
[Pln PI(n-l)
Ptn PZ\n-l)
Pll] [el]
P~l x+ e u
G(5) = [g'I(S) : g'z(5) : ..

[Yq
=
tq (6-53b)
=[el : ez:··· : el
Pqn Pq(n-l) Pq" e
= [el: ez : ... : el
is a realization of (6-52). This can be proved by using the controllable-form
realization of g(s) in (6-14). By comparing (6-53) with (6-26), we see that the
transfer function from u to y¡ is equal to

Then the dynamical eql


which is the ith component of G(s). This ilroves the assertion. Since g;(s) for Xl O O
i = 1, 2, ... , q are assumed to be irreducible, the degree of G(s) is equal to n. Xz 1 O
The dynamical equation (6-53) has dimension n; hence it is a minimal-dimen­ x3 O 1
sional realization of (;(5) in (6-52). We note that if some or all of g;(s) are not
irreducible, then (6~53) is not observable, although it remains to be control­
lable. x n O O
For single-input..single-output transfer functions, we have both controllable­ Y = [O O

form and the observable-fonn realizations. But for column ration'al functions
it is not possible to have the observable-form realization.
REALlZATIONS Of VECTOR PROPER RATIONAL TRANSfER FUNCTIONS 255
:, AND IDENTIFICATION

Example 1
and G into
Consider

(6-51 )
A

G(s)-
_

¡(S s +3 +2)
+ 1)(s
s +4
j =
[OJ
1
+
[s +3j
(s + 1)(s
1
+2)

-- --
per rational function. s+3 s+3
= 1,2, ... , q, say Sil +
OJ 1 [ (s+3f J
= [ 1 +(s+1)(s +2)(s +3) (s +1)(5 +2)

OJ 1 [S2 +6s +9J


: : .... 7pI2nn,] = [ 1 +s3+6s 2 +11s+6 s2+3s+2
(6-52)
Hence a minimal-dimensional realization of G(s) is given by
+ ... +Pqn 1
Xl] [OO
[ ~2 = O
X3 -6 -11
o
~
6
(6-54 )
,+ ° u
y= [ 3
(6-53a)
I
O
1 We study now the realizations of 1 x p proper rational-function matrices.
Since its development is similar to the one for the q x 1 case, we present only

,,]
e u
the result. Consider the 1 x p proper rational matrix
G(s) = [g'l (s) : g~(s) : ... : g'p(s)]

t
eq
(6-53b)
= [el: e2 : ... : e p] +[gl(S) : g2(S) : ... : gp(s)]

= [ e .. e .....
. 'e] +----.-----­ 1
1, 2· . p Sn + a 1 Sn 1 + ' .. + a n
the control1able-form
n sn -
X [ 1-'11- 1 n sn - 2 + . .. + 1-'1/1
+ 1-'12 n sn - 1 + 1-'22­
n .: 1-'21' n sn - 2 +. . .
(6-26), we see that the
+{J2n:"': {JpIS"-¡ +{Jp2 S 2 + .. - + (Jpn]
/l-

(6-55)
Then the dynamical eguation
;ertion. Since gas) for XI O O O -a n f3 In P2n Ppn
e of G(s) is egual to n. X2 1. O O -a n- I PI(n-l) P2(n-l) Pp(n-I)
it is a minimal-dimen­ X3 O 1 O -a n - ., x+ PI (n- 2) P2(n-2)
.~ Pp(n-2) U
e or al1 of g;(s) are not
emains to be control-
xn -al
lave both control1able­ y
°
= [O
O
O O
PII
] x + [el
P21
.. e2
PPI
ep ] u
lmn raÍional functions
(6-56)

¡
·,_._• • • • > __ • U _ •. •.• --.". __._. " . ~ _ ••• _~ , __ ~ _ . ' - . . _ •.• _ ••. ~ ••••
.-._-- -~ ..... _-----­ ..- _... -_ .. '- ..
~---
~._." ..

256 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION REALIZA

is a realization of (6-55). The realization is always observable whether 9;(S), Realization from the
i = 1, 2, ... , p, are irreducible or noto If they are all irreducible, then the realiza­ Consider a q x 1 prop<
tion is controllable as well.
A
It is also possible to find lordan-form realizations for vector proper rational 9
functions. The procedure is similar to the one for the scalar case. We use an
example to iIIustrate the procedure. A
G(s) = ª
"
[

'Example 2 g,

Find a lordan-form realization of the 2 x 1 rational function The Markov paramete


For each g¡(s), we form
composite matrix

r "¡(cx I -+

(6-57)
T~l:;~X
Note that ni, i = 1, 2, .
we draw in Figure 6-6 a block diagram of(6-57). With the state variables chosen number of rows. The i
as shown, we canobtain the lordan-form equation is equal to or lare:er t
{g¡{s), i = 1,2, ... , q} .

X 1J [- 1 1: 0] \01 . . . , q} is not available,


of the denominator of 9
[ ~: = - - ~ ~- +~~ x +t~-J ~
- _:- not be irreducible.
Now we shalI app
y=[ ~ -2 :
2:-IJ° [oJ
' x+
1
u independent rowsofT ii
T and Theorem 6-3, if Ol
T in (6-59), then all sub~
to describe the block diagram. The equation can be readily shownto be con­
(J¡ be the number of Iim
trollable and observable by using Theorem 5-21. Hence it is an irreducible
then the first (X¡ rows 01
lordan-form realization of (6-57).
However these linear inc
in T because they may 1
hav~ (Ji.:s;,C(j, i=2. 3.,
(J¡ =Cl¡. The row of H¡
rows ofT is called the pr
of Iinearly independent
Iinearly dependent row.
r------j-------c~ } - - - - Y¡
the row-searching algori
(x, +1

u --f-------.I k, =[a,,(I)a,,(2)"'a, ,(u,)


k,~ [a" (1 )a2' (2)' .. a2,(IT, )
-2 f-------.;.-{+ } - - - - - Y2

U ,

Figure 6-6 Block diagram of Equation (6-57).


REALIZATIONS OF VECTOR PROPER RATIONAL TRANSFER fUNCTIONS 257
E, AND IDENTIFICATION

Realization from the Hankel matrix


Jservable whether 17;(S),

.ucible, then the realiza- Consider a q x 1 proper rational matrix G(s) expanded as

r vector proper rational


scalar case. We use an

lction
¡
91(Sj
(;(s) = 92.(S)

gq(~)
=
[h1(0)+h1(l)S_1 +h 1(2)s-2

hq(O) +hq(l)s-l +h q(2)s-2

The Markov parameters h¡U) can be obtained recursively as in (6-29) and (6-30).
+"'J
hz{?) +h 2(l)s - 1 + h 2(2)s - 2 + ...

+... .
(6-58)

For each 9;(s), we form a Hankel matrix H¡ defined in (6-31). We then form the
composite matrix

[Hi~/+--( J}
~1_(~~ ~1: p)l) u 1 (no. of Iinearly independent rows)
T~ ~2S~~¿-_1_,p)ll U 2 (6-59)

- jJ) uq
(6-57)
Note that H¡, i = 1, 2, ... , q, have the same number of columns but different
number of rows. The integer IX; is the degree of the denominator of ?j¡(s) and f3
he state variables chosen
is equal to or laflzer th::ln the degree of the least cnmmon cip.nominator of
{9¡(S), i = 1,2, ... ,q}. If the least common denominator of t9¡{s), i = 1, 2, 3,
... , q} is not available, we may choose f3 to be equal to the sum of the degrees
of the denominator of 9¡(S), i = 1, 2, ... , q. Note that in this method, 9;(S) need
not be irreducible. .
Now we shall apply the row searching algorithm to search the linear
independent rows of T in order from top to boltom. Because of the structure of
T and Theorem 6-3, if one row in H¡ is Iinearly dependent on its previous rows of
T in (6-59), then all subsequent rows in H¡ will also be Iinearly dependent. Let
eadily shown to be con­ U¡ be the number of Iinearly independent rows in H¡. If all 17¡(s) are irreducible,
~nce it is an irreducible then the first IX¡ rows of H¡ will be Iinearly independent in H¡ (Theorem 6-3).
However these linear independent rows ofH¡ may not be alllinearly independent
in T because they may become dependent on the rows orH j for j < i; hence we
have (J¡:S;CI.¡, i=2, 3, ... " q. Note that if g1\S) is irreducible, then we do have
u 1 = 1X 1 · The row of H¡ which first becomes Iinearly dependent on its previous
rows of T is caBed the primar y Iinearly dependent row of H¡. If u ¡ is the num ber
of Iinearly independent rows in H¡, then the (u ¡ + 1)th row of H¡ is the primary
Iinearly dependent row. Corresponding to these q primary dependent rows,
the row-searching algorithm will yield
-Y1
<x, +1 (x, +I <X q +1

k'=[UlI(I)UlI(l)"'all(o-¡) 1: O ... O O ···0 O: : O "'0 O O]


k;=[u,,(I)a,,(l)"'a,,(o-¡) O: ad') "''',,(0-,) 1 ···0 Ó: : O ···0 O O]

k q =.[a q,(I)uql (l) "'aql(o-I) O :.lIq,(I) .. ··(/q,(o-,) O '" O O:


.' ..

'-------y---"" '----y------""

U¡ a2
(6-60)

258 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTlFICATION REALIZA"

such that kiT = O. Note that the row vector k j has, except element 1, only Ir (Ji = O for some i, say
((JI +(J2 + ... +(JJ possible nonzero elements. This is a consequence of the
1
row searching algorithm. The k i expresses the primary dependent row of H¡
as a unique linear combination of its previous linearly independent rows of O
T. See Appendix A. e= -(131(1)'" -(1,

Now we c1aim that (;(s) in (6-58) has the following irreducible realization: O
x = Ax + bu (6-61 a)
y = ex + eu (6-61 b) o
with
hl(O)j
h 2 (0)
e= . The dimension of this
[ the rank of H.
hq(O) The assertion that (1
by using the procedure
where, for i = 1, 2, ... , q,
forward, though tediow
O 1 O o
O O 1 o
A¡¡= ((Ji x (J¡) matrix
O O
-a¡¡(2) -a¡¡(3)

for i > j,

O O O

O O O

((J¡ x (JJ matrix

Let Cj be the ith row of (


ciAkb =h¡{.
This estabJishes, followiI
and, for i= 1, 2, ... , q,

b. =
I
h¡IU)
h¡I(2)

j We show next that t!
see that lb Ab .".
in (6-59); hence its rank
verified that the matrix
i

[
h¡I«(J¡)

The matrix e is given by, if (Ji 1=0, for i = 1, 2, ... , q,


1 O ... O: O O ... O: :O
c=
,
O O ... O' 1 o ... O""
"
'o O
O 0j}
O
•• •
1
1.. •
1
I
I
l.
·· .
.
q (6-62a)
[ :: : 1:: : I 1: · .

o o 0:0
,
o 0"
: : 1 O O

'-v---" ~ '-v---"

(JI (J2 (Jq is the identity matrix of (


the proof that (6-61) is an
--_.. _._._--------_._-----------­ --- ----

0, ANO IDENTIFICATION REALlZATIONS OF VECTOR PROPER RATlONAL TRANSFER FUNCTIONS 259

except element 1, only If (Ji =0 for sorne i, say i = 3 for convenience of il1ustration, then
s a consequence of the
y dependent row of H¡
o : O o O: ;00"'0
:O O
O :, 1 O :0 O .". O: :0 O O
Y independent rows of
-a31«J¡) :-a32(1) ... -ad(J2): O O ... O:,.. ':0
, O O
c= O O O :1 O "·0: :00'''0
~educible realization:

(6-61 a) .: ,,
(6-61b) o o o O :0 O .". O:
'--------.....--'

(J¡ (J2
h(0)J
1 (6-62b)

e=

l
h2 (0)

hq(O)
The dimension of this realization is n = (J 1 +(J 2 + .. , +(J q which is equa] to
the rank of H.
The assertion that (6-61) is a realization of G(s) in (6-58) can be established
by using the procedure used in the scalar case. Similar to (6-39), it is straight­
forward, though tedious, to verify
h¡(k + 1)

hl(k +2)

«J¡ x (J¡) matrix

k =0,1,2, ".. (6-63)

hq(k +(Jq)
Let c¡ be the ith row of C. Then we have
c¡Akb=h¡(k+\), k=O, \,2, ... ;i=\,2, ... ,q
This establishes, fol1owing (6-34), the assertion.
We show next that the realization in (6-61) is irreducible. From (6-63), we
see that [h Ah ... A" - 1 bJ consists of ali iinear independent rows of H
in (6-59); hence its rank is n and lA, b} is control1able. Similarly it can be
verified that the matrix

o
O 0J}"
"O
(6-62a)
·· .. q
· '.
O O cq A C1
q- \

~
(Jq is the identity matrix of order n; hence {A, C} is o~servable. This completes
the proof that (6-61) is an irreducible realization of G(s) in (6-58).
'-~ " '.,__ ~ _ ~ _ ~ _ ~ ~ _ ' _ ' ~ ~ _ ' ~ ' ~ ' _ O _ ' " - ~ "__ ~ _ , __
" _ ~ ~ , _ " '_'_~' __' _ " '__
. ' _ ' - ' ~ "~_' ~"~"_ -~_.,,~._ ".__,,'.'-0._ , ., - __.,__
~ ~~~~-',

260 ,,,,"U"" , RCWZA no"" """ ""CM ",U,"A"NCC ANO "''',UACA UON ReAWA I
Example 3
The leftmost matrix is t
Consider we have u I = 4 and u 2 '
using the formula in (F

Gs
()
=[.~':,:-~)2] =[s'J +s'~-JS'5
1 '\'''_,,'')+,,'6_,\'7+ S ' 8 _ ..
6
+5s' -7s' +9s· 8 -\1,,'9
7
+"'J k¡ = [O
k 2= [_1
J
O
_.1
9
'---'--­
s3(~+I) uI

We form the Hankel matrix. with 2¡ =4.2 2 =4. and fi =5. The elements Corres O::
an irreducible realizati~
:(f:
, O 1 1 - J
;0: 1 1 -,3 S o
:, 1:, I -,3 5 -7 O
'\ -,3 5 -7 9 X= O
'- '
O
T=[HdS.5)J= -,3 5
-7 9 -11
H 2 ( 5. S) - : ó: -- -O" - - 6--- -1­ - - -:..:: 1­ l'
3

y=[~
O O 1 -1 \
O 1 -1 1 -1
1 -1 1 -1 1 T~e (u I ~ 4)th and (u 1 +0
-1 1 -\ 1 -\ wlth the slgns reversed. T
If we apply the row-searching algorithm to T, we will fina\ly obtain of9 l(s)andthefirstu 2 Mi
depends only on u 1 and u
1 It is possible to searci
-1 1 order. First we expand (;
3 -\ 1
(;(s) = H(O
-S 3 -1 1
where HU) ha ve the same o
7 -5 3 2 1
--6---0- ---6--- -3¡~ - - O ex i are defined in (6-59). W

r
-1_____________
O O _l
3i> O O 1 _ ~(!~ ----~(~! - I
- ­1 - 1 O 6
J6 O O O H(2) H(3) - - - I
-\ I -1 1 O O O O
T=:
. :
,

1 -\ \ 1
2
O O O O O
l H(ex)
H(ex +1)
H(ex+1)
H(ex +2)
Ii
H
O O :Il 1 There are ex + 1 block rows i
O C[: O -3 } -4
8 U I =4 block row is associated with '
:}) O O \2 -24 ordering of rows, the Hankei
O O O -36 72 .(6-59). Now We shall apply 1
O O O O O mdependent roWS of t in ord
0- ---O ---0- ---6 --- m U2 = 1
block row is linearIy depende:
subsequent block rows are Ji¡
o o o o o
o o o o o
o o o o o 4S~e Ihe discussion in (he Subsecliol
Chaplcr 5. Sce also Schcmes I an(
o o o o o lrollabilily malrix in the second half
REALIZATIONS OF VECTOR PROPER RATIONAL TRANSFER FUNCTIONS 261
E, AND lDENT1FICAT10N

The leftmost matrix is the matrix F defined in (F-9). From the rightmost matrix,
we have (J 1 = 4 and (J 2 = 1. The fifth and seventh rOws of K can be obtained by
using the formula in (F-l1) as
k l =[ O O 1 2 ~:(~ :, O O O O OJ
k 2 =[-.l3 _19 -.1.-
18
_-.1.­
\8 O O ~:l} O O OJ
'- '-v--'
(JI ~4 (J2 =1
The elements :I: correspond to the primary linearly dependent rows. Hence
an irreducible realization of (;(s) is given by
=5,

1 -3
o 1 O O O o
O O 1 O' O O
3 5
5 -7 x= O O O 1 O x+ 1 u
O O -1 -2 O 1
7 9 1 - - - -[ - - - -1­ - - - 1- i - 0­
9 -11 "3 9 -18 18: O
1 - - -...= i ­ o o O: 0J
O : 1 x
1 1 O O
1 -1 The «(JI = 4)th and «(J 1 +(J2 = 5)th rows of A are taken from the k l and k 2 rows
-1 1 with the signs reversed. The vector b consists of the first (J 1 Markov parameters
1 -1 of 91(s) and the first (J 2 Markov parameter of 92(S). The form of e is fixed and
depends only on (J 1 and (J 2­
t finally obtain
It. is possible to search the Iinearly independent rows of T in a dilTerent
order. First we expand (;(s) as
(;(s)=H(O)+H(1)S-1 +H(2)s-2 +H(3)s-3 + ...
where HU) have the same order as (;(s). Let ct. = max {ct.h i = 1,2, ... , q }, \Where
ct.¡ are defined in (6-59). We then form the Hankel matrix

t =[n:~\ ---nm----~~l- --_:::- -~rrc ii --\ lon, block 'DW (&-64)

1 H(ct.)
H(ct.+l)
H(ct. + 1) H(o: +2)
H(ct.+2) H(0:+3) .,.
H(ct. + fi
H(ct.+f3)
-l)J
O
There are ct. + 1 block rows in T; each block has q rows. The ith row of every
~
~}:~ 1 block row is associated with i)¡(s). Except from having more rows and dilTerent
-]
-4
O 8 (Jl =4 ordering of rows, the Hankel matrix in (6-64) is basically the same as the one in
O 12 -24 (6-59). Now we shall apply the row searching algorithm to search the Iinearly
O -36 72· independent rows of t in order from top to bottom. Clearly, if the ith row of a
O O O block row is linearlydependent on its previous rows, then aH the ith rows in the

O
-6 --- -cF
(J2 = 1
subsequent block rows are. linearly dependent on their previous rows. 4 After

O O O
O O O 4See the discussion in the subsections on the controllability indices and observabilily indices in
O Chapter 5. See also Séhemes I ano 2 of lhe scarch of Iinearly inoependent columns of lhe. con­
O O
trollability malrix in the secano half of Section 7-2.
O O O
262 IRREDUCIBLE REALIZATlONS, STRICT SYSTEM EQUIVALENCE, AND lDENTlFICATION
REALIZ.

the completion of this search, we then rearrange the rows of tinto the form of
Example 4
(6-59) as

H¡(C( +1, [3)J] (no. oflinearly independent rows)


Consider the transfer-f

\'¡

T
__ H 2 (a + 1, [3)
(6-65) G(S)=[~J)-3 +G}
H q( a + 1, [3) J \' q We form í with Ct=I
Clearly the total numbers of independent rows of t in (6-64), t in (6-65), and independent rows as
T in (6-59) are aH the same. Let \'¡ be the number of linearly independent
rows of H¡(a + 1, [3) in (6-65). Then we have VI +V2 + ... +v q = (T¡ +(T2 + ... 1
+(Tq, where (T¡ are defined in (6-59). Note that generally we have (T¡ ¡'v¡, for sorne O 1
or all i. Clearly the primary linearly dependent row of H¡ in T is its (v¡ + l)th -1 4 1
row. The major difference between T in (6-59) and t in (6-65) is the orders in -1 2 O
which the linearly independent rows are searched. In 1', the linearly indepen­ 3 -8 -1
gerit rows are searched in the order of the first row of H¡, i = 1, 2, ... , q; the 6
1 -2 -1 3
secQnd row of H¡, i = 1, 2, ... , q, and so forth. For convenience of searching,
we search the rows of t in order from top to bottom and then rearrange it as -5 12 3 -12
T.. -1 2 1 -3
Now from the row searching algorithm and the rearrangement of the coef­ 7 - 16 -5 18
ficients of combinations according to the rearrangements from t to 1', we can 1 -2 -1 3
obtain, similar to (6-60),
~+I ~+I ~+I

. k, =[ulIll) u,,(2) u"I",) 1 o o: u,,(1) u"t",) o o .. · o:··· : (/,.i1~'--O:¡


k, =[u"ll) (/,,12) u"I",) o o o: (/,,(1) u,,(I',) I o ... o:··· : (/,.111'" ",.(",,) o o ... O]

k.=[u.¡(l) (/.,(2) ... u.,I",) o o ... o: u.,II) ... u.,I",) o o .. ' o: .. ·: (/,,,11)'" u".!".) 1 o .. · O]
'-v---- ~ ----v-'
\'1 \'} \'"
(6-66)
such that kiT = O. The major difference between (6-66) and (6-60) is that aJk),
for j > i, are generaHy different from zeros in (6-66). In other words, the (v ¡ + 1)th
row in H¡ in (6-65) depends not only on the.first V¡ rows of H¡ but also on the
first VI rows ofH 2, H 3 , . . . , H q in (6-65); whereas the «(T¡ +l)th row ofH¡ in
(6-59) depends only on the first ()¡ rows of H¡. Similar remarks apply to the
(Vi + l)th row ofH¡ in (6-65). Now we c1aim that the dynamical equation T~e rows of K correspondí
x=Ax +Bu y=Cx (6-67) tamed, by using the formul~

A¡2 ... ~l = [ 6 (
... = [-1
with A= .lA"
A 2¡ A 22
A"l
A 2q
Note that the
k2
k¡ row corres
(

Aq¡ Aq2 ... A qq ~nearly dependent 01', equivi


~2 row_corresponds to the p
e
where A¡i' B, and are given as in (6-61) and (6-62) with (J¡ replaced by V¡ and T and k¡ as
aJk) taken from (6-66). The proof of this statement is similar to the one in
(6-60); it is more tedious, however, because of the necessity of tracking the
rearrangements of coefficients. The proof does not involve any riew idea and
l'

I
r.~¡JT~[ -16
Lk2
will be omitted. Insteadwe give an example to iIIustrate the procedure.
=E, AND IDENTIFICATl0N REALIZATlON OF VECTOR PROPER RATIONAL TRANSFER FUNCTIONS 263

IWS of tinto the form of Example 4

Consider the transfer-function matrix (;(5) in Examp\e 3. We write


?endent rowS)
(6-65)

We form f with o: = max {4, 4} = 4 and f3 = 5, and then search its linearly
independent rows as
in (6-64), t in (6-65), and
: of linearly independent 1 O O 1 -3
t- ... +Vq=O"l +O"z + ... O 1 O O O 1-1
------------------
Iy we have"O"¡ f.v¡, for sorne -1 4 1 O 1 1 -3 5
of H- in T is its (v¡ + l)th
r in (6-65) is the orders in
-1
3
2
-8 -1
O 1
6 1
*
*
O O 1 -1
1 1 -3 5-7
1
------------------
ln t, the linearly indepen­
ofH¡,I= . 1, 2, ... , q', the 1 -2 -1 3 O 1 O 1 -1 1-1
------------------
convenience of searching, -5 12 3 -12 -1 O 1 1 -3 5 -7 9
n and then rearrange it as -1 2 1 -3 -1 O O 1 1 -1 1 -1 1
------------------
7 -16 -5 18 3 O O O 1 -3 5 -7 9-11
rearrangem~t of"the coef­
ments from T to T, we can 1 -2 -1 3 1 O O O O -1 1 -1 1 -1
O O (j) 1 -3
,,+1 O O O :I,-1
~)00'01 6-- -::(:---6--- -6 --- 4
) : . .. : (/ I q~ 1)' l/ 111\ "l/

) ': ."'
:. 1\)'"
(/21/
",,1",)
-
o o ..
.
01 O O O O ~2;
: .. ' : "',(1) .. , ",,11',) I o .' 01
(ti -- -0- - - - 6--- -0- - - - 6
O .~

O O O O O ~
1', (6-66) -------------------
O O O O O ~
;-66) and (6-60) is that aJk),
O O O O O
In otherwords, the (v 1 + l)lh -------------------
roWS ofH l but also on the O O O O O
~ the (O" 1 + 1)th row of H 1 in O O O O O
.imilar remarks apply to the The rows of K corresponding to the zero rows indicated by arrows can be 00­
le dynamical equation tained, by using the formula in (F-11), as
(6-67)
kl =[ 6 O: 4 -18 : -1 O : 1 O: O O]
kz = [-1 0:-1 3: O 1 : O O: O O]

:~:1 Note that the k l row corresponds to the first row of H(k) which first becomes
linearly dependent or, equivalent\y, the primary dependent row of H l , and the
kz row corresponds to the primary dependent row of Hz. Now we rearrange
'qq
f and k¡ as
62) with O"¡ replaced by vi a~d
ment is similar to the one ¡n
the necessity of tracking the
not involve any new idea and
[~Jt~[_~ _~ -~ :~; 0:0
O: O
-V~
;)
. ~ '--v--'
Hustrate the procedure. VI Vz
264 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION IRREDUCIBLE REALIZ,

, I
introduced in CAkB.

=[~J
,,O·, O 1 1 -3
i = 1, 2, ... , q, on digito
:0: 1 1 -3 5 which is q(r:t. + 1) x 13, is
, I

'1'
1- _1
-31 5 -7 x 13; however, in the se;
1 - 3 5 - 7 9 ~Primary dependent row, say the ith row ir
- 3 5 - 7 9 -11 row of H 1 block rows can be skir
x -:6:- -- -0- - - - ó----( -- i
-~ thaJ1 the T in (6-59), th
, I
independent rows in t
'0'.! O 1 -1 1
1-
on computer computat
O 1 - 1 1 -1 ~ Primary dependent We discuss now an i:
1 -1 1 - 1 1 I
row of Hz Consider
-1 1 -1 1 -1J
C(s) = [gl(S) gzl
Hence an irreducible realization of G(s) is given by where H(k) are 1 x p con
O 1 O O O O denominator of aH elem,
O O 1: O O O of g¡(s). Define fi = max
I

-6 -4
O
1
O
1: O 18 x+ 1 u
---------------------
O: O
1
1 -O
O:
I

O -3 ::>
TJ:g
y=[ 1 O O O lH(a
O O O 1 It is a la + 1) x fip matrix
The v1th and (v 1 +vz)th rows of A are taken from the k1 and kz rows with in order from top to bot
the signs reversed. The vector b consists of the first V1 and Vz Markov para­ independent and the (v +'
meters of gl and gz· The form of C is fixed and depends only on V1 and Vz. I Then there exist a(i), i' 1
[a(l) (
Remarks are in order regarding the use of the Hankel matrices in (6-59) 1t is claimed that the dyn¡

!
and (6-64) in the realizations. First we show that the v¡ in (6-65), which are

X~,
obtained by searching the linearly independent rows of T in order from top to
bottom, are the observability indices of any irreducible realization of G(s).
Indeed, from (5-64) and H(k) = CAk-1 B, we have

L- a(1) -(;

where Vk and U k are defined in (5-48) and (5-27). The postmultiplication of


U p - 1 on V~ operates only on the columns of V~, hence the linearly independent y= [ 1 o
rows of T in order from top to bottom are the same as the linearly independent is an irreducible realizatio
rows of V~ in order from top to bottom. Consequently, we conclude that the proof of this statement is s
Vil i = 1, 2, ... , q, are the observability indicesof any irreducible realization of
G(s). This can also be easily verified for the G(s) in Example 4.
Because Vi' i = 1,2, ... , q, are the observability indices, we have *6-5 Irreducible Re,
Matrices: Hankel ME
v~max{v;, i = 1, 2, ... , q} su = max{uiJ= 1,2, ... , q}
Hence the largest order of A¡¡ in (6-67) is smaller than'or equal to the one·in There are many approachl
(6-60). This implies that the data up to CAV+P+1B are used i~ the realization rational matrices. One app
in (6-67); whereas, the data up to CA"+P+ 1B are used in therealization in (6-60). apply the reduction proced
As discussed in Section 5-8, the larger (he power of A, the more errors may be reducible one. We discuss
IRREDUCIBLE REALIZATIONS OF PROPER RATIONAL MATRICES: HANKEL METHODS 265
1, AND IDENTIFICATION

introduced in CAkB. Hence it is more difficult to determine the correct (Ji,


i = 1, 2, ... , q, on digital computer computation. We note that the order of t,
which is q(ex + 1) x {3, is larger than the order of T, which is (ex l +rx2 + ... +ex q )
x {3; however, in the search of linearly independent rows of t, once a dependent
~Primary dependent row, say the ith row, in a block row is found, aH the ith rows in the subsequent
block rows can be skipped. Hence, although the order of t in (6-64) is larger
rOW ofH I
thaJ1 the T in (6-59), the highest power of CAkB used in the search of linearly
independent rows in t is smaller. As a result, it is easier to determine Vi than (Ji
on computer computation, and t should be used.

II
~ Primary dependen!
row ofH z
We discuss now an irreducible realization for a 1 x p proper ralional matrix.
Consider
G(s)= [91(S) 92(S) ... 9 p (S)]=H(0)+H(1)s-I+H(2)s-z+ ...
where H(k) are 1 x p constant matrices. Let ex be the degree ofthe least common
denominator of aH elements of G(s) and let {3i be the degree of the denominator
of 9¡(S). Define {3 =max{{3i' i = 1, 2, ... ,p}. We form the Hankel matrix
H(l)
H(2)
H(2)
H(3)
H({3)
H({3 + 1)
1
T=
f~(ex+1) H(ex +2) ~(ex + {3)
It i5 a (ex + 1) x f3p matrix. ; TOW we search the linearly independent rows of T
in order from top to bottom. 1t is assumed that the first V rows are linearly .
the k1 and kz rows with independent and the (v + l)t& row is linearly dependent on its previous rows.
VI and V 2 Markov para- Then there exist a(i), i' 1,2, , v, such that
Ids only on VI and V2' • [a(1) a(2) a(v) 1 O ... O]T =0

lankel matrices in (6-59)


le Vi in (6-65), which are
)f T in order from top to
;ible realization of G(s).
It is claimed that the dynamical equation

iJ ~ ~ ~ O
O
:
l l x +:
f"(l)
H(2)
u (6-68)
I O O O 1 ! ! H(v - 1)!
L- a (1) -a(2) -'l(3) -a(v)J LH(v) J
~he postmultiplication of y=[ 1 O O O ]x + H(O) u
: the linearly independent
is an irreducible realization of the 1 x p proper rational transfer matrix. The
; the linearly independent
proof of this statement is similar to the column case and is left as an exercise.
Itly, we conclude that the
irreducible realization of
xample4. *6-5 Irreducible Realizations of Proper Rational
dices, we have Matrices: Hankel Methods
= 1, 2, ... , q}
There are many approaches to find irreducible realizations for q x p proper.
an or equal to the one in ¡ rational matrices. One approach is to first finda reducible realization and then
lre used in the realization [.
apply the reduction procedure discussed in Section 5-8 to reduce itto an ir­
"
~t
in the realization in (6-60). reducible one. We discuss this approach first.
~, the more errors may be
266 IRREDUCIBLE REALlZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTlFICATlON IRREDUCIBLE REALI2

Given a q x p proper rational matr~x G(s) = 19¡/s)), if we first find an irre­ +R implies
ducible realization for every element g¡j(s) of G(s), and then combine them
together as in (4-61) or Figure 4-7, then the resulting realization is generally not
controllable and not observable. To reduce this realization requires the
application of the reduction procedure twice.", If we Jind the controllable
canonical-form realization for the ith column, G¡{s), of G(s), say
X¡ = A¡x¡ + b¡u¡ y¡ =C¡x¡ +e¡u¡ and
where A¡, b¡, C¡, and e¡ are of the form shown in (6-53), u¡ is the ith component sVm(s) = ­
of u and y¡ is the q x 1 output vector due to the input U¡, then the composite
These equations impl)
dynamical equation
(sm +
O

l}}l1'
Y = [C¡
A2

O
C2
fIJ and

Consider
C(sI-A)-¡B
A

is a realization of G(s). (Pro ve.) Because of the structure of A¡, b¡ it can be which becomes, after tl
readily verified that the realization is always controllable. It is however gener­
ally not observable. To reduce the realization to an irreducible one requires
C(sI -A)
the application of the reduction procedure only once. A

lto. is possible to obtain difTerent controllable realizations of a proper G(s).


~
A ,.,

Let G(s) = G(s)+G(oo), where G(s) is strictly proper. Let tItes) be the monic This shows that Equati
least common denominator of (;(s) and of the form of A and B, it is easy to
generally not observabJ
tItes) = sm +IX¡ sm - ¡ +1X 2 S m- 2 + ... +lX m (6-69) We note that the reé
Then we can write (;(s) as form realization for a sc
to find a generalization
1
G(s) = t{t(s) [R¡sm-¡ +R 2 sm- 2 + ... +RmJ
A
(6-70) A
G(,
where R¡ are q x p constant matrices. Let I p be the p x p unit matrix and Op where H(i) are q AX P co
be the p x p zero matrix. Then the dynamical equation denominator of C(s} an
we can show

OP]
Op [OP]
Op
x+: u (6-71 a)
H(m +i)= -1X¡H(m +

Op Op Ip Op
-lXm-¡I p -lX m_ 2 I p -1X1I p Ip This is a key equation ir
. Let {A, R, C, E} be a
R¡ J x+ G(oo)u (6-33),
(6-71 b)
G(s) = E +C(sI - A)-¡ R
is a realization of (;(s). To show this,itis sufficient to show C(sI -A)-¡ B = (;(s).
From (6-73) and (~-75),
Define V(s)"~ {sI '-'- A)-IB or (sI - A)V(s) = B. Ves) is a mp x p matrix. If
is arealization of G(s) in
we partition it as V'(s) = [V'l (s) Vi(s)'" V~(s)], where the prime denotes
the transpose and V¡(s) is a p x p matrix, then (sI - A)V(s) = B or sV(s) = AV(s)

¡ HI
IRREOUCIBLE REALIZATlONS OF PROPER RATlONAL MATRICES: HANKEL METHOOS 267
;, AND IDENTIFICATlON

+B implies
if we first find an irre­
Id then combine them SV I(S)=V 2(S)
lization is generally not sV 2(S) = V 3(S) = S2V I(s)
:alization requires the
: find the controllable
G(s), say
and
¡Ui

U¡ is the ith component sVm(s)=-<xmV¡(S)-O:"'_IV2(S)- ... -<xIV",(s)+I p


U¡, then the composite These equations imply
(sm +<XIS
m- 1 + ... +<xm)V1(s)=!/J(s)V¡(s)=I p
. _si-Ilp (6-72)
:-j[uJ and

Consider
V ,(s)- !/J(s) i=I,2, ... ,/1'l

bp p

e p] u C(sI - A( 1B = CV(s) = RmV I(s) + R m-1 Vis) + ... + R 1Vmis)

cture of A¡, b¡ it can be which becomes, after the substitution of (6-72),


,le. lt is however gener­
irreducible one requires C(sI _ ArlB = Rm+sRm _ 1 + ... + sm-I R 1 = G(s)
!/J(s)
:ations of a proper G(s).
This shows that Equation (6-71) is a realization of G(s). Because of the forms
Let lf¡(s) be the monic
of A and B, it is easy to verify that the reálization is controllable. lt is, however,
generally not observable.
(6-69) We note that the realization in (6-71) is a generalization ofthe controllable­
form realization for a scalar transfer function shown in (6-26). lt is also possible
to find a generalization ofthe one in (6-38). Let
(6-70) G(s) = H(O) + H(l)s- I + H(2)s-2 + ... (6-73)

p x p unit matrix and Op where H(i) are q AX P constant matrices. Let lf¡(s) be the monic least common
)ll
denominator of G(s) and of the forro shown in (6-69). Then similar to (6-30),
we can show

: }+[r}
ti I p lp
(6-71.1
This is a key equation in the following ~evelopment.
i = 1, 2, ...

Let {A, B, e, E} be a realization of G(s) in (6-73). Then we have, similar to


(6-74)

R¡ ] x+ (;(oo)u (6-33); .
(6-71b) (6-75)

.how C(sI -A( ¡B =G(s). Fro~ (6-73) and (6-75), we may conclucle, similar to (6-34), that {A, B, e, E}
is a mp x p matrix. Ir is a reaiization of G(s) in (6-73) if and only if E = H(O) and
where the prime denotes
)V(s) = B or sV(s) = AV(s) i=O,I,2, ... (6-76)
268 IRREDUCIBLE REAUZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION IRREDUCIBLE REAU¡

Now we claim that the dynamical equation Hankel matrices

Oq Iq Oq Oq H(1)
Oq O Iq Oq H(2)
x= X+ U (6-77a)
Oq Oq Oq Iq H(m -1)
-C(m1q -G(m_I1q -G(m-2 I q -G( 1 I q H(m)
y= [ I q O O

is a qm-dimensional realization of G(5).


O ] X+ H(O)u
lndeed. using (6-74), we can readily
(6-77b)
.nd T~[::
verify that
Since T and t consist
H(2)] [H(3)] HU + 1)] are q x p matrices, T ar
verified that
AB = ~(3) , AZB = ~(4) , .
.. "' A'B =
H(i+2)
: (6-78)
[
~(m + 1) H(m +2)
[
HU +m) and, in general,
Consequently we have CAiB = HU + 1). This establishes the assertion. The
observability matrix of (6-77) is the unit matrix of order qm; hence (6-77) is
Note that the left-uppe
always observable. lt is, however, generally not controllable.
lt is possible to introduce other either controllable or observable realiza­
k x !(!>k) constant ma
tions. Their introduction however would not introduce any new concept and of order k, and O is the t
can be removed from i\
will be skipped. Instead, we shall discuss in the following two methods which
wi11 yield directly irreducible realizations. Both m'ethods are based on the "HU + 1) = 1q.
Hankel matrices.
where the prime denot
conclude that lA = M,
Method 1. Sin~ular value decomposition. Consider the q xp proper realization of G(s) in (6­
rational matrix (;(s) given in (6-73). Let ¡Jt(s) =j" + IX I S",-1 + ... + IX", be the observable but not nece
least common denominator of all elements of (;(s). Define the qm x qm and
pm x pm matrices c=
A=
r
l
O lq O O l 18=
O Kq IE=
M~ (6-79a)
is a pm-dimensional re:
O
-:m1q
O
-IXm-Il q
O

-IX m - zl q
:. J
-IXll q
realization is a unit mat
realization however is g(
Now we shall use TI
O O O -IXml p an irreducible realizatio
Ip O O -IXm_Il p existence of qm x qm anc
N~ O Ip O -IXm-zl p (6-79b)

O O Ip -IXll p
where 1: = diag "Ir Al' A2,·
where 1" denotes the n x n unit matrix. We also define the following two roots of the eigenvalués
, ANO IDENTIFICATION
IRREDUCIBLE REALlZATIONS OF PROPER RATIONAL MATRICES: HANKEL METHODS 269

Hankel matrices

:J( 1)
~(2)
[ H(l)
T~ ~m
H(2)
H(3)
H(m)
H(m +1)
J (6-80)
u (6-77 a) H(m) H(m +1) Il(2m-l)
H(m -1)
H(m)
H(2) H(3) H(m +1)1
[(O)u (6-77b)
and T~ ~(3) H(4) H(m +2)
(6-81 )
[
(6-74), we can readily H(m + 1) H(m +2) H(2m)

Since T and T consist of m block raws and m block columns, and since H(i)

¡
are q x p matrices, T and tare of order qm x pm. Using (6-74), it can be readily

iB =
HU +
~U + 2)
0J (6-78)
verified that

T=MT=TN (6-82)
H(i +m) and, in general,

les the assertion. The i=O, 1,2, ... (6-83)


ier qm; hence (6-77) is Note that the left-upper-corner element of MiT = TN i is HU + 1). Let Ik,l be a
,llable. k x 1(1 > k) constant matrix of the form Ik,l = [I k OJ, where I k is the unit matrix
or observable realiza­ of order k, and O is t he k x (1 - k) zero matrix. Then the comer element H(i + 1)
e any new concept and can be removed from MiT = TN i as
ing two methods which
lods are based on the i=O, 1,2, ... (6-84)

where the prime denotes the transpose. From this equation and (6-76), we
conclude that ,,.{ A = M, B = TI'p,pm' e = I q .qm , E = H(Ol} is a qm-dimensional
nsider the q x p proper realization of G(s) in (6-73). Note that this realization is the one in (6-77) and is
l Slll - 1 + ... + rt. 1II be the
observable but not necessarily controllable. Similarly, the dynamical equation
)efine the qm x qm and e = I q . qmT = [H(1) H(2) ... H(m)]
A=N (6-85)
R= I'p,pm = [I p o O]
lE = IH(O)
A

(6-79a) is a pm-dimensional realization of (;(5). The controllability matrix of this


realization is a unit matrix; hence the realization is always controllable. The
realization however is generally not observable,
Now we shall use Theorem E-S, the singular value decomposition, to find
an irreducible realization directly fram T and t. Theorem E-S implies the
existence of qm x qm and pm x pm unitary matriCes K and L such that
(6-79b)
I:
T=K [ O 0J'
O L (6-86)

. where I: =diag P'l, Al, , .. , An}and Ai , i = 1,2, .' .. , n, are the positive square
efine the following two roots of the eigenvalues of T*T, where theasterisk stands for the complex
270 IRREDUCIBLE REALlZATlONS, STRICT SYSTEM EQUlVALENCE, AND IDENTlFICATlON IRREDUCIBLE REALl

conjugate transpose. Clearly n is the rank of T. Let K 1 denote the first n Consider
columns of K and L 1 denote the first n rows of L. Then we can write T as
A 2 =(Y+Tl
T= Kl~Ll = Kl~1/2~1/2Ll ~ YU (6-87) =(Y+Ml
=Y+MT
where ~1/2 =diag {JAl, JA2>"" JAn}, y ~Kl~1/2 is a qm x n matrix, and =Y+MT
U = ~1/2Ll is an n x pm matrix. Defines =Y+MM
and (6-88) =Y+M 21

Because of KfK 1 = 1". L¡Lf = 1" (see Problem E-7), we have Repeating the proces~

Y+Y =~-1/2K!Kl~1/2 = In (6-89)


Consider (6-84):
and UU+ =~1/2LIL1*~-1/2 = 1" (6-90)
HU + 1) =1q,qm MiTI'P..
With these preliminaries we can now establish the following theorem. = Iq,qmMiTT +
= Iq,qm TNiT+'
Theorem 6-4 = Iq,qm TT + TJ'~

Consider a q x p... proper rational matrix (;(s) expanded as (;(s) = L:::o H(i)s-i.
= Iq,qm YUU+,
'---v--' '---v-' '­
We form T and T as in (6-80) and (6-81), and factor T as T = YU, and pT = p y = C In
pU, where p denotes the rank, by using the singular value decomposition. =CAiB, i,
Then the {A, B, C, E} defined by
This shows that (6-91)
A=Y+TU+ (6-91 a ) The dimension of
B = first p columns of U = UI'p,pm (6-91b) Am-1B] and V'=[C'
e = first q rows of Y = Iq,qm y (6-91c) T=VÜ and'n=Tsmi
E=H(O) matrix, we conclude ¡
A

is an irreducible realization of G(s). I controllable and obser'

We give first a justification of (6-91) before giving a formal proof. We see The crux of this 1
from (5-65) that Y and U are, respectively, the observability matrix and the Numerically stable co
controllability matrix of {A, B, C}. Consequently, we have (6-91b) and (6-91c). Hence this realization (
Again from (5-65), we can easily verify that t = y AV. The pre- and postmulti­ The decomposition
plication of y+ and U+ to T = y AU and the use of (6-89) and (6~90) yield choose Y = K 1 and U =
y
'f and ~J def]ned in (6-8
(6-91a). Since the dimension of A is equaí io the rank ofT, the irreducibiíiiy of
(6-91a) follows from pY = pU =dim A.
Y*Y==
Now we give a formal proof of the theorem.
A realization whose e
Proof of Theorem 6-4 property V*y = UU* i
speaking, the signal tral
Define
state to the output are s
T+~U+Y+ (6-92) (see Problems 6-22 and
Then we have, by using (6-87), (6-89), and (6-90), In the application, it
model for a given systeJ
TT+T =VUU+Y+YU = YU =T (6-93) to this problem. For e)

5V+ is caBed the pseudoinverse ofV. See Reference 116.

IRREDUCIBLE REALIZATlONS OF PROPER RATIONAL MATRICES: HANKEL METHODS 271


AND IDENTlFICATION

Consider
K1 denote the first n
Ne can write T as A 2 =(Y+TU+)2 = (Y+MTU+f [ using (6-82)J
=(Y+MTU+)(Y+MTU+)
(6-87)
= Y+MTT+TNU+ [using (6-92) and (6-83)]
1 qm x n matrix, and =Y+MTNU+ [using (6-93)J
=Y+MMTU+ [using (6-83)J
=Y+M 2TU+
1 (2 (6-88)
Repeating the process we can show
,ave
Ai=Y+MiTU+ i = 1, 2,3, ...
(6-89)
Consider (6-84):
(6-90)
H(i + 1) = Iq,qmMíTI'p,pm
ling theorem. = Iq,qmMiTT+Tl'p,pm [using (6-93)J
= Iq,qm TNiT+Tl'p,pm [using (6-83)]
= Iq,qmTT+TNiT+Tl'p,pm [using (6-93)J
= 1 YUU+Y+MiTU+Y+YUI' [using (6-87), (6-92), and (6-83)J
lS G(s) = I;':o H(i)s- í. ~ '--v-' '-v-----''-v--'~
C II! Ai II! B

. =YU, and pT= pY =


=CAiB, i=O, 1,2, ...

value decomposition. A

This shows that (6-91) is a realization of (;(s).


(6-91 a) The dimen~ion of A is equal to n, the rank of T. Let Ü = [B AB ...
(6-91b)
Am-1BJ and V'=[C' A'C' ... (A')"'-lC'J. Then from (5-65), we have
(6-91 e) T = VÜ and n =T :o;min(pV, pÜ). 'Since Ü is a n x mp matrix and V is a mq x n
matrix, we conc1ude pÜ =n and pV =n. Hence the realization in (6-91) is
controllable and observable. This completes the proof of the theorem. Q.E.D.
I
The crux of this realization is the singularvalue decomposition of T.
formal proof. We see Numerically stable computer programs are available for this computation.
ability matrix and the Hence this realization can be readily carried out on a digital computer.
ave (6-91b) and (6-91c). The decomposition ofT = YU in (6-87) is not unique. For example, we may
ihe pre- and postmulti­ choose Y = K1 and U = :EL l , and the subsequent derivation still applies. The
(6-89) and (6-90) yield V and U defined in (6-87), however, have the property
fT, the irreducibility oí
y*y = :E 1/2 Kt K l :E 1/2 =:E = :El/2LlLt:El/2 = UU*

A realization whose controllability and observability matrices having the


property y*y = UU* is called an internally balanced realization. Roughly
speaking, the signal transfer effect from the input to the state and that from the
state to the output are similar or balanced in an internally balanced realization
(6-92) (see Problems 6-22 and 6-23).
In t)Je application, it is often desirable to flnd a simplifled or reduced-order
model fora given system. This realization procedurecan be directiy applied
(6-93) to this problem. For example, consider .
272 IRREDUCIBLE REALIZATlONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATlON
IRREDUCIBLE REALI

where 1: = diag {A¡, A2 , • • . An} and Al ;:O:A 2 ;:o: •.. ;:o: Ano If the system is to be bottom. After the Cl
approximated by an m-dimensional model with m < n, then the reduced model into the form
can be obtained from
H¡¡(a+1',B) ]
T = K I diag{A¡, A2 , . .. ,Am}L¡
T= ~2¡(a+1,,B) ]
where K¡ is the first m columns of K and L¡ is the first m rows ofL. See Refe­
rences S141 and S162. [
Hq¡(a +1, f3) ]
To conclude this subsection, we remark that the dimension ofthe irreducible
r~alization is equal to the rank of T. Becallse of Theorem 6-2, we have deg as in (6-65). Using th
G(s) = rank T. ronsequently, the degree of G(s) can also be computed from T. q primary linear1y de
Ifthe degree of Gis known a priori, then the m in (6-80) can be chosen to be the (6-67) can be readily
least integer such that the rank ofT is n. Because ofT = VU, from the definitions servability indices of ~
of controllability index /l and the observability index v, we may conclude that illustrate the procedu
the 1east integer to have pT = n is equal to
Example 1
max{/l, v}
Consider the proper r
Note that /l and v can also be ~ obtained directly from the column and row
degrees of coprime fractions of G(s). See Theorem 6-6 on page 284.

Method 11. Row Searching method. In this subsection we shall search


the linear independent rows of a Hankel matrix of G(s) '" to find an irreducible
realization. This method is a direct extension of the vector case discussed in
We compute the Mark
Section 6-4 and the reader is advised to review the section before proceeding.
Let (;(s) be a q x p proper ration¡l1 matrix. Let a¡ be the degree of the least -2s 2 -3s-2
common denominator of the ¡th row of (;(s) and {3j be the degree of the least (s + 1)2
common denominator of the jth column of (;(s). If their computations are
complicated, they may simply be chosen as the sums ofthe denominator degrees
of the ith row andjth column of (;(s). Define a = max {a¡, i = 1,2, ... , q} and
{3 =max{{3j,j = 1, 2, ... , p}. Similar to (6-59), we may form the Hankel matrix 4s+5
---=4
H¡¡(a¡ +1,{3¡) Hda¡ +1,{32) H¡p(a¡ +1,{3p)1}a¡(no.Of s +1
= H 2¡(a2 +.1, {3¡) H. 22 .(a 2 + 1, {32) H.2p(a 2 +.1, {3 p) }. a 2 linearly
and
-3s-5
T: : : independent s +2
[
H q¡(a q +I,f3¡) H q2 la q +i,{32) ... H qp(a q +l,{3p)J}a p rows)
(6-94)
It is clear that a¡ = 4, a

where Hij(a¡ + 1, {3j) is the Hankel matrix of g¡j(s) defined as in (6-35). Now
we search the linearly independent rows ofT in order from top to bottom. Then
corresponding to the primary linearly dependent rows of [H i ¡ H i2 •.• H ip ],
i = 1, 2, ... , q, we may obtain an irreducible realization similar to the one in
(6-61). Another way is to form the Hankel matrix

t = [:~~~ ::~; :~) +1 1] (6-95)

~(a+1) H(a+2) H(a+{3)


as in (6-64). We then search its linearly independent rows in order from top to
; .

ANO IDENTIFICATION IRREDUCIBLE REALIZAnONs OF PROPER RATIONAL MATRICES: HANKEL METHODS 273

If the system is to be bottom. After the completion of the search, we then rearrange the rows of T
en the reduced model into the form
Hll(O:+l,{3) Hdo:+1,{3) ... HIP(O:+l'{3)]}Vl~nO'OflinearlY
.. H 21 (0: + 1, {3) H 22 (0: +1, {3) .,. H 2io: + 1, {3) mdependent rows)
T= . . .
rowS ofL. See Refe-
[ : : : (6-96)
H q1 (0:+1,{3) Hdo:+1,{3) H qp(o:+l,{3) }v q
Ision ofthe irreducible
as in (6-65). Using the coefficients of linear combinations corresponding to the
rem 6-2, we have deg
q primary linearly dependent rows of T, an irreducible realization similar to
I be computed from T.
(6-67) can be readily obtained. Similar to (6-65), the Vi in (6-96) yield the ob­
an be chosen to be the
servability indices of any irreducible realization of G(s). We use an example to
U, from the definitions
illustrate the procedures.
we may conc1ude that
Example 1

l
Consider the proper rational matrix
the column and row -2S 2 -3s-2

l
)n page 284.
G(s) = (s + 1)2
S2 (6-97)
ection we shall search 4s +5 -3s-5
to find an irreducible s +1 s +2
~ctor case discussed in We compute the Markov parameters:
ion before proceeding.
the degree of the least -2s 2 -3s-2
the degree of the least (s + 1)2
heir computations are
e denominator degrees
;o:¡, i = 1, 2, ... , q} and
)rm the Hankel matrix 4s+5
- - = 4 +S-1_ S-2 +S-3 -S-4 + ...
p~l}
, {3 O" 1 (no. of
s +1
, {3 p) } 0"2 linearly -3s-5
and - - - = -3 +S-l -2s- 2 +4s- 3 -8s- 4 +16s- 5 - •..
independent s +2
Pp) J} O"p rows) It is clear that 0:1 =4, 0: 2 =2, {31 =2 and {32 =3. We form
(6-94)
T=[H l l (O:I +1,{31) HdO:l +1, {32)J
tled as in (6-35). Now
H 21 (0:2 +1, {3¡) Hd0:2 +1, {32)
:n top to bottom. Then
[H i1 H i2 ... H¡pJ, .,--f: -2 :0
I I
,
I 1 O
In similar to the one in
.:-2: , 3 , ,
,1 , O O
,
, 3: -4
,
:0 ,, O O
~~) + 1)]
:-4:
1 ___ I
5 :0
...
_~
, O O
,
5 -6 , O O O
(6-95) ------- - - - - ., ------- - - - - - - - ­
: 1: -1
---
.; 1: -2
1--­
4
[(o: +{3) -1 1 :-2 4 -8
ws in order from top to 1 -1 4 -8 16
........__ ... _=
.....=.....=.. ==

274 IRREDUCIBLE REALIZATIONS, STRICf SYSTEM EQUIVALENCE, AND IDENTIFICATION IRREDUCIBLE REAL!

We note that hJO) are not used in forming T. The application ofthe row-search­ Next we use í ar
ing algorithm to T yields
1

G(s) =
A [-2 0J
+ 4 -3
O 1

O O 1

O O .±
3
1

O O ~ 2 1 :T We hav~ IX =max{IX I ,
2-- f -- ----=-"4- - - -ó-- - -1- - ­
~ ~~ matrix T and apply t
3
42390 210 1
-8- - --=4---=- -17 - ~~ f9- ---O-- =--4- ---0- ---f 2 1
3

1 O-2 1 O O -1 1
-2
3 O
1
-4
3 O
O
O
O
}., ~4 -4
O
2
O
O
O
O 1. O O O 8 -4 O
3
O O O O O O O O
-------------------- -16 8 O
O O O O 4 } ~2 = 1
O O O
O O O O O
32 -16 O
O O O O O
(6-98) 1 O: -2
Clearly we haye 0"1 =4 and 0"2 = 1. Hence the dimension of any irreducible 1 1: -1 -L

realization of G(s) is 5. Corresponding to the fifth and seventh row ofthe right­ -~ 2--- -f -;-- 3--- -e
hand-side matrix of (6-98), we can compute, by using (F-l1), from the leftmost -1 -2: 1 4
matrix in (6-98): : --:3 --- 0- -;--=4 ---()
kIT=[O O 1 2 :I~: O O O]T=O 1 4 :'-1 -8
~ 4--- -ó -; -- 5----<i
and k 2 T=[0
~
O 1 1 O :2
'-v-J
()) O]T=O
---
-1 -8: 1 16 1 __ ­

Hence an irreducible realization of (;(s) is given by 5 O: -6 O


16 : - 1 - l ?
~ ~~:~. li -~ .~--,j
I - ".• '

O Corresponding to the 1

~----~--~~--=--~--l-~ _~~ ~
as shown, we can com]

x= x+ __ u kl = [O 2
O O -1 -1 :-2 1 1
k 2 = [O 2

Y= [ ~ ~ ~ ~ ~ \ ] x +[-~ _ ~ ] u
The elements (Ü corres
rearrange tinto t anc
TheO"lth and (0"1 +0"2)th rows of A are taken from k l and k 2 with the signs
reversed. The matrix B consists of the first O" I Markov parameters of 9I l (s) [~IJt~[O,O
2 .
O1((
O O O
and 9rz(s) and the first 0"2 Markov parameters of 921(S) and 922(S). The form of '-----v----'

e is fixed and depends only on O" I and O" 2' VI =3


.....= =.. _~~ =
.. _= _=.=_.. =--=
~ ~ ~=.=--~. =.=
.. =--=
~ ..=..=.= =._=._=
=. ~~,.~==.

ANO IDENTlFICATION IRREDUCIBLE REALlZATIONS OF PROPER RATIONAL MATRICES: HANKEL METHOOS 275

tion ofthe row-search­ Next we use t and t to find a different irreducible realization. We write

G(s) =[-2
4
0J +[1 0J
-3 1 1
S-l +[-2-1 -21J s -z

+[3 1
0J
4 s
-3 +[-4-1 -8
°JS-4+ ...
We have a=max{a 1 , az} =4 and p=max{Pb pz} =3. We form the Hankel
matrix t and apply the row-searching algorithm to yield
1
2 1
O -1 1

I~ ~ ~ }(J1
O O O
=4
-4
O
2
O
O 1
O -3 1
8 -4 O 3 O 1
lO O O
O O O 4 1 O 1
lO O O
l· - -- - ---- - ­ -16 8 O -7 O O O 1
10 O 4} ~z = 1 O O O -5 -1 O O O 1
10 O O 32 -16 O 15 O O O O O
O O O
(6-98) 1 O -2 1 3 O
3 1 -5 O 7 4
ion 'of any irreducible --------------------------
:venth row ofthe right­ -5 O 8 O -11 -4
1-11), from the leftmost 1 O -1 O 1 O
--------------------------
O O -1 O 2 O
O O O O O O~
=0 --------------------------
O O O O O O <­
=0 O O O O O O
--------------------------
O O O O O O
- Q 1}
0-'
1 Corresponding to the primary Iinearly dependent rows indicated by the arrows
O u as shown, we can compute, by using (F-1l), the coefficients of combinations as
O 2 O 1 1 O (n O O O] (6-99)
1 2 O 3 O 'I: O O O O]
The eleme~ts :X: c,?rrespond to theyrimary linearly dependent rows. Nowwe
rearrange Tinto T and rearrange k¡ accordingly to yield
and kz with the signs
v parameters of 911 (s) [kz~lJt~[O
- ,O
O 1 0.) 0;2 1 O 0J
O O O O: 2 3: f: O O
'-v------" ~
?

,nd gds). The form of
V1 =3 vz=2

1.
... ~:::":,:::,=----::..:'..:.,:",:,, _::...:...,. ::;::' ._--",_..... -::::...7"....:.:....:::.:=;:.• ..
, .~~, .. u·,,_,. _•.. ,_~.,.,'~""" .•• ~'.~ __ "_n-"'.~ __ ~·~ . :.:.:::::.':.:.._._J'~:"_":::':'~:"':'::=--'--....:'.:::._~:'::"":'::":'':':::::::-'':':':'''''''':~::'::::'':';;~:,:":,:_::::_::: ..:::.~:.::::::::~~..::-::::~~;-~~=----=---:::=:.::::::.::.:::.:..:::....::.o=~~:::.....:..::.::;::_:

. . .:.::~_ _~~.-

276 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUlVALENCE, AND IDENTIFICATION IRREDUCIBJ

:" -i -: - 2 3 : O: 1 O That is, n is the sum of al


:-2: 3 -4 : 1: O O
:__ ~_: -4 5, :0: O O
with
-4 5 -6 O O O ~

x ~:t-~{----~-~:---~--:~~---~ =[~J

I 1 I I

'-1'
1
1 1
-1 :'-2' 4 -8 1 ----1
'

1 -1 1: 4 -8 16 ~
- 1 1 - 1 :, - 8 16 - 32 O
1 -1 1: 16 - 32 64 H(s)= O

Hence a different irreducible realization of G(s) is


O O O
O 1 O: O O 1 O

O O 1: O O -2 1

x=
O O -1 :-2 -1 x+ 3 O
u

-----------~------- -------­
O 00:01 11 The matrix H(s) is a p x p
O O O: -2 -3 -1-2 on the diagonal. The m

y=[~ ~ ~ ~ ~Jx+[-~ _~Ju

responding column is a Z'


p x p and p x n constan
expressed as
The v¡th and (v¡ +v2)th rows of A are taken from k¡ and k 2 with the signs re­
versed. The matrix B consists of the first V¡ Markov parameters (excluding
hij(O)) of gIleS) and gds) and the first V2 Markov parameters of g2¡(S) and where N Ic is a q x n coos
g22(S). The form of e is fixed and the positions of 1 depend on V¡ and V2' The The procedure to be
observability indices of this realization are V¡ and V2 (Verify them.) lable canonical-form re;
difference is that the D(s
here. Let v be a p x 1 v,
*6-6 Irreducible Realizations of G(s): Coprime
Fraction Method
and
Controllable-form realization. Consider a q x p proper &rational matrix
~ ~ The substitution of (6- H
G(s). By long division of each element al :8(5)., 'He can "llfite rG(s; ~1.S

G(s) =G( (0) + G(s) and


where G(s) is strictly proper. Since G( (0) gi ves the direct transmission part of a Since D(s) is column re(
realization, we may exclude it in the following discussion. Using the procedure (6-105) can be written a~
discussed in Appendix G, we can factor G(s) as H(s
G(s)=N(s)D~¡(s) Let us define
where N(s) and D(s) are q x p and p x p polynomial matrices and are right
coprime. Furthermore, D(s) is column reduced and 0c¡N(s) < 0c¡D(s) = JJ.¡,
for i= 1, 2, ... , p.. Define .
6This step does not appear il
n=JJ.¡ +JJ.2 + ... +JJ. p rcalization.
CE, AND IDENTIFICATION IRREDUCIBLE REALIZATIONS OFG(S)-COPRIME FRACTION METROD 277

That is, n is the sum of al! the column degrees of D(s). Let us write
D(s) = DhcH(s) + DlcL(s)
with
,
, o :0
=[~J
s
,
, o
,
,
:0
,
,, :.
,
,
Sil! O O O Slll -1 , O , :0
O Sllz O O
_ _ _ _ _ _1 _ _ _ _ _

O
,
, 1
J
., L

:0
_

H(s) =O O Sil 3 O L(s) = ,


,
,
,
,
,
,
•••
, .
,
I :
(6-101 )
O
------------
,
Sllz-1 , :0
O O O Sllp
1 O ~---- _----­
....
O 'O
2 1
3 O u
------ O O
1 1
The matrix H(s) is a p x p diagonal matrix with S to the powers of column degrees
1 -2
on the diagonal. The matrix L(s) is an n x p matrix; if ¡..ti is zero, then the cor­
2 responding column is a zero column. The matrices D hc and D le are, respectively,
4 p x p and p x n constant matrices. Because of bciN(s) < bciD(s), N(s) can be
expressed as
and k2 with the signs re­
(6-102)
IV parameters (excluding
larameters of 921(S) and where N lc is a q x n constant matrix.
~pend on VI and V2' The The procedure to be discussed follows the one for developing the control­
Verify them.) lable canonical-form realization discussed from (6-21) to (6-26). The only
difference is that the D(s) in the scalar case is normalized to be monic, but not
here. Let v be a p x 1 variable defined by ves) = D- 1 (s)u(s). Then we have
lrime
D(s)V(s) = u(s) (6-103)
and YCs) = N(s)v(s) (6-104)
] proper rational matrix
" as The substitution of(6··100) and (6-102) into (6-103) and (6-104) yields
n write G(s)
(DhcH(s) + DlcL(s))v(s) = u(s) (6-105)
and y(s) = N1cL(s)v(s) (6-106)

~ct
transmission part of a Since D(s) is column reduced, the constant matrix D hc is nonsingular. Hence
m. Using the procedure (6-105) can be written as 6
H(s)v(s) = - Dh;, 1 DfcL(s)v(s) + Dh;, 10(S) (6-107)

Let us define
matrices and are right x(s) = L(s)v(s) '. (6-108)
nd ¿¡ciN(s) <¿¡ciD(S) = ¡..ti,
6This step does not appear in the scalar case because D(s) i~ normalized to be ~~nic beFare' ;he
realization.

1
- ... - ...
-.--------­
~ , ~.-­

278 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION IRREDl

We express this explicitly in the time domain:


X 11 V1

x12 ¡j1

V<¡" -1)
Xl!"
------­
XZ1 Vz
= : .(6-109)
V~2 -1)
Xz!' 2
-------
-------
X p1 Vp

V(!'p-1)
XP!'p p

where Vi is the ith component of v, and V\k) = dkVi/dt. From this definition we
have
i = 1, 2, ... , p; k = 1,2, ... , Jli-1 (6-110)

From the definition of H(s) and (6-107), we have


a1
X:"
Xl!',] ~ - ni;,n ,x(t) + ni; 'u(t) ~ r- !"] {b1!"]
~ ~'" x(t) b~" u(t) (6-111 )

r xP!'p

where al!'; denotes the ¡th row of Dh;, 1 D re , and bl!'i denotes the ith row of Dh;, 1.
ap!'p bp!,p
Figure 6-7 Block diagn
A block diagram of(6-11O) and (6-111) is shown in Figure 6-7. It consists of p
chains of integrators. The output of each integrator is chosen as a state variable
as shown. This block diagram is a generalization of the one in Figure 6-3.
From this block diagram or, equivalently, from (6-110) and (6-111), we can
where "x" denotes poss
readily obtain the following dynamical equation.
-al!", the (Jl1 + Jlz)th
O
o
1 O
O ]
O
O ,'
o o
l 10
·0 1 (6-112b) follows directl
We note that if P'i = 0, t;
, 6-7 and the ¡th row of ]J
,
O O O , O
x x x x :x
(-a,",) X ... X: :X x x ... x b,",
- - - - - - - - - - - - -: -0- - I - -0- - - e - - .-. : - ó: ------------­ -0­ - Example 1
:O O 1 O: , O Consider the strictly pro
o l'
I .
:

l ...
I
I
I o x(t) + u(t)
'o O O 1" O -6s-2
X
- - -
wX
- -
X
-. - -
...
- - -
x:
- -4 -
X
- -
...
-
(-a1¡t) ... x: l·
:X
1- - -
X
- -
X
-: - - - - -
X
­ _~1"_' A
_3S2

(S +1)3 (S-;
\J(S) =
-------------1---------------- I :0- -1- - 0- - - - -O ---
, O [
. S
, (s-~
,
, ,'o O 1 O O (s+I)3
o ,
, O , ,:
= [(S + 1)03(S - 2)
, , :0 O O 1
, , (s +1
xpp. X X X X : ..x ( -ap"p) x, ,x x x x bp "
(6-112a)
ANO IDENTlFICATlON
IRREDUCIBLE REALIZATIONS OF G(S)-COPRIME FRACTlON METHOD 279

XllI100'0-1~SI
xII
_~ - (/11 integrators)

'00 -~ (/12 integrators)

(6-109) u

'00-'"
[}Y­
1.
s
PI

)
(/1 p integrators)

~rom this definition we

.. , Jl¡-1 (6-110) y

I'!]
l
1
b
(t) + b~l'l u(t) (6-111)

bpl'p
.. -1
Figure 6-7 Block diagramofG(s)=N(s)D-1(s).
)tes the ith row of Ohc .
ure 6-7. It consists of p
hosen as a state variable
le one in Figure 6-3. y(t)= Cx(t) = N1cx(t) (6-112b)
110) and (6-111), we can where "x" denotes possible nonzero elements. The Jllth row of A is equal to
- 3 11',' the (JlI + Jl2)th row of A is equal to - 3 21'2 and so forth. Equation
(6-112b) follows directly from (6-106) by the substitution of x(s) =L(s)v(s).
1
We note that if Jli =0, then the ith chaLa of integrators wíll not appear in Fig.
o 6-7 and the ith row of O¡;;,IOIe and O¡;;,I wil1 not appear in Equation (6-112).

x X .. , X Example 1
----------
Consider the stricdy proper rational matrix
o(t)
o +

x .~ .~
x(t)

A
_3S 2 -6S-2

(s + 1)' (s -
s' -3.1-1
2)(.1 + 1)' (s -
1
2)(.1 + 1)2
J
",(s) =
.:-----­ [

s
100 O (s + 1)3 (s -2)(.1 +1)'· .: (s -2);.1 +1)2

O 1 O O

.
=.[(.1 + 1)'(.1 -
ü·
2)
(s
O ..
+ 1)'(.1 - 2)
J-'[(
.
~3S2 -6.1 :-2)(.1 -2)
. .1(.1 - 2)
s3.-:3s-1 s +1 ]

s(s + 1)

O O 1
x x x
(6-113)
(6-112a)
-------'- ' . . • ~ . .""c"'
•....:.:c." •. ~ - - - "' .. " " "" _ --.:=
"=:.""=.,,=::. . .

280 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTlFICATlON


IRRED

This fraction is obtained by using the least common denominator of each is a real ization of G(
row and is generally not irreducible. From the coefficients of this fraction, showing that the A, ]
we can obtain, by using the procedure discussed in Appendix G, the following
C(sI -A)-I]
irreducible fraction:
or equivalently,

~Jr +3<t'+1
-1 -l -1
~
G(s) =
[-3S 2
-6s-2 1
s-2 -3
(sI - A)
s O
O J
1
This equation implie~
BD(s)
Bence we have /11 = 3, /12 = 1, and /13 = O and From (6-112), wesee Ü
n=/11 +/12 +/13=4 are zero rows. Furth
(I~= 1 /1m)th row of
We write O(s) and N(s) in the form of(6-100) and (6-102) as

is eq ual to the ith row


O0J
s O +[1O are all zero rows. To
O 1 O s -1
O s

-2 -6 O O
N(s) = [ O -x -x
1

(si ­ A)L(s) =
Note that, beca use of /13 =0, the last colun;m of L(s) is a zero column. We
compute
-x -x

3
O
O
3
O
o
3 -2
O
O
-ll O
L-x -x

Hence a realization of (;(s) is given by

(6-114a)

-2 -6 -3
y= [ \6-114b)
O 1 O
Note that, because of /13 = O, the third row of O~ 1 and Ol~ 1Olcdo nbt appear in
(6-114).
The dynamical equation in (6-112) is developed from the block diagram in
Figure 6-7 which, in turn, is developed from (;(s)=N(S)O-I(S); hence(6-112)
, AND IDENTIFICATION IRREDUCIBLE REALIZATIONS OFG(S)-COPRIME FRACTION METHOD 281

denominator of each is a realization of G(s) = N(s)D -I(S). We shall now formally establish this by
cients af this fraction,
showing that the A, B, and e in (6-112) satisfies
endix G, the fallowing C(sI - A)-I B = N1c(sI - A)-IB = N(s)D-I(s) = N¡J,(s)D- I(S)

or eguivalently,

-1
s-2
-IJ-l
-3
(sI - A) - lB
This eguation implies

= L(s)D - I(S) = L(s)(DhcH(s) + D¡cL(s))- 1 (6-115)

O 1
BD(s) = BDh/H(s) + D¡;/ D¡cL(s~J =(sl - A)L{s) (6-116)
From (6-112), we see that all rows om, except the {L~ = 1 J.!m)th rows, i = 1, 2, ... , p,
are zero rows. Furthermore, these nonzero rows are egua! to D~ 1. Hence the
(L~ = 1 t'm)th row of
!) as BDhc(H{s) +Dh;, 1 D¡~{s))
is egua! to the ith row ofH(s) + Dh;,1 D1cL(s) for i = 1,2, ... , p, and the remainder
are all zero rows. To establish the eguality in (6-116), we write
o o: ,
s
o
-\
-\ o:
• I
.,
,
I

: I ,
, ,
o O o ... -1 : , ,
, , ,
~X -x -x S-XI -x -x o" -XI l-X -x ... -x
- - - - - ~ - - - - - - - - - - - - - - - -:- - - ~ - - =-.- f ------- 0-: :- -- - - - - - - - - - ­
I • ,
, O
, 0 , ,,
(sI - A)L(s) = '
I : I I
, I
is a zera column. We o o -1: :
-
-x
- - -
-x
.. - - - -
-x .,.
- - - - -. - -
-x I -x
- - - - -1- - - - -
-x
- - - - - - - - - -
s-X:
-
:-x
- - •• - - - - -
-x ...
_.- - - - - -
-x
- -­
. .
- - - - - - - - - - - - - - - - - - - - - -:- - - - - - - - - - - - - - - - - -: :- -5- - --:.! - -.-.-. - - --o
:
,
: : o
1 I •
O
I I I :

-ll
I ,
-x -x -x ... -x' -x -x ... -x' I-X -x s-x
3
3
O O -2 I \
, <;
O
~ l
O 0, I

S/JI-I: O:
--o-~---(-:
, ,
o , I S
,
I

(6-114a) x ,
I :
. ,
I

,
I ,
,
O : S,J¡-l: O

\6-114b)

D hc 1 D lc do oot appear in o ' O ,


I
S
I .

)m the block diagram in o O


~(S)D-l(S); hence (6-112) (6-117)
I
I
·282 IRREDUCIBLE REALlZATlONS, STRICf SYSTEM EQUIVALENCE, AND IDENTlFlCATlON IRREDUC I
I
,

From this equation, we can readily verify that all rows, except the eL¡
m= 1 J.!m)th Proof
row, for i = 1, 2, ... , p, of (sI - A)L(s) are zero rows. Because the (¿~ = 1 J.!m)th We first show that if
row of A is equal to the ith row of - D hc 1 D le, as can be seen from (6-111), the
observable. Under t:
(¿~=1 J.!m)th row of (sI-A)L(s)=sL(s)-AL(s) is equal to the ith row of
det D(s) such that G(~ I
H(s) + D,;/ DlcL(s). This establishes the equality in (6-116). Hence the dy­
find a realization of G
namical equation in (6-112) is a realization of (;(s) = N(s)D- 1 (s).
(6-112) is not irreducn
We discuss now the relationship between det(sI -A) and det D(s). From
ever, (6-112) is known
(sI - A)-IB = L(s)D-l(S) in (6-115), we have
shalI establish this st;

(s~ _ A) (Adj (sI - ~(s) L(s) Adj D(s)


D(s) and N(s) are not
det A))B = det such that
where "det" stands for the determant and "Adj" the adjoint. Since (Adj (sI - A))B
and L(s) Adj D(s) are polynomial matrices and since det (sI - A) and det D(s) are
polynomial of degree n, we must have
which implies the exisl
det(sI - A) = k det D(s)
for some constant k. Det (sI -A) has a leading coefficient of 1, and det D(s) and
has a leading coefficient of det Dhc> hence we have k = (det D hJ- 1 and
Because of (6-116), (6-1
det(sI - A) = (det D hJ- 1 det D(s) (6-118)

Realization of N(s)D-l(S) where D(s) and N(s) are not right coprime
or
The realization procedure discussed from (6-100) to (6-112) is applicable
whether G(s) = N(s)D- 1 (s) is a coprime fraction or not as long as D(s) is column Since L(A) consists of a
reduced and (\¡N(s) < (\¡D(s). Now we shall show that the dynamical equation is the n x 1 vector L(A)a
in (6-112) is controllable whether D(s) and N(s) ;lre right coprime or not. This does not have a full ran
can be established by employing the condition rank [B AB ... An - 1 B] = Now we show that
n or the left coprimeness of (sI - A) and B. For this problem, it is easier to right coprime. Under 1
employ the latter. From (6-115), we have on page 231, an eigenva
6-119)

Since L(s) consists of, as can be seen from (6-101), a unit matrix of order p as a
submatrix, we have Now we show that t.hen

rank [~~;)J =p for every s in iC (6-120)


To show this, consider ((
Hence D(s) and L(s) are right coprime (Theorem G-S). From the realization
procedure, we have dim A = deg det(sI - A) = deg det D(s). Hence Corollary
G-S implies that sI - A and B are left coprime or rank [sI - A B] = n, for
every s in iC. ConsequentIy, {A, B} is controIlable (Theorem 5-7). Up to this The composite matrix
point, the coprimeness of D(s)and N(s) has not been used. Theorem 2-5 implies

rank [AI-A
Theorem 6-5
which implies, because {,
The controIlable realization of (;(s) = N(s)D - I(S) in (6-112) is observable if and
only if D(s) and N(s) are right coprime.
IRREDUCIBLE REALlZATIONS OF G(S}---COPRIME FRACTlON METHOD 283
, AND IDENTlFICATlON

Proof
(cept the (I. (11= 1 JLm)th
i

:cause the (I~= 1 JLm)th We first show that if O(s) and N(s) are not right coprime, then (6-112) is not
•seen from (6·1-11), the observable. Under the premise, we can find a 6(s) with deg det 6(s) < deg
lal to the ith row of det D(s) such that (;(s) = Ñ(s)D - 1 (s) = N(S)O-I(S). Usin~ Ñ(s)D -1(S), we can
-116). Hence the dy­ find a realization of (;(s) with a dimension smaller than that of (6-112). Hence
5)0-1(S). (6-112) is not irreducible (controllable and observable) (Theorem 5-19). How­
~) and det D(s). From ever, (6-112) is known to be controllable; hence (6-112) is not observable. We
shall establish this statement once again by using a different argumen1. If
O(s) and N(s) are not right coprime, then there exists at least one s, say s = A,
s) L(s) Adj O(s) such that

11. Since(Adj(sl-A))B O(A)J


rank [ N(A) <p
51- A) and det O(s) are
which implies the existence of a p x 1 nonzero constant vector lX such that
O(A)lX =0
:ient of 1, and det O(s) ud N~~=O
let Dhcr 1 and Because of (6-116), (6-102) and (6-112b), we have
(6-118) (AI-A)L(A)lX=O CL(A)lX=O

lot right coprime or [Al ~AJ L(A)lX=O (6-121 )

I (6-112) is applicable
s long as D(s) is column Since L(A) consists of a unit matrix of order p as a submatrix, if lX is nonzero, so
the dynamical equation is the n x 1 vector L(A)lX. Hence we conclude from (6-121) that [(sI-A)' CJ'
t coprime or no1. This does not have a full rank at s =A, and {A, C} is not observable.
: AB .,. An- 1 BJ = Now we show that if {A, C} is not observable, then D(s) and N(s) are oot
problem, it is easier to right coprime. Under the premise, there exists, from the dual of Problem 5-37
on page 231, an eigenvalue Aof A and its n x 1 eigenvector e such that
6-119)
[AI~AJe=o (6-122)
.t matrix of order p as a
Now we show that there exists a p x 1 vector el! such that
e = L(A)lX and O(A)lX=O (6-123)
(6-120)
To show this, consider (6-116) or
. From the realization
[AI-A L(A)J (6-124)
O(s). Hence Corollary -BJ [ D(A) =0
nk[sl-A BJ=n,for
leorem 5-7). Up to this The composite matrix [H-A - BJ is of dimension n x (n + p); hence
sed. Theorem 2-5 implies

rank [AI-A -BJ +nullitY[Ai-A -BJ=n +p

which implies, because {A, B} is cootrollable,


112) is obserVable if and
nullity[..Ü - A -BJ=p
·_.~_. - .- '.~- ... ~- '. " .... . " .... "".'.' .·_a'.

284 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTIFlCATION IRREDU(

Consequently, the p linear independent columns of [V(A) D' (},)]' form a Proof
basis of the null space of [Al - A - BJ. Because of
Controllability indice
similarity transformat
[U-A
singular matrix. eOll
properties of (;(s). He
the vector [e' O']' is in the null space. Hence there exists a p x 1 vector IX
theorem. Ir we rearra
such that

[OeJ = [L(A)
0(,,1.)
J
IX
where D hc is nonsingl
becomes
This establishes (6-123). The substitution of e = L(A)ex into (6-122) yields (6-121)
"

~
o
which implies, by using (6-116), (6-102), and (6-112b),
O .
8'=
O(A)J ex = O
[ N(A)
O
Hence O(s) and N(s) are not coprime. This completes the proof ofthis theorem.
Q.E.D. l'
where the prime denote~
Theorem 6-5 implies that the realization in (6-112) is irreducible ifand only readily show that the (
if (;(s) = N(S)O-l(S) is a right-coprime fraction. Ir (;(s) = N(s)D-1(s) is a right­ column degrees of D(s).
coprime fraction, then we have, according to Definition 6-1 ,
Observable-Form RI
deg (;(s)=deg det O(s) realization discussed ir
observable form realiza1
Consequently, the realization in (6-112) is irreducible if and only if deg (;(s) = extensively in the next s
dim A. Since all irreducible realizations of (;(s) are equivalent (Theorem 5-20), Consider a strictly p
we conclude that a realization of (;(s) is irreducible if and only if its dimension is
equal to the degree of (;(s). This establishes essentially Theorem 6-2.
The realization in (6-112) is a generalization of the controllable canonical­ where the polynomial Ir
form realization discussed in (6-26). Hence we shall call (6-112) a multivariable reduced. We take the tl
controllable-form realization. We note that the realization depends on the
fraction N(S)O-l(S) used. For example, ifwe use (N(s)U(s»(O(s)U(S»-l, where
U(s) is a unimodular matrix, and D(s)U(s) remains íO be cnturrn reducr;c1 th,,;'-,
we will obtain a different multivariable controllable-form realization. Hence, Furthermore, N'(s)and [
unlike the scalar case, the controllable-form realization in the multivariable
case is not unique. Consequently, the adjective "canonical" is oot used.
of the form in (6-112) ca
Column degrees and controllability indices equation

Before proceeding, we give a relationship between column degrees and control­


lability indices.
is an irreducible realizati
multivariable obser~able
Theorem 6-6 Of course, an observa
The set of the controllability in dices of any irreducible realization of a strictly (6-125). Let the row d~gI
let
proper rational matrix G(s) is equal to the set of the column degrees of O(s) in
any column-reduced right coprime fraction (;(s) = N(S)O-l(S).
....=---=
....=.-c=
=====o===.=, .._-=_
...=....._=
.... =·_=_
....=··.. =.-~. .=....
_ _.=-_
.. =.. ==e-=··.. c,~ .. ===-

IRREDUCIBLE REALlZATlONS OF G(S)-COPRIME FRAcnoN METHOD 285


, AND IDENTIFlCATION

Proof
:V(}.) D'(A)]' form a
Controllability indices are, as shown in Theorem 5-8, invariant under any
similarity transformation and any rearrangement of input vectors by a non­
singular matrix. Column degrees are, as discussed in Theorem G-15', intrinsic
properties of C(s). Hence we may use the realization in (6-112) to establish this
~xists a p x 1 vector a. theorem. If we rearrange the input vector u(t) by
o(t) = Dhcu(t)
where Ohc is nonsingular by assumption, then the new B matrix in (6-112)
becomes
to (6-122) yields (6-121) O"-01:0"'00: :O"'OOj

O O O: O O 1: : O .. , O O

'8'=
r

I ' ••• 1
: : I : : I 1: :

• • I • • I l· ••

0'''00:0'''00::0 01
'---v-----' ~ '---v---'
e proof of this theorem. J1.1 J1.1 J1.p
Q.E.D.
where the prime denotes the transpose_ Using this Üand the A in (6-112), we can
readily show that the controllability indices are J1.l> J1.1' ... , J1. p, which are the
; irreducible if and only column degrees of D(s). This establishes the theorem. Q.E.D.
== N(S)0-1(S) is a right­
6-] , Observable-Form Realization. Similar to the observable canonical-form
realization discussed in (6-18) for scalar transfer functions, we can develop
observable form realizations for transfer matrices. This realization will be used
and only if deg C(s) == extensively in the next section.
ivalent (Theorem 5-20), Consider a strictly proper q x p rational matrix (;(s). We factor it as
i only ifits dimension is C(s) = 0-1(s)N(s) (6-125)
Theorem 6-2.
controllable canonical­ where the polynomial matrices O(s) and N(s) are left coprime and D(s) is row
1 (6-112) a multivariable reduced. We take the transpose of C(s):
zation depends on the C'(s) == N'(s)(D- 1(s))' = N'(s)(D'(s))-1 (6-126)
LJ(s))(D(s)U(S))-I, where
e column reduced. then The rOVlS 01 D(s) beco me the columns oÍ [}J'(s). tlence D'(s) 1S C01UIIlll (¿;clUCé>:L

rm realization. Hence, Furthermore, N'(s) and D'(s) are right coprime. Hence an irreducible realization
)O in the multivariable
z=Az +Bv w=Cz
lical" is not used.
of the form in (6-112) can be developed for N'(s)(D'(s))-1. Consequently, the
equation
x=A'x +C'u y=B'x
mn degrees and control­
is an irreducible Tealization of D -1(s)N(s), and the equation is said to be in a
multivariable observable form_
af course, an observable form realization can also be obtained directly from
(6-125). Let the row degrees of D(s) be Vi, that iS,b,.¡O(s) = Vi, i = 1, 2, ... , q, and
~ realization bf a strictly let
)Iumn degrees of D(s) in
)0-1(S). (6-127)
"'_' .... __ __._ _ .... _.• ,. __ .. .... __ c __

------------

~
~
~
~
.
"
'
_
n
_
.
_
.
~
~
_
~
_
~
_
~
.
~
.
_
.
286 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION

We write C(sI
D(s) = H(s)D h,. + L(s)D¡,. = (H(s) + L(s)D¡,.D,;;.1 )D h ,. (6-128)
which implies
where
SVI O or
_
H(s) =
l O
~
O

This is a key equation


~quation (6-131) will t
G(s).
Dual to a multiv:

l
1 s ... SVI - 1 : O : : O
_ ------O-------:- i --S- - -. ~ ~ - ;;2:'--1 - -: :- - - - - - - - Ó- - - - - - ­ observable-form realiza
L(s) = - - - - - -:- - - - - - - -' - - - - - - -. ~ - - - - - - - -~ '- - - - - - - - - - - - - - - ­ whether D(s) and N(s) ~
.
. . . '. then the realization is e
------O-------;- ----- -'0 --------: :--1- -5- - .- .-. - - ~ vq =- í ­

(6-129)
*6-7 Polynomial
and N(s) = L(s)N¡c (6-130)

Note that D h ,. is a q x q matrix and is nonsingular, for D(s) is row reduced. The Consider the network si
matrix L(s) is a q x n matrix. DI,. and D¡rD';;' 1 are n x q matrices. Thus a and inductor currents
realization of (6-125) is given by Section 3-4, then a d}
network. There are, 1
O O O ,
x: .x', x matical equations to de~
1 O O x', x , x currents ~1(t) and ~2(t) a
O 1 O x: O x O x
. ,
,
O O 1 x: x, x
-------------- ... --------------,
x:OO"'Ox' x 3
,
x , 1 O O x x 2
, , 6s +1
, or
x(t) = O x , O 1 O x , O x x(t) [ -1
,
, to describe the network.
,
x , O O x , x
--------------1-_-------------.,.
,
x , x ,, :- 0- -O------6 - ~x
, , ,
x , x , : 1 O O x This set of equations can
O x ,, O x , , O 1 O x
, ,
, ,
, ,
,
x , x ,, O O 1 x 2h JI
,----rml'-----.-_ ~
+ Nteu(t) (6-131a)

~r~
y= [O O OC h : O O ... Oc : ... : O O ... O
1, 2 V2 I I .

(6-131b)

wherec¡y¡ is the ith column of D,;;.1 and the (L~~ 1 vm)th column ofthe matrix A In
is the ith column of - DlrD,;;.l. Since (6-131) is a realization of (6-125), we have Figure 6-8 A network.
POLYNOMIAL MATRIX DESCRIPTlON 287
;, AND IDENTIFICATlON

C(sI - A)-lB = C(sI -A)-lN 1c


= D-1(s)N(s) = D -l(s)L(s)N lc (6-132)
(6-128)
which implies
D-1(s)L(s) = C(sI - A)-l (6-133)
or L(s)(sI - A) = D(s)C (6-134)

This is a key equation in this realization, and will be used in the next section.
Eq uation (6-131) will be called a multivariable observable-form realization of
(;(s).
Dual to a multivariable controllable-form realization, a multivariable
observable-form realization of (;(s) = D-1(s)N(s) is always observable no matter
whether D(s) and N(s) are left coprime or nol. If D(s) and N(s) are left coprime,
then the realization is controllable as well; otherwise, it is not controllable.

«6-129) *6-7 Polynomial Matrix Description


(6-130)
Consider the network shown in Figure 6-8. If we assign the capacitor voltages
I(s) is row reduced. The and inductor currents as state variables and use the procedure discussed in
[ x q matrices. Thus a Section 3-4, then a dynamical equation can be developed to describe the
network. There are, however, other methods available to develop mathe­
matical equations to describe the network. For example, if we assign the loop
x
currents ¿l(t) and ¿z{t) as shown and use the loop analysis, then we can obtain
x
+ 3s) ~l(S) - ~2(S)=U(s)
1 1
O x (2s 3s

x 1
A
--¿l(S)+
3s 3s s
(1 1
-+-+2s+1 )A~2(S)=0
x

6s2;31s+4JU:~:~J=[~JU(S)
2
x
O x x(t) or [6s _+/
to describe the network. The output y is equal to y(s) = 2S~2(S) or
r-~ (S'1

-----------
)I(s)=[O 2s]l~:(s;j +O'u(s)
O O x
This set of equations can be written in a general form as
O O x
1 O x P(s)~(s) = Q(s)u(s) (6-135a)
y(s) = R(s)~(s) + W(s)u(s) (6-135b)

O 1 x
+ Nrcu(t) «6-131a)
2h II

--r
O .. , O Cqvq]x

(6-131b)
~ 31
~2h
~-
y

In
1 column of the matrix A Figure 6-8 A network.
~ation of(6-125), we have
288 IRREDUCIBLE REALlZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATlON

where P{S), Q(S). R(s). and W{s) are, respectively, m x m, m x p, q x m, and q x p (6-131) such that
polynomial matrices; u{s) is the p x 1 input vector, and y(s) is the q x 1 output
vector. The variable ~(s) is a m x 1 vector and will be called the pseudostate.
In arder to insure a unique solution in (6-135), the square matrix Pis) is assumed lf deg det Pr(s) = n, th¡
to be nonsingular. The set of equations in (6-135) is called the polynomial matrices. Similar to (1
matrix description of a system. In the time domain. this set of equations
becomes and

P(p )~(t) = Q(p )u(t) (6-136a) or

y{t)=R{p)~{t) +W(p)u{t) (6-136b)


where L(s) is defined m
where p stands for d/dt and pi for (¡i¡dt i . From (6-135a) we have· ~(s) =
unit matrix of order m,
P-1(s)Q(s)u(s). The substitution of ~(s) into (6-135b) yields
p .. (s) and L(s) are lefl
y(s) = [R(s)P- 1 (s)Q(s) + W(s)]u(s) Pr(s), CoroIlary G-S in:
sequently, {A, Ca} is ot
Hence the transfer matrix of (6-135) from u to y is equal to
of {Pis), Q(s)} or {P(s), 1
G(s)=R(s)P-1(s)Q(s) +W(s) (6-137) into (6-135b) yields
From this, we can see that the state-variable equation is a special case of y(s) = R(s
(6-135) ifwe identify R(s) = C, P(s) = sI - A, Q(s) = B and W(s) = E. The fraction
This is not exactly in the ~
G(s) = N,.(s)D,:-I(S) is a special case of (6-135} if we identify R(s) = N,.(s), P(s) =
may not be strictly pro
DAs), Q(s) = 1, and W(s) =0. The fraction G(s) = D¡-I(S)N¡(s) is aIso a special
obtain
case of (6-135) if we identify R(s) = 1, P(s) = D1(s), Q(s) = N¡(s), and W(s) =0.
Hence this description is the most general one.
Given a polynomial matrix description (6-135), its transfer function matrix
where X(s) is a poIyriom
G(s) from u to y can be readily computed by using (6-137). Note that the com­
defined as in Corollary (
puted G(s) is not necessarily proper. In the following we discuss a method of
finding a dynamical equation for (6-135). We consider first (6-135a). If Pis) y(s) = C(sI­
is not row reduced, we can find an m x m unimodular matrix M(s) such that
in which qsI - A)-l Bis
M(s)P(s) is row reduced (Theorem G-ll). The premultiplication of M(s) on
both sides of (6-135) yields,
M(s)P(s)~(s) = M(s)Q(s)u(s) (6-138) is a q x p poIynomiaI m.
which implies dynamical equation

~(s) = (M(s)P(S))-1 M(s)Q(s)u(s)~ P,:-I(s)Q,(s)fi(s) (6-139)

where P,.~ M(s)P(s) and Q,.(s)~ M(s)Q(s). Note that deg det Pr(s) = deg det
where p =djdt, is a realiz
Pis). In general P,:-l(S)Q,.(s) is not strictly proper. If so, we carry out the divi­
~r the transfer matrix G(~
sion:
Gis) in (6-137) is strictIy pI
Qr(s) = P,.(s)Y(s) +Q,.(s) (6-140) E(s) defined in (6-149). I
written explicitly as
so that P;I(S)Q..(s) is strictly proper (Theorem G-12'). The substitution of
(6-140) into (6-139) yields y(t) =O
~(s) = P,:- 1(s:Q,(s)u(s) + Y(s)fi(s) (6-141) From (6-144) and Pr(s)
have
Now P,:- 1(s)Qr(s) is strlctly proper; he~ce the realization procedure discussed
from (6-125) to (6-131) can be applied to find{A, B, Ca} of the feirm shown in
POLYNOMIAL MATRIX DESCRIPTION 289
, AND IDENTlFICATlON

(6-131) such that


m x p, q x m. and q x p
y(s) is the q xl output (6-142)
called the pseudostate.
If deg det Pr(s) = n, then A, B, and C o are, respectively, n x n, n x p, and m x n
matrix P(s) is assumed
matrices. Similar to (6-130), (6-133), and (6-134), we have
called the polynomial
this set of equations Qr(s) = L(s)B (6-143)
and Co(sI - A)- 1 = p r- 1(S)L(s) (6-144 )
(6-136a) or Pr(S)C O = L(s)(sI - A) (6-145)
(6-136b)
where L(s) is defined as in (6-129) and is a m x n matrix. Since L(s) contains a
·135a) we have~(s)= unit matrix of order m, the rank of [L(s) : Pr(s)] is m for every s in iC. Hence
ields Pr(s) and L(s) are left coprime (Theorem G-8'). Because dim A = deg det
P,.(s), Corollary G-8 implies that C o and (sI - A) are right coprime, and con­
sequently, {A, C o} is observable. Note that no assumption of the coprimeness
l to of {P(s), Q(s)} or {P(s), R(s)} is used here. The substitution of (6-141) and (6-142)
(6-137) into (6-135b) yields

::m is a special case of y(s) = R(s)Co(sI - A) -1 Bú(s) +(R(s)Y(s) + W(s))u(s) (6-146)

1W(s) = E. The fraction This is notexactly in the state variable form becausethe matrix R(s)Co(sI - A)-1 B
1tify R(s) = Nr(s), P(s) = may not be strictly proper. Ir so, we can apply Corollary G-12 to R(s)C o to
(s)N¡(s) is also a special obtain
¡) = N¡(.s), and W(s) = O.
R(s)C o = X(s)(sI - A) +C (6-147)

:ransfer function matrix where X(s) is a polynomial matrix and C = Rr(A)C o is a q x n constant matrix
37). Note that the com­ defined as in Corollary G-12. Using (6-147), Equation (6-146) becomes
we discuss a method of
~r first (6-135a). If P(s)
y(s) = C(sI - A)-IBu(s) +(X(s)B + R(s)Y(s) + W(s))u(s) (6-14S)

r matrix M(s) such that in which C(sI - A)- 1 Bis a q x p strictly proper rational matrix and
ultiplication of M(s) on
E{s)~ X(s)B + R{s)Y{s) + W{s) (6-149)

(6-138) is a q x p polynomial matrix. From this development, we conc1ude that the


dynamical equation
x{t) = Ax(t) + BiJI(t) (6-150a)
)Q,.(s)u(s) (6-139)
y{t) = Cx{t) + E(p)u(t) (6-15Ob)

. deg det P,.(s) = deg det where p =d/dt, is a realization of the polynomial matrix description in (6-135)
;0, we carry out the divi­ or the transfer matrix G{s) in (6-137). Note that W(s) =0 does not imply that
G(s) in (6-137) is strictly proper. Whether or not G(s) is proper is determined by
(6-140) E(s) defined in (6-149). If E(p) = E ó +E¡p +E 2p 2 + .. " then (6-150b) can be
written explicitly as
2'). The substitution of
y(t) = Cx(t) + Eou{t) +E¡ü(t) +E 2ü(t) + ...

(6-141 ) From (6-144) and Pr(s) = M(s)P(s), where M(s) is a unimodular matrix, we
~~ . .
:ion procedure discussed
:::o} of the form shown in det (sI - A) = k det P(s)
290 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION

for sorne nonzero constant k. This relation is a consequence of the realization Recall that P,.(s) = M
procedure and is independent of the coprimeness arrrong P(s), R(s), and Q(s). The premultiplicatior
The {A, Ca} in (6-144) is of the multivariable observable form and is always yields
observable. The observability of {A, Ca}, however, does not imply the obser­
vabili ty of {A, C} in (6-150). In the following, we shall establish -Ul 1(S) UI2(S)
P(s) M-l(S)L(.
[
{P(s), Q(s) left coprime} if and only if {A, B controllable} (6-151) -0-- - ------=.X(;)­
(P(s), R(s) right coprime} ifand only if {A, C observable} (6-152)

Before doing so, we need sorne preliminary development. We combine


(6-145), (6-140), (6-143), (6-147), and (6-149) to form r- I,,: O

L(s) 0J[SI-A
[ -X(s) I q -C
B J=[ Pr(s) Q,,(S)J[C a -Y(S)J (6-153) =L:T!~~)-
E(s) -R(s) W(s) O Ip
where the leftmost and
They are, respectively, (m + q) x (n + q), (n + q) x (n + p), (m + q) x (m + p), and equation, we are ready
(m +p) x (n +p) matrices. Because of the right coprimeness of {P,.(s), L(s)} and From the first two
the left coprimeness of {si -A, Ca}, (6-145) and Problem G-ll imply the exist­
-U 11 (S) Uds)
ence of U 11 (s) and Uds) such that [ P(s) M-l(S)L(s)

[
-U l l (S) ~12(S)J[ -CaJ=[I"J (6-154 )
where 8(s) denotes the r
P,.(s) L(s) si ­ A O most matrices in (6-158
~--~-~'

unimodular 1
[ 0m (sI
where the leftmost matrix is a unimodular matrix. Using this matrix, we expand
(6-153) as if and only if

o
P(s)
which implies that

[(si - A)
(6-155)
if and only ir [l
orequivalently, [A, lB} is
This is obtained by augmenting the first three matrices in (6-153) as shown 1,;
Simiiariy, we can establi
(6-155) and then searching the fourth matrix in (6-155) to complete the equality. are right coprime.
Since the left-upper-corner matrix in the leftmost matrix in (6-155) is unimodular,
so is theentire leftmost matrix in (6-155). From (6-154), we have U ds)(sl - A) =
1" - U II(S)C O and Definition 6-2

[
-U 11 (S)
1m
UdS)(SI-A)J=[-Ull(S)
Co 1m
1" - U 11 (S)C O
Ca
J The polynomial matrix
only if P(s) and Q(s) are I

=[1" - U 11 (S)J[0 I"J (6-156) Theorem 6-7


0 1111 1111 Co
Since. the two matrices afteT the last equality are unimodular, so is the first Consider the polynomia
Ínatrix ih (6-15"6). Consequently, the rightmost matrix in (6-155) is also uni­ Then an n-dimensional d
modular. or, equivalently, controll~
As the last step, we shall replace P,.(s) and Q/s) in (6-155) by P(s) and Q(s).
POLYNOMIAL MATRIX OESCRIPTlON 291
·, ANO IDENTlFICATlON

Recall that P,(s) = M(s)P(s) and Q,.(s) = M(s)Q(s), where M(s) is unimodular.
uence of the realization
The premultiplication ol' diag (In> M- 1 (s), I q }, a unimodular matrix, to (6-155)
ng P(s), R(s), and Q(s).
yields
ble form and is always
es not imply the obser­
-Uld S) UI2(S) : 0J[lm: O O ]
establish _~!~) ~~l_(~)~J~)_:_~_ -Ó-:-(~i~Á)---B-
[
; control1able} (6-151) O -X(s): Iq O: -C E(s)
C observable} (6-152)

ll'-:- __ ~-----~-] [-UI1(S) Uds)(sl-A): U 12 (S)BJ


l
>pment. We combine
= O : p(s) Q(s) 1m Co : - Y(s) (6-157)
O : - R(s) W(s) - - -O ---------Ó- - - - - : - - -I~ - - ­
\[~o -~p(S)J (6-153)
wherc the leftmost and rightmost matrices remain to be unimodular. With this
»,(m+q)x (m-tp), and eq uation, we are ready to esta blish (6-151) and (6-152).
~ness ol' {p,.(s), L(s)} and From the first two block rows of(6-157), we can readily obtain
m G-ll imply the exist­
[
-U l1 (S)
P(s)
U12(s) J[l m
M- 1(s)L(st O
O 0J
(si -A) B
[II!O P(s)
=
O Q(s)
OJ S(s) (6-158)

where S(s) denotes the rightmost matrix in (6-157). Since the leftmost and right­

:J (6-154) most matrices in (6-158) are unimodular, we conclude that

[; (sI ~ A) ~] has rank n +m for every s in C


Ig this matrix, we expand il' and on!y if
O
Q~S)J has rank n +m for every s in C
-~t]
P(s)
which implies that

[(sI - A) : B] has rank n for every s in C


l) : U 12(S)8]
, - Y(s) (6-155 ) if and only if [P(s) : Q(s)] has rank In for every s in C
._J _

: Ip or equivalentl)', [A, B} is controllable if and only if 1"(5) and Q(s) are left coprime
~s in (6-153) as shown in Similarly, ...ve can eSlablish thal (A, C; is observable ir and oníy il' íP(s) and R(s)
to complete the equality. are right coprime.
,in (6-155) is unimodular,
. we have U 12(S)(sl - A) =
Definition 6-2
The polynomial matrix description in (6-135) is said to be irreducible il' and
only il' P(s) and Q(s) are left coprime and P(s) and R(s) are right coprime.

1,:,1 (S)J[:'" ¿J (6-156) Theorem 6-7

Consider the polynomial matrix description in (6-135) with n = deg det P(s).
1imodular, so is the first .Then an n-dimensional dynamical equation realization ol' (6-135) is irreducible
rix in (6-155) is also uni­ or, equivalently, control1ableand observable ifimd ónly if (6-135) is irreducible.
I
(6-155) by P(s) and Q(s).
292 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTIFICATION

This theorem follows directly from Definition 6-2 and the statements in
given a transfer mat
(6-151) and (6-152). To conclude this section, we discuss the situation where
procedures in Section
(6-135) is not irreducible. If PIs) and Q(s) are not left coprime, then there exists
relationships among
a m x m polynomial matrix H(s) with deg det H(s) > 1 such that
questions can stiJl be
PIs) = H(s)P(s) and Q(s) = H(sQ(s) ~ider a transfer matrix
G(s) = N .(05)0- 1(05)
Consequently. at the roots, l., of det H(s) = 0, we have A ' " or
<.: 1 is strictly proper (
pepe),) Q(A)] = pH(A)[P(),) Q(),)] < m G(s) = NI (05)0 1 1(05) + E
betw~en {N,.(s), D,.(s)} (
where p stands for the rank. lf an n-dimensional dynamical equation is de­
q uestIOns. we rewri te ((
veloped for {P(s), Q(s), R(s), W(s)} with n = deg det PIs), then this n-dimensional
state equation will not be controllable. lfthis equation is decomposed into the
form in (5-54), then the eigenvalues associated with A. c, the uncontrollable part,
wil! be equal to the roots of det H(s). Hence the roots of det H(s) or, equivalently,
[
where ~ is called the PSI
those A in e with p[P(A) Q(A)] < m wil! be called the uncontrollable mode of
{Pis), Q(s)}. These uncontrollable modes are called the input-decoupling
zeros in Reference S185.
Similarly, if Pis) and R(s) are not right coprime, then the roots of the deter­
is called the system l11{{tr
minant of their greatest common right divisor wil! be called the unobservable
modes of {P(s), R(s)}. These roots are called the output-decoupling zeros in
Reference S 185. Hence an irreducible {P(s), Q(s), R(s), W(s)} does not have any
Ir we identify R(s) = e
uncontrollable or unobservable modes.
.identify R(s) = N,.(s), P(s:)
The discussion of the polynomial matrix description in this section is not
meludes (6-159) and (6-1,
complete. We discuss· only its realization problem. For a more detailed
be used to describe any (
discussion, see References S34 and S218. In S218, the description is called the
Consider the system 1
differential operator description.

*6-8 Strict System Equivalence

In this text, we have introduced three types of mathematical descriptions for


where 1 is a unit matrix 0[,
linear time-invariant multivariable system~. They are state-variable equation
than deg det P(s). It is ele

transfer matrix in fractional forms and Ge(s)=R,,(S)P,,-1


(6-160) Hence the input-output b
their en tire dynamical beh~
and polynomial matrix description
Consider two system m
P(s)~(s) = Q(s)u(s)
y(s) = R(s)~(s) +W(s)u(s) (6-161 ) S¡(s)
Equation (6-159) is more general than the one studiedin earlier sections by
where p¡(s), Q¡(s), R¡(s), an
including derivatives ofu in y. By so doing, its transfer matrix can be extended
q x p polynomial matrices.
to include improper case. Given a state-variable equation, the transfer function
ever, because of the discussi
description can be obtained as G(s)=C(sI-A)-lB+(E+E1s+···). lf
to make them egual and re,
G(s) is factored as G(s) = N,.(S)O,:-l(S), then a polynomial matrix description
this reg uirement, the subseg
can be obtained as Q(s) = 1, P(s) = 0,.(05), R(s) = N,.(s), and W(s) = O. Conversely.
S186, and S187.
STRlCT SYSTEM EQUIV ALENCE 293
, AND lDENTlF\CATlüN

and the statements in given a transfer matrix or polynomial matrix description, we may use the
lSS the situation where
procedures in Section 6-7 to deve\op a state variable description. Hence the
prime, then there exists relationships among them have been essentia\\y established. Even so, sorne
questions can still be Ilosed regarding these descriptions. For example. con­
:;uch that
sider a transfer matrix G(s) which is not necessarily proper. It can be factored as
9(s)=N,.(s)D,:-I(S) or it can be decomposed as G(s)=GI(s)+E(s), where
<}:1 is strictly proper and E(s) is a polynominal matrix and then factored as
G(s) = N ¡(S)D¡I (s) + E(s). The question is then: what is the re\ationship
between {N,.(s), D,.(s)} and {N 1 (s), D 1(s), E(s)} ?To answer this and other related
namical equation is de­ questions. we re"'Tite (6-161) as
then this n-dimensional
1 is decomposed into the
P(s)
[ -R(s)
Q(s) -Il
~(S)J
W(st -ú(s) =
[ O ]
-y(s)
(6-162 )
the uncontrollable part,
det H(s)or, equiva\ent\y, where ¡; is called the pseudostate, and the matrix
uncontro\\ab\e mode of
S(S)=[ P(s) Q(S)] (6-163 )
:d the input-decoup\ing -R(s) W(s)

en the roots of the deter­ is ca\\ed the system matrix. 1ts transfer function from u to y is
cal\ed thc unobservab\e G(s)=R(s)P-!(s)Q(s) +W(s) (6-164)
tput-decoupling zeros in
, W(s)} does not have any Ir we identify R(s)=C,P(s)=(sl-A),Q(s)=B, and W(s)=E+E 1 s+ "', or
identify R(s) = N,.(s), P(s) = DAs), Q(s) = 1, and W(s) =0, then the system matrix
ion in this section is not ineludes (6-159) and (6-160) as special cases. Hence the system matrix S(s) can
1. F or a more detai\ed
be used to describe any of the three descriptions Ín (6-159)-(6-161). .
e description is called the
Consider the system matrix S(s) in (6-163). We extend it to

Se(s) =
1
O
O :O
P(s) : Q(s)
J~ [ P e(s) Qe(S)]
[ ---------,----- -Re(s) W(s) (6-165)
O - R(s) , W(s)

1ematica\ descriptions for where I is a unit matrix of any order so that the order of P e(s) is equal to or larger
than deg det P(s). lt is c\ear that
re state-variable equation
'~Let :F'(.{s) = det :2\:;)
and Ge(s) =Rels)P(:-I(S)Q,,(s) +W(s) =R(s)P-I(s)Q(s) +W = G(.s)
Hence the input-ou tpu t behavior of S(s) and that oC Se(s) are ident icaL In fact,
(6-160)
1 (s)N,(s)ú(s) their entire dynamical behaviors are, as will be shown latee equivalent.
Consider two system matrices

S¡(s) =[ p¡(s) Q¡(s) ] i = 1,2 (6-166)


(6-161 ) -R¡(s) W¡(s)
:)
ldied in earlier sections by where PieS), Q¡(s), R¡(s), and Wi(s) are respectively mi x mi, m¡ x p, q x mi, and
sfer matrix can be extended q x p polynomial matrices. Note that mi is not necessarily equal to 1112' How­
uation,the transfer function ever, because ofthe discussion in (6-165), we may extend either mi or m20r both
,)-IB--i-(E+E 1S+··}.. lf to make them equal and require m =m!'=m2 2':deg det PieS), i=l, 2. Without
{nomia\ matrix descnptlOn this requirement, the su bsequent discussion may not hold. See References S125,
,and W(s) =0. Converse\y, S186, and S187.
294 IRREDUCIBLE REALlZATIONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTIFICATION

Definition 6-3
Hence we have
Two system matrices SI (s) and Sz(s) are said to be strictly system equivalent if
and only if there exist m x In unimodular polynomial matrices Ves) and Ves) and
q x m and m x p polynomial matrices X(s) and Ves) such that which, together with 1,

ves) 0J[ PIes) Q¡(S)][V(S) Y(S)]=[ Pz(s) Qz(s) J (6-167)


[ X(s) I q -R¡(s) W¡(s) O Ip -Rz(s) Wz(s)
where k is a nonzero c
I We compute
Since Ves) and Ves) are unimodulars, so are
Gz(s)=R 2 P-¡Q
z +W ­
Ves)
[ X(s)
O]
I
[ves)
O
Y(S)]
Ip
Z
= (R ¡ - XP dV(U:
q '
which, by simple manip
and their inverses. Using this property, it can be readily verified that this equiv­
alence relation has the symmetry property [if S¡(s)~ Sz(s), then Sz(S) ~ S¡(s)], Gz(s) = R
the rel1exitivity property [S ¡ (s) ~ S¡(s)] and the transitivity property [if S¡(s) ~ This completes the proo
Sz(s) and Sz(S)~S3(S), then S¡(S)~S3(S), where ~ denotes strict system equiv­
alence]. In order to establish
The reason of using extended SíeS) in Definition 6-3 can be seen from the we use (6-162) to writeS. 1
realization developed in Section 6-7, especially Equations (6-153) and (6-155).
Without extending Pi (s), the matrices Ves) and Ves) in (6-167) may not be square
and, certainly, may not be unimodular. Using the unimodularity property, we [-:
[-:
can write (6-167) as
and
. [ves) 0J[ PIes) Q¡(S)]=[ Pz(s) Qz(S)][V(S) .Y(S)J (6-168)
X(s) 1 -R¡(s) W¡(s) -Rz(s) Wz(s) O Ip
Note that 8 1 (s) and Sz(s]
where V(s)=V-¡(s) and Y(s) = -V-¡(s)Y(s). Clearly Ves) is unimodular and hence u and y in (6-170) al
Y(s) is a polynomial matrix. A comparison of (6-168) and (6-155) reveals ~¡ and ~z are different in ' 1
immediately that {pes), Q(s), R(s), W(s)} and its realization {A, B, C, D} de­
veloped in Section 6-7 are strictly system equivalent.
Mathematical descriptions which are strictly system equivalent are equivalent
[U 0J[ P ~
X 1 -R: "
in every sense. They have the same transfer matrix; their pseudostates are which can be written as , b)
related by an invertible transformation. Their dynamical-equation realiza­
tions developed by using the procedure discussed in Section 6-7 are zero-inpul r1lT ~"""1 r "
and zero-state equivalent (see Section 4-3). If one realization is controllable
(observable), so is the other, and conversely. These properties will be established
L~ ;Jl-;_
in the following. A comparison of (6-171) an,

Theorem 6-8 ~
Two system matrices. which are strictly system equivalent have the same where Ves) is unimodular, a,
transfer matrix and det p¡ (s) = k det P z(s), where k is a nonzero constant.
t(s)=
Proof
We see that ~¡ and ~z areTe":
We multiply (6-167) out toyie1d ~nd (6-174). Hence, ifS 1 (s)a
VP¡V VP¡Y +VQ, ] [ P z
~n~~t u(t), and any set of i~i
[ -(R¡ -XP,)V (XP¡ -R¡)Y +(XQ¡ +W¡) - -R z
lO'tIal
'd .
conditions in S·
z, an
d'
, entIeal and their pseudosta
:E, AND IDENTIFlCATION
STRICT SYSTEM EQUlVALENCE 295

Hence we have

'ctly system equivalent if U(S)P I(s)V(s) = P 2(S)


atrices u\s) and V(s) and which, together with the unimodularity ofU(s) and Ves), implies
1 that
k det PI(S) =det P 2(s) (6-169)
l
P2(S) Q2(S) (6-167)
where k is a nonzero constant. Hence we have deg det P 1(s)=deg det P 2(s).
- R 2(s) W 2(S)j
We compute
• G2(s)=R 1 Pi l Ql +W 1
=(Rl-XPl)V(UPIVrlU(PIY +Ql)+(XP 1 -R I )Y +(XQl +W I )
which, by simple manipulation, becomes

Iy verified that this equiv­


GAs) = R1 Pi 1Q2 + W 2 = R I P¡IQl + W 1 = G1(s)
oSl(S), then S2(S)~SI(S)],
This completes the proof of this theorem. Q.E.D.
tivity property [if S I(S).~

notes strict system eqUlv­ In order to establish the relationship between the pseudostates ~l and ~2'
we use (6-162) to writeS¡ as
6-3 can be seen from the
ttions (6-153) and (6-155). (6-170)
(6-167)may not besquare
¡¡imodularity property, we
and (6-171 )

¡(s) 11[Y\s) Y(sq (6-168)


Note that SI(S) and S2(S) are two different descriptions of the same system,
2(S)J O lp J hence u and y in (6-170) and (6-171) are the same. However, their pseudostates
y V\s) is unimodular and ~1 and ;2 are different in general. From (6-168), we have

J[ -otJ = ['-R
-168) and (6-155) reveals
lization {A, B, e, D} de­ U
[X
0J[
1 -R
PI
I
Q1
W1
P 2 Q2
2 W2
J[VO 1YJ[ -o; J 1

1 equivalent are eq uivalent which can be written as, by the substitution of (6-170),
ix' their pseudostates are
n~mical-equation realiza­
Section 6-7' are zero-input
realization is controllable
[U0J[ -y0J =[-yJ=
X 1
ol [ ]32
-R 2
Q2JrlV~I--:,'YM---'J'
W2 -u
(6-172)

'operties wi\l be established A comparison of (6-171) and (6-172) yields


;2(S) = V(S);I(S) - Y(s)u(s) (6-173)

where Ves) is unimodular, and


equivalent have the same ;1(S) = V- 1 (S);2(S) +Y- 1 (s)Y(s)u(s)
.a nonzero constant. = V(S);2(S) - Y(s)o(s) (6-174)

We see that ; 1 and ;2 are related by the invertible transformation pair in (6-173)
and (6-174). Hence, ifS 1(s) and S2(S) are strictIy system equivalent, then for any
input u(t), and any set of initial conditions iriS¡, there exists a unique set of
initial conditions in S2, and vice versa, such that the outputs of SI and S2 are
identieal and their pseudostates are related by (6-173) and (6-174). In short, if
296 ¡RREDUCIBLE REALIZATIONS, STR¡CT SYSTEM EQU¡VALENCE, AND IDENTIF¡CAnON

Si, i = 1,2, are strictly system equivalent, there is no difference in their dynamical
Proof
behaviors.
Consider two system matrices S¡(s) and S2(S). lf they are realized by using If {A, B, C, E} and {A, i
the procedure developed in Section 6-7, then the realizations {A¡, B¡. C¡, E¡} constant matrix P such
have the property these, it is straightforwa:
det (sI - A;) = k¡ det p¡(s) i = 1, 2
lf S¡(s) and S2(S) are strictly system equivalent, then det P¡(s)=k det Pz(s).
Hence the dynamical-equation realizations of S¡(s) and S2(S) have the same
dimension and the same characteristic polynomial. Hence their system matr
Now we assume that
Theorem 6-9 valent. Then from (6-16

Coprimeness, controllability, and observability are invariant under the trans­ U(S) O][SI ­
formation of strict system equivalence. [ X(s) I -
q

for sorne unimodular U(s


Proof
the same transfer matrb
Consider (6-167). Because of the presence of the two zero matrices, the first implies
block row of(6-167) can be written as
,
U(S)[p¡(s) Q¡(s)]
[v(S)
O
y(S)]
I = [P 2(s) Q2(S)] (6-175 ) Using CorolIary G-12, Wf
p

and the first block column of (6-167) can be written as


where P is a constant mal
U(S)' O][p¡(S)]V(S)_[ P 2(S)] (6-176)
[ X(s) I R¡(s) - -R 2(s) P(sI ­
L

Since U(s) and V(s) are unimodular, we have, for every s in ic,
which implies P(s)(sI - A
~roper rational matrix, s'
P(s) is a constant matrix.
and to yield

where p denotes the rank in the field of complex numbers. Hence P ¡ (s) and
Q ¡ (s) are left coprime if and only if P 2(S) and Q z(s) are lei'L coprirne. ¿ ;]22\;;) = v¡!hich lmplies if' = jp o and .
sI - A and Q2(S) =E, then P ¡(s) and Q¡(s) are left coprime if and only if {A, E}
is controllable [see Equation (6-151)]. The rest ofthe theorem can be similarly P(sI -A)
proved. Q.E.D. Now we shalI show tl
unimodular, U- ¡(s) is also
In Section 4-3, we introduced the concept of equivalent dynamical equa­
l
tions. Now we shall show that strict system equivalence is a generalization of
this concept. for some polynomial matri
plication of(6~179) and (6-1
Theorem 6-10 U(s)U- 1 (s) = (sI'..:.. A)U(s)(sJ
c,
Two dynamical equations {A, B, C, E} and {A,R, E} are equivalent ifand only which becomes, because of I .
if their system matrices are strictly system equivalent.
I-Pp¡=(sI-;
,._..-._-""-- ...--._,' , .. .,.--" .. -_ ... , .•.. ~ .. _'-" ", ......

:, AND IDENTIFICATION STRIGr SYSTEM EQUIV ALENCE 297

:ence in their dynamical Proof


e,
If {A, B, C, E} and {A, B, E} are equivalent, then there exists a nonsingular
ey are realized by using
izations {A¡, B¡, C i , EJ
e
constant matrix P such that A = PAP - \ B = PB, = CP - 1 and E = E. With
these, it is straightforward to verífy
1
,2 p O][sl-A B-![P- O]=[sl-}
[ o I -C E o I -C
det PieS) = k det P 2 (s).
nd Sz(s) have the same Hence their system matrices are strictly system equivalent.
e,
Now we assume that {A, B, C, D} and {A, B, ñ} are strictly system equi­
valent. Then from (6-168), we have

variant under the trans­


U(S) O
[X(s) I
][SI-C
-A EBJ [SI-C-} = ~][V(S)
E O
Y(S)]
I
(6-177)
q p

for sorne unimodular U(s) and Ves). Since {A, B, C, E} and {A, B, E} have e,
the same transfer matrix (Theorem 6-8), we have E =:E. Equation (6-177)
implies
) zero matrices, the first
U(s)(sl - A) = (sI - A)V(s) (6-178)

Using Coro])ary G-12, we can wríte U(s) as


Q2(S)]
U(s) = (sI - A)U(s) + P (6-179)

where Pis a constant matrix. The substitutíon of (6cI79) into (6-178) yields
(6-176) P(sl ­ A) = (sI - A)[V(s) - U(s)(sl - A)]
~ (sI - A)P(s) (6-180)
.¡ s in e, which implies P(s)(sl-A)-l=(sl-A)-lp. Since (sl-A)-lP is a strictly
s)] proper rational matrix, so must be P(s)(sl - A)-l. This is possible only if
pes) is a constant matrix. Hence we may replace pes) in (6-180) by P(s) = Po
to yield
sP-PA=sPo-AP o
nbers. Hence P 1(s) and
which implies JP' = Po and PA = APo. Hence we have
: left coprime. lf P 2(S) =
,rime if and only if {A, B} P(sl - A) = (sI - A)P and PA=AP (6-181)
theorem can be similarly
Q.E.D. Now we shall show that the P in (6-179) is nonsingular. Since U(s) is
unimodular, U- 1 (s) is also unimodular and can be expressed as

Jivalent dynamical equa­ (6-182)


mce is a generalization of for sorne polynomial matrix U 1 (s) and sorne constant matrix P 1 - The multi­
plication of (6~179) and (6-182) yields
U(s)U- 1 (s) =(sl - A)U(s)(sl - A)U 1 (s) +(sl - A)U(s)P 1 + P(sl- A)U l(S) + PP 1
whichbecomes, because of (6-181) and U(s)U- 1(s) == 1,
are equivalent if and only
1- PP 1 =(sl - A)[U(s)(sl - A)U 1 (s) + U(s)P 1 + PU l(S)]
298 .IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTIFICATION

or D 2(s) = D 1(s)T(s) an
(sl-AT 1(I-PP¡)=Ü(s)(sl-A)U 1(s)+Ü(S)P 1 +PU 1(s)
[~

(6-183)

Since its left-hand side is a strictly proper rational matrix, whereas its right-hand
side is a polynomial matrix, (6-183) holds only if both sides are identically which implies the st
equal to zero. Hence we have I-PP 1 =0 or PP 1 =1. Consequently, P is Now if G(s) is no
nonsingular and is qualified as an equivalence transformation. (;(s) = N(s:
To complete the proof ofthis theorem, we must show 8 = PE and C = CP- 1
From (6-177), we have where Ñ(S)D-1(S)is ~
of
U(s)B =(sl -A)Y(s) +8 D,
which becomes, by using (6-179), [ -N,

(si -A)U(s)B + PB = (si - A)Y(s) +8 We conclude that {í


SimilarIy we can she
or Ü(s)B - Y(s) = (si - A)-1(8 - PB) alent.
What remains to
Its left-hand side is a polynomial matrix, whereas its right-hand side is a rational left-coprime fraction
matrix. This is possible only if B - PB = O. Hence we conclude 8 = PB. D I- 1N/. Tben there
Similarly, we can show C=cp- 1. This completes the proofofthis theorem.
Q.E.D.

System matrices which are strictly system equivalent have the same transfer where the leftmost m
matrix. System matrices which have the same transfer matrix, however, are not D/N,. = NID,., we forn

U U 12 : O] ~I__
necessarily strictly system equivalent. Example 2 of Section 4-3 is an example
of this statement. A different example will be G(s) = N(s)D -l(S) = N(s)D- 1(s),
11

where N(s) and D(s) are right coprime, but N(s) and D(s) are not right coprime. [-1DI 0:1
-
NI 'O O'
- - -

O:
- - - - - _1- _ I

Clearly the two system matrices {D(s), 1, N(s), O} and {D(s), 1, N(s), O} have the
same transfer matrix. However, because of det D(s) 1= det D(s), they are not The second and thirc
strictly system equivalent (Theorem 6-8). matrices of {D,., 1, N r
Although system matrices {p¡(s), Q¡(s), Ri(s), W¡(s)}, which have the same unimodular because i
transfer matrix, are generally not strictly system equivalent, they become (6-186). We show in
strictly system equivalent if the system matrices are irreducible: that is, {F';(s), unimodular. From (6­
Q¡(s)} are left coprime and {P¡(s), Ri(s)} are right coprime. We establish first a U 12 D
special case.
N,.
Since the two matrice
Theorem 6-11 the first matrix in (6- j
matrix in (6-187). (
All coprime fractions of G(s) are strictly system equivalent where G(s) is a
rational matrix, not necessarily proper. {D,., 1, N,., O} and {D h 1
proof of this theorem.

Proof
.7The proor orTheorem G-1
Consider the two right coprime fractionsG(s)=N1(S)Dl1(s)=N2(S)D21(s). to improper G(s).
Theorem G-13 implies the existence of a unimodular matrix T(s) such that 8Th!: procedure is similar to

STRICT SYSTEM EQUIVALENCE 299


AND lDENTIFICATION

Dz(s) = Dl(s)T(s) and Nz(s) = N l (s)T(s).7 Hence we have,

PI + PUl(s) (6-183) 1 0J[ Dl(s) I][T(S) 0J = [ Dz(s) OIJ (6-184)


[ O 1_ -Nl(s) O O I -Nz(s)
whereas its right-hand
which implies the strict system equivalence of {DI' 1, NI' O} and {D z, 1, Nz, O}.
h sides are identically
Now if (;(s) is not strictly proper, then we may also factor (;(s) as
L Consequently, P is
lation.
(;(s) = N(s)O -l(S) + E(s) = (N(s) + E(s)O(s))O -l(S)~ N(s)O - l(S)
B= PBand C=Cp-l

where N(S)O-l(S) is strictly proper and coprime; N(s) = N(s) + E(s)O(s). Because
of

[-~~:~ ~]=[ -~(S) ~I -Zi;~ ~(s)I~~]


(6-185)

We conc1ude that {O, 1, N, E} and {O, 1, N, O} are strictly system equivalent.


Similarly we can show that all left-coprime fractions are strictly system equiv­
alent.
What remains to be shown 8 is the strict system equivalence of right- and
l-hand side is a rational left-coprime fractions of G(s). Consider the coprime fractions (;(s) = N r D; 1 =
we conc1ude B = PB. DI-lN¡. Then there exist U l l and U lZ such that
proof of this theorem.
Q.E.O. Ull Ulz-I[- N"J = [IJ (6-186)
[ DI NI _ D,. O
where the leftmost matrix is a unimodular matrix (Problem G-ll). Now using
. have the same transfer D¡N,. = NI O,., we form the identity
natrix, hówever, are not
cUon 4-3 is an example
~~l ~:z~~]r~-~--~~.--i]=r-ri-¡--~l--~~[~ll U~,D< ~lZ] (6-187)
i(s)D-l(s) = N(s)O-l(s),
:) are not right coprime.
[ =--C---o---:-i L~ :-N
r
O Lo: -1 o] -ó----O---:-I-­
O(s), 1, Ñ(s),O} have the
The second and third matrices from the left in (6-187) are extended system
!o det O(s), they are not
matrices of {D.., 1, N,., O} and {DI, NI' 1, O}. The leftmost matrix in (6-187) is
unimodular because its left-upper-corner matrix is unimodular fol1owing from
}, which have the same

(6-186). We show in the fol1owing that the rightmost matrix in (6-187) is also
luivalent, they become

unimodular. From (6-186), we have - U 11 N, + U 12D.. = 1 and


'educible; that is, {p¡(s),

neo We establish first a


UIll UlzD"J=[Ull UllNI+I]=[U ll C¡[I N..] (6-188)
[ N,. I N.. I 0_ O I
Since the two matrices after the last equality in (6-188) are unimodular, so is
the first matrix in (6-188). This implies the unimodularity of the rightmost
matrix in (6-187). Consequently, we conc1ude from Definition 6-3 that
ivalent where G(s) is a {D,., 1, N,., O} and {DI> NI> 1, O} are strictly system equivalent. This completes the
proof of this theorem. . Q.E. O.

'The proorof Theorem G-13 does nol use lhe properness of (;(5); hence lhe lheorem: is applicable
(s)D1l(s) = Nz(s)Dzl(s).
lo improper (;(5). . .

r matrix T(s) such that


8The procedure is similar lo lhe one from(6-154) lo (6-156).

1
__ .• ~'~_J~'._" .. ,". _ _ ','._'
.. ~._-"---~-

300 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTIFICATION


IDENT

Theorem 6-12
polynomial matrix 1
AH irreducible polynomial matrix descriptions {PieS), Q¡(s), R¡(s), W¡(s)} which These results are di]
have the same transfer matrix are sti-ictly system equivalent. out the foHowing tra
Proof
Every irreducible {Pi, Q¡, R¡, W;} has an irreducible realization {A, B, C, E(p)}
with dim A = deg det PieS). AH irreducible {A, B, C, E(p)} which have the same Transfer matrix
transfer matrix are equivalent (Theorem 5-20) and, consequently, strictly
system equivalent (Theorem 6-10). Hence by the transitivity property, we Dynamical equation
conclude that aH irreducible {Pi Q¡ R¡ W;} of the same transfer matrix
are strictly system equivalent. Q.E. D.
Polynomial matrix
AH irreducible dynamical equations, all coprime fractions and aH irreducible description
polynomial matrix descriptions which have the same transfer matrix are
strictly system equivalent and consequently have, following Theorems 6-2 and System matrix
6-8 and Definition 6-1', the following properties:
~(G(s)) ~ det(sI - A)~ det D(s) ~ det pes) (6-189)
where ~(G(s)) is the characteristic polynomial of C(s) and denotes the Hence the discussion
equality of two polynomials modulo a nonzero constánt factor. Conversely, if time case. Note that
all descriptions have the same transfer matrix and satisfy (6-189), then they the discrete-time case
must be all irreducible. Under the irreducibility assumption, any one of the equivalently, s- I is re]
descriptions can be used, without loss of any essential information, to study There is, however
and design a system. special discussion. e
To conclude this section, we mention that system matrices which are strictly time system expanded
system equivalent to
pes) Q(s) ]
Ses) = [ -R(s) W(s) As in the continuous-l
parameters of g(z). In
where P, Q, R, and W are, respectively, m x m, m x p, q x m, and q x p polynomial
be computed from di g
matrices, can be generated by the foHowing elementary operations:
response of the system
1. Multiplication of any of the first m rows or columns by a nonzero constant. of an impulse as an in
2. Interchange of any two of the first m rows or columns. of g(t) are again impra(
3. Addition of the multiple of any of the first m rows (coiumns) by a poly time case are not real
nomial to any of the m +q rows (the m + p columns). entirely different. If v
initiaHy relaxed linear
These operations can be readily derived from the unimodular matrices
data at the output term
Ves) O] [ves) y(S)] 2, . . .. Hence the real
[ X(s) I q ' O Ip case can be considered.
mathematical descripti
used in the definition of strict system equivalence. ment at the input and (
corrupted by noises.
requires the concepts o
*6-9 Identification of Discrete,.Time Systems
from Noise-Free Data
In the previous sections, we introduced various realization methods for con­

~-----

tinuous-time systems described by transfer matrices. We also introduced 9Consequently melhods ar


M k =I~ tkg(r)dt, k =0, l,2
,D IDENTIFICATION IDENTIFICATION OF OISCRETE-TIME SYSTEMS FROM NOISE-FREE DATA 301

polynomial matrix description and the concept of strict system equivalence.


, Ri(s), W;(S)} which These results are directly applicable to the discrete-time systems if we carry
1. out the following transformations:

Continuous-time systems Discrete-time systems


ition {A, B, e, E(p)}
.vhich have the same Transfer matrix G(s) = N,.(s)D,:-l(S) G(z) = N,(z)D,:-l(Z)

msequently, strictly = D¡-I(S)N¡(s) = D¡-1(Z)N1(z)

;itivity property, we Dynamical equation x(t) = Ax(t) + Bu(t) x(k+ l)=Ax(k)+Bu(k)


ame transfer matrix y(t) = Cx(t) +Eu(t) y(k) = Cx(k) + Eu(k)
G(s)=E +C(sI-A)-IB G(z)=E +C(zI-A)-IB
Q.E.D.
Polynomial matrix P(s);(s) = Q(s)u(s) P(z);(z) = Q(z)u(z)
ns and a1l irreducible description y(s) = R(s );(s) + W(s)u(s) Y(z) = R(z);(z) +W(z)u(z)
transfer matrix are P(s) Q(s) ] P(z) Q(z) ]
19 Theorems 6-2 and System matrix S(s) = [ S(z) = [
-R(s) W(s) -R(z) W(z)

s) (6-189)
and denotes the Hence the discussion of these problems will not be repeated for the discrete­
actor. Conversely, if time case. Note that a1l block diagrams in this chapter are also applicable to
¡fy (6-189), then they the discrete-time case if every integrator is replaced by a unit-delay element or,
ption, any one of the equivalently, S-1 is replaced by Z-I.
.nformation, to study There is, however, one problem in the discrete-time case which deserves
special discussion. Consider a sampled transfer function, g(z), of a discrete­
~ices which are strictly time system expanded as
g(z) = h(O) +h(l)z- I +h(2)z- 2 + ...
As in the continuous-time case, we sha1l ca1l {hU), i = 0, 1, 2, ... } the Markov
parameters of g(z). In the continuous-time case, the Markov parameters must
,and q x p polynomial be computed from dig(t)/dt i , i = 0, 1, 2, ... , at t = 0, where g(t) is the impulse
.perations: response of the system or the inverse Laplace transform of g(s). The generating
of an impulse as an input is not possible in practice; repetitive differentiations
>y a nonzero constant.
of g(t) are again impractical. Hence the Markov parameters in the continuous­
;.
time case are not really available. 9 In the d.iscrete-time case. the situation 1S
(columns) by a poly-
entirely different. If we apply the input {u(O) = 1, u(i) = 0, i = 1, 2, ... } to an
initia1ly relaxed linear time-invariant discrete-time system, then the measured
modular matrices data at the output terminal are the Markov parameters, that is, y(i) = h(i), i = 0,1,
2, . . .. Hence the realization from the Markov parameters in the discrete-time
case can be considered as an identification problem-a problem of determining a
mathematical description of a system from the data obtained by direct measure­
ment at the input and output terminals. In actual measurement, all data will be
I corrupted by noises. A study of the identification problem. with noisy data
requires the concepts of probability and statistics and is outside the scope of this
:ems
I
~
ltion methods for con­ 9Consequently methods are developed lo find realizations by lIsing the moments defined by
We also introduced M k = S~ tkg(t) dt./< =0.1,2, . (see References 528 and 5146). .

l
·., ··--'·····c·······_·--·· ' _.'.' =
'=.'..c='•.""'.,,,,'===.=--=.. =.._='=. -- _.

302 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION IDENTIF.

text. Hence we assume in this text that all data are free of noise and call the In the following, 'i
problem the deterministic identificatían. an arbitrary input-~Ue
Consider a single-variable linear time-invariant discrete-time system with introduced. The prot
transfer function 9(Z). Ir we apply the impulse sequence {u(O) = 1, u(i) = 0, Consider a linear
í = 1, 2, 3, ... }, then the zero-state response yields the Markov parameters transfer matrix (;(z).
h(i)=y(i), i=O, 1,2, .... Clearly the z-transform of {h(i)} yields the transfer
function 9(Z). However, this approach requires the use of an infinite number of
{hU)}. Ir we form a Hankel matrix from {hU)}, and if the system is known to where D(z) and N(z)
have a degree bounded by N,1 o then we need only 2N + 1 of the hU), i = 0, 1, ... , Let the highest degree
2N. From the Hankel matrix, we can readily obtain a dynamical-equation expressed as
description of the system as shown in (6-35) to (6-38). Ir the transfer function
description is desired, we have
and
f3 oza +(J 1 za -1 + ... + f3 a
gA()
Z = where Di> Ni' i = 0, 1, .
. za +(XI Za 1 + . . . + (Xa
apply the p x 1 input SI
where (Xi are obtained from the Hankel matrix as in (6-36) and (Ji can be com­
puted from
°
l
h(O) to the initially relaxed ~
h(1) h(O)

~1(<T) :h(<T-l)
The substitution of I
This matrix equation is just the set of equations in (6-29) arranged in matrix
formo Consequently, the transfer function description of a system can be ob­
tained from a finite number of Markov parameters. This assertion also applies Equating the coefficien
to the multivariable case.
In order to obtain Markov parameters, the system must be initially relaxed.
In a multivariable system with p inputs and q outputs, we apply an impulse
sequence to the first input terminal and no input to all other input terminals,
then the responses at the outputs yield {h k1 (i), k = 1, 2, ... , q; i =0,1,2, ...}.
After the system is at rest again [in theory, after an infinite time; in practice,
after h(i) is practically zero or becomes almost periodic 11 ], we then repeat the
process for the second input terminal and so forth. Hence the measurement of
Markov parameters is possible only ií the system. is at OUí disposal. Ií a system
is in continuous operation, its Markov parameters cannot be measured.

where M = [-No Do
10lfno bOllnd 01 the degree of a system is available. then it is theoretically impossible to identify the matrix Si -Y, 00) has (
system. For example, ifwe have [h(O) =0. h(i) = 1, i = 1,2..... 200) and ifthe degree ofthe system columns. This equatio!
is bOllnded by 10, then we have g(z) = I/(z -1). However the system I/(z -1) + l/z 1oOO may also
generate the given seqllence. Hence if no bOllnd is available. there is no way to identify a system
Hence, given an arbitra
from a finite seqllence of {hU), i =0,1,2, ... , N}. The problem of finding a transfer fllnction to to the search of M to ro
match a finite sequence of Markov parameters is caBed the partial realizatioll problem (see Refe­ rows of nontrivial solu
rences 68, S 126 and S239). Sv( -Y, (0). Since it is,
II ¡·f ~ ·system is not BI BO stable (see Chapter 8), then h(i) will approach infinity or l'emain oscil­
as possible, we search,
Jatory (including approach a nonzero constant). In the former case. the system will satllrate Ol'
bum out and th~ linear model is no longer applicable. In the latter case, the system can be brollght Sv( - Y, (0). Hence the
to rest by resetting. In thebry, thereaÚzati~n ~~ identification is applicable no maller the system dependent rows of Sv('
is stable or not so long as the data are available. similar to the coprime I
--------~---~---_._._. _ ..~-~~._._ ...-._.. __ ...... "

.ND IDENTIFICATION IDENTIFICATION OF OISCRETE-TIME SYSTEMS FROM NOlSE..FREE DATA 303

of noise and call the In the following, we shall discuss a method of identifying a system from
an arbitrary input-output pair. The concept of persistently exciting will be
~te-time system with introduced. The problem of nonzero initiai conditions will also be discussed.
ce {u(O) = 1, u(i)=O, Consider a linear time-invariant discrete-time system with q x p proper
Markov parameters transfer matrix G(z). Let G(z) be factored as
')} yields the transfer G(z) = D-l(z)N(z) (6-190)
an infinite number of
where D(z) and N(z) are, respectively, q x q and q x p polynomial matrices.
: system is known to
Let the highest degree of all entires of D(z) be v. Then D(z) and N(z) can be
)f the hU), i = O, 1, .... ,
expressed as
dynamical-equation
the transfer function D(z) = Do + D 1z + ... + Dvzv (6-191)
and N(z) = No +N 1z + '" +Nvzv (6-192 )

where Di, Ni' i =0, 1, ... , vare q x q and q x p real constant matrices. If we
apply the p x 1 input sequence
;) and f3i can be como.
u(z) = u(O) + U(l)Z-1 + u(2)z- 2 + . . . (6-193)
to the initially relaxed system, then the output is a q x 1 seq uence given by
y(z) = G(z)u(z) = y(O) + y(l)z -1 +y(2)z- 2 + ... (6-194)

The substitution of (6-190) into (6-194) yields D(z)y(z) = N(z)u(z) or


') arranged in matrix - N(z)u(z) + D(z)Y(z) =0 (6-195)
. a system can be ob­
assertion also applies Equating the coefficient of z¡, i = v, v - 1, ... , - 00, to zero yields

st be initially relaxed. o O u(O) u(l)


we apply an impulse O O y(O) y(l)
)ther input terminals, O u(O) u(l) u(2)
, ... , q; i =0,1,2, ...}. [-No Do -NI DI .. , -N v DvJ O y(O) y(l) y(2)
nite time; in practice,
L], we then repeat the
u(O) u( v - 1) u( v) u( v + 1) ...
:e the measurement of
iLJ,¡(n¡ y(v -1) y(v) y(v + 1) ...:
disposal. If a system v,

t be measured. éMSA-v,oo)=O (6-196)

where M = [ - No Do ... - N v D v] is a q x (p + q)(v + 1) matrix and the


I impossible to identiry the matrix SJ - v, 00) has (p +q)( v + 1) number of rows and an infinite number of
1ifthe degree of the system columns. This equation is applicable no matter what input sequence is applied.
°
/(z - 1) + l/z ' OO may also Hence, given an arbitrary input-output pair, the identification problem reduces
lO way to identify a system
ding a transfer function to
to the search of M to meet (6.. 196). There are q rows of M. In order to have q
,lization problell1 (see Refe.. rows of nontrivial s~lutioris in (6-196), we !leed q linearly dependent rows in
Sv( - v, 00). Since it is desirable to have v, the degree of D(z) and N(z), as small
:h infinity or remain oscil.. as possible, we search, roughly speaking, the first q linearly dependent rows of
lhe system will salurate or
S.{-v, 00). Hence the ldentification problem reduces to the search of linearly
. the system can be broughl
cable nomatter the system
dependent rows of Sv( - v, 00) in order from top tobottom. This problem is
similar to the coprime fractiori problem discussed in Section G~4 .

.1..
304 IRREDUCIBLE REALIZATlONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFlCATION
IDENTlFi

The systems to be identified are assumed to be causal and to have proper


form an space of an
rational transfer matrices. Hence if u(k) = O for k < 0, then we have y(k) =0 for
sistently exciting. It r
k<O. Define
ability for the input to
ü(k,/)~[u(k) u(k+1) u(l)] (6-197) Consider So(k, 00).
y(k, I)~ I[y(k) y(k + 1) y(l)] (6-198) search the linearly der
Because ü is persistentI'
and define in y(k, 00 ). Let ro be tb
and q-r o is the num!:
ü(k, 1)
apply the row-searchin!
y(k, 1) ~ependent rows in y(k
----------
ü(k + 1, 1+1) Yi(k, 00) is Iinearly depe
S,,(k, I)~ y(k + 1, I + 1) ()( =0,1,2 ... (6-199) dependent in SI(k, a:.
----------
.' [u(k): ü(k + 1,00)] and.
ü(k +IX, I +IX)
we have 1'1 ~ro. We co
y(k +IX, I +IX)
ü(k,oo)
Y(k, (0)
There are ()( + 1 block rows in S". Each block row consists of p rows of ü and Ü(k +-( 00)'
q rows of y, hence S,,(k, l) is of order (p +q)(1X + 1) x (1- k + 1). We call the
rows formed from ü u rows and the rows formed from y y rows. ~~k_ ~}~ 00)
There are infinitely many equations in (6-196). lt is much more than neces­ Sy(k, 00) =
sary to solve for M. In fact, M can often be solved from MSy(O, l) =0, with I ÜC"k-+~-~ i: ~
chosen so that SvC0, 1) has more columns than rows. The range of data {k, I} Y(k +v -1, C
used in the identification is not critical ir. the system is initially relaxed. How­ Ü(k +-v: 00)­
ever, it becomes critical, as will be discussed later, ifthe nonzero initial conditions
y(k +v, 00)
are to be identified as well.
with 0::$ ro :::$ l' 1 ::$ ... ::$1
j ~ v. Note also that alll
Definition 6-4
assumption of persisten1
An input sequence {u(n)} is called persistently exciting if every u row in (6-199) y rows in S y(k, 00) is clcar
is linearly independent of its previous rows in S,,(k, 1). •
n~number ,
This definition, as such, is not well defined because of its dependence on k, 1,
and Ci.. The integer k is usually chosen as 0, although other value is also per­ =(q-ro)+
mitted. In theory, the integer I should be infinity; in practice, it is chosen so that
S,,(k, 1) has more columns than rows. The integer IX should be equal to or larger
. Let ~\, i = 1, 2, ... , q, I
than, as will be discussed later, the observability index of any irreducible realiza­
mdependent Yi in Sy(k, 00).
tion of G(z) or the largest row degree of row reduced D(z) in any left-coprime ... , y¡(k +v i -1, (0) are lir
fraction of G(z) = D -1(z)N(z). and Yi(k +1,00), I=v· v.'
Then Yi is said to have ;~d~),
The persistent exciting of an input sequence is not defined solely on the input
signal; it also depends on the output it generates. In other words, whether or and
not an input is persistently exciting depends on the system to be identified.
An input sequence which is persistently exciting to a systein may not be so to a
different system. Roughly speaking, an input sequence must be sufficiently
random or .rich in order to identify a system. Since the space spanned by the
rows of Si k; 1) is of a finite dimension, whereas al! possible input sequences 12The subsequentanalysis is símil;
_~~ __.__.. , ..
,~._~~._~_~.~c.~==~~==========

E, AND IDENTIFICATION IDENTIFlCATION OF DlSCRETE-TIME SYSTEMS FROM NOlSE-FREE DATA 305

.sal and to have proper form an space of an infinite dimension, almost aH input sequences are per­
hen we have y(k) =0 for sistently exciting. It means that if an input is generated randomly, the prob­
ability for the input to be persistently exciting is almost lo
Consider So(k, 00).12 We use, for example, the row searching algorithm to
(6-197 ) search the linearly dependent rows of So(k, 00) in order from top to bottom.
(6-198) Because ú is persistently exciting by assumption, aH dependent rows will appear
in y(k, 00). Let Yo be the number of dependent rows in y(k, 00). Clearly Yo'::;; q,
and q - Yo is the number of linearly independent rows in y(k, 00). Next, we
apply the row-searching algorithm to SI(k, 00). Let y[ be the number oflinearly
dependent rows in y(k + 1, 00). Let Yi(k, 00) be the ith row of y(k, 00). If
Yi(k, 00) is linearly dependent in So(k, 00), then y¡(k+ 1,00) will also be linearly
dependent in SI(k, 00). This foHows from the fact that ü(k, 00) =
,2 ... (6-199)
[ u(k) : ü(k + 1, oo)J and y(k, 00) = [y(k): y(k + 1, 00)]. Because of this property,
we have Yl '2:Yo. We continue this process until Yv =q as shown.
ü(k, 00)
~~k.! .ex:} _____ } Yo [no. of dependent rows in y(k, 00)]
nsists of p rows of ü and ü(k +1, 00)
:(l-k+1). WecaHthe y(k + 1, 00)
-----------
} Y[
l y Y rowS. Sv(k, 00) = (6-200)
.s much more than neces­
ü(k +v -1,00)

'rom MS.(O, 1) = O, with 1

y(k+v-1,00)
}Y v -l

The range of data {k, l}


------------
s initiaHy relaxed. How­
ü(k +v, 00)

nonzero initial conditions


y(k +v, 00)
} Yv =q
with O'::;;YO'::;;Yl'::;;'" ::::;Yv-I'::;;Yv=q. Note that if Yv=q, then Yj=q for aH
j '2: v. Note also that aH the ü rows in S.(k, 00) are linearly independent by the
assumption of persistently exciting. The number of linearly independent
y rows in Sv(k, 00) is clearly equal to
.g if every u roW in (6-199)
I
n~ number of linearly independent y rows
v-l
~ of its dependence on k, 1,
'h other value is also per­
=(q-¡'o)+(q-t'¡}+··· +(q-t'v_¡)=vq- I. Yi
i=O
:ractice, it is chosen so that
lOuld be equal to or larger Let Yi' i = 1, ;2, ... , q, be the ith row of y. Let Vi be the number of linearly
. of any irreducible realiza­ independent Yi in Sv(k, 00). By this, we mean that the rows y¡(k, 00), y¡{k + 1, 00),
d D(z) in any left-coprime ... , y¡(k + Vi - 1, 00) are linearly independent of their previous rows in Sv(k, 00),
and Yi(k + 1, 00), 1= Vi, Vi + 1, ... , V- 1 are linearly dependent in S.(k, 00).
defined solely on the input Then Yi issaid to have index Vi' Cle¡uly, we have Vi'::;; V, V= max{ Vi' i = 1,2, ... , q},
n other words, whether or and
.e system to be identified.
n = VI +V 2 + ... +Vq (6-202)
system may not be so to a
jence must be sufficiently
: the space spanned by the
1 possible input sequences 12The subsequent analysis is similar to the one in Section G-4.
--~-
· --- , ,.. ,._"-~ ,~~""-~ _ ",._.., -.~-'''-~'~.''.'_.~ ,_.

306 IRREDUCIBLE REALIZATlONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTIFICATlON


IDENTII

As an example, suppose q = 3 and


ü(k, 00) x
Yl(k, (0)
Y2(k, 00)
y3(k, 00)
---------
ü(k +1, 00) x
Yl(k +1, 00)
~ }r
K (6-203) Because of the assum
Y2(k+1,00) = 1 = 1 d~3 = 1, and all elemei
Y3(k+l,00) x are zeros. In additiOJ
---------
u(k +2, 00) x for 1= 1, 2, 3, .. " Hel
Yl(k +2, (0)
Y2(k +2, 00)
Y3(k +2, 00)
and
where K is a lower triangular matrix with 1 on the diagonal, as discussed in
(A-7) of Appendix A; and x denotes nonzero rows. Then we have VI = 2, V2 = O, d? 1 + d~ lZ + di
and V 3 = 2, and y 1 has index 2, y2 has index 0, and y3 ha$ index 2. D(z)= dg 1 +dLz
[
The Y¡, i = 1, 2, ... , q, which first becomes linearly dependent on its previous d~ 1 + d!31 z + d~
rows in SI' are caBed the primary dependent rows. For example, the Y2 in the This D(z) is clearly ro,
first block row of (6-203) is the primary dependent row; the Y2 in the second however, is not needed
and third block rows are not. The y1 and y 3 in the third block-row of (6-203)
. are .the primary dependent rows. The positions of the primary dependent rows
are clearly determinable from the index of y¡. The primary dependent yi row Because aH elemen ts or
appears in the (Vi + l)th block row of SI" zeros, the row degree of
D(z). Hence D -l(z)N(;
Theorem 6-13 D(z) are left coprime. S
R(z) with deg det R(z) >
Consider an initiaBy relaxed system excited by a persistently exciting input
sequence. We form Sa(k, 00) and search its linearly independent rows in order 1\
[rom top to bottom to yield KS v =Sv, where K is a lower triangular matrix with This implies deg det D(z
1 on the diagonal, and V is the first integer such that aH y-rows in the last block the n = ¿
Vi computed in
row of SI' are linearly dependent. Let [ - No Do:' .. : - N v D.,,] be the q rore we conclude that N(
rows of K corresponding to the q primary dependent rows of SI" Then the fer matrix is equal to del
transfer matrix of the system is given by G(z) = D- 1 (z)N(z), where That D(z) is in the pol
in Appendix G. This co
D(z)=D o +D 1 z+'" +Dvzv (6-204a)

and N(z) = No +N 1 z + ... +Nvzv (6-204b)


An example will be g:
Furthermore, D(z) and N(z) are left coprime, deg (;(z) = deg det D(z) and D(z) is
to iIIustrate the identifi,
row reduced, column reduced, and actually in the polynomial echelon form
matrix, in the coprime fi
(see Appendix G).
equation description can
in Section 6-6,
Proof13 i
I Persistently exciting
In order not to be overwhelmed by notations, we use an example to prove the variable system can b~ a

13The proof is identical to the proof of Theorem G-14.

j 14 The matrix is essenlially in n


.ND IDENTIFlCATION . IDENTIFICATlON OF DISCRETE-TlME SYSTEMS FROM NOlSE-FREE DATA 307

(6-205 )
Because of the assumption VI = 3, V2 = 1, and V 3 = 3, we have df¡ = 1, d~2 = 1,
(6-203) d~3 = 1, and aH elements in (6-205) on the right-hand sides of dfl' dL, and d~3
are zeros. In addition, the columns associated with dit 1, d~;Z, d%;l are zero
for 1= 1,2, 3, .... Hence the Di reduce to 14
3 ' d 111 O dI1 3 '' (11
do11 do12 dIo' O d 21 3 '' :~)
'1"' O O]
D3J=[d~1 d~2 d~3:
[2

[Do DI D 2 dLQ)O : O O O : O O O
q l ' d2
do31 do32 do' , 31 O d 33'' O O'1-'
1 2
~
33 ' d 31 O d 33
1 I I
: ••'

and
d~3 +dt3 Z +d"f3 Z2 J
[
.gonal, as discuss ed in d~1 +dLz+dLz2+z3 d~2
wehavevl =2,v 2 =0, D(z)= d~1 +d~lZ d~2 +z d~3 (6-206)
,index 2. d~1 +d11Z+d~IZ2 d~2 d~3 +d13Z +d~3Z2 +z3
)endent on its previous
example, the y2 in the This D(z) is clearly row reduced. It is also column reduced. (This property,
1; the Y2 in the second
however, is not needed here.) Therefore we have
~d block-row of (6-203) deg det D(z) = VI +v 2 + v3 = n (6-207)
rimary dependent rowS Because aH elements bn the right-hand side of d~2' di l' and d%3 in (6-205) are
nary dependent y¡ row zeros, the row degree of N(z) is at most equal to the corresponding row degree of
D(z). Hence D- 1 (z)N(z) is a proper transfer matrix. We claim that N(z) and
D(z) are left coprime. Suppose not, then there exists a q x q polynomial matrix
R(z) with deg det R(z) > Osuch that
sistently exciting input N(z) = R(z)N(z) D(z) = R(z)D(z) (6-208)
lependent roWS in order This implies deg det D(z) > deg det D(z). However, this is not possible because
r triangular matrix with the n = LVi computed in the algorithm is unique and smaHest possible. There­
y-rows in the last block fore we conclude that N(z) and D(z) are left coprime, and the degree of the trans­
.: -:N, D J be the q fer matrix is equal to deg det D(z) = Vi. L¡
t rows of S,. Then the
That D(z) is in the polynomial echelon form foHows from the definition given
N(z), where in Appendix G. This completes the proof of this theorem. Q.E. D.
(6-204a)
(6-204b)
An example will be given, after the diseussion of nonzero initial conditions,
to illustrate the identification procedure. We note th.at once the transfer
=deg det D(z) and D(z) is matrix, in the coprime fractional fom, of a system is identified, a dynamical­
olynomial echelon form equation description can be readily obtained by using the procedure discussed
in Section 6-6.

Persistently exciting input seq.uences. The identification of a multi­


variable system can be achieved by searching the linearly dependent rows of
. an example to prove the

14 The matrix is essenlially in the echelon formo See Appendix G-l


308 IRREDUCIBLE REALIZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION lDENTIFI

SAk, (0). In practice, we use the matrix Sv(k, l) for a finite l. Clearly, 1must be has a full row rank. '
larger than (v + 1)(P + q) to ensure that there are more columns than rows. In full row rank. The CO]
actual computation, once y¡(k + Vi, l) is Iinearly dependent in SV¡(k, 1), then y¡ is the Hankel matrix o
may be deleted in forming SV¡+ j, j = 1, 2, .... tions for all rows of U(
If u is persistently exciting. then the Ni and Di computed by using the row from the results in Seco
searching algorithm have the properties that D(z) is row reduced and, con­
sequently, nonsingular and D - 1 (z)N(z) is proper. Ifü is not persistently exciting,
then the computed D-I(z)N(z) may not be proper. For example, consider and
g(.:) = 1/(.:: + 1). If we apply u(z) = 1 +.:: - 1, then the output )";(z) is equal to z - l.
We form where u¡(z) is the ¡th COi
tions state that the inp
ifwe useS v (l, 00) in the i
Ü(O, (0)] [1 1 O O O ...]
Hankel matrix of u(z)/"
SI = ht~~-:~ = ~--Ó--Ó-~--6- -:-:~ (6-209) to be linearly independ¡
[
y( 1, (0) 1 O O O O ...
Clearly ü(l, 00) is linearly dependent on its previous rows: hence u(z) is not and
persistently exciting. If we solve (6-209) by using the first linearly dependent
row ofS¡, then the solution is The conditions in (6-21.
ample, one of u¡(z) may
[-1 1: 1 O]SI =0 (6-213). See also Proble
A 1 -z A
case, we may choose an
and y(z) = -1- u(z) (6-210) If the chosen or given i
appear as linear depenc
We see that this g(z) is not proper and is erroneous. Fortunately, this problem automatically whether o
can be automatically detected in the search of linearly independent rows of Si'
1t is possible to obtain a different but equivalent equation of (6-196) by Nonzero initial cond
grouping all u rows at the upper half of S. and all y rows at the lower half. For ofsystems which are not
this example, the equation becomes and observable discrete-I

[ -No DI][~¡~-:¡I~ [-1 (6-211 )

y(l,oolJ

and the solution is also equal to g(z)=(l-z)/l. Since all u rows of SI in y(z) = C(z]
(6-211) are linearly independent by themselves, there is no way to check from
which can be written as
SI that g(z) is erroneous. Hence the persistent exciting of {u(n)} cannot be
defined solely on {u(n)}.
y(z)
A necessary condition for an input sequence to be persistently exciting is
that the matrix where R(z) is a q x 1 poly

U(k' (0)
_ u(k+l,oo)
J to the corresponding ro",
We rewrite(6~217) as
ú(k, (0) - . (6-212)
[
u(k +v, (0) [ -R¡
lDENTlFICATION Of DISCRETE-T1ME SYSTEMS FROM NOISE-FREE DATA 309
, AND lDENTlFICATlON

el. Clearly, lmust be has a ful1 row rank. We give sorne sufficient condition for U(k, 00) to have a
olumns than rows. In full row rank. The condition depends on the value of k. If k = 1, then U(l, 00)
ent in Sv¡(k, 1), then J¡ is the Hankel matrix of u(z). Hence, if u(z) is a rational vector, then the condi­
tions for all rows of U(l, 00) to be linearly independent are, as can be deduced
'uted by using the row from the results in Section 6-4,
ow reduced and, con­
bu(z)::?:p(v +1) (6-213a)
ot persistently exciting,
and bí/¡(z)::?: \' + 1 i = 1,2, ... , p (6-213b)
For example, consider
,ut y(:) is equal to Z-I. where u¡(z) is the ith component of u(z) and b denotes the degree. These condi­
tions state that the input signals must be more complicated than the system
ifwe useS v(l, 00) in the identification. If k = - v, then the matrix U( - v, 00) is the
Hankel matrix of u(z)/zv+ l. Therefore the conditions for al1 rows of (6-212)
.] (6-209) to be linearly independent are

b[U(Z)/ZV+I] ~p(v + 1) (6-214a)


and b[u¡(Z)/ZV+l]::?:v+1 i=1,2, ... ,p (6-214b)
rows; hence í/(z) is not
FIrst linearly dependent The conditions in (6-214) are less stringent than the ones in (6-213). For ex­
ample, one of u¡(z) may have degree Oand satisfies (6-214), but wil1 not satisfy
(6-213). See also Problem 6-26. In practice, we do not have v a priori. In this
case, we may choose an upper bound v* ~ v and use v* in (6-213) or (6-214).
(6-210) Ir the chosen or given input sequence is not persistently exciting, then ü will
appear as linear dependent rows in Sao Therefore the procedure will check
lrtunately, this problem
automatical1y whether or not the input sequence is persistently exciting.
independent roWS of Si'

equation of (6-196) by
Nonzero ¡nitial conditions. In this subsection we study the identification
s at the lower half. For
of systems which are not necessarily initia\ly relaxed. Consider the control1able
and observable discrete-time equation

x(k + 1) = Ax(k) + Bu(k) (6-215a)


y(k) = Cx(k) + Eu(k) (6-215b)
O]SI =O (6-211)
The application of the z-transform to (6-215) yields

y(z) = C(zI - A)-lZX(O) + [C(zI - A)-IB +E]u(z) (6-216)


lce al1 u rows of SI in
s no way to check from which can be written as
ing of {u(n)} cannot be
y(z) = O -l(z)R(z) + O -l(z)N(z)u(z) (6-217)
;: persistently exciting is
where R(z) is a q x 1 polynomial matrix with row degree smaller than or equal
to the corresponding row degree of O(z). Clearly, R(z) is dependent on x(O).
We rewrite (6-217) as
(6-212)
(6-218)
310 IRREDUCIBLE REALlZATIONS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATION
IDENTIFlC

This equation is similar to (6-195) and implies hence the Ni, Di compl
o
O 1 O with the Ni, Di computt
puted by solving success
O ... O u(O) u(l)
In other words the solut
O ... O y(O) y(l)
-----------------
...
--------- MS v( - v, oc!) = O. Hent
O 1 O O and the ones computed J
O ... u(O) u(l) u(2)
[ -Ro -No D o·: ... :. -R v -N v D,.] lar polynomial matrix (1
O .. . y(O) y(1) y(2) matrices are the same.
---------------- ----------
---------------- ---------- Example 1
1 O O O
u(O) u(v -1) u(v) u(v + 1) ... Consider a system with t
y(O) y(v -1) y(v) y(v + 1) .. ,
= MSi - v, (0) = O (6-219)

whereM=[-R o -No Do:"': -R,. -N v DvJ, andSv(-v, (0) denotes


the v(p + 1 + q) x 00 matrix in (6-219). Now we require the input [1 u'(z)]',
where the prime denotes the transpose, to be persistently exciting. This input
[1 u'(z)]' violates the conditions in (6-213b); hence we cannot use Sv(k, 1)
An irreducible realization
with k;:,:: 1, in the identification. This new input however may meet the condi­
tions in (6-214), hence by using Sv( -v, 1), the system with its initial conditions
can be identified. x(k +1
ls it really necessary to identify the initial conditions? In the design, it seems
that initial conditions are never required. What is essential is the transfer
function 01' dynamical equation. Furthermore, if the system is not relaxed at y(k
n =0, and ifthe system is controllable and observable, then we can always find an
input sequence {u(n), n = - no, - no + 1, ... , -1} so that the system is relaxed With the initial statex(O) =
at n = - no and has the given initial state at n = O. Since the initial time is not pair
critical in the time-invariant case, a system with nonzero initial conditions at
n = Ocan be viewed as a system with zero initial conditions at 11 = - no. Con­ k o 1 2
sequently, (6-196) with a proper modification [since u(n) and y(n) are no longer u,(k) -0--0--=-1-­
zero for n < OJ can be used in the identification of the system. Since (6-196) li2(k) O O 1
contains more than enough equations and since we do not have the information .I',(k) 05 3.5 05
of o(n) and y(n) for n < 0, we simply use MSik, ro) = O, fOl" any k 2: 0, in th;;
identification. Consequently, given a system and given {u(n), y(n), i = O, 1,2, ...},
no matter the initial conditions of the system are zero 01' not, its transfer matrix Now we shall try to id
can be obtained by solving MSiO, (0) = O. {u(n), y(n)}.First we aSSurr
row-searching algorithm to
Theorem 6-14
1
Given a linear time-invariant system and given {u(n), y(n), n =0,1,2, ...}, the O 1
transfer matrix of the system obtained from Sv(O, (0) by assuming zero initial _1 -3
condition and the one obtained from Sv( - v, (0) without the assumpÚqn of zero 2 1
initial conditions are the same. -6 -8 -2 1
O - 1 O O 1
Proof O O O O O
Consider (6-196) and (6-219) with (6-196) modified as MSiO, (0) =0. First -3 -4 -1 O
we note that, by deleting the zero rows, the matrix Sv(l, (0) reduces to Sv(l, (0); -6 -9 -2 -1
IDENTlFICATION OF DISCRETE-TIME SYSTEMS FROM NOISE-FREE DATA 311
:, AND IDENTlFlCATION

hence the Ni, Di computed from (6-219) will also satisfy (6-196). Conversely,
with the Ni' Di computed from (6-196), a set of R i, i=O, 1, ... , v, can be com­
1
UrO) u(1)
°
puted by solving successively the first v + 1 column equations ofMS v( -v, (0) =0.
In other words the solution M of MSJO, (0) = O is also a part of the solution of
y(O) ?'~ 1) -­ MS)-v, 00)=0. Bence the D(z) and N(z) computed from MSv(O, 00)=0,
------
O and the ones computed from MS v( - v, (0) = O will differ at most by a nonsingu­
O
u(1) u(2)
lar polynomial matrix (dual of Theorem G-13). Consequently, their transfer
matrices are the same. This completes the proof of the theorem. Q.E.D.
y( 1) y(2)

----------
______ 6 _
Example 1
o o
Consider a system with transfer matrix
-1) u(v) u(v +1) .. ,
y( v + 1) ...

z~ 1]
- 1) y( v)
(-v, (0)=0 (6-219)
(6-220)
and Sv( - v, (0) denotes 2z
~e the input [1 u'(z)]', z-1
lly exciting. Thi~ input
we cannot use Sv(k, 1) An irreducible realization can be found as
ter may meet the condi­
+ 1) =[~
ith its initial conditions

;? In the design, it seems


x(k
°lJX(k)+[ 1
-0.5
1JU(k)
1
(6-221 )

essential is the transfer


;;ystem is not relaxed at
y(k) = [l ~J x(k) + [~ ~] iJ(k)
en we can always find an
With the initial statex(O) = [0.5 3.5]', we can obtain the following input-output
lat the system is relaxed
pair
lce the initial time is not
,ero initial conditions at
k O I 2 3 4 5 6 7 8 9 10 1I
tions at n = - no. Con­ -~---

u¡(k) O O -1 1 O O O 1 -1 O O O
'1) and y(n) are no longer
U1(k) O O 1 O O O 1 O O O O 1
e system. Since (6-196)
.r I (k)
:lOt have the information
.\·,(/d
0.5
,
3.5 0.5 3.5
()
3 3 3 4 5

2.5 55
7
25
1i
"
')
(¡ (: n u 1';,':

=0, for any k ¿O, in the


{u(n), y(n), i =0,1,2, .. '}' Now we shall try to identify the system from the input-output sequence
)r not, its transfer matrix
(u(n), y(n)}. First we assume that the system is initially relaxed and apply the
row-searching algorithm to S2(0, 11) as
1
y(n), n=O, 1,2, .. '}' the O I

by assuming zero initial _1


-3 I
2

Jt the assumption of zero -6 -8 -2 1

O -1 O O 1

-3° -4
° O
-1 °
_1
O 1

as MSJO, (0) = O. First


-6 -9 -2
A
_1 -1
O
-2
° 1
l, (0) reduces to Sv(1, (0); 2
°

312 IRREDUCIBLE REALIZATIONS, STRICf SYSTEM EQUIVALENCE, ANO IDENTIFICATION

o O -1 (1) O O O 1 -1 O O x (ü O O
O O 1 O O (~(~
O O O O O x O 0-1
O O ¡
0.5 3.5 0.5 3.5 Gt: 3 3 4 5 2.5 5.5 x
:5 ; 3 6 6 6 6 8 9 7 9 7 x 0.5 3.5 O.
** Ó- - ­ --.:: ¡ -----f -­ -- -0- - - - - 0- ---6.. -- -1­ - f:; --0- ----ó-----ó­
t~-
x * O 5 3 6
* .T·"--ü- ---ó----6
O :}) O O O 1 O O O O 1 x ,_J _,
3.5 0.5 3.5 3 3 3 4 5 2.5 5.5 2.5 O O 0:-11 1
3 6 6 6 6 8 9 7 9 7 11 O O O '-'
1
0.5 3.5 0.5 3.:
5 3 6 6
where the leftmost matrix is the F defined in (A-9). Its first column is chosen to
make the fourth column, except the first element, of S2(0, 11) a zero column. The last two rows of E
Its second column will make the seventh column, except the first two elements,
of K IS 2(0, 11) a zero column. Note that the location ofthe pivot element in
each row is encircled. Since the computation is carried out by hand, the pivot
element is chosen for convenience in computation. For this problem, we have
ro =0, rl =2 =q, and VI =V 2 = 1. The last two rows of K, which correspond
=[
to the linearly dependent rows ofSz(O, 11), can be readily computed by using the Hence we have
formula in (F-11) as

[ -No Do : -NI DIJ=[ -t 2


.1

-1 :-1
2
1 --34 ',,
_.1 •
2 ,
O
-1
O
-2
1
O ~J which'are the same as
For this example, we c;
Hence we have

N(Z)=L!} ;}+1]
D(z)= [ Z_¡2
+1
:~}J (6-222) Hence the result is corr

It can be readily verified that D-I(z)N(z) = G(z). 6-10 Concluding


Now we assume that the system is not initially relaxed. In this case we must
use (6-219). The application of the row-searching algorithm yields In this chapter, we disc
for proper rational mah
then use the procedure
1
second approach com¡:
O 1 formula in (6-29) and (6­
O O 1 singular value decompm
-0.5 _1 -3 1 realizations. In the las!
2
-5 -6 -8 -2 ¡ coprime fractions of st]
O O O O O 1 . this approach is the COI
.0 Appendix G can be em
O O -1 O O 1
irreducible realization m
O O O O O O' , 1 1
the Jordan-form realizati
_1. ' 1
:-3.5 -3 -4 -1 4 2 O O 1 reader is referred to Ref
-3 -6 -9 -2 _1 -5 1 -2 O
2 for example, References,
CONCLUDING REMARKS 313
AND IDENTIFICATION

o o x
o :}) O O O O O O O O O O x
o o x
O O O -1 eD
O O O 1 -1 O O x
O O O 1 O O O (f: O O O O x
2.5 5.5 x
O 0.5 3.5 0.5 3.5 el: 3 3 4 5 2.5 5.5 x
9 7 x
----Ó-----Ó­ x
** o
-_ ..
5 3 6 6 6 6 8 9 7 (9.~ 7
_-------------------------------------------------­
x
O 1 x
.'Í "
' ... "
O o
-,
o o o o o O o o o x
O O (- 11 1 O O O 1 1 O O O x
5 5.5 2.5 o
O
..
o -'1 :}l~ o o 1 O O O O 1 x
7 11 O
0.5 3.5 0.5 3.5 3 3 3 4 5 2.5 5.5 2.5 O
5 3 6 6 6 6 8 9 7 9 7 11 O
rst column is chosen to

The last two rows of K can be computed as


!(O, 11) a zero column.

. the first two elements,

Jf the pivot element in

O :_l O O:,
~J

l l 3 t 1 I
1
out by hand, the pivot
= I 4 2, 2 -4 : -Z ,
. this problem, we have
[ O'
, l2 -1 : -1 -1':-5:-1 -2: O
f K, which correspond

Hence we have
computed by using the

o O 1
N(z) =[ ¡
z_l
2

-1 -2 O which are the same as (6-222) obtained by disregarding the initial conditions.
For this example, we can also verify that
n-1(z)R(z) = C(zI - A)-l zx(O)
1
2 (6-222)
Hence the result is correct.

6-10 Concluding Remarks
d. In this case we must
rithm yields In this chapter, we discussed three approaches to find irreducible realizations
for proper rational matrices. The first approach is to find a reducible one and
then use the procedure in Section 5-8 to reduce it to an irreducible one. The
second approach computes the Markov parameters by using the recursive
formula in (6-29) and (6-30) and form the Hankel matrix. We can then use the
singular value decomposition or the row searching algorithm to find irreducible
realizations. In the last approach, irreducible realizations are obtained from
coprime fractions of strictly proper rational matrices. The majar effort in
this approach is the computation of coprime fractions and the procedure in
Appendix G can be employed. Allhough we have intro.duced a number of
irreducible realization methods;the treatment is not exhausÜve. For example,
the Jordan-form realization IS not discussed for the general case.' The interested
reader is referred to References S38, S59, and S145. For other methods, see,
1 for example, References, S62, S63, and S191.
,,,., .-y ••• _ - - _ .•. ,. " •• ~._,., • , •. '._.,_.".",~ • . • ."_. _ , .• ~_ . ' • • • . • ~ .'_' ' ." ~ " 0 ' ' ' _ _ "... • ~ •• ~ •• _ • •

.............-­
.~.~~.' -~

.
- - - - ~ - ~_------~--'--~--~----,------~--_

314 IRREDUCIBLE REALIZATlONS, STRICT SYSTEM EQUIVALENCE, ANO IDENTIFICATlON

Now we compare briefly the numerical efficiencies of these three approaches. introduce errors due te
Given a q x p strictly proper rational matrix. Let a¡(s) and flj(s) be respectively will not be discussed.
the (least)common denominators ofthe ith row and jth column of G(s). Because In this chapter, we
the computation ofthe least common denominator is not a simple task, we may established its relation
simply multiply all denominators of the ith row and the jth column of G(s) We showed that if the 1
to yield a¡{s) and flis). For simplicity, we assume p = q and a = deg a¡(s) = deg irreducible, then they a
!3j(s), for all i,j. Then we can find a controllable but not necessarily observable them can be used in the
(or vice versa) realization of dimension ap. To reduce this reducible realiza­ tion.
tion, we must append B and e to A; hence the matrix under operation is roughly The degree of a prc
of order (a + l)p. In the Hankel matrix, the order of the Hankel matrix is degree can be compute
(a + l)p x ap. In the coprime fraction method, if we use the method discussed
in Appendix G to find a coprime fraction, we must form a generalized resultant. 1. Compute the least c
we have deg G(s) = (
Each block row has (p + q) = 2p rows, and we need roughly (a + 1) block rows.
2. Compute the Mark,
Hence the generalized resultant has 2(a + l)p number of rows. Thus we con­
shown in (6-80). n
elude that the matrices used in the first two approaches are roughly of the same
order; the matrix used in the third approach is about twice the sizes of those computed by using t
used in the first two approaches. 3. Find a right fraction
The operation in the first approach requires the triangularization of a matrix deg G(s) = deg det D
by similarity transformations (both column and row operations). In the second coefficient matrices e
and third approaches, we require the triangularization of a matrix by either row We then search the liJ
or column operations. Hence the first approach requires twice the computa­ Then we have
tions of the second and third approaches. Thus we conclude that the Hankel deg G(s) = too
method is probably the most efficient. For a more detailed comparison, see
. Reference S202. We note that all three approaches can.be implementeqby using Similar results can b(
numerically stable methods. 4. Consider the realizat
It is of interest to compare the singular value decomposition and the row­ is controllable but nc
searching method discussed in Section 6-5. They are implemented on UNIVAC QIO =0, Qlj = R,
1110 at Stony Brook in Reference S202. The row searching algorithm is carried QiO =0, Qij =Q(i-lH.
out by applying Householder transformations and the gaussian elimination
with partial pivoting on the columns of Hankel matrices and then computing Using Qij, the observ
the coefficients of combinations by back substitution. The scaling problem is
included in the programo The examples in Reference S202 do not show any
substantial difference between Householder transformations and the gaussian
eiimination with partiai pivoting. The singuíar vaiue decomposition is most
reliable. However, its reliability over the row searching method is not over­
whelming in the examples in Reference S202. This seems to be consistent with Then we have deg (;(5
the remarks in Reference S82, p. 11.23. 5. Find a realization of
The identification of linear time-invariant discrete-time systems is studied. Section 5-8, to an irred
The identification is carried out from a set of arbitrary input-output pair so
long as the input sequence is persistently exciting. Whether the system is These methods can be
initially relaxed or not, the transfer functíon matrix in coprime fractional parison of these method~
form can be obtained by solving MSv(O, ex)) = O. The realization problem can stability seems to be una\
be considered as a special case of the identification problem in which the system To conclude this chal
.. is to be identified from a particular input-output pair, the impulse response or realization in Equation (f
the Markov parameters. The identification method introduced is applicable in algebraic system theor
to continuous-time systeins ir they are first discretized. This, however, will q x p strictly proper ratic
AND IDENTIFICATION
CONCLUDING REMARKS 3]5

hese three approaches. introduce errors due to discretization. This is ou tside the scope of this text and
nd f3 J.(s) be respectively will not be discussed.
A

,lumnofG(s). Because In this chapter, we also introduced the polynomial matrix description and
t a simple task, we ~ay established its relationships with dynaroical equations and transfer matrices.
he jth column of G(s) We showed that if the three descriptions have the same transfer matrix and are
and ex = deg o:;(s) = deg irreducible, then they are aH strictly system equivalent. In thiscase, any one of
necessarily observable them can be used in the analysis and design without loss of any essential informa­
this reducible realiza­ tion.
ler operation is roughly The degree of a proper rational matrix is introduced in this chapter. The
, the Hankel matrix is degree can be computed by using any of the following methods:
;: the method discussed 1. Compute the least common denominator, L\(s), of all minors of (;(s). Then
a generalized resultant. we have deg (;(s) = deg L\(s).
ghly (o: + 1) block rows. 2. Compute the Markov parameters of (;(s) and form the Hankel matrix T
)f rows. Thus we con­ shown in (6-80). Then we have deg (;(s) = rank T. The rank of T can be
are roughly of the same computed by using the singular value decomposition.
twice the sizes of those 3. Find a right fraction G(s) = Ñ(s)D -1(S). Ir the fraction is right coprime, then
deg G(s) = deg det D(s). Ir the fraction is not right coprime, we use the
gularization of a matrix coefficient matrices of D(s) and N(s) to form the matrix Sk shown in (G-67).
:rations). In the second We then search the linear independent rows ofS k in order from top to bottom.
•f a matrix by either row Then we have
ires twice the computa­
Illclude that the Hankel deg G(s) = total number of linear independent N roWS in Soo.
etaHed comparison, see Similar results can be stated for a left fraction of G(s).
)e implemented by using 4. Consider the realizatíon of (;(s) shown in Equation (6-71). The realization
is controllable but not necessarily observable. Define
mposition and the row­
plemented on UNIVAC QIO=O, QIj=R m- j + 1 j=I,2, ... ,m
hing algorithm is carried QiO =0, Qij=Q(i-I)(j-l) -O:m-j+1Q(i-l)rn i =2, 3, ... , m~ j = 1, 2, ... , m
he gaussian elimination Using Q¡j, the observability matrix of (6-71) can be computed as
ces and then computing
The scaling problem is [QH Q12
QI~l
: S202 do not show any
lations and the gaussia n
V~. ~21 Q22 ~2m
. !

: decomposition is most LQml Qm2 QmJ

ing method is not over­


:ms to be consistent with Then we have deg (;(s) = rank V (why?).

5. Find a realization of (;(s) and reduce it, by using the method discussed in
-time systeros is studied. Section 5-8, to an irreducible one. Then we have deg (;(s) = dim A.
uy input-output pair so
Whether the system is These methods can be readily programmed on a digital computer. A com­
x in coprime fracUona\ parison of these methods in terms of computational efficiency and numerical
realization problem can stability seems to be unavailable at present.
blero in which the system To conclude this· chapter, we give a different derivation of the observable
. the impulse response or realization in Equation(6-131). lt wil\ be derived from three maps developed
introduced is applicable in algebraic systeffi theory. We· discuss the discrete-time case. Cons"ider the
~ed. This, however, will q x p strictly proper ratianal matrix G(z) = D-1(z)N(z), where D(z) and N(z)
316 IRREDUCIBLE REALIZATlONS, STRICT SYSTEM EQUIVALENCE, AND IDENTlFICATlON

are respectively q x q and q x p real polynomial matrices. It is assumed that If i = 1, j = 1, then D - 1(.
D(z) is row-reduced and its row degree is Vi' that is, b,.iD(S) = Vi' Clearly we
have b,.¡N(z) < Vi'
Let IK D denote the set of all q x 1 real polynomial vectors with row degrees
smaller than Vi' It can be readily verified that (IK D , IR) is a linear space over the
field of real numbers IR. The dimension of (IK D , IR) is n~ VI + Vz + ... + Vq.
(See Example S, on page 16.) In this polynomial vector space, we shall choose
the n columns of the q x n matrix:
1 z z v, - 1 O O ... This is the first column
O O O
columns of A except tr

J,]
_ O O O 1 z .. . Z "2 - 1 O O
(6-223) (6-131a). Consider now
L(z) = :: :::
.. ...
[ 0D(.H(z)) = D(z)O(
00 O 00 O 1z
Because Dh;.1 is a constal
as a basis. Its columns will be denoted by qij, i = 1, 2, ... , q; j = 1,2, ... , Vi'
O. Because D -1(Z)[(Z~
Let f(z) be a q x 1 rational vector. Define the operator O as
D- 1(z)[(z)D¡,.D';;.I. HeI
O(f(z)) = strictly proper part of f(z) (6-224)
OD(H(z)) =
If f(z) is a polynomial vector, then O(f(z)) = O; if f(z) is strictly proper, then
Note that the ith columl
O(f(z)) = f(z). It is a linear operator which maps (lRq(z), IR) into itself. Next
given by the ith column
we define, for any q x 1 real polynomial,
Now we show that th
OD(h(z)) = D(z)O(D -1(z)h(z)) (6-225) From (6-226), we have
It is a linear operator which maps (lRq[ zJ, IR) into itself. Because the ith row D-l(Z) =Dh;.I[H(z)(1 + f
degree of 0D(h'(z)) is at most V¡- 1, ihe range space of OD(h(z)) is (IK D, IR). =Dh~.I(I+E1Z-l
In algebraic system theory, a realization can be expressed by the following
three maps: for some constant E¡, i =
that if j < Vi, then (D -1(;
B D : IRP----* IK D ; u t-+N(z)u This completes the deriv;
A D : IK D----* IK D; x t-+OD(ZX) (6-226) three maps in (6-226). l
C D : IK D ----* IRq; x----*(D- 1(z)X)_1 maps in (6-226) are useful
SS5 and Section 9-8.
where, ifwe expand D- 1(z)x =lX 1Z- 1 + lXzZ- z + "', then (D- 1(z)X)_1 ~lXl' If
the columns qu in (6-223) are chosen as a basis of IK D , then the map R D is the
representation of N(z) with respect to the basis, that is,
N(z) = [(z)B 6-1 Find the degrees and Ü
matrices.
as shown in (6-130). The ith column of A is the representation of OD(Zqij) with

"ll'~:)'
s +3
respect to the basis in (6-223). Indeed, if we rewrite (6-12S) here as ~-

s +2 s+
D(z) = H(z)D h ,. + [(z)D¡,. (6-227 ) s +1

Then we have (s+3j2 s +4 s

Consider
H(z) = D(z)D,;;.1 - [(z)D¡I.D,;;.1
b.
l ' ~ l' +:)~d2)J
1
(s +2)
1)'

(s+1)(s+2)
_._-_.. -_._---._..- ..
----,-------------_.­
---_•... --~----

AND IDENTIFICATION
PROBLEMS 317

es. It is assumed that Ir i = 1,j = 1, then D- 1(s)Zqll is strictly proper and


iD(S) = Vi' Clearly we
o
ctors with row degrees 1
a linear space over the TID(Zqll) = D(s)D- 1(s)Zq11 = zqll = q12 = L(s) O
; n~v1+V2+'" +vq­
space, we shall choose O

This is the first column of A as shown in (6-131a). Proceeding similarly, all

~
O columns of A except the (¿~ = 1 vm)th columns can be obtained as shown in
O (6-131a). Consider now
] (6-223)
TID(H(z» = D(z)TI(D- 1(z)H(z» = D(z)TI[Dh~l - D- 1(z)L(z)D .D';;.1 ]
z z Vq - 1 "
Because D';; 1 is a constant matrix (a polynomial of degree O), we have TI(D¡~.l) =
... , q; j = 1, 2, ... , Vi' O. Because D-1(Z)L(z)D1J.Dh~1 is strictly proper, TI[D -1(z)L(z)D 1r Dh";.1 ] =
tor TI as D- 1(z)L(z)D l.. D,;;'1. Hence we have
z) (6-224 ) nDCH(z)) = - D(z)D-1(Z)L(z)Dl.. D,~1 = - L(z)D 1,.D¡;;' 1
) is strictly proper, then Note that the ith column of H(z) is zq¡v,. Hence the (2:::n= l)th column of A is
z), IR) into itself. Next given by the ith column of -Dl,.D/~.1, as shown in (6-131a).
Now we show that the ijth column of e in (6-131a) is equal to (D-1(z)qij)_ l'
(6-225) From (6-226), we have

f. Because the ith row D-1(Z) =D¡~.l[H(z)(I + H- 1(z)L(z)Dl")] -1 ~D/~.l(I + H-1(Z)L(zp¡.. t 1iI- 1(z)
D(h(z)) is(IK D , IR). =D,;;.1(I+E 1z- 1 +E 2z- 2 + ·.. )fi- 1(z)
lressed by the following for sorne constant E¡, i = 1, 2, . . .. From this equation, we can readily show
that if j<v¡, then (D-1(Z)qij)_1 =0 and (D-l(z)q¡v¡}=ihe ith column ofD,;;.l.
This completes the derivation of the observable realization in (6-131) from the
(6-226) three maps in (6-226). For a derivation of (6-226), see Reference Si02. The
maps in (6-226) are useful in solving polynomiaJ matrix equations; see Reference
SS5 and Section 9-S.
en (D- 1(z)X)_1 ~Cl1' If
then the map RD is the

6-1 Find the degrees and the characteristic polynomials of the following proper rational
matrices.
~ntation of TID(zq¡J with

'"ll' ~ :)' ~:51


s +3
-128) here as s +2
(6-227) s +1
-
(s+W s +4

bt ;1)'
(s +2)
Id:~d2)1
(s +1)(s +2) .'
318 IRREDUCIBLE REALIZATlÜNS, STRICf SYSTEM EQUIVALENCE, AND lDENTlFICATlüN

6-6 Find an irreducibll


s
S
+3l
+ 1 an unobservable dynami(
S
dynamical-equation reali

S +1
6-7 Find the controll~
6-2 Find the dynamical-equation description of the block diagram with state variables lordan canonical-form d:
chosen as shown in Figure P6-2.
6-8 Set up a linear time
(p3 +7. 1(t),
where pi ~ d/dti.

Figure P6-2
6-9 Find irreducible ce
matrices
6-3 Find the dynamical-equation realizations of the transfer functions

a'l l' +% ~2X' +3)J


S4 +1
a. 4s4 + 2s3 + 2s + 1
S2 +2s +2
S2 -s + 1
s(s +1)2(s +4)
b. SS _ S4 + S3 _ S2 +s- 1
2s +3
Are these realizations irreducible? Find block diagrams for analog computer simulations b. [
(5 +W(s +2) s(s
of these transfer functions.

6-4 Find lordan-canonical-form dynamical-equation realizations of the transfer func­ 6-10 Find irreducible re2
tions

a.
(S + 1)(s + 2)(s +3)

S2 + 1

b. - - " ­
(s +2)3
Use two different methods
S2 +1
c. S2 +2s +2
6-11 Find irreducible rea
lf the lordan-form realizations consist of complex numbers, find their equivalenl dynamical
equations that do not contain any complex numbers.

6-5 Write a dynamical equation for the feedback system shown in Figure P6-S. First
find an overall transfer function and then realize it. Second, realize the open-loop transfer
function and then make the necessary connection. Which realization is more convenient in
computer simulalions if the gain k is to be varied? Use two different methods.

6-12 Find a linea((ime-in


fer function is
(s + 1) (s + 2)(s + 3)

Figure P6-5 I
I
[See Problem 6-3(a).]

I
1
AND IDENTIFICATION PROBLEMS 319

6-6 Find an irreducible realization, an uncontrollable dynamical-equation realization,


an unobservable dynamical-equation realization, and an unobservable and uncontrollable
dynamical-equation realization of 1/(s3 + 1).

6-7 Find the controllable canonical-form, the observable canonical-form, and the
agram with state variables Jordan canonical-form dynamical-equation realizations of l/sO..

6-8 Set up a linear time-varying dynamical equalion for the differential equation
(p3 +a 1(t)p2 +C(z(t)p +C(3(t))y(t) = (f30(t)p2 + f31 (t)p + f3z(t))u(t)
-~
~ ­
6-9 Find irreducible controJlable or observable canonical-form realizations for the
matrices
'unctions

a. 'l(S +l)(s ~2)(s +3)j


S2 +2s +2
s(s + 1)2(5 +4)
2s +3 S2+ 2s +2 ]
lalog computer simulations b. [ (s+1)2(s+2) s(s+V

6-10 Find irreducible realizations of the ralional matrix


ltions of the transfer func­
S+2

~ :3J
Use two different methods.
ls: s +1

1
5

s +2
+1

6-11 Find irreducible realizations of the rational matrix


j their equivalent dynamical

lown in Figure P6-S. First


~alize the open-loop transfer
zation is more convenient in
Use two different methods.

6-12 Find a linear time-invariant discrete-time dynamical eq uation whose sampled trans­
fer Function is

i
I [See Problem 6-3(a).]
1

l.
320 IRREDUCIBLE REALlZATlüNS, STRICT SYSTEM EQUIVALENCE, ANO IDENTlFICATlüN

6-13 Find an irreducible, discrete-time dynamical-equation realization of the sampled 6-17 Show that (6-77) i~
transfer-function matrix k=O, 1,2, ....

Z~3J
Z+2
6-18 Show that (6-77) is
Z +1
Z Z +1 6-19 Consider the q x p
[
Z +1 Z +2

6-14 Consider lt is assumed that there el

x =[)"
O)"
~Jx +[~¡J

u
KoH(v+i)= -K¡H(
or [K
where the overbar denotes the complex conjugate. Verify that by using the transformation where f is defined as in (6­
x=Q¡x, where cal equation
-Xb¡ ~¡J _KVKÜ¡ I
Ql = [ -),E¡ b¡
. -K,,_¡K ü ¡ (
the equation can be transformed into x= :
[ -KzK ü ¡ (J

-KIK ü ¡ O
where
Y=[ Kü ¡ O
~xJ b=[~J
_ [ O
Á= cl =[ -2 Re (Xb¡cd 2 Re (b¡c¡)J
-AX A is a realization of G(s). Sh
controllable.
6-15 Verify that the Jordan-form dynamical equation
6-20 Use Problem 6-19 a
A 1 O O O O b¡
irreducible one. (Hint: Shil
O A 1 O O O bz
O O A O O O b3 O 2:
x= x+ u [O O'
O O O X 1 O El I

O O O O X 1 Ez and then proceed. This pre


O O O O O [ E3 in (6-131) except the rearr;
References S63 and S2ü1.)

can be transformed into


N(s) = N o ­
D(s) = Do'

where D" is assumed to be


dynamical equation
where Á, b, and c¡ are defined in Problem 6-14 and I z is the unit matrix of order 2. [Hint:
Change the order ofstate variables from [Xl X2 X3 X4 Xs X6J' to [XI X4 Xz Xs X3· X6J'
and then apply the equivalence transformation x = Qx, where Q = diag (Q¡, Qz, Q3)·J

6-16 Write an irreducible dynamical equation for the following simultaneous differential
equation:

2(p+l)y¡ +(p+l)yz=pu¡ +uz


(p + l)y¡ +(p +l)yz = (p - l)u¡ y= [ l'

wherep~d/dl. is a controllable realization


PROBLEMS 321
AND IDENTIFICATION

ealization of the sampled 6-17 Show that (6-77) is a realization of G(s) in (6-73) by establishing H(k +l)=CA k B,
k=0,1,2, ....

6-18 Show that (6-77) is always observable but not necessarily controllable.

6-19 Consider the q x p proper rational matrix


(;(s) = H(O) + H(l)s- 1 + H(2)s- 2 + ...

1t is assumed that there exist q x q constant matrices Ki, i =0,1,2, ... , v, such that

KoH(v + i) = - K¡H(v + i -1) -K 2H(v + i - 2) - ... -K,.HU) i= 1, 2, 3, ...


]x
or [K¡ K2 .. , Kv Ko O .,. OJT=O
IY using the transformation where T is defined as in (6-95). lf Ko is nonsingular, show that the vq-dimensional dynami­
cal equation

=[= ~:~,ó~", ~ ~ ...


O
x O]
O ~v
[Ko O
~o O

: x +: :

-K 2 K ü 1 O O 1 K3 K 4

-K 1 Kü l O O O K2 K3
O O OJ x + H(O) u
is a realization of (;(s). Show that the realization is always observable but not necessarily
controllable.

6-20 Use Problem 6-19 and (6-99) to find a realization of (;(s) in (6-97). Reduce it to an
irreducible one. (Him: Shift k 2 in (6-99) to right to forro
O 2:01:10:1 0J
[ O O: O 2: O 3: O 1 = [K¡ K2 K3 KoJ
u , , ,
and then proceed. This procedure can be used to find an irreducible realization of the form
in (6-131) except the rearrangements of state variables as discussed in Section 5-8. See
References S63 and S201.)

6-21Let N(s)D- l (s) be a q x,n strictly proper rational matrix and let
N(s)=N o +N l s+'" +N._ l s·- l

D(s) = Do + D l s +. " +D.s· = D.(D o + Dl s + ... + Is")

where D. is assumed to be nonsingular and Di = D; 1 Di' Show that the pp-dimensional


dynamical equation
lit matrix of order 2. [Hint:
6]' to [XI X4 X2 Xs X3 X6J'
Q=diag(Q1> Q2' Q3)·J

'ing simultaneous differential


hU -Do
1
O

O
-DI -D 2
O
1

O
. ] ["
O
:

-~I'-l'
x+:
O

~,' I
u

y=[ No NI N2 N.- 1 Jx
is a control1able realization of N(s)D-1(s).
'----------_.~-~~~._-

322 IRREDUCIBLE REALIZATlüNS, STRICT SYSTEM EQUIVALENCE, AND IDENTIFICATIüN

6-22 Consider the equivalent time-invariant dynamical equations {A, B, q and {A, B, Cj It is assumed that all eiger
with A=PAP- I, B=PB, and c=cp- 1. Let wer and W Ol be the contr01lability and system
observability grammians defined in Theorems 5-7 and 5-13. Show that
W C1 = PWerp* where
and \VOl =(P-1)*WOlP- 1

6-23 Consider the irreducible dynamical equation lA, B, q. Using Theorem E-4 to
is also internally balanced.
write its grammians as
(Hint: Write W=diag {W 1
and
6-26 Show that ror any sin.
where R*R = I and L 2 =diag [/,i, ;.~, ... , l.;':. Define
is used in the identification
H~L:R:RcLc nonzero input sequence is al
is used?
Using the singular value decomposition (Theorem E-S), we write

H=RHLHQH 6-27 Identify a linear timó

where R¡t;R H = I and QHQ¡t; = 1. Show that if P in Problem 6-22 is chosen as k O [ 2 3


i. Pi" = QHL c- I R¿ ------
u dk) [ [ -[ O
ii. POUl = R¡t;L:R:
u2(k) [ O O -[
iii. P ib = Pi"L¡F = L¡Fp OUl Yl(k) -[ [ 2 [
Then its equivalent dynamical equation {A, B, C} has the fol\owing corresponding gram­ Y2(k) -2 -0.5 [ -0.5
mians:
First identify the system by d
i. W" = 1, W01 = L~ as welI as its initial conditior:
ii. W" = L~, Wo, = I
¡ii. Wer = Wo , = LH 6-28. Let {A, B, q be irre(
and N(s) are right coprime.
They are ca1led, respectively, input-normal, output-normal, and interna//y balanced on
such that C¡(sl-A¡- lB = NI
[0, l] (see References SI61 and SI62).
N1(s)D-I(s), there exists a C I
6-24 Consider the irreducible dynamical equation

x=Ax +Bu y=Cx


[t is assumed that a1l the eigenvalues of A have negative real parts and that the equation
is internally balanced, that is.

w = w coo = W 000 = LH
where W is a diagonal matrix with positive diagonal entries. Show that W is the unique
solution of
AW +WA* = -BB* or WA +A*W = -C*C

(Hinl: Use Corollary F-Ib.)

6-25 Consider the interna1ly balanced system partitioned as


AND IDENTlFICATlON PROBLEMS 323

,ns {A, B, C} and [A, B, c: Jt is assumed that all eigenvalues of A have negative real parts. Show that the reduced
Je the controllability and system
ow that x= Ax +Bu, y= ex
where A=A¡¡ -A 12 Ai}A 21
B= B¡ -A¡zA:;jB z
C=C¡ -C2A221A21
Using Theorem E-4 to is also internally balanced. This result is useful in syslem reduction. See Reference S93.
(Hint: Write W =diag {W¡, W 2 } and use Problem 6-24.)
1*
'o
6-26 Show that for any single-variable system with a proper transfer function, ifS v( - \', IX))
is used in the identification, the impulse sequence l urO) = L u(i) = O, i = 1. 2, L .. ~ or any
nonzero input sequence is always persistently exciting. ls the statement still true ifS,.(O, IX))
is used?
e
6-27 ldentify a linear time-invariant system from the following input-output sequence:

2 is chosen as k . O 1 2 3 4 5 6 7 8 9 !O 11 12 13 14 ..
_
u\(k) 1 1 -1 O O 1 O -\ O \ O -\ -l -2 \
u2(k) \ O O -\ O O O \ -\ O -\ \ O 1 -2
)'I(k) -\ \ 2 1 O O \ \ \ O \ O O -\ -2
)'2(k) -2 -0.5 \ -0.5 -0.5 -0.5 \ \-0.5 -0.5 \ - \ -2.5 -4 -7
ving correspondíng gram­ -----­

First identify the system by disregarding the initial conditions and then identify the system
as well as lts initial conditions.

6-28. Lel {A, B, C} be irreducible and let G(s)=C(sl -A¡- IB= N(s)O-\(s), where O(s)
and N(s) are righl coprime. Show lhat for any C I , there exists a polynomial matrix N 1 (s)
ld internally balal1ced on such that C ¡(si -A)-I B = N I(S)O - I(S). Show also lhal, conversely, for any strictly proper
N 1(5)0-I(S), there exisls a C I sllch that N ¡(5)0-I(S) = C¡(sl - A)- ¡ B.

irts and that the equation

.how that W is lhe unique

= -C*C
form u=r+ Kx on a <
the linear state feedba<
of an uncontrollable d
In state feedback,
outputs. This assump
want to introduce stal

7 mated from the availal


state estirnators undei
asymptotic state estim:
State Feedback Both full-dimensional:
In Section 7-5 we a
and State Estimators We shall establish the :
and state estimator ca
not affect their intendl
not appear in the trans
output.
The inputs and ou
that is, every input COl
trolled by more than 01
7-1 Introduction input controls one and
to be decoupled. We:
In engineering, design techniques are often developed from qualitative analyses system by state feedbac
of systems. For example, the design of feedback control systems by using system can be decouple
Bode's plot was developed from the stability study of feedback systems. In We' study in this el
Chapter 5 we introduced two qualitative properties of dynamical equations: The material is based il
controllability and observability. In this chapter we shall study their practical 78 to 80,84,99,111,11
implications and develop sorne design techniques from them. can be extended to lir
Ir an n-dimensional linear time-invariant dynamical equation is con­ stantaneously controlla
trollable, the controllability matrix [B AB ... An -1 BJ has n linearly to References 10, 98, 99
independent columns. By using these independent columns or their linear
combinations as basis vectors of the state space, various canonical forms can
be obtained. We introduce in Section 7-2 the most usefuI one: the controllable 7-2 Canonical-F<
canonical formo We a!so introduce there the observable canonjeal form
dynamical equation. These two canonical forms are very useful in the designs Single-variable cas«
of state feedback and state estimators. single-variable dynamic
Consider a system, called a plant, and a desired or reference signal; the
control problem is to find a control signal or an actuating signal so that the
output of the plant will be as close as possible to the reference signa1. Ir a
control signal is predetermined and is independent of the actual response of
the plant, the control is called an open-loop control. This type of control is nol where A, b, e, and e are,
satisfactory if there are disturbances or changes in the system. Ir a control matrices. Note that th(
signal depends on the actual response of the system, it is called a feedback or this section we shall dI
closed-loop control. Since the ·state of a system contains all the essential in­ under the controllabilit)
farmation of the system,.if a control signal is designed to be a function of the Let the following dy
state and the reference signal, a reasonably good control can be achieved. In
Section 7-3, we study the effects of introducing a linear state feedbackof the
324
CANON¡CAL-FORM DYNAM¡CAL EQUATlONS 325

form u = r + Kx on a dynamical equation. We show what can be achieved by


the linear state feedback under the assumption of controllability. Stabilization
of an uncontroHable dynamical equation is also discussed.
In state feedback, aH the state variables are assumed to be available as
outputs. This assumption generally does not hold in practice. Therefore if we
want to introduce state feedback, the state vector has to be generated or esti­
mated from the available information. In Section 7-4 we introduce asymptotic
state estimators under the assumption of observability. The outputs of an
asymptotic state estimator give an estimate of the state of the original equation.
Both full-dimensional and reduced-dimensional state estimators are introduced.
In Section 7-5 we apply the state feedback to the output of a state estimator.
We shall establish the separation property which shows that the state feedback
and state estimator can be designed independently and their connection w¡-¡[
not affect their intended designs. We also show that the state estimator wiH
not appear in the transfer-function matrix from the reference input to the plant
output.
The inputs and outputs of a multivariable system are generally coupled;
that is, every input controls more than one output, and every output is con·
trolled by more than one input. If a compensator can be found such that every
input controls one and on]y one output, then the multivariable system is said
to be decoupled. We study in Section 7-6 the decoupling of a multivariable
om qualitative analyses system by state feedback. The necessary and sufficient condition under which a
ntrol systems by using system can be decoupled by linear state feedback is derived.
, feedback systems. In We study in this chapter only linear time-invariant dynamical equations.
f dynamical equations: The material is based in part on References 1, 4, 7, 17, 18, 36, 41, 46, 51, 52, 68,
¡all study their practical 78 to 80, 84, 99, 111, 112, S17, S23, S69, and SllO. The results in this chapter
1 them.
can be extended to linear time-varying dynamical equations if they are in­
lical equation is con­ stantaneously controllable (see Definition 5-3). The interested reader is referred
\."-1B] has n linearly to References 10,98,99,110, and S119.
:olumns or their linear
us canonical forms can
'til one: the controllable 7-2 Canonical~Form Dynamical Equations
rvable canonical forro
~ry useful in the designs Single-variable case. Consider the n-dimensional linear time-invariant,
single-variable dynamical equation:
)r reference signal; the
ating signa1 so that the X=Ax +bu (7-1 a)
y =cx +eu (7-1 b)
~ reference signa\. If a
. the actual response of
where A, b, c, and e are,respectively, n x n, n x 1, 1 x n, and 1 x 1 real constant
Lis type of control is not
matrices. Note that the subscript 1 in FE¡ stands for single variableness. In
e system. If a control
this section we shall derive various equivalent dynamical equations of FE¡
: is called a feedback or
under the controllability or observability assumption.
.íns all the essential in~
Let the followingdynamical equation FE¡
to be a function of the
'0] can be achieved. In
ir state feedback of the
X=Ax +bu (7-2a)
y=cx+eu (7-2b)
326 STATE FEEDBACK AND STATE ESTlMATORS

be an equivalent dymmical equation of FE l , which is obtained by introducing


x= Px = Q - lX, where P ~ Q - 1 is a nonsingular constant matrix. Then eFE l : O
from (4-33) we have O
O
A=PAP- l h=Pb c=eP-l x=
The controllability matrices of FE l and FE l are, respectively, O
-C/.."
U~[b Ab An-lb] (7-3)
y=[ f3n
Ü~[h Ah An-lh]
=P[b Ab An-lb] =PU =Q-lU (7-4)
where (;( lo (;(2,· .. ,Cl n a
and the f3¡'s are to be
Now if the dynamical equations FE l and, consequently, FE l are controllable,1 said to be in the contr
then the controllability matrices U and Ü are nonsingular. Hence, from (7-4)
we have

P=ÜU- l (7-5a) Proof


or Q=UÜ- l (7-5b)
The dynamical equa1
Similarly, ifV and Vare the observability matrices of FE l and FE l , that is, of n x 1 column vecl
quently, the following

l
qn D. b
V= cA
e J
qn-l ~ A
eA~-l q"-2 D. A

ql ~ A
and if FE l and FE l are observable, then
is linearly independen1
or (7-6) from Figure 2-5 that i
¡th column ofthe new
Thus, if two same-dimensionallinear time-invariant, single-variable dynamical
to the basis {ql' q2' ..
equations are known to be equivalent and if they are either controllable or
observable, then the equivalence transformation P between them can be com­ Aql=(
puted by the use of either (7-5) or (7-6). This is also true for multivariable
controllable or observable time-invariant dynamical equations. For multi­
variable equivalent equations, the relation Ü =}FU still holds. In this case U
is not a square matrix; however, from Theorem 2-8, we have ÜU* =PUU*
and P = ÜU*(UU*)-l.
Let the characteristic polynomial of the matrix A in (7-1a) be

Theorem 7-1
If the n-dimensional linear time-invariant, single-variable dynamical equation
FE l is controllable, then it can be transformed, by an equivalence transforma­
tion, into the form

1 See Theorem 5-15.


CANONICAL-FORM DYNAMICAL EQUATIONS 327

)btained by introducing
;:mstant matrix. Then
eFE¡: O 1 O O O O
O O 1 O O O
O O O O O O
x= x+ u (7-7a)
O O O O O
~tively,
-IX" -IX" _¡ -IX"_ Z -lX z -IX¡ 1
(7-3) y=[ /31l f3n- 1 /3" - Z /3z /3¡]x+eu (7-7b)

where IX¡, IXZ' . •. , ct" are the coefficients of the characteristic polynomial of A,
(7-4 )
and the /3¡'s are to be computed from FE¡. The dynamical equation (7-7) is
, FE¡ are control1able,¡
said to be in the controllable canonical formo The transfer function of FE ¡ is
llar. Hence, from (7-4) /3 s"- ¡ +/3 s" - z + ... +/3

g(s) = ¡ z " +e (7-8)


s" +IX¡S" ¡ + ... +IX"_¡S +IX"
(7-5a) Proof

(7-5b)

The dynamical equation F E¡ is control1able by assumption; hence the set


of n x 1 column vectors b, Ab, ... , A"-¡ b is Iinearly independent. Conse­
quently, the fol1owing set of n x 1 vectors
q" ~ b
=VQ qll-¡ ~ Aq" +1X¡q" =Ab +1X¡b
q"-z ~ Aqll-¡ +lX zqll=A zb +1X¡Ab +lXzb (7-9)

q¡ ~ Aqz +1X"_¡q"=A"-¡b +IX¡A"-zb + ... +1X"_¡b


is linearly independent and qualifies as a basis of the state space of FE ¡. Recall
(7-6) from Figure 2-5 that if the vectors {q¡, qz, ... , qll} are used as a basis, then the
¡th column of the new representation Á. is the representation of Aq¡ with respect
19le-variable dynamical to the basis {q¡, qz, ... , qll}' Observe that
~ either control1able or
ween them can be com­ Aq¡=(A"+ IX ¡A"-¡+ ... +lY. ll _¡A+IX"I)b-lX"b
1 true for multivariable

equations. For multi­


I holds. In this case U

.,1 ~ l
we have ÜU* =PUU*
<llll_01X"J
1 (7-la) be

Aq,~q, -,q.~[q,
- •• q, ... q.] lLJ
ble dynamical equation
.................................................

~quivalence transforma­

1
Aqll = qll-¡ -1X¡q" = [ql qz ... q"] [ : ]
I

I _-IX¡
I

j
328 STATE FEEDBACk AND STATE ESTlMATORS

Hence if we choose {q1' qz, ... ,qn} as a new basis of the state space, then A as one might expect.
and b have new representations of the form

o 1 O O O O
O O 1 O O O
O O O O O O
A= b=
(7-10)
O O O O O This can be directly v
-eJ. n -eJ. n - 1 -eJ.n-z -eJ. z -eJ. 1 1 tions FE 1 and CFE 1
hence Q = UÜ- 1 . I
The matrices A and b can also be obtained by using an equivalence transforma­ columns of Q as new
tion. Let Q ~ [q1 qz ... qnJ ~ p- 1, and let x =Px or x·=Qx, then the puting UD-1, we see
dynamical eq uation FE 1 can be transformed into that is,
Q =[q1 q
x=Q- 1AQx +Q- 1bu
We have the folle
y=cQx+eu
observable dynamical
The reader is advised to verify that Q - 1AQ = A or AQ = QA and Q -1 b = b.
The vector e is to be computed from cQ as Theorem 7-2
If the n-dimensional l
(7-11 )
FE 1 is observable, thel
Hence the controllable dynamical equation F El has the equivalent controllable into the form
canonical-form dynamical eq uation C F E l ' This proves the first part of the
theorem.
The dynamical equations FE 1 and eFE 1 are equivalent; hence they have
the same transfer function. lt has been shown in Sections 4-4 and 6-3 that the
transfer function of CFE 1 is equal to g(s) in Equation (7-8). Q.E.D.

One may wonder how we obtain the set of basis vectors given in (7-9). The dynamical eq uati
This is derived in the following. Let D be the controllability matrix of the moreover, the transfer
controllable canonical-form dynamical equation, then
g(s) =
ü ~ [Ri A1bJ ... An- 1 bl J

O O O
O O O el
This theorem can l
theorem of duality (Th
(7-12)
O O 1 en- 3
alence transformation
O 1 el en-z
Equation (7-6): P = V·
1 el ez en- 1
of the observable cano
matrix in (7-12). Hen
where and (7-15) is given by
k-1
ek = - ¿
i=O
eJ.i+ 1e k-i-1 k = 1, 2, ... , n - 1; ea = 1

The controllability matrix Ü is nonsingular for any eJ. 1 , eJ.z, . .. ,eJ.1I" Therefore
the controllable canonical-form dynamical equation is always controllable,
CANONICAL-FORM DYNAMICAL EQUATlONS 329

he state space, then A as one might expect. The inverse of Ü has the following very simple form:

o
a n_ 1
a,,-z
an-Z
a ll - 3
...
••.
al
1
1]
O
O
Ü- I =:: :: ~ A (7-13 )
[ 1 . "
al O O
O
b= (7-10)
1 O O O
This can be directly verified by showing that DÜ- I = 1. The dynamical equa­
O
tions FE I and eFE I are related by the equivalence transformation x=QX;
1
hence Q = uD - l. In the equivalence transformation x = Qx, we use the
columns of Q as new basis vectors of the state space (see Figure 2-S). By com­
:¡uivalence transforma­
puting UÜ-l, we see that the columns of Q are indeed those given in (7-9);
)x or X·= Qx, then the
that is,
Q=[ql qz qll]=[b Ab
oO. An-Ib]ü-l=UA (7-14 )
oO'

We have the following theorem, which is similar to Theorem 7-1, for an


observable dynamical equation.

Theorem 7-2
lf the n-dimensional linear time-invariant, single-variable dynamical equation
(7-11 ) FE lis observable, then it can be transformed, by an equivalence transformation,
equivalent controllable into the form
es the first part of the
1 O
O O -a ll -
n ] .~n]
f3n-l
l
alent; hence they have
X= ~ 1 O -~n-z X + ~n-z u (7-15a)
Ins 4-4 and 6-3 that the [
'-8). Q.E.D. O O 1 -al I
y=[O O O 1 ]x+eu (7-15b)
vectors given in (7-9). The dynamical equation (7-1S) is said to be in the observable canonical form;
Jllability matrix of the moreover, the transfer function of FE¡ is
f3 ¡ sn - ¡ + f3 ZSil - Z + ... + f3 n-¡ s + f3 n +
gA( s) = . 1 e
, :/' -1-0:. 1sl1 - -1-·' +CX¡¡_lS -~-(j~i1

This theorem can be proved either by a dircct verification or by using the


theorem of duality (Theorem S-lO). Its proof is left as an exercise. The equiv­
(7-12 ) alence transformation x = Px between (7-1) and (7-1S) can be obtained by using
Equation (7-6): P = V-IV. It is easy to verify that the observability matrix V
of the observable canonical-form dynamical equation OF E¡ is thesame as the
matrix in (7-12). Hence the equivalence transformation x =Px between (7-1)
and (7-IS) is given by

[a"~, i]~~t,
a ll - z a¡
1; ea = 1 an - z (Xn-3 1
P= : =AV (7-16 )
, az,··· ,.a no Therefore al 1 O
is always controllable, 1 O O O cAn-¡
330 STATE FEEDBACK AND STATE ESTIMATORS

form dynamical equa


We see from (7-9) that the basis ofthe controllable canonical-form dynamical
equation is obtained from a linearcombination ofthe vectors {b, Ab, ... , An-I b}. e,
One may wonder what form we shall obtain if we choose {b, Ab, ... , An - lb} as
the basis. Define ql ~ b, qz ~ Ab, ... , qn ~ An - lb and let x = Q - IX, where
Q ~ [ql qz .. , qnJ; then we can obtain the following new representation
(see Problem 7-2): 01

i~[! :•• g =~::} {r} The transfer function

y=cQx+eu
This equation has the same A matrix as the one in (7-15) and is easier to obtain
because its equivalence transformation Q is simpler. However, the usefulness The controllable
of an equation of this form is not known at present. also be obtained by
equation. The coeff
Example 1 canonical-form eq uati

Transform the following controllable and observable single-variable dynamical The controllable
equation dynamical equations a
They are also useful iJ

x~ [~
2
computer. For exan
-1 (7-17a)
generally need n Z am]
2
and 2n amplifiers and
y=[O O (7-17b) transformed into the c(
into the controllable and observable canonical-form dynamical equations. of (7-16), then the num
The characteristic polynomial of the matrix A in (7-17) is . from n Z + 2n to 2n.

A-1 -2 -lO]
=..1. -9..1.+2 *Multivariable cas(
!l(A)=det [ -~ A~~
3
multivariable dynamic

Hence CX 3 =2, CX z = -9, and (Xi =0. The controllability matrix. U ls

U=[~1 : 1~]

2 12
where A, B, e, and E .
respectively. Let b¡ bf
Ifthe dynamical eq
From Equations (7-5b) and (7-13), we have U = [b¡ b z ... bp

Q=UA =
2 4 16] [-9 O1]
[11 62 128 O1 O1 OO =
[-2 4 2l
-1 6 1J has rank n. Conseque
3 2 1
U. There are many v
The matrix Q can also be obtained from (7-9). From(7-11), we have [rom toe n x np compos
for choosing n linearly
[/33 /3z /3IJ=cQ=[3 2 lJ state space. These two
Hence the equivalent controllable canonical-form and observable canonical-
CANON1CAL-FORM DYNAMICAL EQUATIONS 331

form dynamical equations are, respectively,


lonical-form dynamical
tors {b,Ab, ... ,A"- l b}.
~ {b, Ab, ... , A"- l b} as j=[ -2
~
.d let X = Q - ¡x, where
ing new representation y= [ 3

X~[!
y= [O
The transfer function of (7-17) is thus equal to
A S2 +2s +3
g(s) = S3 -9s +2
and is easier to obtain
The controllable or observable canonical-form dynamical equation can
:owever, the usefulness
also be obtained by first computing the transfer function of the dynamical
equation. The coefficients of the transfer function give immediately the
canonical-form equations. I

gle-variable dynamical The controllable canonical-form and the observable canonical-form


dynamical equations are useful in the study of state feedback and state estimator.
They are also useful in the simulation of a dynamical equation on an analog
computer. For example, in simulating the dynamical equation FE l , we
(7-17a)
generally need n 2 ·amplifiers and attenuators (potentiometers) to simulate A,
and 2n amplifiers and attenuators to simulate b and c. However, if FE¡ is
(7-17b) transformed into the controllable or observable canonical form or into the form
namical equ~tions.
of (7-16), then the number of components required in the simulation is reduced
7)is from n 2 +2n to 2n.

*Multivariable case. Consider the n-dimensional linear time-invariant,


-9)., +2
multivariable dynamical equation
FE: X=Ax +Bu (1-18a)
natrix U ¡s y=Cx +Eu (7-18b)

where A, B, C, and E are n x n, n x p, q x n, and q x p real constant matrices,


respectively. Let b¡ be the ¡th column ofB; that is, B= [b¡ b 2 ••• b p ].
If the dynamical equation FE is controllable, then the controllability matrix

4 2l
62 lJ1
has rank n. Consequently, there are n linearly independent column vectors in
U. There are many ways to choose n linearly independent column vectors
1 (7-11), we have
¡rom the n x np composite matrix U. In the following, we shall give two schemes
for choosing n linearly indepeildent column vectors to forro new bases for the
state space. These two schemes were first discussed in Section S-3.
observable canonical-
-------_...._._--_._-------_. __.__ ._----..__ .__ .-_.. ~ .. _-~_._ ... _._---~--_._ .. _-_._-_._. __._.._._-_. __.. _-------------_._-_._.... _-_.._------.-._ .. -._._------_ .. _._---_._-,----_ .. _._._----_._--~---_ .. __ .__._._ ..._-._._.- _.. _..__ . -----._.. __ ... ~--

332 STATE FEEDBACK AND STATE ESTIMATORS

Seheme 1 OOO"'Ox
We start with the vector b 1 and then proceed to Ab 1, AZb 1, up to AIi,-l b 1 until 100"'O x
the vector AIi'b 1 can be expressed as a linear combination of {b¡, ... , AIi¡-l bd. 010···0 x '
If {1.1 = n, the equation can be control1ed by the first column of B alone. lf
jil < n, we select b z, Ab z, up to Aliz -1 b z, until the vector Alizb z can be expressed OOO···lx·
---- J
as a linear combination of {b 1, ... , AIi¡ -1 b 1, b z, ... ,Aliz -1 b z} (see Problem 7-5). (PI X {1.1) ,
If {1.1 + {1.z < n, we proceed to b 3, Ab 3, ... ,AIi' -1 b 3 and so forth. Assume that
jil +{1.z + {1.3 = n, and the n vectors
A=
are linearly independent. An important property of this set is that the vector
Aliib i can be expressed as a linear combination of the preceding vectors; for
example, Ailzb z can be expressed as a linear combination of {b 1, Ab 1, ... ,
Ail¡-l b 1, b z, Ab z, ... , Ailz -1 bz}.

Seheme 2

The linearly independent vectors are selected in the order of (7-19); that is, we where the X's denote'
start from b 1, b z, ... , b p and then Ab 1, Ah z, ... , Ab p, and then AZb 1, AZb z, zeros. This can be ea~
and so forth, until we obtain n linearly independent vectors. Note that if a representation of Aqi ,
vector, say Ab z , is skipped because of linear dependence on the vectors If the set ofvectors
{b 1, b z, ... , b p, Ab 1}, then aH vectors of the form Akb z, for k ~ 1, can also be will be of the form 3
skipped becatise they must also be dependent on the previous columns. After ,
choosing n linearly independent vectors in this order, we rearrange them as O O x;
1 O x:,
{b 1"·" AJl,-lb'b
1,2,'··'
AJlz-1b"b
2,"" p, ... ,
AJlp-1b}
P (7-21 ) O 1 x,
,
,
where /11 + /1z + ... + /1p = n. Note that the main difference between this O O x :
scheme and Scheme 1 is that in Equation (7-20) Ail'b 1 can be expressed as a
linear combination of {b 1, Ab 1, ... , Ail¡ -1 b 1}, whereas the vector AJl¡ b 1 in
-(~~ ""X ~:f ~- O' - :
(7-21) cannot be expressed as a linear combination of {b 1, ... , AJl¡-l bd; AJl¡ b 1 x :1
x :O
is general1y linearly dependent on al1 vectors in (7-21).z Similar remarks
apply to AJlib¡, for i = 1, 2, ... , p.
Now if the set of vectors in (7-20) is chosen as a basis of the state space of Á= :'(_ _0_ i
FE or, equivalently, let x = Q-1x where (1

Ailz- 1b z
b 3 ... Ail, -1 b 3J
--------------
x,
,
.
then the matrices A and B will be of the forms
x,
x'
.,
,
,
x:
2 More can besaid by using the relative sizes of l1i. For example, if 1115,1125,'" 5,l1p, ihen A!"b 1
depends on{Akb¡, i= 1,.2, ,p; k =0,1, ... ,111 -l}; A"2b z depends on {Akb¡,k=O, 1, ... ;0 111 - 1;
Akb¡,i = 2, 3, ... , p; k =0, 1,0 ,112 -l} and so forth.
3 Using footnote 2, sorne elen
----- --- - ----- - - -

CANONICAL-fORM DYNAMICAL EQUATIONS 333

OOO"'Ox x: x 1 O O

b¡, up to Aiit -¡ b¡ until 100···Ox' x; x O O O

nof{b¡, ... ,Aii'-¡b¡}. 010···0x x:


. , x O O O

:olumn of B alone. lf I : 1

Aii2 b z can be expressed OOO···l.x:


------------J-
x:J x
_ O O O

¡b z} (seeProblem 7-5). (j].¡ x j].¡) : O O O ... O x : x O 1 O

so forth. Assume that :100"'Ox x O O O

:010"'Ox, x O O O

A= ,
l' ••
B= b4 "'bp
:OOO"'lx x OOO
is set is that the vector - - -(~~ -x-fi;)- -- ~ 0- 0- Ó-.-. ~ -Ó -.; OO 1
preceding vectors; for :100"'Ox OOO
.ation of {bl> Ab¡, ... , :010"'Ox OOO
1: : :
I •••

:OOO"'lx OOO
(j].3 x j].3)
(7-22)
ler of (7-19); that is, we where the X's denote possible nonzero elements and unfilled positions are all
" and then AZb¡, AZb z, zeros. This can be easily verified by observing that the ith column of A is the
ectors. Note that if a representation of Aq¡ with respect to the basis vectors [q¡ qz '" qn].
¡dence on the vectors lfthe set of vectors in (7-21) is used as a basis, then the new matrices Á and B
z, for k;:::l, can also be will be of the form 3
'evious columns. After
ve rearrange them as O O x ,, .x', x 1 O O
x ,,
(7-21 )
1 O
O 1 x ,
,
.
x;
x,
,
. x
x
O
O
O
O
O
O
, ,
, ,
lifference between this O O x'! x'J , x
-------­ O O O
can be expressed as a -----------
x : ,,
1 - __ - - - - ­

(111 X )J.l) x: O O x O 1 O
.s the vector AJL, b¡ in
I1¡-lb}' x :1 O X: ,, x O O O
'b ... , A l ' Al1'b 1 ,
x : O 1 x: x O O O
~ 1). z Similar remarks . ,, . ,
, ,

;¡is of the state space of _________ !_l ?__ ? -: ~ ; :: B=


O O
-----------
O
()J.z x )J.z)

Aii2- 1 b z ()J.p x )J.p)


b 3 • •• Aii3 - 1 b 3 ]
-----------"'j-----------¡
X, x, :0--0------; -----------
O O 1
, ,
x, x', :1 O x O O O
x ,, x , :0 1 x O O O
, , ,.
, , ,
.

x , X , :O O
I
x O O O
5,/1" then A"'b!
.,5,/125,'" (7-23)
on{Akb"k=O, 1, ... ,/1,-1;
3 Using footnote 2, sorne elements denoted bY X can be replaced by' zeros.

I
J
334 STATE FEEDBACK AND STATE ESTIMATORS

where the X'S again denote possible nonzero elements and unfilled positions are
all zeros. The matrices Aand B can be verified by inspection. The matrix C
in both cases are to be computed from CQ.
By comparing (7-22) with (7-23), we see immediately the differences in the
matrices A and B due to the different choices of basic vectors according to
Schemes l and 2. The matrix Á has three blocks in the diagonal, whereas the
matrix Á has p blocks. The first three columns of B are very simple, whereas
every column of Bconsists of only one nonzero element.
The usefulness of the forms in (7-22) and (7-23) is not known at present.
The purpose of introducing these two forms is to show that there is no additional
conceptual difficulty in developing canonical forms for multivariable dynamical
equations. By rearranging the vectors in (7-20) or (7-21), different dynamical
equations can be obtained. One of them will be discussed in Section 7-3. For
a survey of various canonical forms, see References S154. Figure 7-1 A state f

7-3 State Feedback Proof

First we show that t


In this section we study the effect of the linear state feedback u = r +Kx on a
FE{. Let X o and Xl 1
linear time-invariant dynamical equation. The variable r denotes the reference
trollabili ty assumptio
input and the K is the feedback gain matrix and is required to be real. In this
in a finite time. No",
study, we implicitly assume that all state variables are available for the feedback.
r(t) = u(t) - kx(t), then
that FE{ is controllac
Single-variable case. Consider the single-variable, linear time-invariant
We see from Figur
dynamical equation
it generates u to contr
FE!: x=Ax +bu (7-24a) other words, if FE I is
y=cx +eu (7-24b)

where X is the n x l state vector, u is the scalar input, y is the scalar output, We see that in the
A is an n x n real constant matrix, b is an n x 1 real constant column vector, invariance are not use
and c is a 1 x n real constant row vector. In state feedback, every state variable
is multiplied by a gain and fed back into the input terminal. Let the gain
between the ¡th state variable and the input be k¡. Define k~ [k l k 2 .•. knJ. Corollary 7-3
Then the dynamical equation of the state-feedback system shown in Figure 7-1 is
The con troI1ability af
FE{: x= (A + bk)x + br (7-25 a ) is invariant under any
y=(c+ek)x+er (7-25b)

which is obtained by replacing u in (7-24) by r +kx, where r is the reference Note that Theoren
input. Note that the dynamical equations (7-24) and (7-25) have the same
dimension and the same state space. Now we shall show that the controllability p[b Ab
of a linear time-invariant dynaniical equation is invariant under any linear
state feedback. for any 1 x n real cons
A1though state fe
Theorem 7-3 equation, it is alway:
dynamical equation b
The state feedback dynamical equation FE{. in (7-25) is controllabie .ror any ( -1/e)c, then the stat(
1 x n real vector k if and only ifthe dynamical equation FE I in (7-24) is con­ is observable. If e =0
trollable. feedback dynamical ec
Problem 7-14).
STATE FEEDBACK 335
unfilled positions are
:ction. The rnatrix e

the differences in the


vectors according to
iiagonal, whereas the
very simple, whereas

ot known at present.
.there is no additional
lltivariable dynarnical
e
), different dynarnical
d in Section 7-3. For
Figure 7-1 A state feedback system.

Proof
First we show that the controllability of FE¡ implies the controllability of
,back u = f +Kx on a FE{. Let Xo and X¡ be two arbitrary states in the state space L. By the con­
denotes the reference trollability assurnption of FE ¡, there exists an input u that will transfer X o to Xl
ed to be real. In this in a finite time. Now for the state-feedback dynamical equation, if we choose
lable for the feedback. r(t) = u(t) - kx(t), then the input r will transfer X o to Xl. Therefore we conclude
that F E{ is controllable.
linear time-invariant We see from Figure 7-1 that the input r does not control the statex directly,
·it generates u to control x. Therefore, ir u cannot control x, neither can r. In
(7-24a) other words, if FE¡ is not controllable, neither is FE{. Q.E.D.
(7-24b)

I is the scalar output, We see that in the proof, the assumptions of single-variableness and time­
lstant column vector, invariance are not used. Therefore we have the following corollary.
k, every state variable
~minal. Let the gain
k~ [k¡ kz ... knJ. Corollary 7 - 3
shown in Figure 7-1 is The controllability oí a multivariable linear Unle-,¡arying dynamical cCjuation.
(7-25a) is invariant under any state feedback of the form u(t) = r(t) +K(t)x(t). •
(7-25b)

lere r is the reference Note that Theorem 7-3 can also be proved by showing that
(7-25) have the sarne
p[b Ab ... An-lb] =p[b (A+ bk)b ... (A+ bk)n-¡b] (7-26)
.hat the controllability
ant under any linear
for any 1 x n real constant vector k (see Problern 7-10)..
Although state feedback preserves the controllability of a dynamical
eq uation, it is always possible to destroy the observabiiity property of a
dynarnical equation by sorne choice of k. For example, if e =/=-0 and if k =
; controllable for any ( - l/e)c, then the state-feedback equation (7-25) is notobservable even if FE ¡
FE ¡ in (7-24) is con- is observable. If e = 0, it is still possible to choose sorne k such that the state­
feedback dynarnical equation will not preserve the observability property (see
Problern 7-14).

1
336 STATE FEEDBACK ANO STATE ESTIMATORS

Example 1 nomial of the matrix (


eigenvalues be
Consider the controllable and observable dynamical equation

x=G
y=[1
n +[~Ju
x

2Jx
If k is chosen as

If we introduce the state feedback


then the sta te-feedback '
u=r +[ -3 -IJx
then the state-feedback equation is
CFE{: o
O

x=[~ ~Jx +[~}-


FE{: O
x=
O
y=[1 2Jx -e
which is controllable but not observable. I

An important property of state feedback is that it can be used to control the


eigenvalues of a dynamical equation.
Since the characteristic
Theorem 7-4
... +am we conclude th
If the single-variable dynamical equation FE 1 given in (7-24) is controllable, values.
then by the state feedback u =:= r +kx, where k is a 1 x n real vector, the eigen­
values of (A + bk) can be arbitrarily assigned, provided that complex conjugate
The gain vector k ir
eigenvalues appear in pair.
u = r +kx. Therefore,'
k = kP, where the colurr
Proof
The matrices A, b, and k
Ifthe dynamical equation FE 1 is controllable, by an equivalence transformation is assigned to the matrix
x = Px, FE I can be transformed into the following controllable canonical The procedure of choosi
form (Theorem 7-1): therefore it can be appl
linear time-invariant dy
I 1

l!
CFE!: O O O

choosing Rr ¡re the fo!1ovvi

i~
O 1 O

O O O

-Cl.fT
O
-a"_1
O
-a,,-z
O
-az
r]i+[!]U
-al 1
Algorithm

Given a controllable {A
1 x n real vector k such
(7-27 a)
its eigenvalues.
y= [fi" fi,,- 1 fi,,- 2 ... f3z filJx +eu (7-27b)
1. Find the characteristic
Let A and b denote the matrices in (7-na), then A = PAP- 1, b = Pb. Because 2. Compute (s -I¡)(s -~ .
. of the equivalence transformation, the state feedback becomes 3. Compute k = [a" ~. a"
.. u = r +kx = r.+kp-Ix ~ r +kx (7-28) 4. Compute q"-i =Aq,,_
5. FormQ=[ql qz '
where k ~kP-I. It is easy to see that the ser Of the eigenvalues of (A + bk) is 6. Find P ~ Q - l.
equal to the set of the eigenvalues of (A + bk). Let the characteristic poly­ 7. k =kP.
---------------_ ..-. .. _. --_._---------_ .. _._.-------
--==c::===_-
-~~~,,===~===- __ ~-_-_-~~.:__-=_-=-=---=-~----------------------------~--------
------.--_.- -- ----,_.~--------_._

-----------------.---­

STATE FEEDBACK 337

nomial of the matrix (A + bk) or, correspondingly, of (A + bk) with desired


eigenvalues be
ation

If k is chosen as

(7-29)

then the state-feedback dynamical equation becomes

CFE{: O 1 O O O O
O O 1 O O O
O O O O O O
x= x+ l'

O O O O 1 O
-Cl. n - ii"_ ¡ -IX" - 2 -ii 2 -IX¡ 1
I (7-30a)
y=[I1" +e(cx"-éX") 11"-¡ +e(cx"_¡-éX"_¡) ... 11¡ +e(cx¡-iX¡)]x +er
be used to control the (7-30b)

Since the characteristic polynomial of the A matrix in (7-30) is s" +iX¡s"-¡ +


... +ii.., we conclude that the state-feedback equation has the desired eigen­
(7-24) is controllable, values. Q.EoD.
real vector, the eigen­
mt complex conjugate The gain vector k in (7-29) is chosen with respect to the state x; that is,
u =1' +kx. Therefore, with respect to the original state x, we have to use
k=kP, where the columns ofP-¡ =Q are those vectors introduced in (7-9).
The matrices A, b, and k are assumed to be real; hence, if a complex eigenvalue
alence transformation is assigned to the matrix (A + bk), its complex conjugate must also be assignedo
ontrollable canonical The procedure of choosing k has nothing to do with the number of outputs;
therefore it can be applied to any controllable single-input, multiple-output,
linear time-invariant dynamical equationo We summarize the procedure of
o rO choosing k in the followingo
o

'+l~
o
u Algorithm
Given a controllable {A, b} and a set of eigenvalues Xl> X2 , o.. , X... Find the
-IX¡ 1 x n real vector k such that the matrix (A +bk) has the set {Xl' X2 , o.. , X"} as
(7-27 a) its eigenvalues.
eu (7-27b)
1. Find the characteristic polynomial of A: det (sl- A) = Sil +IX ¡s" - ¡ + ... +IX...
p-¡, b = Pbo Because 2. Compute (s - X¡)(s -' X2 )· •• (s - Xn ) = s" +iX ¡ s" -¡ -+: ., +ii...
;omes 3. Computek=[cx"-iX" IXn-1-iX n -¡ o.. IXI-~il
(7-28)
4. Compute q.. _ i = Aqn-i+ ¡ + (X¡q", for i = 1,2, . o. , (n-l), with q" = bo
5. Form Q = [q¡ Q2--' o. qnl ­
envalues of (A + bk) is 6. Find P ~ Q - l.
le characteristic poly­ 7. k=kP. I
._._----~-----------_._-----

338 STATE FEEDBACK AND STATE ESTIMATORS

This algorithm can also be developed directly without transforming the case, a differen t rr
equation explicitly into the controllable canonical formo Let ~(s) be the (Appendix F) will
characteristic polynomial of A + bk. Then we have single-variable case
~(s) = det (sI - A - bk) = det [(sI - A)(I - (sI - A)-l bk)J

Stabi Iization
= det (sI - A) det [1 -(sI - A)-lbkJ

Ir a dynamical equa
This can be written as, by using (3-65) and defining ~(s) = det (sI - A),
assigned by the intr
~(s) = ~(s)(1 - k(sl - A)-l b) controllable, one m
or ~(s) - ~(s) = ~(s)k(sl - A)-l b (7-31 ) shown in Theorem :
controllable, by a
From this equation and using Problem 2-39, Equations (7-13) and (7-14), we
transformed into
can readily establish k =kUA =kQ or k =kP.

Example 2
where
Consider the inverted pendulum problem studied in Example 4 on page 185.
Its dynamical equation is, as derived in (5-25),
and the reduced ec

X~l~
1 O form of A, the set 01
O -1 of All and of A22 .
O O A22 is not affected
O 5
r +kx. Therefore;
y = [1 O O
other hand, all the
It is controllable; hence its eigenvalues can be arbitrarily assigned. Let the
{A lb lid is control
desired eigenvalues be - 1, - 2, - 1 ±j. Theri we have ­ can be arbitrarily l­
required to be a re:
~(s) = (s + 1)(s + 2)(s + 1 + j)(s + 1 - j) = S4 + 5s 3 + lOs 2 + lOs +4 in pairs.
The characteristic polynomial of In the design of
eigenvalues (the eig(

A+bkJ ~
1 (the eigenvalues wit
k2 the aboye discussior
O the equation canno
l-2k l -2k 2 transformed into th,
eigenvalues of Azz h
where k = [k l k2 k3 k 4 J can be computed as a dynamical equatio
det (sI - A - bk) = S4 +(2k 4 - k 2 )S3 +(2k 3 - k1 - 5)S2 + 3k 2 s + 3k l form dynamical eqt
eigenvalues are cont
The comparison of the coefficients of det (sI - A - bk) and those of ~(s) yields of Statement 4 of T
k1=t k 2 =lf k 3 =s,¡ k 4 =li if and only ifthe ma­
that Al l and An in
Hence the introduction of the state feedback
u(t)=r(t)+[i lf ~ liJx Effect on the nume
will place the eigenvalues of the resulting equation at - 1, - 2, -1 ±j. I Consider the dyna
c(sl - A) - 1 b +e. Si
Even for this simple {A, b}, direct computation of det (sI - A - bk) is very 7-4 we conclude tha
complicated and is not suitable foi computer computation. Hence, for large duction ofstate feec
n, the use of the algorithm is much more convenient. In the multivariable g(s) are shifted by st:
-_._._"~-~-----~,~._---.=.",.-===-====--=-::==-=----~'-':'~=-:::"-:::=:':'''....--::::'''~==-.==::::--==-----~ . _---_.­
-':::'--:::.='::::-- -==--=--=-.-.:.......===...-=--===-:::..:=....::.::=.::~=='::::.====--------_

STATE FEEDBACK 339

out transforming the case, a different method of computing k by solving a Lyapunov equation
m. Let li(s) be the (Appendix F) wilI be introduced. The method is directly applicable to the
single-variable case. To avoid repetition, it will not be discussed here.

Stabilization
Ir a dynamical equation is controlIable, then alI the eigenvalues can be arbitrarily
1= det (sI - A),
assigned by the introduction of state feedback. If a dynamical equation is not
controlIable, one may wonder how many eigenvalues can be controlIed. It is
(7-31 ) shown in Theorem 5-16 that if a linear time-invariant dynamical equation is not
controllable, by a proper choice of basis vectors, the state equation can be
(7-13) and (7-14), we
transformed into
x=Ax +bu (7-32)

where -_[AO A
A- ­
A
11 12 ]
(7-33)
lmple 4 on page 185. 22

and the reduced equation Xl = A11 Xl + b1 u is controllable. Because of the


form of A, the set of eigenvalues of A is the union of the sets of the eigenvalues
of A11 and of A22 . In view of the form of b, it is easy to see that the matrix
A22 is not affected by the introduction of any state feedback of the form u =
r +kx. Therefore alI the eigenvalues of A22 cannot be controlIed. On the
other hand, aH the eigenvalues of A11 can be arbitrarily assigned because
{A 11 , bd is controllable. Hence we conclude that the eigenvalues of(A +bk)
ily assigned. Let the can be arbitrarily assigned if and only if {A, b} is controlIable. Since k is
required to be a real vector, complex conjugate eigenvalues must be assigned
10s 2 +10s +4 in pairs.
In the design of a system, sometimes it is required only to change unstable
eigenvalues (the eigenvalues with nonnegative real parts) to stable eigenvalues
(the eigenvalues with negative real parts). This is calIed stabilizatíon. From
the aboye discussion we see that ifthe matrix A 22 has unstable eigenvalues, then
the equation cannot be stabilized. Hence we may conclude that if {A, b} is
transformed into the form in (7-33), and if {Al 1> bd is controlIable and alI the
eigenvalues of .4. 22 have negative real parts, then {A, b} is stabilizable. Whether
a dynamical equation is stabilizable can also be seen from its equivalent Jordan­
form dynamical equation. If alI the Jordan blocks associated with unstable
'S2 + 3k 2 s + 3k 1
eigenvalues are controllable, then the equation is stabilizable. From the proof
id those of li(s) yields of Statement 4 of Theorem 5-7 on page 206, we can conclude that }, is in A22
:z¡ if and only if the matrix [..1.1 - A B] does not have a full row rank. We note
that Al! and A22 in (7-33) may have same eigenvalues.

Effec(on thé numerator of g(s)


1, -2, -1 ±j. • Considerthe dynamical equation (7-24). Its transfer function is g(s)=
c(sl - Af 1 b +e. Since every pole. of g(s) is an eigenvalue of A, from Theorem
:t (si - A - bk) is very 7-4 we conclude that the poles of g(s) can be arbitrarily assigned by the intro­
.on. Hence, for large duction of state feedback. It is of interest to note that although the poles of
In the multivariable g(s) are shifted by state feedback, the zeros of g(s) are not affected. That is, the
340 STATE FEEDBACK AND STATE ESTIMATORS

zeros of g(S) remain unchanged after the introduction of state feedback. We which becomes, in t
prove this by showing that the numerator of the transfer function of FE 1 in
(7-24) is equal to the numerator of the transfer function of eFE 1 in (7-30). y(t) = (e
The transfer function of FE 1 is for t ~ O. If all eige
A() 131S,,-1 +132S,,-2 + ... +f3" + approaches zero as
gs = e Ir {A, b} is controlla
s" +O:I S" 1 + ... +0:"
by a proper choice
es" +(131 +eO:l)s,,-1 + +(13" +eo:,,) deeper inside the le
(7-34 )
s" +O:IS,,-1 + +0:" However, in this cas
The transfer function of (7-30) is saturation. Further
the choice of proper
[131 +e(O:I -&I)]S,,-1 + ... +[13" +e(o:" - IX,)] by a compromise be
g¡(s) +e
s" +O:I S" -1 + ... +0:" As discussed in .,
es" +(131 +e0: 1)s,,-1 + +(13" +eo:,,) by state feedback; h
(7-35 ) we may choose p = l
s" +O:I S" 1 + +á"
which has the same numerator as g(s) in (7-34). This proves that the zeros of
g(s) remain unchanged after the introduction of state feedback. and the design is COI
This property can be used to explain why a state feedback may alter the {A, b} is not con trol
observability property of an equation. Ir sorne of the poles are shifted to have complete contr,
coincide with zeros of g(s), then the degree of g¡(s) in (7-35) is less than n, and Much more can b
the dynamical equation in (7-25) is not irreducible. In this case, (7-25) must It appears that the d
be unobservable beca use it is controllable (Theorem 7-3). function and will be
discussed further in t
Asymptotic tracking problem-nonzero set point
In this subsection we use state feedback to design a system so that its output *Multivariable ca:
will track the step reference input r(t) =rd, for t ~O. Ir rd =0, the problem is multivariable dynam
called a regulator problem; if rd =1= O, it is a special case of the asymptotic tracking
problem. A step reference input can be set by the position of a potentiometer
and is often referred to as set point. In this problem, the input u is chosen as
where A, B, e, and l
u(t) = pr(t) +kx(t) (7-36) real matrices. In sta
where p is a constant gain and k is the feedback gain vector. The substitution
of (7-36) into (7-24) and then the application of the Laplace transform to the
resulting eq uation yields where r stands for a r
called the feedback g
;(s) = (e + ek)[(sl - A - bkt l X (O) + (si - A - bk)-1 bpr(s)] + epr(s) (7-37)

Ir r(t) = rd, for t ~ O, then res) = rJs. The substitution of the identity
(si - A - bk)-I S-1 =( -A - bk)-lS-1 +(sl -A - bk)-I(A + bk)-1 In the following,
trollable, then the ei
[see also Equation (H-I)] and
proper choice of K.
g¡(s)=(e +ek)(sl-A-bk):-lb +e (7-38)
Method 1
into (7-37) yields
In this method we e
;(s) = (e +ek)(sl - A - bk)-1 [x(O)1 +(A + bk)-1 bprd] +g¡(Ü)prds-1 problem and then ap!
STATE FEEDBACK 341

state feedback. We which becomes, in the time domain,


r function of FE! in
y(t)=(c +ek)eIA+bk)1 [x(O)I +(A +bk)-lbprd] +g¡(O)prd (7-39)
1 of eFE! in (7-30).
for t 2: O. If aH eigenvalues of (A + bk) have negative real parts, then e lA +1>1<),
approaches zero as t->oo and consequently y(t) approaches g¡(O)prd as t->Cú.
If {A, b} is controHable, aH eigenvalues of (A +bk) can be arbitrarily assigned
by a proper choice of k. Clearly, if the eigenvalues of (A + bk) are shifted
(7-34 )
deeper inside the left-half s plane, the faster the output approaches g¡(O)prd'
However, in this case the required feedback gain wil1 be larger and may cause
saturation. Furthermore, the system may be more susceptible to noises. Thus
the choice of proper eigenvalues is not a simple task and is generally reached
--ex,,)]
- e
+ by a compromise between conflicting criteria.
As discussed in the previous subsection, the zeros of g(s) are not affected
by state feedback; hence we have g¡(O) f.O if and only if g(O) f.O. If g(O) f. O,
(7-35) we may choose p = l/gJ(O). In this case, we have

¡ves that the zeros of as t-> 00

Jack. and the design is completed. We note that the design can also be achieved if
dback may alter the {A, b} is not control1able but is stabilizable. In this case, however, we do not
poles are shifted to have complete control over the rate of convergence of y(t) to rd'
5) is less than n, and Much more can be said regarding the design ofasymptotic tracking problem.
his case, (7-25) must It appears that the design can be carried out more easily by using the transfer
function and will be discussed in Section 9-6. Hence this problem will not be
discussed further in this chapter.

:m so that its output *Multivariable case. Consider the n-dimensional linear time-invariant,
multivariable dynamical equation
d = 0, the problem is

e asymptotic tracking FE: x=Ax+Bu (7-40a)


n of a potentiometer y=Cx+Eu (7-40b)
n.put u is chosen as
where A, B, C, and E are, respectively, n x n, n x p, q x n, and q x p constant
(7-36) real matrices. In state feedback, the input u in FE is replaced by
)1". The substitution
ace transform to the
where r stands for a reference input vector and K is a p x n real constant matrix,
caBed the feedback gain matrix; and Equation (7-40) becomes
ores)] + epr(s) (7-37)
FEJ: x =(A +BK)x +Br (7-41 a )
le identity y=(C+EK)x+ Er (7-41b)

In the following, we shall show that if the dynamical equation FE is con­


trollable, then the eigenvalues of (A + BK) can be arbitrarily assigned by a
proper choice of K. This wil1 be established by using three different methods.
(7-38) .
Method 1
In this method we change the multivariable problem into a single-variable
problem and then apply the result in the previous subsection.
---_ .._----_.. _--_._-----_._-----------.-----_._-_._-----------_._._---_._-------_._---_._-_._ .. _._-------_.-,_._----------~------------_._----_ .. _---~--_._-_. __.--_._---_. ---~-_._-----_ .. _------_. __..•--_._ .._--_._._._._-_._. __ _-_._._-_.
..

342 STATE FEEDBACK AND STATE ESTlMATORS

A matrix A is called cyclic if its characteristic polynomial is equal to its Theorem 7-6 4
minimal polynomial. From Theorem 2-12 or Example 1 on page 48, we If {A, B} is controlla
can conclude immediately that A is cyclic if and only if the Jordan canonical the eigenvalues of A
form of A has one and only one Jordan block associated with each distinct
eigenvalue (Problem 2-45). The term "cyclicity" arises from the property that if Proof
A is cyclic, then there exists a vector b such that the vectors b, Ab, ... ,An -1 b
span the n-dimensional real space or, equivalently, {A, b} is controllable. This Let the characteristic
property can be easily deduced from Corollary 5-21.

Theorem 7-5 where a¡, i= 1, 2, ... ,


tion of X(s) with respt
lf {A, B} is controllable and if A is cyclic, then for almost any p x 1 real vector v,
the single-input pair {A, Bv} is controllable. L1'1
If L1(s) has repeated re
Proof sufficient condition fe
Controllability is invariant under any equivalence transformation, hence we
may assume A to be in the Jordan canonical formo To see the basic idea, we an an-
use the following example: O °n

O O
2 21 O: O]
O O
det --------.
O 1:0 an - I 2a n_
A= O O 2: O O
O a'l-
[
~--~--~-r~--~ O O

(See Appendix G). V


There is only one Jordan block associated with each distinct eigenvalue; hence
nomial of kij. There i
A is cyclic. The condition for {A, B} to be controllable is that the two rows in
B encircled by dotted lines in (7-42) are nonzeros (Theorem 5-21). of {ku} as a vector in 1
The necessary and sufficient condition for {A, Bv} to be controllable is that clear that the solutio
almost all K, we hay.
ex 1=0 and f3 1=0 in (7-42). ex = Oand f3 =0 if and only if VI =0 and vtfvz = - 2/1.
establishes the theoreI
Hence any v other than VI =0 and vI/vz = - 2 will make {A, Bv} controllable.
This establishes the theorem. Q.E.D.
V,;/e are D.OV/ ready

The cyclicity assumption in this theorem is essential. Without this assump­ Theorem 7-7
tion, the theorem does not hold. For example, the {A, B} in lfthe dynamical equat
of the form u = r + Kx.
of (A + BK) can be arbi
appear in pairs.

IS controllable. However, there is no v such that {A, Bv} is controllable. 4 This theorem can be extel
lf allthe eigenvalues of A are distinct, then there is onlyone Jordan block real constant matrix H, th
associated with each eigenvalue. Hence a sufficient condition for A to be cyclic cyclic. The matrix A + BI
is that all the eigenvalues of A are distinct Theorem 7-6 is equally apl
--_ .. _--_._--_._._.-----------­ ------------- --_._----
-------------- --.-_.--­
-=.~~::=.:----=-~~--
--_._--~-----
---~--~--------_._-------_._~
---~-- -~-- - -­ ~- --~-----

STATE FEEDBACK 343

Theorem 7 _6 4
lmial is equal to its
: 1 on page 48, we If {A, B} is controllable, then for almost any p x n real constant matrix K, all
le Jordan canonical the eigenvalues of A + BK are distinct and consequently (A +BK) is cyclic.
i with each distinct
1 the property that if Proof

)rs b, Ab, ... ,An - 1 b


s controllable. This
Let the characteristic polynomial of A + BK be

where a¡, i = 1, 2, ... , n, are functions of aH elements, kij, of K. The differentia­


tion of Li(s) with respect to s yields
IY P x 1 real vector v,
Li'(s)=nsn-l +(n- 1)ad'-z + ... +an-l

If Li(s) has repeated roots, then Li(s) and Li'(s) are not coprime. A necessary and
sufficient condition for Li(s) and Li'(s) to be not coprime is
'ormation, henee we
O O
ee the basic idea, we
1 O

o O a an-I 1
det ------------------ n------------------ ~ y(kij) =0
a,,-l 2a n-2 n O O O
O an-I (n - l)al n O O

O O O an-l n

(See Appendix G). We note that y(kij) is generally a nonhomogeneous poly­


lct eigenvalue; hence nomial of kij. There is a total of p x n number of kij in K. Hence we may think
that the two rows in of {kij} as a vector in the (p x n)-dimensional real vector space (IR P )(", IR). It is
n 5-21). clear that the solution, kij, of y(kij) = O is a subset of (lIF )(n, IR). Hence for
e controHable is that almost aH K, we have y(k¡) +0 and all the roots of Li(s) are distinct. This
=0 and Vl/VZ = - 2/1. establishes the theorem. See Problem 7-8. Q.E. D.
{A, Bv} controllable.
Q.E.D. We are now ready to establish the following main theorem.

Theorem 7-7
Vithout this assump­
in Ir the dynamical equation FE in (7-40) is controllable, by a linear state feedback
of the form u = r + Kx, where K is a p x n real constant matrix, the eigenvalues
of (A + BK) can be arbitrarily assigned provided complex conjugate eigenvalues
appear in pairs.

is controllable. 4This theorem can be extended as follows: If {A, B,C} is irreducible, then for almost any p x q
lly one Jordan block real constant matrix H, the eigenvalues of A + BHC are distinct and consequently (A+ BHC) is
lion for A to be cyclic cyclic. The matrix A + BHC r~sults from the constant output feedback u = r + Hy. The proofor
Theorem 7-6 is equally applicable here.
-._._-_.~ .. _,_.. ~-~~_ __ _----_._-~ .. __.. ~------_ ..-----_.. __.__.._._--_._--_._------_._-_.. ~--"_._ _---~ _.~- .. _.. _ .. "-~ .. _~ _~---_._.,._----_.- __ _.__._-_._._._.,_ _---_._._.__
. .. ._--_._-----~-_ .. _.. _-_._-_._.- .. - ..'.
, __ .',_.'---_ __ .-

344 STATE FEEDBACK ANO STATE ESTIMATORS

Proof If {A, B} is not cont


If A is not cyclic, we introduce u = w + K¡x such that Á ~ A + BK! in

i =(A +BK¡)x +Bw (7-43)


and the eigenvalues of ~
is cyclic. Since {A, B} is controllable, so is {A, B} (Corollary 7-3). Hence conclude that the eiger
there exists a p x 1 real vector v such that {A, Bv} is controllable. 5 Now we only if {A, B} is control:
introduce another state feedback w = r +Kzx, as shown in Figure 7-2, with K z
chosen of the form Method 11
In this method, we shall
in (5-93) and (6-112) al
where k is a 1 x n real vector, then (7-43) becomes similar to the single-inpl
we assume n =9 and p =
i=(A +BKz)x +Br =(A +Bvk)x +Br are Jl¡ = 3, Jlz = 2, and Jl
Since {A, Bv} is controllable, the eigenvalues of (A + Bvk) can be arbitrarily M =[b¡ Ah!
assigned by a proper choice of k (Theorem 7-4). By combining the state feed­
(see Section 5-3) is nonsi
back u = w + K¡x and the state feedback w = r + Kzx as

u =r +(K¡ +Kz)x ~ r +Kx

the theorem is proved. Q.E.D.

5 The choices of K , and vare not unique. They can be choserr arbitrarily, and the probability is
almost I that they will meet the requirements. In Theorem 7-5 of Reference S38, a procedure is
given to choose K , and v without any uncertainty. The computation required however is very
complicated.

From M-IM= 1, w
k = O, 1; e 13 Az h 1 = 1; el.
we do not have e 13 A2 bz =
1,2,3 and A 2b¡} and e l3
matrix 6

+ y

6Tills mat~ix can be obtained e


Figure 7-2 State feedback of a multivariable dynamical equation. computing explicitly M - l. ~
----.:.-==--=-:::.::----===-====- -=-=-=---====-.-====-=..=.===-..:==:=::--==------- --------- -- - _._--------_.__.__. -._-_.
- - - _ ••- - - - _ ••• _ _. - - - _ •• _ . _ - - - - _ . _ - - - - -
-~
o
---_. __
.
----_._----_._-----_. .. .-----_ ... -._--
_~----._----------_
_... ­ _-----_._----­
.
_---._~--~ ..
_~

STATE FEEDBACK 345

Ir {A, B} is not controllable, they can be transformed into

(7-43)
[A~l i::J [~lJ

and the eigenvalues of A22 are not affected by any state feedback. Hence we
Corollary 7-3). Hence conclude that the eigenvalues of (A + BK) can be arbitrarily assigned if and
:ontrollable. 5 Now we only if {A, B} is controllable.
i1 in Figure 7-2, with K z
Method 11
In this method, we shall transform {A, B} into the controllable form discussed
in (5-93) and (6-112) and then compute the required K. The procedure is
similar to the single-input case. In order not to be overwhelmed by notations,
we assume n =9 and p = 3. lt is also assumed that the controllability indices
+-Br are MI = 3, Mz = 2, and M3 = 4. Then the matrix
.Bvk) can be arbitrarily (7-44)
Jmbining the state feed- l
(see Section 5-3) is nonsingular. We compute M- and name its rows as
ell
e lZ
Q.E.D. e l3 ~ell'l

eZ l
e zz ~eZ1'2

it;arily, and the probability is e 3l


Reference S38, a procedure is e 3Z
tion required however is very e3 3
e34 ~e31'J

From M-lM=I, we can obtain sorne relationships such as e 13 Ak b l =O,


k = 0,1; e 13A z b l = 1; e 13 A k b Z= O, k= O, 1;e 13 A k b 3 = O, k= O, 1,2,3. Note that
we do not have e 13 A z b z = Obecause AZb z is a linear combination of {bj , Abj,j=
1,2,3 and AZb l } and e 13 A2b l = 1. Now we use eil'i' i = 1,2,3, to form the square
matrix 6
en

e 13 A

e 13 A z
~elI'IA/ll-l

e 22

P = e 22 A
(7-45)
e 3 4

e 34 A

e 34 A 2

e 34 A3.

óThis matrix can be obtained directly from the coefficients of linear combinations of A~' b¡ without
[uation. computing explicitly M-l. See Reference S178.
346 STATE FEEDBACK AND STATE ESTIMATORS

Using the aforementioned relationship eijAkb =0 or 1, it is straightforward, possible to obtain sorne


though very tedious, to verify that P is nonsingular and 7 form on the diagonal o
c~mp~~ion form on tl
o 1 O : o O O
(A +BK) in (7-47) are
O O 1: I
:,
O O O
X X X'X x'x x x x 1 X O ( s3 -a 3 s Z - azs .
.. __ ...... __ J ........ __ 1_ .. _ ........ _ ...... _
------­
:O 1: O O O and S9 - dgs S _ dsS 7
A= PAP- 1 = X X X: X X: X X X X B=PB= O 1 O Since a¡, b¡, C¡, and d. can
------­
--------~-----~-ó--i--6--0-
, , O O O Theorem 7-7. 1

: : O O 1 O O O O
: : O O O 1 O O O Method '"
X X x:x x:x X X X O O 1
We introduce a method o
(7-46)
forming A into a controlI~
where the unfilled positions are all zeros. The pair {A, B} is said to be in a equation. See Appendix j
multivariable controllable formo The introduction of u = r + Kx, where K is a
3 x 9 real constant matrix, yields Algorithm
x=(A +BK)x +Br Consider a controlIable {
Because of the form of B, all rows of A except the three rows denoted by strings n x P constant matrices. 1
values.
of x are not affected by the state feedback. Because the three nonzero rows
of B are linearly independent, the three rows of A denoted by strings of x can 1. Choose an arbitrary n >
be arbitrarily assigned. For example, we may choose K so that CA +BK) is with those of A.
of the form 2. Choose an arbitrary p x
O 1 O: 01000:0000 3. Solve the unique T in thc
O O 1 : .
.
00100:0000 4. If T is nonsingular the
000.10:0000 ~igenvalues as those'of F.
al az a3 :
--------:-0- -i -:-: - - ------­ - - -

000:01:0000
- .... - _1- _ .. _ _ .. 1_ .. ........ __

K and repea t the process.


:b bz : or O O O: O O: 1 O O O We justify first the algori
-6--6 -0- ~ -o -0-: -Ó- -( - Ó--O
l
-------~-----~----------
: :O 1 O O AT-TF= -Bi( implies
: :O O 1 O 000;00;0010 (A+ BKT-I)l
:, :O O O 1 000:00:0001 Hence A + BK and F are sin
L C.,11 d
'2 d 3 ' d 11(7 l..-1.3
! ' ú6
~4 {ts A a9 l
I j
F can be arbitrarily chosen so
., (7-47)
the eigenvalues of A +BK cal
The left-hand-side matrix has three blocks of companion form on the diagonal; As discussed in Appendix
the right-hand-side matrix has only one companion formo Of course, it is solution T always exists in Al
F lIave common eigenvalues
BK
. . T o remove this uncert~ '

l x x] elgenvalues. A necessary co
7 lf JI, > JIz > JI3, the three nonzero rows of O become O 1 x. lf JII ;5;JIz ;5;JI3, they become controlI~ble and {F, K} obser
lOO] 001 . for the smgle-variable case (p =
O I O In our example, JIl = 3, JIz = 2, and JI3 = 4, and the three nonzero royvs ofO are Theorem 7-10; the proof will
[ O O 1 Hence the assertion will not bl:

r.~l
~ 0'1" o~~J A simple pattern can be deduced from these examples. B Notethat we have d(T)"'AT -TI
d(T) are ,litA) - JIAF).
STATE FEEDBACK 347

possible to obtain sorne other forms, for example, two blocks of companion
1, 1't l'S straightforward,
form on the diagonal or a block triangular form with two or three blocks of
d7 companion form on the diagonal. The characteristic polynomials of the
o O O (A + SK) in (7-47) are
O O O (S3 - a3 s2 - a2 s - al)(s2 - b 2 s - b¡)(S4 - C4S3 - C3S2 - C2S - Cl)
1 x O and S9 - d 9sS - dsS 7 - ... - d2s - d1
0- - Ó- - Ó
S=PB= O 1 O Since a¡, b¡, C¡, and di can be arbitrarily assigned, we have established once again
O--O--Ó Theorem 7-7.
O O O
Method 1I1
O O O
O O 1 We introduce a method of computing the feedback gain matrix without trans­
(7-46) forming A into a controllable formo It will be achieved by solving a Lyapunov
equation. See Appendix F.
{A, B} is_said to be_i~ a
If u = r + Kx, where K lS a
Algorithm
Consider a controllable {A, B}, where A and B are, respectively, n x n and
n x p constant matrices. Find a K so that A + BK has a set of desired eigen­
¡:e rows denoted by strings values.
se the three nonzero rows
1. Choose an arbitrary n x n matrix F which has no eigenvalues in common
enoted by string~ of ! _can
with those of A.
ose K so that (A +BK) is
2. Choose an arbitrary p x n matrix K such that {F, K} is observable.
3. Solve the unique T in the Lyapunov equation AT - TF = - BK.
0 0 : 0 0 00 4. Ir T is nonsingular, then we have K =KT- l , and A +BK has the same
00:000 0 eigenvalues as those of F. Ir T is singular, choose a different F or a different
, 1 0 : 0 0 00 K and repeat the process. I
:- 0- -i -:- 0- -Ó- -Ó - 0­ We justify first the algorithm. Ir T is nonsingular, the Lyapunov equation
:00:100 O AT - TF = - BK implies
:-ó-o-~-ó--(-o--Ó
(A+BKT-l)T=TF or A+BK=TFT-l
:00:0010
Hence A + BK and F are similar and have the same set of eigenvalues. Since
: O O: O O O 1 F can be arbitrarily chosen so long as its eigenvalues are distinct from those of ¡le,
: d 4 d s : d 6 d7 ds d9
the eigenvalues of A + BK can almost be arbitrarily assigned.
" (7-47)
As discussed in Appendix F, if A and F have no common eigenvalues,s a
anion form on the diagonal; solution T always exists in AT - TF = - BK for any K and is unique. Ir A and
lÍon formo Of course, it is F have common eigenvalues, a solution T may or may not exist depending on
BK. To remove this uncertainty, we require A and F to have no common
eigenvalues. A necessary condition for T to be nonsingular is that {A, B}

.~ J
xl ir Il¡ ::;; 112 ::f 1l3' they become
controllable and {F, K} observable. The condition becomes sufficient as well
for the single-variable case (p = 1). The dual of this assertion will be proved in
Theorem 7-10; the proof will again be used in the proof of Theorem 7-11.
three nonzero roWS ofB are Hence the assertion will not be proved here.

8 Notethat we have d(T) =AT - TF rather than d(T) = AT +TF. Hence the eigenvalues of
mples. d(T) are A;(A) - Il}F).
348 STATE FEEDBACK AND STATE ESTlMATORS

Nonuniqueness of feedback gain matrix K


elements of K2 , which
From the design procedures discussed in the previous subsections, we see most of the assigned e
that many different feedback gain matrices can yield the same eigenvalues of cIear that the larger the
(A + BK). Hence it is natural to ask which K should we choose? Before dis­ tudes of the feedback g;
cussing this problem, we give an example. Next we compare tI
yield the same eigenva
Exan'lple 3 transfer matrix C(sI _)
input and zero-state re
Consider the dynamical equation
compare only their ze

x=[L_Ltjj] +[LJ]u
O
O
00:01
O O :-1 -2
x
O
O
O
1
(7-48)
(h) the responses of
[2 1 O -1 _
K¡ is roughly three tim
gains and yields smaller

Y = [1 -1 3: 2 O) x For a companion-for
is shown in Reference S
The equation is in a multivariable controllable formo The problem is to find a proportional to
K so that the eigenvalues of the resulting matrix are -1, -2 ±j, and -1 ± 2j.
We compute
where /Xlmax is its largest t
Ll(5) =5 5 +75 4 +245 3 +485 2 +555 +25
gains are also proportion
If we choose K as

then we have
-
K¡=
.
[-2
-25
O
-55
O
-48
-1
-23
-lJ
-5 10

1 O O

U j]
O
A +BK¡ = O
O
1
O
O
O
1
O
-55 -48 -24
If we choose K as
7 -s
1(2 =l-
r
O
-9
O O
-1
-4 -~J
then we have

..
A+BK2=[_~~ ~tJjj] O
__
00:01
. .. O O O :-5
, -2

This matrix has twó. blocks of companion formo One has eigenvalues
:.... 1, - 2 ±j; the "otherhas -1 ± 2j. These two blocks are noninteracting.
Now we are ready to compare these two K's.. The eIements of Kr , which
yields one block of companion form, are much larger in magnitude than the Figure 7-3 Transient respe
__________________ _ .:::::=--=.:.:.:.._:::..---=.__ ==-_ ..:..-::.:::=-__---::::...----::::..--==-::. ---=-=-=-- - ---------."
v ~ __ ~ _~ ,_~ ~, "_~ ~~

STATE FEEDBACK 349

elements of 1(2' which preserves the original blocks of companion formo If


most of the assigned eigenvalues have magnitudes larger than one, then it is
)us subsections, we see clear that the larger the order of a companion-form block, the larger the magni­
the same eigenvalues of tudes of the feedback gains.
we choose? Before dis­ Next we compare the responses of the two systems. Although 1(1 and 1(2
yield the same eigenvalues, the transfer matrix C(sI - A- 81(1)-18 and the
transfer matrix C(sI - A- 81(2) - 1 B are generally different. Thus, their zero­
input and zero-state responses are generally different. For convenience, we
compare only their zero-input responses. We plot in Figure 7-3(a) and
(b) the responses of ¡=(A+8K 1 )x and ¡=(A+8K 2 )x due to x'(O)=

"j~u
[2 1 O -1 -2]. We see that the largest magnitude in transient in
1(1 is roughly three times larger than the one in 1(2' Hence 1(2 has smalier
(7-48)
gains and yields smaller magnitudes in transient and is preferred. I
1 O
) 1 F or a companion-form matrix of arder m, if all its eigenval ues are distinct, it
is shown in Reference 517 that the largest magnitude in transient is roughly
proportional to
The problem is to find ~

-1, - 2±j, and -1 ±2}.


(IIlmaxt -1
where IIlmax is its largest eigenvalue in magnitude. The magnitudes of feedback
gains are also proportional to m. Hence, in order to have small feedback gains
s +25

-1
·23
.-1]
-5
10

24
1
O

O j] O

-5
~2J:;t~~¿~5~~~4;====t"""'---i6---------- sec

-10 (a) K¡

,J]
, 1
-2
-5

lrm. One has eigenvalues


lcks are nonintera0ing. .
The elements of K 1 , whlch
uger in magnitude than the Figure 7-3 Transient responses.
350 STATE FEEDBACK AND STATE ESTIMATORS

and small transient, the order of the largest companion-form block should be and of eigenvalues ­
kept as small as possible. gains required in (7-5
The orders of companion-form blocks are determined by the lengths of shown by a computer
chains ofvectors b¡, Ab¡, ... , Akb¡ in (7-44). Ifthey are chosen by using Scheme 2 (7-50) is much much s
discussed in (7-21), the orders of companion-form blocks are equal to the We recapitulate t]
controllability indices of {A, B}. In this case, the largest order is equal to the formation of {A, B}
controllability index f1. If we use Scheme 1 discussed in (7-20) or any other Scheme 2 discussed i
scheme, the length of the longest chain, and consequently, the order of the [H AH ... An- 1 B
largest companion-form block will be larger than, at best equal to, f1. Hence in {Á, B} are equal to t
order to have small feedback gains and small transient, we shall use Scheme 2 the order ofthe larges1
to search the linearly independent columns in [B AB ... An-lB] and the among aH transforma!
orders of the resulting companion-forrn blocks are f11, f12, ... , f1 p , the con­ diagonal form with c<
trollability indices. this is not possible du
The introduction of state feedback can increase the order of a companion­ Problem 7-20 to choo
form block, as shown in Example 3, but can never decrease it. Hence in finding K =KP may yield a sy
a K:, we should preserve the orders of the original companion-form blocks. The design procedt
Even so, there are stiB a lot of flexibility in choosing K. We may choose K so single-input system; h<
that the resulting matrix is in block diagonal form or in block triangular form3. the method should not
Furthermore different grouping of desired eigenvalues will yield different K. However, the method 1
If we require the resulting matrix to be block diagonal and group the desired be discussed in Section
eigenvalues to minimize Up to this point, w
mas {(IXll max )!'! - \ (iX 2Imax)!'2 - \ ... , (IXplmax)!'r l} eigenvalues. This dep
where IXdmax is the largest eigenvalue in magnitud~ in the ith companion­ such as the rise time, se1
form block of order f1¡, then the flexibility bf choosing K is considerablyreduced. signals, and so forth. )
A problem may sometimes arise in preserving the orders of companion­ simpIe answer to the I
form blocks. For example, let f1¡ = 3, i = 1, 2, and let the desired eigenvalues simulations. Of cours
be - 3 ± 3j, - 4 ± 2j, and - 3 ±j. In this case, it is not possible to choose a not be unique. The 01
real Kso that A + BK preserve f1¡ and has a block diagonal or a block triangular consequently, a unique
formo Ifwe combine the two companion-form blocks into one, then the result­ formance index
ing A +BK is
J

I ~l

O 1 O O O

O O 1 O O

The K can be uniqueJy


O O O O' (1-49) This is outside the scop
l
-lJ
O O O 1

-3J O O
-4800
We see that the feedback gains are very large. _
O O O
-2804 -920 -180
and S4, or any other bo,

Assignment of eigenva
In this subsection, we ,
Now we use the procedure discussed in Problem 7-20 to find a K so that eigenvectors of the resu
the resulting Á +:DK is of the form assigned eigenvalues, Xi,
O 1 O O O O associated with Xi, that i
O O 1 O O O
-54 -36 -9 : -20 -8 -1
(A
- -

O
-

O
- -

O , O
,
- -

1
- - - - - - -
.
_1- _ _ _ _ _ _ _ _ _ _ _ _ _ _

O
(7-50) or BI

O O O , O O 1
which implies that
18 6 1 : -60 -44 -11
BK[e
STATE FEEDBACK 351

n-form block should be and 01' eigenvalues - 3 ± 3j, - 4 ± 2j, and - 3 ±j. We see that the feedback
gains required in (7-50) are much smaller than those in (7-49). lt can also be
nined by the lengths 01' shown by a computer simulation that the largest magnitude in the transient of
110sen by using Scheme 2 (7-50) is much much smaller than the one in (7-49). See Reference S17.
Ilocks are equal to the We recapitulate the preceding discussion in the following. In the trans­
est order is equal to the formation 01' {A, B} into a multivariable controllable form, we should use
i in (7-20) or any other Scheme 2 discussed in (7-21) to search the linearly independent columns in
uent1y, the order 01' the [B AB ... An-l B]. The orders 01' companion-form blocks in the resulting
:st equal to, fl. Hence in {A, B} are equal to the controllability indices, fl¡, 1, 2, ... , p, 01' {A, B} and
t we shall use Scheme 2 the order 01' the largest companion-form block in {A, B} is the smallest possible
¡J .,. An-1B] and the among all transformations. We then choose a K so that (A +BK) is 01' block
ll, fl2' . . . , fl P'
the con­ diagonal form with companion-form blocks 01' order fli on the diagonal. lf
this is not possible due to complex eigenvalues, we may use the procedure in
e order 01' a companion­ Problem 7-20 to choose a K. This process 01' choosing K and consequently
ease it. Hence in finding K = KP may yield a system with small feedback gains and small transient.
companion-form blo_cks. The design procedure discussed in Method I transforms the problem into a
K. We may choose K so single-input system; hence the order 01' the companion-form block is n. Thus
In block triangular form!. the method should not be used to avoid large feedback gains and large transient.
les will yield different K. However, the method may require a smaller-dimensional state estimator as will
lal and group the desired be discussed in Sections 7-5 and 9-5.
Up to this point, we have not yet discussed how to choose a set 01' desired
eigenvalues. This depends highly on the performance criteria 01' the design
p\max)"p-l}

such as the rise time, settling time, overshoot, largest magnitude 01' the actuating
e in the ¡th companion­
signals, and so forth. Even if these criteria are precisely specified, there is no
K is considerably reduced.
simple answer to the posed problem.One way to proceed is by computer
he orders 01' companion­
simulations. 01' course, the set 01' eigenvalues obtained by this process will
~t the desired eigenvalues
not be unique. The only known systematic way 01' finding a unique K and,
. not possib1e to choose a
consequently, a unique set 01' eigenvalues is by minimizing the quadratic per­
'onal or a block triangular
formance index
~s into one, then the result-

J= 10'" [x*(t)Qx(t) +u*(t)Ru(t)] dt


o O
O O The K can be uniquely determined by solving an algebraic Riccati equation.
O O (7-49) This is outside the scope 01' this text and the reader is referred to References 3
1 O and S4, or any other book on optimal control.
O 1
-180 -20 Assignment of eigenvalues and eigenvectors
In this subsection, we discuss the assignment 01' eigenvalues as well as the
m 7-20 to find a K so that eigenvectors 01' the resulting A +BK. For convenience, we assume that the
assigned eigenvalues, Xi' are all distinct. Let e i be an eigenvector 01' (A + BK)
associated with Xi' that is
1 O
) O (A +BK)ei= Xie i i=1,2, ... ,n
~ -1 or BKe¡ = (XiI - A)e¡ ~ f i
(7~50)
(----Ó
which implies that
) 1
~ -11
- ,_._-_.- .... _-_ .... _~~--~." .__ ._._._._~_._---~-~_._,-_ .. _~---_ .. _- ._~.~,--~--
~ __ • . _ •• __ •• _ ••• , ... _ . ~ , • • • • •• ~ , ~ _ .__ ••• ~ . • __ • ~. _ _ • • , •• u . . . ~ ~ • .•• ~_g •••' _ ' _ _ ~~~~_' ' ~'_'._'~_.~' • • _ _ •• ~ ' _ _ '_~AA_". ----- - . - . - - - . - ••- . - ..... - - - - - .._ ....- _ . __ •• - ._~"- .•

352 STATE FEEDBACK ANO STATE ESTIMATORS

Now if {e i , i = 1,2, ... , n} are linearly independent, then we have


BK = [f 1 f2 . .. f n] [e 1 e2 . . . e n] - 1 (7-51 )

Unless all columns of the matrix on the right-hand side of (7-51) are inside the ~(s)
range space of B, no K will satisfy the equation. Hence the assignment of
eigenvectors cannot be entirely independent from the assignment of eigenvalues.
For a general matrix A, there is no way to predict what form an eigenvector
will assume. However if a matrix is of the companion forms shown in (7-47),
then its eigenvectors are of the form [1 Xi x¡ O O O O O O]',
[O O O 1 Xi O O O O]', or [O O O O O 1 Xi x¡ Xn for the
Figure 7-4 The effect '
left-hand-side matrix in (7-47) and [1 Xi Xl "5.r "5.t "5.( "5.P "5.? "5.~]' for
the right-hand-side matrix in (7-47) (see Problem 2-26). Hence for the {A, S}
in (7-46), after the eigenvalues are chosen, if the eigenvectors are chosen as which implies
discussed, then a solution K exists in (7-51). In fact, the solution is unique an¿
u(s) = (1- ~
can be solved by using the three nonzero rows of S in (7-46).
This process of choosing eigenvectors is far from an arbitrary assignment. The substitution of (7-5:
Once we choose the structure of A +SK (one, two, or three companion-form
y(s) = (N(,
blocks), and once we assign the eigenvalues for each block, the eigenvectors are
practically uniquely determined. Hence if the eigenvectors are chosen as Hence the transfer matr
discussed, there is no essential difference between the assignment of eigen­
vectors and the assignment of structure of (Á + SK) discussed in the previous
Gf (
subsection. For further discussion of the assignment of eigenvectors, the reader
is referred to References S91, S125, S160. We see that the numera
state feedback. The col
Effect on the Numerator matrix of G(s) matrix of the denominó
any state feedback. Th(
In the single-variable case, we showed that the numerator of 9(s) is not affected assigned by a proper ch,
by any state feedback, whereas its denominator can be arbitrarily assigned. case.
Now we establish a similar statement for the multivariable case. Before pro­
If G(s) is factored a1
ceeding, the reader should review the material in Section 6-6, especially Equa­ affected by any state feed
tions (6-100) through (6-117). ~tor of every element of
Once {A, B, C} is transformed into the form in (7-46), by reversing the G(s) wiJI be affected by st
procedure from (6-100) to (6-117), we can obtain feedback gain matrices I
C(s) =C(sIT - A)-l 10 +E = N(s)lD -1 (s) +E = (N(s) + ED(s))JD - j (5) differel1t traj].sient respon

with Computational problem


D(s) = D'.eH(s) + DlcL(s)
N(s) =CL(s) In this subsection, we dis<
III for state feedback. 1
where H(s) and L(s) are defined as in (6-101). We call N(s) +ED(s) the numer­ {A, B} into a multivariabl,
ator matrix and D(s) denominator matrix of G(s). If we plot G(s) as in Figure to achieve this are given i
7-4 and call the output of D -les) the pseudostate a(s), then the state x(s) is, as AkB for k = 1,2, ... , n - )
established in (6-108), equal to x(s) = L(s );(s). Hence the state feedback o(s) = ditioned problem, as discu
r(s) + Kx(s) can be achieved by applying Kas shown in Figure 7-4. From not satisfactory from a co
Figure 7-4 we have

u(s) = r(s) + KL(s)~(s) = r(s) + KL(s)D-l(S)U(S) 9 This observation was providec


STATE FEEDBACK 353

we have E
>
-n
]-1 (7-51 )
A
of (7-51) are inside the + y(s)
mce the assignment of
+
ignment of eigenvalues.
1at form an eigenvector
forms shown in (7-47),
O O O O O O]',
1 X" F X3]' for the Figure 7-4 The effect of state feedback.
~ Xr' Ip Xi~ IP]' for
Hence for the {A, B}
which implies
nvectors are ehosen as
~ solution is unique ane o(s) = (1- KL(s)D-l(S))-l r (S)= D(s)(D(s)- KL(s)tlr(s) (7-52)
7-46). The substitution of (7-52) into y(s) = G(s)ñ(s) = (N(s) + ED(s))D- 1(s)íi(s) yields
n arbitrary assignment.
. three companion-form y(s) = (N(s) + ED(s))D-l(S)D(s)(D(s) - KL(s))-l r (S)
)ck, the eigenvectors are
Hence the transfer matrix from r(s) to y(s) is
lVectors are ehosen as
le assignment of eigen­ G¡(s) = (N(s) + ED(s))(D(s) - KL(s))-1
iscussed in the previous = (N(s) + ED(s))[DheH(s) + (DIe - K)L{s)] -1 (7-53)
i eigenvectors, the reader
We see that the numerator matrix N(s) +ED(s) of G¡(s) is not affected by any
state feedback. The column degrees of H(s) and the column-degree-coefficient
matrix of the denominator"matrix D¡(s) á D(;)-KL(s) are not affected by
any state feedback. The lower degree part of D ¡(s) however can be arbitrarily
.tor of g(s) is not affeeted assigned by a proper choice of K. This result is similar to the single-variable
be arbitrarily assigned. case.
riable case. Before pro­ If G(s) is factored as N(s)D -1(S), the numerator matrix N(s) will not be
ion 6-6, especially Equa­ affected by any state feedback. This, however, does not imply that the numer­
ator of every element of G(s) will be unaffected. In fact, zeros of element of
(7-46), by reversing the G(s) will be affected by state feedback. This is one of the reasons why different
feedback gain matrices K, which yield the same set of poles, may yield vastly
different transient responses. 9
(s) +ED(s))D- 1 (s)

Computational problems
In this subsection, we diseuss sorne computational problems in Methods Il and
III for state feedback. The major step in Method Il is the transformation of
N(s) +ED(s) the numer­ {A, B} into a multivariable controllable formo The equivalence transformations
we plot G(s) as in Figure to achieve this are given in (7-14) and (7-45). They require the computation of
), then the state X(S)Ais, as AkB for k = 1, 2, ... , n - 1 and may change the problem into a less well-con­
the state feedback u(s) = ditioned problem, as discussed in Section 5-8. Hence this procedure is generally
vn in Figure 14 From not satisfactory frorri a computer computational point of view,

'-I(S)O(S) 9 This observation was provided by Professor D. Z. Zheng ofTsinghua University, Beijing, China
- .• - ._- ----.. ". -_•. ,.•".- ·._·,u· •._._.•__ ••. ".~ •..• ·••· ·_.~L_ ~._.,, .. "_~._. ~ _._ .__ .__.._.. ~_ ~_ ..__ .__ ._._~ .~ .. __ . .__ _.__ ~~ __ .__._ . _._._. _ .._. __ ~~_ __.._.___ .__ .._ . _ M .,," ._ _ __._.__ ..
--~----~-------------~~~~~

354 STATE FEEDBACK ANO STATE ESTIMATORS

In Section 5-8, we discussed a different method of transforming an equation state observer. In thi
into a controllable form without explicitIy computing AkB. The method is first full-dimensional and
to transform A into a Hessenberg form as in (5-89), and then transform it into dual to the design 01
a controllable form as in (5-93). (Note the different ordering of the state vari­ single-variable case \\
ables.) The first step is carried out by a sequence of orthogonal transforma­ special case of the mu
tions or gaussian eliminations with partial pivoting and is numerical1y stable. In this section we
The second step must be carried out by gaussian eliminations without any ofthe variable. For e
pivoting and is numerically unstable. Hence this transformation method is
again not satisfactory. Full-dimensional ~
In Section 6-6, we introduced an irreducible realization of G(s) = N(S)O-l(S) time-invariant dynam
in a controllable formo Hence, if N(S)O-l(S) can be computed from {A, B, C},
then a controlIable form dynamical equation can be obtained. In Reference
S17l, an algorithm is presented to compute a coprime N{S)O-l{S) from an
Hessenberg-form dynamical equation. However in the design, we still need where A, R, and e are,
the equivalence transformation. Thus it is not clear whether this design pro­ For simplicity, the dir
cedure will be satisfactory. In References S156 and S157 a dynamical equation assume now that the s
is first transformed into the Hessenberg form, the pole assignment is then state variables are nol
carried out by using the QR algorithm. The methods used in References S156 completely known. J
and S157 are alI numericalIy stable. Once a dynamical equation is transformed x(t) from the availablt
into a Hessenberg form, the Hyman method in Reference S2l2 can also be matrices A, B, and C.
employed to carry out the pole-assignment. For the 20 x 20 matrixon page 219 original system as shc
with e =0, the Hyman method yields a comparable result as the QR method. estimator. Now if the
See Reference S202. initial state and are dr
The procedure in Method 111 does nqt require the transformation of an will be equal to the act
equation into a controlIable formo It requires merely the solution of a Lyapunov is how to find the initi:
equation. The solution of Lyapunov equations has been extensively studied that state. This proble
in the literature, see, for example, References S9, S18, and Sl07. The algorithm dynamical equation Fj
in Reference Sl07 is claimed to be efficient and numericalIy stable and may be from its input and ou
employed in our designo an open-loop estimatOl
A limited comparison between Method n and Method nI in terms of There are, howevel
operational efficiency and numerical stability was carried out in Reference First, the initial state m
S202. Because of the complexities of the problem, no clear-cut concl usion was
reached in Reference S202.
lJ

7-4 State Estimators

In the previous section we introduced state feedback under the assumption


that all the state variables are available as outputs. This assumption often does
not hold in practice, however, either because the state variables are not acces­
sible ror direct measurement or because the number of measuring devices is
limited. Thus, in order to apply state feedback to stabilize, to optimize, or to
decouple (see Section 7-6) a system, a reasonablesubstitute for the state vector
often has to be found. In this sedion we shall show how the available inputs
and outputs of a dynamical equation can be used to drive a device so that the
outputs of the device will a¡)proximate the state vector. The device that con­
strucís an approximation of the state vector is calIed a state estimato,. or a Figure 7-5 An open-Io
STATE ESTIMATORS 355

ansforming an equation state observer. In this section we shal1 introduce two types of state estimators:
.kB. The method is first ful1-dimensional and reduced-dimensional. The design procedure is basical1y
d then transform it into dual to the design of state feedback; hence its díscussion will be brief. The
'dering of the state vari­ single-variable case will not be discussed separately and will be included as a
orthogonal transforma­ special case of the multivariable case.
Id is numerically stable. In this section we use the circumflex over a variable to denote an estimate
íminations without any x
of the variable. For example, is an estimate ofx; i is an estimate of X.
msformation method is
Full-dimensional state estimator. Consider the n-dimensional linear
:ion of G(s) = N(s)D-1(s) time-invariant dynamical equation
)mputed from {A, B, C}, FE: X=Ax +Bu (7-54a)
obtained. In Reference y=Cx (7-54b)
.me N(s)D - 1 (s) from an
the design, we still need where A, B, and C are, respectively, n x n, n x p, and q x n real constant matrices.
whether this design pro­ For simplicity, the direct transmission part has been assumed to be zero. We
57 a dynamical equation assume now that the state variables are not accessible. Note that although the
pole assignment is then state variables are not accessible, the matrices A, B, and C are assumed to be
; used in References Sl56 completely known. Hence the problem is that of estimating or generating
1equation is transformed x(t) from the available input u and the output y with the knowledge of the
erence S2t2 can also be matrices A, B, and C. If we know the matrices A and B, we can duplicate the
Ox 20 matrixon page 219 original system as shown in Figure 7-5. We called the system an open-Ioop
~esult as the QR method. estimator. Now if the original equation FE and the estimator have the same
initial state and are driven by the same input, the output x(t) of the estimator
the transformation of an will be eq.ual to the actual state x(t) for al1 t. Therefore, the remaining question
he solution ofa Lyapunov is how to find the initial state of FE and set the initial state of the estimator to
been extensively studied that state. This problem was solved in Section 5-4. It is shown there that if the
and St07. The algorithm dynamical equation FE is observable, the initial state of FE can be computed
~rically stable and may be from its input and output. Consequently, if the equation FE is observable,
an open-Ioop estimator can be used to generate the state vector.
Method IIIin terms of There are, however, two disadvantages in using an open-Ioop estimator.
carried out ih Reference First, the initial state must be computed and set each time we use the estimator.
) clear-cut conclusion was

lck under the assumption


~his assum ption often does
.te variables are not acces­
~r of measuring devices is
.tabilize, to optimize, or to I

)stitute for the state vector I

!/ .how the available inputs


\ . I
I A I
) drive a device so that the L ~~~~ ~ J
:tor. The devicethatcon­
Figure 7-5 An open-loop state estimator.
lled a state estimator or a

356 STATE FEEDBACK AND STATE ESTIMATORS

This is very inconvenient. Second, and more seriously, if the matrix A has which can be written
eigenval ues with positive real parts, then even for a very smal1 difference between
x(t o) and x(to) at sorne to, which may be caused by disturbance or incorrect
estimation of the initial state, the difference between the actual x(t) and the
estimated x(t) will increase with time. Therefore an open-loop estimator is, in The asymptotic e;
general, not satisfactory. either by a block diag
Another possible way to generate the n-dimensional state vector is to
differentiate the output and the input n - 1 times. If the dynamical equation is
observable, then from u(t), y(t), and their derivatives, the state vector can be Clearly x is the error
computed (see Problem 5-17). However, pure differentiators are not easy to tracting (7-56) from (7
build. Furthermore, the estimated state might be severely distorted by noises
if pure differentiators are used.
We see from Figure 7-5 that although the input and the output of FE are
available, we use only the input in the open-loop estimator. It is conceivable If the eigenvalues of (A
that if both the output and input are utilized, the performance of an estimator error x can be control
can be improved. negative real parts sm
Consider the state estimator shown in Figure 7-6. The estimator is driven zero at rates faster thar
by the input as wel1 as the output of the original system. The output of FE, x(to) and x(t o) at initi2
Y=Cx, is compared with y ~ Cx, and their difference is used to serve as a the eigenvalues of A ­
correcting termo The difference of y and Cx, y -Cx, is multiplied by an n x q is much more desirabl
real constant matrix L and fed into the input of the integators of the estimator. In the fol1owing \
This estimator wil1 be cal1ed an asymptotic state estimator, for a reason to estimator.
be seen later.
The dynamical equation of the asymptotic state estimator shown in Figure Method I
7-6 is given by
In this method, we ap
i =Ax +Bu +L(y-Cx) (7-55 ) state estimators.

l
,------------------1
I + I
I ,---------­
I
I
I

I
x

I I
I I
I I
I A I
I L
:
L- --.JI I
L _
Figure 7-6 An asymptotic state estimator. Figure 7-7 An asympt
STATE ESTlMATORS 357

y, if the matrix A has which can be written as


nal1 difference between
sturbance or incorrect i. =(A -Le)x +Ly +Bu (7-56 )
he actual x(t) and the
The asymptotic estimator in Figure 7-6 can be redrawn as in Figure 7-7
~n-Ioop estimator is, in
either by a block diagram manipulation or from (7-56). Define
Iilal state vector is to
: dynamical equation is
x~ x-x
he state vector can be Clearly x is the error between the actual state and the estimated state. Sub­
.tiators are not easy to tracting (7-56) from (7-54), we obtain
~ely distorted by noises
i=(A-LC)x (7-57)
d the output of FE are If the eigenvalues of (A - Le) can be chosen arbitrarily, then the behavior of the
.ator. It is conceivable
error x can be control1ed. For example, if al1 the eigenvalues of (A - Le) have
rmance of an estimator negative real parts smal1er than - CJ, then al1 the elements of x will approach
zero at rates faster than e-a'. Consequently, even ifthere is a large error between
The estimator is driven x
x(to) and x(t o) at initial time t o, the vector will approach x rapidly. Thus, if
:m. The output of FE, the eigenvalues of A - Le can be chosen properly, an asymptotic state estimator
e is used to serve as a is much more desirable than an open-loop estimator.
: multiplied by an n x q In the following we discuss two different methods of designing a state
:gators of the estimator. estimator.
'mator, for a reason to

imator shown in Figure Method I


In this method, we apply the design procedures for state feedback to design
(7-55) state estimators.

,----------------------,
I I
I
I I
I I
I I ~

I
I
I
I
IL I
~

Figure 7-7 An asymptotic state estimator.


==_._._
~ ~ --===-_ _.__ _._._._.__ __"
~. __.. _~.~. .
~~"_~ .. ..-.=--::::="-::::""=_.::::::::.:: .._:::,.::::
-.~._ =-~ =--~"-=":::: :_ __ ,:,::
.~:::, __",._" ._"__.. __". __..__
._._~ ..
._,_.u~_
· .. _._ .~. ~~ __ ~~ •• __ __•••
._.~ _.~_~. ..•,•••.••••_. .~_u •• _ ~ __ . . ..••_.• _~ •• _..._ ...._.__ •.._. __ u, •••" "
-;­ _

358 STATE FEEDBACK AND STATE ESTIMATORS

Theorem 7-8 Proof

If the n-dimensional dynamical equation FE in (7-54) is observable, its state Define


can be estimated by using the n-dimensional state estimator
Then we have
~ = (A ­ LC)x + Ly + Bu

with the error x =x ­ xgoverned by


If the three conditions
and u(t). Hence z(t) is
and all the eigenvalues of A - Le can be arbitrarily assigned, provided complex Now we show the
conjugate eigenvalues appear in pairs. • x(O) = O and u(t) == O, \
find a u(t) to make e(t
To show the theorem, it is sufficient to show that if {A, C} is observable, trollable, we can find ;
then all the eigenvalues of A - Le can be arbitrarily assigned. lndeed, if This establishes the ne(
{A, C} is observable, then {A', C}, where the prime denotes the transpose, is
controllable (Theorem 5-10). Hence for any set of n eigenvalues, we can find With this theorem,
a real K so that A' +CK has the set as its eigenvalues (Theorem 7-7). Since
the eigenvalues of A + K'C and those of A' +C'K are the same, the theorem is Algorithm
established by choosing L = - K'.
It is clear that all the design procedures for state feedback can be used to 1. Choose an F so tha
design state estimators. Hence their discussions will not be repeated. We disjoint from those (
mention only that, in order to have small gains in L and smal1 transient in 2. Choose a G so that
i = (A ­ Le)x, the largest order of companion-forrn blocks in the transformed 3. Solve the unique Ti
A matrix' should be as smal1 as possible. Dual to the state-feedback part, the 4. lf T is nónsingular,
smallest such order is the observability index of {A, C}. F, G, and H is an est
different F and/or (
Method 11
We give sorne remar J
Consider the n-dimensional dynamical equation eigenvalue, then a solu
(Appendix F). lf A and
FE: X=Ax +Bu (7-58a) may not exist in TA­
y=Cx (7-58b) and A to have no comn
The controllability (
where A, B, and e are, respectively, n x n, n x p, and q x n real constant matrices.
as will be established in
It is assumed that FE is irreducible. Define the n-dimensional dynamical
equation
Theorem 7 -1 O
z=Fz +Gy +Hu (7-59)
lf A and F have no com
of a nonsingular T in '
where F, G, and H are, respectively, n x n, n x q, and n x p real constant matrices.
trollable. For the singlt
Theorem 7-9 Proof
Thestate z(t) in (7-59) is anestimateofTx(t) for sornen x n. real constant matrix T
Let the characteristic p.
in the sense that z(t) ­ Tx(t)-+ O as t -+ 00, for any x(O), z(O), ánd u(t) if and only if
Ó(s) = det
1. TA-Ff=GC
2. H=TB (7-60) Clearly we have Ó(A) =
3. All the eigenvalues of F have negative real parts. F, then Ó(A.¡) is an eiger
STATE ESTIMATORS 359

Proof
is observable, its state Define e ~ z-Tx
ltor Then we have

é =:i - Ti = Fz + Gy + Hu - TAx - TBu


= Fe +(FT - TA + GC)x +(H - TB)u

°
Ir the three conditions in (7-60) are met, then e(t) = eFte(O) -+ for any x(O), z(O),
and u(t). Hence z(t) is an estimate of Tx(t).
ned, provided complex Now we show the necessity of the conditions. Ir 3 is not met, then for
• x(O) =0 and u(t) =0, we have e(t)=eF1z(0)+0 as t-+oo. IfHi=TB, we can
find a u(t) to make e(t)+O as t-H/J. Ir TA - FT i= GC and if {A, B} is con­
f {A, e} is observable, trollable, we can find a u(t) to generate a x(t) which makes e(t)+O as t-+ oo.
'assigned. Indeed, if This establishes the necessity part of the theorem. Q.E. D.
notes the transpose, is
genvalues, we can find With this theorem, we can now propose a design algorithm.
(Theorem 7-7). Since
le same, the theorem is Algorithm

edback can be used to 1. Choose an F so that all of its eigenvalues have negative real parts and are
not be repeated. We disjoint from those of A.
and small transient in 2. Choose a G so that {F, G} is controllable.
cks in the transformed 3. Solve the unique T in TA - FT = Gc.
tate-feedback part, the 4. Ir T is nonsingular, computeH =TB. The equation in (7-59) with these
F, G, and H is an estimate ofTx(t) or x(t)=T- 1 z(t). Ir T is singular, choose
different F and/or G and repeat the process. •

We give sorne remarks regarding this algorithm. Ir A and F have no common


eigenvalue, then a solution T always exists in TA - FT = GC and is unique
(Appendix F). Ir A and F have common eigenvalues, then a solution T may or
fl-58a) may not exist in TA - FT = GC. To remove this uncertainty, we require F
(7-58b) and A to have no common eigenvalues.
The controllability of {F" G} is needed for the existence of a nonsingular T
!real constant matrices. as will be established in the following theorem.
limensional dynamical
Theorem 7 -1 O
fl-59)
Ir A and F have no common eigenvalues, necessary conditions for the existence
of a nonsingular T in TA-FT=GC are {A,C} observable and {F, G} con­
1 real constant matrices.
trollable. For the single-output case (q = 1), the conditions are sufficient as well.

Proof
n real constant matrix T Let the characteristic polynomial of A be
O), and u(t) if and only if
Ll(s)=det(sI-A)=sn +a¡sn-¡'+azsn- z + ..'. +an
fl-60) °
Clearly we have Ll(A) = (Cayley-Hamilton theorem). Ir A¡ is aneigenvalue of
F, then Ll(AJ is an eigenvalue of Ll(F) (Problem 2-32). Since A and F have no
360 STATE FEEDBACK ANO STATE ESTIMATORS

common eigenvalue, we have Ó,(Ai) +0 for al1 eigenvalues Ai of F and dynamical equation
n
no means simple nUD
det Ó,(F) = TI ó,(},¡) + O It is not clear at presen
i= 1 efficiency and numeril
(Problem 2-22). Hence Ó,(F) is nonsingular.
The substitution of T A = FT + GC into TA 2 - F 2T yields Reduced-dimensio
T A2 - F 2T = (FT + GC)A - F 2T = F(TA - FT) + GCA = FGC + GCA Method I
Proceeding similarly, we can obtain the fol1owing set ofequalities: Consider the n-dimen~

TI-IT=O
TA-FT=GC
TA 2 - F 2T = GCA + FGC where A, B, and C are,
TA 3 - F 3 T = GCA 2 + FGCA + F 2GC
In this section we assUl
TAn-FnT=GCAn-1 +FGCAn-2 + ... +FI- 2GCA+F"- I GC
We multiply the first equation by !X n , the second equation by !X n - I , ... , the last
equation by 1, and then sum them up. After sorne manipulation, we finally where R is an (n -q) x
obtain as P is nonsingular. V

where QI and Q2 are r.


o
O In =PQ =[
(7-61 ) Now we transform
x=Px,
where UF is the control1ability matrix of {F, G}, and VA is the observability
matrix of {A, C}. The matrix A q is an nq X nq matrix and is similar to the one in
(7-13). Since Ó,(A) =0 and Ó,(F) is nonsingular, Equation (7-61) implies
T=_ó,-I(F)UFAqV A (7-62) which can be partitione
From this equation we conclude that if ranJe UF < n or rank V A < n, then rank ';[¡1 r
T < n (Theorem 2-6) and T is singular. However, rank UF = n and rank V A = n
do DOt imply the nonsingularity ofT. Hence they are only necessary conditions
Lx J=l
2

for T to be nonsingular. y =[
Ir q = 1, UF' A q, and V A are n X n square matrices. A q is always nonsingular.
Hence T is nonsingular if and only ifU F and VA are nonsingular or, equivalently, where XI consists of thl
{F, G} is control1able and {A, C} is observable. This establishes the theorem. A12 , A2!> and A22 are,
Q.E.D. (n -q) matrices;.B¡ and
that y = Xl ' Hence on
Although the unique solution T in TA - FT = GC can be computed from Consequehtly we need 01
(7-62), the methódmay not be desirable because it requires the .computation n-dimensional estimatol
of UF and VA (see Section S-S). The computational problem of TA - FT = GC Usiilg XI =Y,we wri
has been studied exterisivelyin the literature; see, for example, References S9,
S18, and S107.
The design procedure in Method 1 requires the transformation of a
STATE ESTIMATORS 361

dynamical equation into a multivariable observable formo This step is by


les }.¡ of F and
no means simple numerically. 5ee the discussions in the state-feedback part.
It is notclear at present which method, 1 or I1, is better in terms of computational
efficiency and numerical stability.

Reduced-dimensional state estimator


yields
;CA = FGC + GCA Method I

f equalities: Consider the n-dimensional dynamical equation


x=Ax +Bu (7-63a)
y=Cx (7-63b)
where A, B, and C are, respectively, n x n, n x p, and q x n real constant matrices.
In this section we assume that C has a full row rank, that is, rank C = q. Define
....... - .

GCA +F"-IGC

P~ [~J (7-64 )
on by C1. n - l " .. , the last
lanipulation, we finally
where R is an (n - q) x n real constant matrix and is entirely arbitrary so long
as P is nonsingular. We compute the inverse ofP as

(7-65 )

where Ql and Qz are n xq and n x (n - q) matrices. Clearly we have

o
O
III=PQ=[~J[Ql Qz]=[~~: ~~:J=[~q ~-J (7-66)

Now we transform Equation (7-63) into, by the equivalence transformation


(7-61 )
x=Px,
V A is the o bservability j =PAP-Ix +PBu
d is similar to the one in y=CP-lx=CQx=[Iq O]x
on (7-61) implies

(7-62) which can be partitioned as

rank V A < n, then rank (1'-67 a )


UF = n and rank V A = n
lly necessary conditions
(7-67b)
'q is always nonsingular.
ingular or, equivalently, ~here_Xl consis!s of the first q elements of X and Xz the remainder of x; Á ll ,
~stablishes the theorem. A IZ ' A ZI ' and A 22 are, respectively, q x q, q x (n- q), (n - q) x q, and (n - q) x
Q.E.D. (n - q) matrices; B l and Bz are partitioned accordingly. We see from (7-67b)
that y = x¡. Hence only the last n --' q elements of x need to be estimated.
can be computed from
Consequently we need only an (n - q)-diinensional state estimator rather than an
quiresthe' computation
n-dimensional estimator as discussed in ihe' previous subsection.

lblem ofTA - FT = GC
Using Xl = y, we write (7-67a) as .

~xample, References 59, y = At1 y + A12 x 2 + BID

e transformation of a i z =A 22 x Z +AzlY +Bzo (7-68a)

.J
362 STATE FEEDBACK AND STATE ESTIMATORS

which become, by defining ü = A 2¡y + B 2u and w =y - A¡¡y - B¡ U, Since x = Px, We have


x 2 =A 22 x 2 +ü
W =A 12x2 (7-68b)

We note that ü and w are functions of known signals u and y. Now if the This gives an estimate
dynamical equation in (7-68) is observable, an estimator of x 2 can be constructed. ~iag.ram ofthe (n - q)-d
In Figure 7-8.
Theorem 7 -11
A comparison betw
The pair {A, C} in Equation (7 -63) or, equivalently, the pair {A, C} in Equation estimators is in order.
(7-67) is observable if and only if the pair {A 22 , A 12 } in (7-68) is observable. • alence transformation d
required in the reduced­
The controllability part of this theorem was implicitIy established in Section dimensional case. In th
5-8. See also Problem 5-34. Thus the proof ofthis theorem is left as an exercise. of integrators. In the I
Ir {A, C} is observable, then {A 22 , A¡2} is observable. Consequently, there the constant matrix Q t
.h ¡
exists an (n - q)-dimensional state estimator of x 2 of the form Wlt noises, the nbises VI
dimensional estimator t1
(7-69 ) noises in y will be sup~n
such that the eigenvalues of (A 22 - LA 12) can be arbitrarily assigned by a Method 11
proper choice ofL (Theorem 7-8). The substitution ofü and w into (7-69) yields
Consider the n-dimensior
(7-70)

This equation contains the derivative of y. This can be elir~lÍnated by defining


z =i 2 -Ly (7-71) where A, B, and C are res¡;
Then (7-70) becomes
It is assumed that niis ir

z =(A 22 -LAd(z +Ly) +(A 2¡ -LA l1 )y +(B 2 -LBdu


=(A 22 -LAdz + [(A 22 -LAdL +(A 2¡ -LA l1 )]y +(B 2 -LBdu (7-72) be an (n - q)-dimensional ,
(n - q) x (n - q), (n - q) x q,
This is an (n - q)-dimensional dynamical equation with u and y as the inputs
and can be readily constructed by using, for example, operational amplifier
circuits. From (7-71), we see that z +Ly is an estimate of x 2. Indeed, if we
define e = x 2 - (z +Ly), then we nave

é =i 2 -(z +Li¡) =A 2¡x¡ +A 22 x 2 +B 2u -(A 22 -LAd(z +Lx¡)


-(A 2¡ -LA l1 )x¡ -(B 2 -LBdu-LA l1 x¡ -LA12x2 -LB¡u
= (A 22 - LA12)(X 2 - Z - Lx¡)
or é = (A 22 - LA 12)e (7-73)

Since the eigenvalues of (A 22 - LA 12) can be arbitrarily assigned, the rate of


e(t) approaching zero or, equivalentIy, the rate of (z + Ly) approaching x 2 can
be determined by thedesigner. Hence z +Ly yields an estimate of X2'
Nowwe combine x¡ =y ;=i¡ with i 2 =z +Ly to form

Figure 7-8 An(n -q)-dimen:


=='--__ -
=:::_=-:c-_--~
- ---- _

STATE ESTIMATORS 363

Since x =Px, we have x = p- 1 x = Qx or

(7-68b)
X=Q~=[Q¡ Q2J[L/+Z]=[Q¡ Q2J[~ I~-JGJ (7-74 )

This gives an estimate of the original n-dimensional state vector x. A block


" and y. Now if the
diagram of the (n - q)-dimensional state estimator in (7-72) and (7-74) is plotted
'X 2 can be constructed.
in Figure 7-8.
A comparison between an (n - q)-dimensional and an n-dimensional state
estimators is in order. The reduced-dimensional estimator requires the equiv­
air {A,C} in Equation alence transformation discussed in (7-64). Excluding this step, the computation
7-68) is observable. • required in the reduced-dimensional case is c1eady less than the one in the full­
dimensional case. In the implementation, the former also requires less number
. established in Section of integrators. In the reduced-dimensional case, the signal y is fed through
~m is left as an exercise. the constant matrix Q 1 to the output of the estimator. Hence, if y is corrupted
. Consequently, there with noises, the noises wil\ appear in the output of the estimator. In the full­
form dimensional estimator, the signal y is integrated or filtered; hence high-frequency
(7-69)
noises in y will be suppressed in the full-dimensional estimator.

itrarily assigned by a
Method 11
md w into (7-69) yields
Consider the n-dimensional dynamical equation
(7-70) FE: x=Ax +Bu (7-75a)
y=Cx (7-75b)
eliminated by defining
where Á, B, and C are, respectively, n x n, n x p, and q x n real constant matrices.
(7-71 ) 1t is assumed that FE is irreducible and rank C = q. Let
z=Fz+Gy +Hu (7-76)

be an (n -q)-dimensional dynamical equation. F, G, and H are, respectively,


-(B 2 - LBd" (7-72) (n - q) x (n - q), (n - q) x q, and (n - q) x p real constant matrices to be designed.
l U and y as the inputs
, operational amplifier
:e of x2 . Indeed, if we
-----------------------,

-LAd{z + LX 1)

--~
~A12X2 - LB 1"

(7-73)
h
X

ly assigned, the rate of


~y) approaching X2 can
[1 estimate of X2'

rm

Figure 7-8 An (n -q)-dimensional state estimator.


.. _- -_ .. _... --.--~._-----_ ...._--- ---_.. _.. __._..-_ .._--_. _._-_.-.-._--_._----- ----_. __ .._.__ .__...-

364 STATE FEEDBACK ANO STATE ESTIMATORS

The procedure of computing F, G, and H is similar to Method 11 of the full­ Now we consider tl
dimensional case. theorem by showing th;

Algorithm
1. Choose an (n - q) x (n - q) real constant matrix F so that al! of its eigen­ Define Il = Al VAf = LB]
values have negative real parts and are di~tinct from those of A.
2. Choose a G so that {F, G} is controllable, that is
rank [G FG ... Fn-q-1G]=n-q

3. Solve the unique T in TA - FT = GC. Note that T is an (n - q) x n matrix.


4. If the square matrix of order n where x denotes nonze:

P=[~J

then we have f31l=O. F


[7-77)

is singular, go back to step 1 and/or step 2 and repeat the process. If P is


nonsingular, compute H =TB. Then the z in Equation (7-76) is an estimate which together with the
of Tx and the original state x can be estimated by Hence (7-79) and rank 1
Because Al is nons:
(7-78) with the observability (
I observable and {F, G}
the proof of this theore:
The proof of Theorem 7-9 can be used to establish that z is an estimate
ofTx. The combination ofy =ci and z =Ti yields immediately (7-78). The discussion in ]
computational probler
Theorem 7 -1 2 dimensional case regar.
equally applicable here
If A and F have no common eigenvalues, necessary conditions for the existence To conclude this se(
of a full rank T in TA - FT = GC such that
Theorem 7 -1 3
P=[~J
If the n-dimensional d:
is nonsingular are {A,e} observable and {F, G} controllable. For the single­ sional state estimator, a:
output case (q = 1), the conditions are sufficient as well. cles1,ed é:igenva!ues (pr
can be constructed, Whl
Proof
First we note that Equation (7-62) still holds for the present case. The only 7-5 Connection.
differences are that F is an (n - q) x (n - q) matrix and UF is an (n - q) x nq
matrix. We write Consider the n-dimensi

P= [~J =[ -Ll -l(~UFAqV [~


If {F, G} is not controllable, thenrank UF < n - q. Hence we have rank
J= -Ll J
~l(F)J[UF~qV
where A, B, and C are, r'
It was shown in Sectio
T < n - q and P is singular. If {A, C} is not observable, there exists a nonzero
the state feedback u =
n x 1 vector f such that VAf = O; This implies Cf = Oand Pr = O. HenceP is
desired positions. It '"
singular. This establishes the necessity of the theorem.
CONNECTIÚN OF STATE FEEDBACK AND STATE ESTIMATOR 365

1 Method II of the full- Now we consider the single-output case. We show the sufficiency of the
theorem by showing that r = O is the only possible solution in

C ] r-O (7-79)
[ UFAIV A ­
so that aH of its eigen­ ¡1n]', ar
Ithose of A.

l
~~ IlllXn-
2 =
1 . IX
IX n.- 2
n- 2
IX n.- 3
IXI
1
n-q .
. .. ..
is an (n - q) x n matrix. fJn 1 O O
where x denotes nonzero elements. From this equation, we see that if Cr = O,
then we have ~II=O. From UFA 1 VAr=UdJ=O, we have, by using ~n =0,
0-77)
" n- 1
Uf·A. V Ar= L fi¡F i - 1
G= L fiiFi- 1 G =0 (7-80)
eat the process. If P is j-= 1 i= I
ion (7 -76) is an estimate which together with the controllability of {F, G} imply Pi = O, i = 1, 2, ... , n - 1.
Hence (7-79) and rank UF = n -1 imply p = O.
Because Al is nonsingular, p=AIVAr=O implies VAr=O, which together
(7-78) with the observability of {A, C} imply r = O. This establishes that if {A, C} is

• observable and {F, G} is controllable, then P is nonsingular. This completes


the proof of this theorem. Q.E. D.
;h that z is an estimate
nmediately (7"78). The discussion in Method II of the full-dimensional case regarding the
computational problem and the discussion in Method 1 of the reduced­
dimensional case regarding the comparison with the full-dimensional case are
equally applicable here; hence they will not be repeated.
ditions for the existence To conclude this section, we summarize the preceding results as a theorem.

Theorem 7 -13
If the n-dimensional dynamical equation FE is observable, an (n -q)-dimen­
sional state estimator, as given in (7-72)and (7-74)or in (7-76) and (7-78), with any
ollable. For the single-
desired eigenvalues (provided complex conjugate eigenvalues appear in pair)
can be constructed, where q is the rank of matrix C in FE. 11

present case. The only 7-5 Connection of State Feedback and State Estimator
Id UF is an (n - q) x nq
Consider the n-dimensional dynamical equation
FE: x=Ax +Bu (7-81 a)

(F)][U:qVJ y=Cx
where A, B, and C are, respectively, n x n, n x p, and q x n real constant matrices.
(7-81 b)

Hence we have rank


It was shown in Section 7-3 that if {A, B} is controHable, the introduction of
a
e, there .exists· ~onzero
the state feedback u = r + Kx can place the eigenval ues of (A+ BK) in any
and Pr = O. Hence P is
desired positions. It was shown in Section 7-4 that if {A,C} is observable, a
---_ _ __ .• "'_._._-.. _._--_ _-~ , .._ ,_ .. __ _._. __ .. _------".~---_._.- _. ---_ .. __ .. _-_. ._..__ .. _
__ ~ -.-.--" .. ..
~-_ _-_ .. ~ _ _~.~ ..

366 STATE FEEDBACK ANDSTATE ESTIMATORS

state estimator, with arbitrary eigenvalues, of dimension n or n - q can be To simplify (7-85), w


constructed by using u and y as the inputs. In the fol1owing, we consider only
the reduced-dimensional case; the ful1-dimensional case is left as an exercise.
Consider the (n - q)-dimensional state estimator developed in (7-76) and (7-78):
After sorne manipu\.
z=Fz +Gy +Hu (7-82a)
final1y obtain the fol1
(7-82b)

with TA -FT = GC, H =TB and

[Q1 Q2J [~] =Q1 C +Q2T =1 (7-83)

The eigenvalues of a
Ir the eigenvalues of F are chosen to have negative real parts, the estimated
transformation; henc
state x(t) will approach the actual state x(t) exponential1y.
system in Figure 7-9
The state feedback is designed with respect to the actual state of FE. Ir the
that the eigenvalues (
actual state x(t) is not available for feedback, it is natural to apply the feedback
as far as the eigenval i
gain K to the estimated state x(t), that is,
from the estimated st:
u =r +Kx (7-84) of state feedback ane
pendently and the eig'
as shown in Figure 7-9. Three questions may be raised in this connection:
feedback and those (
(1) In the state feedback" u = r + Kx, the eigenvalues ofthe resulting equation are
separation property.
given by the eigenvalues of A + BK. In the estimated state feedback u = r + Kx,
We discuss now Ü
do we stil1 have the same set of eigenvalues? (2) Will the eigenvalues of the
shown in (5-54). He]
state estimator be affected by the feedback u = r + Kx? (3) What is the effect
7-9 can be computed
of the estimator on the transfer-function matrix from r to y? To answer these
questions, we form the composite dynamical equation of the system in Figure
7-9 as as

[i] [Ai -
+BKQ1C
GC +HKQ1C
BKQ2
F +HKQ2
][x] [B]
z + H r
(7-85a)
(Theorem 5-16). Thi~
the use of a state estim
and does not appear
y =[C OJ [:] (7-85b) explanation. In the c,
to be zero. Consequ
This is obtained by the subsütution of (7-84) and (7-820) into (7-81) aad (¡-¿La). aH t. Hence, aS fu; do
difference whether a ~
Ir x(O) f- x(O), the
values of the estimate
parts of these eigenva
cause larger gains in
state feedback case.
mator will act as a dil
and large transient W.
simple answer to the
suggested in the liten
two or three times fas
ofthe negative real pa
Figure 7-9 Feedback rrom the estimated state. of those of A + BK. '
CONNECTION OF STATE FEEDBACK ANO STATE ESTIMATOR 367

Ion n or n - q can be To simplify (7-85), we introduce the following equivalence transformation


wing, we consider only
~ is left as an exercise.
'ed in (7-76) and (7-78):
After sorne manipulation and using T A - FT = Ge, H = TB and (7-83), we
(7-82a)
finally obtain the following equivalent dynamical equation
(7-82b)
[:]=[A ~BK BK
F
Q2][:] +[:]r (7-86a)

(7-83)
y=[C OJ [:] (7-86b)

The eigenvalues of a dynamical equation are invariant under any equivalence


11 parts, the estimated
transformatian; hence we conclude from (7-86a) that the eigenvalues of the
y.
system in Figure 7-9 are the unian of those of A +BK and those of F. We see
:ual state of FE. If the
that the eigenval ues of the state estimator are not affected by the feedback and,
1to apply the feedback
as far as the eigenvalues are concerned, there is no difference in state feedback
from the estimated state xor from the actual state x. Consequently, the design
of state feedback and the design of a state estimator can be carried out inde­
pendently and the eigenvalues of the entire system are the union of those of state
ed in this cannection:
feedback and those of the state estimator. This property is often called the
: resulting equation are
separation property.
te feedback u = r + Kx,
We discuss now the transfer matrix from r to y. Equation (7-86) is ofthe form
the eigenvalues of the
shown in (5-54). Hence the transfer matrix of (7-86) or of the system in Figure
(3) What is the effect
7-9 can be computed from
o y? To answer these
lf the system in Figure x = (A + BK)x + Br y = Cx
as G¡(s) =C(sI -A -BKt lB
(Theorem 5-16): This is the transfer matrix ofthe perfect state feedback without
(7-85a)
the use ofa state estimator. In other words, the estimator is completely canceled
and does not appear in the transfer matrix from r to y.This has a simple
(7-85b) explanation. In the computation oftransfer matrix, aH initial states are assumed
to be zero. Consequently, we have x(O) = x(O) =0, which implies x(t) =x(t) for
into (7-81) and (7-82a). al! t. Hence, as rar as the transfer matrix fram r to y is concemed, tnere is no
difference whether a state estimator is employed or no1.
If x(O) i=- x(O), the rate of x(t) approaching x(t) is determined by the eigen­
values of the estimator. Clear1y, the larger the magnitudes of the negative real
parts of these eigenvalues, the faster x(t) approaches x(t). However, these will
cause larger gains in L and generate larger magnitudes in transient, as in the
state feedback case. As the eigenvalues approach negative infinity, the esti­
mator will act as a differentiator and will be susceptible to noises. Large gains
and large transient will also cause the estimator to saturate. Hence there is no
simple answer to the question of what are the best eigenvalues. Ithasbeen
suggested in the literature that the eigenvalues of A - LC or F be chosen to be
two or three times faster than the eigenvalues of A + BK; that is, the magnitudes
ofthe negative real parts ofthe eigenvalues of A - LC or F betwo or three times
of those of A +BK. This seems to be a simple and reasonable guideline. lf a
368 STATE FEEDBACK ANO STATE ESTlMATORS

dynamical equation is corrupted by noises and modeled as a stochastic process,


the matrix L can be computed by minimizing the mean square error. The
interested reader is referred to References S4 and S10.

Example 1
Now we may apply th.
Consider the inverted pendulum problem studied in Exmple 2 on page 338.
Its dynamical equation is

l~ :
d-!---
~ I~- -~ -~l + r-! l
oJO J 5 -2
This completes the des

Functional estimat
an estimator of dimens
not necessary to recom
y = [1 : O O OJ x some functions of the
As discussed there, if we introduce the state feedback In these cases, the dim
In the following, we d
u(t) = r(t) + \j lf- ~ ~Jx kx, where k is an arbitr
then the resulting equation has eigenvalues -1, -2, and -1 ±j. functional estimator or
Now if the state x is not available for feedback, we must design a state esti­ Consider the dynaI
mator. We use Method 1 of the reduced-dimensional state estimator to design is to design an estimat
a three-dimensional estimator. Since the equation is already in the form of we transform (7-63) in
(7-67), the step of the equivalence transformation in (7-64) can be skipped.
Clearly we have
-1 y=
O
5 !] where x = Px with P ~
output w(t) will be cho:
Let us choose arbitrarily the eigenvalues of the estimator as - 3 and - 3 ± 2j.
Then we have to find a L such that the eigenvalues of A22 - LA l2 are - 3 and
- 3 ± 2j. For this problem, we shall solve L by brute force. Let L' = [11 /2 /3J,
where the prime denotes the transpose. Then we have where F is a m x m m
problem is to design Sl
A ~ A22 -L~~12 = =~~ r -6 ~] that w(t) will approach I
7-9, we can show inaé i
_ L-/3 5 O
and det(51 - A) = 53 +/152 -(5 +/2)5 - (/3 + 5/ 1 )
By equating its coefficients with those of
and a11 eigenvalues of F
(5 +3)(5 +3 +2j)(5 +3 -2j)=S3 +95 2 +315 +39 matrix, then z(t) appro
we obtain to meet

/2 = - 36
Hence a three-dimensional state estimator is, following (7-72) and (7-74), then w(t) approaches k,

O] [ZI] [-45] [1]


1 Z2 + 240 y + O u
•o [I q oJ can be replaced by
O Z3 576 -2 developmenl, with slight m
CONNECTION OF STATE FEEDBACK AND STATE ESTIMATOR 369

lS a stochastic process, 1 O O

LO square error. The " 9 1 O

x= -36
r -84
O
O
Now we may apply the state feedback from
1
O
xas
mple 2 on page 338.
u= r + [~ Jf- \:1. -1f-Jx
This completes the designo I

u Functional estimators. If all the n state variables are to be reconstructed,


an estimator of dimension at least n - q is required. In many cases, however, it is
not necessary to reconstruct al1 the state variables; what is needed is to generate
sorne functions of the state, for example, to generate Kx in the state feedback.
In these cases, the dimensions of the estimators may be reduced considerably.
In the following, we discuss a method of designing an estimator to generate
kx, where k is an arbitrary 1 x n constant vector. Such an estimator is caBed a
Id -1 ±j.
functiona/ estimator or observer.
ust design a state esti­
Consider the dynamical equation in (7-63) with rank e =q. The problem
lte estimator to design
is to design an estimator so that its output wil\ approach kx as t -H:JJ. First
lready in the form of
we transform (7-63) into the form in (7-67), rewritten here for convenience,lO
7-64) can be skipped.

[:J = [i~: i~:][:J +[:J u ~ Ax +Bu


y = [Iq O]x ~ Cx
O] where x = Px with P given in (7-64). The functional estimator with a single
output w(t) will be chosen of the form
r as - 3 and - 3 ± 2j.
!2 - LA 12 are - 3 and z=Fz +Gy +Hu (1-87a)
w = Mz + Ny (1-81b)
. LetL' =[/1 /2 /3]'
where F is a m x m matrix; G, m x q; H, m x p; M, 1 x m and N,l x q. The
problem is to design such an estimator of a dimension as smal1 as possible so
that w(t) will approach kx = kP-1x =kx, where k = kP-1. Following Theorem
7-9, we can show that if
-5/¡) TA-FT=GC (7-88a)
H=TB (7-88b)

and all eigenvalues of F have negative real parts, where T is a m x n real constant
ls +39 matrix, then z(t) approaches Tx(t) as t~ oo. Hence, ir M and N are designed
to meet
(1-88c)

7-72) and (7-74), then w(t) approaches kx(t) as t~oo.

- ¿]u 'o [lq O] can be replaced by [C, O], where C, is.aq x q nonsingularmatrix, and the subsequent
-2 development, with slight modification, still holds. This su bsection follows closely Reference SIlO.
370 STATE FEEDBACK AND STATE ESTIMATORS

Now we partition T as T = [T¡ T 2]' where T ¡ and T 2 are, respectively, The mq x mq ma trix
m x q and m x (n - q) matrices. Then (7-88a) can be written as, using = e Theorem 2-4 that, fe
[I q O], if and only if the mq
T¡A¡¡ +T 2A 2¡ -FT¡ =G (7-89a)
T¡A¡2 +T 2A 22 - FT 2 =0 (7-89b)

lf we also partition k as k = [k¡ kz], where k¡ and k 2 are, respectively, 1 x q


and 1 x (n -q) vectors, then (7-88c) becomes
MT¡ +N=k¡ (7-89c) has rank n -q. No
and MT 2 =k 2 (7-89d) is observable. Furt
where v is the observ;
Now we choose F and M to be of the forms then V has rank n­

! 1 O Once T ¡ is known, v
O 1 N from (7-89a), (7-8:

JJ
estimator of kx. Th;
F= [
o o
-rx. m Theorem 7-14
M=[ O O O ] lf the dynamical eqt
then for any 1 x n r
where rx.i can be arbitrarily assigned. Note that {F, M} is observable. Let
functional estimator
tij,j = 1, 2, ... , m, be the jth row of Ti, i = 1, 2. Then because of the forms of F
approaches kx(t) exp
and M, (7-89d) and (7-89b) imply
t2 ¡ = k2 The dimension (lO
t ll A 12 +t 2¡A 22 =t 22
example, if n = 12 an
t 12 A 12 +t 22 A 22 =t 23 (7-90a)
index v could be ¡; =
then n - q = v - 1.
lf a p x n feedbac
where vis ap xl veCl
These equations, except the last, can be written as
7-2, then Kx can als.
t 2 ¡ =k 2 estimator. For a ge
t 22 =k 2A 22 +t¡¡Á¡2 eigenvalues of the fu
t 23 = k2Á~2 + t¡¡Á¡2 Á 22 + t¡2 Á ¡2 (7-9Ob) arbitrarily chosen. 1
lt mc~Y be possible í_(
In the design, if we ;
dimension of estimat
The substitution of these t 2j ,j = 1,2, ... ,m, into the last equation of (7-90a)
5131 and 5184.
yields, after sorne straightforward manipulation,

k~k2Á'í'2 +rx.¡k 2Á'í';¡ + ... +rx.m-¡k2Á22 +rx.m k2


*7-6 Decouplin
rx.~1
.I
Consider a p-input p­
= - [t¡m t¡.m-¡ t¡2 tll] rx.~1 rx.¡1 I
eq uation description
[
rx.1II_¡1
(7-91 )
The vector k is a 1 x (n - q) vector and is known oncek is giveri and rx. i are chosen. where u is the p x 1 iJ
DECOUPLING BY STATE FEEDBACK 371

Id T 2 are, respectively, The rnq x rnq matrix in (7-91) is nonsingular for any (Xi' Thus we conclude from
written as, using = e Theorem 2-4 that, for any ka solution [t 1m t 1,m-1 . .. t l l ] exists in (7-91)
if and only if the mq x (n - q) matrix
(7-89a)
(7-89b)
V= A12
A1Z:A 22 J
are, respectively, 1 x q
r Á1zAz21
has rank n -q. Now {A, C} in (7-63) is observable if and only if {A 22 , Á 12 }
(7-89c)
is observable. Furthermore, the observability index of {A zz , A1Z } is v - 1,
(7-89d)
where vis the observability index of {A, C} (Problem 7-27). Hence, if m = v - 1,
then V has rank n - q, and for any k and any (Xi, a solution T 1 exists in (7-91),
Once T 1 is known, we compute T z from (7-9üb) and then compute G, H, and
N from (7-89a), (7-88b), and (7-89c). The resulting estimator is a functional

I,J
estimator of kx. This is stated as a theorem.

Theorem 7 -14

o ] lf the dynamical equation in (7-63) is observable with observability index v,


then for any l x n real constant vector k, there exists a (v - l)-dimensional
vi} is observable. Let functional estimator with arbitrarily assignable eigenvalues so that its output
:cause of the forms of F approaches kx(t) exponentially. •

The dimension (v -1) could be much smaller than n -q for large q. For
example, if n = 12 and q = 4, then we have n - q = 8; whereas the observabili(y
(7-90a) 1;
index v could be =3 [see Equation (5-51)], and we have v -1 =2. If q = 1,
then n - q = v - 1.
lf a p x n feedback gain matrix K is of rank 1, it can be written as K = vk,
where v is a p x 1 vector and k is a 1 x n vector as the K z in the design in Figure
7-2, then Kx can also be generated by using a (v -1)-dimensional functiona1
estimator. For a general K, the situation is much more complicated. The
eigenvalues of the functional estimators discussed aboye are permitted to be
(7-9Ob) arhitrarily chosen. For a given set of eigenvalues, if they are chosen properly,
it may be possible to further reduce the dimension of a fl.lnctional estim;;üo;­
\.ZZ+tl. m-
In the design, if we augument the matrix C in (7-63b) as [e' K'T, then the
1Á 1z
dimension of estimators may also be reduced. See References S92, S97, Sl10,
ast equation of (7-9üa) Sl31 and S184.

*7-6 Decoupling by State Feedback

Consider a p-input p-output system with the linear, time-invariant, dynamical­


equation description
FE: x=Ax+Bu ·(7-92a)
(7-91 )
y=Cx' (7-92b)

; given and (Xi are chosen. where u is the p x 1 input vector, y is the p x 1 output vector; A, B, and e are

...... _
_._- _._-_.... .... • • __ . '.• _~'_'._ • • • _ •••• _ .~n
••
: :_
._. :.~_.::..:.
_ _::..:_.~ •• . . . . _.-~._.
• •_ - - - . __ - __
• • •••• -:- _

372 STATE FEEDBACK AND STATE ESTIMAlDRS

n X n, n X p, and p X n real constant matrices, respectively. It is assumed that The transfer function
p :::; n. The transfer function of the system is
G(s)=C(sI-A)-lB (7-93)
We shall derive in th,
Clearly G(s) is a p X P rational-function matrixo If the system is initially in can be decoupled by s
the zero state, its inputs and outputs are related by d¡ ~ min (the differen
Y1 (s) = 911 (S)U1 (s) + gds)uz(s) + o. o + glP(S)Up(s) of each entr
yz(s) = gZl (S)U1 (s) + gds)uz(s) + . o o + gzp(s)uP(S) and the 1 X P constan
(7-94)
. Yp(s) =gp1 (S)U1(S) + gpz (s)uz (s) + ... +gpp(s)up(s)
where gij is the ijth element of G(s). We see from (7-94) that every input controls where G¡(s) is the ¡th
more than one output and that every output is controlled by more than one
i,lPUt. Because of this phenomenon, which is called coupling or interacting, Example 1
it is generally very difficult to control a multivariable systemo For example,
suppose we wish to control Y1 (s) without affecting yAs), Y3(S), .. o, Yp(s); the Consider
required inputs U1(S), uz(s), .. o, u/s) cannot be readily foundo Therefore, in
sorne cases we like to introduce sorne compensator so that a coupled multi­
variable system may become decoupled in the sense that every input controls
only one output and every output is controlled by only one input. Conse­
quently, a decoupled system can be considered as consisting of a set of inde­
pendent single-variable systemso It is clear that if the transfer-function matrix
of a multivariable system is diagonal, then the system is decoupled. The differences in deg

Definition 7-1
A multivariable system is said to be decoupled if its transfer-function matrix is
The differences in deg
diagonal and nonsingular. I

In this section we shall study the problem of the decoupling ofmultivariable Ez = 1


s
systems by linear state feedback of the form
u(t)=Kx +Hr (7-95 ) Theorem 7 -15
where K is a p x n real constant matrix, H is a p >< p real constant nonsingular
matrix, and r denotes the p x 1 reference input. The state feedback is shown in feedback of the form u
Figure 7-10. Substituting (7-95) into (7-92), we obtain
FE!: x=(A+BK)x+BHr (7-96a)
y=Cx (7-96b)

is nonsingular.

H ..... ~~~ e We see from this 1


is a property of its tra
tion comes into playa
forethe controllabilit y
Fig u re 7 -10 Decoupling by state feedback_ ofthe system are imm¡
... -_----_ _._--_.. _.. ------ __._----- .- -- _- --._-- ------:==::::::::::=::::-._--­

DECOUPLlNG BY STATE FEEDBACK 373

Iy. It is assumed that The transfer function of the state-feedback system is


(;¡(s, K, H) ~ C(sI - A -BK)-IBH (7-97)
(7-93)
We shall derive in the following the condition on (;(s) under which the system
~ system is initiaHy in can be decoupled by state feedback. Define the nonnegative integer d¡ as
di ~ min (the difference ofthe degree in s ofthe denominator and the numerator
p(s)Up(s) of each entry of the ith row of (;(s))-I
p(s)up(s) and the 1 x p constant row vector E¡ as
(7-94)
pp(s)uP.(s) E¡ ~ lim Sd;+ l (;¡(s)
S-oo

at every input controls where (;¡(s) is the ith row of (;(s).


led by more than one
oupling or interacting, Example 1
¡ystem. For example,
(s), )\(s), . .. , yp(s); the Consider

~:~1
found. Therefore, in
that a coupled multi­
Lt every input controls (;(s) = [ " "+; +2]
Iy one input. Conse­
isting of a set of inde­ s2+~s+1 s2+ s +4
ansfer-function matrix
; decoupled. The differences in degree of the first row of (;(s) are l and 2; hence dI =0 and

El = lim s [
s'" 00 S2
s +2
+s +1 S2
l
+s +2
J= [1 OJ
sfer-function matrix is
• The differences in degree of the second row of (;(s) are 2 and 2, hence d 2 = 1 and

E2 = hm

s2 [ 1
~2----'---- 3J 11
lpling ofmultivariable s""oo s +2s +1

(7-95) Theorem 7 -1 5
A

i constant nonsingular A system. with the transfer-function rnairix G(s) can be decoupted by state
e feedback is shown in feedback of the form u = Kx + Hr if and only if the constant matrix

(7-96a)
(7-96b)

y
=>
is nonsingular.

We see from this theorem that wh~ther or not a system can be decoupled
is a property of its transfer-function matrix. The dynamical"equation descrip­
tion comes into playa role onlywhen thegain matrix K is to be found. There­
fore the controlIability and observability of the dynamical-equation description
of the system are immaterial here. Let (; ¡(s) and C¡ be the ith row of (;(s) and C,

.L
374 STATE FEEDBACK AND STATE ESTlMATORS

respectively. Then, fram (7-93) and using (2-3), we have and


G¡(s) =C¡(sl - A)-lB (7-98)

In the following, we shall establish the relations between the integer di, the
vector E¡, and the matrices C¡, A, and B. First, we expand (7-98) into (see
then
Prablem 2-39)
A 1 CBH=O, (
G¡(s) = Ll(s) Ci[sn-11+R1Sn- Z + ... + Rn-1]B and

1 Weestablish in the
= Ll( ) [CBsn- 1 + CR1Bsn- z + ... + CRd¡Bsn-d¡-l + ... + CRn-1B] E¡andE i.
s (7-99)
Theorem 7 -16
where
Ll(s) ~ det (si - A) ~ sn +CX1S n- 1 + ... +cx n For any K and any no
(7-100)

and R1=A+cx11 Proof


Rz=AR 1 +cxzI=A z +cx1A+cxzI
(7-101 )
It is easy to verify th
C¡Ad¡-lB=Oimply th:
R n- 1 = AR n- z +cxll-11 = An-1 +cx1A"- z + ... +cxn-11
C¡(,
Since di + 1 is the smallest difference in degree between the numerator and the
and C(A + BK)'
denominator of each entry of Gi(s), we conclude from (7-99) that the coefficients
associated with s"-d¡,sn-d¡+l, ... ,and sn-1 vanish, but the coefficient matrix Consequently, we havt
associated with sn-d¡-l is different fram zero. Hence we have
CiB=O, CiR1B=0, ... , C¡Rd¡_lB=O (7-102) and
and 11 Ei=CiRd¡B",O (7-103 )
Since H is nonsingulaJ
Fram (7-101) it is easy to see that conditions (7-102) imply that we conclude that di = (.
CiB=O, CiAB=O, ... , CAd¡-lB=O (7-104) With these prelimir
and Equation (7-103) becomes
Proof of Theorem 7-1
(7-105)
Necessity: Suppose th~
Therefore, di can also be defined from a dynamical equation as the sma!1est and nOllsingular The:
integer such that C¡Ad¡B "'0. If C¡AkB =0, for all k ::;n, then di ~ n -1. (See
Prablem 7-28.)
Similarly, we may ~define d¡ and E i for the transfer-function matrix G J in
(7-97). If we expand GJi into
... 1 - -
G Ji = ~(s) [C¡BHsn- 1 +CiR1BHs"- z + ... +CRII-1BH] (7-106) is a diagonal constant
is nonsingular by assu;
where
Sufficiency: If the mah
~(s) ~ det (si - A - BK) ~ Sil +~lSn-1 + ... +~n (7-107) Define

11 E¡ and C¡Rd¡B may dirrer by a constant because ó>(s) is factored out in (7-99). For convenience,
this diITerence, if ~ny, i~ neglected.
•. __ '
"_~'_W'_"._.'_' "~ •• J ••• •• _ _ •• _ _ , . ~ . •__ .~ __ ••• _ _ •• _"_ •• __ • __ ..
w_._~._.,, _··_"··~·_"'_ -_.. - . __.- .. _.~ '.'-­ ·...••••__ •. _.W __ ··_··_ _' .• ~ ,,_._ ~y . . . . _ •• _. __ ••••• __ • • • . • • • • • ~ " _ " _ •• - _ . _ . - - • • • •" ••• ~ .•••..• - - _ . . • ••• _ _ . _ . _. . -

DECOUPLING BY STATE FEEDBACK 375

and R1 =(A +BK) +&11


R2 =(A +BK)2 +&l(A +BK) +&21
(7-98) (7-108)
:en the integer di, the
.pand (7-98) into (see then
CBH=O, C(A +BK)BH=O, ... , C(A +BK)d¡-lBH=O (7-109)
and E =C(A +BKf¡BH feO (7-110)

We establish in the following the relationships between d¡ and Ji and between


l + ... +CR n- 1B] E i and E¡.
(7-99 )
Theorem 7 -16

For any K and any nonsingular H, we have di = d¡ and Ei = E¡H.

+0: 11
(7-100)

Proof
It is easy to verify that, for each i, the conditions C¡B = O, C¡AB = O, ..
(7-101 )
C¡Ad¡-l B = Oimply that
. +O:n-l 1
CitA + BK)k =CiAk k = 0,1,2, ... , di (7-111 )
he numerator and the and CitA +BK)k =CiAd¡(A +BK)k-d¡ k =d¡ +l,d¡ +2, ... (7-112)
19) that the coefficients
the coefficient matrix Consequently, we have
have C(A +BK)kBH =0 k =0,1, ... , di~ 1
(7-102) and CitA +BK)d¡BH=CiAd¡BH =E¡H
(7-103)
Since H is nonsingular by assumption, if E i is nonzero, so is E¡H. Therefore
Iy that we conclude that di = di and Ei = KH. Q.E. D.

=0 (7-104) With these preliminaries, we are ready to prove Theorem 7-15.

Proof of Theorem 7-1 5


(7-105)
Necessity: Suppose that there exists K and H such that G¡(s, K, H) is diagonal
lation as the smallest <md nonsingular. Then
then d¡ ~ n -1. (See

'unction matrix Gf in

is a diagonal constant matrix and nonsingular. Since E = EH and since H


is nonsingular by assumption, we conclude that E is nonsingular.
Sufficiency: Ir the matrix E is nonsingular, then the system can be decoupled.
Define

(7-113)
in (7-99). For convenience,
376 STATE FEEDBACK ANO STATE ESTlMATORS

We show that if K = - E - 1F and H = E - 1, then the system can be decoupled e¡, we immediately obta.
and the transfer function of the decoupled system is decoupled by using K =
l/S(d1 + 1) O

or, equivalently,
1 1
G¡(s, -E- F,E- )=
l 6O 1/s (d2

O
+ 1) Although a system c
the resulting system is n
system are at the origino
to move these poles to ti-
If the dynamical equ~
values by state feedback.
(7-114)
the number of eigenvall
where e¡ is a row vector with 1 in the ith place and zeros elsewhere. First we reduced. For a complete
show that C¡(A +BK)d¡+l =0. From (7-105), (7-112) and (7-113) and using compute K and H to have
K= -E- 1F and E¡E- 1 =e¡, we obtain 36,41, 112, S218, and S2:
CiCA +BK)d¡+l =C¡Ad'(A +BK)=C¡Ad¡+l +C¡Ad¡BK
=F¡ -E¡E- 1F = F¡ -e¡F =0
7-7 Concluding Rf
where F¡ and E¡ are the ith row ofF and E, respectively. Hence we conclude that
In this chapter we studi,
CiCA +BK)d¡+k =0 for any positive integer k (7-115)
observability. We showe
Since di =d¡, Equation (7-106) reduces to is controllable, we can, t
eigenvalues of the resultin
1 - 1­
C¡(sl - A - BK)-lBH ==- [CRd¡Bsn-d¡ - + C¡Rd¡+ 1Bs,,-d¡-2 + ... of controllable canonicaJ
. Ms) + Cít-1 BJH (7-116) state feedback, all the sta
asymptotic state estimatoJ
Now, from (7-105), (7-108), (7-112), (7-115), and the fact CiCA +BK)kB =0; for
observable, a state estima
k =0,1, ... , d¡ -1, which follows from (7-109) and the nonsingularity of H,
structed. The constructic
it is straightforward to verify that
servable canonical-form d)
Cj~.d¡B =C¡(A + BK)d¡B = C¡Ad¡B = E¡ TA-FT=GC.
CR d¡+ 1B= C[(A + BK)d¡ +l + !Y. 1(A+ BK)d¡JB = &l E ¡ We studied in this ch,
Section 7-2, 7-3, and 7-4 c,
equation is instantaneousl
trollability assumption, th
Consequently, Equation (7-116) becomes
dynamical equation has a fl
C¡(sl-A-BK)-lBH=L\~S)[sn-d¡-l +&lSn-d¡-2 + ... +&n-1-dJE¡H (7-117)
transformation, a controll
obtained. Consequently
remarks apply to the obser"
What is left to be shown is that
We also studied the der..
L\(s)=sn +éi 1s n- 1 +&2Sn-2 + ... +&n state feedback. The cone
= Sd¡ +l(sn-d¡-l + &¡sn-d¡- 2+ ... + &n-d¡- ¡) (7-118) required in this part of Stl
decoupling also used the st.
From the Cayley-Hamiltonthebrem, we have The combination of stat.
(A +BK)n +&¡(A +BKr 1 + ... +&nl =0 (7-119) in Section 9-5 directly fram t
the state-variable approad
By multiplying CiCA +BK)d¡ to(7-119) and using (7-115), we can obtain discussed there.
&nC¡(A+BK)d¡=O which implies that &n=O. Next by multiplying The introduction of con
CiCA + BK)d¡ -1, we can show that&" _1= O. Proceeding similarly, we can prove constant gain output feedbac
Equation (7-118). By substituting (7-118) into (7-117) and using E¡H = E¡E- 1 = cannot arbitrarily assign all
---:~---~--~~-----~------~. ------~----------.------------- - ~

CONCLUDING REMARKS 377

stem can be decoupled e¡, we immediately obtain Equation (7-114). Consequently, the system can be
decoupled by using K = - E- I F and H = E-l. Q.E.D.

Although a system can be decoupled by using K = - E - 1 F and H = E - 1,


the resulting system is not satisfactory because all the poles of the decoupled
system are at the origino However, we may introduce additional state feedback
to move these poles to the desired location.
If the dynamical equation is controllable, then we can control all the eigen­
values by state feedback. Now, if in addition we like to decouple the system,
(7-114) the number of eigenvalues that can be controlled by state feedback will be
·os elsewhere. First we reduced. For a complete discussion of this problem and others, such as how to
and (7-113) and using compute K and H to have a decoupled system with desired poles, see References
36,41, 112, S218, and S227.

7-7 Concluding Remarks


Hence we conclude that
In this chapter we studied the practical implications of controllability and
integer k (7-115) observability. We showed that if a linear time-invariant dynamical equation
is controllable, we can, by introducing state feedback, arbitrarily assign the
eigenvalues of the resulting dynamical equation. This was achieved by the use
BS n - d ¡-2 + ... of controllable canonical-form dynamical equations. When we introduce
state feedback, all the state variables must be available. If they are not, an
_C¡Rn-¡BJH (7-116)
asymptotic state estimator has to be constructed. If a dynamical equation is
lct C¡(A +-BK)kB =0, for observable, a state estimator with a set of arbitrary eigenvalues can be con­
he nonsingularity of H, structed. The construction of state estimators is achieved by the use of ob­
servable canonical-form dynamical equation or by the solution of the equation
TA-FT=GC.
E¡ We studied in this chapter only the time-invarlant case. The results in
BK)d'JB= rilK
Section 7-2, 7-3, and 7-4 can be extended to the linear time-varying case if the
equation is instantaneously controllable since, under the instantaneous con­
trollability assumption, the controllability matrix of the linear time-varying
dynamical equation has a full rank for all t. Hence by using it in the equivalence
transformation, a controllable canonical-form dynamical equation can be
obtained. Consequently the state feedback can be introduced. Similar
remarks apply to the observability part.
We also studied the decoupling of a transfer function matrix by introducing
state feedback. The concepts of controllability and observability are not
in
required in this part of study. It was discussed in this chapter because the
. + !XII-d¡-¡)
decoupling also used the state feedback as in Section 7-3.
The corilbination of state feedback and state estimator will again be studied
in Section 9-5 directly from the transfer-function matrix. Comparisons between
-riJ=O ­ (7-119) the state-variable approach and the transfer-function approach will also be
discussed there.
(7~1l5), we canobtain
The introduction of constant gain feedback from the output is called the
Next by multiplying
constant gain output feedback. Unlike the state feedback, the output feedback
ng similatly, we can prove
cannot arbitrarily assign all the eigenvalues of the resulting equation. Roughly
and usingE¡H=E¡E-¡ =
378 STATE FEEDBACK AND STATE ESTlMATORS

speaking, the number of eigenvalues which a constant gain output feedback


can assign is equal to min {n, p + q - 1}, where p = rank B and q = rank C. In 7-7 Consider the state fe
this assignment, no repeated eigenvalues are permitted. Furtherrnore, the Show that for any Constan
Under what condition on 1
assignment cannot be exact hut only arbitrarily close. See References S31,
and H are q x 11, P X 11, al1(
S71, S129, Sl30, and Sl31. See also Reference S165.
To conclude this chapter, we mention that al1 the results in this chapter that n '¿g.) (Answer: A solu
can he applied direct1y, without any modification, to the discrete-time case.
Consequently, a solutíon 1
We mention that the discrete-time state estimator should be modeled as singular.) .
x(k + 1) = Ax(k) + Bu(k) + L(y(k) - Cx(k))

In other words, the estimated x(k + 1) is reconstructed from {y(O), y(l), ... , y(k)}.
If x(k + 1) is reconstructed from {y(O), y(1), ... , y(k), y(k + 1)} and modeled as
x(k + 1) = Ax(k) +Bu(k) + L(y(k + 1) -Cx(k + 1))

then the situation wil1 be different. The reader is referred to Reference S215.
Figure P7-7

Problems
7-8 Let
7-1 The equivalence transformatíon Q ofi = Px = Q - IX in Theorem 7-1 can be obtained
either by using Q = [ql ql ... qn], where the q¡'s are given in (7-9), or by using Q =
UÜ- l , where U is the controllabílíty matrix of FE I and Ü- I is given in (7-13). From a
computatíonal point of view, which one is easier to use? Fínd the values of k = [k l I
Fínd the values of h such thé
7-2 Consider the controllable dynamical equatíon concIude that for almost all k

FE I : x=Ax +bu 7-9 Use state feedback to tra


y=cx 7-4to -1, -2,and -2. Drav
Ir the vectors {b, Ab, ... , An -1 b} are used as basis vectors, what is its equivalent dynamical and then add the requíred stat,
equation? Does the newc bear a simple relation to the coefficients ofthe transfer function
7-10 Show that
of FE I as it does in the controllable canonical form? (Hint: See Problem 2-15.)
p[b Ab .. " An-l
7-3 Find the matrix Q which transforms (7-17) into the observable canonical-form
dynamical equation. for any 1 x n constant vector k.

7-11
7-4 Transform the equatíon

if~ =~ ]x+[~}
.
x= [-~~
y= [1 1 O]x
into the controllable canonical-form dynamical equation. What is its transfer functíon?
whích has an unstable eígenvaJw
7-5 Let f.l.l be the largest integer such that {b[, Ab[, ... , A~l - Ib l } is a linearly independent subequation associated w¡'th .
seto Let f.l.l be the largest integer such that {b[, Ab[," .. , A~l -1 b[, b1 , Ab 1 , " .. , A~2 -1 b1 } st a bI"l" Ize the equatl'on by " elge
. usmg sto
is a Iinearly independent seto Show that Anb 1 for all n ¿f.l.l is linearly dependent on closed-Ioop equatíon has eigen~a':
{b[, ... , A~l -Ib[, b1 , ... , A~2 -·1 b1 }.
7~12 Gíven
7-6 Transform the dynamical equation in Problem 7-4 into the observable canonical­
form dynamical equation.
PROBLEMS 379

gain output feedback 7-7 Consider the state feedback and the output feedback systems shown in Figure P7-7.
B and q = rank C. In Show that for any constant matrix H, there exists a constant matrix K such that Kx = HCx.
ed. Furthermore, the Under what condition on C will there exist a matrix H such that K = HC for any K? (C, K
See References S31, and H are q x n, p x n, mul p x q constant matrices, respectively. It is general1y assumed

that n¿q.) (Answer: A solution H exists in K=HC for a given K if and only if p [~J=Pc.
results in tbis cbapter
tbe discrete-time case. Consequently, a solution H exists in K = HC for any K if and only if e is square and non­
singular.)
.Id be modeled as
Ix(k))
Ilom {y(O), y(l), ... , y(k)}.
I + 1)} and modeled as
I:x(k + 1))
\red to Reference S215. Figure P7-7

7-8 Let

[heorern 7-1 can be obtained A=[~ ~J b=[~J c=[1 O]


Ien in (7-9), or by using Q =
'1 is given in (7-13). From a Find the values of k = [k 1 kz] such that the matrix (A + bk) has repeated eigenvalues.
Find the values of h such that the matrix (A + bhc) has repeated eigenvalues. Can you
conclude that for almost al1 k and h, the two matrices have distinct eigenvalues?

7-9 Use state feedback to transfer the eigenvalues of the dynamical equation in Problem
7-4 to - 1, - 2, and - 2. Oraw a block diagram for the dynamical equation in Problem 7-4
and then add the required state feedback.
tat is its equivalent dynamical

\cien ts of the transfer function


7-10 Show that
!See Problem 2-15.)
p[b Ab .,. An-1b] =p[b (A - bk)b
I,e observable canonical-form for any 1 x n constant vector k.

7-11 Consider the lordan-form dynamical equation

1'lVhat is its transfer functíon? which has an unstable eigenvalue 1. The dynamical equation is not controllable, but the
subequation associated' with eigenvalue 1 is control1able. Do you think it is possible to
1, -1 b l } is a linearly independent stabilize the equation by using state feedback? Ir yes, find the gain vector k such that the
IA",-Ib¡, bz, Ab z, ... . '. A"' -1 b z} closed-loop equation has eigenvalues - 1, - 1, - 2, - 2, and - 2.
¿!lz is linearly dependent on
7-12 Given

lnto the observable canonical-


------.;=:::::-.--._.==:-.::__
.-_--.. ---=-=~.~-~.:::::.-.-:::::--:::::-.--:::::.---:..:..:. -:::::--::::::::::--:::::-
.. -:::::
. --::::::: : .- : : :-.-: : : : : "': : :"-:: : "-', : : ---:: : -- ------.---"'-'--'--'"--.---.----.- - - - - - - - - . - - . - . ------.---.---.-.--.----_.. '~"'--'" - - - 1

380 STATE FEEDBACK AND STATE ESTIMATüRS

Find the gain vector [k l k 2 ] such that the state-feedback system has - 1 and - 2 as its
7-21 Verify that the ma
eigenvalues. Compute k l , k 2 directly without using any equivalence transformation.
7-22 Let
7-13 Consider the uncontrollable state equation
1 O
2 O
O -1
O O
Is it possible to find a gain vector k such that the equation with state feedback u = kx + r Find two different real co
has eigenvalues -2, -2, -1, -1? Is it possible to have eigenvalues -2, -2, -2, -1? values - 4 ± 3j and - 5 ±
Howabout -2, -2, -2, -2? (Answers: Yes; yes; no.)
7-23 Show that the state
7-14 The observability of a dynamical equation is not invariant under any state feedback.
Where does the argument fail if the argument used in the proof of Theorem 7-3 is used to 7-24 Show that the (11 - <:¡
prove the observability part? 7-9.

7-15 Find a three-dimensional state estimator with eigenvalues - 2, - 2, and - 3 for the 7-25 Establish the sepan
dynamical equation in Problem 7-4. Use two different methods.
7-26 Consider the dynarr
7-16 Find a two-dimensional state estimator with eigenvalues - 2 and - 3 for the dynam­
ical equation in Problem 7-4. Use two different methods.

7-17 Consider a system with the transfer function


(s - l)(s +2)
(s + l)(s- 2)(s + 3)
ls it possible to change the transfer function to
y=
s-l
(s+2)(s+3) Find a full-dimensional es
estimator for kx(t) with
by state feedback? 1f yes, how?

7-18 In Problem 7-17, if the state is not available for feedback, can you carry out the
design by using a state estimator? Choose the eigenvalues ofthe estimator as -4 and - 4. 7-27 Prove Theorem 7-11.
What is its overall transfer function? observability índices of (A,­
l _~

7-19 Prove Theorem 7-11.


1
7-20 Consider the matrix in Problem 2-30. Show that the eigenvalues of the matrix are
Al, A2" .. , AII , - ¡, CI. ±jj3, ')'1' ')'20' .. , ')'112 _ 1 if aij are chosen as
"1- 1
and note that if linearly ind
811(S)=s"l+alllSn,-1 + ... +alln,=(s-IX)·TI (s-A¡)
i=l
order from top to bottom, tI­
"1 -1 ability matrices are identical.
821(s)=a21ISn,-I+ ... +a21I1,=-j3 TI (s-}ct)'
i=l
7-28 Consider the irreducib
"2 - 1
8 22 (s) =S"2 +a221Sn2 - 1 + ... +a22112 =(s -,- IX) II (5 - ')'¡) .
i
i=Ol
1
I
,
/12 -
where A is an n x n matrix, b i:
8ds)=aI21 Sn2 - 1 +... +aI2112=fi TI (s-y;) i function. Show that gis) ha~

.~.
i=l
PROBLEMS 381

:m has -1 and -2 as it, 7-21 Verify thatthe matrix in (7-50) has eigenvalues - 3 ± ~i, - 3 ±j and - 4 ± 2j.
i1ence transformation.
7-22 Let

A= [~ ¿ ~ ~l
-3 1 2 3
2100
~~ ~l
B= 1 2
02

Find two different real constant 2 x 4 matrices K such that the matrix (A + BK) has eigen­
1 state feedback u = kx + r
values -4±3j and -5 ±4j.
lValues -2, -2, -2, -1?
7-23 Show that the state estimator in Theorem 7-8 meets Theorem 7-9.
t under any state feedback.
7-24 Show that the (11 - q)-dimensional state estimator in (7-72) and (7-74)meets Theorem
¡ of Theorem 7-3 is used to
7-9.

7-25 Establish the separation property by using a full-dimensional state estimator.


:s - 2, - 2, and - 3 for the
1s. 7-26 Consider the dynamical eguation
_ 2 and - 3 for the dynam­
o 1 O
O O O
1 O 1
1 1 1
O 1 O
O O 1
O O O
Find a full-dimensional estimator, a reduced-dimensional cstimator, and a functional
estimator for kx(t) with
k=[O 2 -IJ
)ack, can you carry out the 7-27 Prove Theorem 7-11. Let Vi be the observabilíty indices of {A, C}. Show that the
he estimator as - 4 and - 4. observabilíty indices of {A n , A12 } are v¡-1. (Hint: Compute

:igenvalues of the matrix are -A l l

111 - 1
and note that if linearly independent rows of the observability matrices are searched in .
TI (s -A¡) order from top to bottom, the locations of the linearly independent rows of both observ­
¡= 1 ability matrices are identica!.)

7-28 Consider the irreducible single-variable dynamical eguation


/12 - 1
x=Ax +bu y=cx
TI (s-Y,)
i= 1
where A is an 11 x 11 matrix, bis an n.x 1 vector, ande is a 1 x 11 vector. Let gbe its transfer
function. Show that g(s) has m zeras-in other words, that the numerator of g(s) has
382 STATE FEEDBACK AND STATE ESTlMATORS

degree m-if and only if


Find a k so that (A + bk
for i=O, 1,2, ... , n-m-2
and cAn-m-1 b f= O. Or equivalently, the difference between the degree of the denominator
and the degree of the numerator of g(5) is J = n-m if and only if
cAd-1bf=O cAib=O fori=0,1,2, ... ,d-2 Hin!: Use the method or
(Hin!: Show that cAib is invariant under any equivalence transformation and then use the
controllable canonical-form dynamical equation.)

7-29 a. Can a multivariable system with the transfer function

~
52 +1

be decoupled by state feedback?


25 +1
53 +5 +1 "['J
b. Can a system with the dynamical-equation description

X~[~ -~} +[~ ~}


1
O
1 -1 O

y=[~ ~J x
-1
2
be decoupled by state feedback?

7-30 Consider the linear time-invariant discrete-time equation

-1 -2
x(k +1)= ~ -1
[ O
y(k) = [ 1 1

Find a feedback gain k so that all the eigenvalues of the resulting equation are zeros.
Show that, for any initial sta te, the zero-input response of the equation becomes identicany
zero for k "2.3. This is caBed a dead bea! control. See Problem 2-50.

7-31 Establish a counterpart of Theorem 7-9 for the discrete-time case.

7-32 Show that the controllability and observability indices of {A + BHC, B, C} are the
same as those of.{A, B, C} for any H. Consequently, the controllability and observability
indices are invariant under any constant gain output feedback.

7-33 Given
~::':=:-~:_:;'--=-==:":::="===;::_'_'_" ' "'__ ~'L _ _ ' _ •• _ . .~ •_ _• • • ._ . _•• _ ~ __ _ __ ~ • • • _

._-----_._-----------~-----,-_._._-~----~._~-----~-~.--,------"

PROBLEMS 383

Find a k so that (A + bk) is similar to


2 -1
degree ofthe denominator F= [ ~
if
.,d-2 Hint: Use the method on page 347. Choose k = [6 12 20J, then
formation and then use the 1 1
T= -3 -4 -51 ]
[
-1/2 -1/3 -1/4

ion

resulting equation are zeros.


e equation becomes identical1y
oblem 2-50.

ete-time case.

;es of {A + BRe, B, C} are the


ontrollability and observability
·ack.
STABIl

8-2 Stabilityel
Descri ption

Time-varying case
systems with only on
shown in Chapter 3 t

8 satisfy the homogenei


then the input u and t

y(t) =
Stability of Linear Systems
where g(t, T) is the im
output measured at tir
In the qualitative ~
perhaps the only gues
properties, under wha
For example, if the inp

\u(t
under what condition e
output y satisfies

8-1 Introduction
for allt in ( - 00, oo)?
Controllability and observability, introduced in Chapter 5, are two important
gualitative properties of linear systems. In this chapter we shall introduce
another qualitative property of systems-namely, stability. The concept of
stability is extremely important, because almost every workable system is
designed to be stable. Ir a system is not stable, it is usually ofno use in practice. does there exist a consí.
We have introduced the input-output description and the dynamical-equation
description of systems; hence it is natural to study stability in terms of these
two descriptions separately. In Section 8-2 the bounded-input-bounded­
output (BIBO) stability of systems is introduced in terms of the input-output If the bpul approache
description. We show that if the impulse response of a system is absolutely output approach anotl
integrable, then the system is BIBO stable. The stability condition in terms of input approaches a cons
rational transfer functions is also given in this section. In Section 8-3 we ing to these various pro
introduce the Routh-Hurwitz criterion which can be used to check whether or for the system. We sha;
not all the roots of a polynomial have negative real parts. We study in Section used in linear systems: t
8-4 the stability ofa system in terms ofthe state-variable description. We intro­ Recall that the inpl
duce there the concepts of equilibrium state, stability in the sense of Lyapunov. when the system is initia
asymptotic stability, and total stability. The relationships between these relaxed system. Hence t
concep.ts.are established. In Section 8-5, we introduce a Lyapunov theorem description is applicable
and then use it to establish the Routh-Hurwitz criterion. In the last section,
the stability of linear discrete-time systems is discussed. Definition 8-1
The references for this chapter are 58; 59, 65, 76, 90, 102, 116, S79, S 111,
S135, and S153. A relaxed system is said
if and only if for any bOL
384
J
STABILITY CRITERIA IN TERMS OF THE INPUT-OUTPUT DESCRIPTION 385

8-2 Stability Criteria in Terms of the Input-Output


Description

Time-varying case. We conslder first single-variable systems, that is,


systems with only one input terminal and only one output terminal. It was
shown in Chapter 3 that if the input and output of an initially relaxed system
satisfy the homogeneity and additivity properties, and if the system is causal,
then the input u and the output y of the system can be related by

y(t) = Loo g(t, "C)U("C) d, for aH t in ( - co, oc!) (8-1)

s where g(t, "C) is the impulse response of the system and is, by definition, the
output measured at time t due to an impulse function input applied at time "c.
In the qualitative study of a system from the input and output terminals,
perhaps the only question which can be asked is that if the input has certain
properties, under what condition will the output have the same properties?
For example, if the input u is bounded, that is,

for all t in (- co, co) (8-2)

under what condition on the system does there exist a constant kz such that the
output y satisfies
ly(t)I.:::; kz < OC!

for aH t in ( - co, oc!)? Ir the input is of finite energy, that is,


,ter 5, are two important
lpter we shall introduce (8-3)
ability. The concept of
'ery workable system is does there exist a constant k 4 such that
lal1y ofno use in practice.
d the dynamical-equation
tability in terms of these
lounded-input-bounded­
erms of the input-output lf the input approaches a periodic function . under Wh3.t concilic);} \vill the
of a system is absolutely output approach another periodic function with the same period? If the
.litY condition in terms of input approaches a constant, will the output approach sorne constant? Accord­
lion. In Section 8-3 we ing to these various properties, we may introduce different stability definitions
used to check whether or for the system. We shall introduce here only the one which is most commonly
lrts. We study in Section used in linear systems: the bounded-input-bounded-output stability.
,le description. We intro­ Recall that the input-output descl'iption of a system is applicable only
in the sense of Lyapuno v ,
when the system is initially relaxed. A system that is initially l'elaxed is called a
ltionships between these
relaxed system. Hence the stability that is defined in terms of the input-output
uce a Lyapunov theorem
description is applicable only to relaxed systems.
rion. In the last section,

Definitíon 8.-1.
:d.

" 90, 102, 116, S79, Slll,


A relaxed system is said to be BIBO (bounded-input-bounded-output) stable
if and only ir fol' any bounded input, the output is bounded. •
386 STABlLITY OF LINEAR SYSTEMS STAB!

We illustrate the importance of the qualification "relaxedness" by showing output. Let us choo
that a system, being BIBü stable under the relaxedness assumption, may not
be BIBü stable if it is not initially relaxed.
where
Example 1
Consider the network shown in Figure 8-1. lf the system is initially relaxed,
that is, the initial voltage across the capacitor is zero, then y(t) = u(t)/2 for all t.
Therefore, for any bounded input, the output is also bounded. However, if Clearly u is bounded.
the ·initial voltage across the capacitor is not zero, because of the negative
capacitance, the output will increase to infinity even if no input is applied. I

Theorem 8-1 is not bounded.


A relaxed single-variable system that is described by
We consider now
terminals and q outPl
y(t) = Loo g(t, r)u(r) dr
is BIBü stable if and only if there exists a finite number k such that
where u is the p x 1 in
Loo Ig(t, r)\ dr =:;;,k< 00 q x p impulse-responso
for all t in (- 00,00).

Proof G(t,

Sujficiency: Let u be an arbitrary input and let lu(t)1 =:;;, k l for all t in ( - 00, 00).
Then
Then 9ii is the impul~
ly(t)1 = \ Loo g(t, r)u(r) dr I=:;;, Loo Ig(t, r)llu(r)i dr =:;;, k l Loolg(t, r)1 dr =:;;, kk l
output terminal. Sim
system is defined to be
for all t in ( - 00, 00). Necessity: A rigorous proof of this part is rather involved. the output vector is ba
ponent of the vector
We shall exhibit the basic idea by showing that if
pair of input and outp
number of bounded fu
Loo Ig(t, r)1 dr = 00
theorem.
for sorne t, say tI' then we can find a bounded input that excites an unbounded
Theorem 8-2

A relaxed multivariable

In In
¡­y

~
+
u"-'
-i F
is BIBü stable ir and e
In
entry of G,
In

Figure 8-1 A system whose output is bounded for any bounded input if the initial
voltage across the capacitor is zero. for all t in ( - 00, (0).
STABILlTY CRITERIA IN TERMS Of THE INPUT-OUTPUT DESCRIPTION 387

axedness" by showing output. Let us choose


; assumption, may not u(t) =sgn [g(t 1 , t)] (8-4 )
where

={ ~
if x =0
sgn x if x>O
em is initially relaxed, -1 if x <O
en y(t) =u(t)/2 for all t.
Clearly u is bounded. However, the output excited by this input,
Jounded. However, if
ecause of the negative
o input is applied. •
y(t 1 ) = Loo g(t 1 , r)u(r) dr = Loo \g(t 1, r)1 dr = 00

is oot bouoded. Q.E.D.

We consider now multivariable systems. Consider a system with p input


terminals and q output terminals, which is described by

y(t)= foo G(t, r)u(r)dr (8-5 )


r k such that
where u is the p x 1 input vector, y is the q x 1 output vector, and G(t, r) is the
q x p impulse-response matrix of the system. Let

9l1(t,r) glZ(t,r) glP(t,r)l


G(t, r) = gZl~t, r) gzz\t, r) gzp\t, r) . (8-6)

k 1 for all t in (-00,00). 1 gql(t, r) gdt, r) gqp(t, r)

Then gij is the impulse response between the jth input terminal and the ith
output terminal. Similar to single-variable systems, a relaxed multivariable
l [00 Ig(t, r)i dr ~ kk 1 system is defined to be BIBO stable if and only if for any bounded-input vector,
the output vector is bounded. By a bounded vector, we mean that every com­
lis part is rather involved. ponent of the vector is bounded. Applying Theorem 8-1 to every possible
pair of input and output terminals, and using the fact that the sum of a finite
number of bounded functions is bounded, we have immediately the following
theorem.

)at excites an unbounded Theorem 8-2


A relaxed multivariable system that is described by

y(t) = foo G(t, r)u(r) dr


is BIBO stable if and only if there exists a finite number k such that, for every
entry of G,

bounded input if the inítia\


for all t in ( - 00, co).

--_ .. ..- -- .. _-----­
_-----------~------_._-------_
----------_
-.---~---.-_._-_

._-­

388 STABlLlTY Of LINEAR SYSTEMS STABIL

Time-invariant case. Consider a single-variable system with the following


input-output description:

y(t) = J>(t-L)U(L)dL= Ig(L)U(t-L)dL (8-7)

where g(t) is the impulse response of the system. Recall that in order to have a
description of the form (8-7), the input-output pairs of the system must satisfy
linearity, causality, and time-invariance properties. In addition, the system is
assumed to be relaxed at t = O.
234
Corollary 8-1
A relaxed single-variable system which is described by

y(t) = I g(t - L)U(L) dL


Figure 8-2 An absolul
zero as t -> oo.

is BIBü stable if and only if


If gis absolutely ir
tOO [g(t)\ dt sk < 00

for sorne constant k.


as o: -> oo. We shall u:
Proof
This follows directly from Theorem 8-1 by observing that
Theoreril 8-3

I Ig(t, L)I do = I Ig(t - L)I dL = I Ig(o:)1 do: s 100 Ig(o:)1 do: (8-8) Consider a relaxed si!
related by
For the time-invariant case, the initial time is chosen at = O; hence the inte­
°
gration in (8-8) starts [rom instead of - oo.
t
Q.E.D.

A function 9 is said to be absolutely integrable on [0, (0) if

rOO Ig(t}1 dt sk < 00


Jo I Since
Graphically, it says that the total area under Igl is finite. The fact that 9 is
absolutely integrable does not imply that 9 is bounded on [0,(0) nor that g(t)
approaches zero as t -> oo. Indeed, consider the function defined by and since

4
-1
n +(. t - n)n
for - 3 < (t -n)sO
f(t-n)= .
n

{ is nondecreasing as ex increa
1
n -(t - n)n 4 forOs(t-n)s3
n
for n = 2,3,4, . . .. The function .r
is depicted in Figure 8-2. lt is easy to Consequently,
verify that f is absolutely integrable; however, f is neither bounded on [0, (0)
nor approaching zero ast -> oo.
STABILlTY CRITERIA IN TERMS OF THE INPUT-OUTPUT DESCRIPTION 389

tem with the following ,


I

Id, (8-7)
n
that in order to have a
he system must satisfy
addition, the system is
II
2 3 4

Figure 8-2 An absolutely integrable function that is neither bounded nor tending to
zero as (---+00.

If g is absolutely integrable on [0, (0), then 1

(8-9)

as a ----+ oo. We shall use this faet in the proof of the following theorem.

at Theorem 8-3
Consider a relaxed single-variable system whose input u and output y are
100 Ig(a)1 da (8-8)
related by
it t = O; henee the inte­
Q.E.D. y(t) = J~ g(t - ,)u(,) d,

0, (0) if

l Since

Lite. The faet that g is


l on [0,(0) nor that g(t)
ro
\g(/)1 dI = r r
o
Ig(/)1 dI +
a

Ig(/)1 dI

)ll defined by and since

-n)~O I Ig(t)1 d(

is nondecreasing as fY. increases,

liro
a- 00
Jao Ig(/)\ dt ..... J.'o" 19(1)1 dt
gure 8-2. ltis easy to Conseq uen tI y.
ither bounded on [0, (0)
390 STABILITY OF LINEAR SYSTEMS STABIl

If 2. We prove this par

[ ' [g(t)1 dt =::; k < 00 y

for sorne constant k, then we have the following: Let UM ~ rnax lu(t)
I

1. If u is a periodic function with period T -that is, u(t) = u(t + T) for all
t ~O-then the output y tends to a periodic function with the sarne period
(not necessarily of the sarne waveforrn).
Ir g(r)u(

2. If u is bounded and tends to a constant, then the output will tend to a constant. which appraaches .
3. Z If u is of finite energy, that is, ciently large, (8-12)

(rO \u(t)lz dt}/Z =::;k l < 00

then the output is also of finite energy; that is, there exists a finite kz that
for all t ~ t l. As car:
depends on k l such that
is approxirnately eq
becornes, for all t ~
(IX> [y(t)i2 dt}/z =::; kz < 00

Proof
1. We shall show that if u(t) = u(t + T) for all t ~O, then y(t)-4 y(t + T) as t -4 oo. which is independen
It is c1ear that 3. It can be shown tha
with the property
y(t) = J>(r)u(t - r) dr (8-10)

and
r + l T
r,+ T for sorne constant k
y(t+T)= Jo g(r)u(t+T-r)dr= Jo g(r)u(t-r)dr (8-11 ) space. In this space
(see Prablern 2-49):
Subtracting (8-10) frarn (8-11), we obtain
IIJII ~ (
Ir+
T
¡y(t +T) - y(t)1 = g(r)u(t -r)drl
The Schwartz inequa
r'+T rl + T
=::; J, ig(r)llu(t -r)\ dr =::;UM J. Ig(-r)\ dr
~ 1"' ü'-'

11 J(t)gi
where UM ~ rnax lu(t)l. It follows frarn (8-9) that Iy(t) - y(t + T)i-4 O as
OSIST
t-4 00, or y(t)-4 y(t + T) as t -4 oo.

2 This can be eXlended lo as fol1ows: For any real number p in [1. co]. if . ,

(J '"
g (r)
)I/P
o lu{t)lpdt ::;;k¡<co lJl -¡-\- .,..-__
then there exists a finite k 2 .such that

(r [y(t)!P dt}'P ::;;k 2 < co


. tI

and the system is said to be Lp-stable. Figure 8--3: The convolut


STABILITY CRITERIA IN TERMS OF THE INPUT-OUTPUT DESCRIPTION 391

2. We prove this part by an intuitive argumento Consider

y(t) =
1',
o
g('1:)u(t - '1:) d'1: + J' g('1:)u(t - '1:) d'1:
'1

(8-12 )

Let UM ~ max
, Iu(t)l· Then we have, with t > t b

u(t) = u(t + T) for aH


with the same period
\rg('1:)u(t -'1:)d'1: \ s r Ig('1:)llu(t -'1:)1 d'1: SUM

which approaches zero as t 1 -'> 00, following from (8-9). Hence if t 1 is suffi.­
r Ig('1:)\ d'1:

wil1 tend to a constant. ciently large, (8-12) can be approximated by

y(t) =i= r g('1:)u(t - '1:) d'1: (8-13 )

: exists a finite k'}. that for aH t 2: t ,. As can be seen from Figure 8-3, if t is m uch larger than t l, u(t - '1:)
is approximately equal to a constant, say (f., for aH '1: in [0, t ,]. Hence (8-13)
becomes, for all t ~ t 1 ~ 0,

y(t) =i= (f.

which is independent of t. This completes the proof.


r g('1:) d'1:

y(t)--> y(t + T) as t --> oo.


3. It can be shown that the set of aH real-valued functions defined over [O, (0)
with the property
(8-10)
(ro If(t)i2 dt) l/'}. sk < ro .

for some constant k forms a linear space over IR, It is called the L'}. function
space. In this space we may define a norm and an inner product as follows
)u(t - -c) d'1: (8-11 )
(see Problem 2-49):

Ilfll ~ ([' If(t)i2 dt),l'). (1, g) {¿ ro f(t)g(t) dt

The Schwartz inequality (Theorem 2-14) reads as

r)\ d'1:
I[} f(t)g(t) dt Is (t
dl

[fU)\2 dt) 1/2 (I" Ig(t)i2 dtYí


2

t Iy(t) - y(t + T)\ --> o as

]. ir u (t - T)

tY:+-:~-....,....----------
---f-::--~----;I'''''''''--t------~-------I-_T

Figure 8-3~ The convolution of g and u.

i
392 STABlLITY Of LINEAR SYSTEMS STABlI

With these preliminaries, we are ready to proceed with the proof. Consider lf

¡y(OI = I f~ g(r)u(t - r) dr \ ~ [ ' ig(r)llu(t - r)1 dr

which can be written as for sorne k and if u(t)

ly(t}1 ~ too (ig(r)11/2)(ig(r)/1/2Iu(t - rm dr (8-14 )


where () = tan -1 (lm ¿
Applying the Schwartz inequality, (8-14) becomes and 1m denote the re~

Proof
Since sin (t - r) = sin t
Consider now

too ly(t)i2 dt ~k too too Ig(r)llu(t -r)i2 dr dt

y(t) = I g(r)u(t-r)dr

= sin wt too g(r) ce


= k tOO (too lu(t - r)i2 dt) Ig(r)l dr (8-15 )
1t is clear that
where we have changed the order of integration. By assumption, u is of
11
00

fini te energy; hence there exists a finite k 1 such that g(r)sinw(

too.1u(t - r)12 dt ~k 1
Hence, from (8-9) we (

Hence (8-15) implies that y(t)->sin w

~kk1
By definition, g(s) is th
too ly(t)12 dt too Ig(r)1 dr =k 2k 1

In other words, if g is absolutely integrable on [O, 00) and if u is of finite


energy, then the output y is also of finite energy. Q.E. O. which implies that

We see from Corollary 8-1 and Theorem 8-3 that if a system that is describ­ g(iw) = roo g(¡
able by Jo

y(t) = f>(t - r)u(r) dr


Since g(t) is real-valuec

is BIBO stable, then if the input has certain property, the output will have the
same property. For the time-varying case, this is not necessarily true. The
following is a very important corollary of Theorem 8-3.
Hence (8-17) becomes
Corollary 8-3 y(t) -> sin wt (R
Consider a relaxed single-variable system whose input u and output y are where () = tan -1 (1m g(i,
related by

y(t) = f>(t - r)u(r) dr


This corol1ary sho\\­
system, ifthe input is a si
STABILlTY CRlTERIA IN TERMS Of THE INPUT-OUTPUT DESCRIPTION 393

1 the proof. Consider If

-,JI eh IX> [g(t)1 dt ::; k < 00

for sorne k and if u(t) = sin wt, for I ::::: 0, then


y(t)---> \g(iw)\ sin (WI +8) as t ---> 00 (8-16)
)d, (8-14)
where 8 = tan -1 (1m g(iw)(Re g(iw)) and g(s) is the Laplace transform of g('); Re
and 1m denote the real part and the imaginary part, respectively.

.",
Proof
ig('r)l[u(t _,)1 2 d,
o Since sin (t - ,) = sin t cos ¡; - cos I sin ¡;, we have

y(t) = t g(,)u(t - ¡;) d¡; = I g(,)[ sin wt cos w¡; - cos WI sin w¡;] d¡;
d¡; dI
=sinwt to g(,)cosw,d¡;-coswt t'" g(¡;)sinw¡;d¡;- f" g(¡;)sinw(t-¡;)d-c
(8-15)
It is clear that

ly assumption, u is of Ir g(,) sin w(t - ,) d¡; \::; r [g(¡;)[[sin w(t - ')1 d,::; r Ig(,)1 d¡;

Hence, from (8-9) we concl ude that as t ---> 00, we obtain

y(t)---> ~in wt t'" g(,) cos w, d¡; - COS WI t'" g(,) sin w'"[(h (8-17)

By definition, g(s) is the Laplace transform of g('); that is,

'fJ) and if u is of finite


Q.E.D. which implies that

system that is describ­ g(iw)= r'" g(t)e-


Jo
iwr
dt = f'" g(t) cos wt lit - i r'" g(t) sin
Jo JO
wl di

:le output will have the


Since g(t) is real-valued, we have
Re g(iw) = r g(t)coswtdt (8-18a)

necessarily true. The 1m g(iw) = - 10'" g(t)sinwtdt (8-18b)

Hence (8-17) becomes


y(t)---> sin wt (Re gUro)) +cos wt (1m g(iro)) = Ig(iw)\ sin (wt +8)
lt u and output y are where 8 = tan - 1 (Img(iw)(Re g(iro)). Q.E.D.

This corollary shows that for a BIBO-stable, .linear time-invariant relaxed


system, ifthe input is a sinusoidal function, after the transient dies out, the output
394 STABILITY Of LINEAR SYSTEMS

is also a sinusoidal function. Furthermore, from this sinusoidal output, the


is BIBO stable if and
magnitude and phase of the transfer function at that frequency can be read out entry of G,
directly. This fact is often used in practice to measure the transfer function of a
linear time-invariant relaxed system.
Linear time-invariant systems are often described by transfer functions;
hence it is useful to study the stability conditions in terms of transfer functions. If the Laplace tra
If g(s) is a proper rational function of s, the stability condition can be easily function matrix then
stated in terms of g(s). If q(s) is not a rational function, then the situation is ofC(s). '
much more complicated; see Reference S32.
Theorem 8-5
Theorem 8-4 A relaxed multivariab
A relaxed single-variable system that is described by a proper rational function is a proper rational-fu
9(S) is BIBO stable if and only if all the poles of 9(S) are in the open left-half of every entry of C(s) 1
s plane or, equivalently, all the poles of 9(S) have negative real parts. •
Corollary 8-2 and
By the open left-half s plane, we mean the left-half s plane exc1uding the and 8-4 if we consideJ
imaginaryaxis. On the other hand, the c10sed left-half s plane is the left-half input-output pairo
s plane inc1uding the imaginary axis. Note that stabilíty of a system is inde­
pendent of the zeros of 9(S). 8-3 Routh-Hurv
Proof of Theorem 8-4 If the transfer functior
If g(s) is a proper rational function, it can be expanded by partial fraction stability of the system
expansion into a sum of finite number of terms of the form is irreducible, that is, ti
·nator and numerator, t
f3 nator of g(s). Hence I
(s-Af system is determined b
and possibly of a constant, where A¡ is a pole of 9(s). Consequently, g(t) is a polynomial is called a
sum of finite number of the term t k - ) eA,! and possibly of a 6-function. lt is parts. Hence a systen
easy to show that t k - ) eA,! is absolutely integrable if and only if A¡ has a negative irreducible transfer fun
real part. Hence we conc1ude that the relaxed system is BIBO stable if and a polynomial is Hurwi
only if all the poles of g(s) have negative real parts. Q.E.D. are computed. Howev
computa tia n of the roo
Example 2 the exact locations of
Therefore it is desirab
Consider the system with a transfer function g(s) = l/s. The pole of 9(S) is on without solving for thl
the imaginary axis. Hence the system is not BIBO stable. This can also be method: the Routh-Hu
shown from the definition. Let the input be a unit step function, then u(s) = l/s. Consider the polyn(
Corresponding to this bounded input, the output is fE -) [g(s)u(s)] =
fE -) [1/s 2 ] = t, which is not bounded. Hence the system is not BIBü stable. • D(s) =aos n +
where the a;'s are real nu
For multivariable systems, we have the following results. we shall give some nec
Hurwitz polynomial-t
Corollary 8-2
then D(s) can be factore
A relaxed multivariable .system that is c;lescribed .by
D(s) =c
y(t) = LG(t - .)u(r) d. =(,
ROUTH-HURWITZ CRITERION 395

inusoidal output, the is BIBO stable if and only if there exists a finite number k such that, for every
ilency can be read out entry of G,
: transfer function of a
[' \gJt)\ dt .:s;k< 00
'Y transfer functions;
; of transfer functions. If the Laplace transform of G(t), denoted by G(s), is a proper rational­
mdition can be easily functíon matrix, then the BIBO stability condition can also be stated in terms
, then the situation is of G(s).

Theorem 8-5
A relaxed multivariable system that is described by y(s) = G(s)u(s), where G(s)
oper rational function is a proper rational-function matrix, is BIBO stable if and only if all'the poles
e in the open left-half of every entry of G(s) have negative real parts. I
~ real parts. •
Corollary 8-2 and Theorem 8-S follow immediately from Theorems 8-2
s plane excluding the and 8-4 if we consider every entry of G as the impulse response of a certain
input-output pair.
s plane is the left-half
:y of a system is inde­
8-3 Routh-Hurwitz Criterion

If the transfer function of a system is a rational function of s, then the BIBO


ed by partíal fraction stability of the system is completely determined by the poles of g(s). If g(s)
is irreducible, that is, there is no nontrivial common factor between its denomi­
:m
nator and numerator, then the poles of g(s) are equal to the roots of the denomi­
nator of g(s). Hence under the irreducibility of g(s), the BIBO stability of a
system is determined by the roots of a polynomial, the denominator of g(s). A
::::onsequently, g(t) is a polynomial is called a Hurwitz polynomial if all of its roots have negative real
of a b-functíon. It is parts. Hence a system is BIBO stable if and only if the denominator of its
mly if A¡ has a negative irreducible transfer function is a Hurwitz polynomial (Theorem 8-4). Whether
is BIBO stable if and a polynomial is Hurwitz or not can be readily determined once all of its roots
Q.E.D. are computed. However, if the degree of the polynomial is three or higher, the
computation of the roots is not a simple task. Furthermore, the knowledge of
the exact locations of the roots is not needed in determining the stability.
Therefore it is desirable to have sorne method of determining the stability
The pole of g(s) is on without solving for the roots. In this section we shall introduce one such
.ble. This can also be method: the Routh-Hurwitz criterion.
mctíon, then u(s) = l/s. Consider the polynomial
t is ;t' - 1 [g(s )u(s)] =
ao>O (S-19)
is not BIBO stable. •
where the a¡'s are real numbers. Before developing the Routh-Hurwitz criterion,
.ults. we shall give sorne necessary condition for D(s) to be Hurwitz. If D(s) is a
Hurwitz polynomial-that is, if all the roots of D(s) have negative real parts-
then D(s) can be factored as ..
D(s) =ao n (s +ctdn(s +f3j +iy)(s +f3j - iy) ..
k j

= ao n
k
(s +ctd n
j
(S2 + 2f3 jS + f3; +yJ) (S-20)
_. _.. . .._ ... __, . . _ .. _.J_ ..__. .. .__ . .. . . . .

~
~
.
~
.
,
.
~
.
~
_
~
_
.
_
.
.
.
,
.
.
.
-
~
-
-
.
~
_
=
-
-
-
-
.
-
-
-
~
-
-
_
_
=
:
:
:
:
.
:
_
.
:
:
_
.
.
.
.
:
=
=
_
=
:
=
=
=
'
~
=
:
:
:
:
:
:
_
_
:
:
-
-
-
'
:
.
:
-
-
396 STABILITY Of LINEAR SYSTEMS

where Cf.k > 0, fJ j > Ú, and iZ = - 1. Since aH the coefficients of the factors in
the right-hand side of (S-20) are positive, we conclude that if D(s) is a H urwitz
polynomial, its coefficients ai> i = 1, 2, ... ,n must be al! positive. Hence given a
polynomial with a positive leading coefficient, if some oIits coefficients are negative
or zero then the polynomial is not a Hurwitz polynomial. The condition that aH
coefficients of a polynomial be positive is only a necessary condition for the
polynomial to be Hurwitz. A polynomial with positive coefficients may still Since the four numbe
not be a Hurwitz polynomial; for example, the polynomial with positive co­ conclude from Theoren
efficients .
In order to ohtain o
S3 + SZ + lis + 51 = (s + 3)(s - 1 +4i)(s -1 - 4i) These processes can b{
is not a Hurwitz polynomial. an even number and
Consider the polynomial D(s) given in (S-19). We form the fol!owing with n' ~ n/2,
polynomials
Do(s) =aos n +azsn - Z + ... (S-21a)
D1(s) = als n - 1 +a3s"- 3 + o •• (S-21b)
Note that the number
that is, if n is even, Do(s) consists of the even part of D(s) and D1(s) consists of is even; they are equa
the odd part of D(s); if n is odd, Do(s) consists of the odd part of D(s) and D1(s) shown in Table S-1.
consists of the even part of D(s). Observe that the degree of Do(s) is always one two rows are just the
degree higher than that of D1(s). Now we expand Do(s)/D1(s) in the foHowing remaining rows are 01
Stieljes continued fraetion expansion:

1
1
Cf.ZS +-------------­
1 (S-22)
Cf.3 S +---------­
Table S.1 Continuous-

+---­ s" a~O) a)O)

s
11-1
a~l a\l)
-----~---

Sll- 2 a\il a\2)


Theorem 8-6
The polynomial D(s) in (S-19) is a Hurwitz polynomial if and only if the n Sn-3 ab"l a\3)
numbers Cf.l> Cf.z, ••• , Cf. n in (S-22) are all positive. •

This theorem will be proved in Section S-5. Here we give an example to a\j'-2l a·1(II-2)t,
.
illustrate its application.
Si a\j'-ll:
Exarilple 1
-'--------:
SO a~l:
,
Consider D(s) = S4 + 2s + 6s z + 4s + 1.
3 We have Do(s) = S4 + 6s" + 1, D1(s) =
2s 3 +4s,and t Ir n is odd, the first two ro
the pattern is identical.
ROUTH-HURWITZ CRITERION 397
ients of the factors in
at if D(s) is a Hurwitz Do(s) 1 1 ¡
--=-s+ 3 ="2s+----­
)sitive. Hence given a D¡(s) 2 2s +4s . 1
oejJicients are negative
ts+--- (8-23)
4s 2 + 1 .B. 1
The condition that al! 7 S +"7
"2s
;ary condition for the
: coefficients may still Since the four numbers 1X¡=~,1X2=~'(.(3=~' and 1X 4 are all positive, we =f
nial with positive co­ conclude from Theorem 8-6 that the polynomial D(s) is a Hurwitz polynomial. I

In order to ohtain IX ¡, 1X 2 , ... ,IX", a series oflong division must be performed.


l-4i) These processes can be carried out in tabular formo First we assume that n is
an even number and rename the coefficients of Do(s) and D¡(s) in (8-21) as,
¡: form the fol!owing with n' ~ n/2,
Do(s) = abO)s" +a~Ols"- 2 + ... +a~9~ ¡S2 +a~9)
(8-24 )
(8-21a) D¡(s)=a~)s"-¡ +dl ls"-3 + ... +a~~~¡s
(8-21b)
Note that the number of coefficients in D¡(s) is one less than that of Do(s) if n
) and D¡(s) consists of is even; they are equal if n is odd. Using these coefficients we form the table
part of D(s) and D¡(s) shown in Table 8-1. lt is caBed the Routh tableo The coefficients in the first
of Do(s) is always one two rows are just the coefficients of Do(s) and D¡(s). The coefficients in the
'D¡(s) in the following remaining rows are obtained from its previous two rows by using the formula
a(k+ ¡la(k) _ a(k)a(k + ¡)
a\k+2)= ° <+~k+¡lo <+1 =a\~¡ -1Xk+¡a\\+/l (8-25)
ao

(8-22)

Table 8.1 Continuous-Time Stability Table (for n even)t


,
Sil abO) a\O) a~O) a~?~ a~9) :
1 I
,
n-¡S +­ ,
IX"S
s11-1
abl ) a\l) a~l) a(l)
"'-1
,
I

----------------- ----

5'1-2 ab2 ) a\Z) a~Z) a(Z) :


"'-1 I

,
I if and only if the n 5"-3 aff) a\3) a~3) a(3)
n'- 2
,
I

,
'e give an exampIe to a(M-Z) ,
1 ,
, ag'-Z)
(n-l)1
ao , 0("_1 =-'­
ag'-I)
Cf.,,=~)
ao

.t Ir n is odd, the first two rows have the same number of coefficients. Otberwise,
the pattern is idenlical.
-- _ _- _._---~-._~_ .. .
_-------_.~--_._---_._-------_ .. _-----_._--_.~-_._-_
~ .. _ - - - - - - - - - - - - - - ­
_--_._--_._----_._._------_._---._--_._--------"~-.~._._------------------_._--._-_._-.-._--
----_.~--~-_ ... ~._-_._-_

398 STABILITY OF LINEAR SYSTEMS

Using this equation


where (8-26)
a\n- 2) > O. Similarly
Table 8-1 are all po
There is a simple pattern in (8-25). Using this pattern, the Routh table can be
i = 1, 2, ... ,17, then al
readily computed. We note that the number of coefficients will decrease by
the theorem.
one at every odd power of s at the leftmost column of Table 8-1. The (J.¡ are
defined as the ratios of two subsequent coefficients in the first column of the
In forming the R
table. We shall show that these (J.¡ are equal to the (J.¡ in (8-22). In order to do so,
zero appears, we ma)
we use the coefficients in Table 8-1 to form the polynomials, for k = O, 1, ... , 17,
is not a Hurwitz po]
Hurwitz polynomial,
coefficients are positi
n-k
if n-k is even
J2
where
-l
k'-
n-~-l if n-k is odd
Example 2

Consider the polynor

Then it is straightforward to verify, by using (8-25) and (8-26),

which implies
AH the coefficients ar'
(8-27) Example 3
Consider the polynor
form
This equation holds for k=O, 1, ... , n-2. For k=n-1, we haveDn(s)jDn-¡(s)=
a~)ja~-¡)s = 1j(J.ns. Using (8-27), we can write Do(s)jD¡(s) into the form in (8-22).
This establishes that the (J.¡ in (8-22) can indeed be computed from the coefficients
in the first column of Table 8-1 as shown.

Theorem 8-7 (Routh-Hurwitz criterion) A zero appears in the


The polynomial D(s) of degree n in (8-19) 1s a Hurwitz polynmnial if and onl:;
Example 4
if the n terms ag), i = 1, 2, ... , n, in the first column of Table 8-1 are aH positive
or if and only if aH the coefficients aj¡) in Table 8-1 are positive. Consider 2s 4 + 2s 3 + s

Proof
The assumption ao = abO) > Oin (8-19) and the relations (J.¡ = ag - 1 )jag), i = 1,2, ... ,
n, imply (J.¡ > O, i= 1, 2, ... , n if and only if ag) > O, i = 1,2, ... , n. Hence the
A negative number ap;
first part of the theorem foHows directly from Theorem 8-6. To show the
polynomial.
second part, we show that aW > O, i = 1, 2, ... , n if and only if aH the coefficients
in Table 8-1 are positive. The sufficient part is obvious. To show the necessary
We note that in co
part, we write (8-25) as
- are not áffecüid if a ro'
a(k) - a(k, +2) + (J.k +1 a(k1 +1+1)
1+1 - the cbmputation of Ta
ROUTH-HURWITZ CRITERION 399

(8-26) a~'- 2) >


°
Using this equation we can show that if a~) > 0, a~-I) > and (/.n-l > 0, then
O. Similarly we can show that the coefficients associated with 53 in
Table 8-1 are aH positive. Proceeding upward, we can show that if ag) > 0,
le Routh table can be i = 1,2, ... , n, then aH the coefficients in Table 8-1 are positive. This establishes
ents will decrease by the theorem. Q.E.D.
'able 8-1. The (/.¡ are
le first column of the In forming the Routh table for a polynomial, if a negative number or a
~2). In order to do so, zero appears, we may stop the computation and conclude that the polynomial
als, for k = 0, 1, ... , n, is not a Hurwitz polynomial. However, to conclude that a polynomial is a
n- 2k' Hurwitz polynomial, we must complete the table and then check whether aH
eoefficients are positive.

Example 2
Consider the polynomial in Example 1. We form
1 6 1
\-26), 2 4
4 1
3.5
1

AH the coefficients are positive; hence the polynomial is Hurwitz. I

(8-27) Example 3
Consider the polynomial D(5) = 35 7 +25 6 + 25 5 + 54 + 35 3 + 52 + 1.55 + 1. We
form
e haveD,,(s)/D,,-¡(s)=
into the form in (8-22). 3 2 3 1.5
:d from the eoeffieients 2 1 1 1
1 1 O
2 2

A zero appears in the table; henee the polynomial is not Hurwitz. •

,olynomial if and only Example 4


ble 8-1 are aH positive
ositive. Consider 25 4 +25 3 +S2 +35 +2. We have

== ag-1l/ag), i = 1,2, ... ,


l, 2, . _. , n: Hence the A negative number appears in the table; henee the polynomial is not a Hurwitz
:ro 8-6. To show the polynomial. •
Iy if aH the coefficients
To show the necessary We note that in eomputing Table 8-1, the signs of the numbers in the table
are not affected if a rowis multiplied bY'a positive number. By using this fact,
the computation of Table 8-1 may often be simplified.
400 STABILITY OF LINEAR SYSTEMS

Example 5 where G(t, r) ~ C(t),


E. What we have d
Consider D(s) = 2s 4 + 5s 3 + 5s 2 + 2s + 1. We forrn the following table: applied here. We s
(see Section 2-8).
2 5
5 2
Theorem 8-8
21 5 (after the multiplication of 5)
17 (after the multiplication of 21) The zero-state respol
5 if there exists a finite
There are four positive numbers in the first column (excluding the first number);
hence D(s) is a Hurwitz polynomial. •
for any t o and for all
To conclude this section, we mention that the Routh table can also be used
to determine the number of roots in the right-half s planeo If ag> f.O, i =
The norm used in
1,2, ... ,n, then the number of roots in the right-half s plane is equal to the
of time, 11 u(t)11 can b(
number of sign changes in {ab1l , ab2 J, ... , ag'l}. See, for example, References
sponding to these dif
S41 and S44. If ag> =0 for sorne i, the situation becomes complicated. See
Reference S37 and a number of papers discussing this topic in Reference S239.
has different values.
be used.
8-4 Stability of Linear Dynamical Equations
Next we study th
we study the responsl
Time-varying case. Consider the n-dimensional linear time-varying dy­
namical equation
E: x= A(t)x + B(t)u (8-28a) due to any initial st~
y =C(t)x (8-28b) x(to) = X o is given by

where x is the n x 1 state vector, u is the p x 1 input vector, y is the q x 1 output


vector; and A, B, and C are n x n, n x p, and q x n matrices, respectively. It is It is clear that the B:
assumed that every entry of A, B, and C are continuous functions of t in ( - 00,(0). m ust introduce diffen
No assumption as to boundedness of these entries is made. We have assumed of equilibrium state.
that there is no direct transmission part in (8-28) (that is, E = O) because it does
not play any role in the stability study. Definition 8-2

The response of the state equation (8-28a) can always be decomposed into A state Xe of a dynam
the zero-input response and the zero-state response as only if

x(t)=$(t; t o, x o, u)=$(t; t o, xo,O)+$(t; t o, O, u)


for all t 2. too
Hence it is very convenient to study the stabilities of the zero-input response
and of the zero-state .response separately. Combining these results, we shall We see fram this (
immediately obtain the stability properties of the entire dynamical equation. and if no input is appli
First we consider the stability of the zero-state response. The response of
Hence at any equili
E with Oas the initial state at time t o is given by . . . equilibrium state of x

y(t) = l'
lO
C(t)<Il(t, r)B(r)u(r) d'l:
definition, an equilibr
xe =
= l'
lO
G(t,r)u(r)dr for all t 2. too Clearly,
STABILITY OF LINEAR DYNAMICAL EQUATIONS 401

where G(t, T) ~ C(t)$(t, T)B(T) is, by definition, the impulse-response matrix. of


E. What we have discussed in Section 8-2 (in particular, Theorem 8-2) can be
,lIowing table: applied here. We shall rephrase Theorem 8-2 by using the notion of norm
(see Section 2-8).

Theorem 8-8
n of 5)
n of 21) The zero-state response ofthe dynamical equation E is BIBO stable if and only
if there exists a finite number k such that

ling thefirst number);


• L IIC(t)$(t, T)B(T)II dT sk< 00 (8-29)

:able can also be used


for any t o and for all t ;::: to. •
plane. If agl 1= 0, i = The norm used in (8-29) is defined in terms ofthe norm ofu. At any instant
plane is equal to the
example, References
of time, Ilu(OII can be chosen as Li
lu¡(t)l, max¡ lu¡(t)l, or (Li
lu¡(t)12)l/2. Corre­
sponding to these different norms of u,
les complicated. See
}ic in Reference S239. Ilqt)Q)(t, T)B(T)II
has different values. However, as far as stability is concerned, any norm can
be used.
15 Next we study the stability of the zera-input response. More specifically,
we study the response of
ear time-varying dy­
x=A(t)x
(8-28a) due to any initial state. The response of x = A(t)x(t) due to the initial state
(8-28b) x(to) = Xo is given by

" Y is the q x 1 output x(t) ~ cjJ(t; to, xo, O) = $(t, to)iX o


es, respectively. It is It is c1ear that the BIBO stability can no longer be applied here. Hence we
;tions of t in ( - 00, 00). must introduce different kinds of stability. Before doing so we need the concept
e. We have assumed of equilibrium state.
1<:=0) becauseit does
Definition 8-2

; be decomposed into A state X e of a dynamical equation is said to be an equilibrium state at t o if and


only if

t o, O, u)
~ zero-input response
for all t ;::: to. •
hese results, we shall We see fram this definition that if a trajectory reaches an equilibrium state
ynamical equation. and ifno input is applied, the trajectory will stay .at the equilibrium state forever.
nse. The response of Hence at any equilibrium state, xe(t) = O for .all t;::: too Consequently, an
equilibrium state of x =A(t)x is a solution of A(t)x';"O for aH t ::::(0' Or fram
definition, an equilibrium state of x = A(t)x is a solution of
or
for all t ;::: to. Clearly, the zera state, O, is always an equilibrium state ofx = A(t)x.
402 STABlLITY OF LINEAR SYSTEMS

We shall now define the stability of an equilibrium state in terms of the Definition 8-4
zero-input response.
An equilibrium stat<
i.S.L. at t o and if eve
Definition 8-3
t --> oo. More precise I

An equilibrium state X e is said to be stable in the sense of Lyapunov at t o if and any e> 0, there exist
only if for every ¡; > 0, there exists a positive number Dwhich depends on ¡; and
t o such that if Ilx o- xell ~ D, then
for all t :2:t l +T lf
1I<I»(t; to, xo,O) - xell ~¡; pendent of ti> the eql
over [to,co).
for aH teto. It is said to be uniformly stable i.s.L. (in the sense of Lyapunov)
over [t o, (0) if and only if for every ¡; > 0, there exists a positive Dwhich depends
The concept of a
on ¡; but not on t o such that if Ilxo- xell ~ D, then
that the stabilities e
1I<I»(t; ti> xo,O) - xell ~6 because we do not kl
should be chosen. 1
for any t 1 e t o and for all t e t l' • property, ¿¡ and y can
rium state of a linear
If an equilibrium state is uniformly stable i.s.L., then it is stable i.s.L. How­ in the large.
ever, the converse may not be true. For example, the zero state of
Theorem 8-9
x(t) = (6t sin t - 2t)x(t)
Every equilibrium st¡ ,
is stable i.s.L., but not uniformly stable i.s.L. See Reference S206. Roughly and only if there exis
speaking, an equilibrium state X e is stable i.s.L. if the response due to any initial
state that is sufficiently near to X e will not move far away from X e . For time­
invariant systems, there is no difference between stability and uniforrn stability. for all t :2: too Ir k is i

Example 1 Proof
Consider the pendulum system shown in Figure P3-17 (page 127). The applica­ Sufficiency: Let Xe be
tion of Newton's law yields
u(t) cos e- mg sin e = mil}
Let Xl = e, and X2 = e. Then we have
Xl =X2

. = (g)

X2 -¡ sin Xl
COS Xl
+-----;;:;¡ U

Its equilibrium states are the solutions of Xl =


yield
° and X2 =0 with u{t) =0 which

k=O, ± 1, ±2, ...

The equilibrium states [kn O]', k = 0, ± 2, ± 4, , are uniformly


stable i.s. L.
The equilibrium states [kn O]', k = ± 1, ± 3, , however, are not stable
i.s.L. (Why?) Figure 8-4 Asympto
STABILITY OF LINEAR DYNAMICAL EQUATIONS 403

state in terms of the Definition 8-4


An equilibriurn state X e is said to be asymptotically stable at t o if it is stable
i.s.L. at t o and if every rnotion starting sufficiently near Xe converges to Xe as
t-"oo. More precisely, there is sorne y>O such that ifllx(tl)-xell.=s;y, then for
Lyapunov at t o if and any e> 0, there exists a positive T which depends on e, y, and ti such that
jch depends on f. and 11«P(t; t1> x(t 1), O) -xell.::;e
for aH t c. t 1 + T. If it is unifonnly stable Ls.L. over [to, 00) and if T is inde­
pendent of ti' the equilibriurn state is said to be uniformly asymptotically stable
over [to, 00). •
e sense of Lyapunov)
itive ¿j which depends The concept of asyrnptotic stability is illustrated in Figure 8-4. We see
that the stabilities defined in Definitions 8-3 and 8-4 are local properties,
because we do not know how srnall o in Definition 8-3 and y in Definition 8-4
should be chosen. However, for linear systerns, because of the hornogeneity
• property, ¿j and y can be extended to the entire state space. Hence, if an equilib­
riurn state of a linear equation is stable at all, it will be globally stable or stable
is stable i.s.L. How­ in the large.
) state of
Theorem 8-9
Every equilibriurn state of x= A(t)x is stable in the sense of Lyapunov at to if
ence S20G. Roughly and only if there exists sorne constant k which depends on t o such that
)Use due to any initial
y frorn X e. F or tirne­
1\ «I>(t, to)11 sk < 00

and uniforrn stability. for aH t c. too Ir k is independent of t o, it is uniforrnly stable Ls.L.


Proof
ge 127). The applica­ Sujficiency: Let Xe be an equilibriurn state of x = A(t)x; that is
Xe = «I>(t, tO)x e for all t c. t o

'o with u(t)=O which

tí +T
miformly stable Ls.L.
.ivever, are not stable
Figure 8-4 Asymptotic stability.

.J
404 STABILlTY OF LINEAR SYSTEMS

Then we have Lyapunov. Consegu


for any t o and aH t.
x(t) - Xe = cI>(t, tO)XO - Xe = cI>(t, tO)(X O- Xe) (8-30)
}' > 0, and for every E
lt follows from (2-90) that

¡¡x(t) - xell ~ 11 cI>(t, to)llllx o - xell ~ kllxo - xell


for allllxoll~'l', and
for all t. Henee, for any e, if we ehoose b = e/k, then Ilx o - xell ~b implies that 11 cI>(to + T, to)xoll = 11 e
that
Ilx(t)-xell~e forallt~to
Necessity: If Xe is stable i.s.L., then cI>(t, to) is bounded. We prove this by
eontradietion. Suppose that Xe is stable i.s.L. and that cI>(t, to) is not bounded, This eondition and 11
then for sorne to, say to, at least one element of cI>(t, to), say 4>ij(t, to), becomes
11 cI>(t, to)11 ~ k 3 fOl
arbitrarily large as t-4 oo. Let us choose Xo at time to such that all the eom­
ponents of (x o - x e) are zero exeept the jth component, whieh is egual to cx. 11 cI>(t, to)11 = 11 cI>(t, t o +
Then the ith component of (x(t) -xe) is egual to 4>Jt, to)'cx, whieh becomes k3
arbitrarily large as t-4 00, no matter how small cx is. Hence Xe is not stable i.s.L. ~2 fOl
This is a eontradiction. Hence, if Xe is stable i.s.L., then cI>(t, t o) is bounded.
11 cI>(t, to)11 ~ 11 cI>(t, t o +
Q.E.D.
k3
The zero state is, as mentioned earlier, an eguilibrium state. If the zero ~22 for
state is asymptotically stable, then the zero state is the only equilibrium state of
and so forth, as showl
x = A(t)x. lndeed, if there were another eguilibrium state different from the let k 1 = 2k 3 . Then fr,
zero state, then by choosing that eguilibrium state as an initial state, the response
wouldnot approach the zero state. Hence, if x = A(t)x is asymptotically
stable, the zero state is the only eguilibrium state.
for any t o and for a1l1

Theorem 8-10 If the zero state of


The zero state of x = A(t)x is asymptotically stable at t o if and only if initial state, the zero­
°
11 cI>(t, to)11 ~ k(t o) < 00 and 11 cI>(t, t o)ll-4 as t -400. The zero state is uniformly
asymptotieally stable over [0, (0) if and only if there exist positive numbers
for linear eguations, j
also said to be expone
k1 and k 2 sueh that We have diseussed
stability of the zero-sl
11 <1>(t, t o)\ 1 ~ k¡ e- k2 (1-lo) (8-31 )

for any t o ~ ° and for all t ~ too

Proof
We prove only the second part of the theorem.

Sufficiency: lf

then
¡lx(t)11 = 11cI>(t, to)x(to)11 ~ 11cI>(t, to)llllxoll ~kle-k2(1-'o)llxol¡ lO + í.

which implies that Ilx(t)II-4O at t-4oo, uniformlyin too Necessity: Ifthe zero
state is asymptotical1y stable, then by definition it is stable in the sense of Figure 8-5 11<I>(t, t 0)11
STABILITY Of LINEAR DYNAMICAL EQUATIONS 405

Lyapunov. Consequently, there exists a finite number k 3 such that 11 <I>(t, to)11 :::;;k 3
for any t o and all t ~ t o (Theorem 8-9). From Definition 8-4, there is sorne
(8-30)
y > O, and for every e> Othere exists a positive T such that
Ilx(to +T)II = 11<I>(to +T, to)xoll:::;;e (8-32)

-xell for aH Ilxoll sy, and for any to. Now choose an Xo such that Ilxoll ='1 and
- xe \\:::;;6 implies that Il<I>(to +T, to)xoll = 11<I>(to +T,to)lll!xoll and choose f. ='1/2; then (8-32) implies
that
1I <I>(t o + T, to)ll :::;; ~ for any t o
1. We prove this by
>(t, to) is not bounded, This condition and 11 <I>(t, to)11 s k3 imply that
say c/J¡/t, lo), becomes I! <I>(t, to)!1 :::;; k 3 for aH t in [to, t o + T)

:uch that aH the com­


\\<I>(t, t o)\\ = \I<I>(t, t o +T)<I>(t o +T, t o)\\ :::;;1\<I>(t, t o +T)I\\\<I>(to +T, t o)\\

, which is equal to rx.


k
to)'rx, which becomes :::;;23 for al1 t in [to +T,to +2T)
;e X e is not stable Ls.L.
n CI>(t, lo) is bounded. 1I<I>(t, to)11 :::;;II <I>(t, t o +2T)IIII<I>(to +2T, t o +T)IIII<I>(to +T, to)11
Q.E.D.
:::;;~; foralltin[t o +2T,t o +3T)
Jm state. lf the zero
ly equilibrium state of and so forth, as shown in Figure 8-5. Let us choose k2 such that e- k,T =~ and
ate different from the let k 1 = 2k 3 . Then from Figure 8-5, we see immediately that
,tial state, the response
f)x is asymptotical1y IICI>(t, to)11 :::;;k¡e- k, (r-ro)
for any t o and for al1 t ~ too Q.E.D.

lf the zero state of X: = A(t)x is uniformly asymptotically stable, then for any
at to if and only if initial state, the zero-input responsewi1l tend to zero exponentially. Hence,
ero state is uniformly for linear equations, if the zero state is uniformly asymptotically stable, it is
xist positive numbers also said to be exponentially stable.
We have discussed separately the stability ofthe zero-input response and the
stability of the zero-state response. By combining these results, we may give
(8-31 )

......
IIk3/2 - - - - - - ­
~k-4--- .
[ 31 ---- k 18
I 3 -­

k2(HOl\lx o ll lO +T i o + 2T lO + 3T

Necessity: lf the zero


,table in the sense of Figure 8-5 II <1>(t, to)11 bounded by an exponentialIy decreasing function.

L
::==::::--=.:.~.::-:::".:.-===::::::.=====-===:-._-~--'.'--_ .. _-_._-----"---~--~. __ ._._-----,-_._-----------~-- -~ _-_
..- -... _--_._--_._------_. __ ._-------_._~,-~. __ .-.-.--.~_

------------_.~_.~_._--~---.---.-_.--'-------~~------.
.... _._._-----_.
_.-------.._--­

406 STABILlTY OF LINEAR SYSTEMS

various definitions and theorems for the stability of the entire dynamical of the zero-state resp
equation E. Before doing SO, we shall discuss the relation between the stability study the entire respl
of the zero-state response and the stability of the zero-input response.
The necessary and sufficient condition for the zero-state response of E Definition 8-5
to be BIBO stable is that, for sorne finite k,
A linear dynamical e
L IIC(t)<1>(t, e)B(e)11 de :5; k < 00 if and only if for any
as all the state variab
for any t o and for all t ~ to. It has been shown by the function given in Figure
We see that the e
8-2 that an absolutely integrable function is not necessarily bounded. Con­
BIBO stability; they
versely, a bounded function need not be absolutely integrable. Hence, the
of all state variables;
stability Ls.L. of an equilibrium state, in general, does not imply nor is implied
also for any initial ~
by the BIBO stability of the zero-state response. A function that approaches
function properly, bel
zero as t--+ 00 may not be absolutely integrable; hence asymptotic stability
and the system will 1
may not imply BIBO stability. If a system is uniformly asymptotically stable,
every system is requil
then $(t, e) is bounded and absolutely integrable as implied by (8-31); hence
with sorne conditions on B and C, uniformly asymptotic stability may imply Theorem 8 -13
BIBO stability.
A system that is descr
Theorem 8-11 stable if and only ifC
Consider the dynamical equation E given in Equation (8-28). If the matrices
B and C are bounded on (- 00,00), then the uniformly asymptotic stability
of the zero state implies the BIBO stability of the zero-state response. I
for an~ t o and for all i

Proof Proof
This theorem fol1ows directly from the fact that The response of (8-28;

JIIC(t)$(t, e)B(e)11 de:5; JIIC(t)IIII<l>(t, e)IIII B(e)11 de ~klk2 JII$(t, e)il de XI

where IIB(t)11 ~kl' 11C(t)11 ~k2 for all t. Q.E.D. Hence we conclude f
initial state and any
The converse problem of Theorem 8-11~that of determining the COD­ <l>(t, e)B(T) is absolutel
ditions under which the BIBO stability of the zero-state response implies the bounded if and oniy ii
asymptotic stability of the zero state-is much more difficult. In order to solve
this problem, the concepts of uniform controllability and uniform observ­ Total stability cle~
ability (Definitions 5-4 and 5-8) are needed. We state only the result; its proof true. A systell,I may
can be found in Reference 102. If B and C are bour
stability. The situatic
*Theorem 8-12
rime~invariant ca~
Consider the dynamical equation E given in (8-28). If the matrices A, B, and C . tained for linear, time
are bounded on (- 00,00) and if E is uniformly controllable and uniformly because aH the conditi
observable, then thé zero state of E is asymptotically stable (under the zero- . are very difficult, if TI(
input respOnse) if and only if its zero-state response is BIBO s t a b l e . 1 time-invatiant .dynam
matrix is, however, no
We have studied the stability of the zero-input response and the stability the matrix A.
STABILlTY Of LINEAR OYNAMICAL EQUATIONS 407

:he entire dynamical of the zero-state response. Their relations are also established. We shall now
between the stability study the entire response.
mt response.
,-state response of E Definition 8-5
A linear dynamical equation is said to be totally stable, or T-stable for short,
if and only if for any initial state and for any bounded input, the output as well
as all the state variables are bounded. •
lction given in Figure We see that the conditions of T-stability are more stringent than those of
Hily bounded. Con­ BIBO stability; they require not only the boundedness of the output but also
tegrable. Hence, the of all state variables; the boundedness must hold not only for the zero state but
t imply nor is implied also for any initial state. A system that is BIBO stable sometimes cannot
~tion that approaches function properly, because sorne of the state variables might increase with time,
: asymptotic stability and the system will bum out or at least be saturated. Therefore, in practice
asymptotically stable, every system is required to be T-stable.
plied by (8-31); hence
c stability may imply Theorem 8-13
A system that is described by the linear dynamical equation E in (8-28) is totally
stable if and only if C(to) and <l>(t, t o) are bounded and

8-28). If the matrices [II<Il(t, ,)B(,)II d, sk< el)


y asymptotic stability
tate response. • for any t o and for aH t '2: to.

Proof
The response of (8-28a) is

r <Il(t, ,)B(,)u(,) d,
t

x(t) =<Il(t, to)x o +


',¡kz JI1cD(t, ')11 do Jo
Q.E.D. Hence we conclude frum Theorems 8-8 and 8-9 that x is bounded for any
initial state and any bounded input if and only if <l>(t, to) is bounded and
<l>(t, ,)B(,) is absolutely integrable. From y(t) = C(t)x(t), we conclude that y is
determining the con­
bounded if and only if C(t) is bounded. Q.E. D.
~ response implies the
cult. In order to solve
Total stability clearly implies BIBO stability; the converse however is not
and uniforrn observ­
true. A systeIl} may be totaHy stable without being asymptotically stable.
lly the result; its proof
If B and C are bounded, then uniformly asymptotic stability implies total
stability. The situation is similar to Theorem 8-11.

Time-invariant case. Although various stability conditions have been ob­


Le matrices A, B, and C tained for linear, time-varying dynamical equations, they can hardly be used,
oHable and uniformly because aH the conditions are stated in terms of state transition matrices, which
table (under the zero­ are very difficult, if not impossible, to obtaio. In the stability study of linear
BO stable. • time-invariant dynamical equatioos, the knowledge of the state transition
matrixis, however, not needed. The stability can be determined directly from
lonse and the stability the matrix A.
408 STABILITY Of LINEAR SYSTEMS

Consider the n-dimensional linear time-invariant dynamical equation


FE: x=Ax +Bu (S-33a)
y=Cx (S-33b)
Its equilibrium states
where A, B, C are n x n, n x p, q x n real constant matrices, respectively. As in
words, every point in tl
the time-varying case, we study first the zero-state response and the zero-input
of the matrix are - 1, (
response and then the entire response. The zero-state response of FE is charac­
value O, which has a ze:
terized by
Hence every equilibriu
G(s) =C(sI - A)-lB
Example 3
From Theorem 8-5, the zero-state response oj FE is BIBO stable if and only if al!
the poles ofevery entry oj(;(s) have negative real parts. The zero-input response Consider
of FE is governed by x = Ax or x(t) =eA1xO' Recall that an equilibrium state
of x = Ax is a solution of x = O or Ax = O, and has the property X e = eAtx e for
all t ~O. Note also that stability implies uniform stability in the time-invariant
case.
It has eigenvalues -1
Theorem 8-14 eigenvalue O is not a
Every equilibrium state of x = Ax is stable in the sense of Lyapunov if and only equilibrium states of th
if all the eigenvalues of A have nonpositive (negative or zero) real parts and those
with zero real parts are distinct roots of the minimal polynomial of A.3 Theorem 8 -1 5

The zero state of x = Ax


Proof
of A have negative real
Let X e be an equilibrium state of x = Ax. Then x(t) - X e = eAI(x o - xJ Hence
every equilibrium state is stable í.S.L. if and only if there is a constant k such Proof
that IleAtll.::; k < <X) for aH t ~ O. Let P be the nonsingular matrix such that
A= PAp-l and Ais in the lordan formo Since e AI = PeAlp-l, then In order for t he zero statl
ness of JleAtll, it is requi
IleAtll.::;llplllleAtllllp-lll that IleAtll~O as t---+oo.
Consequen,tly, if I\eAtll is bounqed, so is IleAtll. Conversely, from the equation conclude that IleAtl1 ~O
e At =p-1eAtp, we see that if IleAtl1 is bounded, so is IIeAtll. f!ence W'Y conclude consequently of A have
that every equili~rium state is stable í.s.L. if and only if IIeAtl1 i~ bounded on
[O, 00). Now IleAtl1 is bounded if and only if every entry of e A1 is bounded. lf a linear time-inv,
Since Á is in the lordan form, every entry of e At is of the form leajt +irojt, where response will approach z
el. j + iro j is an eigenval ue of A(see Section 2-7). Ir el. j is negative, it is easy to see
stable. This is consiste]
that leajt +irojt is bounded on [O, 00) for any integer k. Ir el.j = O, the function the time-invariant case i
le irojt is bounded if and only if k = O-that is, the order of the lordan block The eigenvalues of
associated with the eigenvalue with el. j = O is 1. Q.E. D. A, det(sI - A) =0. We
Hurwitz criterion to che
Example 2 negative real parts. Her
of x = Ax can be easily
Consider nomial of A and then ap
The BIBü stability (
determined by the pales
3 Equivalently, if A is transformed ¡nto the Jordan form, the order of every Jordan blocks associated
wÚh eigenvalues with zero real parts is l. Note that this condition does not imply that the eigen­
values of A with zero real parts are distinct G(s) =C(sI·
STABILITY Of LINEAR DYNAMICAL EQUATIONS 409

llamical equation
(8-33a)
(S-33b)
Its equilibrium states are [x le X2e O]', for any Xl e -1=0, X2e -1=0. In other
es,respectively. As in words, every point in the Xl - X2 plane is an equilibrium state. The eigenvalues
llse and the zero-input of the matrix are -1, O, and O. Its minimal polynomial is s(s + 1). The eigen­
;ponse of FE is charac­ value O, which has a zero real part, is a distinct root of the minimal polynomial.
Hence every equilibrium state is stable i.s.L. •

Example 3
, stable if and only if alI
he zero-input response Consider
lt an equilibrium state
r
property X e = e Al x e 101'
:y in the time-invariant
~O -1~] x(t)

It has eigenvalues -1, O, and O. Its minimal polynomial is S2(S + 1). The
eigenvalue O is not a distinct root of the minimal polynomial. Hence the
equilibrium states of the equation are not stable i.s.L. I
fLyapunov ifand only
:ro) real parts and those
lynomial of A. 3 Theorem 8-15
The zero state ofi = Ax is asymptotically stable if and only if all the eigenvalues
of A have negative real parts.
Hence
• = eA'(x o - x e ).
~e is a constant k such Proof
?;ular matrix such that In order for the zero state to be asymptotically stable, in addition to the bounded­
Alp.-l, then
ness of JIeAlII, it is required that IleAl11 tends to ~ero as t-+ ro, 01' equivalently,
that IleA'II-+ Oas ! -+ ro. Since every entry of eA' is of the form tkeajl +':j" we
conclude that I\eAlII-+O as t -+ ro if and only if all the eigenvalues of A, and
sely, from the equation consequently of A have negative real parts. Q.E.D.
'11. I-!ence w~ conc1ude
if Ile A '\ I i~ bounded on Ir a linear time-invariant system is asymplOtically stable, its zero input
Iltry of eA' is bounded. response will approach zero exponentially; thus it is also said to be exponentialIy
e form tkeajl +iOJj" where
stable. This is consistent with Theorem S-lO because asymptotic stability in
legative, it is easy to see the time-invariant case implies uniformly asymptotic stability.
If aj=O, the function The eigenvalues of A are the roots of the characteristic equation of
.er of the J ordan block A, det (sI - A) = O. We have introduced in the previous section the Routh­
Q.E.D.
Hurwitz criterion to check whether 01' not all the roots of a polynomial have
negative real parts. Hence the asymptotic stability of the zero-input response
of i = Ax can be easily determined by first forming the characÚ:ristic poly­
nomial of A ando then applying the Routh-Hurwitz criterion.
TheBIBO stability of the linear time-invariant dynamical equation FE is
determined by the poies of G(s). Since
very lordan blocks associated

det (s~ -A) e [Adj (sI -A)JB


does not imply that the eigen-
G(s) =C(sI -A( lB
410 STABILITY Of LINEAR SYSTEMS

every pole of (;(s) is an eigenvalue of A (the converse is not true). Consequently, It is clear that tol
if the zero state of FE is asymptotically stable, the zero-state response of FE ever may not imply 1
will also be BIBO stable. (This fact can also be deduced directly from Theorem the BIBO stability.
8-11.) Conversely, the BIBO stability of the zero-state response in general asymptotic stability
does not imply the asymptotic stability of the zero state, because the zero­ asymptol ic stability is
state response is determined by the transfer function, which, however, describes If a linear time-in
only the controllable and observable part of a dynamical equation. able, then the charac
?olynOmial 2f G(s) (1
Example 4 IS a pole of G(s), and (
we have the fOllowing
Consider a system with the following dynamical equation description:

x=G _~Jx+[~Ju Theorem 8-17

If a linear time-invarian
y=[1 lJx
then the following state
Its transfer function is
1. The dynamical eq ua
g(s) = [1 lJ [S-1
-1 s+1
° J-l [OJ =s+11
1
2.
3.
The zero-state respol
The zero state of FE j
4. AH the poles of the tI
Hence the zero-state response of the dynamical equation is BIBO stable;
5. AH the eigenval ues of
however, the zero state is not asymptotically stable, because there is a positive
eigenvalue. I
A system is said to h
Theorem 8-16 function matrix if the d
troHable' and observable.
Let pletely characterized by i,
x= [~e ~;2 Jx +[~eJ u (8-34a)
the system can be detern
need of considering the d
y = [ Ce Ce J x (8-34b) A remark is in order 1
is independent of t and if
be an equivalent dynamical equation of the dynamical equation in (8-33) the zero state is asymptoti
with {Ae, Be} controllable. Then the dynamical equation in (8-33) is totally that if fo~ each t, aH the eig
stable if and only if al! the eigenvalues of Ae have negative real parts and all the state of x = A(t)x is asym r
eigenvalues of Al' have negative or zero real parts and those with zero real parts the folIowing exampk.
are distinct roots of the minimal polynomial of Al"
Example 5
Proof
Consider the linear time-v:
The application of the Laplace transform to (8-34a) yields

[Si -A
x(s) = e - A~? J- \(0) +[Si -Ae - A~2 -1[BeJu(s) J-
O si - Al' O si - Ac O
[(SI -AJ-l ...~ -IJX(O) +[(Sl - AJ- BeJu(s) The characteristic polynom
1
=
O l$.-Ac) O
where M= (si- Ae)-IA I2 (sl -Áe)-I. From this eql,lation, we may conclude det [,1.1- A
that x is bounded for ahy initialstate and any bounded u if and only if the
conditions in the theorem hold. lf x is bounded, so is y. This establishes the Hence the eigenvalues of A
theorem. Q.E.D. ofthe equation is neither asy
STABILITY OF LINEAR DYNAMICAL EQUATIONS 411

It is clear that total stability implies BISO stability. BIEO stability how­
,t tme). Consequently,
ever may not imply total stability because no condition is imposed on Ae in
o-state response of FE the BIEO stability. A comparison of Theorems 8-15 and 8-16 yields that
directly from Theorem asymptotic stability implies total stability but not conversely. Hence
.te response in general asymptotic stability is the most stringent among these three different stabilities.
tate, because the zero­ If a linear time-invariant dynamical eguation is controllable and observ­
LÍch, however, describes able, then the characteristic polynomial of A is egual to the characteristic
al equation. polynomial of G(s) (Theorem 6-2). This implies that every eigenvalue of A
is a pole of G(s), and every pole of G(s) is an eigenvalue of A. Consequently,
we have the following theorem.
on description:
Theorem 8-17
If a linear time-invariant dynamical equation FE is controllable and observable,
then the following statements are equivalent:
1. The dynamical equation is totally stable.
2. The zero-state response of FE is BIEO stable.
1
3. The zero state of FE is asymptotically stable (under the zero-input response).
- s +1
4. AH the poles of the transfer function matrix of FE have negative real parts.
uation is BIEO stable;
5. AH the eigenvalues of the matrix A of FE have negative real parts. I
::cause there is a positive

I A system is said to be completely or faithfully characterized by its transfer­


function matrix if the dynamical-equation description of the system is con­
troHable and observable. We see from Theorem 8-17 that if a system is com­
pletely characterized by its transferjunction matrix, then asymptotic stability of
the system can be determined from its transfer-function matrix alane with no
(8-34a) need of considering the dynamical-equation description of the system.
A remark is in order concerning the stability of x = A(t)x. If the matrix A
(8-34b) is independent of t and if aH the eigenvalues of A have negative real parts, then
the zero state is asymptotically stable. Hence one might be tempted to suggest
mical equation in (8-33)
that iffor each l, aH the eigenvalues of A(l) have negative real parts, then the zero
.lation in (8-33) is totally
state of x= A(t)x is asymptotically stable. This is not so, as can be seen from
ltive real parts and all the
the following example
those with zero real parts
Example 5
Consider the linear time-varying equation
¡ields . [-1 J
x= e
O -1 x
2l

- A~.? llrBc1 u(s)


sl-AcJ Lo
The characteristic polynomial of the matrix A at each t i.s given by
_~c)-lBc1 u(s)

det [ Al ~ A] .= det [A. O


+1
quation, we may conclude
unded u if and only if the Hence the eigenvalues of A are - 1 and - 1 for aH l. However, the zero state
) is y. This establishesthe of the equation is neither asymptotically stable nor stable i.s.L., because the state
Q.E.O.
412 STABILlTY OF LINEAR SYSTEMS

transition matrix of the equation is where the ,l;'s are the

lI>(t'O)=[e~' , '\'¿ 1-
)~min
n
I
Xii
i= 1
(as in Problem 4-1 or by direct verification), whose norm tends to infinity
as t----> OCJ. •
The fact that p- I =P

*8-5 Lyapunov Theorem


Hence, the inequality
The asymptotic stability of x = Ax can be determined by first computing the
characteristic polynomial of A and then applying the Routh-Hurwitz criterion.
If al1 the roots of the characteristic polynomial have negative real parts, then Definition 8-6
the zero state of x = Ax is asymptotical1y stable. There is one more method of A hermitian matrix l\
checking the asymptotic stability of x = Ax without computing explicitly the for all nonzero x in en
eigenvalues of A. We shal1 discuss such a method in this section and then or nonnegative definite
apply it to establish the Routh-Hurwitz criterion. holds for sorne nonzer
Before proceeding, we need the concept of positive definite and positive
semidefinite matrices. An n x n matrix M with elements in the field of complex
numbers is said to be a hermitian matrix if M* = M, where M* is the complex Theorem 8-19
conjugate transpose of M. If M is a real matrix, M is said to be symmetric. A hermitian matrix IV
The matrix M can be considered as an operator that maps (en, 1[:) into itself. any one of the fol1owiJ
It is shown in Theorem E-l that al1 the eigenvalues of a hermitian matrix are
real, and that there exists a nonsinguIar matrix P, cal1ed a unitary matrix, such l. AIl the eigenvalues
that p-I = P* and M =PMP*, where M is a diagonal matrix with eigenvalues 2. AH the leading prin
on the diagonal (Theorem E-4). We shall use this fact to establish the following of M are nonnegati
theorem. 3. There exists a nom
N*N. 6
Theorem 8-18
Let M be a hermitian matrix and let Amin and }'max be the smallest and largest
eigenvalues of M, respectively. Then
4 The prillcipal millors of the
AminllxW .::;;x*Mx '::;;Am¡¡xllxW (8-35)

for any x in the n-dimensional complex vector space 1[:/1, where

IlxW ~ (x, x) ~ x*x = ¿ Ixd2


are m,,_ m22_ m))_ det [
m,
i= I
m2
and Xi is the ith component of x. minors whose diagonal ele:

millors of M are m". det [


Proof
last k columns and the last
Note that x*Mx is a real number fO( any x in en. Let P be the nonsingular 5 It is shown in Reference 39
matrix such that P - I = P* and M =PMP*, where M is a diagonal matrix the principal minors are pe
with eigenvalues of M on the diagonal. Let x = Px or x = P - IX = P*x, then are nonnegative, then all
Problem 8-32b.
n
" Ir N is an upper triangular
x*Mx = x*PMP*x = x*Mx = ¿ AJX'¡2 able in LlNPACK and lB~
i= 1
LYAPUNOV THEOREM 413

where the },¡'s are the eigenval ues of M. It fol1ows that


n

A min ~
, 1-1 2::S;X *Mx -__x*Mx -
1... Xi A __ '\-\2 ::s; A'
1... A¡ X¡
"
max (8-36)
i=l i= 1
norm tends to infinity
I
The fact that p-l = P* implies that
n

IlxW=x*x=x*x= I IX¡1 2

i= 1

Hence, the inequality (8-36) implies (8-35). Q.E.D.


by first computing the
outh-Hurwitz criterion.
Definition 8-6
egative real parts, then
: is one more method of A hermitian matrix M is said to be positive definite if and only if x*Mx > O
)mputing explicitly the for all nonzero x in 1[:". A hermitian matrix M is said to be positive semidefinite
1 this section and then or nonnegative definite if and only if x*Mx::::::O for all x in 1[:", and the equality
holds for some nonzero x in C". I
'e definite and positive
s in the field of complex
Theorem 8-19
here M* is the complex
IS said to be symmetric. A hermitian matrix M is positive definite (positive semidefinite) if and only if
maps (C", iC) into itself. any one ofthe following conditions holds:
a hermitian matrix are
j a unitary matrix, such 1. Al! the eigenvalues of M are positive (nonnegative).
matrix with eigenvalues 2. All the leading principal minors 4 of M are positive (al1 the principal minors
) establish the following of M are nonnegative).s
3. There exists a nonsingular matrix N (a singular matrix N) such that M =
N*N. 6

the smallest and largest


4The principal minor, of lhe malrix

" where
M = j:::: ;::~ ;::~lJ'
Lm.31 m32 nt:33

m!2'J'], det[m m,,], det[m n


11
are m". mn. m.'.,. det [m\l m,,], and detM. that ¡s, the
m21 m22 m31 m.:n m:n m33
minors whose diagonal elements are also diagonal elements of the matrix. The leading principal

minors or M are m", det [m"


m21
mIl], and det M, thal ¡s, the minors obtained by deleling the
m22 .
last k columns and lhe last k rows, for k = 2, 1. and O.
~t P be the nonsingular 5 lt is shown in Referenee 39 that if a11 the leading principal minors of a malrix are positive, then all
the principal minors are positive. However, it is nol true thal ir all the leading.prÍncipal minors
~ is a diagonal matrix
are nonnegative, then 'all the principal minors are nonnegative. For a counterexample, lry
,r x = P-1x = P*x, then Problem 8-32b. .
(, Ir N is an upper lriangular malrix, i~ is called lhe Cholesk)' decomposilion. Subroulines arc avail­
able in LlNPACK and IBMScientific Subroutine Package to carry out this decomposition.
,
1
I

k
414 STABlLITY OF LINEAR SYSTEMS

Proof will be bowl shaped,


values taken by Val<
Condition 1 follows directly from Theorem 8-18. A proof of condition 2 can value of V will incre
be found, for example, in References 5 and 39. For a proof of condition 3, trajectory. Taking ti­
see Problem 8-24. Q.E. O. x = Ax, we obtain
With these preliminaries, we are ready to introduce the Lyapunov theorem d d
- V(x(t» = ­
and its extension. They will be used to prove the Routh-Hurwitz criterion. dt dt
=x*
Theorem 8-20 (Lyapunov theorem)

AlI the eigenvalues of A have negative real parts or, equivalently, the zero state where N ~ -(A*M·
of x = Ax is asymptotically stable if and only if for any given positive definite along any trajectory
hermitian matrix N, the matrix equation - x*(t)Nx(t) is always
A*M+MA= -N (8-37) cally with time along
approach zero as t -> (
has a unique hermitian solution M and M is positive definite. I only at x =0; hence v
M and N that are relat
Corollary 8-20 the zero state as t -> (
AlI the eigenvalues of A have negative real parts, or equivalently, the zero state x=Ax. A Lyapuno'
of x = Ax is asymptotically stable, if and only if for any given positive semi­ concept of distance or
definite hermitian matrix N with the property {A, N} observable, the matrix of x = Ax decreases w
equation
Proof of Theorem 8-;
A'!'M+MA= -N
Sufficienc y: Consider .
has a unique hermitian solution M and M is positive definite. I

The implication of Theorem 8-20 and Corollary 8-20 is that if A is


asymptotically stable and if N is positive definite or positive semidefinite, along any trajectory o
then the solution M of (8-37) must be positive definite. However, it does not
say that if A is asymptotically stable and if M is positive definite, then the
matrix N computed from (8-37) is positive definite or positive semidefinite.
Before proving the Lyapunov theorem, we make a few comments. Since
Theorem 8-20 holds for any positive definite hermitian matrix N, the matrix
N in (8-37) is often chosen to be a unit matrix. Since M is a hermitian matrix,
there are n 2 unknown numbers in M to be solved. If M is a real symmetric
matrix there are n(n + 1)/2 unknown numbers in M to be solved. Hence the
matrix equation (8-37) actually consists of n 2 linear algebraic equations. To
apply Theorem 8-20, we first solve these n 2 equations for M, and then check
whether or not M is positive definite. This is not an easy task. Hence Theorem
8-20 and its corollary are generally not used in determining the stability of
x = Ax. However, they are very important in the stability study of nonlinear
time-varying systems by using the so-called second method of Lyapunov.
Furthermore, we shall use it to prove the Routh-Hurwitz criterion.
We give now a physical interpretation of the Lyapunov theorem. If the
hermitian matrix M is positive definite, the plot of V(x) Xl

V(x) ~ x*Mx (8-38) Figure 8-6 A Lyapun,


LYAPUNOV THEOREM 415

will be bowl shaped, as shown in Figure 8-6. Consider now the successive
values taken by V along a trajectory of x= Ax. We like to know whether the
:oof of condition 2 can value of V wil\ increase or decrease with time as the state moving along the
l proof of condition 3, trajectory. Taking the derivative of V with respect to t along any trajectory of
Q.E.D. x= Ax, we obtain
the Lyapunov theorem :t V(x(t)) = {~t (x*(t)Mx(t)) = (:t x*(t)) Mx(t) +x*(t)M (:t X(t))
:h-Hurwitz criterion.
=x*(t)A*Mx(t) +x*(t)MAx(t)=x*(t)(A*M +MA)x(t)
= -x*(t)Nx(t) (8-39)

ivalently, the zero state where N ~ - (A*M + MA). This equation gives the rate of change of V(x)
. given positive definite along any trajectory of x= Ax. Now if N is positive definite, the function
- x*(t)Nx(t) is always negative. This implies that V(x(t)) decreases monotoni­
cally with time along any trajectory of x = Ax; hence V(x(t)) will eventually
(8-37)
approach zero as t -> oo. Now since V(x) is positive definite, we have V(x) = O
efinite. I only at x = O; hence we conclude that if we can find positive definite matrices
M and N that are related by (8-37), then every trajectory ofi = Ax will approach
the zero state as t -> oo. The function V(x) is caBed a Lyapunov function of
i = Ax. A Lyapunov function can be considered as a generalization of the
Livalently, the zero state
concept of distance or energy. Ir the "distance" of the state along any trajectory
ny given positive semi­
of i = Ax decreases with time, then x(t) must tend to O as t-> oo.
observable, the matrix
Proof of Theorem 8-20
Sufficienc y: Consider V(x) = x*Mx. Then we have
efinite. I
. d
V(x) ~ dt V(x) = - x*Nx
8-20 is that if A is
r positive semidefinite, along any trajectory of i = Ax. From Theorem 8-18, we have
However, it does not V x*Nx (AN)min
,itive definite, then the -=---<---- (8-40)
V x*Mx - (AM)max
positive semidefinite.
few comments. Since
V(x)
n matrix N, the matrix
ti is a hermitian matrix,
. M is a real symmetric
I be solved. Hence the
1gebraic equations. To
for M, and then check
y task. Hence Theorem
rmining the stability of
>ility study of nonlinear
method of Lyapunov..
itz criterion.
punov theorem. lfthe
)
Figure 8-6 A Lyapunov function V(x).
(8-38)
\
!
\
J
416 STARII.ITY (lF I.INEAR SYSTEMS

where (AN)min is the smal1est eigenvalue of N and (AM)max is the largest eigenvalue t -" oo. This shows t:
of M. From Theorem 8-19 and from the assumption that the matrices M and similarly proved as ir
°
N are positive definite, we have (AN)min > and (AM)max > O. lf we define Theorem 8-20 ar
asymptotic stability
CI. ~ (AN)min right and are basic
- (AM)inax provide a simple pro.
then inequality (8-40) becomes V:::;; -Cl.v, which implies that V(t):::;;e-atV(O).
lt is clear that CI. > O; hence V decreases exponentially to zero on every trajectory A proof of the Ro
°
of x = Ax. Now V(x) = only at x = O; hence we conclude that the response of
x = Ax due to any initial state X o tends to Oas t -" oo. This proves that the zero
D(s) = (

state of x = Ax is asymptotical1y stable. N ecessity: If the zero state of x = Ax with real coefficients .
is asymptotically stable, then all the eigenvalues of A have negative real parts.
Consequently, for any N, there exists a unique matrix M satisfying
A*M +MA=-N and compute
and M can be expressed as Do(s)
D (s) =CI.¡:
M = L'" eA 'INe A' dt (8-41 ) 1

(see Appendix F). Now we show that if N is positive definite, so is M. Let H


be a nonsingular matrix such that N = H*H (Theorem 8-19). Consider

xóMxo = L'" XÓ eA *'H*HeA'xo dt = L'" IIHeAtxoW dt (8-42)

Since H is nonsingular and e At is nonsingular for all t, we have HeAtx o =1=- O for
°
all t unless X o =0. Hence we conclude that xóMxo > for all X o =/=-0, and M
is positive definite. This completes the proof of this theorem. Q.E.D.
For convenience, we s

Theorem 8-6

In order to establish Corollary 8-20, we show that if N is positive semidefinite The polynomial D(s) i
CI.¡, Cl. z , .. . , Cl. are posit
and if {A, N} is observable, then x*(t)Nx(t) cannot be identically zero along any 1I

nontrivial trajectory ofx = Ax (any solution due to any nonzero initial state xo)'
First we use Theorem 8-19 to write N as N = H*H. Then the observability ProoF
of {A, N} implies the observability of {A, H} (Problem 5-36). Consider First we assume that ~
x*(l)Nx\t) ,.~ x(¡cH"ik';x¡", lilleA(xoi¡2 function

Since {A, H} is observable, all rows of He At are linearly independent on [0, 00). g(s) ~
Hence we have that HeA1x o = O for all t if and only if X o = O. Because e A1 is
analytic over [0,00), we conclude that for any X o =/=- O, HeAlxo can never be The assumption Cl. i =/=-1
identically zero over any finite interval, no matter how small; otherwise it factor between Do(s)
would be identically zero over [0, 00). See Theorem B-l. Note that HeA1x o = O, between D¡(s) and D(s;
at sorne discrete instants of time, is permitted. Cohsider the block
With the preceding discussion, we are ready to establish Corollary 8-20. function from u to y is !
Consider the Lyapunov function V(x) defined in (8~38)and dV(x)/dt = equivalently, from the
- x*(t)Nx(t) in (8-39). If X o =1=- O, dV(x)/dt:::;;O and theequality holds only at
sorne discrete instants of time; hence V(x(t)) will decrease with time, not neces­
sarily monotonic at every instant of time, and will eventually.approach zero as 7 This rollows Reference 90.
.. =~

LYAPUNOV THEOREM 417

t --+ oo. This shows the sufficiency of the corollary. The necessary part can be
the largest eigenvalue
3.t the matrices M and similarly proved as in Theorem 8-20 by using (8-42).
O. If we define
Theorem 8-20 and its corollary are generally not used in checking the
asymptotic stability of x = Ax. However, they are important by their own
right and are basic in the stability study of nonlinear systems. They also
provide a simple proof of the Routh-Hurwitz criterion.
s that V(t) s: e-'" V(O).
ero on every trajectory A proof of the Routh-Hurwitz criterion. Consider the polynomial
je that the response of ao>O
lis proves that the zero
with real coefficients a¡, i =0, 1,2, ... , n. We form the polynomials
le zero state of X = Ax
ve negative real parts. Do(s)=aos n +azs n - Z + ...
[ satisfying D 1(s)=a1 Sn - 1 +a3 Sn - 3 + ...
and compute
Do(s)
--=<XlS +-----------:-------­
D 1 (s)
(8-41 ) <XzS +----------:-----­
1
<X3 S + - - - - - - - - - - ­

:finite, so is M. Let H
1
8-19). Consider +--­
(842)

+-
ve have HeAlxo O for For convenience, we shall restate the theorem here.
) for all Xo =/= O, and M
orem. Q.E. D. Theorem 8-6
The polynomial D(s) is a Hurwitz polynomial if and only if aH the n numbers
is positive semidefinite <X l' a z , ... , <XII are positive.
ntically zero along any
onzero initial state xo)·
ProoF
rhen the observability
)-36). Consider First we assume that a11 the iX/S are different from zero. Consider the rationa!
function
"011' 1
indepeodent 00 [O, el»).
X o = O. Because e
Al'IS
',HeA'xo can oever be The assumption <Xi =/=0, for i = 1, 2, ... , n, implies that there is no common
)w smaH; otherwise it factor between Do(s) and D 1(s). Consequent1y, there is no common factor
Note that HeAtx o = O, between Dl(s) and D(s); in other words, g(s) is irreducible.
Consider the block diagram shownin Figure 8-7. Weshow that the transfer
:ablish Corollary 8-20. function from u to y is 9(s). Let h 1(s) be the transfer function from X n to X n -1 or
(8-38) and dV(x)/dt = equivalently, from the terminal E to the terminal F, as shown in Figure 8-7.
equality holds only at
;e with time, not neces­
7 This follows Reference 90.
ually approach zero as
418 STABILITY Of LINEAR SYSTEMS

y Let the M rnatrix in (

Figure.8-7 A block diagram of g(s). Then it is easy to veri

Then
A*M -t
y(s) 1/a l s 1
D(s) = 1 +[1 +hl(s)J/als = 1 +als +hl(s)

Let hz(s) be the transfer function from X,,-l to x,,_ z, then h l (s) can be written as It is clear that N is .
verify that {A, N} is o
the zero state of x =)
hl(s) = Yazs = 1~ positive definite, or eq
1 + hz(s)/azs a 2 s +hz(s)
the zero state of x = A
Proceeding forward, we can show easily that the transfer function frorn u to y of A, or equivalently
is indeed g(s). With the state variables chosen as shown, we can readily write words, D(s) is a Hurwi
the dynamical equation of the block diagram as are positive.
Consider now the
In other words, sorne
Xl O O O O O Xl O equal to zero. Suppo
an
8-1 are equal to zero,
-1 factor. The cornmon
Xz O O O O Xz
O
an- 1 an- 1 say fes). Then D(s) c
-1 even function or an (
X3 O
an- Z
O O O O X3
+
O u Hurwitz polynornial.
I : zero, we may replace

~:lj lxj l:J


complete Table 8-1.

l:
-1
Xn-l O O O Ir sorne a i is negative, ~
az by e) has a positive re
-1 functions of its coeffi.1
xn O O O
positive or zero real p.
al
(8-43)
y=[ O O O O O lJx

Irreducibility of (8-43) can be verified either by showing that it is controUable *8-6 Linear Tim
and observable or by the fact that its dimension is equal to the degree of the
denorninator of g(s). Consequently, the characteristic polynomial of the The stability concepts
rnatrix A in (8A3) is ~qual to the denorninator of g(s) (Theorem 6-2). Now we 'applicable'to the disCl
shall derive the condition. for the zero state of (8-43) to be asymptotically st¡¡.ble. quite different. In thi
DISCRETE-TIME SYSTEMS 419

Let the M matrix in Corollary 8-20 be chosen as

M~ ~'
O O
Ci. 1I _¡ O

O
O
Ci. z

O J] (8-44)

Then it is easy to verify that

O O O O O
O O O O O
O O O O O
A*M +MA=­ ~-N (8-45)

O O O O O
¡ (s)

O O O O 2

1 h¡ (s) can be written as


lt is clear that N is a positive semidefinite matrix. lt is straightforward to
verify that {A, N} is observable. Hence from Corollary 8-20 we conclude that
the zero state of x = Ax is asymptotically stable if and only if the matrix M is
positive definite, or equivalently, the n numbers Ci.¡, Ci. z, ... , Ci. 1I are positive. Now
the zero state of x = Ax is asymptotically stable if and only if all the eigenvalues
:r function from u to y of A, or equivalently all the roots of D(s), have negative real parts. In other
n, we can readily wríte words, D(s) is a Hurwitz polynornial if and only if all the n numbers Ci.¡, Ci. z , ... ,Ci. 1I
are positive.
Consider now the case in which not all the Ci.;'S are different from zero.
In other words, sorne of the coefficients in the first column of Table 8-1 are
X¡ O eq ual to zero. Suppose abZ ) = O. If all the coefficients in the Sil - Z row of Table
8-1 are equal to zero, it irnplies that Do(s) and D¡(s) have at least one cornrnon
factor. The common factor is clearly either an even or an odd function of S,
X z O
say f(s). Then D(s) can be factored as f(s)i5(s). Since not all the roots of an
even function or an odd function can have negative real parts, D(s) is not a
X3 + O u Hurwitz polynomial. lf not all thecoefficients in the s"-Z row are equal to
. . zero, we may replace a\?l by a very small positive number ¡; and continue to
complete Table 8-1. In this case it can be seen that sorne Ci.¡ will be negative.
~ \

l X:-¡ If sorne Ci.¡ is negative, at least one root of the modified D(s) (since abZ ) is replaced
by ¡¡) has a positive real part. Now the roots of a polynornial are continuous
functions of its coefficients. Hence, as ¡¡--->O, at least one root of D(s) has a
XII 1J positive or zero real part. Q.E. D.
(8-43)
]x

g that it is controllable *8-6 Linear Time-Invariant Discrete-Time systems


lal to the degree of the
tic polynomial of the The stability concepts introduced for the cootinuous-tirne systems are directly
'heorem 6-2). Now we applicable to the discrete-timecase. However, the conditions of stability are
e asyrnptotically stable. quite different. In this section we shall discuss sorne of these conditions.
420 STABILITY OF LINEAR SYSTEMS

Consider a relaxed linear time-invariant di serete-time system described by Table 8.2 Discrete-tim

k
y(k) = ¿ g(k - m)u(m) (8-46) Cl~) a\O)
m=O _) kOCl~,O) koa~,o~ 1

Then for any bounded-input sequence {u(k)}, (that is, there exists a finite h Clbl ) a\l)

such that lu(k)1 < h for k =0, 1,2, ...), the output sequence {y(k)} is bounded, if -)klCl~,1~l kla~IJ.2
2 a\2)
and only if Clb )

C()
a~l-l) aY'-l)
¿ Ig(k)1 < ca (8-47)
_)k"_laY'-I)
k=O
ag')
that is, {g(k)} is absolutely summerable. 8 The proof of (8-47) is similar to the
continuous-time case and is left as an exercise. The z-transform of (8-46) yields
Y(z) = g(z)u(z) is the difference of its t
Ir g(z) is a rational function of z, then the system is BlBO stable if and only if all by the same procedu
the poles of g(z) have magnitudes less than 1, or equivalently, aH the poles of define (Xi = agJ, i = 0,1.
g(z) lie inside the unit circle of the z planeo This can be readily proved by
noting the z-transform pair Theorem 8-21
All the roots of D(z) i
k =0,1,2, ... <:> g(z) =fl[g(k)] =~ numbers (Xi ~ ag), i = 1
z-b
where bis a real or a complex number. lf Ibl < 1, then We shall prove thi:
discrete-time systenis.
ical equation

Otherwise, it diverges.
lf g(z) is irreducible, the poles of g(z) are equal to the roots ofits denominator.
lf the degree of the denominator is three or higher, the computation of the roots The concepts of equi
is complicated. We introduce in the following a method of checking whether asymptotic stability ~
or not aH the roots of a polynomial are inside the unit circle without computing Theorem 8-14, every ~
explicitly the roots. The method is a counterpart of the Routh-Hurwitz only if aH the eigenvall
criterion. \vith magnitudes eq ua
Consider the polynomial with real coefficients Similar to Theorem 8
stable if and only if al1
D(z) = aoz n +alz"- 1 + ... +an-1z +a ll ao>O (8-48)
Lyapunov theorem fOl
We define ala) = ai, i = 0, 1, ... , n, and form the table in Table 8-2. The first
row is just the coefficients of D(z). The constant k o is the quotient of its last Theorem 8-22
and first elements. The second "row is obtained by multiplying k o on the first
row, except the first element, and then reversing its order. The third row All the eigenvalues of 1
positive definite herm
hermitian matrix N wi
8 An absolutely integrable function is neither rrecessarily bounded nor necessarily"approaches zero as
(-+ 00as shown in Figure 8-2. An absolutely summerable sequence however is always bOllnded
and approaches zero as k-+ oo. Thus the stability problem in the discrete-time case is simpler
than the one in the continuing time case. has a unique hermitiar
OISCRETE-TIME SYSTEMS 421

Table 8.2 Discrete-time Stability Table


le system described by

(8-46) abo l a\O) a~?~ 2 Q~IO~ 1 a(O)


"
_) koa~,O) koa~,o~ 1 koaS-°) koa\O)

there exists a finite h


ab )
1 a\l) a~,l~ 2 a~,l~ 1

- )kla~i~ 1 kla~1]2 k 1a\1)


:e {y(k)} is bounded, if
ab 2
) a\2) a~12~ 2 k 2= d,2] 2/a62 )

(8-47)

(8-47) is similar to the


nsform of (8-46) yields
is the difference of its two previous rows. The remainder of the table is obtained
by the same proeedure until n numbers {abl >, ab2l , ... , a~)} are obtained. We
stable if and only if aH
define (J.¡ = agl, i = O, 1, ... , n.
llentIy, all the poles of
be readily proved by
Theorem 8-21

z AH the roots of D(z) in (8-48) have magnitudes less than I if and only if the n
]=- numbers (J.¡ ~ ag), i = 1, 2, ... , n, computed in Table 8-2 are aH positive. I
z-b
We shall prove this after the establishment of the Lyapunov theorem for the
diserete-time systems. Consider the linear time-invariant diserete-time dynam­
ieal equation

x(k + 1) = Ax(k) + Bu(k)


y(k) =Cx(k)
,ots of its denominator.
mputation of the roots The eoneepts of equilibrium state, stability in the sense of Lyapunov and
d of eheeking whether asymptotie stability are identical to the eontinuous-time case. Similar to
ele without eomputing Theorem 8-14, every equilibrium state of x(k + 1) = Ax(k) is stable i.s.L. if and
)f the Routh-Hurwitz only if all the eigenvalues of A have magnitudes equal to or less than 1 and those
with magnitudes equal to 1 are distinet roots af the minimal polynomial af A.
Similar to Theorem 8-15, the zero state of x(k + 1) = Ax(k) is asymptotical1y
stable if and only if all the eigenvalues of A have magnitudes less than l. The
ao >0 (8-48)
Lyapunov theorem for the diserete-time case reads as:
ITable 8-2. The first
the quotient of its last Theorem 8·22
tiplying ko on the first
All the eigenvalues of A have magnitudes less than 1 if and only if for any given
lrder. The third row
positive definite hermitian matrix N or for any given positive semidefinite
hermitian matrix N with the property {A, N} observable, thematrix equation
ecessarily approaches zero as
however is always bounded
A*MA-M= -N
iiscrete-time case is simpler has a unique hermitiansolution M and M is positive definite. I
422 STABILITY OF LINEAR SYSTEMS

This theorem can be proved by defining

and computing
V(x(k» = x*(k)Mx(k) 1­
~ V(x(k»~ V(x(k + 1»
- V(x(k» = x*(k)A*MAx(k) - x*(k)Mx(k) 2 +
cxi _ [
= x*(k)(A*MA - M)x(k)
= - x*(k)Nx(k)

and is left as an exercise (Problem 8-36).

A Proof of Theorem 8-21 9

Figure S-8 A block di


Now we shall use Theorem 8-22 to prove Theorem 8-21. Define, for i =

0,1, ... , n. for i = 1,2, ... ,n. The


apply this diagram re
Di(z) =aglz/l- i +a~)z/l-i-¡ + ... +a~~ 1 k i =a~~Jag) (S-49)
Do(z)/Do(z) shown in
and Di(z) =a~i~ ¡Z"-i +a~i~¡_¡z"-i-l + ... +alilz +ag) (S-50) diagram, there is a tot,
where D¡(z) is the reciprocal of D¡(z). These polynomials can be defined re­ z- 1 as a state variable a
cursively by
D¡-I(Z)-ki-¡Di-¡(Z)
Di()
Z =-------­ i=I,2, ... ,n (S-51 ) with
Z

with Do(z) = D(z). Note that the coefficients of Di(z) are the a~¡), j =0,1, ... , n - i c=
defined in Table 8-2. It can be verified that the reciprocal of Di(z) can be
ex pressed as b=

i = 1, 2, ... , n (S-52) and

From Table 8-2, we have ag+ 1 ) =ag) - kia::~i = ag)(1-kf) or -kakI 1


-k ak 2 cx 2 /CJ. I -k
(i+ 1) IX
k2=1-~=I-~ (S-53) A= - k ak 3cx 3/CJ.¡ -k·
1ag) lXi

-kak,,_¡CJ./l_¡/CJ.¡ -k¡
Simple manipulation among (8-51), (8-52), and (8-53) yields -kaCJ..!cx¡ -k¡
- IX· ¡
Di· 1(z) = (zD ¡( z) + k i _ 1 D;(z) ---'-=-- (8-54 )

This state equation is ob1
- - IX· ¡
the dimension of A; henc
D i - ¡ (z) = (ki_¡zD¡(z) + D;(z)) -'-- (S-55 )
lXi F or this ma trix A, if we e
These recursive equations are valid for i = 1,2, ... ,n. Note that D/l(z) = D/l(z) =
a\í'l. Consider now the transfer function
D i - 1 (Z) k i - 1 zD i(z) + Di(z) (1 - kt_l)D¡(z) then it is straightforward
--- = = k· 1 +----.:...--.:...-=..:.~,..:­
D¡-I(Z) zD¡(z) +k i - 1 Di(z) 1­ zDi(z) +k i - 1 D i(z)
which becomes, by usiri.g.(8~53), A*MA-

Di-¡(z);"'k _¡ + IX¡Di(Z)/lXi-1D¡(z) By assumption, we have a


i (S~56)
Di_¡(z) z+ki-¡Di(z)jDi(Z) it has the property that
eigenvalues of A and,cons
9 This follows Rererence SIl!. if and only if M is positi ve (
OISCRETE-TIME SYSTEMS 423

- x*(k)Mx(k)

Figure S-S A block diagram of 15¡_ ¡(z)/D¡_I(z).

8-21. Define, l'or i = l'or i = 1, 2, ... , n. The block diagram ol' (8-56) is shown in Figure 8-8. lf we
apply this diagram repetitively, we can final1y obtain the block diagram ol'
<¡ = a~:~ Ja~) (S-49) Do(z)/Do(z) shown in Figure 8-9. Note that Dn(z)/Dn(z) = 1. In this block
+a~) (8-50) diagram, there is a total ol' n unit delay elements. lf we assign the output ol'
Z-I as a state variable as shown, then we will obtain the l'ollowing state equation
tials can be defined re­
x(k + 1) = Ax(k) + bu(k)
y(k) =cx(k) +kou(k)
!, ... , n (8-51) with
e = [1 O O .. , O O]
thea)i), j =0,1, ... , n - i
iprocal ol' Di(z) can be b=[~6
<xo
<xnJ'
<XO

and
2, ... , n (S-52)
-kok, 1 O O O
k"f) or - k ok 2rx 2/rx¡ -k¡k2 1 O O
- k ok 3rx 3 /rx¡ -k¡k 3rx 3/rx 2 -k 2k 3 O O
(S-53) A=
- kok n ­ ¡rx" - ¡/rx ¡ -k¡k,,_¡IY.,,_ Jrx 2 - k 2 k" _ ¡rx" - ¡/rx 3 -k n ­ 2 k,,-¡
-kolY.,jlY.¡ -k¡IY.,j1Y. 2 - k2rx ,jrx 3 -k Zcx,ja n ­
ll - 1
ields
(S-57)
(8-54·)
This state equation is obtained l'rom D(z)/D(z), and the degree 01' D(z) is equal to
the dimension ol' A; hence the roots ol' D(z) are identical to the eigenvalues ol' A.
- 1 (8-55) For this matrix A, if we choose M as

Note that Dn(z) = .5,,(z) = M=diag{~, ~, ... ,~}


al ()(2 ()(n

.kt- I)D¡(z) then it is straightl'orward to veril'y

A *MA'~ M = - N = - diag {~ , 0, O, ... , O}


) +k¡_ ¡Í)¡(z)
ao .'
By assumption, we haveao > O; hence N is positive seniidefinite. Furthermore
z) (S-56) it has the property that {A, N} is observable. Hence we conclude that aH
:(z) eigenvalues ol' A and, consequently, al1 rootsol' D(z)have magnitudes less than 1
il' and only il' Mis positive definite or, equivalently, ()(i > O, i = 1,2,. ., n. Q.E. D.
-====:===== . ..-._..._-
_.~---------------~._ ~ ------=-_==::::_~.::_c==c::.:_.:.=: ===. . . . . -=.---=­ =----=-===
~

8-7 Concluding
r-------------~

--------------------l
I In this chapter we in
I and the stability i.s.L
I
i
zero-input response.
+ I being uniformly stab

"" I distinetion between u

--------------, I neeessary and suffieie

I-~ -..-{ ,. I

I they ean hardly be e

I not available.

I
For the time-inva
I
I
I
the transfer funetion (
I
I

I a11 the eigenvalues of

I
the Routh~Hurwitzcrj
I

I
the Lyapunov theorer
I

I by using the Leverriel


I
I
I
eomputational errors.
I
I

I form by a numerica11:
I

I
I then be more easily Cl
I
I I
form is also used in tl
I
I I
I
I I
I Reference SI 07. One,

-;- ~I,,--..~ I-~~


I
tion needed in the ROl
I

: iT-
I i dJi 1
Lyapunov equation a
I
: L""r-J ... i ¡.~
1 I
; --:
1
-- I
I .
J-~
I

l .••
""-] : 1

Routh-Hurwitz methc.
in checking the stabili
the former may also b
The coneepts of s1
I I : _ 1;1;
: I :
1

I L J'Q<Q'I I
The stability conditio
I I I
I
function He inside the '
1 - I I
BIBO stable; whereas
I I
I

I 1 unit circle of the z pla


I -;:1-;: I

'--J relationship ean be es


I

I
I

. - _,Jo} I

L__---1~tr- I~I~:
I
I

, I
whieh maps the left~ha
I

I
may transform a dise
I

I
using the bilinear trar
I

I
IL
_ on diserete-time equati
I
------------ ...J
1

I
I ~
1 _o I
~
:
1
I ?:J
1

T
?:J ;1; :;
" I el Problems
.~ - - ~-- ~ ~~ ~ -l
o
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - u..
8-1 1s a system with th
How about g(t, r) = sin te
424

CONCLUDING REMARKS 425


I
8-7 Concluding Remarks
I----------l In this chapter we introduced the BIBO stability for the zero-state response
I : and the stability í.s.L. and asymptotie stability for the equilibrium state of the

I~:

zero-input response, For the time-varying case, a system may be stable withou t
being uniformly stable. For the time-invariant case, there is, however, no
distinction between uniform stabilities and (nonuniform) stabilities. Although
1----, I
I
i

necessary and sufficient eonditions are established for the time-varying case,
they can hardly be employed because state transition matrices are generally
II I I
not available.
I I
F or the time-invariant case, the stability can be checked from the poles of
I

the transfer function or fram the eigenvalues of the matrix A. Whether or not
I

I
all the eigenvalues of A have negative real parts can be cheeked by applying
I
the Routh-Hurwitz criterion to the characteristic polynomial of A or by applying
I
the Lyapunov theorem. The characteristic polynomial of A can be eomputed
I

I
by using the Leverrier algorithm (Problem 2-39), which however is sensitive to
I
computational errars. If the matrix A is first transformed into a Hessenberg
I
form by a numerieally stable method, the characteristic polynomial of A can
I

1
then be more easily eomputed. See Referenees S90 and S2l2. A Hessenberg
I
form is also used in the efficient method of solving the Lyapunov equation in
I

Reference S107. Once A is transformed into a Hessenberg form, the computa­


I

tion needed in the Routh-Hurwitz method is mueh less than that in solving the
I
Lyapunov equation and eheeking the positive definiteness of M. Henee the
I
Routh-Hurwitz method is simpler eomputationally than the Lyapunov method
I

in checking the stability of A. Although no comparison has been carried out,


the former may also be more stable numerically than the latter.
I
The concepts of stability are equally applicable to the diserete-time case.
I

I
The stability eonditions, however, are different. lf all the poles of a transfer
I
function lie inside the open left-half s plane, then the continuous-time system is
I

I
BIBü stable; whereas if all the poles of a transfer function lie inside the open
I
unit circle of the z plane, then the discrete-time system is BIBO stable. Their
I
1"
relationship can be established by using the bilinear transformation
:~
! \--1 .2- 1

¡ s =_._­
I

I 'Q z +1
lO
I

I
I 6
which maps the left-half s plane into the unit circle in the z planeo Although we
I
I f':
I
1 ~ may transform a diserete-time prablem into a eontinuous-time problem by
,;.o
~I~ I 1..><: using the bilinear transformation, it is simpler to check the stability directly

1
I~~:--.1_ _ I g
1:;5
on discrete-time equations.
l<e
I

1CJ'l

I ,

\~
IClO

~I~
~ ~ I :J Problems
IQ Q I en
___________ ...1 u.
8-' ls a system with the impulse responses g(l,r)=e-21t1-hl, for l'2:r, B[BO stable?
How about g(l. r)=sin te-(I-<) cos r?
426 STABlLlTY OF LINEAR SYSTEMS

8-2 Is the network shown in Figure P8-2 BIBO stable? Ir not, find a bounded input that 8-11 Prove parts ! ane
will excite an unbounded output. systems describable by r

ti 1F
I
Y
8-12 Consider the dyn,

~
Figure P8-2

8-3 Consider a system with the transfer function g(s) that is not necessarily a rational Y'
function of S, Show that a necessary condition for the system to be BIBO stable is that Is the zero state asympto
19(s)1 is finite for all Re s ~ O. stable? Is the equation 1

8-4 Consider a system with the impulse response shown in Figure P8-4. Ir the input 8-13 Consider
u(t) = sin 2nt, for t ~O, is applied, what is the waveform of the output? After how many
seconds will the output reach its steady state?

g(t}

Find all the equilibrium


Is it asymptotically stable'
stable?
o

Figure P8-4 8-14 Check the BIBO st

8-5 Is a system with the impulse response g(t) = 1/(1 + t) BIBO stable?

8-6 Is a system with the transfer function g(s) = e-s/(s + 1) BIBO stable?
8-15 Find the rangcs of I
8-7 Use the Routh-Hurwitz criterion to determine which of the rollowing polynomials
are Hurwitz polynomials.
a. S5 +4s 4 + !Os 3 +2,,2 +5, +6
b. SS +S4 +2s 3 +2s 2 +5s +5 is BIBO stable.

c. -2s4 -7s 3 -4s 2 -5s-10


8-16 It is known that the

8-8 Can you determine without solving the roots that the real parts of all the roots of
S4 +14s3 +71s 2 +154s + 120 are smaller than -1? (Hint: Lets=s' -1.)

8-9 Give the necessary and sufficient conditions for the following polynomials .lo be
Hurwitz polynomials:

.a. aos 2 +a¡s +a2


b. aos 3 +a¡s2 +a 2s +a3
is BIBO stable. Can we e

8-10 Find the dynamical-ecjuation description of the· network shown in Problem 8-2.
Find the equilibrium states of the equation. Is the equilibrium state stable in the sense of 8-17 Showthatx=Oísti

Lyapunov? Is it asymptotically stable? t --> OO.


PROBLEMS 427

nd a bounded input that 8-11 Prove parts 1 and 2 ofTheorem 8-3 by using the Laplace transform for the c1ass of
systems describable by rational transfer functions.

8-12 Consider the dynamical equation


1 O O

X-l O
-1
¡ -1
O
O
O
-2
1
O
O
O
1
O
- 10
!}+W
-4 -IJ
y=[ O O O O OJx
ot necessari\y a rational Is the zera state asymptotically stable (for the case u =,O)? Is its zero-state response BIBO
o be BIBO stable is that stable? Is the equation totally stable?

igure P8-4. If the input 8-13 Consider


utput') After how many

)"=[1 1Jx

Find all the equilibrium states Of the equation. Is every equilibrium state stablc í.s.L.?
Is it asymptoticaIly stable? Is its zero-state response BlBO stable? Is the equation totany
stable?

8-14 Check the BlBO stability of a system with the transfer function
25 2 -1
stable?

o stable')
8-15 Find the ranges of k¡ and k 2 such that the system with the transfer function
he foIlowing polynomials 5 +k¡

is BIBO stable.

8-16 It is known that the dynamical equation

iI parts of a1\ the roots of


s=s'-I.)

,owing polynomials to be

is B1BO stable. Can we conclude that the real part of A is negative? Why?
:k shown in Problem 8-2.
state stable in the sense Of 8-17 Show that x = O is the only solution satisfying (<1>(t, to) - I)x =,0 if 1I <1>(t, to)"":"O as
t---+OCi.
428 STABlllTY Of LINEAR SYSTEMS

8-18 Consider the following linear time-varying dynamical equation:

E : * =2ex +u
y=x
Sho\\l that the zero state of E is nol stable i.s.l.(under the zero-input response).

8-19 Consider the equivalent equation of E in Problem 8-18 obtained by the equivalence
transformation.;; = PU)x. where P(C) =e- r ': Figure P8-26
E: .\: =(2ee- r' _2ee- r'jer'.;; +e-r'u =0 +e-r'u
y=el:!.x

Show that the zero state of Eis stable i.s. L. (under the zero-input response). From Problems 8-27 Consider a discrete-t
8-18 and 8-19. we conclude that (//1 equivalelJce eralJsforll1ac ionlleecl'lOe preserre che scabilitr
af che zero scaee. Does an equivalence transformation preserve the BI BO stability of the
zero-state response?
Show that any bounded-in
8-20 Show that stability i.s.L. and asymptotic stability of the zero state of Ji: = A(C)x are if and only if
invariant under any Lyapunov transformation (see Definition 4-6). Is the transformation
P(c) = e-/ in Problem 8-19 a Lyapunov transformation?

8-21 Show that if Ji: = A(t)x is stable i.s. L. at ca' then it is stable i.s. L. at every tI "2: ca·
8-28 Consider a discrete-
[Hinl: Use <1>(1, e¡)= «1>(1,10)<1>-'(1" lo) and note the boundedness of <I>-I(e" to) for any
finite 1, and 10.J

8-22 Consider a sy'stem with the following dynamical-equation description:


Show that any bounded-ii

x{i
-1
1

O
if and only if

y=[ 2 3 'J x
Is the zero state asymptotically stable? Is the zero-state response BIBO stable? Is the 8-29 Consider a system \
system T-stable?

8-23 Prove Theorem 8-13.

Show that the system is B


8-24 Prov·~ condirion 3 oí' Thcorerr. ~~-1 9. 1-] inL. U:)(; f/rr :==

8-25 Consider the linear time-invariant controllable state equation

and
X =Ax +Bu

Show that if u = - B*W - '(T)x, where 8-30 Prove Corollary 8­

W(T)= I T
e-A'BB*eAo'dr T is an arbiti-ary ,positive number
have proper rationa\ tran

8-31 Is the function


then the overall system is asymptotically stable. Furthermore, V(x(c)) = x*(e)W- '(T)x(e)

is a' suitable Lyapunov function for the closed-loop system.

8-26 Are the networks shown in Figure P8-26 totally stable? AIJSlrers: No; yeso

positive definite or semi¿


PROBLEMS 429

uation:

)-input response).

btaíned by the equivalence


Figure P8-26

'esponse) From Problems 8-27 Consider a discrete-time system that is described by


el! /101 preserve ¡/¡e slabilily
: the BI80 stability of the y(n) = I g(n, m)u(m)

Show that any bounded-input sequenee {u(n)} excites a bounded-output sequenee {y(n)}
zero state of x = A(I)x are if and only if
-ó). ls the transformation
for all n
m=-oo

ible i.s.L. at every 1 1 ;:>: lo.


less of <I>-I(ll, lo) for any 8-28 Consíder a diserete-time system that is deseribed by

y(n) = I g(n - m)u(m)


m=O
>n description:
Show that any ~ounded~input sequence {u(n)} ~xeites a bounded-output sequenee {y(n)}
if and only if

I Ig(m)1 ~k < 00
m=O

mse BIBO stable? Is the 8-29 Consider a system with the impulse response

g(l)=g¡(I) + I C/¡o(l -T¡)


i=O

Show that the system is BIBO stable ifand only if

ltion
r 19¡(I)1 dI ~kl < 00
and I la,J~k2 < 00
i=O

8-30 Prove Corollary 8-3by usíng partía\ fraetion expansion for the class of systems that
have proper rationa\ transfercfunetion deseriptions.
;itive number
8-31 Is the funetion

[Xl Xl X 3] [~ ~ n6J r~0


AnslI'ers: No; yeso
2 2 G~
positivedefinite or semidefinite? [Hinl: Use Equation (E-2) in Appendix E.]
430 STABILlTY üF LINEAR SYSTEMS

8-38 Consider the inI.


8-32 Which of the following hermitian (symmetric) matrices are positive definite or
positive semidefinite?

a.

[¡ ~]
O
3
1
b.


O
O
O ~]
[",a, and W = diag [W I , W 2 ]

[! ~
c. O d. alaz

O
1 aloz
a l a3
",a:l
ozaz aZ a 3
aZ a 3 a3 a 3
W 2 ha ve no diagona [ er

--'

where a¡, i = 1,2,3 are any real numbers.


for i = 1, 2. are asympt
8-33 Let Al = - 1, Az = - 2, A3 = - 3 and let a 1> az, a3 be arbitrary real numbers, not References SI62 and SI
zero. Prove by using Corollary 8-20 that the matrix
8-39 Consider tbe poi:
af alaz ala3
---- ----
2A 1 }'I +},z Al + )'3
,
azal a2 aZ a 3
M= - - - - ---­
},z + Al 2A 2 +
Az }'3

a3 a l a3 a Z a 3Z
---- ----
Al + A3 },z + }'3 21 3 and the matrix A in (8­
Verify that Ac=Q-1AQ
is a positive definite matrix. [Hint: Let A = diag (A ¡, Az, Az).J

8-34 A real matrix M (not necessarily symmetric) is defined to' be, as in Definition 8-6,
positive definite if x'Mx > O for all nonzero x in IR". Is it tme that the matrix M is positive
definite if all the eigenvalues of M are positive real or if all the leading principal minors are
positive? If not, how do you check its positive definiteness? Hil1l: Try
(Due to Y. P. Ham)

8-40 Consider tbe systc


y¡= C;x¡+ Eill i . Use lb
scription of the system \'
8-35 Let M be a hermitian matrix of order 11. Let e¡, i = !, 2.. .11, be a set of linearly is i.he ¡TuHr\j; A, of tb..i.s ór:
jndependent vectcrs. ,,~. Li ' .:~ i:;:;';;;
v\lhmls LOe eondlLlon for
definite? (Answer: False.)
that lhe system be BIBO

8-36 Prove Theorem 8-22.

l
8-37 Determine the asymptotic stability of the discrete-time equation

~
O I O

. O O 1
. O

x(k + 1):= O O O .O
l ' x(k)

0.008 0.008 -0.79 -0.8 .

(1) by computing itsch~racterislic polynomial and the'n' ap'plyingTheo;:em 8"21 and (2)
Figure P8-40
by solving.the Lyapunov equation in Theorem 8-22.
PROBLEMS 431

8-38 Consider the internally balanced system in Problems 6-23 to 6-25 partitioned as
are positive definite or
I2 J[X 1J+[8 IJU
[ x~lJ
2
= [Al!
A 21
A
A 22 _ x2 82

y = [CI C2] [::J


and W = diag {W [, W 2}' Show that if the eguation is asymptotically stable and ifW 1 and
W 2 have no diagonal entries in common. then the subsystems
Xi =A¡¡x¡ +8¡u
y¡=C¡x¡
for i = 1, 2, are asymptotical1y stable. This result is useful in system reductíon. See
References S 162 and S 176.
trary real numbers, not
8-39 Consider the polynomial in (8-48) with the following companion matrix

~~[~ -~J
I O
'3 O 1

O O
'3 -a ll -Q"-l -Q"-2

and the matrix A in (8-57) obtained from the discrete-time stability table in Table 8-2.
Verify that A c = Q - [AQ or QA c = AQ, wherc

Q~[l
all) a~) a!i1 2
ab11 a\2) a~1.!3 al.".
a~12J 2 ]
be, as in Definition 8-6,
the matrix M is positive O O
ag,-l) aY,:-I)
ling principal minors are O O O ag')

int: Try
(Due to Y. P. Harn.)

8-40 Consider the system shown in Figure P8-40, where S¡ is described by Xi = Á¡Xi+ R¡u¡,
y¡ = c,x¡ + E¡u¡. Use the composite state x' = [X'1 x~J to develop a sta te variable de­
scription of the system with [r'l r2]' as the input and [Y'I y'l u'¡ U'l]' as the output.
. ,n, be a set of linearly 15 the matrix Á of this descriplion the salTle ':\s the ane in the tjm~-inwlr¡ajll. c,:,.se of 1}.fA\')
. ,11,then pv:iI is posilive What is the condition for the system to be asymptotically stable'! Will thecondition ensure
that the system be BIBO stable from any input-output, pair?

latíon
f[ + .u¡ YI
SI
~
I I

(k) Y2 u2 + .. f2
S2
+

Ig Theorem 8-21 and (2) Figure P8-40


(al

9 Figure 9-1 (a) Unity


Linear Time-I nvariant
Composite Systems: up from three basic e
Hence we shal1 restfi.
Characterization, Stability, aH the subsystems tha
it is easy to show th.
and Designs system. For linear CI
can be directly appliel
linear time-inyariant :
and a transfer functio!
feedback connection,
9-1 Introduction output stability of the
many questions can:
The design of control systems can be formulated as fol1ows. Giyen a plant, ample, what is the il
design an oyerall system to meet certain design objectiyes_ Because of the 91(s)andg 2 (s)? Is it­
presence of noises, the compensators are required to have proper rational from 91 and 92 witho¡
functions or matrices. In order to reduce sensitiYity due to the plant yariations questions for the sir
and load disturbances, the configuration of the oyerall system must be of feed­ answered. Before prc
back or closed-loop type. There are many possible feedback configurations.
In this chapter we study only the two configurations shown in Figure 9-1. The Definition 9-1
one in Figure 9-1(a) will be called the unityfeedback system. The one in Figure
A system is said to be
9-1(b) will be caHed the plant input-outputfeedback system or, simply, the input­
matrix if and only if t
output feedback system,! for the feedbacks are introduced from both the input
trollab!e and observal
and output of the plant g(s). This configuration arises from the c'Jr.nbiD.<tUcr
of the state-feedback and state-estimator studied in Chapter 7. Before pro­
The motivation of
ceeding, the reader should review the time-inyariant part of Section 3-6.
description of a systerr
A control system is, as can be seen from Figure 9-1, basically a composite
function matrix of th
system. By a composite system, we mean that the system consists of two or
controllab le from the
more subsystems. Rather than plunging directly into the design problem, we
function does not desc:
study first sorne basic problems associated with composite systems. We
equation description
assume thatthere is no loading effect in any connection of two suhsystems;
information obtained
that is, the transfer function of each subsystem remains unchanged after the
- transfer function of ti
connection (see Reference S46). - ­
case the system is saic
Although there are many forms of composite systems, theyare mainly built - matrix.
- Eyery dynamicale.
1 This terminology is by no means universal. Another possible name is the Luenberger-- or state­
one; heilce when we a¡:
estimator-type configuration_
what its dynamical-eq
432
INTRODUCTlON 433

Ca)

Cb)

Figure 9-1 (a) Unity feedback system. (b) lnput-output feedback system.

up from three basic connections: parallel, tandem, and feedback connections.


Hence we shall restrict ourselves to the studies of these three connections. Ir
aH the subsystems that form a composite system are alllinear and time-invariant,
it is easy to show that the composite system is again a linear time-invariant
system. For linear composite systems, all the results in the previous chapters
can be directly applied. For example, consider the feedback connection of two
linear time-invariant systems with a transfer function g1 (s) in the forward path
and a transfer function gz(s) in the feedback path. Ir the transfer function of the
feedback connection, g¡(s) = (l +g 1(s)gZ(S))-l g 1(s), is computed, then the input­
output stability of the feedback system can be determined from g¡(s). However,
many questions can still be raised regarding the composite system. For ex­
ample, what is the implication if there are pole-zero cancellations between
:>ws. Given a plant,
gl(S) and gz(s)? Is it possible to determine the stability of the feedback system
ves. Because of the
from g 1 and gz without computing g¡(s)? In the first part of this chapter, these
lave proper rational
questions for the single-variable as well as multivariable systems will be
:> the plant variations
answered. Before proceeding, we introduce a definition.
;tem must be of feed­
lback configurations.
n in Figure 9-1; The Definition 9-1
1. The one in Figure
A system is said to be completely characterized by its rational transfer-function
or, simply, the input­
matrix if and only if the dynamical-equation description of the system is con­
from both the input
trollable and observable. ;;n
'om the combination
Lpter 7. Before pro­
The motivation of this definition is as follows. Ir the dynamical-equation
:>f Section 3-6.
description of a system is uncontrollable and/or unobservable, then the transfer­
>asically a composite
function matrix of the system describes only the part of the system which is
m consists of two or
controllable from the input and observable at the output; hence the transfer
~ design problem, we
function does not describe the system fully. On the other hand, ifthe dynamical­
posite systems. We
equation desciiption of a system is controllable and observable, then the
i oftwo subsystems;
information obtained from the dynamical equation and the one from the
unchanged afier the
transfer functionof the system will be essentí:illy the same. Hence, in this
case the system is said to be completely charact~rized by its transfer-function
they are mainly built
matrix.

the Luenberger- or state-


a
Every dynamícal equation can be reduced to controiiab](~ and observable
one; hence when we apply Definition 9-1 to a composite system, we m,ist c1arify
what its dynamical-equation description is. Let Si, i = 1,2, be two subsystems
434 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS COMPLETE (

with state vectors Xi, i = 1, 2. Then the state vector of any connection of SI It is assumed that th
and S2 will be defined as x' = [X'1 x~]. With this definition, there will be no transfer-function m[
confusion in the state-variable description of any composite system. The equations (9-1) are c
state-variable descriptions of the tandem, parallel, and feedback connections that the transfer-fun
of two systems are derived in (3-62) to (3-64). (;1(S) +G 2 (s); the tI
All the subsystems that form a composite system will be assumed to be followed by S2 is (
completely characterized by their transfer-function matrices. This assumption, connection of SI W'
however, does nat imply that a composite system is completely characterized ([ + Gl (S)G 2 (S))-IGi
by its composite transfer-function matrix. In Section 9-2 we study the con­ the systems SI and S
ditions of complete characterization of composite systems. For single-variable function G(s) comple
systems, the conditions are very simple; ifthere is no common pole in the parallel
connection, or no pole-zero cancellation in the tandem and the feedback con­ Example 1
nections, then the composite system is completely characterized by its transfer
Consider the paralle
function. In Section 9-3 we extend the results of Section 9-2 to the multi­
whose dynamical-eql
variable case, with the condition of pole-zero cancellation replaced by the
condition of coprimeness. In Section 9-4 we study the stability problem of
composite systems. The stability conditions are stated in terms of the transfer
matrices of the subsystems. The remainder of the chapter is devoted to the and
design problem. In Section 9-5 we study the design of compensators in the
unity feedback system to achieve arbitrary pole placement and arbitrary
Their transfer functio
denominator matrix. The problem of pole placement and that of arbitrary
denominator are identical in the single-variable case. They are, however,
different in the multivariable case; the compensator required for the latter is
much more complicated than that required for the former. In Section 9-6 we
The composite transf(
design robust control systems to achieve asymptotic tracking and disturbance
rejection. The static decoupling problem is also discussed; both robust and
nonrobust designs are considered. In the last section, we study the design of
compensators in the input-output feedback system. The results are more
1t is clear that g(s) = 1
general than those obtained in the state-variable approach discussed in
beca use g(5) does not
Chapter 7.
be checked from the
The references for this chapter are S2, Sil, S19, S34, S35, S4Ü, S49 to S51, S54,
nection, we have u I =
S55,S64toS66, S75,S81,S85,S93, S94,S98,SI74,SI85, SI99,S218,S237,and
equation is
S23?,

9-2 Complete Characterization of Single-Variable

Composite Systems

It is easy to check Ü
observable; hence frOl
Consider two systems Si> for i = 1,2, with the dynamical-equation descriptions
charaderized by its ce
x¡=Á¡x¡ +B¡"¡ (9-1 a)
y¡ =C;x¡ + E¡"¡ (9-1b) In the. following
transfer functions co
where Xi, U¡, and y¡ are,· respectively,the state, the input, and the output of the
problem directly fre
system Si. A¡, B¡, C¡,.and ·E¡ are real constant matrices. The transfer-function
dynamical equations.
matrix of S¡ is
function matrix, then
(9-2) of the systemis equal t
COMPLETE CHARACTERIZATlON OF SINGLE-VARIABLE COMPOSITE SYSTEMS 435

my connection of SI lt is assumed that the systel'Ils SI and SJ. are completely characterized by their
tion, there will be no transfer-function matrices G l (s) and G 2 (s); 01', equivalently, the dynamical
lposite system. The eq uations (9-1) are controllable and observable. It was shown in Section 3-6
eedback connections that the transfer-function matrix of the parallel connection of SI and S2 is
6 1(s) +6 2(s); the transfer-function matrix of the tandem connection of Si
ill be assumed to be followed by S2 is 6 2(5)6 1(5); the transfer-function matrix of the feedback
es. This assumption, connection of SI with S2 in the feedback path is 6 1(5)(1+6 2 (5)(;1(5))-1=
lpletely characterized (1 +6 1 (s)6 2(s))-1(;I(S). Although (;1(S) and (;2(S) completely characterize
-2 we study the con­ the systems SI and S 2, respectively, it does not follow that a composite tnmsfer
. For single-variable function G(s) completely ~haracterizes a composite system.
on pole in the parallel
nd the feedback con­ Example 1
terized by its transfer Consider the parallel connection of two single-variable systems SI and Sl
.on 9-2 to the multi­ whose dynamical-equation descriptions are, respectively,
tion replaced by the
stability problem of FE1: Xl =X l +Ul
1 terms of the transfer Yl =X 1 +Ul
lter is devoted to the and FE!: X2=X2- U2
compensators in the Y2 =X2
ement and arbitrary Their transfer functions are
and that of arbitrary
They are, however, A s A -1
91(S)=-­ and 92(S) =-1
uired for the latter is s-1 s-
;r. In Section 9-6 we The composite transfer function of the parallel connection of SI and S 2 is
:king and disturbance
sed; both robust and A A 5 -1
g(s) =91(S) +g2(S) = - - +- = 1
le study the design of 8-1 s-1
fhe results are more It is clear that g(5) = 1 does not characterize completely the composite system,
,proach discussed in because g(s) does not reveal the unstable mode e' in the system. This can also
be checked from the composite dynamical equation. In the paral1el con­
5, S40, S49 to S51, S54, nection, we have u 1 = U 2 = u and y = y 1 + y 2; hence the composite d ynamical
SI99,S218,S237,and equation is

l~:J=l~ ~Jl::J +l-~Ju

y=[1 lJx +u
Variable
It is easy to check that the composite equation is not controllable and not
observable; hence from Definition 9-1, the composite system is not completely
·equation descriptions characterized by its composite transfer function. I
(9-1 a )
(9-1b) In the followíng we shall study the cOl)ditions under which composite
transfer functioIls completely describe composite systems. We study this
and the output of the problem directly from transfer-function. matrices without looking into
The transfer-function dynamicalequations. If a system is c6rilpl~tely chara~ierized by its transfer­
function matrix, then the dimension. of the dynamical-equation description
(9-2) of the system isequal to the degree of its transfer-function matrix (Theorem 6-2).
COMPLETE (
436 LINEAR TlME-INVARIANT COMPOSITE SYSTEMS

Therefore, whether or not a system is completely characterized by its transfer­ 1. The parallel conr
function matrix can be checked from the number of state variables of the 91(S) +92(S) ifano
system. Ir a system is an RLC network,2 then the number of state variables is 2. The tandem conr
equal to the number of energy-storage elements (inductors and capacitors); 92(s)9 1 (s) if and 01
hence an RLC network 2 is completely characterized by its transferlunction 3. The feedback cor
matrix if and only if the number oi energy storage elements is equal lO the degree characterized by !
oi its transJer-Junct ion matrix. Consider now two RLC networks SI and S2, canceled by any z
which are completely characterized by their -transfer-function matrices Gl (s) Pfoof
and G2(s), respectively. The number of energy-stora~e elemellts in any c~m­
posite connection of SI and S2 is clearly equal to 15G l (s) +15G 2(s). Let G(s) 1. It is obvious tha
be the transfer-function matrix of the composite connection of SI and S2. 15 9 < bg l +1592. 1
Now the composite s),stem consists of(15 G1(s) +15G 2(s)) energy-storage elements;
hence, in order for G(s) to characterize the composite system completely, it is
necessary and sufficient to have 15G(s) =15G l (s) +15G 2(s). This is stated as a
theorem. We show now tha
15 91 +(592. We pr
TheQrem 9-1 there is at least o
there is a comm(
Consider two systems SI and S2, which are completely characterized by their assumption that ti
proper transfer-function matrices Gl (s) and G2(s), respectively. Any composite there is a common
connection of SI and S2 is completely characterized by its composite transfer­ tion that 91 is irre(
function matrix G(s) if and only if in common, then
il connection of S 1 a
2. The proof of this 1
.... 3. The transfer funct
This theorem can also be verified from the dynamical-equation descriptions
of systems. Recall from Section 3-6 that the state space of any composite
connection of SI and S2 is chosen to be the direct sum of the state spaces of
SI and S2; consequently, the dimension of the composite dynamical equation
is the sum of the dimensions of the dynamical-equation descriptions of SI and By the irreducibili
S2- Hence Theorem 9-1 follows directly from Definition 9-1 and Theorem 6-2. have D 2 and N 2.
In order to apply Theorem 9-1 we must first compute the transfer-function if and only if D 2 al
matrix of a composite system. This is not desirable, particularly in the design
of feedback control systems. Hence the conditions in terms of G1 and Gzfor G Example 2
to characterize completely the composite connections of SI and S2 will be Consider the tandem
studied. We study in this section only single-variable systems. The multi­
variable systems will be studied in the next section.
The transfer function of a single-variable system is a scalar, and its degree
is just the degree of its denominator if the transfer function is irreducible. We as shown in Figure 9-.
assume in this section that all transfer functions are irreducible; that is, their
denominators and numerators are coprime.

Theorem 9-2

Consider two single-variable systems SI and S2, which are completely charac­
terized by their proper rational transfer functions 91(S) and 92(S).

Figure 9-2 A single-vi


2 We assume thal there are no capacitors-only loops and inductors-only cutsets in the network.
COMPLETE CHARACTERIZATlON OF SINGLE-VARIABLE COMPOSITE SYSTEMS 437

1. The parallel connection of SI and S2 is completely characterized by 9(5) =


:rized by its transfer­
91(5) +92(5) if and only if gl(5) and 92(5) do not have any pole in common.
:ate variables of the
2. The tandem connection of SI and 52 is completely characterized by 9(5) =
r of state variables is
92(5191 (5) if and only if there is no pole-zero cancellation between 91 and g2.
ors and capacitors);
3. The feedback connection of SI and 52 shown in Figure 9-2 is completely
it5 tran5ferjunction
characterized by 9(5) = (1 +9192) - 191 if and only if there is no pole oí 92(5)
i5 equal to the degree canceled by any zero of 91 (5 j.
networks SI and S2,
ction matrices (;1(5) Proof
lements in any com­ 1. It is obvious that if 91 and 92 have at least one pole in common, then
.) +<5G 2 (5). Let G(5) lJ9 < lJg 1 +lJ92' Let 9i = N ¡(Di' for i = 1, 2; then
:ction of SI and 52'
rgy-storage elements; A A A N I D2 +N 2 D I
g=gl+g2= .
;tem completely, it is DI D2
This is stated as a
We show now that if 91 and g2 do not have any pole in common, then lJg =
lJ91 +lJ92' We prove this by contradiction. Suppose 8g <8g 1 +8g 2, then
there is at least one common factor between N ID 2 +N 2D I and D ID 2. If
there is a common factor, say, between N ID 2 + N 2DI and DI' then the
haracterized by their assumption that there is no common factor between DI and D 2 implies that
vely. Any composite there is a common factor between NI and DI' This contradicts the assump­
s composite transfer- tion that 91 is irreducible. Hence we conclude that if g 1 and g2 have no pole
in common, then 9(5) = 91(5) +92(5) characterizes completely the paral1el
i
connection of SI and 52·
2. The prQofof this part is obvious and is omitted.
3. The transfer function of the feedback system shown in Figure 9-2 is
~quation descriptions
ce of any composite
of the state spaces of D 1 D 2 +N IN 2
: dynamical equation By the irreducibility assumption, DI and NI have no common factor, nor
lescriptions of SI and have D z and N 2 • Hence D 2 N] and D I D 2 +NIN z have common factors
;)-1 and Theorem 6-2. if and only if D 2 and NI have common factors. Q.E.D.
: the transfer-function
ticularly in the design Example 2
as of G1and (;,Z for G
of SI and S2 will be Consider the tandem connection of SI and S2 with transfer functions
systems. The multi- 1 5-1
and
5 -1 s + 1

scalar, and its degree


)n is irreducible. We as shown in Figure 9-3(a). There is a pole-zera cancellation between 91 and 9z·

:ducible; that is, their

re completely charac­
ldg z (5).

Figure 9-2 A single-variable feedback system.


y cutsets in ¡he netwof.k.
438 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS ce

On the other han


path of Figure 9-4(b)
back path, the transf<
terizes the feedback ~
(a)

Consider a specü
g2(S) = k, where k is a
by gI (s), the transfer I
(b)

Figure 9-3 Tandem connection of 5 1and 52' which ís not characterized completely by
=9291 = 1/(5 + 1).
g(5)
always characterizes 1
Hence the composite transfer function
A A A 1
g(s) = g2(S)g I(S) =--1
S+
9-3 Controllabil
does not completely characterize the tandem connection. This can be seen by of Composite Sy
applying a unit step input to the composite system; although the output of the
tandem connection is bounded, the output of SI increases exponentially with Let gj(s) = N¡(s)D j- I (s)
time, as shown in Figure 9-3(b). III N j and Di, Theorem (
SI and S2 is completE
Example 3 D 2 (s) are coprime. l
Consider the feedback connections shown in Figure 9-4. In Figure 9-4(a), the g2(S)g¡(s) if and only
pole of the transfer function in the feedback path is canceled by the zero of the coprime. In this secti
transfer function in the forward path. Hence the transfer function of the composite systems. B
feedback system does not completely describe the feedback system. Indeed, 6-8, on strict system ec
its transfer function is A system can be de
s-1 matrix in fractional fo
s+l ,')-1
g(5) = - - - - ­
s-1 1 s+2
1 +---­
s+ls-1 where .u and y are th,
Its degree is smaller than 2. descriptions describe t
If {A, B, C} is not irred
{D¡, NI}' and {pes), Q
strictly system equivale
system equivalent, and
For our problem, it tI
convenient. Hence th
We recall from (6-151
{pes), Q(s)} is left COpl
(a) (b) right coprime. These
Figure 9-4 Feedback systems. complete characterizat
.. .• _---_ .. _----­
_,._~
-------~._-~-_ .. _-~-~-_ .. --- ­

CONTROLLABIUTY AND OBSERVABIUTY Of COMPOSlTE SYSTEMS 439

On the other hand, although the pole ofthe transfer function in the forward
path of Figure 9-4(b) is canceled by the zero of the transfer function in the feed­
back path, the transfer function of the feedback system still completely charac­
terizes the feedback system. Its transfer function is
A s +1
lIlI
g(s) = (s _ 1)(s +2)

-l
Consider a special case of the feedback system shown in Figure 9-2 with
g2(S) = k, where k is a real constant. Since there is no pole in g2(S) to be canceled
g
by ¡(s), the transfer function

A
g(s)= [ A ] ¡A
1 +kg[(s) - g[(s)= N[k
acterized completely by D[ + N[
always characterizes the feedback system completely.

9-3 Controllabilityand Observability


This can be seen by of Composite Systems
ugh the output of the
~s exponentially with Let g¡(s) = N ¡(s)D¡-¡(s), i = 1, 2, and let N ¡(s) and D¡(s) be coprime. In terms of
I Ni and D¡, Theorem 9-2 can be stated as follows: The parallel connection of
S¡and S2 is completely characterized by g¡(s) +g2(S) if and only if D¡(s) and
D2 (s) are coprime. The tandem connection is completely characterized by
g2(S)g¡(S) if and only if D¡(s) and N 2(s) are coprime and D 2(s) and N ¡(s) are
In Figure 9-4(a), the
coprime. In this section we shall extend these conditions to the multivariable
led by the zero of the
composite systems. Before proceeding, the reader is advised to review Section
(lsfer function of the
6-8, on strict system equivalence.
,ack system. Indeed,
A system can be described by a dynamical equation {A, B, C, E}, a transfer
matrix in fractional forms C(s) = N,.(s)D,:-¡(s) =D¡-¡(s)NI(s), or a system matrix
r pes) Q(s) TS(s)l r [b!
L-R(s) w(stlL -íi(s)J = L-y(sU
where u and y are the input and output and ~ is the pseudostate. If these
descriptions describe the same system, they all have the same transfer matrix.
If {A, 8, e} is not irreducible (oot controllable or not observable) or if {D" Nr },
{D" NI}, and {pes), Q(s), R(s)} are oot coprime, then they are generally not
strictly system equivalent. Howevcr, if they are all irreducible, they are strictly
system equivalent, and any one of them can be used in the analysis and designo
For our problem, it turnsout that the use of the system matrix is the most
convenient. Hence the system matrix will be used extensively inthis section.
We recall from (6-151) aod (6-152) that {A,B} is controllable ifand only ir
{P(s), Q(s)} is left coprime; {A, C} is observable if aod only if {pes), R(s)} is
fight coprime. These properties will be used to establish the conditions for
complete characterization for the multivariable case.
440 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS
CON

Let G¡(S) = Nr¡(s)Dri ¡(S) = D¡¡ ¡(s)N¡¡(s) be, respectively, right and left co­
O,.¡(s) and D..z(s) are l.
prime fraction of (;¡(s). Then system Si can also be described by
nection is controllable
D,.¡(s)
[ - N,.¡(s) O
1][ - t(S)] [ O]
u¡(s) - -Yi(S)
i = 1,2 (9-3) Ir we use (9-4). then

where ~¡(s), Ui(S) and y¡(s) are, respectively, Pi x 1, Pi X 1, and qi x 1 vectors, or

D¡i(S) Nl¡(S)][ ~¡(S)J = [ O] A


i = 1,2 (9-4 )
[ -1 O -u¡(s) -Yi(S) By a similar argument i
wheres¡(s), u¡(s), and y¡(s) are, respectively, q¡ xl, p¡ x 1, and qi xl vectors. if and only if D l1 (s) a~(
results as a theorem.
Parallel connection. For the parallel connection shown in Figure 9-5, we Theorem 9-3
have P¡=P2,q¡=q2'U¡=U 2 =u, and Y=Y¡ +Y2, where u and y denote the
input and output of the overall system. Using these equations and (9-3), we Consider lwo systems 1
can obtain matrices G¡(s) and G2 «
~ ~

fractions of G¡(s). Ther

~~'~~s: I?;~~)__)] [t~;~-J


l-N,.¡(s) -N..z(s):O -u(s)
~
= [ __ __ ]
-y(s)
(9-5)
trollable if and only if 1
nection is observable if a

The roots of det O,.,(s;


to describe the parallel connection. Although (9-3) are irreducible by assump­ H). lf the poles of G¡I
tion, the system matrix in (9-5) is not necessarily irreducible. In order to obtain [D,.¡(s) Dds)] and the r.
a simpler result, we carry out the following strict system equivalence trans­ (why?). Hence a suffici<
formation: trollable and observable
This condition however
Ir G¡(s) and G 2 (s) ar~
the folIowing: The parall
only if their denominators
This transformation merely substracts the second block row of the system proof of statement 1 of Th,
matrix in (9-5) from the first block row. Since controllability and coprimeness 9-3, we have the necessary
are invariant under the transformation of strict system equivalence (Theorem terize completely the para
6-9), we conclude that the dynamical equation description of the parallel
connection is controllable if and only if Tandem connection.
sb.oyvn. in Figu.re 9··(; i;;I~ !~.
lJi)r\(S) ~ : n
l O O..z(s) : 1 J or (9-1Oi ) using these relations and I

has a full rank for every s in e Because of the block triangular form, the O"o(s)
second matrix has a full rank for every s in e if and only if the matrix
[D,.¡(s) - D..z(s)] has a full rank for every s in e or, following Theorem G-8',
r

- N,.¡(s)
----Ó-----=

SIZ
u
,-.,------------­
I
+

.'UI. 'YI Uz Y;

I .
s, . .
S,
­
IL _

Figure 9-5 The parallel connection of SI and S 2'


Figure gc6 S12. the tandern (
CONTROLLABIUTY AND OBSERVABILlTY OF COMPOSITE SYSTEMS 441

ly, right and left co­ O"I(S) and 0"2(S) are left coprime. Hence we conc1ude that the paraHel con­
'ibed by nection is controllable if and only if Or[ (s) and 0,.2(S) are left coprime.
lf we use (9-4), then the system matrix of the parallel connection is given by
1,2 (9-3)
= OII(S) O: NII(S)]

O 012(S) : N I2 (S)

and q¡ x 1 vector s, or [---------- •-- -- - (9-7 )


-1 -1: O
i = 1, 2 (9-4) By a similar argument, it can be shown that the parallel connection is observable
if and only if 011(S) and Dds) are right coprime. \Ve recapitulate the aboye
md q¡ x 1 vectors. results as a theorem.
Theorem 9-3
lown in Figure 9-5, we
'e u and y denote the Consider two systems which are completely characterized by their transfer
quations and (9-3), we matrices GI(s) and G 2 (s). Let G¡(s) = 0i;- l(s)Nl¡(s) = N,.¡(s)O,~ I(S) be coprime
fractions of (;;(s). Then the parallel connection of these two systems is con­
trollable if and only if O"I(S) and Ods) are left coprime. The parallel con­

j¡,l] (9-5)
nection is observable if and only if OII(S) and 012(S) are right coprime.

H).
Iil

The roots of det O"j.,(s) or det O¡¡(s) are ca)led the potes of G¡(s) (see Appendix
lf the poles of G I (s) and those of G 2 (s) are disjoint, then the matrix
irreducible by assump­
ible. In order to obtain [Orl(s) 0"2(S)] and the matrix [Oíl(S) D;2(S)]' have a full rank for every s in t:
;tem equivalence trans­ (why?). Hence a sufficient condition for the parallel connection to be con­
trollable and observable is that GI(s) and (;2(S) have no pole in common.
This condition, hüwever, is not a necessary condition (see Problem 9-4).
Or[(S)
O
-O'2(S): 0J

O,.z(s) : 1

If (;I(S) and (;2{S) are 1 xl rational functions, then this theorem reduces to
the following: The parallel connection is controllable and observable if and
=--Ñ:I(s) -- ~-Ñ'~2(;) : 0­ only if theír denominators have no roots in common. This provides a different
proof of statement 1 ofTheorem 9-2. By combining the conditions in Theorem
lock row of the system
9-3, we have the necessary and sufficient conditions for (;\(s) + (;2(S) to charac­
lability and coprimeness
terize completely the parallel connection of the two systems.
TI equivalence (Theorem
:cription of the parallel Tandem connection. For the tandem connection of SI followed by S2
shown in Figure 9-6, we have q¡ = P2, u(s) = M] (s), an 2 \s) = YI (s), aod y(s) = Y·/el.
~~2_(~)_ ~ ?-I (9-6)
Using lhese reiations and (9-3), we can obtaín
D'2{S) : lJ
ock triangular form, the
and only if the matrix (9-8)
following Theorem 0-8/,

Sl2
r----------~--l

I I

u IU¡ Y2Ly

I I

IL I

--J

Figure 9-6 Sil' the tandem connection of SI followed by Sl'


442 LINEAR TlME-INVARIANT COMI'OSITE SYSTEMS ce

to describe the tandem connection. Note that Yl(S)=U 2 (s) is a part of the fol1owing three pain
pseudostate of the tandem connection. It is clear that the tandem connection and N'2(S), or D,.¡(s) ,
is controllable if and only ir. for every s in e, the matrix
Let G;(s) be a q¡ x
o N'i(S)D,:¡ l(S)= Di! ¡(5)
O'2(S) (9-9)
called the po/es of G¡(
o '3!in (Pi, q J, where p ,
has a ful1 rank for every s in C. By adding the second block row to the third Gj(s) (see AppendixH
block row,3 (9-9) becomes
Corollary 9-4
O,,(s) O: O IJ A sufficient condition
O D'2(S) : 1 O
[ ------~-----~-~---- P"l ?q¡, q2 ?P2' and q
- Nrt (s) O,-z(.~)' O O
G2~S) is a transmissiol
which implies that the matrix in (9-9) has a full rank for every s in e if and only of G 2 (s)].
if [ - N,¡(s) D,z(sl] has a full rank for every s in C. Hence we conclude that
the tandem conneclion is control1able if and only if O'2(S) and N'.I(S) are, Proof
fol1owing Theorem G-8', left coprime.
Let (;l(S) = Nr¡(s)D,:¡¡

lt _
If we use (9-3) to describe SI and (9-4) to describe 52, then the tandem con­
polynomial matrix ane
nection is described by
P2 at evelY s except the

O,¡(s) O Ijr ~¡(S)j


O: O j zero of G ¡ (s), we have

~ ~'-: ~s~ _~I~~S) __ ~~t~ j_~ _-J:~:~ =


matrix [D,.2(s) N,.¡(sf
[ __ (9-10) coprime. Hence if no 1
O 1 O: O . - u(s) - y~s) connection of S ¡ .follolA
observability part can t
If we add the product of the second block row and N l2 to the third block row,3
then the system matrix in (9-10) becomes The combination (

Orl(s)
-NrI(s)
O
O
0:11
-1:0
Theorem 9-4 yields the
tandem connection. 1
Theorem 9-4 reduces te
o:
r - N I2 (S)N'.I(S) D I2 (S)
- - - - -ó-- - - - - - - i - ---- 0-:- ó
Hence we conc1ude that the tandem connection is controllable if and only if
O
is controllable (observaI
of g¡(s) [no pole of g¡(
characterizes complete!
N l2 (S)N r1 (s} and D l2 (S} are lefi coprir(~~:;. Ey!:1. :..:lrrülar ~lrgU;:.G'~;:;';'1 -,:"j~:. ~:;(.'~~'" :;:::'.":\;:: ~nd. (Y;;:!~i ~~ :.:~~~.-::; ):' ­

the fol1owing theorem. provides a different proc


The concept of pole­
Theorem 9-4 the matrix case if it is ca,
the tandem connection
Consider two systems Si which are completeIy characterized by their transfer <5G(s) = c5( G 2 (s)G¡ (s)) = <5
matrices G;(s), i = 1, 2. Let Gi(s) = Dií 1 (s)N¡¡(s) = N,i(S)D,:¡ l(S) be coprime tandem connection is eitl
fractions of G;(s). Then the tandem connection of SI fol1owed by 52 is con­
trol1able if and only if any one of the following three pairs of polynomial <5G(s
matrices, 0ds) and NrI(s), 0 11 (slDds) and N l1 (s), or O/2(s) and N¡z(s)N'I(S), are I
I - In this case, we may defir
left coprime. The tandem connection is observable if and onfy if apy one of the ! For example, if

) This is a-lransformation of slriclsyslem equivalence_ See lhe slalemenl al lhe end ofSeclion 6-8.
CONTROLLABILITY AND OBSERVABILITY Of COMPOSITE SYSTEMS 443

following three pairs of polynomial matrices, D 11 (S) and Nds), D/[(s)D.. 2 (s)
02(S) is a part of the
and Nds), or D,.¡(s) and Nds)N¡-[(s), are right coprime. Ií
he tandem connection
Let G¡(s) be a q¡ x p¡ rational matrix and have the coprime fractions G¡(s)=
N .. ¡(s)D,~ 1(S)= D/i 1 (s)Nu(s). Then the roots of det D ..i(s) or det Du(s) are
(9-9) called the poles of (;¡(s) and those s for whích pN.. ¡(s) < min (p¡. q¡) or pI"'Us) <
mín (p¡, qJ, where p stands for the rank, are called the tral1smissiol1 zeros of
(;¡(s) (see Appendix H).
block row to the third
Corollary 9-4
A sufficient condition for the tandem connection of SI followed by S2 with
PI "2:.Q1>Q2 "2:.P2' and ql =P2 to be control1able (observable) is that no pole of
(;2(S) ís a transmíssíon zero of (;1(S) [no pole of (;I(S) ís a transmíssíon zero
of (;2(S)].
every s in e if and only
lence we conc1ude that Proof
D"2(S) and N.. 1 (s) are,
Let (;1(S) = N¡-[(s)D,:¡I(S) and G2(s) = N"2(S)D,:;I(S), where N"1(S) is ql X PI
~, thcn the tandem con­ polynomial matrix and Dds) is a P2 x P2 polynomial matrix. D"2(S) has rank
P2 at every s except the roots of det D,As). If no pole of (;2(S) ís a transmissíon
zero of (; I(S), we have pN"I(S) = ql = P2 at the roots of det Dds). Hence the

Jl-~y(s)i 1J (9-10)
matrix [Dds) N r1 (s)] has rank P2 at every s in e and, consequently, is left
coprime. Hence ifno pole of (;2(S) ís a transmission zero of (;1(S), the tandem
connection of S1 followed by S2 is, fol1owing TheQrem 9-4, controllable. The
observability part can be similarly proved. ' Q.E. D.
to the third block row,3
The combination of the controll~bílitx and observability condítions in
Theorem 9-4 yields the condition for G 2 (S)G 1 (s) to completely characterize the
tandem connection. If (;1(S) and G2 (s) are 1 xl rational functions, then
Theorem 9-4 reduces to that the tandem connection of g1(S) fol1owed by g2(S)
is control1able (observable) if and only if no pole of g2(S) is canceled by any zero
of gl(S) [no pole of gl(S) is canceled by any zero of g2(S)]. Hence g2(S)gl(S)
characterizes completely the tandem connection of gl(S) followed by 92(S) if
ntrollable if and only if é1.nd only if there is no pole-zcro cancellgu.c~~.' '/";';:'1,''-:''': ..: ,~::; ( ". ~l.~í
; I ";'-".'

gument, we can establish provides a different proof of statement 2 of Theorem 9-2.


The concept of pole-zero cancellation in the scalar case can be extended to
the matrix case if it is carefully defined. If G 2 (S)G 1 (s) characterizes completely
the tandem connection in Figure 9-6, then we have, following Theorem 9-1,
:terized by their transfer <'5G(s) = <'5(G 2 (s)G 1(s)) = <'5G 2 (s)+ <'5G 1 (s), where <'5 denotes the degree. Ir the
~ri(s)D;¡ 1(8) be coprime tandem connection is either uncontrollable or unobservable, then we have
\ followed by S2 is con­ <'5 (;(s) = <'5(G 2 (s)G I(S)) < <'5.G 2(s) + <'5 G l (s)
Tee pairs of polynomial
lds) and Nds)Nrl(s), are In this case, we may define that there are polé-zera cancellations in G 2 (S)G 1 (s).
md only if anyone of the For example, if

. 'l'
~s-l
s
G 1 (o5)= _Í_ . j'
emenl al lhe end o[ Section 6-8. 05-1
444 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS Cal

then b(G 2(s)G I (S))=O < bG 2(s)+ bGI(s)= 1 and there is one pole-zero cancel­ Proof
lation in G 2(S)G I(s). Clearly, the existence of pole-zero cancellations in
We shall prove this ti'
G 2(s)G 1(s) does not imply the existence of pole-zero cancellations in GI (s)G 2(s),
ability. Let x be the e
as can be verified from the example. Unlike the scalar case, pole-zero cancel­
Xi is the state of Si' :
lations in the multivariable case may not involve actual cancellations of poles
composite state space
and zeros.
time. Let Y2 be the Oi
Let ~(s) and ~i(S) be, respectively, the characteristic polynomials of G(s)
S J, this input will tran
and G¡(s). Ir G(s) = G 2(S)G I (s) has no pole-zero cancellation, then we have
is SJ' Conversely, if
~(s)= ~1(S)~2(S), as can be easily seen from Equation (3-63a). Ir (;2(S)(;I(S)
uI=u-Y· HenceSJ;
has pole-zero cancellations, then deg ~(s) < deg ~ 1(s) + deg ~2(S), and the roots
Now we consider
of ~1(S)~2(S)/~(S) are called the cancelled poles. If A is a cancelled pole, then
S J due to sorne unkno
from Theorem 9-4, we have the following equivalent conditions:
if - Yo is applied to S
1. [DdA) N"l(A)] and/or [DNll(~))] have no full rank. S21 is Yo· If S21 is ob
output Yo, the initial sl
12(;"
able. lfS 21 isnotobs~
D ll V,)D,z(A)] output Yo. Consequer
2. [Dll(A)DdA) N1I(A)] and/or [ Nd},) have no ful1 rank.
its zero-input response
this theorem.
3. [DdA) NdA)N"I(A)] and/or [ N D"I(A)]
')N have no ful1 rank.
dA ,.¡(A)
In general, the COI
For a direct proof of this statement, see Reference S2. See also the discussion condition of observab
of the input-decoupling zero on page 292. If al1 cancelled poles have negative two conditions are ider
real parts, then G 2(S)G I (s) is said to have no unstable pole~zero cancellation. of 91(S), then 91(s) follo
is observable. Hence \\o
Feedback connection. Consider the feedback connection of SI and S2 connection in Figure <;
shown in Figure 9-7. It is assumed that S¡ is completely characterized by e¡(s). controllable or, equiva
It is also assumed that det (1 + el
(s)Gz(s)) =1= oat s = <Xl to ensure that the feed­ This checks with what '
back transfer matrix is wel1 defined and proper (Theorem 3-5). Let S 17. denote We consider a speci:
the tandem connection of SI followed by S2; $21, the tandem connection of le We cal1 this a consl,
S2 followed by SI; and SJ' the feedback syslem in Figure 9-7. any new state variable,
observable, so is K fol
Theorem 9-5 \Nith a constant gain ¡¡¡:,
ano only ii SI is con1
Consider two systems Si which are completely characterized by their transfer observability are invari,
)e
matrices (;¡(s), i = 1, 2. It is assumed that det (1 + GI( <Xl 2(00)) 1= O. Then with the constant state
the feedback system S J is controllable (observable) if and only if S 12 is con­ is preserved under com
trollable (S21 is observable). preserved.
We discuss further !
be described by irred uc
r . ­
,A i , Bi ¡ C¡, E;}, where G
u + ut
rationa,1 matrices. By a
~[G¡(sr
and deg G¡(s)"" ~
where ~[.] denotes the
Figure 9-7 Feedback conneclion of SI and S2' denotes equality of polyr
CONTROLLABIUTY AND OBSERV ABIUTY Of COMPOSITE SYSTEMS 445

one pole-zero cancel­ Proof


zero cancellations in We shal1 prove this theorem from the definition of controllability and observ­
el1ations in (;1(S)G 2(s), ability. Let" be t he composite state of SI and S2; that is, X = [X'I x~J', where
;:;ase, pole-zero cancel- Xi is the state of Sj. If S12 is controllable, then for any Xo and any XI in the
cancellations of poles composite state space, an input u I to SI2 exists to transfer Xo to Xl in a finite
time. LetY2 betheoutputofS 12 duetou 1. Nowifwechooseu=u l +Y2for
e polynomials of G(s) S f' this input will transfer Xo to x l' This proves that if 5 12 is controllable, so
:l1ation, then we have is Sf' Conversely, if Sf is control1able, corresponding to any u, we choose
(3-63a). Ir Gz(s)G I(s) "1=" - y. Hence 5 12 is controllable.
eg i!2(S), and the roots Now we consider the observability part. Let the zero-input response of
a cancelled pole, then Sf due to sorne unknown state "o be Yo. If S21 is in the same initial state and
lditions: if - Yo is applied to S21, then from Figure 9-7, we can see that the output of
S21 is Yo. If 5 21 is observable, from the knowledge of the input - Yo and the
output Yo, the initial state Xo can be determined. Consequently, S f is observ­
able. If S21 is not observable, we cannot determine Xo from its input -Yo and
output Yo. Consequently, we cannot determine the initial state "o of 5f from
full rank.
its zero-input response, and Sf is not observable. This completes the proof of
this theorem. Q.E. D.
) full rank.
In general, the condition of control1ability of 5 12 is different from the
See also the discussion candition of observability of S21' However, in the single-variable case the
ed poles have negative two canditions are identical. Indeed if no pole of 92(S) is canceled by any zero
.Ie-zero cancellation. of 91(s), then 91(s) followed by 92(S) is controllable, and 92(S) followed by gl(s)
is observable. Hence we conclude that, for the single-variable case, the feedback
connection in Figure 9-7 is controllable and observable if and only if SI2 is
mection of SI and S2
characterized by (;¡(s). controllable or, equivalently, no pole of 92(S) is canceled by any zero of 91(S),
This checks with what we have in part 3 of Theorem 9-2.
to ensure that the feed­
We consider a special case of Figure 9-7 in which Gz{s) is a constant matrix
n 3-5). Let SI? denote
K We cal! this a constant outputfeedback system. Since K does not introduce
tandem connection of
any new state variable, if SI is control1able, so is SI followed by K. If SI is
Ire9-7.
observable, so is K fol1owed by S l' Hence the feedback connection of SI
with a constant gain K in the feedback. oath 1s contro!Iabie 8.p.d I)bserv'~.b¡c; ¡r
and oniy if SI is controilable and observable. Hence controllability and
~rized by their transfer observability are invariant under any constant output feedback. This contrasts
(00)G 2(00» fO. Then with the constant state feedback discussed in Chapter 7 where controllability
md only ir SI 2 is con- is preserved under constant state feedback, but observability is generally not
preserved.
We discuss further the feedback system in Figure 9-7. Let the system Si
be described by irreducibleG¡(s)= N..¡(s)D,~ I (s) = Di; I (s)Nti(s) and irreducible
{A j, B¡, e¡, E;}, where GI(s)and G 2(s) are, respectively,'q xp and p x q proper
rational matrices. By assumption, we have .
i![ G¡(s)] ~ det D,.¡(s) ~ det Dli(S)~ det(sI - A;) .

and deg (;¡(s) = deg de.t Dr¡(s) == deg det Dli(S) = dím A¡= 11¡

where i![' ] denotes the characteristic polynomial of a rational matrix and ~


denotes equality ofpolynomials modulo a nonzero constant factor. The overall
446 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS COI

transfer matrix ofthe feedback system in Figure 9-7 is, as derived in (3-68), hence, we have
GJ(s) = G¡(s)(1 + G 2 (s)G¡(s))-¡
= N,.¡(s)D,~ ¡(s)[1 + Dii ¡(s)Nds)N,o¡(s)Dri ¡(s)] -¡
= N,o¡ (s)[Dds)D,.¡ (s) + Nds)N,.¡(s)] -¡Dds) (9-11)
The state equatior
The matrix A of its state-variable description is, as derived in (3-64),
__ [A¡-B¡Y 2 E 2 C¡ -B¡Y2C2 ]
AJ ( 9-12)
B 2 Y¡C¡ A 2 -B 2Y¡E¡C 2
with Y¡ =(1 + E¡E z )-¡ and Y z =(1 + E 2E¡)-¡. Ir the feedback system is not
completely characterized by GJ(s), then +
+
L\.[ GJ(s)] det (s 1- A J) lts A matrix is identic
and deg GJ(s) < dim AJ=n¡ + nz
Theorem 9-6
where +
denotes not equal modulo a nonzero constant factor. Conversely, if
the equalities hold, then the (n¡ + n2)-dimensional dynamical equation de­ Consider the system in
scription of the feedback system in Figure 9-7 is controllable and observable. O. Then we have
Note that we have
L\.(H(s))
L\.[ GJ(s)] +det [Dds)D,.¡(s) + Nds)N,.¡(s)]
because the factorization in (9-11) is not necessarily irreducible.
The controllability and observability of a system depends on the assignment
of the input and output. Although the system in Figure 9-7 may be uncon­
trollable or unobservable, if we assign the additional input and output as shown
in Figure 9-8, then it is easy to show, by using the argument in the proof of
Theorem 9-5, that the feedback system is always controllable and observable
from the input [r'l r~]' and the output [Y'I y'zJ. Similariy, the system is
always controllable and observable from the input [r') r~]' and the output
[u'¡ u~J. We study now the transfer matrix H(s) from [r'l r'J' to [u'¡ u'zJ.
From u¡(s)=r l (S)-Y2(s)=r,(s)-G 2(S)U 2 (s) and lJ 2 (s)=r 2 (s)+YI(s)=r 2 (s)+ where ~ denotes the
G¡(s)u¡(s), we have factor.

!.....;; I ~:. J

By definition, We remark first that th,


in Section 3-6, the pro¡::
[ uz(s)
~ (S)J¡ = H(s) [: I(S)J
r z(s)
H(s) may have poles a
right fraction

+ . [. ~1
fl
H(s)=
- G ¡ (s)

. 1I2 ~ r.., =[D,o~(s)

Figure 9-8 Feedback system with additional input and output. which can be easily veri
CONTROLLABILITY AND OBSERVABILITY OF COMPOSITE SYSTEMS 447

; derived in (3-68), hence, we have

(9-11)
The state equation of the system in Figure 9-8 can be readily computed as
ved in (3-64),

(9-12 ) [ ~IJz
X
= [Al - B I Y ~EzCI
BzYIC[
- B I Y zC, ]
Az-BzYIEICz
[Xl]
Xz

feedback system is not +[ B¡Y z


BzY,E t
-B 1 Y zE z
BzY\ _ r2
J[r[J
Its A matrix is idenlical lo lhe A r in (9-12).

Theorem 9-6
factor. Conversely, if
rnamical equation de­ Consider the system in Figure 9-8 wilh lhe assumpl ion det [1 + G[( ex:;. )G 2 ( ex:;. JJ =F
)l1able and observable. O. Then we have

6(H(s)) ~ del [D/[(s)Dds) + N Il (s)N,z(sIJ (9-13a)


rl(s)] ~ del [Ods)D,[(s) + N¡z(s)N,.¡(s)J (9-13b)

ducible. ~ det [ °'1(.1) NdS)] (9-13c)


ends on the assignment - N,.¡(s) Dds)
.lre 9-7 may be uncon­
ut and output as shown ~ det [ °ds) N/z(s)] (9-13d)
"ument in the proof of - NIl(s) D/1(s)
~l1able and observable
;imilarly, the system is ~ det (.1'1- A j ) (9-13e)
'1 r'2]' and the outpul ~ 6[ G 1 (s)J6[ G z(s)J del [1 + {;, (.I')G :!(sl] (9-13f)
1[r'l r'z]' t o [u'¡ u'z]'.
where ~ denotes the egua lity of polynomials modulo a nonzero constanl
l=r z(.I)+YI(s)=r 2 (s)+
factor.

We remark first that the condition det [1 + GI (ex:;. )(;z( ex:;.)] =F Oensures, as shown
in Section 3-6, the properness of al! elements of H(s). Without the assumption,
H(s) may have poles at s = ex:;. and the theorem may not hold. Consider the
right fraction

H(s) = [ ! (;2(S)J-['=[. 1_[ NdS)O,:;I(S)J- [


. ~ .
. -G¡(s) l. -N,.[(s)O"I(s)

[
='
0 .. [(.1)
O
0][ ~
Ods)
o,[(s)·.N,.z(S)]-1
N,.¡ (s). Uds)

which can be easily verified. The fraction is right coprime as canbeseen from,
tpul.
448 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

by carrying out elementary row operations, totically stable if the

r state as t -+ oo. A s'

l
S if the zero-state res po;
O"I(S) Nd )] O
-N'I(S) Ods) -+ -Nr¡(s) asymptotically stable,
0"1 (S) O Or¡ (S) However, ir a system
O Ods) O asymptotic stability ir
Consider two lineé
and the right coprime assumption of N..¡(s)O,~ I(S) (Theorem G-8). Thus, we
characterized by thei
have established, following (6-189), the equality in (9-13c).
transfer matrix G(s) (
Again by direct verification, we have
stability or the compe
H(s) = [ 1 N,.z(S)O,:-/(S)]-1 asymptotic stability (
- Oil 1(s)N l1 (s) 1 system is known to \;

1 O][1
- [O O,.z(s) - N l1 (s)
NdS ) ] ­
0 11 (s)O,z(s)
1 [1 O]
O O'I(S)

task to compute com]


nection. Even if the
tedious to check whei
It is easy to show that the factorization is irreducible. Hence we have terized. Therefore, it .

1 N,z(S)] ]
osite systems from e
~(H(s))=det [ -N (s) DI 1(s)Ods) =det[Ol1(s)O,z(s)+N l1 (s)N,z(s) function matrices. W(
l1 For the parallel an
where we have used, in the last step, the formula simple. The parallel
stable if and only ir 5
det[~ ~]=det[-dP-l ~][~ ~]=det[~ V-~p-IU] asymptotic stability is
(3-62) and (3-63), we ~
=det P det (V - QP- 1U) nection of SI and 52 a
assertion follows.
This establishes (9-13a). The equality ~[H(s)] ~ det (sI - A f ) follows directly
For the feedback
from (6- 189) and the fact that the state-variable equation is controllable and feedback system may
observable from the input [r'l r~]' and the output [U'1 u~]'. back system may be un
From (9-13a), we have
variable case and then
det {Ol1[I + 0 11 l(s)Nl1 (s)N"2(S)Or2 1(s)]O,z{s)}

= det D l1 (s) det [1 + (;1(S)(;2(S)] det D,z(s)


Single-variable fee
back system shown in F
This implies (9-13[). Q.E.D.

This theorem provides five formulas for computing det(sI-A f ) and is


9f (s) =1­-+
useful in stability study. In the single-variable case, there is no distinction
where 9¡(s) ~ Ni(S)/Di(~
between right and left fractions, and the theorem reduces to
Recall that every transf
det (sI - A f )=D¡(s)D 2 (s) + N 1 (s)N 2 (s) we speak of 9f(S)' we n
that if NI and D 2 havt
where 9¡(S) = Ni(S)/Di(S), i= 1,2.
syst~m js completely el
tween N ¡ and D 2 , ther
system is not completel
9-4 Stability of Feedback Systems
Theorem 9-7
Every control 'system is designecf to be stable. Therefore it is important to
study the stability problem of composite systems. We recall from Chapter 8 Consider the feedback ~
that a system or, mOre precisely, the zero-input response of a system is asymp­ 52 arecompletely char
STABILITY OF FEEDBACK SYSTEMS 449

totically stable if the response due to every initial state approaches the zero
state as t-+ oo. A system is bounded-input-bounded-output (BIBO) stable
if the zero-state response due to every bounded input is bounded. If a system is
'S)l asymptotically stable, then the system is also BIBO stable, but not conversely.
However, if a system is completely characterized by its transfer matrix, then
(s) asymptotic stability implies and is implied by BIBO stability.
Consider two linear time-invariant systems SI and 52 which are completely
:)[em G-8). Thus, we characterized by their transfer matrices G1 (s) and G2 (s). If the composite
). transfer matrix (;(s) of any connection of S, and S2 is computed, the BIBO
stability of the composite system can be determined from (;(s). However, its
asymptotic stability cannot be determined from (;(s) unless the composite
system is known to be completely characterized by (;(s). It is not a simple
task to compute composite transfer matrices, especially, in the feedback con­
nection. Even if the composite transfer function is obtained, it is still quite
tedious to check whether or not the composite system is completely charac­
ence wehave terized. Therefore, it is desirable to be able to determine the stability of comp­
osite systems from GI(s) and (;2(S) without computing composite-transfer­
function matrices. We shall study this problem in this section.
For the parallel and tandem connections of S, and S2' the problem is very
simple. The parallt::1 or tandem connection of S, and S 2 is asymptolically
stable if and only if S, and S 2 are asymptotically stable. The condition for
asymptotic stability is that all eigenvalues of A have negative real parts. From
(3-62) and (3-63), we see that the eigenvalues of the parallel or tandem con­
nection of SI and S 2 are the union of those of S, and those of S2- Hence the
assertion follows.
[- A f) follows directly
For the feedback system, the situation is much more complicated. A
.on is controllable and
feedback system may be stable with unstable subsystems; conversely, a feed­
"2, J'. back system may be unstable with stable subsystems. We study first the single­
variable case and then the multivariable case.
Ods)}
Single-variable feedback systems. Consider the single-variable feed­
O"2(S)
back system shown in Figure 9-9. The transfer function ofthe feedback system is
Q.E.D.
g (s) = : ~ (:; = i'f ' Q ?JI;s;
ng det(sI - A f ) and is f 1 +g,(S)g2(S) D,(s)D 2(s) +N,(s)N 2(s)- Df(s)
there is no distinction where g¡(s) ~ N¡(s)/D¡(s) for i = 1, 2 and N f and Df are assumed to be coprime.

;:s to Recall that every transfer function is assumed to be irreducible; therefore, when

¡(s) we speak of g¡(s), we mean gf(S) = Nf(s)/D¡(s). It was shown in Theorem 9-2

that if N, and D 2 have no common factor, then D f =D,D 2 +N I N 2, and the

. system is completely characterized by g¡(s). If there are common factors be­

-tween.N, and D 2, then D j consists of only a part of D,D 2 +N I N 2, and the

system is not completely characterized by gf(S).

Thebr€m.9.-7.
:fore it is important to
~. recall from Chapter 8 Consider the feedback system shown in Figure 9-9. lt is assumed that S, and
;e of a system is asymp- 52 are completely characterized by their proper transfer functions g,(s) and
450 LINEAR TlME-INVARIANT COMPOSITE SYSTEMS

system is asymptoti.
N ¡(s)N 2(5) have nega
Because of possil
D¡(s)D 2(s) + N¡(s)N z(
the theorem is only
Consequently, the S1
Figure 9-9 A feedback system. asymptotically stable
gzCs). Itisalsoassumedthatl +g¡(<Xl)92(<Xl) 7'=0. Letg¡(s)= N¡(s)/D¡(s). Then
the feedback system is asymptotically stable if and only if all the roots of This theorem car
D¡(s)Dz(s) + N¡(s)N z(s) have negative real parts. The condition is sufficient but proved here, howeve
Prom (9-15) we see t
not necessary for the systern to be BIBO stable.
independent of wheth
Proof tion of the feedback

°
First we note that the condition 1 + g1 (<Xl)9 2 (<Xl) -=/=- will ensure the properness
of gf(s). Let x = [x'¡ X'2J' be the composite state vector of the feedback
whether the feed ba(
g
I + ¡(<Xl )g z (<Xl) = O, e
real parts, the thec
system, where X¡ is the state vector of Si, and let x = Afx + bfu, y = efx +efU be (- S2 + s + 1)/S 2 , 92 s,
A ( •

the dynamical equation describing the feedback system [see Equation (3-64)]. (- Sz + s + I )/(s + 1) \1
Then the characteristic polynomial, det (s 1 - A f), of Af is a polynomial of N1(s)Nz(s)=(s+ 1) h,
degree n¡ +nz, where degD¡(s) =n¡, i = 1,2. input sin (2 will excitt
Nowwe show that ifl +g¡(<Xl )g 2(<Xl )7'=0, then deg [D¡(s)Dz(S) + N ¡(s)N2(S)] = ofthe feed back system
n¡ +n2' In other words, the term sn, -r n 2 in D¡(s)D 2(s) will not be canceled by 0, Theorem 9-7 does r
any term in N¡(s)N 2(s). Let If yz(s) in Figure 9-'
b· sn¡ +b· sn¡-¡ + ... output feedback systt
g.(s) = 'o I¡ controllable and obse
1 a¡Osn, +an sn ¡ ¡ + ...
implied by BIBO stab
Then we have
Corollary 9-7
· [1 +A ()A (')]-1 +b¡ob 20 _Q¡OQ20+ b lO b 20 (9-14 )
l 1m 91 s 92 S - ­
Consider the feedback
,~O() Q¡OQ20 Q¡OQ20
D¡(s)D 2 (s) +N¡(s)N2(s)=aIOa20Sn,+n2 + ... +bIOb20Sn,+rt2 + ... that S ¡ is completely
and
=(QIOQ20 +b lO b 20 )sn, +n, I + kg ¡( <Xl) -=/=- O. Then
(9-15 }
stable if and only if al
+terms of lower degrees
From (9-14) and (9-15) we see immediately that if 1 +g¡(<Xl)92(<Xl) 7'= 0, then
deg [D¡(s)D 2(s) + NI (s)N 2 (s)] = n¡ + n z = deg det (sl- A f ). Using this fact and Multivariable feed
the relation case is much more ca
Consider the multivari
g¡(s)=c¡(sI-Af)-¡bf+ef=d (; A ) [cfAdj(sI-A f )bf ]+ef S ¡ is assumed to be COI
et s - J
G¡(s). The assumptioJ
connection. Let Gf(S
NI (s)Dz(S)

D I"(s)D 2(s) + N¡(s)N 2(s)

u
we conclude that
det (sI - A f ) =k(D¡"(s)D 2(s) + N¡(s)Nz(s)) (9-16)
for sorne constant k. In other words, the set of the eigenvalues of A f and the
set of the roots of D¡(s)Dls)+ N ¡(s)N2(S) are the same. Hence the feedback Figure 9"10 A multiv,
STAB1UTY OF FEEDBACK SYSTEMS 451

system is asymptotically stable if and only if all the roots of D¡ls)D 2ls) +
N ¡ ls)N 2ls) have negative real parts.
Because of possible cancellations between N¡(s) and D 2(s), some roots of
D¡(s)D 2(s) + N¡(s)N 2(s) may not appear as poles of g¡(s); hence the condition in
the theorem is only a sufficient condition for the system to be BIBü stable.
Consequently, the system in Figure 9-9 can be BIBü stable without being
asymptotically stable. Q.E.D.

,) = N ¡(s)/D ¡(s). Then This theorem can in fact be deduced directly from Theorem 9-6. It is
Iy if all the roots of proved here, however, by using explicitly the condition I +g¡(00)g2(00)+0.
dirion is sufficient but
From (9-15) we see that (9-16) holds if and only if I +g¡(00)g2(00)+0; it is
independent of whether the (n¡ + n 2 )-dimensional dynamical equation descrip­
tion of the feedback system is irreducible or not. It is also independent of
whether the feedback system is completely characterized by g¡(s). If
ensure the properness I +g¡(00)g2(00)=0, even if all roots of D¡(s)D 2(s) + N¡(s)N 2(s) have negative
ctor of the feedback real parts, the theorem stili does not hold. For example, if g¡(s)=
I-b¡u, Y=cfx +e¡u be (-s2+ s +1)/s2, g2(s)=I, then we have l+g¡(00)g2(00)=0 and 9¡(S)=
:see Equation (3-64)]. (-S2+ S+ I)/(s+ 1) which is improper. Although the root of D¡(s)D 2(s)+
~ is a polynomial of N¡{s)N 2(s)=(s+ 1) has a negative real part, the application of the bounded
input sin {2 will excite a term of 2{ cos {2, which is not bounded, at the output
)D2(S) + N ¡ (s)N 2 (s)] = ofthe feedback system. Thus without imposing thecondition I + g¡ (00 )g2(00) +
ill not be canceled by 0, Theorem 9-7 does not hold in general.
If y2(S) in Figure 9-9 is a consta~nt k, the feedback system reduces to a constant
output feedback system and is, as discussed in the previous section, always
controllable and observable. In this case, asymptotic stabiiity implies and is
implied by BIBO stability.

Corollary 9-7
(9-14 )
Consider the feedback system shown in Figure 9-9 with {¡](s) = k. lt is assumed
Q2Ü
that S ¡ is completely characterized by its proper transrer runction yI (.1') and
b 2 0s"¡ +"2 +
I +kg¡(oo)+O. Then the feedback system is BIBü stable and asymptotically
stable ir and only if all zeros of 1 +kg¡(s) or, equivalently, all roots of D,(s)+
s J,,; t'll (.;-;) h8_V~ nega tiv~ ·~·~2.1 :-::'.:;"X í::::..

g¡(00)g2(00) 1=0, then


). Using this fact and Multivariable feedback systems. The situation in the multivariable
case is much more complex, hence we study first the simplest possible case.
Consider the multivariable feedback system shown in Figure 9-10. The system
sI - A¡) bf ] + e¡ ~¡ is assumed to be completel)' characterized by its p x p proper rational matrix
G¡(s). The assumption that G¡(s) is square is necessary for the proper feedback
connection. Let G¡(s) be the transfer matrix of the feedback system. Then
SI

N 2(S)) , .
nvalues of A¡ and the
Hence the feedback
(9-16)
~r
Figure 9-10 A multivariable feedback system.
452 LINEAR TlME-INVARIANT COMPOSITE SYSTEMS

we have First we note


det [1+ G1 (<Xl)J:¡bü (
L1 1 (s)det [1+ (; I (s)] i

where Adj(') denotes the adjoint of a matrix. Since SI is completely charac­


terized by G ¡ (s), it fol!ows from Theorem" 9-5 that the feedbas.k system in
where E(s) and F(s) a
Figure 9-10 is completely characterized by G¡(s). lf the order of G¡ (s) reduces
to 1 xl, then det [1 +G 1 (s)] becomes 1 +g¡(s) and the stability depends only
on the zeros of 1 +g¡(s). Hence, one might be tempted to assert that the
stability of multivariable feedback systems depends only on the zeros of
To show that ,1¡(s) d<
det [1 +G¡(s)]. 6.¡(s) is divisible with
following identity:
Example 1
Consider the multivariable feedback system shown in Figure 9-10 with
-s _s_l

l
where p is the order (
"
G¡(s) =
s-1
1 s +1
-2
J (;¡(s) of order i; for
elements of (;¡(s). TI
s +1 Now 6.¡(s) is the least
lt can be easily verified that F(s) is at most equal
minors of (;¡(s) folle
2S

l s~Sl without remainder b


"
G¡(s) == s +1. s-2
J Theorem 9-8 can be (
-1 portance, we shall pro
s-l
Proof of Theorem 9­
which is unstable. However, we have det [1 +G¡(s)J = -1, which does not
have any right-half-plane zeros. Hence the stability of the feedback system We shal! use the dyn.
cannot be determined solely from the zeros of det [1 + G¡ (s)]. !!l first part of the theore

In order to study the stability of multivariable feedback systems, the concept


of the characteristic polynomial of a proper rational matrix ~ needed. Recall
;: 1 thn" t'D~ r.hnrnnro-'si·'r ·rnj .. ·.·.r_r.' .. ; .r. r ..' ,:,~,,-,_
f··~_
itvííl L ..·.".,ünfoon 0-
"o':; ".' .
ú:H le ..... i c..~a.''-'L'..... !¡ >'i" !?V~Y,L:'Jl':'''..i.c.;.: h c ';':"':·;1\~1):, . j~.' ••.' ' .
.!:, ... 'p .J'

,1¡(S), is the least common denominator of al! the minors of G¡(s). lf G¡(s) has
be tne dynamical-eq ua
the coprime fraction N¡(s)D¡¡(s), then we also have 6.¡(s)=kdetD¡(s), for
some nonzero constant k.
Note that the dynam
Theorem 9-8 its knowledge is not
assumption that S¡ is
Consider the feedback system shown in Figure 9-10. It is assumed that S ¡
control!able and obst
is completely characterized by its proper rational matrix G¡(s) and that
6.¡ (s) of G ¡ (s) is equal t
det [1 + (; ¡ (00)] :¡b O. Let (; ¡ (s) = N,.¡ (s)D,:¡ ¡ (s) = D¡l~(s)Nl1 (s) be coprime frac­
tions and 6.¡ (s) be the characteristic polynomial of G¡ (s). Then the feedback
system is asymptotical!y stable and BIBO stableif and only if al! the roots of
By the substitution 0[.
any of the three polynomials 6.¡(s)det[I+·G¡(s)],det[Ó,.;(s)+N rl (s)J and
det [D1¡(s) + N¡¡(s)] have negative real parts. I
STABILITY OF FEEDBACK SYSTEMS 453

First we note that e As) is proper because of the assumption


det [1 + G1 (ro)] =1=- o (Theorem 3-5). Before proving the theorem, we show that
~[(s)det [1+ e[(s)J is a polynomial. Define
A E(s)
det [1 +G[(s)] ~ F(s) (9-18)
is completely charac­
e feedback system in where E(s) and F(s) are assumed to be coprime. Then we have
order of el (s) reduces
stability depends only ~[(s)det [1 +e[(s)1 = ~[ls)E(s)
ted to assert that the - F(s)
)nly on the zeros of To show that ~[(s) det [1 + e[ (s)J is a polynomial is the same as showing that
~[(s) is divisible without remainder by F(s). This will be shown by using the
following identity:
A P

gure 9-10 with det [1 +G[(s)J = ¿ (X¡(s) +1 (9-19)


i; 1

where p is the order of eleS) and (Xi is the sum of all the principal minors of
eleS) of order i; for example, (Xp(s)=det (;[(s), (X[(s)=sum of the diagonal
elements of eleS). The verification of (9-19) is straightforward and is omitted.
Now ~[(s) is the least common denominator of all the minors of G 1 (s), whereas
F(s) is at most equal to the least common denominator of aH the principal
minors of (;¡(s) following (9-19); hence we conc1ude that ~¡(s) is divisible
without remainder by F(s) and that ~¡(s)det [1+ (;¡(s)] is a polynomial.
Theorem 9-8 can be deduced from Theorem 9-6; however, because of its im­
portance, we shall prove it directly in the following.

Proof of Theorem 9-8


= - 1, which does not
)f the feedback system Weshall use the dynamical-equation description of the system to pro ve the
;¡(s)]. III first part of the theorem. Let

FE: x¡ = A¡x¡ + 8¡1I¡ (9-20a)


ck systems, the concept
y¡ =C¡x¡ + E¡"¡ (9-20b)
itrix is needed. Recall
1 oí G1(s), denoted by be the dynamical-equation description oí the system .)"¡. it is ciear lhal
s of G¡(s). If G¡(s) has
~¡(s)=kdetD¡(s), for (9-21 )

Note that the dynamical equation FE is used only implicitly in the proof;
its knowledge is not required in the application of the theorem. Now the
assumption that S ¡ iscompletely characterized by e ¡(s) implies that FE is
1t is assumed that S ¡ controllable and observable. Consequently, the characteristic polynomial
natrix G¡ (s) and that ~ ¡(s) of e 1(s) is equal to the characteristic polynomial of A¡; that is,
IN l ¡ (s) be coprime frac­
:s). Then the feedback ~¡ (s) ~ ~[G¡(s)J = det(sl - A;) . (9-22)
only if all the roots of By the substitutionof ti 1.= u - y ¡, y = y[ and the use óf the identity
:t[D, ¡(s) + Nrt(s)J and
• 1-'---(1 +Etl-[E[ =(1 +Etl-¡
"-.

454 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

the dynamical-equation description 01' the feedback system can be obtained as Example 2

FE¡: X¡ =[A¡-BI(I+E¡)-lC¡Jx¡ +BI(I+E¡)-lu (9-23a) Determine the stabil


y = (1+ E1r IClx¡ + (1 + E¡)- IElu (9-23b)

The dynamical equation FE ¡ is controllable and observable following Theorem


9-5. Hence we conclude that the feedback system is BIBO and asymptotically
stable if and only if all the eigenvalues 01' Al -B¡(I +E¡)-¡C¡ ha ve negative
real parts. We show in the following that the set 01' the eigenvalues 01' The characteristic pI
Al -BI(I +E¡}- IC I is equal to the set of the zeros of c,¡(s)det [1 +G¡(s)]. verify that det [1 + G
Consider only if aH the roots o
parts. This is not th
nor asymptotically st
=det {(.d - Al )[1 +(sl- A¡ )-¡ 8 1(1 + El )-ICIJ: (9-24 ) lt is 01' interest to
It is well known that det AB =det A det 8: hence (9-24) becomes
Óf(s) = det (sI - Ad det [1 +(sl - A¡}-¡BI(I +E¡)-¡C¡J (9-25)
lt is shown in Theorem 3-2 that det [1 + G¡(s)G 2(S)] = det [1 + (; 2(S)(; I(s)J and, where E(s) and F(s) h
since Ó¡(.~) =det (sI - At!. Equatíon (9-25) implies that feedback system intr
poles which are posse
Ó¡ls)=Ó¡ls)det[1 +CI(sl-A¡}-IBI(1 +E¡}-IJ system. For exampl,
=Ó¡ls)det [[(t +E¡) +C¡(sl -A¡)-IBIJ(I +E¡}-I} (1 + 1)(s - 1); hence n
=ÓI(s)det [1 + (;¡(s)J det (1 + E¡)-¡ (9-26) G¡(s)and G1(s)have
Since det (1 + E¡) - ¡ = det [1 + G¡l co)] - I =FO. we conclude from (9-26) that the Roughly speaking
set of the eigenvalues of Al - BllI + E I )- le l is eguaI to the set 01' the zeros 01' of det [1 + (;¡(s)]. T
Ó lls) det [1 + G¡(s)]. This proves the first part 01' the theorem. any common factors i
Now we show the second part. Let G¡{s)= N,. I(s)D':-1 I (s). Then we have might be determinabl
necessarily correct, as
(;¡(s) = G1(.1')[1 + (; 1(.1')] - I = N.. ¡(s) 0':-1 I (s)[1 + N"I(S)O,~ I (s)] - 1

= N,. ¡(s)D,:j I(S){ [D"I (s) + N .. I (s)]D,:j I (s)] - 1 Example 3


= Nr¡(s)[D,.¡(s)+N'·I(s)]-1 (9-27)
Consider the feedbacl
We show that if D"1 (s) and Nr¡ (s) are right coprime.. so are iD,. l.,..• -: ]'>1 i ,) ?n-j
NI[ (s). The coprimeness of D q (s) and N.. I(s) implíes the existence 01' polynomiaí
matrices X(s) and Y(s) such that

X(s)Dr¡ (s) + Y(s)N,.¡ (s) = I


It is straightforward t,
(Theorem G-8). This equation implies

X(s)[D,.¡ (s) + N.. ¡(s)] + [Y(s) - X(s)JNr¡ (s);' I

Since X(s) and Y(s) - X(s) are both polynomial matrices, Theorem G-8 implies
the rightcoprimeness 01' D'1(s)+ N.. ¡(s) and Nr¡(s). Conseq.uently, G¡(s)=
N,.¡(s)[D d (s) + N .. 1(s)J -¡ is a coprime fraction and ó[ Gf(S)J~ det [D,.¡(s) +
. N,.¡(s)] ~ Ó¡(s). Hence the feedback system is BIBü and asymptotically stable 11' we do not cancel o
if and only if all the roots 01' detCD,.¡ (s) + N"I (s)] have negative real parts. half-plane zero. Thw
Q.E.D. not stable.
STABIUTY Of FEEDBACK SYSTEMS 455

m can be obtained as Example 2

(9-23a)
(9-23b)

le fol1owing Theorem
) and asymptotical1y
A

GI(s)=

l l
Determine the stability of the feedback system in Figure 9-10, with

-s
s-1 s
5+1
-2
s +1
¡)-le l have negative
f the eigenvalues of The characteristic polynomial of G¡(s) is oÓ¡(s)=(s-l)(s +1). It is easv to
oÓ¡(s)det [1 + (;1(5l]. verify that det [1 + (;¡(s)] = -1. Hence the feedback system is stable if'and
on1y if al1 the roots of oÓ ¡ (s) det [1 + (; ¡(s)J = - (s - 1)(s + 1) have negative real
parts. This is not the case; hence the feedback system is neither BlBO stab1e
nor asymptotically stable. ~

(9-24 ) lt is of interest to give sorne interpretation of the roots of the p01ynomial


~comes A E(s)
oÓ¡(s)det [1 +GI(sj] =oÓ¡(s)­
E1)-IC1J (9-25)
F(s)
where E(s) and F(s) have no common factor. The roots of E(s) are poles of the
[1 + 6 2(s)(; I(s)J and, feedback system introduced by feedback, whereas the roots of oÓ 1(s)j F(s) are
poles which are possessed by the open-loop system S las well as by the feedback
system. For example, in Example 1 we have E(s) = - 1,F(s) = 1, oÓ¡(s)jF(s) =
'1 + 1)(s -1); hence no new pole is introduccd in the feedback system G¡(s);
(9-26) G ¡(s) and G 1(s) have the same poles 1 and -1.
Roughly speaking, the polynomial oÓ1(s)/ F(s) takes care of the missing zeros
: from (9-26) that the
of det [1 + (;I(S)]. Therefore one rnight suggest that if we do not cancel out
he set of the zeros of
any cornmon factors in det [1 +6¡(s)J, then the stability ofthe feedback system
Drem.
might be determinable from the zeros of det [1 + 6 1(s)]. This is however not
I(S). Then we have
necessarily correct, as can be seen from the fo\lowing example.

Example 3
(9-27)
Consider the feedback system shown in Figure 9-10 with
1 1
·e Dr¡(s) + N,¡(.s) and
istence of polynomial s-0.5
1
s -0.5 J
lt is straightforward to verify that

(s + 0.S)(S2 + s + 0.25) s+0.5

)=1 (s+ 1.5)(s-0.5)2 (s+ 1.5)(s- 0.5)2

[heorem G-S implies (s +0.5)(5 - 0.5)(s + 1.5)


)nsequently, G¡(s) = (s - 0.5)(s - 0.5)(5 + 1.5)
t; ¡(s)] ~ det [D,.¡ (s) +
asyrnptotical1y stable If we do not cancel out the common factor, then det [1 + G¡(s)J has a right­
legative real parts. half-plane zero. Thus we may suggest erroneously that the feedback system is
Q.E.D. not stable.
456 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

The determinant of (;l(S) is given by and conseguently,


A 1-(s+0.5) -1 det (.
det G 1(s) = =---:-­
(s + 1.5)(s - O.W (s + 1.5)(s - 0.5) if and only if
Hence the characteristic polynomial, L1 1 , of G1(s) is
Ll 1 (s) =(s +1.5)(s -0.5)
Tosave space, this as
Since we have E(s)/ F(s) = (s +0.5)/(s - 0.5) and L1 1(s)E(s)/ F(s) = (s + 1.5)(s +0.5), of (9-29) is independ¡
the feedback system is asymptotica11y stable and BIBO stable. 111 characterized by
G¡(s) =
We consider now the case where a dynamical system appears in the feedback
path as shown in Figure 9-11. This feedback system is generally not completely
characterized by its composite transfer matrix; hence the situation here is more
complicated than the unity feedback system shown in Figure 9-10. The factorization in (9

Theorem 9-9 L1[G¡(s)] f


For this reason, the c(
Consider the feedback system shown in Figure 9-11. It is assumed that S 1 and
system in Figure 9-11
~z are com,Pletely characterized by their q x p an~p x q proper transfer m~trices
+
G 1(s) and Gz(s). It is also assumed that det [1 + G 1(00 )G z(00)] O. Let G¡(s) =
the feedback system i~
terized by G¡(s) and
N..¡(s)D,:¡ l(S)= Di¡ 1(s)Nli(s), i = 1,2, be coprime fractions and L1,(s) be the
Theorem 9-9 beco mes
characteristic polynomial of ¿¡(s). Then the feedback system is!symytotically
stable if and only if a11 the roots of L1 1(s)L1 z(s)det[I+G 1(s)G Z (s)], or Now we consider
and output as shown i
DI 1(s)Drz(s) + N 11 (S)N,·z(s), or any other of (9-13) have negative real parts. The

["l(S)J=
condition is sufficient but not necessarily for the system to be BIBO stable.
Ir Gz(s) = K, a p x q constant matrix, the condition is necessary as we11 for the "z(s)
system to be BIBO stable. 1
By direct computatior
This theorem can be deduced directly from Theorem 9-6. It can also be
proved by using its dynamical eguation description to establish, as in (9-26),
H(s) = [~" Gz(é
-G 1 (s) t¡
L1 (s)
f
~ det (si - A ) = L1 (s)L1 (s) det [1 +A G1(s)?z(S)]
f 1 z (9-28) [1" + G2 (s)d
det [1 + G1(oo)Gzioo)] = [ ¿l(s)[i p +G z(
riecause the cnaracter;
referred to Rererence 49. Similar to the proof in (9-14) and (9-15), it is possible egual to det (si - A f ),
to establish, by assuming the column reducedness or Drz(s) and the row re­ sufficient for the systen
d ucedness of 0 11 (s), that that there exist G 1 (s) ¡
deg det [Dll(s)D,.z(s) + N 11 (S)N.. z(s)] = deg ¿l(S) + deg Gz(s) = n1 + nz stable but the fourth er
instability of the en trito
In the design of fee
+ section ofwe11-posedn(
pairs to be proper. SiJ
al! possible input-outp¡
is not BIBO stable, an
input at rz wi11 cause th
in practice to reguire e'
Figure 9~11 A multivariable feedback system. input-output pairs.
STABILITY OF FEEDBACK SYSTEMS 457

and consequently.
(9-29 )
-0.5) if and only if
det [1 + GI(co)GAcoJ] +O
To save space, this assertion wil! not be proved here. We note that the validity
4'(.1') = (s +1.5)(s +0.5), of (9-29) is independent of whether 01' not the feedback systern is completely
~able. 11 characterized by
G ¡(s) = [1 + G¡(s)Gz(s)] - I GI(s)
ppears in the feedback = [1 + Di¡ I(s)N II (s}N,.z(s)D,:;I(SJ] - I Di¡l(s)N lI (s)
lerally not completely = Dds)[DlI(s)Dds) + NII(s)N,.z(s)] - I NlI(s) (9-30)
situation here is more
sure 9-10. The factorization in (9-30) is not necessarily irreducible; hence general!y we have
~[G ¡(s)] f det [DI I(s)Dds) + NII(s)Nds)] ~ det (s 1- A¡) (9-31 )

For this reason, the condition in Theorem 9-9 is only sufficient for G¡(s) 01' the
, assumed that SI and system in Figure 9-11 to be BIBO stable. Ir Gz(s) = K, a p x q constant matrix,
oper transfer m~trices the feedback system is, as discussed in the previous section, completely charac­
(co)J '1'=0. Let G¡(s) = terized by G¡(s) and ~[G¡(s)] ~det [DII(s)+ N¡¡(s)KJ; hence the condition in
ns and ~i(S) be the Theorem 9-9 becomes necessary as wel! for G¡(s) to be BIBO stable.
,tem is}syrnytotical!y Now we consider the same system in Figure 9-11 but with additional input
t [1 + GI(s)GZ(s)J, or and Olltput as shown in Figure 9-8. As derived in (9-12), we have
sative real parts. The
n to be BIBO stable. GZ(s)J-I [~I(S)J
::essary as wel! for the Iq 1'z(s)
1 By direct computation, we have

n 9-6. It can also be H(s) = [~p GZ(S)J-I


;tablish, as in (9-26), -GI(s) Iq
I (.1')(; z(s)J
(9-28)
[lp + Gz(s)G¡(s)] - 1 - Gz(s)[I'!- + G~(s)Gz(s)] - I J (9-32)
co )(;z(co)] = _GI(s)[l p + GZ(s)GI(S)]-1
[ [I q + GI(s)GZ(S)]-1
he interested reader lS Because (he characteristic polynomiai 01 H(s) is, as dcveloped in fÍleorem 9-6,
nd (9-15), it is possible equal to det (sI - A¡), the condition in Theorem 9-9 is now necessary and
>rz(s) and the row re- sufficient for the system in Figure 9-8 to be BIBO stable. It is important to note
that there exist GI(s) and Gz(s) so that any three entries of H(s) can be BIBO
stable but the fourth entry is noL For al! possible combinations of stability and
instability of the entries of H(s), see Reference S75.
In the design of feedback systems, we shál! require, as discussed in the sub­
section of well-posedness, every transfer function from ál! possible input-output
pairs to be proper. Similar1y we shall require a system tobe BIBO stable from
al! possible input-output pairs. For example, ifthe transfer matrix from r z tO"1
isnot BIBO stable, any load disturbances Which can be 11l0deleq as a nonzero .
input at l' z will cause the system in Figure9-8 to saturaú~. '. Hence it is rea"sonable
inpractice to require every feedback system to be BIBO stable from al! possible
input-outPllt pairs.
458 LINEAR T1ME-INVARIANT COMPOSITE SYSTEMS

Consider again the system in Figure 9-8 with [r'l ra as the input and
[Y'I y'2T as the output. Clearly, we have

YI(S)J (¿ Hy(s)
[ Y2(S) -
[~I(S)J [~I(S)~I(S)J= [G
= 1 (S)

r 2(s) G 2(S)U 2(s) O

which ímplies, by using (9-32),

HAS)=[A GI~)[lp+G2~)GI~)]-1 - G~(s)G2(s)[I~+ G~(S)G2(S)J -IJ Figurb 9-12 Unity fee

G 2(s)G 1(s)[l p + G 2(s)G 1(s)] - 1 G 2(s)[I q + G 1 (s)G 2(s)] - 1


sator C(s) = N,(s)/Dc(s
Using the identity is a real constant. W,
1- G¡(s)Gj(s)[1 + G¡(s)Gj(s)] -1 = [1 + G¡(s)GAS)]-1 for i,j= 1, 2 (9-34) without imposing any,
of C(s) is imposed to é
and defining Let gf(S) be the trar
(9-35)

we can readily establish ¡::;-


Lrom gf s +gf(S)C(s)¡;
A () A ,

(9-36) N(s)/D(s), we can write


From this relationship, we see that, because of the nonsingularity of J, every
f\.
pole of H(s) is a pole of HAs) and vice versa. Hence H(s) is BIBO stable if and C(s) (¿ ­
-[
only if HAs) is BIBO stable. This fact can also be seen from Figure 9-8. Because
OfY2 = r l - U1 and Yl = U2 - r 2, if r¡(t) and U¡(t) are bounded, so arey¡(t), i= 1,2. Now we c\aim that fo
inequality
Corollary 9-9
The feedback system in Figure 9-8 is BIBO stable if and only if the feedback
system in Figure 9-11 is asymptotically stable. I where ¿j denotes the de,
exist to achieve the desi
This corollary can also be conc1uded from Problem 8-40. Thus, if a feedback
system is designed to be asymptotically stable, then the system is BIBO stable
from al1 possible input-output pairs. On the other hand, if a system is BIBO which, together with (9­
stable from one input-output pair and if there is no unstable hidden mode or,
(j D e(s) - ¿j N e!
roughly speaking, no unstable pole-zero cancellation, lhen the syslem is asymp-·
totically stable. In practice, every control system should be designed to be rlence C(s) is proper. Ti
asymptotically stable. a proper compensator e
not only the poles but ;
assigned.
There is, however, a
9-5 Design of Compensators: Unity Feedback Systems see that C(s) contains t
nection of C(s) fol1ow(
Single-variabJe case. In Chapter 7, we studied the problem of state feed­ Furthermore, the cancel
back and state.estimator and showed that the eigenvalues of the original system has no control over thest
can be arbitrarily assigned withOut affecting its numerator. In this section, we right-halfs plane, then tI
shall study a similar problem by using transfer functions in fractional forms. Hence this design is no
We study" onlyuníty feedbacksystems in this section; the configuration devel­ undesirable pole-zero ca
opedin Chapter 7 wi11 be studj(~d inSection 9-7. polesand zeros are no\
Consider the unityfeedback system shown in Figure 9-12. The compen- from (9-37).
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 459

as the input and Compensator Plan!

I'~~lll
--j
k
~"
----... ces) ~i g(s) =.V(s)/D{sl l---;--..
1- ~L! I !

L __._. .=.J
FigurE> 9-12 Unily feedback syslem.

sator C(s) = N,(s)/Dc(s) is required to be a proper rational function. The gain k


is a real constant. We shall study what can be achieved in this configuration
without imposing any constraint on C(s) so long as it is proper. The properness
'01' i,j= 1,2 (9-34)
of C(s) is imposed to avoid amplification of high-frequency noises.
Let 9¡(S) be the transfer function from r to y in Figure 9-12. Then we have
kC(s)g(s)
(9-35) g¡(.s) A

1 +C(s)g(s)

From g¡(s) +g¡(s)C(s)9(s) = kC(s)g(s), and using gj(s) = N¡(s)/O¡(s) and g(s) =
(9-36) N(s)fD(s), we can write C(s) as

;ingularity of J, every O(s)N ¡(s)


(9-37)
lis BIBO stable if and N(s)(kO¡(s) - N¡(s»
n Figure9-8. Because
ed, so are y¡(t), i= 1,2. Now we claim that for any gj(s) = Nj(s)/Dj(s) meeting the pole-zero excess
inequality
(jD j(s) -(jN¡(s) ';2: (jD(s) - (jN(s) (9-38)
d only if the feédback
I
where (j denotes the degree, a constant gain k and a proper compensator C(s)
exist to achieve the designo lndeed, by a proper choice of k, we have
~O. Thus, if a feedback (j[kO ¡(s) - N ¡(s)] = (jD j(s)
system is BIBO stable
d, if a system is BIBO which, together with (9-37) and (9-38), imply
:table hidden mode al', bDc(s) - bN ,(s) = [bO¡(s) - bN ¡(s)] - [bD(s) - bN(s)] ';2: O
:n the system is asymp­
uld be designed to be Hence C(s) is proper. Thus for any ij¡(s) meeting the pole-zero excess inequality,
a proper compensator exists in Figure 9-12 to meet the designo We note that
not only the poles but also the zeros of the overall system can be arbitrarily
assigned.
There is, however, a very serious problem in the designo From (9-37), we
back Systems see that C(s) contains the factor l/g(s) = O(s)/N(s). Hence the tandem con­
nectiori. of .C(s) followed by g(s) always involves poie-zero cancellations.
problem of state feed­ Furthe'rmore, the canceled poles are dictated by the given plant; the designer
s of the original system has no cO'n'trol ayer these poles. Ir the 'given 5(s) has polesor zeros in the open
or. In this section, we right-half splane, then the design wm" cOntain unstable pole-zero cancellations.
ns in fractional forms. Hence thisdesign is not always permissible inpractice. In ·oth.er words, if
le corifiguration devel­ undesirable pole~zero cancellations are notpermitted, ai-bitrary assignments of
polesand zeros are not always possible in Figure 9-12 by computing C(s)
re 9-12. The compen- from (9-37).
460 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

In the following, we discuss a method of computing C(s) which will not Equating the coeffici
involve any undesirable pole-zero cancel1ations. However, we must paya
price for achieving this. The zeros of N¡(s) of the resulting system cannot be
arbitrarily chosen; they are dictated by the zeros of C(s) and g(s). If we wish to with
control the poles as well as the zeros of gAs), then we must choose a different
Do DI ...

configuration (see Problem 9-10 and Reference S34).


= =
The substitution of g¡ N ¡/D¡, g N/D, and C NJD c into g¡(s) =
kC(s)§(s)/[l + C(s)§(s)] yields
= No NI···

N¡(s)
D¡(s)
kNc(s)N(s)
Dc(s)D(s) + Nc(s)N(s)
(9-39) SnI=l:~ );::::­
-- - --------
In the arbitrary pole placement, we study only the polynomial equation O O .
(9-40) O O .

In the fol1owing, we shall study the condition on D(s) and N(s) under which a This is a set oflinear ~
set of solution {Dc(s), Nc(s)} exists in (9-40) for any DAs). We sha1l also study between the polynorr
the degrees of Dc(s) and Nc(s) in order to achieve arbitrary pole placement. 4 hence it is permissit
By arbitrary pole placement, we always assume implicitly that complex con­ Sm consists of m + 1 1
jugate poles appear in pair. Otherwise, the compensator will have complex coefficients of g(s) anl
coefficients and cannot be implemented in practice. right by one column.
The polynomial equation in (9-40) is called the Diophantine equation in application of Theor
honor of Diophantus of the third century in Reference S139. 1t is ca1led the (9-44) has a solution
compensator equation in Reference S34. This is the most important equation S", has a full column
in the remainder of this chapter. ·Instead of studying· it directly (see Problem rank is that 2(m + 1);
G-14), we shall translate itinto a set of linear algebraic equations. From the compensator to achi
set of algebraic equations, we will then develop all properties of the Diophantine G-4, we have that S"
equation. The set of algebraic equations can also be used to compute the In fact, from the proo
compensator C(s). Let us define for m:2:n -1 ifand on
D(s)=Do+Dls+··· +D"sn Dn-l=0 (9-41 a) the following theoren
N(s)=N o +NIs + +N"s" (9-41 b)
Dc(s)=D co +Dels + + Dc",s"' (9-42a) Theorem 9-10
Nc(s)=N co +N c1s + +Ncmsm 19-42b \

and define
(9-43)
with deg N(s)~deg D
where Di, Ni, Dci> N ci , and F¡ are real constant, not necessarily a1l nonzero. The for every DAs) of degI
substitution of (9-41) to (9-43) into (9-40) yields and only if D(s) and !'
F o + F IS + ... + Fn+msn+m =(D co + DeiS + +Dcmsm)(D o + D1s + +Dnsn)
+(N co + NeiS + +Ncmsm)(N o + N IS + +N"s") This theorem statl
to meet (9-46). Noth

4
00 Matbemalieally, Ihe finding of polynomial solutions in (9-40) is equivalent to the finding of inleger
solutions x, y in ax +by = f, wherea,b, and r are giOven integerso This is a topie in number theory 5 Ir D(s)and N(s) are note.
or continued ofraetlonso only if the grealest como;
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 461

· C(s) which will not Equating the coefficients of the same power of s yields
:ver, we must paya [DeoNeo:Dc1Nc1:"':DemNem]Sm=[Fo F¡ Fz (9-44 )
Img system cannot be
ld g(s). If we wish to with
I 1St choose a different

rIt"_o__!!_¡__ ~._._ J!::-: ~ __ ,,!_,,


Do D¡ '" D,,_¡ D" O
?
O
0
... 011
o_.~ __ ~_ 1st block row
O Do .. o D,,-z D,,_¡ D" O .. o v 1
O No'" N,,-z N,,_¡
--------------------------------------
N" O ... O~
(9-39)
: : (9-45)
lOmial equation -O - --O - -'-0-0- - --0- - - - -i>~ -- -i>~ - -i>; --.-.~ --D,~ } (m+ l)th block row
(9-40) O O ... O No N¡ N z ... N"

i N(s) under which a This is a set of linear algebraic equations. There is a one-to-one correspondence
We shalI also study between the polynomial equation in (9-40) and the algebraic equations in (9-44);
lry pole placement. 4 hence it is permissibk to study the former by using the latter. The matrix
y that complex con­ Sm consists of m + I block rows; each block row has two rows formed from the
JI' will have complex coefficients of g(s) and can be obtained by shifting its previous block row to the
right by one columno lt is clear that Sm is a 2(m + 1) x (n +m + 1) matrixo The
phantine equation in application of Theorem 2-4 to the transpose of (9-44) reveals that for every F,
,139. 1t is called the (9-44) has a solution {De;, N ei , i = 0,1,. o. ,m} if and only if pSm = n +m + 1 or
t important equation Sm has a fuIl column rank. A necess~ry condition for Sm.to have a fuIl column
jirectly (see Problem rank is that 2(m + l);::::n +m + 1, 01' m;:::: n -1. Hence the smallest degn~e o(a
quationso From the compensator to achieve arbitrary pole placement is n -1. From CoroIlary
es of the Diophantine G-4, we have that S"-1 is nonsingular if and only if D(s) and N(s) are coprime.
lsed to compute the In fact, from the proof of CoroIlary G-4, we have that Sm has a full column rank
for m ~n - 1 if and only if D(s) and N(s) are coprimeo Hence we haveestablished
(9-41 a)
the foIlowing theoremo s
O
(9-41 b)
(9A2a) Theorem 9-10
(9-42b 1

D¡(s) = Dls)D(s) + Ne(s)N(s) (9-46)


I-m (9-43)
with deg N(s) ::;deg D(s) = n. Let De(s) and Nc(s) be of degree m 01' lesso Then
rily aIl nonzero. The for every D¡(s) of degree n + mor less, there exist DC<s) and N ls) to meet (9-46) if
and only ir D(s) and N(s) are coprime and m ~n -1. I

:) +D¡s + .. +D"s")
o

:) +N¡s + .... +N,¡S") This theorem states only the conditions for the existence of DC<s) and NC<s)
'. to meet (9-46). Nothing hasbeen said regarding the properness of Ne(s)/De(s).

ent to ¡he find ing of integer


5 lf D(s) ami 'N(s) arenot eoprime, 'then a solution {D,(s), N~(s)} exists in (9-46) for any DAs) if and
is a topie in number theory
only if the greatest comrrion divisor of D(s) and N(s) is a divisor of D /(s). See Problem G-14.
462 LINEAR TlME-INVARIANT COMPOSITE SYSTEMS

This question will be answered in the fol1owing two theorems by removing the The employment
condition deg D¡(s) < n + m fram Theorem 9-10. set of linear algebraic
whether g(s) is strictl
Theorem 9-11 required compensato
Consider the feedback system shown in Figure 9-12 with g(s) = N(s)/D(s) and
deg N(s) < deg D(s) = n. Then for any D ¡(s) of degree n +m, a praper compen­ Example 1
sator C(s) of degree m exists so that the feedback system has transfer function Consider a plant witl
N(s)Dj 1(s)Ne(s) if and on1y if D(s) and N(s) are coprime and m e. n - 1.

Proof
If we show that D em in (9-44) is different from zero, then this theorem follows
Let C(s) be chosen as
directly from Theorem 9-10. If g(s) is strictly proper, then N" =0. Conse­
quently, fram the last column of (9-44), we have Dem=Fn+,JD n. Hence if
F n +m is different from zero, so is Dem . This proves that the compensator C(s)
is proper. Q.E.D.
Then the equation in
Theorem 9-11'
Consider the feedback system shown in Figure 9-12 with 9(S) = N(s)/D(s) and
deg N(s)~deg D(s)=n. Then for any D¡(s) of degree n+m, a strictly proper
compensator C(s) of degree m exists so that the feedback system has transfer
function N(s)Dj l(s)Ne(s) if and only if D(s) and N(s) are coprime and m c.n.
Since the square ma
Proof solution exists in (9-4~
For a proper 9(S), the solution {Dei, NeJ in (9-44) exists if and only if N(s) and sator of degree 1 alwa:
D(s) are coprime and m c..n -1. However, if m = n -1, the solution is unique, (9-48) may become
and there is no guarantee that Dem i=- O, and the compensator may become the s01ution in (9-48)
or C(s) = (3s + 1)/(-(
improper. If m c.n, the number of unknown {De;, NeJ in (9-44) is larger than
the number of equations. Since the last row of Sm, with m c.n, is linearly function.
Now we choose
dependent on its previous rows (see Section G-2), we may choose N e ", = O.
For this choice, we have De", = Fn+,JDn- Hence the compensator is strictly
proper. This establishes the theorem. Q.E.D.

In Theorem 9-11, ir m = n -1 and D¡(s) = P(s)D(s), where P(s) is an arbitrary


polynomial of degree m, then the unique solution of (9-46) is Dis) = P(s) and
Ne(s) = O. In this case, no pole of the plant is to be altered, and the compensator
is zero. This is a degenerated case. Similarly, in Theorem 9-11', if m = n and
D¡(s) = P(s)D(s), and if the compensator is required to be strictly proper, then
Ne(s)=O, and the compensator is zero. Clearly, Theorem 9-11' still holds if
the compensator is permitted too be .proper. As discussed in Section 3-6, the
well-posedness problem of feedback systems .should be checked in the designo
In Theorem 9~Ü, 9(s) is strictly proper and C(s) is proper; hence 1 + C( 00 )g(oo) i=­
O, and the feedback system is wdl pósed. 'In .Theorem 9-11', 9(S) is proper, but
C(s) is strictly proper: hence 1 +C(co)g(co)i=-O, and the feedback system is The first five rows of t
again well posed. the design mayinvolve polezerü cancel1ations between a ful1 column rank;
D¡(s) and N(s )Ne(s). H owever these poles are al1assignable by the designer; De2 = F 4 and N e2 = O
hence they will not cause any problem in the designo strictly proper transf(
DEsIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 463

rems by removing the The employment of Theorems 9-11 and 9-11' is very simple. We form the
set of linear algebraic equations in (9-44) with m = n - 1 or m = n, depending on
whether g(s) is strictly proper or proper. Its solution yields immediately the
required compensator. This is illustrated by an example.
h g(s)=N(s)/D(s) and
Example 1
-m, a proper compen­
has transfer function Consider a plant with transfer function
mdm¿n-l.
A +1
S2
g\s) .2
s +2 s_ 2
. this theorem follows
Let C(s) be chosen as
then N n = O. Conse­
= Fn+",/D". Hence if
the compensator C(s) (9-47 )
Q.E.D.
Then the equation in (9-44) becomes

th g(s) =N(s)/D(s) and


+ m, a strictly proper
~k system has transfer
coprime and m ¿no
Since the square matrix in (9-48) is nonsingular, for any F¡, i = 0,1,2,3, a
solution exists in (9-48). In other w.ords, for any three arbitrary poles;a compen­
If and only if N(s) and sator of degree 1 always exists. However, for sorne set of chosen poles, the Del in
:he solution is unique, (9-48) may become zero. For example, if D¡(s) =2 +2s +0.5s 2 +3s 3 , then
lensator may become the solution in (9-48) can be computed as Dco = - 0.5, Del = O, N cO = 1, N el = 3,
n (9-44) is larger than or C(s) = (3s + 0/( - 0.5). This compensator does not have a proper transfer
vith m ¿n, is linearly function.
N ow we choose
may choose N cm = O.
Jmpensator is strictly 2
C(s) = N co +Nels +N c2 S
Q.E.D. Dco +Dels +D c2 S 2

;:re P(s) is an arbitrary in this case, (c;-44) becomes


46) is Dc(s) = P(s) and -2 2 1 O O
, and the compensator
1 O 100
em 9-11', ir m = n and ---------------------
;: strictly proper, then O -2 2 1 O
em 9-11' still holds if O 1 O 1 O
---------------------
ed in Section 3-6, the O O -2 2 1
;hecked in the designo O O 1 .0 1
hence 1'+ C( 00 )g(oo) i= =[F o F¡ F2 F3 F 4J (9-49)
11', g(.s)is proper, but
,e feedback system is The first five rQws of the 6 x 5 matrix in (9-49) are linearly independent and have
~anceÚations betv:.een a full column rank; hence for any F¡,i=O, 1, ... ,4, a solution {Dci,N ci with
lable by the designer; Dc2 = F~ aúd N c2 = O} exists in (9A9).ln other words, the compensator has a
strictly proper transfer function. This confirms with Theorem 9-11'. •
464 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

In Theorem 9-11, n - 1 is the smallest degree which a compensator must scribed by the q x 1
have in order to achieve arbitrary pole placement. In other words, for every
D ¡(s), a compensator C(s) of degree n - 1 always exists so that the unity feedback
system in Figure 9-12 has D¡(s) as its denominator. lf Df(s) is prechosen, then
it may be possible to find a compensator of degree smaller than 11 - 1 to achieve
the designo For example, if Df(s) is chosen as
D f(S) = 4s 2 + 2s + 1
in the example, then we have
[DeO Neo]So = [1 2 4] where D(s) the least (
and DeO = 1, NeO = 3. In other words, we can find a compensator C(s) =f of D(s)
degree O to achieve this particular pole placement. Hence, if the degree of a N(s)
compensator is extremely important, we_ may proceed as follows: First, we
where Dí are constar
form So and check whether the chosen D f(S) of degree n is in the row space
find a compensator
ofS o. lfyes, the design can be completed by using a compensator of degree O.
number of poles of 1
lfnot, we formSI and increase the degree of Df(s) by one to Df(s) = 15 f (s)(s +k).
assigned. Furtherml
Ir Df(s) is in the row space of SI for sorne k, and if the pole s + k of this k is
small as possible. FI
acceptable, then the design can be accomplished by using a compensator of
degree 1. Ir D¡(s) is not in the row space of SI' then we must increase m by 1
and repeat the process. For a study of this problem, the reader is referred to
Hence the transfer rr
References S97, S184, S21O, and S223.
equal to
The larger the degree of a compensator, the larger the number of parameters
available for adjustment to achieve design purposes. Ir the number of para­
meters is larger than the minimum required for arbitrary pole placemeni, the
Let us write the com~
spared parameters can be used to achieve other design objectives such as the
assignment of zeros or the minimization of sensit ivity functions. The interested 1
C(s)=-I
reader is referred to References S25 and S83 and Problem 9-29. DJ5)
In Theorems 9-11 and 9-11', if D¡(s) is chosen as a Hurwitz polynomial, then with
the feedback system is stable. Hence we conc1ude that every plant with a
proper transfer function can be stabilized by using a compensator with a proper
transfer function in the unity feedback configuration shown in Figure 9-12. where Dí are scalars
19-50) and (9-53) i'1tc (
5~g-og~e-Üli'1pijJr~ or sfirugoe-QlLf[Put case. In this section vveüiscüss i.·i-lé¿C.s¡5~·,;
rE; f(s) = N(s),
of compensators to achieve pole placement for single-input multiple-output and
multiple-input single-output systems. The general case will be postponed to
= [De(s
the next subsection. Because N(s) and N e (
Consider the feedback system shown in Figure 9-13(a). The plant is de­ and N(s)Ne(s) is a q x

Hence the problem of


This equatiori is agem
to (9-40), we shall trar:

Di(' .
(al (b)

Figure 9-13 Unity reedbacksystemswith single-input or single-output plant. 6 Note lhat, beca use of pos
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 465

1 compensator must scribed by the q x 1 proper rational matrix


her words, for every N'I(s)
at the unity feedback D'I(s)
:.'1) is prechosen, then
than n -1 to achieve G (s) =
A
N~(s) 1 6. I
D~(s) - D(s) N 2(S) = N(s)D - (s) (9-50)

N~(s)
D~(s)
where D(s) the least common denominator of all elements of C(s). We assume

lpensator C(s) of =t D(s) = Do +DIs +D 2 s2 + + DlIs n D n #=0

::e, if the degree of a N(s) = No + NIs + N 2 s 2 + + Nnsn (9-51 )

LS follows: First, we
where Di are constants and Ni are q x 1 constant vectors. The problem is to
! is in the row space
find a compensator with a proper transfer matrix of degree m so that n +m
pensator of degree O. number of poles of the feedback system in Figure 9-13(a) can be arbitrarily
o D fes) =15 f(s)(s +k). assigned. Furthermore, the degree m of the compensator is required to be as
,ole s +k of this k is small as possible. From Figure 9-13(a) we have
19 a compensator of
nust increase m by 1 u(s) = (1 + C(s)G(s))- I C(s)r(s)
reader is referred to Hence the transfer matrix of the overall feedback system in Figure 9-13(a) is
equal to
umber ofparameters
the number of para- G fes) = G(s)[1 +C(s)G(s)] - IC(S) (9-52)
pole placement, the Let us write the compensator C(s) as
bjectives such as the

ions. The interested


(9-53)
:1 9-29.

litz polynomial, then


with De(s) = Deo + Deis + + Demsm
every p\ant with a NcCs) = Neo + N eI S + + Nemsrn (9-54 )
:nsator with a proper
own in Figure 9-12.
where Di are scalars and N ci are 1 x q constant vectors. The substitution of
(9-50) and (9-53) into (9-52) yields

we d iscuss the design Gfes) = N(s)D- I(s)[1 '+ D e- I (s)Ne(s)N(s)D - I (s)] - ID; I (s)Ne(s)

multiple-output and
= [DcCs)D(s) + NcCs)N(s)] -1 N(s)NcCs) (9-55)
'Viii be postponed to
Because N(s) and Ne(s) are q x 1 and 1 x q vectors, Ne(s)N(s) is a 1 x 1 matrix
1). The plant is de- and N(s)Ne(s) is a q x q matrix. Hence Gfes) is a q x q rational matrix. Define 6
Df(s) = DcCs)D(s) + Ne(s)N(s) (9-56)

Hence the problemof pole placement reduces to the solving of Equation (9-56).
'This equationis a generalization ofthe Diophantineequation in (9-40). Similar
. to (9-40), we shall translate it into a set of linear algebraic equations. Let
(9-57)

(b)

~-Olltpllt plan!. 6 Note thal, because of possible pole-zero cancellalions, nol all rools óf D¡(s) are poles of G¡(s).
466 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

The substitution of (9-51), (9-54), and (9-57) into (9-56) and equating the co­ Proof
efficients of the same power of s yield
The design problen
F,,+mJ ~ F application of Theo
(9-58) there exists a set of
if the matrix Sm has
with definition of r¡ and t

Do DI D" O O
} one block row
No NI N" O O
----------------------------
O Do ... O Hence the conditior
D"_I D"
S",= O .. . ... O (9-59) This is the case if al
No N"_I NI!
---------------------------- (Corollary G-14). 1
---------------------------- any {F;}.
O O O Do DI Dn lf (;(s) is strictly
O O O No NI ...
N" solved as De", = FIl + m
is proper and if m =
We call the rows formed from {DJ D rows and the rows formed from {Ni} m;::: v, then the last q
N rows. Then every block row in (9-59) has one D row and q N rows. The Consequently, we ID;
matrix S", consists of m + 1 block rows; every block row is the shifting to the Fn+mlD". Hence the
right by one column of its previous block row. The matrix S,. is c1early a proof of this theorem
(1 + q)(m+ 1) x (n + 1 + m) matrix. Now we search the Iinearly independent
rows ofS,. in order from top to bottom by using, for example, the row-searching The application o
algorithm discussed in Appendix A. Then as discussed in Theorem G-13, aH in (9-50) have no COI
D rows in Sm wiH be linearly independent, and sorne N rows will be Iinearly form S", by using the
dependent on their previous rows in Sm' Let r¡ be the number of Iinearly independent rows in
dependent N rows in the i + 1 block row ofS",. Then because of the structure algorithm. Once v is
ofS",. we have O:::::; ro:::::; 1'1 :::::; :::::;q. Let l' be the integer such that,.o:::::; 1'1 :::::; •.. :::::; (;(s) is strictly proper.
r,. _ 1 < q and 1', = 1',+ 1 = = q. We cal1 l' the row index of C(.~). Then l' is Its solution yields th
the largest row degree of A(s) in any left-coprime fractíon of C(s) = A - 1 (s)B(s) solution of (9-58) can
with A(s) row reduced [see (G-80) and (G-81)J or the observability index of any
irreducible realization of (;(s) (dual ofTheorem 6-6).

TheOl'em S- -¡ T v'/t; éfúpn.asizt: ((le

Consider the feedback system shown in Figure 9-13(a) with the plant described ability index, the con
by a q x 1 strictly proper (proper) rational matrix C(s) = N(s)D-I(s) with hinges on the search
deg D(s) =n. Then for any D¡(s) of degree n +m, there exists a 1 xq proper S,. is a (l +q)(m + 1)
(strictly proper) compensator C(s) = D; I(s)Ne(s) with deg De (s) = m so that the we need (1 +q)(m + 1
feedback system has q x q transfer matrix N(s)Dj I(s)Ne(s) if and only if D(s) may start from the SI
and N(s) are right coprime and m;::: v - 1 (m;::: v), where v is the row index of note that aH remarb
(;(s), or the observai?i1ity index of any ir~educible realization of (;(s), or the For example, for a giv,
largest row degree of A(s) in any left~coprime fraction of (;(s) = A -1(s)B(s) with degree smaller than v
A(s)row reduced. of a compensator, it
other design objectivt
Dual to Theorem
71[ q = 1, this theorem reduces to Theorems 9-11 and 9-1\' and l' is equal.to 11. system shown in Figl
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 467

md equating the co­ Proof


The design problem is equivalent to the solving of Equation (9-58). The
application of Theorem 2-4 to the transpose of (9-58) yields that for every F,
Fn+mJ Do F
(9-58)
there exists a set of {Dei> NeJ or, equivalently, a compensator C(s) if and only
if the matrix Sm has a fu11 column rank, that is, pSm = n +m + 1. From the
definition of ri and the linear independence of a11 D rows, we have
m

pSm = (m + 1) + I (q - t"¡) (9-60)


le block row i~O

Hence the condition for Sm to have a rank of n + m + 1 is n = ¿?~ o (q - r¡).


This is the case if and only if D(s) and N(s) are right coprime and m;? v - 1
(9-59)
(Coro11ary G-14). This establishes the existence of Dei and N ei in (9-58) for
any {FJ
lf C(s) is strictly proper, then N" = O. Consequently, Dem in (9-58) can be
solved as D em =F"+,,jD n, and the compensator C(s) in (9-53) is proper. lf (;(s)
is proper and if m = v -1, then Dem in (9-58) may become zero. However, ir
vs rormed from {Ni} m;? v, then the last q rows of SOl are linearly dependent on their previous rows.
and q N rows. The Consequently, we may choose N em = O. For this choice, D em becomes D em =
is the shifting to the F"+,,jDI!" Hence the compensator C(s) is strictly proper. This completes the
latrix Sm is clearly a proof of this theorem. Q.E.D.
linearly independent
>Ie, the row-searching The application of this theorem is very simple. If D(s), N ¡(s), N 2(S), ... , N q(s)
in Theorem G-13, all in (9-50) have no common roots, then D(s) and N(s) are right coprime. We
row's will be linearly form SOl by using the coefficients of D(s) and N(s), and then search its linearly
~ number of linearly independent rows in order from top to bottom by using the row-searching
;;ause of the structure algorithm. Once v is found, we choose n + v - 1 poles ror the overa11 systems ir
;hthatro~"I:::;'" :::; C(s) is strictly proper. From these poles, we form Equation (9-58) with m = \'-1.
:x of G(.~). Then \' is Its solution yields the compensator. We note that in hand calculation. the
1 of (;(s) = A -1(s)8(s) solution of (9-58) can be obtained by continuing the row searching to
rvability index ofany

We emphasize that aithougnTneorem 'J-l L. iS stalCG in terms oí the observ­


ability index, the concept is not needed in applying the theorem. The theorem
:h the plant described
hinges on the search of m so that Sm in (9-59) has a full column rank. Since
(s) = N(s)D -I(S) with
Sm is a (l +q)(m +1) x (n +m +1) matrix, in order to have a full column rank,
exists a 1 x q proper
we need (l +q)(m + 1) ;?(n + m + 1) or m ;?(n/q) - 1. Hence in searching m, we
,De(s)= m so that the
may start from the smallest integer m ;?(n/q) - 1 rather than from m = O. We
(s) ir and only if D(s)
note that a11 remarks for the .single-variable case are equa11y applicable here.
v is the row index of
For example, for a given set ofpoles, it may be possible to find a compensator ofa
~ation of C(s), or the
degree smaller than v- 1tb achieve pole placement. By increasing the degree
G(s) = A -1 (s)B(s) with
of a compensator, it is possibJe to achieve, in addition to the pole placement,
other dcsign objectives. .
Dual to Theorem 9-12, we have the fo11owing theorem ror the feedback
.\ to 11. system shown in Figure 9-13(b).
468 LINEAR TI\1E-INVARIANT COMPOSITE SYSTEMS

Theorem 9-12' denominator of all é


of C(S).8 Let 6.(s) b(
Consider the feedback system shown in Figure 9-13(b) with the plant described
common denominat
by a strictly proper (proper) 1 xp rational matrix C(s)=D-l(s)N(s) with
Definition 6-1). C[¡
deg D(s) = n. Then for any D¡(s) of degree n + m, there exists a p x 1 proper
nomial h(s). lf Ó-(s)
(strictly proper) compensator CCs)=Ne(s)D;l(S) with degDe(s)=m so that the
rational matrix. Fo
feedback system has 1 x 1 transfer function N(s)Dj 1 (s)Ne(s) if and only if D(s)
and N(s) are left coprime and m ¿/l-l(m ¿/l), where /l is the column index of
C(s), or the controllability index of any irreducible realization of C(s), or the 1_1_ 2

Gl(s)= ·lS_1_
largest column degree of A(s) in any right-coprime factorization of C(s) = A +1 5+
B(s)A -les) with A(s) column reduced. III
o
s +1
The polynomial equation arises in this theorem is of the form then it can be readil
Every 1 x por q xiI
D¡(s) = D(s)De(s) + N(s)Ne(s) (9-61 )
(;(s) = tjJ - 1 (s)N(s)N e(.
Note that De and N e are on the right-hand side of D(s) and N(s), rather than the matrices, then C(s) i:
!eft-hand side as in (9-56). Equation (9-61) can be solved indirectly by taking its elements gij(S) of G(.I
transpose to become the form of (9-56) or solved directly as fol1ows: Using Consider the cycl
the coefficient matrices of D(s), N(s), Dcls), Ne(s), and D¡(s), we form the linear
algebraic equation
, ,
Do No O O O O
Deo , Deo Fo
DI NI , Do No , O O
Neo Neo Fl
,o
Del Del F2
,o O O We see that if al = ­
N el , , N cl
Tm ~ DII Nn - l··
, Do
°1
No (9-62) nomials of C2 (s) and
N" Dn -
: 1 1 ,
,
O O Dn N" DI NI teristic polynomials, l
D em Dem
the characteristic poI:
N em , N em Fn + m shows that in any liJ
O O O O D II
N II

m + 1 block columns canceling a pole is ve


theorem.
This equation is essential1y the transpose of (9-58) (see Theorems G-14 and
i : ':,;;"....:: .... ~.; . .._

to right. Let /l be the least integer such that the last p N columns of TI' are all
Consider a q x p cycl
linearly dependent of their left-hand-side columns. The /l will be called the
and 1 x q real consta
column index of C(s). It is equal to the controllability index of any irreducible
realization of C(s) or the largest column degree of the column-reduced A(s)
in any right-coprime factorization of C(s) = B(s)A -l(S). The proof ofTheorem
where 6.(. ) denotes tt
9-12' is similar to the one of Theorem 9-12 and will not be repeated.

Multivariable case: Arbitrary pole assignment. In this section, the


. design technique de'leloped in the orevious silb~ection will be exten1ed to 8 Lel [A, B, C, E} be an ¡rre

.general proper ratiOnlJ.l matrices. We extend it .first to a special class of rational polynomial of A. Comp.

matrices;called cydic ratiónaJmatrices, ando thentothe general case. matrix in Prob\em 2-45.

Consider a q x pproper rational matrix G(s). Let tjJ(s) be the least common 9 Compare this theorem w

DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 469

denominator of all elements of C(s) or, equivalently, of all minors of order 1


of C(S)8 Let Li(s) be the characteristic polynomial of C(s) defined as tht least
h the plant described
common denominator of all minors of orders 1,2, ... , min (p, q), of C(.}) (see
~s)=D-¡(s)N(s) with
Definition 6-1). Clearly, in general, we have Li(s) = lf¡(s)h(s) for sorne poly­
exists a p x 1 proper
nomial hes). Ir Li(s) = lf¡(s) k for sorne constant k, then C(s) is called a cyclic
Dc(s) = m so that the
rational matrix. For example, if
.(s) if and only if D(s)
. the column index of
~ation of C(s), or the 1

l
í_1_
~1 -J
:torization of C(s) =

:le form
I
A

l S
GI(s)= _1_
s
+1

+ 1

A
Gz(s)=
s+1
_1_
s
s +2

s +1 s +1
then it can be readily verified that Cz(s) and C,(s) are cyclic, but C¡(s) is noL
Every 1 x p or q x 1 proper rational matrix is cy;lic. If C(s) can be expressed as
(9-61 )
C(s) = lf¡ -1(s)N(s)NcCs), where N(s) and Nc(s) are q x 1 and 1 x p polynomial
N(s), rather than the matrices, then C(s) is cyc1ic (why?). Ir no A. appears as a pole of two or more
idirectly by taking its elements gij(s) of C(s), then C(s) is cycli<l (Problem 9-14).
Iy as fol!ows: Using Consider the cyc1ic rational matrix Gz(s). We form
), we form the linear

o
o
Fo
F¡ .
(;z(s)a

.
~
-
Gz(s) [al]
az
= l a~ :~z J
azs +(2az
s +1
+al~

Fz
We see that if al = -az, then Gz(s)a = [O -az]' and the characteristic poly­
(9-62) nomials of Gz(s) and Gz(s)a are differenL For a with al i= - az, their charac­
teristic polynomials, however, are eq ua!. Recall that for cyc1ic rational matrices,
the characteristic polynomial is eqlial to the minimal polynomia!. This example
Fn + m shows that in any linear combination of the columns of Gz(s), the chance of
canceling a pole is very smal!. This is true in general as state in the following
theorem.
Theorems G-14 and
Tm in order fram lelt
~olumns of T" are all
Consider a q x p cyc1ic proper rational matrix C(s). Then for almost al! p x 1
Ji. will be called the
and 1 x q real constant vectors t ¡ and t z, we have
ex of any irreducible
:olumn-reduced A(s) Li[C(s)] = Li[C(s)t¡] = Li[tzC(s)] (9-63)
he proof of Theorem
where Li(') denotes the characteristic polynomial of a rational matrix. 9
repeated.

In this section, the


will be exten1ed to 8 Let {A, R, C, El be an irreducible realization of G(5). Then it can be shown that t/J(s) is the m¡"nímal
e<;:ial c1ass ofrational polynomial of A. Compare the definítion of cydic.raÚonal matrix with theone for'a constan! A
:neral case. matrix in Problem 2-45.
be the least common 9 Compare this theorem with Theorem 7-5.
470 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

Proof Theorem 9-14


Let us write Consider the feedba
by a q x p cycIic stri,

G(s) =_1 N(s) =_1


l{I(s) l{I(s):
l~~i~~l compensator is aSSUI
C(s) of degree m. If
of the unity feedbac
Ni s) respectively. the cor
where l{I(s) is the least common denominator of all elements of (;(s); N(s) is a realization of G(.s) 01
q x p polynomial matrix, and N¡(s) is the ith row of N(s). We assume that
every element of G(s) is irreducible. The cyclicity assumption of G(s) implies Proof
8[G(s)J = l{I(s). Let A¡, i = 1, 2, ... ,m, be the distinct roots of l{I(s). First we Since Gts) is cyclis. b
assume that A¡, i= 1,2, ... , m, are real. Then we have that 8[ G(s)J = 8[ G(.'
N(A¡} 1=0 fori=1,2, ... ,m
Hence the rank of N(A¡) is at least 1. Consequently, the null space of N(A¡) has
Then Theorem 9-12
a dimension of at most p - l (Theorem 2-5). In other words, the set of vectors t
C(s) = D c- l(s)N..{s) wi
in the p-dimensional real vector space ~ P which satisfy N(A¡}t = Ois a linear space
alln +m roots of
of dimension at most [J - 1. To help to visualize what will be discussed. we
assume p = 2. Then those t satisfying N(A¡)t = O, for each i, is at most a straight
line in ~ 2. Consequently, there are at most m straight lines on which N(A¡)t = O
for sorne i. The tI not on these straight Iines have the property can be arbitrary assi¡
of the feedback systel
for all i = 1,2, ... , m
This implies that l{I(s) and N¡(s)t¡, i = 1, 2, ... , q, in

~
N ,IS)t¡]
'\. _ ~ N 2 (s)t 1
GI.\)t¡ - ,',
'I'(s) :
Nq(s)t l
have no common factor. Hence l{I(s) is the charactcristic polynomial of G(s)t¡.
This establishes 8[ G(s)] = 8[ G(s)t lJ for the case where aH roots of l{I(s) are real.
ir sorne of ;'j, i = 1, 2, ... , in, are not real, the preceding proof stil! holds if we
use N(A¡} +N(},t). This completes the proof of .1[G(s)J =.1[G(s)t¡J.J:"óe i:
other part of the theorem can be similarly proved.
Now we show that (9-63) holds for almost al! tI in ~p. For easy visualization,
we assume p = 2 and consider only tI with IIt l l1 2 = 1. Then from the preceding
proof, we see that at most 2m points on the unit circle of ~ 2 will not meet (9-63),
the rest ofthe unit circle (infinitely many points) will meet (9-63). Hence almost
every vector tI in ~2 has the property 8[(;(S)] =8[G(s)t l J. Hence ifa vector
tI is. chosen arbitrarily or generated randomly, th~ probability of having
8[G(s)] == 8[G(s)tlJ is almost équal to 1. This establishes the theorem. Q.E.D.

The cy~licity of (;(s) is essenltal in this theorem. Ir (;(5)' is not cyclic, the
theüi-em d~esnot h61d in general (tr"y G 4 (s) in Problem 9'-13).. UsingTheorem
9-13, we can now extend the design procedure in Theorems9-12 and 9-12' to
cyclic rational matrices. Figure 9-14 Design (
DESIGN Of COMPENSATORS: UNITY FEEDBACK SYSTEMS 471

Theorem 9-14
Consider the feedback system shown in Figure 9-14 with the plant described
by a q x p cyc1ic strictly proper (proper) rational matrix G(s) of degree n. The
compensator is assumed to have a p x q proper (strictly proper) rational matrix
C(s) ofdegree m. lf m ~min (jl-l.l' - 1) [m ~min (}l, v)], then aH n +m poles
of the unity feedback system can be arbitrarily assigned, where jJ. and vare,
respectively, the controllability and observability indices of any irreducible
realization of G(s} or the coJumn index and row index of G(s).
:nts of G(S); N(s} is a
s). We assume that
Proof
ption of (;(s) implies
)ts of ¡Jt(s). First we Since Gis) is cyc1ic by assumption. there exists a f' x l constant vector tI such
that Ll[ G(s)] = Ll[ G(s)t 1 J. Lel us wrile l he q x I ral ional matrix e(s)t I as
GIs)t¡ = NIs)D-11s)

u\! space of N(A.¡) has Then Theorem 9-12 imp\ies lhe exislence of a 1 XI/ proper rational matrix
ds, the set of vectors t (:(s) = D c- 1 (s)N,(s) with deg Ós) = 111 ~ l' - I if G(.~) is strict\y proper. such that
¡)t = O isa linear space all n +m roots of
will be discussed. we
L, is at most a straight D Il~) = D.(s)D(s) + Nrl-~)N(s)

:s on which N(A.i)t = O can be arbitrary assigned. Now we show that the roots of Df(s) give the poles
aperty of the feedback system in Figure 9-14(a). Indeed, from Figure 9-14(a), we have
h(s) = (:(s)C(s) =C(s)[r(s) - (;(s)t¡h(s)]

,----------1
¡ ¡
e I I Y

I
1
polynomial of G(s)t¡. I
L
I
-.J
roats af 1/1 (s) are rea\.
)roof sti\\ ho\ds if we
sl] = Ll[G(S)t¡J. The iI
(al
¡:< ar easy visualization,
~n from the preceding
2 will not meet (9-63),

(9-63). Hence almost


1]. Hence if a vector
I
lrobability of having I I
L .J
the thearem. Q.E. D.

G(s) is not cyc1ic, the


-i3). Using Theorem lb)
~ms 9-12 and 9-12' to
Figure 9-·14 Design of compensators for plant with cyclic proper rational matrix.
472 LINEAR T1ME-INVARIANT COMPOSITE SYSTEMS

which implies Theorem 9-15


hes) = [1 + C(s)G(s)t¡] - ¡ C(s)r(s) Consider a q x p pro pI
and G¡(s)=G(s)t¡[1 +C(s)G(s)t¡]-¡C(s) (9-64) every p x q constant rr
A

The substitution of G(s)t¡ = N(s)D- ¡(s) and C(s) = D,-¡(s)N,(s) into (9-64) yields G(s) =
G feS) = N(s)D- ¡(s)[l + D,- ¡(s)N,(s)N(s)D- ¡(s)] -¡ D c- ¡(s)NcCs) is proper (strictly pro~
= N(s)[D,(s)D(s) + Nc(s)N(s)] -¡ Nc(s) (9-65)
10
Proof
where we have used the fact that Dc(s) and D(s) are l xl polynomial matrices.

From (9-65) we conclude that the roots of D feS) give the poles of (9-65). Hence We show that the roc

the qxp compensator defined by cts)=t¡C(s)=D;¡(s)t¡Nc(s) can achieve distinct for almost all ]

arbitrary pole placement.

Now we show that the observability index of G(s)t¡ is equal to the one of
C(s). If G(s) is factored as G(s) = A -l(s)B(s), where A(s) and B(s) are left coprime where Gi, i = 1, 2, ... ,n
and A(s) is row reduced, then the observability index of G(s) is equal to the tion of 3.(s) with respe(
largest row degree of A(s) (see Theorem 6-6). Consider G(s)t 1 = A -¡(s)B(s)t¡.
The condition ~[G(s)] = Ll[G(S)t1] =det A(s) implies that A(s) and B(s)t¡ are 3.' (s)
left coprime. Hence the observability index of G(S)t1 is also equal to the largest If ~(s) has repeated roc
row degree of A(s). This establishes that the observability index of G(s) and sufficient condition for
that of G(s)t 1 are the saine. Hence we have deg C(s) ¿ v - 1, where v is the
observability index of G(s) or C(s)t 1. Since cts) = t¡ C(s) is cyclic, we have
deg cts) = deg C(s) ¿ v - 1. This completes the proof of one part of the theorem.
The rest can be similarly proved. Q.E.D.
det O
From (9-65), we see that the transfer matrix from r to y is of the form
N(s)Dc-¡(s)Nets), as in the vector case. However, the N(s) and Nc(s) in this
design are not unique; they depend on the choice oft¡. Although the degree of o
compensators in Theorems 9-12 and 9-12' are minimal to achieve pole place­ (See Appendix G). W
ment, the degree in Theorem 9-14 may not be minima\. In other words, it may nomial of kij. There is
be possible to design a compensator of degree less than min (J1- 1, v -1) to of {kij} as a vector in
achieve arbitrary pole placement for a q x p cyclic proper rational matrix. clear that the solution,
What is the minimum degree seems to be a difficult problem. almost every k¡j, we ha'
With Theorem 9-14, we can now discuss the design of compensators for rrhis Ís Si!:.~.712.r t0 th2J
general proper rational matrices. The procedure consists of two steps: First are distinct.) Hence G
change a noncyclic rational matrix into a c}'clic one and then apply Theorem If G(s) is strictly prc
9-14. Consider a proper rational matrix G(s). Let Ll(s) be its characteristic the condition for G( "
polynomia\. We claim that if all roots of Ms) are distinct, then G(s) is cyclic. det (1 + C( 00 )K) '" O. fl
Let if¡(s) be the least common denominator of all elements of G(s). Then we proper, so is G(s) " for al
have Ll(s) = if¡(s)h(s) for sorne polynomial hes). If all roots of Ll(s) are distinct,
thell we have Ll(s)=if¡(s)k for sorne constant k, forAif¡(s) must contain every root
of G(s). Hence G(s) iscyclic. Note that a cyclic G(s) may have repeated poies. With this theorem,
Hence the condition thatall roots of ~(s) are distinct is a sufficient but not placement for general
necessary condition for G(s) to be cyclic. This properiy ~v.i11 be used to establish gajn output feedback to
that every noncyclic proper rational matrix can be transformed into a cyclic .
one by introducinga constant giliri feedback from the outpUt tothe input, as
stated in the following theorem. ¡o The proof is identical lo lh
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 473

Theorem 9-15

Consider a q x p proper (strictly proper) rational matrix G(s). Then for almost
} (9-64) every p x q constant matrix K, the q x p rational matrix
'lAs) into (9-64) yields G(s) =[1 +G(s)K]-IG(s)=G(s)[1 +KG(S)]-l
1D c- 1(s)Nc(s) is proper (strictly proper) and cyclic.
(9-65)

Proof'O

polynomial matrices. A

oles of (9-65). Hence We show that the roots of the characteristic polynomial, ~(s), of G(s) are all
5)t l Nc(s) can achieve distinct for almost all K. Let
~(s)=aOsn +als"-l + ... +a"
s equal to the one of

d B(s) are left coprime where a¡, i = 1, 2, ... , n, are functions of all elements, kij, of K. The differentia­

f G(s) is eq ual to the tion of ~(s) with respect to s yields

G(s)t l =A -1(s)B(s)tl'

3.'(s) = naos"-l + (n - l )als"- Z + ... + a,,- 1


It A(s) and B(s)t l are
;0 equal to the largest If ~(s) has repeated roots, then ~(s) and ~'(s) are not coprime. A necessary and
ity index of (;(s) and sufficient condition for ~(s) and ~'(s) to be not coprime is
.v - 1, where v is the O O
a" a n- l al ao
(s) is cyclic, we have
O an az al ao O
le part ofthe theorem.
Q.E.D. det O O a a u -: 1 =y(k¡¡) =0.
- - - - - - - - - --- - - - - - -n- - - - - - - - - - - - - - - - - -ao
. to y is of the form an-l 2a n- z nao O O
~(s) and N~(s) in this
.lthough the degree of O O a tl - 1 2a,,_z nao
o achieve pole place­ (See Appendix G). We note that y(kij) is generally a nonhomogeneous poly­
n other words, it may nomial of kij' There is a total of p x q number of kij in K. Hence we may think
1 min (¡.t - 1, v -1) to of {kij} as a vector in the (p x q)-dimensional real vector space ~PXq It is
)per rational matrix. c1ear that the solution, k¡j, of y(k¡j) = O is a subset in ~ p x q. In other words, for
lem. almost every k¡j, we have y(kij) f- O. Consequently, all roots of ~(s) are distinct.
of compensators fOl" (This is similar to tha!.for ?J.rnast ~·dl C~:., C~~~, ~lr.-cl u., t~i..~ '((";'0 1',3 ~f(x~:,'2 +C~2·:.'+··"'·
ts of two steps: First are dis}inct.) Hence G(s) is cyclic" "
then apply Theorem lf G(s) is stríctly PIoper, so is G(s) for every K. (Prove it.) If G(s) is proper,
) be its characteristic the condition for G(s) to be proper, as discussed in Section 3-6, is
ct, thell (;(s) is cyclic. det(1 + G( 00 )K) 1= O. Almost all K satisfy this condition. Hence if G(s) is
ltS of G(s). Then we '" for almost all K. This completes the proof of this theorem.
proper, so is G(s)
ts of ~(s) are distinct, Q.E.D.
1st contain every root
y have repeated poles. a
Wíth thís theorem, the desígn of compensator to achíeve arbitrary pole
s a sufficierit but not . placement foro general (;(s) becomes obvious. We first introduce a constant
. ~. Á A

ill be used to establish . gain outP\lt feedback to make G(s) = [1 + G(s)K] -IG(S) cyclic. We then apply
sformed :intoa cyclic
utput to the input, as
10 The proof is identical lo the one of Theorem 7-6.
474 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

Theorem 9-14 to design a compensator;.. C(s) of degree m=?:min(¡1-1,v-l) given by


or In =?: min (¡1, v) depending on whether C(s) is strictly proper or proper, where
Ji. and vare, respectively, controlIability and observability indices of G(s). Gn (s) = [1 +
Hence all poles of the feedback sy¡;tem in Figure 9-15(a) can be arbitrarily =[1 +1
assigned. =[1 +,
The feedback configuration in Figure 9-15(a) is not exactly a unity feedback
The substitution of G(
system. HO';Yever, if we are concerned only with the poles of CAs), not the
structure of GAs), we may combine the parallel connection of C(s) and K into G o,
J l.
(s)1-- in-
t.iL...l' j

C(s) and the configuration reduces to the unity feedback system shown in X D­
Figure 9-15(b). = [D(s).

By a similar manipulal
Theorem 9-16
can be computed as
Consider the feedback system shown in Figure 9-15ib) with the plant described
Gfes) = [D(s)De(s) + N(~
by a q x p strictly proper (proper) rational matrix G(s) of degree n. The com­
pensator C(s) is assumed to have a p x q proper (strictly proper) rational matrix From (9-66) and (9-67),
of degree m. If m =?:min (J.1-1, v -1)[m =?:min (J.1, v)], then al1 n +m poles ll of have the same set of po!
the unity feedback system in Figure 9-15(b) can be arbitrarily assigned, where J.1 assigned by a proper ch
and vare, respectively, the control1ability and observability indices of any C(s) + K, the poles of n
irreducible realization of G(s), or the column index and row index of C(s). Now we c1aim that d,
be an irreducible realiza
Proof x = Ax +Bu, y = ex + (J
Hence we have deg de
First we show that if C(s) = C(s) + K and if C(s) can be written as C(s) =
proved by using Theo rel
De-l(s)tl Nis), where t l is a p x 1 constant vector and NcCs}is a 1 x q polynomial implies that the parallel (
matrix, then the poles of the system in Figure 9-15(a) and those in Figure 9-15(b) Hence we ha ve deg det «
are the same. The transfer matrix of the feedback system in Figure 9-15(a) is
What remains to bé
indices of C(s) and (;(s)
11 Because of possible pole-zero cancellations, not all n + m of these roots will be the poles of the for almost every consta
resulting system. However, for convenience, we shall call all of them poles. The strictly proper N(s) and D(s) are right c
part of this theorem was first established in Reference S26 in the state variable approach and in lability index of any irree
Reference S51 in the transfer-function approach. degree of D(s) (see Theol
'" = l\!(s)lDJ-
G(s) !r·
Since N(s) and D(s) are
(;(5)

A
and Ves) such that X(s)O
¡----------l

I + I y
X(s)[DI
Hence Theorem G-S imp
I
I
column degrees of D(s) al
I for all K if N(s)D -les) is
I

I I
proper. .Hence the contr
L -.J
facl <;an also be proved b
can besimilarly proved.
(al (b)
The tr:ansfer matrix G
Figure 9-15 Design of compensator.
Unlike the cases of single
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 475

m ~ min (ji - 1, V- 1) given by


oper or proper, w~ere GJI (S) = [1 + G(s)C(s)] - 1 G(s)C(s)
lbility indices of G(s). = [1 + (1 + G(s)K) - 1G(s)C(s)] - 1(1 + G(s) K) - 1G(s)C(s)
{a) can be arbitrarily
= [1 + G(s)K + G(s)C(s)]- 1G(s)C(s)

:actly a unity feedback The substitution of G(s)=D- 1(s)N(s) and C(S)=D;I(S)t 1 N e (s) yields
10Ies of GAs), not the GJI (S) = {D - 1 (s)[D(s)De(s) + N(s)KDe(s)+ N (s)t 1 Nc(s)]D c' 1 (s)} -1
ion of C(s) and K into X D-I(s)N(s)D e- 1 t 1 N e(s)
ack system shown in = [D(s)De(s) + N(s)KDe(s) + N(s)t 1 Nb)]' 1 N(S)tl Nls) (9-66)

By a similar manipulation, the transfer matrix of the system in Figure 9-1S(b)


can be computed as

ith the plant described G J(s) = [D(s)De(s) + N(s)KDe(s) + N(s)t 1Ne(s)] -1 N(s)[t 1NAs) + KDe(s)] (9~67)
)f degree n. The com­ From (9·66) and (9-67), we conclude that the systems in Figure 9-15(a) and (b)
)roper) rational matrix have the same set of poJes. 11 Since the poles in Figure 9-15(a) can be arbitrarily
11
;en all n+m poles of assigned by a proper choice of C(s) and K, we conclude that by choosing (:(s) =
'arily assigned, where J1 C(s) +K, the poles of the system in Figure 9-15(b) can be arbítrarily assigned.
ability indice~ of any Now we claim that deg det C(s) = deg det Ces). Let x= Ax + Bu, y=Cx +Eu
row index of G(s). be an irreducible realization of C(s); then we have deg det C(s) = dim A. Clearly
x = Ax + Bu, y = Cx +(E +K)u is an irreducible realization of (:(s) = C(s) +K.
Hence we have deg det C(s) = dim A = deg det (:(s). This fact can also be
be written as C(s) = proved by using Theorem 9-3. If we :vrite K = KI.¡-I = 1; lK, then Theorem 9-3
s) is al x q' polynomial implies that the paral1el connection of C(s) and K is controllable and observable.
those in Figure 9-15(b) Hence we have deg det (:(s) = deg det C(s) +deg det K = deg det C(s).
em in Figure 9·1~(a) is What remains to be proved is that the control1ability and observability
indices of (;(s) and (;(s) = [1+ G(s)K] - 1G(s)= G(s)[I+KG(s)] - 1 are the same
for almost every constant K. Ir G(s) is factored as G(s) = N(s)D-I(S), where
'oots wíll be the poles of the N(s) and D(s) are right coprime and D(s) is column reduced, then the control­
m poles. The strictiy prop~r
lte varíable approach and ID
lability index of any irreducible realization of 6(s) is equfll to the largest column
degree of D(s) (see Theorem 6-6). Using this fraction, (;(s) becomes
G(s) = N(s)D-1(s)[I +KN(s)D- 1 (s)] -1 = N(s)[D(s) +KN(s)]-\
Since N(s) and D(s) are right coprime, there exist polynomial matrices X(s)

and Y(s) such that X(s)D(s) + Y(s)N(s) = I (Theorem G-S). We modify it as

X(s)[D(s) + KN(s)] + [Y(s) - X(s)K]N(s) = I


Hence Theorem 0-8 implies that N(s) and D(s) +KN(s) are right coprime. The
column degrees of D(s) are dearly equal to the column degrees of D(s) +KN(s)
for a11 K if N(s)D-I(s) is strictly proper, and Joralrno~t ~.ll K if N(s)D-1(s) is
proper. Hence the controllability índices of G(s) and G(s) are the same. This
fact CilO also be proved by using dynamical equations:, The observability part
can besimilarly proved. This completes the proof of this'iheorem. Q.E.D. ,

Thetransfer matrix G¡(s) of the resulting feedback system isglven in (9-:67).


Unlike the cases of single input, single output, and cyclic plant, G/s) is not in
476 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

the form of N(s)D¡ 1 (s)NAs) with scalar D f(s). However, from the design We compute
procedure, we may conclude that G¡(S) in (9-67) iscyclic. See also Problem 9-16.
We recapture the design procedure in the following:
(;(s) = [1 + G(s)K
Step 1. Find a K such that G(s) = (1 + G(s)Kr 1 G(s) is cyclic.

Step 2. Find a",tI so that ~[G(s)] = ~[G(S)tI].


The minimal polynon
Step 3. Write G(s)t I = N(s)D-I(s) =(N o + NIs+'" + Nllsn)(D o + DIs + ... +D"s")- 1
teristic polynomial of
and form the SIlI in (9-59). Find the least integer \' such that Sv- ¡ has a [ull 1
column rank. The integer vis called the row index or the obseruability indexo (S3 + W [(s-+
Step 4. Choose n +1.' - 1 number of poles and compute
1
~¡(s) = Fo +F IS + F lSl + ... +FIl+v_ISn+v-I (53 +3)1 [(s'¡
1
Step 5. Solve Dci and N ci from (;"3+W [s(s-

[DcoN co : ... : Dc(v-I)Nc(V-ll]Sv-1 =[F o F I F1 Hence the characteri~


minimal polynomial.
Then the compensator is given by
two outputs; hence t
C(s) =tlC(s) +K observability indexo \

whereC(s)=(Dco+Dc¡S+'" +Dc(v- I)SV-I )-I[N co + Nc¡s+ ... + NC(V-llSV-I]


111
A A

In step 2, if we find a t z such that ~[G(s)] = ~[tzG(s)], then we must modify


steps 3 and 5. In step 3, we form T m as in (9-62) and search its linearly inde­
pendent eolumns in order from left to right. Let J1 be the least integer such
that T¡t-I has a full row rank. The integer J1 is called the column index or the
controllability indexo In step 5, we solve C(s) from the linear algebraic equation
in (9-62). Then the compensator is given by C(s) = C(s)t 1 + K. The matrix T o has a fu
Consider a q xp proper rational matrix G(s) of degree n. Ir its control­ iSJ1-1=O. Let
lability and observability indices are roughly of equal lengths, then we have
J1 znjp and \' znjq (see Equations (5-31) and (5-51)). Hence if p "2:.q, we use
t 1 ; otherwise, use t l' By so choosing, we may achieve min {J1- 1, v -l} without
computing explicitly both~L and V.
vi dei'e .l.ll.é.. s Lqjtt'SCC i Pl~

Example 2 poles are eh osen arbitr


Consider ~f(S)

The solutions of

(;(s)
.
l ~ ~1
= si,
.O O ­s

lt is a noncyclie rational matrix of degree3. Rather arbitrarily, we choose the are D co = 1, Mo ==1, N~
constant ·output feedback gain as 9-IS(b) is given by .

C(s) = C(s)t 2 -t
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 477

:ver, from the design We compute


See a1so Problem 9-16.
(;(s) = [1 + G(S)K]-IG(S) = ~3 [s + 1
S3 + _2
s(s + 1)
- 2s
1]
S2 - s + 1
·clic. The minimal polynomi,¡l.l of (;(s) is clearly equal to S3 + 3. To find the charac­
.Do+D1s+"'+ D "s")-1 teristic po1ynomia1 of (;(s), we compute its minors of order 2 as
,ch that S. _ 1 has a fun 1
the observability indexo
(S3 + W [(s + 1)( -2s) - ses + 1)( -2)] =0
1 [2 ] 1 3 1
(S3+W (s+1)(s -s+1)+2 =(S3+W(S +3)=(S3+3)
1 1 s
(' W[s(s+1)(s2- s + 1)+2s]= 3-3)2 S(S3+ 3 )=_,-­
s' + (s + (s' +3)
A

2"- F'I+V-¡] Hence the characteristic polynomial of (;(s) is S3 + 3, which is equal to the
minima1 po1ynomial. Hence (;(s) is cyclic. The system has three inputs and
two outputs; hence the controllability index is probab1y smaller than the
observability indexo We choose t 2 = [1 O]. Then we have
... 1
t 2 G(s)=-,-[s+1 s(s+l) 1]
;+ ... + Nc(v-IlS .-1] s' +3
I and L1[ (;(s)] = L1[t 2 (;(s)]. We form, as in (9-62),
], then we must modify
earch its linear1y inde­
: the 1east integer such T=l~~~ ~
o 0:0 1 O b
le column index or the
lear a1gebraic equation
1:O O O J
t 2 + K. The matrix T o has a full row rank; hence J1 = 1, and the degree of compensator
:gree n. Ir its control­ is J1 - 1 = O. Let
1engths, then we have
NJO]
[ N~o
Hence if p '2:q, we use C(S)=f
n {J1-1, v -l} without cO N3
co
wh¡;re [be; superscrlfils denote the componems oí Nc(.s).iile n +UJ. - l} = j

poles are chosen arbitrarily as -1, -1, and - 2. Hcnce we have


L1 f (s) =(s + lf(s + 2) = 2 + 5s + 4s 2 +S3

The solutions of

Jitrarily, wechoose the Hence the compensator in Figure

-1] r2 ., . , 1J
O = 3 O
1 O 1
478 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

It is a compensator of degree O. As a check, we compute We shall do the same


Let G(5) = N(s)D­
1 [(S+1)(3S+2) 1 ]
A

G¡(s)=[I+G(s)C(s)]
A N _lA c-.J
G(S)C(s)=(s+1)2(S+2) O (s+1)(s+2) G/(s) = N(s)I
= N(s)[
Its poles are -1, -1, and -2. We note that G¡ is cyclic (verify it). I
Define the polynomis
To conclude this subsection, we mentíon that although we can place arbitrary
the poles, the resulting G¡(s) is always cyclic. 12 In other words, the structure
Then we have G¡(s) =
of GAs) is restricted; we have no control over it. This restriction will be re­ Given D(s) and N(s) a
moved in the next subsection. This is the matrix ver
(9-40), we shall transla
Multivariable case: Arbitrary denominator matrix assignment. 1 3
In this subsection, we study the design of compensators to achieve arbitrary
denominator matrices. If we can assign an entire denominator matrix, then
certainly we can achieve arbitrary pole assignment. Hence this problem
accomplishes more than the pole assignment discussed in the previous subsec­ We also write
tion. Consequentiy, the degrees of compensators required for arbitrary denomi­
nator matrix are generally much larger than the ones required for pole assign­
mento
Consider the unity feedback system in Figure 9-16. The plant is described and
by a q x p proper rational matrix G(s). The compensator to be designed is
required to have a p x q proper rational matrix C(s). Let G ¡(s) be the transfer The substitution ofthe
matrix of the overall system. Then we have
G ¡(s) = [1 + G(s)C(s)] - 1G(s)C(s)
= G(s)C(s)[1 + G(S)C(S)]-1
where
= G(s)[1 + C(s)G(s)] - 1C(S) (9-68)

The first equality is obtained from y(s) = G(s)C(s)[r(s) - y(s)]; the second one Do D1 .
from e(s) = res) - G(s)C(s)e(s); and the third one from o(s) = C(s)[r(s) - G(s)u(s)]. No NI .
(Verify them. Compare also with Theorem 3-3.) In the single-variable case, S
if we assign both poles and zeros of the unity feedback system, then the design
= [ O Do'" [
O No'" l'

----- ---------
ni

will generally involve undesirable pole-zero cancellations. In arder tú avoict ! - - - - --------­


"i:hese canceilatioús, v<lé Clssigií only t~le. poles and leave the ZéfOS UGSpéCIÍi(;u.

13
12 Compare with Method 1 of the multivariable case in the design of state feedback.

Follows closely References 549 and 5237.

l
O
O
The matrix Sm has m +
'"
...
O

from Di and q rows fon


N rows. This matrix Sn¡
that if N(s)D- 1 (s) is pn
independent of their pi
may be lineariy depenc
linearly dependent N ro
ofS m ,. we have ro :$1'1 ::'S:

14 The integer J1 is the largest .


Figure 9-16 Multivariable feedback system. single-input case, however ,
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 479

We shall do the same for the multivariable case.


Let G(s) = N(s)D- I(S) and let C(s) = D; l(s)N e(s). Then (9-68) implies
3s +2) G¡(s) =N(s)D- 1 (s)[I + D; l(s)N e(s)N(s)D- 1 (s)] -1 D e- 1 (s)N e(s)
=N(s)[De(s)D(s) +Ne(s)N(s)]-INclS) (9-69)
ic (verify it). Define the polynomial matrix

.we can place arbitrary D¡(s) = Dcls)D(s) + Ne(s)N(s) (9-70)


:r words, the structure Then we have G¡(s) = N(s)Dj 1 (s)Ne(s). Hence the design problem becomes:
restriction wil\ be re­ Given D(s) and N(s) and an arbitrary D¡(s), find De(s) and Ne(s) to meet (9-70).
This is the matrix version of the Diophantine equation in (9-40). Similar to
(9-40), we shall translate it into a set oflinear algebraic equations. Let us write 14
trix assignment.' 3

'S to achieve arbitrary D(s)=D o +D 1 s + ... +Dl'sl' (9-71 a )


Jminator matrix, then N(s)=N o +N 1 s+'" +Nl'sl' (9-71b)
Hence this problem
n the previous subsec­ We also write
j for arbitrary denomi­
quired for pole assign- De(s) = Deo + Dcls + + Demsm (9-72a)

Ne(s) = Neo + Ncls + + Nemsm (9-72b)

The plant is described and D¡(s) = F o + F 1s+ F 2S 2 + ... + F mw S '" +1' (9-73)

ator to be designed is
~t G¡(s) be the transfer The substitution of these into (9-70) yields
[Deo Neo: Dc¡ Nc¡: ... : D em Nem]Sm=[F o F 1 F2 ... Fmw ] ~ F
(9-74)

where
(9-68)

. y(s)]; the secpnd one Do DI DI' O O ... O }ro(numberofdependent


= C(s)[r(s) - G(s)u(s)]. No NI NI' O O .,. O
le single-variable case,
;ystem, then the design S,.,= [ ~ ~o ::: ~1'-1 ~I' ~ ::: ~ }r/.o WS) (9-75)
-- O 1'- 1 l' 1

-----~ - -e:---- ---:--~:~:-~:k


ns. In order to avoid -- ------------------------------

the zeros unspecified.


l-~
,tate feedback.
The rnatrix Sm has m + 1 block rows; each block row consists of p rows formed
from Di and q rows formed frorn Ni' We call the former D rows and the latter
N rows. This rnatrixS mis studied in Appendix G. It is shown in Theorem G-13
that if N(s)D- 1 (s)is proper, then all D rows in Sm, m =0,1,2, ... , are Iinearly .
independent of their previous rows. Sorne N rows in each block, however,
y
rnay be linearly dependent on their previous rows. Let r¡ be the number of

r
linearly depertdent N rows in the (i + 1)th block. Then because of the structure
ofS m, we have ro S;rl S;'" S;r m S;q. Let v be the least integer such that r~ =q.

14 The integer J1 is thelargest colunm degree of D(s). It is different from deg det D(s) = n. In the
single-input case, however we have J1 = n.
480 LINEAR TIME-INV ARIANT COMPOSITE SYSTEMS

See Equation (G-70). Then we have The rank of S,. _1 is


m
(9-76),
rank Sm = (m + l)p + ¿ (q -1) for m < v-1
j=O
v- 1
and rankS m=(m +l)p + ¿ (q -I)=(m +l)p +n for m 2. v-1 A comparison of (9·
j= o co1umn rank if and (
(9-76)

where n ~ L;~~ (q - rj) is the degree of C(s). We call v the r~w index of C(s).
It is the 1argest rowdegree of A(s) in any leftcoprime fraction of G(s) = A -l(s)B(s)
with A(s) row reduced. It is also equal to the observability index of any irreduci­ This equality holds if
ble realization of C(s). See Equation (G-8l) and Theorem 6-6. O(s) and N(s) are rig
rank of Sk and the n
Theorem 9-17 from v - 1, we conc1'
only if D(s) and N(s).
Consider a q x p proper rational matrix with the fraction G(s) = N(S)O-I(S).
If Df(s) is of colt
Let /l¡, i = 1, 2, ... , p, be the column degrees of O(s), and let v be the row index zero columns in the
ofC(s). Ifm 2.v -1, then for any D¡(s)withcolumndegrees m +/l¡, i = 1, 2, .. . ,p, zero columns coincid
or less, there exist Dc(s) and N/s) of row degree m or less to meet
must be inside the ro'
o fes) = 0c(s)D(s) + Nc(s)N(s) (9-74), or equivalentl
completes the proof e
if and only if D(s) and N(s) are right coprime and O(s) is column reduced. 15
This theorem stal
Proof meet (9-70), but state~
To study this questic
Let /l =max {/l¡, i = 1,2, ... , p}. Since C(s) is proper, the column degrees of proper or proper. Th
N(s) are equal to or smaller than the corresponding column degrees of D(s). 9-11 and 9-11'). Ho
Consequently, the matrix coefficient matrix D/L (
Before proceeding
S _[0 0 01 °/L-1 O/L]
0- No NI N/L-I NI'
¿;
has at least a total of = 1 (/l - /l¡) zero columns. In the matrix SI' some new and
zero column will be created in the rightmost block column; however, sorne
zero colulnns in So 'NHl cHsappeó~r froro. § 1- I-Iencc the nurnb~r oi zere cJI~;,,_:..L_' lheorem 9-Hs
in SI remains to be
p p
Consider a q x p stric
C(s) = N(s)D- 1 (s). L
¿ (/l - /l¡) = P/l - ¿ /l¡
i= 1 i=l v be the row indeJ
In fact, this is the minimum number of zero columns in Si, i = 2,3, . . .. Let m¡ 2. v - 1 (m¡ 2. v) for :
8. _1 be the matrix S. _ 1 after deleting these zero columns. Since the number of
columns inS miS(/l + 1 +m)p, the number of columns in 8.- 1 is equal to
p p
(/l +v)p-(P/l- L /lJ=vp +¿ /li (9-77 ) exists and is nonsing­
(;;1 i=l meet

15 This theorem reduces to Theorem 9-10 for the single variable case. In the single-variable case.
we have J1 = v=degD(s) = 11, and the search of the row 'index becomesunnecessary. if and only if D(s) and
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 481

The rank of S"-1 is clearly equal to the rank of S"-I' Hence we have, from
(9-76),

rankS"_l =rankS"_I =vp +n (9-78)

for m?:. v - 1 A comparison of (9-78) with (9-77) reveals immediately that S,. _1 has a full
column rank if and only if
(9-76 )
p

the row index of G(s). I J1i=n


i= 1
on of G(s) = A - 1 (s)B(s)
y index of any irreduci­ This equality holds if and only if, following Corollary G-14 and Definition G-4,
:m6-6. D(s) and N(s) are right coprime and D(s) is column reduced. Since both the
rank of Sk and the number of columns in Sk increase by p as k increases by 1
from v -1, we conclude that S"" for m?:.v -1, has a full column rank if and
only if D(s) and N(s) are right coprime and D(s) is column reduced.
ion G(s) = N(s)D -1(S). Ir DAs) is of column degree m +J1i, then there are at least I~=1 (fl-flJ
let v be the l'ow index zel'o columns in the F matl'ix in (9-74). Furthel'more, the positions of these
esm +fl¡, i = 1,2, .. . ,p, zel'o columns coincide with those ofS",. Since S", has a full column l'ank, the F
to rneet must be inside the row space of Sm' Hence a set of solutions {Dei, N ei } exists in
(9-74), 01' equivalently, a set of solutions {De(s), Ne(s)} exists in (9-70). This
completes the proof of the theol'em. Q.E.O.
\ is column reduced. 15
This theorem states the condition fol' the existence of De(s) and Nis) to
meet (9-70), but states nothing l'egarding whethel' D; 1 (s)Ne(s) is Rl'oper or not.
To study this question, we consider separately the case where G(s) is strictly
:he column degrees of proper 01' proper. The situation is similar to the single-variable case (Theorems
llumn degrees of D(s). 9-11 and 9-11'). However, the proof must be modified because the leading
coefficient matrix DI' of D(s) is generally not nonsingular.
Before proceeding, we define
] H(s) = diag {Sl'l, Sil', , Sil,,) (9-79)
e matrix SI' sorne new and He (s) = diag {s'" I , s"'z, , s",p} (9-S0)
,Iumn; however, sorne
1mber ofzero columns Theorem 9-'¡ 8
Consider a q x p strictly proper (proper) rational matrix G(s) with the fl'action
G(s) = N(s)D- l (s). Let fli, i= 1,2, ... ,p, be the column degrees of D(s) and let
v be the row index of G(s). Let mi be the row degrees of D,(s). [f
mi ~ v - 1 (mi?:. v) for all i, then for an y D f(S) with the property that
in Si' i =2,3,.... Let
;. Since the number of lim He-1(s)D¡(s)H-l(S)=J (9~S1)
§,,_ 1 is equal to s ..... 00 .

exists and is nonsingular, there exists proper (strictly proper) De-:-1(s)Ne(s) to


(9-77 )
meet

(9-82)
In the single-variable case,
.les unnecessary. ir and only if D(s) and N(s) are right coprime and D(s) is column reduced.
482 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

Proof
proper) rational mal
feedback system in I
Let m = max {mi} and J1 = max {J1;}. Consider the ith row equation of (9-74):
are right coprime ano

This theorem fo1l


where D,cj denotes the ith row of D cj and so forth. Since mi;::: l' - 1, the resultant (9-70). Several remé
¿
Sl1li' excluding (J1- J1í) number of zero columns, has a full column rank. The the case where G(S)
assumption of (9-81) implies that the ith row of D [(s) has column degrees D [(s) = P(s)D(s), thel
at most mi + Ilj. Hence F í has ¿
(J1- Ilj) number of zero elements whose In other words, if th(
positions coincide with those oí S"", Hence we conclude that for any D [(s) compensator C(s) = ,
meeting (9-81), solutions Dc(s) and Nc(s) of row degrees at most mí exist in (9-82). remark concerns the
We write (G-57) as M(s) = [M lJc + M¡(s)JHc(s), where M¡(s)~ M1c(s)H c- 1 (s) D [(s) and between 1
is a strictly proper rational matrix and M¡( (0) = O. Similarly, we write Nc(s) and Dc(s) in q
D(s) = [D h + D/(s)JH(s)
tions involve the pol(
N(s) = [N h + N¡(s)JH(s)
these cancel1ations al
Dc(s) = He(s)[DclJ + DcI(s)J
and Nc(s) have a COI
(9-83) Hence the poIe-zero
Nc(s) = He(s)[NclJ + NcI(s)J
D [(s) = He(s)[D [h + D [1 (s)] H(s) assignable poles. Th
system. Since (;(s) is,
where D/(s), DcI(s), N 1(s), Ncl(s) and D [1(S) are strictly proper rational matrices. I. Hence the unity fe
The constant matrix D IJ is the column degree coefficient matrix of D(s) and is if Dc(s) and Nc(s) are \(:
nonsingular by assumption. The substitution of (9-83) into (9-82) yields, at which is much larger t
s=oo, We state the dual
Oel,OIJ + NeI,N" = D[IJ (9-84)
Ir G(s) is strictly proper, N,,=O and D ch = D[hD,;-1. By assumption, D[IJ=J Corollary 9-19
is nonsingular; hence Dc(s) is row reduced and 0c- l(s)N c(s) is proper. Ir G(s)
~onsider a q x p stric
is proper, we have N h i= O. However, the row of Sm corresponding to N ícll
G(s) =D-l(S)N(s). L<
are linearly dependent if m;::: v, Hence we may choose N íclJ = O. For this choice,
be thecolumn index 01
we have Dich=Dí[hDh-l and D ch =D[I,D,;-1. Hence Dc(s) is row reduced.
Because of Ni'h = O, Dc- 1 (s)N c(s) is strictly proper. Q.E. D. Ir mi ;:::¡l - 1 (mi ;:::Jl) fe

lim diag(s··\l,.
It is important to note that if mi;::: l' for some i, the solutions of (9-74) 01' s .... 00

(9-82) are not unique and there may be improper solutions. However, there
exists) unde~· f.b.e 2.ssunl.ption of {9-81)., at least ODe 3,,:1: ')f ~~:~.. :::y.>~: :)'.. :.;t:~s'!:I~/ exists !:-tnd lS llf)7lsi '1';­ 1

proper solutions. Ir O[(s) meets (9-81), it is said to be row-column-reduced in (strictly proper) ratio~
Reference S34. Every O [(s) can be transformed, by elementary operations, r to y in Figure 9-1 7 i,
into a row-column-reduced one. are left coprime and [
Now we may apply Theorem 9-18 to the design problem.

Theorem 9-19 e
1-----­
Consider a q x p strictly proper (proper) rational matrix (;(s) with the fraction
(;(s) N(s)D-l(S). Letllí, i= 1,2,.'.. ,p be thecolumn deg~ees of O(s) and let v
0;= _ _o -'/~'-J-_--l-;-/~ Dc 1 ~
be lbe row index of G(s). Let the row degrees of Dc(s} be mi, i;=; 1,2, ... , p. Ir L _
mi;::: l' ~.l (mi;::: v) for al1 i, thenfor any O [(s) with the property thilt

lim Ht 1 (s)D[(s)H- 1(s)=J


s- 00
Figure 9-17 Unity feec
exists and is nonsingular, thereexists a compensator with a p x q proper (strictly
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 483

proper) rational matrix D: l(s)N c(s) such that the transfer matrix of the unity
feedback system in Figure 9-16 is N(S)O¡l(S)N c (s) ifand only if O(s) and N(s)
>w equation of (9-74): are right coprime and O(s) is column reduced. il

This theorem follows directly from Theorem 9-18 and Equations (9-69) and
li 2: v - 1, the resultant (9-70). Several remarks are in order regarding this theorem. We discuss only
.Jll column rank. The the case where (;(s) is strictly proper. Ir mi = v - 1, for all i, and if we choose
, has column degrees O [(s) = P(s)O(s), then the unique solution of (9-70) is Oc(s) = pes) and Nc(s) = O.
zero e1ements whose In other words, if the denominator matrix of the plant is not to be altered, the
de that for any Dr(s) compensator C(s)= O,:"l(s)N c(s) is O. This is a degenerated case. The second
most mi exist in (9-82). remark concerns the possible existence of common divisors between N(s) and
: MI(s)é M¡Js)H: 1 (s) O[(s) and between D[(s) and Nc(s) in (;r(s) = N(s)D¡ l(s)N c(s); and between
¡lady, we write Nc(s) and Dc(s) in C(s) = Oc- 1 (s)N c(s). In the first case, the pole-zero cancella­
tions involve the poles which the designer has the freedom in placing, therefore
these cancellations are permitted in practice. From (9-70), we see that if O,.{s)
and Nc(s) have a common left divisor, then it is also a left divisor of D [(s).
(9-83) Hence the pole-zero cancel1ations between Dc(s) and Nc(s) involve again only
assignable poi es. The final remark concerns the well posedness of the feedback
system. Since (;(s) is strictly proper and C(s) is proper, we have 1 + (;( 00)C( ro)=
)per rational matrices.
I. Hence the unity feedback system in Theorem 9-18 is well posed. Note that
matrix of D(s) and is
ifDe(s) and Ne(s) are left coprime, thedegreeofthecompensator is ¿mi 2:p(v- 1),
) into (9-82) yields, at
which is much larger than the one required for arbitrary pole assignment.
We state the dual of Theorem 9-19 as a corollary.
(9-84 )

( assumption, D [11 = J Corollary 9-19


[e(S) is proper. Ir (;(s) Consider a q x p strictly proper (proper) rational matrix G(s) with the fraction
~orresponding to N icl1
C(s) = O -l(s)N(s). Let Vi, i = 1, 2, ... , q, be the row degrees of D(s) and let J1
el¡ = O. For this choice, be the column index of C(s). Let thecolumn degrees of De(s) be mi, i = 1,2, ... , q.
Oe(s) is row reduced. If mi 2: 1I - 1 (mi :?-/I) for aH i, then for any D [(s) with the property that
Q.E.D.
liro diag {s - \\ s- \\ ... , s - "q} D [(s) diag [s· m" S -m>, ... , S-m(/} = J
solutions of (9-74) 01' S""" 00

ions. However, there


exists ano is nons~nguh1f. th~~:,:: :~x.js~~ '~\ r::r)n:;~:-=:":SF:t·~;~- ~//itb :.0. ';.' ',¡' -~_
¿"" •• ..: ,

of proper or strictly
(strictly proper) rational matrix Nc(s)D c- 1 (s) such that the transfer matrix from
)w-column-reduced in
r to y in Figure 9-17 is equal to 1- Dc(s)D¡ I (s)O(s) if and only if O(s) and N(s)
lementary operations,
are left coprime and D(s) is row reduced.
,Iem.

(;(.1') with the fraction


egrees of D(s) and let v
,e mi, i= 1,2, ... ,p. If
)perty that
'-'---------------------~
----~---_._-----------'

Figure .9-17Unity feedback system.


a p x q proper (strictly
484 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

The substitution of G(S) = D-l(S)N(s) and C(S) = N c(s)D c- l (s) into the first
equality of (9-68) yields

Gf(s) = [1 + D- l (s)N(s)N c(s)D c- l (s)] -1 D- l (s)N(s)N c(s)D; l(S)


= Dc(s)[D(s)Dls) + N(s)Nc(s)] -1 N(s)N c(s)D c- l (s) (9-85)

Define

(9-86)

Then G¡(s) becomes


G¡(s) = Dc(s)D¡ l(s)[D¡(s) - D(s)Dc(s)]D; l(S) which yields
= 1- Dc(s)D¡ 1 (s)D(s)

This shows that the transfer matrix from r to y in Figure 9-17 is equal to
[1 - Dc(s)D¡ 1 (s)D(s)]. The desígn ín Corollary 9-19 hínges on solvíng (9-86).
Note that the transpose of (9-86) becomes (9-70); left copríme and row reduced
become ríght copríme and column reduced. Hence Theorem 9-18 can be
applied directly to the transpose of (9-86). Of course Equation (9-86) can also We see that Dc(s) is :
be solved directly. We use the coefficient matrices of D(s) and N(s) to form Hence if G(s) ís prc
T k as shown in Equation (9-62) and then search linearly independent columns compensator.
in order from left to right. The least integer Jl such that all N columns in the If we choose mi =
last block column of T¡.¡ ·are linearly dependent, is called the column index of
(;(s). Dual to the row index, the column index is equal to the largest column D
degree of the column reduced Dr(s) in any right coprime fraction of G(s)=
N r (s)D,:-l(S). It is also equal to the controllability index of any irreducible
then the compensato
realization of (;(s). The proof ofCorollary 9-19 is similar to the one ofTheorem
9-19 and will not be repeated.

Example 3
We gíve an example to iJlustrate the appíicatíon oí Theorem 9-i9. o....:onsiuer
the proper rational matrix Remarks are in 01
and the design of ar
~ = N(s)D- l (s) = [S2 O+ 1
G(s)
S2
S J[s2-1
+s + 1 O S2
O
-1
J-l strictly proper, the
min (Jl- 1, v - 1) (Th,
p(v -1) (Theorem 9-1
We form So, Si' ... , and search their linearly dependeni rows in order from top ment assigns only the
to bottom. For this example, we can readily obtain that v = 2. Clearly we always yields a cycL
have Jll = Jl2 = 2. Let /111 =/112 = V - 1 = 1. We choose mairix yields general
·of thecyclic and non
applicable here, the
·placement design are
denominatormatrix (
Then the compensator is the solution of of these two designs i
DESIGN OF COMPENSATORS: UNITY FEEDBACK SYSTEMS 485

s)D c- 1 (s) into the first -1 O O O 1 O, O O


O -1 O O O 1, O O
100 110 O O
;)Nc(s)D; I(S) O 1010 1,0 O
,; I(S) (9-85) NclJ - -6 ---0- : ~ í --- 0- ---6 ---0- -- -( ---ó
O O O -1 O O O 1
O O, 1 O O 1 1 O
(9-86)
O O O 1 O 1 O tJ
= [~ ~ ~ ~ ~ ~ ~ ~J
which yields

gure 9-17 is equal to


1ges on solving (9-86).
_2s;21

rime and row reduced


rheorem 9-18 can be
s+d

uation (9-86) can also We see that Dc(s) is singular and the compensator Dc-l(s)Nc(s) is not defined.
[)(s) and N(s) to form Hence if G(s) is proper, the choice of mi = v - 1 may not yield the required
independent columns eompensator.
: all N eolumns in the Ir we choose mi = v and choose
1 the colúmn index of
to the largest column
me fraction of (;(s) =
lex of any irreducible then the compensator can be computed as
to the one ofTheorem

DJs)=l O
rs2 +3 4~s - 1)/3J NJs) =[
4S +4
O
-4(s + 1)/3J
s- +s + 1 2(s + 1)

and D c- l(s)Nc(s) is strictly proper. The degree of Dc(s) is equal to 4. lIi

orem 9-19. Consider


Remarks are in order regarding the design of arbitrary assignment of poles

S2
O
-1
J-l and the design of arbitrary assignment of denominator matrices. Ir (;(s) is
strictly proper, the minimal degree of compensator for the former is
min (J1-1, v -1) (Theorem 9-16); whereas the minimal degree for the latter is
p(v -1) (Theorem 9-18) or q(J1-1) (Theorem 9-19). The design ofpole assign­
ows ín order from top ment assigns only the poles or the determinant of the denominator matrix and
lat v= 2. Clearly we always yields a eyc!icoverall transfer matrix. The design of denominator
matrix yields generaHy· a noncyclie overall transfer matrix. if the diseussion
of the cyclic and noncyc!ic designs in the multivariable case of Section 7-3 is

[O30J [1O 0J1 s applicable here, the transient responses of the, system obtained by the pole­
t - ' s 2' + 3
ptacement design areprobablyworse than those of the system obtained by the
2
denominator matrix design with the same set of poles. A detailed comparison
of these two designs is not available at present.
486 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

any unstable pole-z(


Decoupling. The results in Theorem 9-19 may be used to design a unity
i = 1,2, ... , p, and th
feedback system so that its transfer matrix GJ(s) is diagonal and nonsingular.
Such a sxstem is said to be decoupled (see Definition 7-1). If the plant transfer consequently the nur
matrix G(s) = N(s)D-1(S) is square and nonsingular, we may choose DJ(s) = is proper. Once the
DJ(s)N(s), where fiJ(s) is diagonal, in Theorem 9-19. Then the overall transfer In this design, it is a
matrix in Figure 9-16 is both (;(¡(s) and (Ct.¡(s) -!
Ir det N(s) is not
GJ(s) = N(s)D i l(s)N c(s) = N(s)(D¡(s)N(sW 1 Nc(S) = fi j 1 (s)Nc(S) pole-zera cancellatio
N- 1(5). [Recal! that
Now if the degree of compensator is sufficiently large. we may be able to choose
a Nc(s) which is diagonal. In this case, G¡(s) becomes diagonal and the system poles in 0-1(5).J To
is decoupled. This design is achieved by canceling N(s) and may involve
undesirable pole-zero cancellations. Hence this decoupling method is not
always satisfactory. has no open right-ha
In the following, we discuss a different method of achieving decoupling. 16 to be the least comml
The method involves direct cancellations and is applicable only to stable plants. N- 1 (s). Once f3¡(s) i~
Con~ider a p x p proper rational matrix G(s) = N(s)D -l(S). 1t is assumed that (9-87) is proper. Nc
G(s) is nonsingular and det D(s) is a Hurwitz polynomial. Let P(s) in Figure 9-18 proper. Let G-1(S)
be a diagonal rational matrix of the form that if
p
P( s) =d'lag {f31(S) , f32(S) , ... , f3 (S)} deg iX¡(S
(;(l(S) (;(2(S) (;(p(s)
We see that the transfer matrix of the tandem connection of the compensator fol' i = 1,2, ... ,p, thel
followed by the plant is G(s)C(s) = N(s)D- \s)D(s)N -l(S)p(S) = P(s). Because ith column of P(s) is 1
G(s)C(s) = P(s) is diagonal, the feedback system in Figure 9-18 reducesto a 'set ithcolumn of G- \s),
of p number of decoupled single-loop systems. Hence the system is decoupled. is also equal to the p
Now the matrix P(s) is to be chosen so that the compensator G-l(S) is equal to the
zero excess inequa1ity
C(s) = D(s)N -1(S)P(S) = G-1(S)P(S) (9-87)
The poles of GJ(s)
is a proper rational matrix and the overall transfer matrix (;(¡(s) are determined fr
GJ(s), we can readily ,
G ()-d" {131(5) 132(S) ~)_}
(9-87) is proper and th,
Js - lag (;(1(5) +f3l(5)' (;(2(5) +/32(5)"'" iXp(S) +/3p(s)

The poles of GJ(5) are


is BIBO stable. !f det N(s) is H urwitz, the design in Figure 9-18 wil! n.ot j,_y:-:1'i~ r;.I~~e:.t2d by 'Ll:.t ,:;1U;1r:;c:'
phase zeros of G(s). ~
cancel1ations. In othe
16 Follows Reference S76. canceled and should b
Ir a plant is not st
decoupling procedure.
Compcnsator
,----------------l lations, although aH s
\ I
C(5), the property of d(
rT------'-'""' I 1--,---1----::-"<1

1 "'1
----." '-/--~l I ---'-.,r¡

1--'---' compensator is usuall~


I
L -l
I
is very sensitive to par

l7 Lel gis) = N(s)( D(s). We


the zero-pole excess of g(i
Figure 9-18 Decoupling of a plant.
DESIGN 01' COMPENSATORS: UNITY FEEDBA.CK SYSTEMS 487

any unstable pole-zera cancellations. In this case, we may choose f3¡(s) = 1,


ed to design a unity
i = 1, 2, ... ,p, and the poles of Gf(S) becomes o::¡(s) + L The degree of o:¡(s) and
,nal and nonsingular.
consequently the number of poles of GAs) are to be chosen so that C(s) in (9-87)
If the plant transfer
is praper. Once the poles of G¡(s) are specified, (;(¡(s) can be readily computed.
may choose D¡(s) =
In this design, it is also possible to design a stable compensator by requiring
n the overall transfer
both (;(¡(s) and (o:¡(s) + 1) to be Hurwitz.
If det N(s) is not Hurwitz, then the preceding design will involve unstable
= Dj 1 (s)Nc(s) pole-zero cancellations. These cancel1ations arise from the unstable poles 01
N- 1(s). [Recall that (;(s) is assumed to be stab!e, hence there are no unstable
nay be able to choose
poles in D- 1 (s).] To avoid these, we choose f3¡(s) so that the rational matrix
gonal and the system
(s) and may involve (9-88)
pling method is not
has no open right-half s plane poles. This is accomplished by choosing f3¡(s)
1ieving decoupling. 16 to be the least common denominator af the unstable poles of the ith column of
only to stable plants. N- 1(s). Once P¡{s) is ehosen, the degree of (;(¡(s) is to be chosen so that C(s) in
). It is assumed that (9-87) is praper. Note that although (;(s) is proper, G-1(S) is generally not
~et P(s) in Figure 9-18 proper. Let G- I(S) ~ (ñ¡}s)/dij(s)). Then fram (9-87), we can readily verify
that if

} deg (;(J~) - deg f3¡(s) ~ max [deg i1¡}S) - deg dij(s)]


}
(9-89)

1of the compensator for i = 1, 2, ... ,p, then C(s) is proper. That is, if the pole-zera excess 17 of the
P(s) = P(s). Because ith column al P(s) is equal to 01' larger. than the largest zero-pole excess of the
9-18 reduces to a set ¡th colum~ of (; -1(S), the~ C(s)' is proper. Note that the left-hand side of (9-89)
system is decoupled. ~ also equal to the pole-zera excess of G¡(s). Ir p = 1, the zero-pole excess 01
nsator G -1(S) is equal to the pole-zera excess of G(s), and (9-89) reduces to the poIe­
zera excess inequality discussed in (9-38) for the single-variable case.
(9-87)
The poles of G¡(s) are the zeras of (;(¡(s) + P¡(s). Once f3¡(s) and the degrees of
(;(¡(s) are determined fram (9-88) and (9-89), fram the assignment of the poles of
G¡(s), we can readily compute o::¡(s). Using these cx¡(s), thc compensator C(s) in
(9-87) is prape! and the unity feedback system G¡(s) in Figure 9-18 is de~oupled.
The poles of G f(S) are assignable by the designer. The zeros, p¡(s), of G f(s) are
9- J 8 win not involvE: dictatecl by the closecl right-half p!ane mots of del I''i(.», cailed [he nomninimum­
phase zeros of (;(5). They are chosen as in (9-88) to avoid unstable pole-zera
cancellations. In other words, nonminimum phase zeras of (;(5) should not be
canceled and should be retained in Gf(5).
If a plant is not stable, it must be stabilized before the application of the
decoupling pracedure. We see that the decoupling is achieved by exact cancel­
lations, although all stable ones. If there are any perturbations in (;(s) and
C(s), the property af decoupling will be destroyecl. Furthermore, the degree of
compensator is usually very large for an exact decoupling. Hence decoupling
is very sensitive to parameter variations and is expensive to implement.

17 Let gIs) = N(s)/D(s),


We cal! dég D(s) - deg N(s) ¡he po\e-zero excess and deg N(s) - dég D(s)
¡he zero-pole excess of gIs).
488 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

9-6 Asymptotic Tracking and Disturbance Rejection Befare proceedin~


If r(t) and w(t) both g<
Single-variable case. Consider the design of the control system shown in the feedback system
Figure 9-19(a). The plant with transfer function g(s) is given, the problem is to or w(t) does not go to
find a compensator with a proper transfer function C(s) so that the feedback nature, then it is not
system is asymptotically stable and meets some other specifications. One of rejection. Hence we J
the important specifications is to require the output of the plant y(l) to track the designo We assun
the reference signal r(t). Because of physical limitations, it is not possible to
design a feedback system so that y(t) = r(t) for all t (see Reference S46). The best
we can achieve is that
lim e(t) = lim [r(t) - y(t)] = O
' .... 00 l-co and
This is called the asymptotic tracking. It is well known that if r(l) is a step
function, or res) = 2[r(t)] = l/s and if C(s)g(s) is a type 1 system, that is, C(s)g(s) where the polynomiai
has one pole at the origin, and if the feedback system in Figure 9-19(a) is and Nw(s) are howeve
asymptotically stable, then the output y(t) wi1l track r(t) asymptotically. In equivalent to the assu:
this section, this statement and its generalization will be established.
Consider the feedback system shown in Figure 9-19(b). The plant is written
as g(s) = D-1(s)N(s). At the input of D-1(s), a signal w(t) is injected into the
plant. The signal W(l) will be called the disturbance signa!. Now in addition and
to asymptotic tracking, we also require that the affect of w(t) at the output
a pproaches zero as t -> 00 ; that is, lim y w(t) -> Oas t -> 00, where y w{t) is the out put
with some unknown ir
of the feedback system in Figure 9-19(b) due to the application of w(t) and
A,. and A", are D,(s) al
r(t) == O. This is called the disturbance rejection. Hence, if we succeed In finding
t-> 00 have no effect e
a compensator C(s) = Nc(s)/Dc(s) in Figure 9-19(b) so that for any r(t) and any
and D",(s) have zero o
w(t), we have
nominator of the unsl
lim e(t) = lim [r(t) - y(t)] = O (9-90) zero or positive real p;
{-co l-CO

then the feedback system achieves asymptotic tracking and disturbance rejec­
Theorem 9-20
tion. In this section we shall study the design of this problem.
Consider the feedbac
-:;omplete!y characteri:
slgnai r(t) ano disturí
w(s) = N w(s)/Dw(s). Le
poles of res) and w(s).
sator with a proper 1
(a)
asymptotically stable
rejection.

Proof
[f no root óf cP,(s) is a
transfer function.l/ cjJ(s
. able (Theorem 9~2).
(b)
coprime, and there exi
Figure 9-19 Design or control systems. back system shown in
ASYMPTOTlC TRACKING AND DISTURBANCE REJECTION 489

Rejection Befóre proceeding, we discuss first the nature of the signals r(t) and w(t).
Ir r(t) and w(t) both go to zera as t -+ 00, then (9-90) will be automatically rnet if
ptrol systern shown in the feedback system in Figure 9-19(b) is asymptotical1y stable. If either r(t)
ven, the prablem is to or w(t) does not go to zera, and if we have no knowledge whatsoever about its
¡'so thatthe fuedback nature, then it is not possible to achieve asyrnptotic tracking and disturbance
oecifications. One of rejection. Hence we need sorne information of r(t) and w(t) before carrying out
I:he plant y(t) to track the designo We aSSume that the Laplace transforrns of r(t) and w(t) are given by
1;, it is not possible to
lerence 546). The best A) =.;Lro[r(t )J = -
res N,.(S)
- (9-91 )
D,.(s)

and
A
w(s) = 2 [w(t)
J = -(-)
N w(s) (9-92)
Dws
11 that if r(t) is a step
Iystem, that is, C(s)g(s) where the polynomials D,(s) and Dw(s) are known and the polynomials N,.(s)
1I in Figure 9-19(a) is and N,Js) are however arbitrary so long as res) and w(s) are proper. This is
1\:) asyrnptotical1y. In equivalent to the assumption that r(t) and w(t) are generated by
established.
x,.=A,X,. (9-93a)

. The plant is written


r(t) = C,.X, (9-93b)

t) is injected into the


and xw(t) = Awx w (9-94a)

al. Now in addition


w(t) =CwX w (9-94b)

lof w(t) at the output


11ere yw(t) is the output with some unknown initial states x,(O) and xw(O). The minimal polynomials of
'Plication of w(t) and A, and A w are D,(s) and Dw(s). The parts of t(t) and w(t) which go to zera as
we succeed in finding t --+ 00 have no effect on y as t -+ 00, hence we assume that sorne roots of D,(s)
l
t for any r(t) and any and Dw(s) have zera or positive real parts. Let cjJ(s) be the least common de­
nominator of the unstable poIes of res) and w(s). Then aH roots of cjJ(s) have
(9-90) zera or positive real parts.

¡nd disturbance rejec­ Theorem 9-20


blern.
Consider the feedback system shown in Figure 9-19(b), where the plant is
complete1y characterized by its proper transfer function g(s). The referenr;<>:
signa! r(l) and disturbance signal w(t) are modeled as res) = N,.(s)fD,.(s) and
w(s) = N w(s)/Dw(s). Let cjJ(s) be the least common denominator of the unstable
poles of res) and w(s). If no root of cjJ(s) is a zera of g(s), there exists a compen­
sator with a proper transfer function so that the unity feedback systern is
asymptotical1y stable and achieves asymptotic tracking and disturbance
rejection.

y(t)
Proof

J
[f no rootof cjJ(s) is a zero of g(s), the tandem connection of the systern with
transfer. function llc/J(s) followed by 9(S) = N(s)/Q(s) iscontroIlable artd observ­
able (Theore"m 9-2). Consequently, the polynomials N(Is) and D(s)cjJ(s) are
coprinie, and there exists a compensator C(s) = Nc(s)/Dc(s) such that the feeo­
back system shown in Figure 9-20 is asymptotically stable (Theorems 9-11 and
490 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

9-11') or, equivalently, all roots of for sorne reference


tracking without int
D I(s) ~ Dc(s)D(s)4>(s) + Ncls)N(s) of the internal mode
have negative real parts. variations of 9(s) = JI
Now we clairn that the feedback system in Figure 9-20 with C(s) = robusto This will be
N,ls)/Dc(s)cjJ(s) will achieve asyrnptotic tracKmg and disturbance rejection. 1. It is weH known
Indeed, the output y(t) excited by w(t) and r(t) == O is equal to have a zero stea(

)~w(s) = - ew(s) = 1 + Nc(S)~~;~!(S)4>(S)D(S) íl{S)


input), the plant
has one poJe (tw
function), then w{
Dc(s)4>(s) N ",(s)
transfer function
Dcls)4>(s)D(s) + Ncls)N(s) D",(s) designo
Dcls)Nw(s) 4> (s) 2. From (9-95) and
(9-95)
D c(s)D(s)4>(s) + N c(s)N(s) D",(s) rejection are achi
reference and dü
Since aH unstable roots of D,/s) are canceled by 4>(s), all the poles of Yw(s) have those modes 1/4>(.
negative real parts. Hence we have yw(t) = -ew(t)->O as t->oo. We see that transfer functiom
even though w(t) does not go to zero as t ........ 00, its effect on y(t) diminishes as rnent of the inter
(->OO.
the undesirable rr
Let y,(t) be the output excited exclusively by r(t). Then we have 3. The location of th
r(s) - y (s) =
,.
(1 _ Nc(s)N(s)/Dc(s)D(s)c/J(s) ) r(s)
1 +Nc(s)N(s)/D c(s)D(s)4>(s)
case so long as 1/
from w to y. lf il
Dc(s)D(s)c/J(s) N ,(s) will not appear in
4. Because of aging
D c(s)D(s)4>(s) + Nc(s)N(s) D,.(s)
9(s) or, equivalent
Dcls)D(s)N,.(s) 4>(s) This is caHed the
(9-96)
D c(s)D(s)4>(s) + Nc(s)N(s) D..(s) perturbation, eve
Again, aH the poles ofr(s) - y,.(s) have negative real parts; hence we have permitted so long
r(t)-y,.(t)->O as t->oo. Because of linearity, we have y(t)=y)t) +y,.(t) and
e(t) = r(t) - y(t)->O as t ........ oo. This establishes the theorem. Q.E.D.
rernain to have n
is not permitted.]
The design procedure developed in this proof consists of tvvo steps: ¡ntrc··
duction of 1/4>(s), a rnodei of the reference and disturbance signals, inside ihe
loop, and the stabilization of the feedback system by introducing the compen­ The robustnes
sator Nc(s)/Dc(s). The duplication of the dynamic or rnodel, 114>(s), inside the case where r(s) = 1
with transfer fune
loop is often referred as the internal model principie. As will be discussed later,
A
g

and consider the 1

Figure 9-20 Asymptotic tracking and disturbanee rejection. Figure 9-21 Asympll
ASYMPTOTIC TRACKING AND DISTURBANCE REJECTION 491

for sorne reference input and w(t) = 0, it is possible to achieve asymptotic


tracking without introducing an internal model. However, the employment
of the internal model will make the feedback system insensitive to parameter
variations of g(s) = N(s)/D(s) and N c(s)/Dc(s). Hence this design is said to be
re 9-20 with C(s) = robusto This will be expanded in the following remarks:
Iisturbance rejection. 1. It is well known in the design of unity feedback systems that, in order to
I to have a zero steady-state error for a step reference input (a ramp reference
input), the plant must be of type 1 (type 2) transfer function, that is, g(s)
has one pole (two poles) at the origino Ir r(t) is a step function (a ramp
function), then we have cfJ(s) = s [cfJ(s) = S2] and (lN(s)g(s) is oftype I (type 2)
transfer function. Hence the well-known result is a special case of this
designo
2. From (9-95) and (9-96), we see that asymptotic tracking and disturbance
(9-95) rejection are achieved by exact cancellations of the unstable modes of the
reference and disturbance signals. This is accomplished by duplicating
:he poles of yw(s) have those modes 1N(s) inside the loop so that cfJ(s) appears as numerators of the
s t-+ oo. We see that transfer functions from r to e and from w to y. In other words, the employ­
on y(t) diminishes as ment of the internal model is to create the required numerator to cancel
the undesirable modes.
hen we have 3. The location ofthe internal modeI11cP(s) is not critical in the single-variable
case so long as l/cfJ(s) does not appear in the forward paths from r to e and
-) r(s) from w to y. If it appears in the forward path from, say, w to y, then cfJ(s)
,)
will not appear in tlle numerator of the transfer function from w to y.
4. Because of aging or the variation of the load, the plant transfer function
g(s) or, equivalently, the coefficients of N(s) and D(s) may change with time.
This is called the parameter perturbation. In this design, the parameter
(9-96)
perturbation, even large perturbation, of N(s), D(s), NJs), and Dc(s) are
permitted so long as all roots of
hence we have
y(t) = yw(t) + y,.(t) and Dc(s)D(s)<jJ(s) + NJs)N(s)
n. Q.E.D.
remain to have negative real parts. [Note that the perturbation of <jJ(s)
is not permitted.] Hence this design is insensitive to péH"ameter perturb2c­
:s of two steps: intro­
lion ana is said io be rooust.
lce signals, inside the
The robustness is due to the presence ofthe internal model. We use the
·oducing the compen­
case where r(s) = l/s and w(s) = O to ilIustrate this point. Consider a plant
ldel, lN(s), inside the
with transfer function
'1i1l be discussed later,
A Nnsn+NII_¡sn-¡ + +N¡s+N o
g(s) = Dns" +Dn_¡sn 1 + +Dls +D o
v(t) and consider the feedback system shown in Figure 9-21, where a constant

y(t)
y{t)

Figure 9~21 Asymptotic tracking with and without an internal model.


492 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

6. Unlike De(s), N e(s


gain k is placed at the input. Let not permitted in
N sm + . are achieved by ,
C(s) = Dems m + . by the roots of
em
and expensive te
Then the transfer function, 9¡(s), from r to y is given by study in the follo'
~ kg(s)C(s) k(NIINems"+1lI + + NoN eo ) assume
g¡(s) = 1 +g(s)C(s) = (DIID em + NIINem)s"+ 111 + +(DoDeo+NoNeo)
lt is assumed that g¡(s) IS asymptotical1y stable. Ir r(s) = l/s, the application
of the final-value theorem yields and

lim (r(t)- y(t)) = lim s(r(s) - 9¡(s)r(s)) = 1 - 9¡(0) where a ~ O and


t-+ ce .,,-+0 perturbation or
from r to e in Fi¡
= 1- kNoN eo (9-97)
DoD eo + N oN eo
If C(s) contains the internal model l/s and if s is not a zero of g(s), then we where all poles e
have Deo =0, Neo f.O and N O f. O. In this case, if k = 1, then r(1) - y(t)->O r(s) and w(s) =0 i:
as t -> oo. As long as Deo = O, Neo =1= Oand 9¡(s) remains to be asymptotically
stable, we always have r(t) - y(t)-> O as t-> 00 for all parameter perturba­ e(s) = (
tions of g(s) and C(s). Hence the design is robust.
Now we assume that g(s)C(s) does not contain the internal mode! l/s.
Then we have Do f.O and Deo f.O. In this case, if we choose k in (9-97) as s
(
with k¡ =­

We see that k 1 i:
then we have r(t) - y(t) -> O as t -> oo. In order to have a finite k, we need exact, then k 1 is 2
No =1= Oand Neo =1= O. In the design of compensators, it is always possib1e to Ir k 1 is nonz~
find a C(s) with Neo =1= O. Ir Do =/=-0, we have No =/=- O if and only if <fJ(s) = s k ¡ea'. If a is po>
is not a zero of g(s). Hence we condude that, under the condition stated in output y(t) will r
Theorem 9-20, ifr(s)= l/s, il is also possible to design a feedback system, track r(t) but wit
without introducing the internal model, to achieve asymptotic tracking. with m > 1, and i
This design, however, is Dot robust. Ir there are perturbations in ?ny ;." again not possibl.
No, Do, Neo, and Deo, we do nol have r(t)-y(t)->O as t->oo. Hence the This, howevet
design which does not employ the internal model principIe is not robust. interest. In the d
We emphasize that the condition that no root of </J(s) is a zero of 9(S) is result will be evet
needed in both robust and nonrobust designs. this case, if the inl
5. From the discussion in item 4, we may condude that the condition stated in errors in impleml
Theorem 9-20 is necessary as well. If we introduce the interm!.l model and but with a finite S1
if anúoot of <fJ(s) is a zero of g(s), then the root, which is unstable, becomes a of <fJ(s), the smalh:
hidden mode and will not be affected by any compensation. Hence the 7. The design of N k
unity feedback system can never 'be asymptotically stable. Ir nointernal of the polynomia
model is employed, even though g(s) is of the form </J(s)Ñ(s)/D(s), it is straight-­
forwúd to show from (9-95) and (9-96) ttiat <fJ(s) will not appear as anumer~
atar oC the transfer functions f~om w to' y and r to e; hence asymptotic Under theconditi
tracking and disturbance rejection cannot be achieved.
ASYMPTOTIC TRACKING AND D1STURBANCE REJECTION 493

6. Unlike Db), Nc(s), D(s), and N(s), the variation of the coefficients of <f;(s) is
not permitted in robust design, because tracking and disturbance rejection
are achieved by exact cancellation of the unstable modes of r(t) and w(t)
by the roots of <f;(s). In practice, exact cancellations are very difficult
and expensive to achieve; hence inexact cancellations often occur. We
study in the following the effect of inexact cancellation. For simplicity, we
assume
-(DoD co +NoNco ) ~ 1 ~
r(s) = - - ro(s)
= l/s, the application s-a
and <f;(s) = (s - a +e)<f;o(s)
where a:2:0 and e is a small real number which denotes the amollnt of
perturbation or inexact implementation of <f;(s). The transfer function
from r to e in Figure 9-20 then has the form
(9-97)
g,.(s) =(s -a +e)g,.o(s)
l zero of g(s), then we where all poles of g,.o(s) have negative real parts. The signal e(s) due to
: 1, then r(t) -y(t)-->O r(s) and w(s) =0 is equal to
¡ to be asymptotically
~ ~ 1 ~
parameter perturba­ e(s) = (s - a +e)g,·o(s) - - ro(s)
s-a
e internal model l/s. ~
k¡ +terms due to t he poles of g,.o(s) ~ )
choose k in (9-97) as =-- and ro(s
s-a

with k ¡
(s - a +e)
s-a
A ~ I A

g,·o(s)ro(s)(s-a) s=a =eg,.o(a)ro(a)


A

We see that k¡ is proportional to the deviation e. lf the cancellalion is


'e a finite k, we need
exact, then k ¡ is zero; otherwise, k¡ is nonzero.
is aIways possible to
If k¡ is nonzero, even though very small, then e(t) contains the term
f and only if <f;(s) = s
k ¡ ea'. If a is positive, this term will approach infinity as t --> 00, and the
e condition stated in
output y(t) will not track r[t) asymptotically. If a is zero, then y(t) will
n a feedback system,
track r(t) but with a finite deviation. If r(s) has a repeated pole (s - at,
lsym ptotic tracking.
with m> 1, and if cancellations are not exact, then asymptotic tracking í;:;
·turbations in any of
again nol possibie.
s t----> oo. Hence the
This, however, does not mean that Theorem 9-20 is only of theoretical
nciple is not robust.
interest. In the design, if the internal model principIe is not empIoyed, the
;) is a zero of g(s) is
resuIt will be even worse. In practice, most r(t) and w(t) are bounded. In
this case, if the internal model principIe is employed, even though there are
e condition stated in
errors in implementing <f;(t), the output will still track the reference signal
: internal model and
but with a finite steady-staú: error. The more aceurate the implementation
unstable, becomes a
of <f;(s), the smaller the errO"r.
nsation. Hence the
7. The design of N c(s)/Dc(s) tostabilize the feedback system requires the solving
able. If no internal
of the polynomial equation. .
(s)/D(s), it is straight­
t appear as a numer­ D¡(s)=Dc(s)D(s)<f;{s) + Nc(s)N(s)
?; hence asymptotic
Under the condition ofTheorem 9-20, D(s)<f;(s) and N(s)are coprime. If we
494 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

use Theorem 9-10, the degree of DJ~) is deg D(s) +deg cjJ(s) -1 and the total differential eguat
degree of the compensator N /s)/Dc(s)cjJ(s) is deg D(s) + 2 deg cjJ(s) - 1. A discussion of this
different procedure is to solve the eguation
'Multivariable ca~
DJ(s) = Dc(s)D(s) + N c(s)N(s) system shown in Figl
with the constraints that Dc(s) contains the factor cjJ(s) and Nc(s)/Dc(s) is matrix G(s) factored f
proper. By so doing, the degree of compensator can be considerably q xq and q xp poly
reduced. See References S34 and S238. signal r(t) and the q )<
8. We give a state-variable interpretation of the disturbance w(t) in the p!ant
shown in Figure 9-20. The output _v(.~) is egual to and
y(s) = D - 1 (s)~¡;(s) + D - I (s)N(s)u(s) where D,.(s) and D w (!
1
Let D(S)=S"+DI_1S"- + ... +D o and N(s)=NI_IS"- 1 +N I _2 S"- 2 + q x 1 polynomial ma
... + No· Then YlI(S) = N(s)D- 1 (s)u(s) can be realized as any N,.(s) and any N,

x= Ax + bllu(t) YlI = ex (9-98)

with A and e of the forms shown in (6-38). Similarly, y",(s) = D- I(S)W(S) This is the problem e
can be realized as, by using (6-8), Following the sil
x= Ax + bww(t) y=ex (9-99) introduction of an il
There is, however, on
where A and e are identiGal to the A, e in (9-98). Hence we may combine is critical in the m u lt
(9-98) and (9-99) to yield w(t) by assuming r(t)'
shown in Figure 9-23
x= Ax + bllu(l) + bww(t) y=ex
ew(s)
From the eguation, we see that the disturbance in Figure 9-19 can be
considered as imposed on the state of the system. This eguation can be or en.(s)
generalized to

x= Ax + bllu(t) + b",w(t)
and w(t) is callcd an additiue disturbance. This is the type of disturbancc ,. + ~-----Jr""
~L1Jll'IlS;
most often studied in dynamical eguations.
, I
9. There are two types of specifications in the design of control systems. One
is calíed the steady-state performance, the other the transient performance.
The steady-state performance is specified for the response as t -> (jJ. Hence
asymptotic tracking and disturbance rejection belong to this type of
lb---­
Figure 9-22 Multiv8
specification. The transient performance is specified for the response right
after the application of the reference signal and disturbance. Typical
specifications are rise time, settling time, and overshoot. They are governed
mainly by the location of the poles of the overall systems or, eguivalently,
the roots of Dc(s)D(s)<f>.(s) + Nc(s)N(s). The relationship between these poles
. and the transient performance is generally complicated. For a discussion
of this problem, séeReference S46.
10. To conclude this subsection, we remark that disturbances can be roughly
. classified as nóíse-type aÍld waveform-structured disturbances. The former
L-__, ­
reglJires the statisticaJ description and is studied in stochastic control
theory. See, for example, Reference S10. The latter is describable by Figure 9-23 Placem(
ASYMPTOTlC TRACKING AND DISTURBANCE REJECnON 495

.c/J(s) - 1 and the total differential equations such as the ones in (9-93) and (9-94). For an extensive
;) +2 degc/J(s)-l. A discussion of this type of disturbances, see References S121 and S122.

'Multivariable case. Consider the design of the multivariable control


system shown in Figure 9-22. The plant is described by a q x p proper rational
'>(s) and Nc(s)jDAs) is matrix G(s) factored as G(s) = 0-I(s)N(s), where D(s) and N(s) are, respectively,
can be considerably q x q and q x p polynomial matrices. lt is assumed that the q x 1 reference
signal r(t) and the q x 1 disturbance signal w(t) are modeled as
ance \.I'(t) in the plant r(s) = O,:-l(s)N,(s) (9-100)
and w(s) = O,~ l(s)N,)s) (9-101 )

where O,(s) and O",(s) are q x q polynomial matrices and Nr(s) and Nw(s) are
_ 1s" - 1 +N" _2S " - 2 + q x 1 polynomial matrices. The problem is to find a compensator so that, for
as any N,(s) and any N,)s),

(9-98) e(t) = lim (rU) - y(t)) = O (9-102)

-Iy, Yw(s) = D -1 (s)w(s) This is the problem of asymptotic tracking and disturbance rejection.
Following the single-variable case, the design will consist of two steps:
(9-99) introduction of an internal model and stabilization of the feedback system.
There is, however, one important difference: the location of the internal model
nce we may combine is critica! in the multivariable case. To iIIustrate this point, we consider only
w(t) by assuming r(t) = O. The model D:; I(S) of the disturbance w(t) is placed as
shown in Figure 9-23. Let ew(s) be the input of D - 1 (s) as shown. Then we have
" ew(s) = w(s) - N(s)O,~ l(s)NAs)D; l(s)D- 1(s)ew(s)
Figure 9-19 can be
fhis equation can be 01' e,)s) = [1 + N(s)O;;: l(s)N c(s)D; I(S)O-I(SW 1-W(S)

) +eww(t) -------1
[

'-if~"'"''''''
I
: type of disturbance
f----------'-___/)
)ntrol systems. One : i \
3.nsient performance.
Ii - - - - - - - - - - - - - - - - - - ­1I
•iI ' .
nse as t -> co. Hence
mg to this type of
Figure 9-22 Multivariable feedback system.
01' the response right
isturbance. Typical
. They are governed ~(S)
:::ms 01', equivalently,
I between these poles

~
y.(S)
d. For a discussion I (.n .

~.~~~-I
nces can be roughly
·bances. The former
1 stochastic control
:::1' is describable by Figure 9-23 Placement or internal model for disturbance rejection.
496 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

This equation and (9-101) imply if no root of cjJ(s) is a

Y,..(s) = D - l(s)[1 + N(s)D~ l (s)Nc(s)D c- 1 (s)D - I(S)J -1 D;;: l(s)N\vC~)


= Dc(s)[D(s)Dc(s) + N{s)D,~ I (s) Nc(s)J -1 D;;: 1(s)N",{s) (9-103 )
ran k[IJ-A
-c
Although there is D,~ I{S) inside the parentheses, because ofthe noncommutative where {A, R, C, E} is
of N{s) and D,~ 1 (S), it cannot be brought out of the parentheses to cancel D,~ 1 (s). of A or the degree of
Hence there is no guarantee that the internal model D,~ 1 located as shown will
accomplish the disturbance rejection. Proof
Now we shall place the internal model D;;: l(S) at the input of the summing
point where w{t) enters the system. In other words, we interchange the positions Using Theorem 9-4.
ofthe blocks D;;: l{S) and N(s) in Figure 9-23, or replace N(s)D;;: 1 (s) by D,~ 1 (s)N(s) lowed by C(s) = DI"':
in (9-I03). Then we have trollable if and only
Since D¡(s) and N/(s)
y".{s) = Dc{s)[D(s)Dc(s) +D;;: l(s)N(s)N c(S)] -1 D,~ 1 (s)Nw(S) is always controllabh
= Dc{s){D;;: 1(s)[Dw(s)D(s)DAs) + N(s)Nc(s)J} -1 D;;: l(s)N w (s)
The tandem conr
= Dc(s)[D,..(s)D(s)Dc(S) + N{s)Nc{s)J - l Dw(s)D;;: 1 (s)N",(S) (9-104) N 12 (s) ~ N/(s) are ri¡
We see that ir there is no cancellation between Dw(s)D{s)DAs) + N(s)Nc{s) and matrix
Dw(s), then the modes of w(t) are completely canceled by the internal model.
Therefore, in the multivariable case, the position of the internal model is very
important; it must be placed at the summation junction where the disturbance
signal enters the loop. In the scalar case, because of the commutative property, has rank p for every !
the position of the internal model is not critical, so long as it is not in the forward cjJ{X) =1= O, and rank M
paths from r and e and from w to y. M(A) = rank N¡(A). H
By a similar argument, in orderto achieve asymptotic tracking, we must rank NI()c)=p. By DI
place D,:- fes) at the input of the summing point where the reference signal r mission zero of G(s).
enters the loop. In other words, we must place D r- 1 (s) after D- 1 (s) or in the this part of the theorej
feedback path. Ir the internal model is required to locate inside the compen­ by [cjJ{s)I q ] - I is simila
sator shown in Figure 9-22, then it must commute with N(s) to achieve distur­
bance rejection and commute with N(s) and D- 1 (s) to achieve asymptotic With this result, \\
tracking. For a general q x p rational matrix C(s) = D- 1 (s)N(s), the internal compensators to achi
model which has these commutative properties must be very restricted. Ir the is a generalization of
internal model is chosen of the form a(s)I, where a{s) is a polynomial and 1 is a
unit matrix, then it has all the commutative properties. Furthermore, these
commutative properties hold even if there are parameter perturbations in
C(s). Hence, by choosing the internal model of the form a{s)l, the design will Consider the feedbacl
be robust. pletely characterized
Consider the reference signal and disturbance signal given in (9-100) ancl that the reference sig
(9-101). Let cjJ(s) be the least common denominator of the unstable poles of
every element of D; 1 (s) and D;;; l(S). Let C(s) = D¡-l(S)N¡(s) = N(s)D- 1{s),
where D¡(s) and N/(s) are left coprimeand N(s) and D(s) are right coprime.
Then Á is called a transmissian zera of' C{s) ir rank N(Á) < min (p, q) or rank
N/(X) < min (p, q). See Appendix H.

Theorem 9-21
The taQdem connection of cjJ -l(s)l p followed by the q x p proper rational
Figure 9-24 M ullivari
matrixC(s)( C(s) followed by cjJ - 1 (s)I q ) is controllable and observable if and only
ASYMPTOTlC TRACKING AND DlSTURBANCE REJECTlON 497

if no root of c/J(s) is a transmission zero of (;(s) and q ~p (p ~ q) 01', equivalently,

l(s)N .,,(5) IJ-A


rank :]=Il+P (l1+q) for every root l of cP(5) (9-105)
(s) (9-103) [ -c
fthe noncornmutative where {A, B, C, E} is any irreducible realization of (;(s) and n is the dimension
leses to cancel O,~ 1 (s). of A 01' the degree of (;(5).
located as shown wil\
Proof
input of the surnming
~rchange the positions
Using Theorem 9-4, the tandem connection of lp(cP(s)I p)- 1 = (cP(s)I p)-1 I p fol­
;)0;;: 1(5) by O,: 1(s)N(s) lowed by (;(s) = O¡-I(s)N¡(s), where O¡(s) and N¡(5) are left coprime, is co,,·
trol\able if and only if Ods) ~ O¡(s) and N1Z(s)N,.¡ ~ N1(s)I p are left coprime.
Since 0¡(5) and N¡(s) are left coprime by assumption, the tandem connection
s) is always controllable.
1 (s)N,.(5) The tandem connection is observable if and only if 011(S) ~ (cP(s)I p) and
N",(s) (9-104) N¡z(s) ~ N¡(s) are right coprime 01', equivalently, the (p +q) x p polynomial
:)Oc(5) +N(s)Nc(s) and
matrix
'Y the internal model.
nternal model is very M(s) ~ [cP(S)I p] (9-106)
N¡(s)
where the disturbance
omrn utative property, has rank p for every s in iC (Theorem G-8). Ir A is not a root of cP(s) = 0, then
it is not in the forward cP(I~) i= 0, and rank M()~) = p. If), is a root of cP(s) = O, then <f¡(A) = O and rank
M(A)=rank N¡(A). Hence the tandem conn~ction is observable.ifand only.if
tic tracking, we tnust rank N¡(A) = p. By Oefinition H-3, rank N¡(A) = p if and only if A is not a trans­
the reference signal r mission zero of G(s). This together with Theorem H-6 completes the proof of
after O-l(S) 01' in the this part of the theorem. The proof of the tandem connection of (;(s) followed
te inside the compen­ by [<f¡(s)I q ] - 1 is similar and is omitted. Q.E.D.
'Il(s) to achieve distur­
) achieve asymptotic With this result, we are ready to develop the condition for the existence of
-l(s)N(s), the internal compensators to achieve tracking and disturbance rejection. The condition
very restricted. Ir the is a generalization of the scalar case in Theorem 9-20.
polynomial and 1 is a
Furthermore., these Theore¡-n 9-22
eter perturbations in
Consider the feedback system shown in figure 9-24 where the plant is com­
1 a(s)I, the design will
pletely characterized by its q x p proper rational matrix (;(s). It is assumed
given in (9-100) ancl that the reference signal r(t) and the disturbance signal w(t) are modeled as
the unstable poles of
1(s)N¡(s) = N(S)O-l(S), w

(s) are righ t coprime.


~j < min (p, q) 01' rank

lX P proper rational
Figure 9-24 Multivariable feedback system.
observable if and only
498 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

res) = D,:-I(S)N,.(s) and w(s) = D;;; 1(s)Nw(S). Let <jJ(s) be the least common The design consi~
denominator of the unstable poles of every element of D,:-I(S) and D\~ I(S). If and stabilization of
no root of <jJ(s) is a transmission zero of G(s) and p ¿q or, equivalentIy, Hence the total comr
model <jJ(s) appears ~
rank [JJ--CA BJE = n +q for every root 1, of <jJ(s) are called, accordin¡
modes of r(t) and w
there exists a compensator with a p x q proper rational matrix such that the established in Theon
feedback system is asymptotical1y stable and achieves asymptotic tracking appear in e(t). In ere
and disturbance rejection. zero of the plant G(
observable mode or (
Proof will make the system
tracking and disturb~
Let G(s) = D- 1(s)N(s) be a left-coprime fraction and let <jJ -1(s)l q be the internal
C(s) and all of <jJ(s) m
mode\. The compensator Nc(s)D c- 1(s) is to be designed to stabilize the unity
We note that the j
feedback system. From Figure 9-24, we have of the reference and
e(s) =r(s) - D- 1(s)N(s)NcCs)D; l(s)(<jJ(s)l )-le(S)
q
be brought out of the
the blocking zeros of
Hence the transfer matrix from r to e is equal to De(s), N(s) and Ne(s)
Ge,.(s) = [1 + D- 1(s)N(s)N c(s)D; 1(s)(<jJ(s)lqr 1] - 1 (9-107) and (9-108). 1
= {D- 1(s)[D(s)(<jJ(s)l q )D c(s) + N(s)Nc(s)]D; l(s)(<jJ(s)I q )-I} -1 (9-107) of robustness, see Re!
= <jJ (s) DcCs)[<jJ(s)D(s)Dc(s) + N(s)Ne(s)] -1 D(s) We discuss in the
single copy of the ref(
Similarly, the transfer matrix from w to y can be computed as wil1 be general1y not
Gyw(s) = <jJ(s)De(s)[<jJ(s)D(s)DcCs) + N(s)N~(s)] -1 (9-108) 0-I(s)N(s)NcCs)D e- 1(1
Then we have
If p ¿ q and if no root of <jJ (s) is a transmission zero of G(s), then G(s) = D - 1(s) N (s)
followed by <jJ -1(s)Iq is controllable and observable (Theorem 9-21). Hence e(s) = [1 -+
we have deg [<jJ-l(s)lqD-l(S)N(s)] =deg [<jJ-l(s)I q ] +deg det D(s)(Theorem 9-1) We assume, without I(
which implies that <jJ(s)D(s) and N(s) are left coprime. Consequently the roots real parts. Since alll
of the determinant of of e(s) have negative
D ¡(s) ~ <jJ(s)D(s)De(s) + N(s)Ne(s) matrix. If D,.(s) and !"
matrix if and only if]
can be arbitrarily placed. in particular, placed in the open left-half plane by a t.hat D(s) = D(s)j[}l(sl 1,
proper choice af JDc{s) and NAs) (Theorem 9-1ó Oi 9-19). i-lence llii;; wúilj is possible if and only
feedback system is asymptotically stable.
Al1 roots of <jJ(s) have zero or positive real parts, hence no cancel1ation G(s)C(s)
between the roots of <jJ(s)D(s)Dcls) + N(s)Ne(s) and the roots of <jJ(s) will occur.
<';:onsequentIy, <jJ(s) will app~ar as a nUlperator of every element of Ge,.(s) and This is consistent witl
Gyw(s). !'i0te that wAe have Gew(s) = - Gyw(s). placed at the point wl
With Ge,·{s) and Gew(s), the response e(t) due to r(t) and w(t) can be written as this design, we req uil
However, the comput
e(s) = Ge,.(s)r(s) + Gew(s)w(s) = Ge,.(s)D; 1(s)N,.(s) + Gew(s)D; l(s)N w(s)
C(s) is found fora gi
(9-109)
subtracted and move
Since cjJ(s) appearsas zeros of every element of Ge,.(s) and Gew(s), áll unstable turbatiol1s inG(s).. H
poles of D,:-I(S) and D;;; I(S) are canceled by <jJ(s). What remains in e(s) are all design, the design' pr
stable poles. Hence, for any Nw(s) and N,.(s), we have e(t)-->O as t--> 00: This straightforward and ti
proves the theorem. Q.E.D. is m uch larger.
ASYMPTOTlC TRACKING AND DlSTURBANCE REJECTION 499

~ the least common The design consists of two steps: introduction of the internal model cP -1(s)lq
),:-I(S) and O:I(S). Ir and stabilization of the feedback system by the compensator N c(s)D c- 1 (s).
:, equivalently, Hence the total compensator is given by Nc(s)(Oc(s)p(S))-I. B~cause the internal
model cP(s) appears as zeros of every element of Ge,.(s) and Gew(s), these zeros
-t ), of cP(s) are called, according to Definition H-4, the blocking zeros. Ir the unstable
modes of r(/) and W(/) are the blocking zeros of Ce,(s) and Cew(s), then, as
matrix such lhal the established in Theorem H-7, for any initial state these unstable modes will not
asyrnptotic tracking appear in e(/). In creating these blocking zeros if any of them are a transmission
zero of the plant (;(s), then this zero will become an uncontro!lable 01' un­
observable mode 01' a hidden mode of the system. This unstable hidden mode
will make the system useless in practice. Hence, in order to achieve asymptotic
tracking and disturbance rejection, no root of cP(s) can be a transmission zero of
-1(s)l q be the internal (;(s) and aH of cP(s) must be the blocking zeros of C ...(s) and CeAs).
to stabilize the unity We note that the internal model consists of q copies cP -1(s)l q of the dynamic
of the reference and disturbance signals. Because of these q copies, cP(s) can
be brought out of the parenthes~s, as shown in (9-107) and (9-108), and become
the blocking zeros of Ge,·(s) and Gyw(s). Consequently,the perturbations of O(s),
0cCs), N(s) and NcCs) will not affect these blocking zeros as can be seen from
(9-107) and (9-108). Hence the design is robust. For a more detailed discussion
(9-107) of robustness, see Reference S81.
We discuss in the following that the design can also be achieved by using a
single copy of the reference signal as an internal mode!. The design, however,
ed as wil1 be generally not robust. Consider Figure 9-22. Define (;(s) = C(s)C(s) =
sn -1 (9-108) D -1(s)N(s)NcCs)D c- 1 (s)é O-I(s)N(s), where O(s) and N(s) are left coprime.
Then we have
hen (;(s) = 0- 1(s)N(s)
leorern 9-21). Hence c(s) = [1 + G(s)] -lr(S) = [O(s) + N(s)ll D(S)O,-I (s)N,(s)
det O(s)(Theorern 9-1) We assume, without loss of generality, that aH poles of 0,:- 1 (s) have nonnegative
onsequently the roots real parts. Since all poles of (O(s) + N(s)) - 1 have negative real parts, all poles
of c(s) have negative real parts if and only if O(S)D,:-I(S)N,(s) is a polynomial
matrix. Ir D,.(s) and N,.(s) are left coprime, then O(s)D,:- 1(s)N,(s) is a polynomial
matrix if and only if 0,.(5) is a right divisor of 0(5), 01' there exists a D(s) such
~n left-half plane by a that D(s) = 6(s)O,.(s) (see Problems 9-19 lo 9-21). Henc~ asymptoOr. tn3cJdng
9). Hence the unily is possible if and only if a compensator C(s) can be found such that

lence no cancellation C(s)C(s) = 0- l(s)N(s)NcCs)Oc-1(s) = (D(s)O,(S))-1 Ñ(s)


ots of cP(s) will occur. =0,-I(s)D-1(s)N(s) (9-110)
elernenl of Ce,.(s) and This is consistent with the statements that thc internalmodel, O,:-I(S), must be
placed at the point where the reference signal enters the loop. We see that, in
. w(t) can be written as this design, we require only one copy of the dynamic of the reference signa!.
However, the computing of C(s) to meet (9-110) is complicated. Even if such a
,.(s)D: l(s)N w(s) C(s) is found fora given C(s), there is no guarantee that O,:-I(S) can still be
(9-109)
subtracted and moVed to the leftmost posilion as in (9-110) if there are per­
d Cew(s), all unstable turbations in C(.~) ..Herice this c1esign will not be robust. In contrast with this
:emains in e(s) are all design, the design procedure of employing cP - 1 (s)lq asan internalmodel is
1).....,,0 as 1-+00. This straightforward and the design is robust. However, the degree ofcompensators
Q.E.D. is much larger.
500 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

Similar to the single-variable case, the perturbation of the internal model Consider the feedbac
is not permitted in the robust designo All other remarks in the single-variable
case are equally applicable here and will not be repeated.
Before moving to the next topic, we remark on the necessity of the condition
p ?:.q; that is, the number of plant inputs be larger than or equal to the number
of plant outputs. If q > p, the tandem condltion of <jJ -l(s)l p followed by G(s)
is controllable and observable, although G(s) followed by <jJ -l(s)lq is not
It is a plant with one
(Theorem 9-21). Consequently, we may introduce the internal model as shown
No matter how the f~
in Figure 9-25. The transfer matrix from r to e in Figure 9-25 can be computed as
at the outputs, the in
Ge,.(s) = [1 + 0- 1(s)N(s)<jJ - 1(s)lpNc(s)O; 1(S)J -1 (9-111 )
A k
If we write 0-I(s)N(s)<jJ-1(S)l p =0-1(S)<jJ-l(S)N(s)=(<jJ(S)0(S))-IN(s), then u(s) =- +(t
s
(9-111) can be simplified as
Hence we have, by u:
(9-112 )

This equation is similar to (9-107). Although <jJ(s) in (9-107) may become lirr
blocking zeros of Ge,.(s), the <jJ(s) in (9-112) can never be all blocking zeros of
We see that, unless 1
G.,.(s). This can be seen from the step of writing
achieve asymptotic f
0-1(s)N(s)<jJ -1(s)l p = 0-1(S)<jJ -1 (s)IqN(s) = [<jJ(s)O(s)J -1 N(s) asymptotic tracking,
equal to the q-dimen:
We see that p copies of <jJ(s) is increased to q copies. Because <jJ -1(s)I p fol1owed
q x p matrix, in order
by G(s) is controllable and observable, we have deg (O -1(s)N(s)<jJ - 1(s)I p ) =
of plant inputs must
deg det O(s) + p deg <jJ(s) (Theorem 9-1). The degree of det <jJ(s)O(s) is equal to
Using this argurr
q deg <jJ(s) +deg det O(s), which is, because q > p, .larger· than
Theorem 9-22. If r(
deg «<jJ(s)0(s))-1 N(s)). Hence <jJ(s)O(s) and N(s) are not left coprime. Con­
tracking, we need ran
sequently, not al1 the roots of the determinant of
for every root ). of ,
<jJ(s)O(s)Oc(s) + N(s)Nc(s) and if }, is not a pole e
is not a transmission
can be arbitrarily assigned (Corollary 9-19). Furthermore sorne roots of <jJ(s)
will appear as unassignable roots, and there is always cancellation between
Static decoupling
<J¡(s)O(s)Oc(s) + N(s)Nc(s) and 4) (s). Hence not all roots of <jJ(s) will be the
feedback system shO\
blocking zeros of Ger(S) in (9-112). Consequently, if q > p, asymptotic tracking
terized by its q x p prc
is not always possible. A similar remark applies to disturbance rejection.
~¡1t l~Jr~J.i -?{.J'} == ;~:" ..
We use a different argument to establish once again the necessity ofp ?:.q.l J
reference signals are s
output as t -HfJ is cal
18 This argument was provided to the author by Proressor C. A. Desoer. application of r(t) is
matrix from r to y t

Figure 9-25 Multivariable feedback system with q> p. Figure 9-26 Design (
ASYMPTOTIC TRACKING AND DISTURBANCE REJECnON 501

;)f the internal model Consider the feedback system shown in Figure 9-25 with
in the single-variable

~ssity ofthe condition


~ equal to the numper
G(J;~,~l
(s)I p fo\lowed by G(s)
j by cP-1(s)I q is not
L+3J
It is a plant with one input and two outputs and is BIBO stable. Let r(s) = ro/s.
ernal model as shown No matter how the feedback system is designed, in order to have step functions
25 can be compu ted as
at the outputs, the input to the plant must be of the form
:s)J -1 (9-111)
k
u(s) =- +(terms with poles inside the open left-half s plane)
p(s)D(s)t 1 N(s), then s
Hence we have, by using the final-value theorem,
>(s)4>(s) (9-112)

(9-107) may become


lim y(t)= limsG(s)u(s)= G(O)k =
t-oo s-o
[-O°.5Jk
al\ bloeking zeros of
We see that, unless ro is inside the range space of (;(0), it is not possible to
achieve asymptotic tracking. Hence for any ro in IRq, in order to achieve
)D(s)] -1 N(s) asymptotic tracking, it is necessary to have the range space, gjl(G(O)), of G(O)
equal to the q-dimensional real space (IR q, IR) ol' rank (;(0) = q. Since (;(s) is a
LUse 4> -1(s)I p followed
q x p matrix, in order to have rank (;(0) = q, we need p ?:.q; that is, the number
¡r l(s)N(s)4>-I(S)I p ) =
of plant inputs must be greater than or equal to the number of plant outputs.
et 4>(s)D(s) is equal to Using this argument, we can now establish intuitively the conditions in
> p, larger· than
Theorem 9-22. Ir r(s) = ro/(s - A) with A?:.O, in order to achieve asymptotic
t left coprime. Con­
tracking, we need rank (;(..1.) = q. For the general case, we require rank (;(..1.) = q
for every root A of </>(s). Ir we use the copl'ime fraetion G(s) = N(s)D- 1 (s)
and if Ais not a pole of (;(s), then rank (;(..1.) = q if and only if rank N(A) = q, or }.
. is not a transmission zero of (;(s). This is essential\y Theorem 9-22.
)re sorne roots of 4>(s)
cancel\ation between
Static decoupling: Robust and nonrobust designs. Considel' the
s of cP(s) wil1 be the feedback system shown in Figure 9-26 where the plant is completely charac­
7, asymptotic traeking
tel'ized by its q x p proper rational matrix G(s). Let the reference signal r(s) be of
¡rbance rejection.
the fornl r(s)== ds- 1, where <rn is an arbitrary q)( i constant -'Ector; ~úa~ :J:> ~~'l'-'
le necessity of p ?:. q. 18
reference signals are step functions of various magnitudes. The response at the
output as t-" 00 is caBed the steady-state response. The response right after the
[.
application of r(t) is caBed the transient response. Let the overall transfel'
matrix from r to y be G[(s). Ir G[(s) is BIBO stable, then the steady-state

+ y

~
-~-~
+

Figure 9-26 Design of unity feedback system.

L.
502 LINEAR TIME-INVARIANT COMPOSlTE SYSTEMS

response due to r(s1= ds- 1 can becomputed as, by using the final-value theorem, discussed in Figure S
hence we set P = 1 in]
lim y(e) = lim sG¡(s)((s) = lim sG ¡(s)ds- ¡ = G¡(O)d (9-113)
We find a C(s) so tha
t-cc s-o s-o
Now if G¡(O) is diagonal and nonsingular, in particular, a unit matrix, then
the feedback system is said to be seaeically decoupled. Indeed, if G¡(O) =
have negative real pa
diag {h¡, h z , ... , h q }, then (9-113) implies
computed, by using ('
lim h(t) =hid i i = 1, 2, ... , q
t-"" G¡(s)
where Yi and di are the ith components of y and d. Hence the steady-state
response at the ¡th output of a statically decoupled system depends solely on the Hence we have
ith reference input and is decoupled from the other inputs. Ir there is a change
in the magnitude of the ith reference step input, it will cause responses at all
output terminals. However, as time approaches infinity or as the transient Since al! the poles of
dies out, it will cause only a change at the ith output and no change at other is nonsingular. Ir s
outputs. Hence in a staticalIy decoupled system, only the steady-state responses Now we may design
are decoupled, but not the transient responses. This differs from the decoupled constant matrix N(O
system discussed in Section 7-6 where all the responses, transient as well as we may choose
steady state, are decoupled. Furthermore, decoupling is defined for any
reference signal; whereas, static decoupling is defined only for step reference
inputs. The class of step reference signals, however, is very important and is
often encountered in practice. For example, temperature and humidity controls
and G¡(O) becomes
of a room are this type of reference signals. Maintaining an aircraft at a fixed
altitude is another example.
Hence the system in f
Asymptotic tracking actually achieves decoupling as e---+ 00; the steady state
to the one in Figure (
of Yi(e) tracks r¡(e) and is independent of r ie), for j i= i. Hence the design for
Nc(O) and D ¡(O), then
asymptotic tracking can be directly applied to static decoupling. In this case,
marize the preceding
we have <b(s) = s. Let K = Iq and P = <b - l(s)I q = S-1 I q in Figure 9-26. We then
design a compensator C(s) in Figure 9-26 to stabilize the feedback system. As
shown in Theorem 9-22, if P'2:q and if s is not a transmission zero of G(s), then s Corollary 9-22
will appear as a zero of every element of (;er(S), the transfer matrix from r to e. Consider the feedback
Hence we have (;er(O) =0 and charact~~ized by its q
~.~:~r:' ·~:-f f;J ,,\ ~~.~'
lim e(i) = lim [¡r(ei - Y(é)] = Hm sGe.. (S)~S . . \ = Ge ,(O);1 = 0 ri --­
"

t- 00 t- 00 s-o
Because of e(s) = r(s) - y(s) or Y(s) = r(s) - (;e,.(s)r(s), the transfer matrix, G¡(s),
[rom r to y is equal to
G¡(s) = 1 - (;er(S) then there exists a com
feedback system is ¡
Consequently, we have (;¡(O'....= Iand the feedback system is statically decoupled. {A, B, e, E} is any im
Note that every element of G¡(s); except those on the diagonal, has s as a zeJ,"o
in its numerator. We remark once
Because of the presence of the internal model, the design is robust. .Tl:iat internal model or.nonl
is, the system remains to be statically decóu~led with perturbations, even large the remark in (5) on ti
perturbations, of the parameters of C(s) and G(s), so long as the feedback system· To check· the conditi
remains to be asymptotically stable. In the following, we introduce a design of (;(s) and is complica
which is not robust. The design is an extension of the single-variable case 9-22 is eq uivalent to r
ASYMPTOTIC TRACKING AND D1STURBANCE REJECTION 503

:final-value theorem, discussed in Figure 9-21. In this desilin, no internal model will be employed;
hence we set P = 1 in Figure 9-26. Let G(s) = N(s)D- 1(s) and C(s) = D e- 1(s)N e(s).
(9-113) We find a C(s) so that all the roots of the determinant of the polynomial matrix

a unit matrix, then D¡(s) = Dis)D(s) + Ne(s)N(s)


Indeed, if G¡(O) = have negative real parts. The transfer matrix from r to y in Figure 9-26 can be
computed, by using (9-68), as
G ¡(s) = G(s)[I + C(s)G(s)] - lC(s)K
= N(s)[De(s)D(s) + Ne(s)N{s)] -1 Ne(s)K = N(s)Dj 1 (s)Ncls)K
:nce the steady-state
jepends so lely on the Hence we have
If there is a change (9-114 )
ause responses at al!
{ or as the transient Since all the potes of D j I(S) have negative real parts, the constant matrix D¡(O)
j no change at other is nonsingular. If s is not a transmission zero of G(s), then rank N(O) = q.
teady-state responses Now we may design a compensator C(s) so that rank Ne(O) = q and the q x q
'S from the decoupled constant matrix N(O)Dj I(O)N e(O) is nonsingular. Under these assumptions,
transient as well as we may choose
; is defined for any
(9-115 )
lly for step reference
ery important and is and G¡(O) becomes
¡nd humidity controls
an aircraft at a fixed G¡{O) = 1
Hence the system in Figure 9-26 is statically decoupled. This design is, similar
the steady state
--> OC! ; to the one in Figure 9-21, not robusto Ir there are any perturbations in N(O),
Hence the design for NiO) and D¡(O), then the system will not be statically decoupled. We sum­
IUpling. In this case, marize the preceding results as a corollary.
~igure 9~26. We then
feedback system. As
Jn zero of (;(s), then s Corollary 9-22
"er matrix from r to e. Consider the feedback system shown in Figure 9-26 where the plant is completely
charactelized by its q x p proper rational matrix G(s). If s is not a transmission
zero of G(s) and p ¿Cj or, eqEl·v2JeEt~.:~/:,

ransfer matrix, G¡(s),

then there exists a compensator with a q x p proper rational matrix such that the
feedback system is asymptotically stable and statically decoupled, where
s statical\y decoupled. {A, B, C, E} is any irreducible realization of G(s). . •
gonal, has s as a zero
We remark once again that the design can be robust by introducing an
:sign is robusto That internal model or nonrobust without introducing an internal model. Similar to
turbations, even large the remark in (5) on the single~variable case, the condition is necessary as well.
LS the feedback system
To check the condition in Corollary 9-22, we must find a copriine fraction
ve introduce a design of G(s)and is complicated. If G(s)hasno pole at s =0, thecondition in Corollary
e single-variable case 9-22 is equivalent to rank G(O) = q, which can be easily checked.
504 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

State-variable approach. In this subsection, we shall discuss the design


of robust control systems to achieve asymptotic tracking and disturbance
rejection by using sta te-variable equations. The discussion wil\ be brief
becallse the procedure is similar to the transfer fllnction approach. Consider a
plant described by
X =Ax +Bu + Bww(t)
y =Cx +Eu +Eww(t)
where A, R C. K B"" and E w are. respectively, 11 x n. 11 x p. q x n, q x p. 11 X q. and
q x q constant matrices. It is assllmed that {A, B} is controllable and {A, e}
is observable. It is also assumed that the disturbance signar w{t) is generated by

with some unknown initial state. The problem is to design a robust control
system so that the output of the plant will track asymptotica11y the reference
signal r(t) generated by Figure 9-27 Design of r
X,.(t) = A,.x,.{t) r(t) = C,.x,.(t)
with some unknown initial state. Let </>,)05) and </>,.(s) be the minimal poly­ This connection was sh(
nomials of A w and A,., respectively, and let if and only if p 2: q and r
we reestablish directly t
be the least common multiple of the closed right-half s plane roots of </>,.(05) and ,11­
rank
</>,.(05). Thus a11 roots of </>(05) have nonnegative real parts. The internal model [ -(
</>-l(05)I q can be realized as
Theorerh 5-7 implies th~
Xc = Aex e + 8 ee
Ye =X e rank V(s) ~

where A e = block diag Ir, r, ... , r} for every S in e Ir fA,


every s in e Ir s is not é
'-------v---'
q-tuple
~ rank(05I-Ae)=mq. He
Be = block diag {t, t, ... , t} not a root of </>(05), then v
ro o
ü 1
Next we show that unde
with
at the roots of </>(05). We
lJ.. o o

and e = r - y as shown in Figure 9-27. FEe is an mq-dimensional equation. This


V(s) = [~n o
internal model is cal\ed the o5ervocompensator in References S64 to S66. Note
tha.t the- output of the servocompensator consists of al\ qm numbers of state The first factor has rank
variables. This is possible if (he servocompensator- is implemented by using second factor has rank n
operational amplifiers; resistors, and capacitors. -Now consider the tandem (9-117). Hence Sylvester
connection of the plant fol\owed' by. the servocompensator. Its composite rank V(s) 2: (n
dynamical equation is, as derivedin (3-63),
. Since V(s) is an (11 +mq) )
(9-116 )
ASYMPTOTIC TRACKING AND DISTURBANCE REJECTlON 505

w
lall discuss the design
¡
dng and disturbance
Icussion will be brief
lpproach. Consider a

,q X 11, q X p, 11 X q. and
lmtrollable and {A, C}
l1al w(t) is generated by

::sign a robust control


,totically the reference
Figure 9-27 Design of robust system to achieve tracking and disturbance rejcction.

be the minimal poly­ This connection was shown in Theorem 9-21 to be controllable and observable
if and only if p 2.q and no root of </J(s) is a transmission zero of the plant. Now
we reestablish directly that (9-116) is controllable if

,Iane roots of </Jw(s) and rank [ AI-A BJ =/l +q for every root ). of </J{s) (9-117)
s. The internal model -C E
Theorem 5-7 implies that (9-116) is controllable if and only if

rank V(s) ~ rank [


sI -A
BcC
O :
:
si - A c • -
B
B
J
cE
= n +mq

for every s in e Ir {A, B} is controllable, we havc rank [s 1- A B] = n for


every s in e Ir s is not an eigenvalue of Ac or, equivalently, a root of ¡P(A), then
rank (sI - A r ) = mq. Hence we concludc from the structure of V(s) that ir <; is
nol a root of </J(s), then we have
rankV(s)=n -!-:ne
Next we show that under the condition of (9-117), Equation (9-118) still holds
at the roots of </J(s). We write

~nsional eq uation. This


VI') ~[~ :, ,[ ~AJ['T [~,. ~EJ fn +q +mq
nces 564 to 566. Note
11 qm numbers of state The first factor has tank iz + mq by the irreducible realization of </J -1(s)I q • The
second factor has rank 11 +q +mq at every root of </J(s),by the assumption of
implemented by using
IJ consider the tandem
(9-117). Hence 5ylvester's irtequality (Theorem 2-6) implies :
nsator. Its composite rank V(s)2.(n +mq) +(n +q +mq) - (n +q +mq) = n +mq
Since V(s) is an (n + mq) x (n +mq +q) matrix, we conclude that
(9-116 ) rank V(s) = n +mq
506 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS
DE~

at every root of 1>(S). Consequently, (9-118) holds at every s in e and (9-116) is


control1able.
[f (9-116) is control1able, then the eigenvalues of the composite system can
be arbitrarily assigned by the state feedback

u=[K KeJ[:J=KX +Kexe


Hence by a proper choice of K and KC' the feedback system in Figure 9-27 can
be stabilized. If the state of the plant is not available, we can design a state
estimator and then apply the feedback gain K at the output ofthe state estimator.
This completes the design of a robust control system to achieve asymptotic (a)
tracking and disturbance rejection. For a more detailed discussion, see Refer­ Figure 9-28 Input-out
ence S81.

9-7 Design 01' Compensators: Input-Output Then 9¡(5) becomes


Feed back Systems
Single-variable case. In this section, we study the design problem in
Section 9-5 for a different configuration. Specifical\y, we study the condition Theorem 9-23'
on the plant for the existence of proper compensators in the input-output
feedback system shown in Figure 9-l(b), to achieve arbitrary denominator Consider a plant with tr:
matrix. We also study the minimum degrees ofcompensators to achieve the 11. For any DAs) of deg
designo We note that the input-output feedback configuration is deve\oped proper Com pensators L
frorn the design of state feedback and state estimator -discussed in Chapter 7. Figure 9-28(a) has tram
Consider the system in Figure 9-1(b). We study first the case C(s) = 1. The D(s) and N(s) are coprirr
compensators Co(s) and C I(S) are required to have the same denominator as
Proof
Co(s) = L(s) (9-119 )
DAs) For any DAs) of degree lí
D ¡(s) - DAs)D(s) is of de
M(s)
and C(s)=-­ (9-120) to (9-122) yields that (9- i
DAs) and only if D(s) and N(s)
Because of this restriction, the feedback system in Figure 9-l(b) can be redrawn
as shown in Figure 9-28(a). if deg DJ~) = m, the ¡;( 2 [",liana! ,{lB."'>
[Co(s) C I(S)] is of degree In and can be realized by using In integrators. Ir We compare first thi
the configuration in Figure 9-28(a) is reduced to the single-loop system shown order to enSure the prope¡
in Figure 9-28(b), then the denominators of the two compensators are different. in Theorems 9-11 and 9-1
Hence their implementations require twice as many integrators as the one in and require the degree of
Figure 9-28(a). Hence, although the two configurations in Figure 9-28 are Theorem 9-23' is chosen 8
equivalent mathematical\y, they are different in actual implementation. proper whether D¡(s) is (
The transfer functionfrom r to y in Figure 9-28(a) can be computed as function of the unity feed
we can control only D r(s
. N(s)De(s) feedback system in Figu
9 (s) = 9(S) A
(9-121 )
¡ 1 + Co(S) + C I(s)g(s) DAs)D(s) + L(s)D(s) + M(s)N(s) Dr(s) as well as De(s). The
.Define Figures 9-12 and 9-28(a).
and the compensator D'
D¡(s) = De(s)D(s) +L(s)D(s) + M(s)N(s) degree, the former has t~
or D¡(s) - De(s)D(s)= L(s)D(s) + M(s)N(s) (9-122) three sets of parameters D el
.;

DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 507

~ry s in e and (9-116) is y

: composite system can

(b)
;tem in Figure 9-27 can
, we can design a state
ut of the state estimator. (3)
to achieye asymptolic
Figure 9-28 lnpul-olllPlll feedback syslem.
:d discussion, see Refer-

Then 9/(s) becomes


put
(9-123)
the design problem in

we study the condition


Theorem 9-23'
)[s in the input-output
Consider a plant with transfer functiong(s) =N(s)/D(s) and deg N(s).:::;;deg D(s) =
arbitrary denominator n. For any DAs} of degree m and any D¡(s) of degree n +m or less, there exist
Jensators to achieve the proper compensators L(s)/Dcls) and M(s)fDe(s) so that the feedback system in
lfiguration is deyeloped Figure 9-28(a) has transrer function N(s)D¡ l(s)D e(s) from r to y if and only if
iscussed in Chapter 7. D(s) and N(s) are coprime and m '2. n -1.
,t the case C(s) = 1. The
le same denominator as Proof

(9-119 ) For any De(s) of degree m and any D ¡(5) of degree n + mor less, the polynomial
D /(05) - D/s)O(s) is of degree n +111 or less. The application of Theorem 9-10
to (9-122) yields that (9-122) has solutions L(s} and M(s} of degrees mor lcss if
(9-120)
and only if D(s} and N(s) are eoprime and m '2.11 - 1. This proyes the theorem.
Q.ED.
re 9-1(b}can be redrawn
1 x 2 rational matrÍz We compare first this theorem with Theorems 9-1 1 and 9-11'. First, in
using m integrators. 1f order to ensure the praperness of the compensators, we m lIst consider separately
ingle-loop system shown in Theorems 9-11 and 9-11' the cases where the plant is proper or strictly praper
mpensators are different. and require the degree of O /(s) to be exactly equal to n +m. Since the D,(s) in
IOtegrators as the one in Theorem 9-23' is chosen a priori to haye degree m, the compensators are always
:ions in Figure 9-28 are proper whether D¡(s) is of degree n + 111 or not. Second, the oyerall transfer
1 implementation. function of the unity feedback system in Figure 9-12 is N(s)D¡ l(s)N ,(s), where
can be computed as we can contraJ only D¡(s). The oyerall transfer function of the input-output
feedback system in Figure 9-28(a) is N(s)D¡ 1 (s)De(s), where we can control
IDe(s) (9-121 )
D¡(s) as well as De(s). The reason for haYing this extra freedom can be seen from
JD(s) + M(s)N(s)
Figures 9- 12 and 9-28(a).. Although the compensator D; 1(s)NAs) in Figure 9-12
and the compensator D e- l(s)[L(s) M(s)] in Figure9-28(a) have the same
degree, the foimer has two setsof parameters Del' N el' whereas the lalter has
N(s)
N(s) (9-122 ) three sets üf paraineters Dei' L¡, M j. Since ítreq lIires only two sets ofparameters
S08 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS
DE!

to meet (9-40) and (9-122), either De(s) or L(s) in (9-122) can be arbitrarily
+
assigned. We choose to assign Dc(s) because it dictates the poles of the com­
pensator.
The employment of Theorem 9-23' requires the solving of (9-122) or
E(s) ~ D¡(s) - Dc(s)D(s) = L(s)D(s) + M(s)N(s) (9-124 )

This is the Diophantine equation in (9-40) and, hence, can be translated into a
set of linear algebraic eq uations. Let
Figure 9-29 An input
D(s)=D o +D¡s + +D"s"
N(s)=N o +N¡s + +N"s"
Dc(s)=D co +Dc¡s + +Dcmsm
L(s)=L o +L¡s + +Lmsm (9-125) Although al! comI
M(s)=Mo+M¡s+'" +Mms m serious problem. Th.
D¡(s)=F o +F¡s + ... +F"+ms"+ m function and is, as dis
and E(s) = E o + E¡s + . _. + E Il + msrz+m feedback systems, we
the resulting system t
Then Equation (9-124) is equivalent to the algebraic equation posed system and sho
[Lo M o : L¡ M ¡ : ... : L m MmJSm = [E o E¡ ... EIl+mJ ~ E (9-126)
Theorem 9-23
where SIIl is defined as in (9-45). Ir m:::::n - 1, then SIIl has a ful! column rank.
Hence for any E, solutions {L i , M;} exist in (9-126). The solutions yield im­ Consider a plant with 1
mediately the proper compensators Co(s) and C¡(s). n. For any Dc(s) of dI
compensators L(s)/D c(:
Example 1 9-28(a) is well posed an
only if D(s) and N(s) al
Consider a plant with transfer function g(s) = N(s)/D(s) = (s -1)/s(s - 2).
Clearly, we have n = 2 and m "2:: n - 1 = 1. Let us choose arbitrarily Dc(s) = s + 1, We see that by re
and D¡(s) = S2 + 2s + 2. Note that the degree of D¡(s) is smal!er than n + m = 3. then assert the wel! pos
We compute feedback system is wel
E(s) =D¡(s)-Dc(s)D(s) =S2 +2s +2-(s +1)s(s-2) from O at s = OC! if deg
= 2 + 4s + 2s - S3
2
De l(s)D f (s)[;
and form
":-:.':" '.;;:; ~-;<~ ':"/.'. ¡--" ~~:>. ,"
value of De-¡(s)D¡(s)D­
[Lo Mo: L¡ 4 2 -1J for S-' 00,
D; ¡(s)D¡(s)D-¡(s)·
Its solutions are Lo = - 6, L¡ = - 1, M 0= - 2 and M ¡ = 6. Hence the com­ Hence if FIl+m=O or (
pensators are the system is not well]
Co(s) = L(s) = -6 -s 1 +Co(oo) +C¡(oo)g(oo
D/s) 1 +s . note that if deg D¡(s) < j
We consider now a
and C¡(s) = M(s) = - 2 +6s
Dc(s) 1 +s .. Corollarv 9-23
They are al! proper rational functions. The block diagram of the feedback Consider a plarit with tr
system is shown in Figure 9-29. • /1. For any i5 ¡(s) of d(
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 509

122) can be arbitrarily


1the poles of the com­

Ing of (9-122) or
1

I;)N(S) (9-124)

Iw be translated into a

Figure 9-29 An input-output feedback system which is not well posed.

(9-125 )
Although al! compensators in Figure 9-29 are proper, the system has one
serious problem. The system has a loop which yields an improper transfer
function and is, as discussed in Section 3-6, not well-posed. In the design of
feedback systems, we require not only al! compensators to be proper but also
the resulting system to be well-posed. Theorem 9-23' may not yield a well­
posed system and should be replaced by the fol!owing theorem.
En+mJ ~ E (9-126)
Theorem 9-23
,¡aSa ful! column rank.
Consider a plant with transfer functiong(s) = N(s)/D(s) and deg N(s):::; deg D(s) =
I[he solutions yield im­
n. For any Dc(s) of degree m and any D[(s) of degree n +m, there exist proper
compensators L(s)/Dc(s) and M(s)/Dc(s) such that the feedback system in Figure
9-28(a) is wel! posed and has a transfer function N(s)Dj l(s)Dc(s) from r to y if and
only if D(s) andN(s) are coprime and m ~ n - 1 . . 11
s)/D(s) = (s -1)/s(s - 2).
arbitrarily Dc(s) = s + 1, We see that by removing deg DAs) < n +m from Theorem 9-23', we can
smaller than n +m = 3. then assert the well posedness in Theorem 9-23. We show that the input-output
feedback system is well posed by showing that 1 + Co(s) + C,(s)g(s) is different
from O at s = 00 if deg D¡(s) = n +m. From (9-122), we have
-1)s(s - 2)
D; '(s)D¡(s)D - '(s) = 1 + D,- '(s)L(s) + Dc-'(s)M(s)N(s)D- '(s)
=1+Co(s)+C,(s)g(s) (9-127)

value of Dc-'(s)D¡(s)D-'(s) at s=oo. Using the notations of(9-125), we have,


= [2 4 2 - 1] for s -> 00,
Dc- '(s)D¡(s)D - '(s) -> Dc~' s -m(FII+msn+m + Fn+m _ ,Sn+ m-, + ... )s-IID;;'
, = 6. Hence the com­ Hence if Fn +m=0 or deg DAs) < n +m, then 1 +Co(oo) +C,(oo)g(oo) =0 and
the system is not wel! posed. However, if F II + m 1=0 or deg D¡(s) = n +m, then
1 + Co( 00) + e, (oo)¡]( 00) 1= O and the system is wel! posed. It is of interest to
note that if deg DAs) < n +m, the pole-zero excess jnequality in (9-38) is violated.
We consider now a speciai case ofTheorem 9-23.. :

Corollar·y 9-23
iagram of the feedback Consider a plant with transfer function g(s) = N(s)/ D(s) and deg N(s) :S deg D(s) =
I n. For any i5 As) of degree n, there exist proper compensators L(s)jD/s) and
510 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS DE~

M(s)/Dc(s) of degree n - 1 and with arbitrarily assignable poles such that the system to minimize Ü
feedback system in Figure 9-28(a) is wel1 posed and has transfer function
N(s)i5 ¡ 1 (s) if and only if D(s) and N(s) are coprime. Ili

where q> O is a weigr


This corol1ary foIlows directly from Theorem 9-23 by choosing D f(s) = function, then the opt
Dc(s)i5 As) and noting

9fes) = N(s)D ¡ 1(s)Dc(s) = N(s)i5 ¡ l(s)D; 1(s)Dc(s) = N(s)i5 ¡ l(S)

This design always involves the canceIlation of Dc(s), which can be chosen by
where i5 f (s) is a Hur'
the designer, however. The degree of Dc(s) in the coroIlary can be larger than
N(s) in the numerator
n - 1; however, it does not seem to serve any design purpose because DcCs) is
function 9(s). See R
completely canceled in the designo
requires the solution (
We compare now CoroIlary 9-23 with the result obtained in the state­
variable approach. By state feedback, we can assign the eigenvalues of A or
Multivariable case
the poles of g(s) = N(s)/D(s) as the roots of i5 As) without affecting N(s). An will be extended to the
(n -1)-dimensional state estimator with arbitrary eigenvalues can be con­
system shown in Fig
structed to generate an estimate of the state. The connection of the feedback rational matrix (;(s) ,
gain from the output of the estimator yields the overal1 transfer function p x p proper rational
N(s)ji5As) (see Section 7-5). Hence Corollary 9-23 establishes essentially the matrix C¡(s)= Dc-l(s)1
result of state feedback and state estimator. However, the result in Corollary be computed as
9-23 is slightly more general: the plant is permitted to have a proper transfer
function. In Chapter 7, we design state estimators only for strictly proper plants Gis) = (;(s)[1
or dynamical equations wi.th the direct transmission parts equal to zero. In = N(s)D
Corol1ary 9-23, we require deg i5¡(s) = deg D(s); in the design of state feedback, = N(s)[I
we require deg i5fes) = deg D(s) and i5 f" = D" (that is, their leading coefficients
are equal). Ir deg i5¡(s) = deg D(s) and i5 f" = D", the compensator Co(s) is Define
always strictly proper for g(s) strictly proper. This can be verified from the last
column .equation of (9-126) (Problem 9-28). Hence we always have D
1 +Co(oo) +CI(oo)§(oo) #=0 for the class ofsystems studied in the state-variable or E
approach, and consequently, the well-posedness problem does not arise in the
approach. Hence the result in Corollary 9-23 is more general than the result of
state feedback and state estimator.
In the transfer~function approach, we require only the concept of coprime­
ness. In the state-variable approach, we require the concepts of control1ability
and observability. In the former approach, the design consists of forming a
linear algebraic equation and its solutions yield immediately the required
compensators. In the latter approach, the design requires one similarity
transformation to compute the feedback gain, and requires one similarity
transfonnation. or one solution of a Lyapunov matrix equation to find a state
estimator. Hencefor the single-variable case, it appears that the design in
the transfer-function approach is simpler conceptua:Ily and computationaIly
than the one in the state-variable approach.
To conclude this subsection, we remarkthat Corollary 9-23 can be used in
the design of optimal systems. Consider a plant with transfer function g(s) =
N(s)jD(s) with input u(t) and output y(t). ltis required to design an overal1 Figure 9-30 Input-out]
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 5U

system to minimize the quadratic performance index


lle poles such that the
has transfer function
11 J = [' {q[y(t) - r(l)] 2 + u 2 (t)} dl (9-128)

where q> O is a weighting factor and r(t) is the reference signa\. lf r(l) is a step
l by choosing D ¡(s) = function. then the optimal system which has the smallest J is of the forro

~ (s) = ~(s)
1= N(s)i5¡ 1(S) g¡ D¡(s)

Ihich can be chosen by where 15¡(s) is a Hurwitz polynomial and has the same degree as D(s). The
lary can be larger than N(s) in the numerator of 9¡(s) is the same as the numerator of the plant transfer
urpose because D/s) is function 9(s). See Reference S46. Hence the design of the optimal system
requires the solution of Corollary 9-23.
obtained in the state­

:he eigenvalues of A or
Multivariable case. In this su bsection, the results in the single-variable case
out affecting N(s). An
wil1 be extended to the multivariable case. Consider the input-output feedback
7envalues can be con­
system shown in Figure 9-30. The plant is described by the q x p proper
~ection of the feedback
rational matrix (;(s) = N(s)D- 1(s). The compensators are denoted by the
..eran transfer function
p x p proper rational matrix Co(s) = D; 1(s)L(s) and the p x q proper rational
:ablishes essential1y the
matrix C 1(s)= D;-I(S)M(s). The transfer matrix from r to y in Figure 9-30 can
. the result in Corollary
be computed as
have a proper transfer

'or strictly proper plants


(;¡(s) = (;(s)[1 +Co(s) +C 1(s)G(S)]-1
)arts equal to zero. In
= N(s)D- 1(s)[1 + D; 1(s)L(s) + D; 1(s)M(s)N(s)D - 1(s)] -1
lesign of state feedback,
= N(s)[Dc(s)D(s) + L(s)D(s) + M(s)N(s)] -1 Dc(s) (9-129)
heir \eading coefficients

Define
: compensator Co(s) is

be verified from the last D ¡(s) =Dc<s)D(s) + L(s)D(s) +M(s)N(s) (9-130)


:nce we a\ways have E(s) ~ D¡(s)- D,(s)D(s)= L(s)D(s) + M(s)N(s) (9-131 )
or
lied in the state-variable
~m does not arise in the
;eneral than the result of

the concept of coprime­ +


ncepts of controllability
n consists of forming a
mediÍltely the req uired
requires one similarity
requires one similarity
equation to find a state
lears thaJ the design in
Iy and computationally

l~ry9-23 can be used in


transfer function g(s) =
Figure 9-30 lnput-output feedback system.
ed to design an overall
512 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS DI

Then C¡(S) beco mes input-olltput feedbae


po sed.
(9-132 )

Define In the design, tht


chosen as
H(s)= diag [S"I, su', ... , SU p
} (9-133 )
He(s) = diag {Sil", S111" • •• ,Slllp} (9-134 )

where dei (s) are arbit


Theorem 9 c 24
mi are odd integers, ti
Consider a plant with q x p proper rational matrix C(s) = N(s)D-1(s). Let In this case, if we che
J.1i, i = 1,2, ... , p, be the column degrees of O(s), and let v be the row index of the d ifficu 1ty of assigr
C(s). Let mi ~ v - 1, for i = 1, 2, ... ,p. Then for any D,(s) of row degrees mi
and row red uced, and any D ¡(s) of the property that Corollary 9-24

lim R:-1(s)D¡(s)H-I(s)=J (9-135) Consider a plant wil


s~ 00
J1¡, i = 1, 2, ... ,p, be tt
Ihen for any De(s) of
exists and is nonsingular, there exist compensators with proper rational matrices
D ¡(s) of column deg
D; 1(s)L(s) and De-I(S)M(s) such that the feedback system in Figure 9-30 is well
compensators with p:
posed and has transfer matrix N(s)D ¡ 1(s)D,(s) if and only if D(s) and N(s) are that the feedback sysl
right coprime and D(s) is column reduced. N(s)Dj I(S) if and onl
reduced.
Proof
Consider the E(s) defined in (9-131). Clearly we have Proof

lim H;I(s)E(s)H-l(s)=J-Iim H;I(s)DcCs)D(s)H-I(s)=J-DchD h (9-136) Let D ¡es) = De(s)D¡(s).


s~ 00
9-24 are met under thf
where D,,, is the row degree coefficient matrix of D,(s) and D" is the column and G¡(s) become
degree coefficient matrix of D(s). Because J - O,,,D /• exists, we conclude from
the proof of Theorem 9-18 that solutions L(s) and M(s) of row degrees at most
and
mi exist in (9-131). Consequently D,-I(S)L(s) and D,-I(S)M(s) are proper fol­
lowing the assllmption of DcC.s). Note that whether J - D c " D h is nonsinglllar This cstablishes the ca
or not is immaterial here.
We show that the system is well posed by showing that !+Co(co)+ The applic3.tion o
el (co )C( ex)) is nOl1singuiar. F rorn (9- dO j, Wé.; navc:, sirniía¡ LV Ci- iL."/j, r, ¡ \r!t, use Lúe coed
lO; L

shown in (9-75). We t
D c- l(s)D¡(s)D -I(S)= 1 + D,- I (s)L(s)+ D e- l(s)M(s)N(s)D-I(S)
from top to bottom. I
= 1 + Co(s)+ C¡(s)C(s) (9-137)
block row of S, are lir
We write, similar to (9-83), convenience, we assum
L(s) =L o + L
D(s) = [D /I + D/(s)]H(s) (9-138)
De(s) = H,(s)[D eI, + DcI(s)] (9-139 ) M(s) = M o +~
and E(s)=O/s)­
where D/(s) and Dcl(s) are strictly proper. Theri we have
where J1 = max {,ui, i =
lim D; 1 (s)D ¡(s)D(s) = lim [Del. + DcI(:s)] -1 He- l(s)D¡(s)H - 1 (s)[D" + O/(S)]-I (9-131).yields
s-+ co s-+ 00 .

(9-140)
[Lo Mo : L I MI

which is nonsingular by the assumptions of O,(s), D(s), and (9-135). Henee the
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 513

input-output feedback system in Figure 9-30 is, following Theorem 3-6, well
posed. Q.E.D.
(9-132)

In the design, the denominator matrix De(s) of the compensators can be


chosen as
(9-133)
(9-134) (9-141 )

where dei(s) are arbitrary Hurwitz polynomials of degrees mi. Ir some O!' al1
mi are odd integers, then we may not be able to assign complex conjugate roots.
;(s) = N(s)D -les). Let In this case, if we choose DcCs) to be of the form shown in Problem 2-29, then
v be the row index of
the difficulty of assigning complex conjugate roots will not arise.
)e(S) of row degrees mi
Corollary 9-24
Consider a plant with q x p proper rational matrix G(s)= N(s)D-'(s). Let
(9-135 )
Il¡, i = 1,2, ... ,p, be the column degrees ofD(s) and let v be the row index of G(s).
Then for any De(s) of row degrees aH equal to v - 1 and row reduced, and any
roper rational matrices fi ¡(s) of column degrees Ili, i = 1, 2, ... , p, and column reduced, there exist
n in Figure 9-30 is well compensators with proper rational matrices D; '(s)L(s) and D; l(s)M(s) such
11y ir D(s) and N(s) are that the feedback system in Figure 9-30 is well posed and has transfer matrix
N(s)fi¡ '(s) if and only if D(s) and N(s) are right coprime and D(s) is column
reduced.

Proof
Let Df(s) = De(s)D¡(s). Clearly the degree requirements ofD¡(s) in Theorem
9-24 ~re met under the assumptions ofDcCs)and DAs). With this D¡(s), (9-131)
and D" is the column and G ¡(s) become
üsts, we conclude from DcCs)[fi¡(s) - D(s)] = L(s)D(s) + M(s)N(s) (9-142)
of row degrees at most and G feS) = N(s)[DcCs)D ¡(s)] -, De(s) = N(s)D¡ '(s) (9-143)
(s)M(s) are proper fol­
- D dl D iI is nonsingular This establishes the corollary. Q.E.D.

ving that 1 + C o(ro) + The application of Theorem 9-24 and its corol1ary is straightfol'W8.cd.
ilar to (9- i 27), First we use the coefficient matrices 01 D(s) and N(s) to form the matrix Sm
shown in (9-75). We then search the linearly independent rows ofS m in arder
l(s)N(s)D- '(s) from top to bottom. Let v be the least integer such that aH N rows in the last
(9-137) block row of S. are linearly dependent. This v is the row index of G(s). For
convenience, we assume mi = v - 1 for aH i in Theorem 9-24. Let
L(s)=Lo+L,s+'" +L._,s·-' (9-144)
(9-138)
M(s)=M o +M,s + ... +M._,s·-' (9-145)
(9-139)
arid E(s) = D ¡(s) - DcCs)D(s) =Eo + Els + ... + El' h ' _ ¡S" +v -l (9-146)

where Il =max {Il¡, i = 1,2, ... , p}. The substitufion of (9-144) to (9-146) into
IH - '(s)[D" + D¡(s)]-' (9-131) yields
(9-140) [Lo Mo:L l Ml:"<L'- 1 M,-,]S'-l=[E o El ... El'h-l]
(9-147)
and (9-135). Hence the
514 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS DI

The solution of this set of linear algebraic equations yields the required com­ Using G(s)= N(s)D­
pensators. be written as
It is ofinterest to compare the result in Corollary 9-24 with the one developed
G¡(s) = N
in the state-variable approach. If D(s) and N(s) are right coprime and D(s) is
column reduced, then an irreducible realization can be readily found for (;(s) = Ir we define
N(s)D- 1(s) (Section 6-6). By state feedback, we can achieve N(s)Dj I(S), where
D¡(s) and D(s) have the same column degrees and the same column-degree­
coefficient matrix. Note that the latter condition is not required in Corollary and E(s) ~
9-24. If the state is not available for feedback, we may design a state estimator
then G¡(s) = N(s)D ¡
with arbitrary eigenvalues, which are equivalent to the roots of det Dc(s), to
Theorem 9-24 to acb
generate an estimate of the state. Now the application of the state feedback
ever, the row degrees
from the output of the state estimator yields N(s)Dj I(S). Note that the eigen­
degrees of Dc(s) will t
values of the estimator are not control1able as can be seen from (7-86) and will
introductionofQ-l(:
not appear in the transfer matrix from r to y.
A remark is in order regarding the degreesof compensators. In the state­
Theorem 9-25
variable approach, the dimension ofthe state estimator is n -q (Theorem 7-13).
In the transfer-function approach, the degree of the compensators is p(v -1). Consider a plant wil
Note that the row index v of (;(s) is equal to the observability index of any assumed that D(s) an<
irreducible realization of (;(s), and has the property v 2.n/q. If p =q = 1, then column degrees f1.i, i =
n = v and the dimension of the state estimator is equal to the degree of com­ P x P arbitrary polynl
pensators. Ir p 2.q, then p(v -1) 2.n - q; if p < q, p(v -1) can be greater than, and be row reduced.
equal to, or less than n - q. Hence the results in the state-variable and transfer­
function approaches are not exactly identical. Similar to the single-variable
case, the design procedure in the transfer-function approach appears. to ·be is a proper rational
simpler, conceptual1y and computationally, than the one in the state-variable D c-l(S)M(s), and Q­
approach. posed and has trans
In the fol1owing, the input-output feedback system in Figure 9-30 will be computed from
extended to the one shown in Figure 9-31, in which Q(s) is a polynomial matrix
and Q -1(S) is required to be proper. The transfer matrix from r to y in Figure
9-31 can be readily computed as with (SciR 1(s) < ()ci O(s)
G¡(s) = (;(s)Q-l(s)[I +C O(S)Q-l(S) +C 1(s)(;(S)Q-l(S)]-1 (9-148)

Proof

The column degrees O


and M(s) of row deg
D c-l(S)L(s) and D c- I(S
Next we. show th
(9-152). We use (9-13:
D(s)D¡I(S)D
Since D 11 + D¡(s)and D
Figure 9-31 An input-output feedback system. is proper.
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 515

ds the required com­ Using (;(s) = N(s)D-I(S), Co(s) = D,:-I(S)L(s) and C1(S)= Dc-I(S)M(s), G¡(s) can
be written as
'ith the one developed G¡(s) = N(s)[Dc(s)Q(s)D(s) + L(s)O(s) + M(s)N(s)] -11\(s) (9-149 )
t coprime and 'P(s) is
Ldily found for G(s) = Ir we define
:ve N(s)Dj 1(s), where D¡(s) ~ Dc(s)Q(s)D(s) +L(s)D(s) +M(s)N(s) (9-150)
,ame coluron-degree­
required in Corol1ary and E(s) ~ D¡(s)- Oc(s)Q(s)D(s)= L(s)O(s)+ M(s)N(s) (9-151)
sign a state estimator then Gj(s) = N(s)Dj l(s)Dc(s) and has the same form as (9-132). The design in
roots of det Dc(s), to Theorem 9-24 to achieve (9-132) is accomplished without using Q-I(S); how­
of the state feedback ever, the row degrees of Dc(s) are general1y difTerent. In the fol1owing, al1 row
Note that the eigen­ degrees of Dc(s) wil1 be required to be the same. This is possible because of the
n froro (7-86) and wil\ introduction ofQ - 1(s). In this case, R(s) in (9-134) becomes diag {Slll, Slll, ... ,s"'}.

lsators. In the state­ Theorem 9-25


l1-q (Theorem 7-13).
lpensators is p(v -1). Consider a plant with q x p proper rational matrix (;(s) = N(s)D-1(s). It is
"Vability index of any assumed that D(s) and N(s) are right coprime and D(s) is column reduced and of
l/q. lfp=q=1,then column degrees ¡J¡, i = 1,2, ... ,p. Let v be the row index of (;(s). Let Dc(s) be a
lo the degree of com­ p x p arbitrary polynomial matrix of row degrees all equal to 111 with m;::: v - 1
) can be greater than, and be row reduced. Then for any D¡(s) with the property that
-variable and transfer­
H(s)D ¡ ¡ (s)H,(s) (9-152 )
to the single-variable
)roach appears to be is a proper rationa¡" matrix, there exist proper compensators Dc-l(s)L(s),
¡: in the state-variable Dc-I(S)M(s), and Q-I(S) so that the feedback system in Figure 9-31 is well
posed and has transfer matrix N(s)D¡ 1(s)Dc(s). Furthermore, Q(s) can be
m Figure 9-30 will be computed from
s a polynoroial matrix
( froro r to y in Figure D ¡(s) = Q¡(s)D(s) + R1(s) (9-153)

with c)ciR¡(s) < (\¡D(s) = p¡, and

QI (s) = Dc(s)Q(s) + R 2 (s) (9-154 )

with o,.¡ 1f«2{S) < o,.¡ Dc(s) = m, and íL(s) and IVK(sj are soiutlons 01

R 2 (s)D(s)+ R1(s)= L(s)D(s)+ M(s)N(s) (9-155 )

Proof
The column degrees of R 2 (s)D(s) + R 1(s) are c1early at most 111 + ¡Ji. Hence L(s)
and M(s) of row degrees at most m exist in (9-155) (Theorem 9-17). Thus
Dc-I(S)L(s) and Dc-I(S)M(s) are proper.
Next we show that D(s)D ¡ 1(s) Dc(s) is proper under theassumption of
(9-152). We use (9-138) and (9-139) with m = mi, for all i, to write
D(s)D¡ 1(s)Dc(s) = [O" + D¡(s)]H(s)D¡ 1(s)Hc(s)[D c" + Dcl(s)] (9-156)

Since D" + D¡(s)and Dc/,+ D,¡(s) are prope"r, if(9-152) holds, then D(s)D¡ 1(s) Dc(s)
IS proper.
516 LINEAR TIME-INVARIANT COMPOSlTE SYSTEMS

From (9-153) and (9-154), we have proper rational mat]


problem will be studi
Of(S) = Oc(s)Q(s)O(s)+ R 2 (s)0(s)+ RI(s) (9-157)

which implies Implementations


loop system shown
O; 1(s)O f(s)O - 1(s) = Q(s) + Oc- l (s)R 2 (s)+ Oc- 1(s)R¡(s)O -I(S) (9-158) proper rational matr
Because O(s) is column reduced and Oci O(s) > ociRI(s), RI(S)O-I(S) is strictly this subsection, we d
proper. Because Oc(s) is row reduced and o,-¡ Oc (s) > o.-iR 2 (s), Oc- l (s)1 is proper feedback configurati
and Oc- l(s)R 2 (s) is strictly proper. Hence the polynomial part of N(S)O-l(S) and ifT(s)
O;I(S)Of(S)O-I(s) is Q(s). Since [Oc- I(S)Of(S)O-I(S)]-l is proper, it follows transfer matrix is G a
from Theorem 3-4 that Q-I(S) is proper. trary assignment of e
The feedback system is well posed if and only if the matrix is square and nonsinf

P- I(S) ~ [1 + Oc- 1(s)L(s)Q - I(S) + Oc- l(s)M(s)N(s)O- 1(s)Q - 1(s)] - 1 T(


= Q(s)O(s)[ Oc(s)Q(.I')O(s) + L(s) 0(.1') + M(s)N(s)] - 10c(s) (9-159) where di(s) are Hun
is proper (Theorem 3-6). The substitution of (9-150) into (9-159) yields T(s) proper, then the
discussed may find se
p-I(S)= Q(s)O(s)O¡ l(s)Dc(s) (9-160) LetG(s)= N(s)O­
which implies Then the open-Ioop ti

pes) = O; I(S)Of(s)D - 1(s)Q - 1(s) (9-161 ) Go(s) = I

where
Using (9-153) and (9-154), we have
pes) = 0c- 1(s)[OAs)Q(s)O(s) + R2(S)0(S) + R 1(s)] O - 1(s)Q -I(S) and D ¡(s) and Nfes)
= 1+ D c- 1(s)R 2 (s)Q -'1 (s) + Oc- 1(s)R 1(sjO - 1(s)Q - I(S) (9-162) shown in Figure 9-16
Because O; 1(s)R 2(s) and RI(s)D - 1(s) are strictly proper, D c-1(S) and Q - I(S) are is to be solved from a
proper, the polynomial part ofP(s) is I. Hence P- I(S) is, following Theorem 3-4, hence the unity feedb
proper. This establishes the well-posedness of the feedback system. Q.E.D. the open-loop system
9-31, the overall tran
We remark the condition in (9-152). One way to check the properness of assignable. A compa
H(s)O¡ l(s)Rc(s) is by direct computation. The computation of the inverse 01 bility of implementinl:
Of(S) is, however, complicated. Instead, we may compute and O f(S) are meto .
introdüce a differcnt ¡

which can be obtained by inspection because of the forms of Rc(s) and H(s). lmplementation l. e
Clearly Y(s) is much simpler than D feS). Now H(s)O¡ 1(s)Hc(s) is proper if and Figure 9-31. Its oyera
only if y-I(S) is proper (Theorem 3-4). This is a simpler way of checking the Gfes) = G(s)Q-I(s)[1
condition in (9-152). Because of Hc(s) = diag {sm, s"', ... ,Slll}, a sufficient con­
dition for y-I(S) to be proper is that Df(s) has column degrees m+ /li and is
column reduced. Even if O feS) is not column reduced, it is still possibl~ for 19 This probIem was first 1
Reference SSO.
Y(s) to have a proper inverse. See the footnote on page 115.
The designs in Theorems 9-24 and 9-25 yield
G feS) = ~(s)D ¡ ](;)'Oc(s) = N(s)O - 1(s)D(s)D¡ 1(s)Oc(s)
= G(s)T(s) (9-163)

where T(s) ~ 0(.1')0¡l(S)Oc(.S) is, as proved in (9-156), a proper rational matrix.


Itturns out that every system expressible as G(s)T(s), where T(s) is an arbitrary Figure9-32 An open­
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 517

proper rational matrix, can be implemented as shown in Figure 9-31. This


problem wil1 be studied in the next section.
:.1') (9-157 )
Implementations of open-Ioop compensators. 19 Consider theopen­
loop system shown in Figure 9-32, where the plant is denoted by the q x p
)R 1(s)D- 1 (s) (9-158) proper rational matrix (;(s) and T(s) is an arbitrary proper compensator. In
this subsection, we discuss the implementations of this open loop system by
~l(s)D-l(S) is strictly
feedback configurations. Befare proceeding, we mention that if (;(s) =
¡(s), D;- 1(.1')1 is proper
N(s)D - 1 (s) and if Tis) is ch~sen as D(s)D¡ 1(s) and is proper, then the open-loop
)olynomial part of
transfer matrix is Go(s) = G(s)T(s) = N(s)Di I(S). This is the problem of arbi­
1 is proper, it follows
trary assignment of denominator matrix discussed in Corollary 9-24. If C(s)
is square and nonsingular and if we choose T(s) as
atrix
'I(S)Q-l(S)] -1
T(s) = (; -1(S) diag {dI (s), d 2(s), ... ,dp(s)} (9-164)

~)] - 1 D,(s) (9-159) where d¡(s) are Hurwitz polynomials of smallest possible degrees to make
T(s) proper, then the resulting system is decoupled. Thus the problem to be
(9-159) yields
discussed may find several applications in the design of multivariable systems.
(S-15D) LetC(s)= N(s)D-l(s) be a coprime fractíon with D(s) column reduced.
Then the open-loop transfer matrix is
(;o(s) = (;(s)T(s) = N(s)D- l (s)T(s) ~ N(s)Dj l(s)N ¡(s) (9-165 )
(9-161 )
where D- l (s)T(s) = D¡ l(s)N ¡(s) (9-166)

[)-I(S)Q-l(S) and D ¡(s) and N ¡(s) are left coprime. If we use the utlity feedback system
~)Q-l(S) (9-162) shown in Figure 9-16, the overall transfer matrix is N(s)D¡ l(s)Nis) and Nc(s)
Dc- 1 (s)and Q-l(s)are
is to be solved from a Diophantine equation and cannot be arbitrarily chosen;
hence the unity feedback system in Figure 9-16 cannot be used to implement
lllowing Theorem 3-4,
the open-loop system. Ir we use the input-output feedback systemin Figure
ck system. Q.E.D.
9-31, the overall transfer matrix is N(s)D¡I(s)D c (s), where Dc(s) is arbitrarily
assignable. A comparison of this with (9-165) reveals immediately the possi­
eck the properness of
bility of implementing Co(s) in Figure 9-31 as long as the conditions on O,(s)
ltíon of the inverse of
and D¡(s) are met. This will be done in Implementation l. We shall also
te
introduce a different implementation in Implementation n.
y)H - ¡ (s)
Implementation l. Consider the input-output feedback system shown in
ms of R(s) and H(s).
Figure 9-31. Its overall transfer matrix is, from (9-148) and (9-149),
.)Hc(s) is proper if and
r way of checking the (;¡(s) = G(s)Q -1(s)[1 + D - l (s)L(s)Q - I(S) + D; 1 (s)M(s)N(s)D -1 (s)Q -1(S)] - 1
c
,sm}, a sufficient con­ (9-167)
degrees m + J1¡ and is
it is still possible for 19 This problem was firsl f¿rmulaled and solved in Reference S218. This presentation fo\lows
Reference SSO.
115.

(9-163)

roper rational matrix.


Figure 9-32 An open-Ioopsystem.
ere T(s) is ah arbitrary
518 LINEAR T1ME-INVARIANT COMPOSITE SYSTEMS D

or G¡(S) = N(s)[Dc(s)Q(s)D(s) + L(s)D(s) +M(s)N(s)] -1 D/s) (9-168) Proof

where Dc(s) can be arbitrarily chosen. Hence this configuration may be used The substitution of(
to implement G(s)T(s) = N(s)D; l(s)N ¡(s). In this implementation, we require G¡(S) = 1
T(s) to be nonsingular. The nonsingularity ofT(s) implies the nonsingularity of
Nfes). Consequently we can always find a unimodular U(s) such that U(s)N¡(s) Hence G¡(S) implem
is row reduced and U(s)D¡(s) and U(s)N¡(s) remain to be let coprime (Theorem Equations (9-169
G-11). Thus we assume without 10ss of generality that N¡(s) is row reduced. T(s) =
A In the following, we present a procedure so th,Ú G[(s) will implement
G(s)T(s). Note that Dc(s)D ¡(¿
U sing (9-138) and (9·
Step 1. Compute a fraction (;(s) = N(s)D-l(s), where D(s) ann N(s) are right
coprime and D(s) is column reduced. Compute [D1.+ D¡(s)]
D-l(s)T(s) = D; 1 (s)N ÁS) (9-169) Since Oh and Del,
[Del. + Dcl(s)] - ¡ are,
where D ¡(s) and N ¡(s) are left coprime and N Ás) is row reduced. Let the condition in (9­
DriN ¡(s) = mi, i = 1,2, ... ,p Theorem 9-25.
where Dri denotes the ith row degree. We discuss now .
Step 2. Compute the row index, v, of (;(s). Define depends on T(s). Ifl
det D¡(s) in (9-169) is
will also be BIBO s1<
Let det Dc(s) which are 1
.o/s) = diag {Cl. l (s), Cl.2(S), ... , Cl.p(s)} Hurwitz, then the fee
where lXi(S) is an arbitrary polynomial of degnie m - mi' Then the matrix The decompositi(
algorithms in Referel
Dc(s) = .o/s)N¡(s) (9-170)
design, the condition
has row degrees all equal to m:;:::: v - 1 and is row reduced. (9-172) by
Step 3. lf
i= 1, 2, ... ,p (9-171 )
with Dc¡R¡(s):::;;DciD(s
set Q(s) = 1 and go to step 4. If not, compute example, from the un
D ¡(s) = Ql(s)D(s) +Rl(s) (9-172) Q¡{
with C\¡Kl(S) < ÓciÜ(s) = j1i, for aH i, and compute ¡Ol" dúy COllstalil rnatr

Ql(S)= N¡(s)Q(s) +R 2 (s) (9-173)

with Dri R 2 (s) < DriN ¡(s) = mi, i = 1, 2, ... ,p. These decompositions are with D,J{2(S) :::;;Dr¡N¡l
unique. proper. However, f(
Step 4. Solve L(s) and M(s) from Furthermore, the res'
degree conditions in
.oc(s)[D fes) - N ¡(s)D(s)] ;= L(s)D(s) + M(s)N(s) (9-174a)
degrees at most m stil
or .o/s)[R 2 (s)D(s) + R 1 (s)] = L(~)D(s) + M(s)N(s) (9-174b) .
1mplementatioll 11.
Theorem 9-26 section. In this iinph
The input-output feedback system in Figure 9-31 with Dc(5), Q(s), L(s), andM(s), . square and may not t
cAomputed from (9-170) and(9-174) implements the open-Ioop system (;o(s) = in FigLire 9-31 excep
G(s)T(s) = N(s)D; l(s)N fes) and is well posed. . terminal as shúwhin
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 519

(9-168) Proof
The substitution of (9-170) to (9-174) ¡nto (9-168) yields
guration may be used
nentation, we require G¡(s) = N(s) [fic(s)D¡(s)J -lfic(s)N¡(s) = N(s)Dj l (s)N ¡(s) (9-175 )
s the nonsingu\arity of
(s) such that U(s)N ¡(s)
Hence G¡(s) implements the open-Ioop system.
let coprime (Theorem Equations (9-169) and (9-170) imply
~¡(s) is row reduced. T(s) = D(s)O¡ l(s)N¡(s) = O(s)[fic(s)D¡(s)] - 1 Oc(S) (9-176 )
G¡(s) will implement
Note that Dc(s)D¡(s) in (9-175) corresponds to the D¡(s) in Theorem 9-25.
Using (9-138) and (9-139) with mi = m, for all i, we can write (9-176) as
t(s) ann N(s) are right
[D h+ D/(s)J - lT(s)[D ch + Dcl(s)J - l = H(s)[fic(s)D¡(s)J - l Hc(s) (9-177)

(9-169) Since D" and D,.!> are nonsingular by assumption, [Oh + D,(s)J - 1 and
[Och + DcI(s)J - l are, following Corollary 3-4, proper. Hence, if T(s) is proper,
ow reduced. Let the condition in (9-152) is satisfied and the theorem fol1ows directly from
Theorem 9-25. Q.E.D.

We discuss now the stability of the implementation. Clear\y, the stability


depends on T(s). lfT(s) is chosen so that T(s) and G(s)T(s) are BIBü stable, then
det D¡(s) in (9-169) is a Hurwitz po1ynomial, and the feedback implementation
wil1 also be BIBü stable. The design involves the cancellation of the roots of
det fi,(s) which are however arbitrarily assignable. lf det Dc(s) is chosen as
)} Hurwitz, then the feedback implementation is asymptotically stable as well.
.mi' Then the matrix The decomposltions in (9-172) and (9-173) can be carried out by using the
algorithms in References S34, S 137 and S236. See also Problem G-15. In the
(9-170)
design, the conditions in (9-172) and (9-173) can be l'elaxed. We may replace
uced. (9-172) by
(9-178)
... , P (9-171) with ¿¡c¡RI(s) ::s:;oc¡D(s). In this case, the decomposition is not unique. Fol'
example, from the unique QI(S) and RI(s) in (9-172), we can obtain

(9-172) RI(s) = R(s) - BO(s) (9-179)

101 any constanl matrix ¿, LO meeL (9- i 78 j. 3íl11liélnj, (9-1 ¡ j) can be replaced by
(9-180)
(9-173)

decompositions are with (j,.¡Rz(s)::S:;¿¡riN¡(s). In this case, we cannot expect every Q¡I(S) to be
proper. However, for almost all QI(S) so computed, Q¡I(S) is still proper.
Furthermore, the resulting D,(s)[D ¡(s) - N¡(s)Q(s)D(s)] still meets the column
degree conditions in Theorem 9-17; hence solutions L(s) and M(s) of row
s)N(s) (9-174a)
degrees at most m still exist in (9-174).
s)N(s) (9-174b)
lmplementatiori jI. We introduce a different implementation in this sub­
section. In this implemen.tation, the op~n-IoopcompensatorT(s) may be non­
(s), Q(s), L(s), and M(s), squáre and inay n6t haveá full rank. The configuration is identícal to the one
n-loop system Go(s) = in Figure 9-31except that acorripensator Ole. l(s)K(s) is placedat the input
terminal as shown in Figure 9-33. The transfer matrix of the system is clearly
520 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS DE

Step 4. Define

Ir J: - f.J..¡ - In 2. O, ~
Hurwitz polynorr
b,.¡W(s) ~m.
Step 5. If

set Q(s) = 1 and g.


Figure 9-33 A plant input-output feedback system.

equal to, fol\owing (9-167) and (9-168),


GAs) = G(s)Q -1(s)[1 + D e- 1(s)L(s)Q- I(S) + D e- l(s)M(s)G(S)Q-1(S)] - 1D e- 1(s)K(s)
(9-181 )
or G¡(s)= N(s)[De(s)Q(s)D(s) +L(s)D(s) +M(s)N(s)]-IK(s) (9-182) with 8,.¡I~2(S) < 6,.¡
Step 6. Let
The implementation eonsists of the fol\owing steps:
Step 1. Compute a fraetíon (;(s) = N(s)D- 1(s), where D(s) and N(s) are right
eoprime and D(s) is eolumn redueed. Compute and salve L(s) and
(9-183) W(s
where DJ(s) and N J(s) are left eoprill).e. Furthermore, we require DAs)
to be eolumn redueed and to have D Jh = 1 (the eolumn-degree-eoeffieient Theorem 9-27
matrix is the unit matrix). The feedback system i
Step 2. Let 6ci D J(s) = J¡, and be¡D(s) = Ji¡, i = 1, 2, , p. Define the open-Ioop system
V(s) =diag {vI(s), V2(S), , vp(s)} (9-184)
Proof
Ir ¡;?:. {ti' set Vi(S) = 1. If J¡ < {ti' set vi(s) as a monie arbitrary H urwitz
polynomial of degree Ji¡ - J¡. We then write The sllbstitlltion. of JK

D -I(s)T(s) = D;: 1 (s)V - l(s)V(s)N J(s) = (V(s)Dr(S))-l(V(s)N As))


~fij1(S)N¡(s) (9-Hl5)
This shows that the fe
where f> J(s) ~ V(s)D J(s) and NJ(s) = V(s)NJ(s). Let be¡f> J(s) = f¡. Beeause
loop eompensator T(s
of the assumption D JI. = 1, we may eonclude Ji ?:.Jii. 20
From (9-188) and I
Step 3. Compute the row index, v, of (;(s). Let De(s) be an arbitrary polynomial
matrix of row degrees aH egual to In and row redueed sueh that m?:. v - 1 W(s)D J
and D; 1N¡(s) is proper.
Beeause b,.¡ R 2(s) < /11, (~

20 Because theproperness of T(s) = D(s)D¡ '(s)N¡(s) does no! imply the properness of D(s)D¡ 1 (s)
(see' Problem 3-42), we cannot conclude that f; ":?:.J1i. Consequently, we must introduce v¡(s) to Henee under the assu
conclude ¡; ~Jl¡. However, in most cases, for example, when T(s)has a fult rank, we have v¡(s) = 1, reduced anci m?:. v - 1
for al! i. (9-191). Hence the eo
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 521

Step 4. Define
W(s) = diag {Wl (s), W2(S), > • > , Wp(s)} (9-186 )

If J: - ¡.t¡ - m:?: O, set w¡(s) = 1. IfJ: - ¡.t¡ - m < O, set w¡(s) as a monic arbitrary
Hurwitz polynomial of degree (¡.t¡ +m - J:). Because of ¡; :?:¡.t¡, we have
ór¡W(s)':::::m.
Step 5. If
i = 1,2, .. >, P (9-187 )

set Q(s) = 1 and go to step 6. If not, compute


D¡(s) = Ql(s)D(s) +R¡(s) (9-188)

with óc¡R¡(s)<Óc¡D(s)=¡.t¡, and compute


)Q - ¡ (s)J - ¡ D; ¡(s)K(s) W(s)Q¡(s) = Dc(s)Q(s) + R 2 (s) (9-189)
(9-181 )
with ó,¡R 2 (s) < Ó,>;Dc(s) = m.
¡K(s) (9-182)
Step 6. Let

K(s) = W(s)Ñ J(s) = W(s)V(s)N J(s) (9-190)


s) and N(s) are right
and solve L(s) and M(s) fram

(9-183) W(s)D¡(s) - Dc(s)Q(s)D(s) = L(s)D(s) + M(s)N(s) (9-191)

)re, we require Df(s)


Theorem 9-27
mn-degree-coefficient
The feedback system in Figure 9-33 obtained from (9-184) to (9-191) implements
Define the open-Ioop system G(s)T(s) = N(s)Dj ¡(s)N¡(s) and is wel! posed.
(9-184 )
Proof
ic arbitrar y Hurwitz Thc substitulion of K(s) = W(s)N J(s), (9-186), and (9-191) into (9-182) yields

))-l(V(s)N¡(s))
GJ(s) = N(s)[W(s)D¡(s)J - 1W(s)Ñ ¡(s) = N(s)Df l(S)Ñ ¡(s)
;-_:: l\I(s)[j- J·(,))T(:;} ~-;~ ·:G(:;')1'(:;:
(9-185)
This shows that the feedback system in Figure 9-33 does implement the open­
C¡DJ(S)=!i. Because loop compensator T(s).
I

From (9-188) and (9-189), we have


"arbitrary polynomial
d such that m:?: v - 1 W(s)D J(s) - Dc(s)Q(s)D(s) = R 2 (s)D(s) + W(s)R¡ (s) (9-192)

ÓC¡[W(s)D I(s) - Dc(s)Q(s)D(sJ] < m + ¡.ti for al! i (9-193)


: prpperness of D(s)Dj 1(s)
Hence under" the assumption that D(s) and N(s) right coprime, D(s) column
> wr:, must introduc~ v¡(s) \0
1 fun ránk, we ha_e v,(s) = 1,
reduced and m:?: v -l,solutions L(s)"and M(s) of degrees at most m exist in
(9 c I9l).Hencethe compensators D c- l(s)L(s) and D; 1 (s)M(s) are proper.
522 LINEAR TlME-INVARIANT COMPOSITE SYSTEMS
DE

The polynomial equation in (9-192) can be written as A remark is in or


D; l(s)W(s)Df(s)D -1(S) = Q(s) + D; l(s)R 2(s) +D; l(s)W(s)R 1(s)D- I(S) In general, the degre
(9-194 ) previous subsection.
the same denominatl
1 1
Because D e- (s)R 2 (s) and R 1 (s)D- (s) are strictly proper and D;I(S)W(S) is to deg det De(s). He
proper, the polynomial part of D e- 1(s)W(s)D f (s)D- 1(s) is Q(s). To show deg det Q(s) + deg de
Q-l(S) is proper, it is sufficient to show, by using Theorem 3-4, that previous subsection.
(D e-l(S)W(s)D f (s)D- 1(s))-1 is proper. Using (9-133), (9-134), (9-138), (9-139), the implementation i:
and similar formulas for D f(S) and W(s), we have The remarks in (
previous subsection ,
D(s)D¡ 1 (s)W - 1 (s)De(s) = [D h+ Dt(s)]H(s)H [;,1(.1')[1 + D fl(S)] - 1
section and will not t
[1+ W¡(s)]-IH¡:¡;I(S)H e(S)[D eh + DcI(s)] (9-195)
In addition to the
Because D¡(s), D ft(S), W¡(S), and Dcl(s) are all strictly proper, we have, as s-> 00, back implementation
may use the single-lo
D(s)D¡ 1 (s)W - l(s)D e(s) -> DhH(s)H[;/ (s)H w(s)He(s)D eh
1 (9-196)
feedback system in I
which approaches a zero or nonzero finite constant matrix. Hence The basic idea and (
D(s)D¡ 1(.1')'" -1(s)De(s) is proper, and consequently, Q - 1(.1') is proper. mentation 1; however
Now we show that the compensator D; 1(s)K(s) = D e- 1 (s)W(s)N f(S) is steps in (9-172) and (~
proper. Consider see Reference S218.
T(s) = D(s)D j \s)N f(S) Applications. 21 11
= [D + D¡(s)]H(s)H[;/(s)[1 + D fl(S)] -1 H,v is)[Nfh + N ft(s)]
'I will be employed to dI
As .1'->00, we have T(s)-> D hH(s)H b¡1 (s)HÑ¡(s)N fh . Because T(s) is proper, we problem and then de
conclude bance rejection. Fim

(\¡ N¡(s):5: J¡ - Jt¡ Decoupling. Conside


The row degree m of Dis) is chosen so that D; l(s)N ¡(s) is proper. Now we N(s)D- 1 (s). Ifan Opé
claim that D e- 1 (s)W(s)N ((s) remains to be proper. If J¡ - Il¡ - m 2::0, br¡W(s) = 0, where ñf(s) = diag {el¡
minimum degrees to r
and br;(W(s)Nf(S)) = briN'f(s) :5:b,.¡D e(s). If J¡ - Il¡ - m < 0, b,¡W(s) = Ili + m - J¡,
the input-output feedl
and we have bri(W(s)Nf(s)):5:Il¡+m-];+J¡-Il¡=m=briDcls). Hence
D e- l(s)K(s) is proper. Dj 1(.1'). We note that
We have shown that all compensators are proper. What remains to be of det (D f(s)N(s)); hen(
shown is that the overaH system is well posed. The overal1 syst,,:,n ,~ '1,,,,11 Since the input-outou
if and only if the matrix the disappearance oi ]
cancellations. Thus j
P- I(S) ~ [1 + D e- \s)L(s)Q - I(S) + D e- 1 (s)M(s)N(s)D - I(S)Q - I(S)] - 1 involve unstable pe
= Q(s)D(s)[De(s)Q(s)D(s) + L(s)D(s) + M(s)N(s)] - 1 Dcls) (9-197) D(s)N -1(s)D¡ 1(.1') can
We discuss in the
is proper. The substitution of (9-191) into (9-197) yields
without involving an
P-l(S) = Q(s)D(s)[W(s)D ¡(s)] -ID e (s) nonsingular proper ra
which implies,by using (9-192), right coprime and D(s

P(s) = De""' l(s)[D e(s)Q(s)D(s) + R 2(s)D(s) +W(s)R 1 (s)]D - I(S)Q - I(S) . 1


= I + D; l(s)R 2 (s)Q - I(S) + D e- 1 (s)W(s)R 1 (s)D - I(S)Q - I(S) with N

Because D e- 1 (s)R 2 (s), R 1(s)D-l(s) are strictlyproperand because Q-'I(S) and


D e- 1 (S)W(s) are proper, the polynomial part of P(s) is 1, which is nonsingular.
Hence P-l(S) is proper. . Q.E.D. 21 Follows Reference S54.
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 523

A remark is in order regarding the degrees of compensators in this designo


1 In general, the degree of det Q{s) in this design is smaller than the one in the
;)W{s)R 1{s)D- {s)
previous subsection. Because D c- l{s)K{s), D c- 1{s)L{s) and D c- 1(s)M(s) all have
(9-194 )
the same denominator matrix, they can be implemented with a degree equal
1
er and 0c- {s)W{s) is to deg det Dc{s). Hence the total degree of compensators in this design is
;) is Q{s). To show deg detQ(s) + deg det Dc{s), which is generally smaller than the ones in the
~ Theorem 3-4, that previous subsection. If det D¡(s), det V{s), det Dc{s) and det W(s) are Hurwitz,
-134), (9-138), (9-139), the implementation is, similar to Implementation 1, asymptotically stable.
The remarks in (9- 17S) and (9-1 SO) regarding (9-172) and (9- 173) in the
previous subsection are equally applicable to (9-ISS) and (9-1S9) in this sub­
t- f> fl{S)] - 1
section and will not be repeated.
,+ Dcl(s)] (9-195)

In addition to the two implementations, it is possible to obtain other feed­


ler, we have, as s -> 00, back implementations of the open-Ioop compensator T{s). For example, we
may use the single-loop feedback system in Figure P9-32 or the input-output
(9-196 )
feedback system in Figure 9-30 [without using Q -l{S)] to implement T{s).
tant matrix. Hence The basic idea and design procedures will be similar to the ones in Imple­
(s) is proper. mentation 1; however, they may require less computation because the division
= D; 1{s)W{s)N f{s) is steps in (9-172) and (9-173) are not needed. For yet another implementation,
see Reference S21S.

Applications. ZI In this subsection, the results of the previous subsection


)[N fh + Ñ fl{s)] will be employed to design feedback systems. We consider first ·the decoupling
problem and then decoupling together with asymptotic tracking and distur­
¡use T{s) is proper, we
bance rejection. Finally, we study themodel matching problem.

Decoupling. Consider a p x p nonsingular proper rational matrix G{s) =


N{s)D- 1 {s). If an open-Ioop compensator T{s) is chosen as D{s)N- 1 {s)i>jl{S),
;) is proper. Now we
where O¡(s) = diag {dI (s), dz{s), . .. , dp{s)} and d¡{s) are H urwitz polyr;.omials of
U¡ - m 2:0, <\¡W{s) = O,
minimum degrees to make T{s) proper, then the implementation of G{s)T{s) in
o,.¡ W{s) = j.l¡ + m - ¡;,
1,
the input-output feedback system in Figure 9-31 will yield a decoupled system
m = 6,.¡Ok;). Hence
Oj 1(s). We note that in this design, the roots of det D{s) are shi}ted to the roots
ofdet (O¡(s)N(s)); hence there are no cancellations involving the roots of det O(s).
What remains to be
Since the input-output feedback has no effect on the numerator matrix N(s),
11 system is well posed
the disappearance of N(s) from the decoupied system is accompíished by exact
cancellations. Thus if det N{s) is not a Hurwitz polynomial, the design will
:s)Q - I{S)] - 1 involve unstable pole-zero cancellations. Thus the choice of T(s)=
1- 1 D c{s) (9-197) D{s)N - I(S)O jI (s) cannot always be employed to design decoupled systems.
We discuss in the following a method of designing decoupled systems
without involving any unstable pole-zero cancellations. Consider a p x p
nonsingular proper rational matrixG{s) = N{s)D-1{s), where D{s) and N{s) are
right coprime and D(s) is column reduced. We factor N(s) as

)]D - 1{s)Q - I{S)


N{s) == N¡{s)NzCs)

I{S)Q-l{S)
witb N I{S) = diag {/311 (5), /31 is), ... , /31 p{S)}

d because Q- 1 (s) and


which is nonsingular.
Q.E.D. 21 Follows Rererence S54.
524 LINEAR T1ME-INVARIANT COMPOSITE SYSTEMS DE

where 13 ¡¡(S) is the greatest common divisor of the ith row of N(s). Let 13 2i(S) -
be the least common denominator of the unstable poles of the ith column of D(s)N 2 (s) = L'r,
N Z1(s). Define N2d(s)~diag {f321(S), f3ds), .. . , f32P(S)} and define
Hence we may choos
N2(S)~ N Z1(s)N 2d (S)
It is a rational matrix with only stable poles, that is, poles with negative real
parts. Then we have
The implementation
N 2(s)N 2(s) = N 2d (S) = diag {f321(S), f3ds), ... , f32P(S)} matrix
Now we choose an open-loop compensator T(s) as
T(s) = D(s)N 2(s)D,-1(S) (9-198)
with D,(s) =diag {1X 1(S), 1X 2(S), ... , IXp(S)}

where IX¡(S) are Hurwitz polynomials of minimum degrees to make T(s) proper.

Then we have

Implementation I. V<
G(s)T(s) = N [(s)N 2(s)D -1(s)D(s)N 2(s)D,- l(S) = N [(s)N 2d(S)D t- 1(s)

-d' {f31(S) f32(S) f3 p(S)} (9-199)


- lag ( )' 1X ()
1X 1 S S , ... , IXpS()
2 S2 +2.s
_ 4
where f3¡(s) = f3¡¡(S)f32i(S). Hence the implementation of G(s)T(s) will yield a - [ _l s +
4
decoupled system. We note that this design involves only the cancellations of
the stable roots of det N 2(s), and the resulting feedback system will be asym­ where D res) and N r(~
ptotically stable. ced ure discussed in A
v = 2, Clearly we ha'
Example 2 choosing Dc(s) arbitrE

Consider the plant

G(S)=N(S)D- 1(S)=L s:l s ~lJ[s2;1 ~T1


which has row degree:
We compute f311(S) = 1, and f312(S) =s +1. Hence we may factor N(s) as 6d (D c (s)D r (s)),:::; Ji¡ + 17

N(s)=Nl(S)N2(S)=I~ . . .J
0,_,1ls,2 11
...J

Wecompute
Thus, the L(s) and M(
_[
N 2 (s) = -------:­ 1 [1
(s+l)(s-1) -1
Hence we have

N 2As) =diag {(s -1), (s -1)} (9-200)

and 1
'N 2(s) = Ni (s)N 2As) = [_ ~ ~21 ] s ~ 1
Now we choose T(s) =D(s)N 2(s)D,-1(S), withD,(s) = diag {1X 1(S), 1X 2(S)}, where
IX¡(S) are Hurwitz polynomials of minimumdegrees to make T(s) proper. We = [-3//4 ~27/4
compute 54 13/4 '
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 525

of N(s). Let ~Zi(S)


~J[ -~
)W
of the ith column of
,nd define
Hence we may choose

,les with negative real D,(s) = [s 0+


2
(s
O
+ 1)2
J (9-201 )

The implementation of G(s)T(s) will yield a decoupled system with transfer


, f3 zP(s)} matrix
GIS)TIs) = dia {s - 1 (s + 1)(s ­
\ \ g S + 2' (s + 1)2
l)}
(9-198)
. {s - s- 1}
=dlag - -1. - ­ (9-202)
s+2's+1
5 to make T(s) proper.
Implementation I. We implement G(s)T(s) in Figure 9-31. We compute
s)N zAs)D,-I(S)
D-11)TI)=Ñ()D-"II)=[
\s \s 2
S
, \s, -1
1 -IJ[(S+2)(S+I)
sZ_ O
O J-l
(s+l?
(9-199)
=[SZ+~S+~ -i-s-~ J-l[ 1 -i-J~D-l(S)N (s)
r G(s)T(s) will yield a _l s +.1
4 4
SZ +lls +2
4
-1
4
s _.1 -
4
J J
Iy the cancellations of
where D J(s) and N J(s) are left coprime. This is computed by using the pro­
system will be asym­ cedure discussed in Appendix G. The row index of G(s) can be computed as
v = 2. Clearly we have /11 = max {v - l,ml, /11z} = max {2 -1, 0, 1} = 1 and, by
choosing Dc(s) arbitrarily as Dc(s) = diag {s + 3, I},

-
Dc(s) = DC<s)NAs) =
[s +3
O
1
-i- J
0J[ 1 s-­
-1 '4
1 = r3
_
+
1

which has row degrees all equaJ to m = 1 and is row reduced. Clearly we havc
.y factor N(s) as ¿¡c¡(f>c(s)DJ(S»~¡Li+l11,i= 1,2, hence we may set Q(s)= l. We compute

~l
lj Js +';' J
Thus, the L(s) and M(s) in (9-174) can be solved from
1 1 O O 1 O
O O O 1 O O
O· . 1- 6- ·0- -1- -Ó
1 1 1 1 O O
---------------------
(9-200)
1 1 O O 1 O
1 000100
+1
0--(-6--0--1--0
1 1 1 1 O O
19 {CC1(S),<xZ(s)}, where .
iake T(s) proper. We =[-3//4 -27/4: 13/2 -9/4: 9/4 O: O 00J
54 13/4 : - 1/4 3 :1 O: O
DE~
526 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

unstable pole-zero ca
as

L(s)= [
(3l/2)S+33/4
(-7/2)s- 2
27/4
-1/4
J _ [( -31/2)s- 6
M(s)- (1/2)s+2
-9J
13/4
This completes the first implementation. where #¡(s) are uniql
to be chosen to make
Implementation JI. We implement G(s)T(s) in Figure 9-33. We have to be decoupled onl)'
D¡/e =1 and J¡ ?:.p¡; hence V(s) =1, D¡(s) =D¡(s), and Ñj(s) = N¡(s). We choose designed, in additior
m = v -1 = 1and rejection, iX¡(S) cannot
the additional feedba
bance rejection.
In order to achie'v
be robust with respe
Because JI - PI - m = 2 - 2 - 1 = -1, we choose WI(S) arbitrarily as s +2, of discussed in Section 9
degree 1. Because.f2 - P2 - m = 2 - 1 - 1 = O, we set w 2 (s) = 1. Hence we have Ifwe introduce a diagl
with h¡(s)/cjJ(s) proper

~J
S +2
W(s)= [ O reduces to p number e
transfer function of th
Clearly, we have 6ci (W(s)O ¡(s)):::; Il¡ +m, i = 1, 2; hence we set Q(s) = 1 and

s+2 3(s
-~~ +2)J
~ Its denominator is a 1
r
K(s) = W(s)N ¡(s) =
-1 s--,¡ h¡(s)/</>(s) to be strictly
h¡{s); there are deg a¡{.
The L(s) and M(s) in (9-191) can be solved as and #¡(s) are coprime,
proper choices of h¡(s) ¡
L(s) = [llS + 21/4 (-3/4)S+3J -1IS-4 -27/4J and disturbance reject
(- 5/2)s - 2 -5/2 M(s) = [ (5/2)s + 2 9/4 or, beca use N(s) in G
#¡(s) is a factor of det r
This completes the second implemental ion. 111

For this example, the total degrees of compensators of these two imple­
+
mentations are the same. !n general, the total degree of com1Jensators in the ·r~
----/I~-
second implementation is less than or equai to the one in the rirst implementa­
tion.
The design procedure discussed in this subsection can be modified in
several ways. For example, if a stab1e root of det N 2 (s) is very close to the
imaginary axis, it may be retained in #2;(S), instead of being canceled. Instead
of decoupling the plant for each pair of input and output, we may decouple it
for a group of inputs and a group of olitputs. In this case, the plant is to be
decoupled into a block diagonal mat~ix. These modifications ar~ straight­
forward and will not be discussed.
'? E
Asymptotic Tracking, DistUl'bance Rejection, and Decoupling . . 111 this .sub­
section, we shall design a robust system· foachieve decoüpling,asyú1ptotic I

tracking and disturbance rejection. Let C(s) = N(S)O-I(S) be a JI x JI 110n­


singular proper transfer matrix: It can be decoupled, without involving any Figure 9-34 Oesign of
DESIGN Of COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 527

unstable pole-zero cancelIations, as

{~l(S) ~2(S) ~p(S)}


2)s- 6
1-2 13/4
-9J d.
lag CY.¡(s)' iX2(S)'···' O:p(s)
where ~i(S) are uniquely determinable from N(s) and the degrees of iX;(S) are
to be chosen to make the open-Ioop compensator T(s) proper. Ir the plant is
ure 9-33. We have to be decoupled only, CY.¡(s) can be arbitrarily assigned. Ir the plant is to be
)= N f(s). \Ve choose designed, in addition to decoupling, to achieve tracking and disturbance
rejection, (X;(s) cannot be arbitrarily assigned. They must be used to stabiJize
the additional feedback connection to be introduced for tracking and distur­
bance rejection.
In order to achieve asymptotic tracking and disturbance rejection, and to
be robust with respect to parameter perturbations, we must introduce, as
rbitrarily as s + 2, of discussed in Section 9-6, an internal model <f> -'(s)I p as shown in Figure 9-34(a).
1=1- Hence we have Ir we introduce a diagonal polynomial matrix H(s) = diag {h, (s), h 2(s), ... , hp(s)}
with h¡(s)/<f>(s) proper or strictly proper, then the system in Figure 9-34(a)
reduces to p number of single-variable systems shown in Figure 9-34(b). The
transfer function of the system in Figure 9-34(b) is clear1y equal to
set Q(s) = 1 and h;(s)~¡(s) 6. h¡(s)~¡(s)
(9-203)
<f>(S)iX¡(S) +h¡(s)/J¡(s) = [¡(s)

Its denominator is a polynomial of degree deg <f>(s) +deg iX¡(S). Ir we require


h¡(s)/<!>(s) to be strictly proper, there are deg <f>(s) number of free parameters in
hi(s); there are deg iX¡(S) + 1 number or free parameters in o:¡(s). Hence ir <I>(s)
and /J¡(s) are coprime, the roots of f¡(s) in (9-203) can be arbitrarily assigned by
proper choices of h¡(s) and o:;(s). The condition for asymptotic stability, tracking
-4 -27 /4J and disturbance rejection is that no root of <f>(s) is a transmission zero of (;(s)
+-2 914 or, because N(s) in (;(s) = N(s)O -1(S) is square, a root of det N(s). Because
~¡(s) is a factor of det N(s), we conclude that if no root of <1>(8) is a transmission
I

of these two imple­

compensators in the

the first implementa­

can be modified in

I is very close to the

19 canceled. Instead

, we may decouple it
(a)

se, the plant is to be

cations ate straight­ +

pfillg. In this su b­

:oupling, asymptotic

l(S) be a p xp non­ (b)

ithout iúvolving any Figure 9-34 Design of robus! sys!em.

528 LINEAR T1ME-INVARIANT COMPOSITE SYSTEMS D

zero of (;(s), then <fJ(s) and f3¡(s) are coprime, and the roots of f¡(s) can be arbi­ loop compensator.
trarilyassigned. From the assignment of f¡(s), we can compute h¡(s) and lXi(S).
We then design an input-output feedback system to decouple the plant as
diag {f31(S)/1X 1(S), f32(S)/1X 2(S), .. " f3 p (S)/lX p (S)}. The resulting system is asymp­ Now ir a solution T(s
totical1y stable and is decoupled. It will also track asymptotically the reference design can be accoml
input and reject plant-entered disturbances. In this design, if there are param­ configuration shown
eter perturbations in the plant and compensators (excluding the internal on the solution of (9­
model), the decoupling property will be destroyed. However, if the overall Let T¡(s) and Cm;1
system remains to be asymptotica11y stable, the property of asymptotic tracking be written, following
and disturbance rejection will be preserved. is a linear algebraic e
real rational functior
Example 3 necessary and sufficie
Consider the plant in Example 2 on page 524. We have f31 (s) =s - 1, f3 2(S) =S2 - 1,
deg IXI(S) = I and deg Cf.2(S) = 2. Suppose the plant is to be designcd to track
step inputs l/s and reject plant-entered disturbance ofthe form e2', then we have over lR(s). The soluti
In the model matchir
<fJ(s) = s(s - 2) proper rational matri
Since <fJ(s) and f3 ¡(s) are coprime, h¡(s) and IX¡(S) can be found to stabilize the sma11 as possible. Tb
to meet (9-204) is call
system in Figure 9-34(b). Rather arbitrarily, we choose 11(S) =(s +2)3. Then
SI72 and S210.
the solutions of
Before proceedin¡
<fJ(SP1(S) + f31 (s)h l(S) = 11 (s) = (s + 2)3 A(s) and B(s) are left
are h 1(s) =36s -8 and 1X 1(S) =s -28. Ir 12(S)=(S +2)4, the solutions of is unimod ular, that is,
is, unimodular. Folle
<fJ(S)1X 2(S) + f32(s)h 2(s) = 12(S) = (s + 2)4 matrix M(s) tb be roH'
are h 2(S)=1;2 S -16 and 1X2(S)=S2_1;2S_1;4. Next we replace (9-201) and if the R(s) in any fa
Theorem G-8', M(s) i
(9-202) by
in iC. We may also ex
O,(s) = [Cf.l(S)
O
O J=[S-28
1X2(S) O
O
S2_1~2S-~J
l A square polynomial
sum of a11 row degree:
if and only if its row-,
and (;(s)T(s) =diag {s
s-
1
-28'
s
2 s~:~ 1 124 1í
--3-S--3-
we define a q x n poi)
coefficient matrix is e
ii' iVK(s) is 01' ¡"uii row (¡
Once G(sr¡C(s) is implemented as an input-output feedback system inside the
U(s) such that U(s)1V
box in Figure 9-34(a), the design is comp1eted. This part of design is similar
irreducibility and co
to the one in Example 2 and will not be repeated. •
With this preliminary
Model matchillg. The design problcms discussed so far concern only with the We discuss first th
assignment of poles and denominator matrix; the numerator matrix is left un­ Ir (;(s) is square and n
only if (; -1(S)(;",(s) Ü
specified. In this subsection, we discuss the assignment of denominator matrix
~ore complicated. . L
as well as the numerator matrix 01', equivaJently, rhe entire overall transfer
G(s) and Gm(s). The I
'matrix. This problem i~ often referred to as the exact módel matching problem.
, Consider a plant with q x:p proper rational matrix (;(s). The desired mQdel
is assumed to have the q x r pr.oper rational matrix(;m(.~)' The problem is to
find a'configuration and co'mperrsators lor the plailt so that the resulting overall n If Gm(s)= 1, then the solut
reduces also to the minim
systcm has (;m(S) as its transfermatrix. We study theproblem by usingan open-
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 529

ts of f¡(s) can be arbi­ loop compensator. Let T(s) be a p x r open-Ioop compensator so that
)mpute h¡(5) and íX¡(5).
G(s)T(s) = G",(s) (9-204)
ecouple the plant as
ng system is asymp­ Now if a solution T(s) exists in (9-204) and is a proper rational matrix, then the
,totically the reference design can be accomplished by implementing T(s) in the input-output feedback
~n, if there are param­ configuration shown in Figure 9-31 or 9-33. Thus the design problem hinges
tcluding the internal on the solution of (9-204).
;)wever, if the overall Let Ti(s) and Gmi(s) be the ith column of T(s) and Gm(s). Then (9-204) can
)f asymptotic tracking be written, fol1owing (2-2), as G(s)T¡(s) = Gmi(s), i = 1,2, ... ,r. Each equation
is a linear algebraic equation studied in (2-36) with entries in lR(s), the field of
real rational functions. Hence Theorem 2-4 is directly applicable. Thus the
necessary and sufficient condition for the existence of a solution T(s) in (9-204) is
,)=s-1,P2(s)=s2-1, rank G(s) = rank [G(s) G",(s)J
be designed to track
form él, then we have over lR(s). The solution T(s) is general1y a rational matrix, proper or improper.
In the model matching problem, we are interested in only solutions which are
proper rational matrices. Furthermore, we require the degree of T(s) to be as
found to stabilize the small as possible. This problem of finding a proper T(s) with a minimal degree
11(5) =(s +2)3. Then to meet (9-204) is cal1ed the minimal design problem 22 in References S95, S125,
SI72 and S21 O.
Before proceeding, we digress to introduce some concepts. Recall that
A(s) and B(s) are left coprime if and only if their greatest common left divisor
he solutions of is unimodular, that is, the R(s) in any factorization [A(s) B(s)] = R(s)[ Á(s) B(s)J
is unimodular. Following this, we may define a q xn, with n ?q, polynomial
matrix M(s) to be row irreducible, or its column submatrices to be left coprime,
'e replace (9-201) and if the R(s) in any factorization M(s) = R(s)M(s) is unimodular. Fol1owing
Theorem G-8', M(s) is row irreducible if and only if rank M(s) = q for every s
in C. We may also extend the concept of row reducedness to nonsquare matrix.
A square polynomial matrix A(s) is row reduced if deg det A(s) is equal to the
sum or all row degrees or A(s). This definition implies that A(s) is row reduced
if and only if its row-degree-coeflicient matrix is nonsingular. Following this,
we define a q x n polynomial matrix M(s) to be row reduced if its row-degree­
coefficient matrix is of rank q or of ful1 row rank. Similar to Theorem G-l1,
lck system inside the ir M(s) is 01' fun row rank in iR (S), [hefe ex;::;,::; 8. unimodular polynoiHici< maÜiX
rt of design is similar U(s) such that U(s)M(s) is row reduced. We may similarly define column
I irreducibility and column reducedness for nonsquare polynomial matrices.
With this pre!iminary, we are ready to study the minimal design problem.
concern only with the We discuss first the condition for the solution T(s) in (9-204) to be proper.
ator matrix is left un­ Ir G(s) is square and nonsingular, the answer is very simple: T(s) is proper if and
f denominator matrix only ir G-1(S)G m (s) is proper. Ir G(s) is not square, the situation is slightly
ntire overal1 transfer ~ore complicated. Let 4J(s) be the least coml1)on denominator of al1 entries of
del matching problem. G(s) and Gm(s). The inultiplication of 4J(s) to (9-204) yields
l The desired model A(s)T(s)=B(s) (9-205)
l The pro blem is to
.t theresulting overall 22 Ir Gm(s) ~ I~ then the solution '-i(s) in (9-204; is a rig~t invers~of ~(sj.· Hencetheinverse problem
lem by usirig an open- reduces also to the minimum design problem. .
530 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS [

where A(s) = </J(S)G(S) and B(s) = </J(s)G", (S) are q x p and q x r polynomial does have a full ce
matrices. If G(s) and, consequently, A(s) are of ful1 row rank, there exists a algorithm.
unimodular polynomial matrix U(s) such that U(s)A(s) is row reduced. Consider
U(s)A(s)T(s)= U(s)B(s) (9-206) Theorem 9 -28
IrY(s) in (9-209) is a rr
We assert that if G(s) is of full row rank in lR(s), then T(s) is proper if and only if G-14', then NT(s)Di
b..¡(U(s)A(s)) :::O:b'i(U(s)BCs)) i= 1, 2, .. "q (9-207)
Proof
where b,.¡ denotes the ith row degree. Ir A(s) is square, this assertion is essenti­
Ir Y'¡'e has rank r, ther
ally Theorem G-IO. The proof for the nonsquare case is similar to the one of
hence N T(s)Di 1 (s) is
Theorem G-l Oand will be omitted.
The properness condition in (9-207) does not tell us how to find a T(s) with a T(s) =
minimum degree. In the following, we shall introduce such a method. The
method also gives the properness condition of T(s); hence the condition in IrY'¡'e has no ful1 COIUl
(9-207) is not really needed. Let T(s) = NT(s)Di 1 (s). Then (9-205) becomes algorithm, at least on
A(s)NT(s) = B(s)DT(s) or and Y/,,) = 0, where Y'
basis, Yhe has a fui! ce
[A(s) B(s)] [ - ~~i;~J = O (9-208)
jth column degree of

This is the polynomial equation studied in Equation (G-90); hence all discussion
in Section G-6 is directly applicable. Let (W, lR(s)) denote the right null space IfT(s) is proper, as s­
of (9-208). Its dimension is, following Theorem 2-5, equal to Hence ir Y¡'e has no
the theorem.
r~ p +r - rank [A(s) B(s)] = p + r - rank [G(s) Gm(S)]
In this theorem,
Now we may apply Theorem G-14' to solve (9-208). We form the generalized minimal basis is eSSé
resultant T k from the coefficient matrices of A(s) and B(s) as in Theorem G-14'. 9-33). The theorem i
We then search its linearly dependent columns in order from left to right by case. Ir r < r, clearl:
using the column searching algorithm. There will be exactly r primary depend­ If ;::::O:r, Y(s) and \'¡"
ent columns in T k • Let the (p + r) x;: polynomial matrix Y(s) be the solutions holds. In this case,
corresponding to these r primary dependent columns. Then Y(s) is, as in have increasing colur
Theorem G-14', column irreducible and column reduced, and is a minima! the 11rst 1" coiumns VI
polynomial basis of the right nuií space oí" (9-208). of (9-204). This com
Let Y he be the column-degree-coefficient matrix of Y(s). For convenience
r
of discussion, we assume = r and partition Yhe and Y(s) as Example 4
Find a 3 x 2 minima
- NT(S)] (9-209)
Y(s) = [ DT(s)

where Y/le and NT(s) are p x r and Y'hc and DT(s) are r x r matrices. " We note
that, because of the properness assumption of G(s) = A - l(s)B(s), the
[ - N'(s) D'(s)]' in Theorem G-14' always has the properties that bciN(S):::;;
be; D(s) and D(s) is column reduced orD he has a full column rank. In the minimal
design problem, A(s) in (9-208) is not necessarily square and A - 1 (s)B(s) may not Yhe is not computed 1
23"if
be defined; therefore there is no guarantee that DT/le, the coiumn-degree­ sorne j. However, ¡fral
coefficient matrix of DT(s) in (9-209), is of full columri rank. We note that Yhe elernentary colurnn ope
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 531
1d q x r polynomial does have a ful1 column rank as a consequence of the coiumn searching
Irank, there exists a algorithm.
w reduced. Consider
(9-206) Theorem 9-28
UY(s) in (9-209) is a minimal polynomial basis of (9-208) obtained as in Theorem
; proper if and only if
G-l4', then NT(s)D·¡ les) is proper if and only if Y¡'e has rank r.
(9-207)
Proof
s assertion is essenti­
U Y¡'e has rank r, then D'r(s) is column reduced and (jei N T(S) :S;(jei DT(s), for al1 i;
similar to the one of
hence NT(s)Di les) is proper. To show the converse, we consider
N to find a T(s) with a or (9-21 O}
uch a method. The
.1ce the condition in UY¡'e has no full column rank, there is, as a consequence ofthe column searching
hen (9-205) becomes algorithm, at least one column, say thejth column, such that óejNT(s) > (jej DT(s)
and Yhej = O, where Y¡'ej denotes the jth colu!!1n of Y¡'c- 23 §ince Ves) is a minimal
basis, V he has a ful1 column rank. Hence ifY hej =0, then Yhej=/=O. Let skj be the
(9-208) jth column degree of Ves). Consider the jth column equation of(9-21O):

T(s)DTj{s)S-k j = NTAs)S-k j
); hence al1 discussion
e the right null space UT(s) is proper, as s---> 00, we have T(ca )Y/,ej = O = Yhej =/= O. This is not possible.
al to Hence if Y/,e has no ful1 column rank, then T(s) is improper. This establishes
the theorem. Q.E.D.

In this theorem, the condition that Ylre is of ful1 column rank or Ves) is a
form the generalized minimal basis is essential; otherwise the theorem does not hold (see Problem
as in Theorem G-14'. r
9-33). The theorem is developed for the case = r. We discuss now the general
from left to right by r
case. U < r, c1early no solution T(s), proper or improper, exists in (9-204).
tly r primary depend­ If r? r, Ves) and Y¡'e are (p + r) x r matrices of rank r, and Theorem 9-28 still
Ves) be the solutions holds. In this case, in order to obtain a minimal solution, we arrange Ves) to
Then Ves) is, as in have increasing column degrees, that is, (jcl Y(s).::s; (je2 Y(s):S; ... :s;(jcf Ves). Then
d, and is a minimal the nrst r columns with a nonsigular \\e wil1 give the minimal proper so!ution
of (9-204). This compíetes the discussion oí' the minunai design proolelÚ.
s). For convenience
lS Example 4

(9-209)

"matrices. We note
(s) = A -l(s)B(s), the
.. lerties that ÓciN(s).::s;
Find a 3 x 2 minimal proper solution of

-s
s+3
-3s-7
s+3
l
rank. In the minimal
jA -:-.1 (s)B(s) may not
23 Ir Yh,· i~ not computed by· using the column searching algorithm, we may not have Yh,j = O, for
the column-degree­ can
sorne j. H·owever, ifrank Y,,, < r, bne of its cciluinn be transformed into a zero colurnn by
:k. We note that Ylle elementary coluinn opeiations.
532 LINEAR TIME"INVARIANT COMPOSITE SYSTEMS D

The multiplication of S(S + 1)(S + 3) and the substitution of T(s) = NT(s)Di 1 (S) to the three primar)
yields to compute

Clearly we have r = p + r - rank [G(s) Gm(s)] = 3 + 2 - 2 = 3. We form TI as


in Theorem G-14' and then apply the column searching algorithm to compute

o O O: O O:
o
3 O O: O O:
:3 - - - 0- -- -0- -: - 6---- -0- -:- - ó--- -0- - - - Ó- : - -0- - - - - Ó which yields immedi;
4 -6 -3 : 1 -7: 3 O O: O
-----------L---------~------------l--------
O r

y(,)~ 1
1 O -3 : 1 - 1 : 3 O O, O O
1 -8 -4 : 2 -10 :L 4 -6 -3 ,: 1
___________ L
-7_
O O -1 : 1 -1: 1 O -3 ,1 -1
O -2 -1 :
___________ L 1 - 3 : L 1 -8 -4 : 2 -10_ ~-----

• O O -1 1 -1
o ,
"., O -2 -1 1 -3
Y(s) has rank 3 for ev¡
reduced, for Y,,~ is 01
100 O O O O O O O basis of the right nul
rank 2; hence we ha"
O 1 -1/2 1/6 -7/6: 1/2 O O O O

O O 1 1 -1 1 O -3 1 -1

O O O 1 1 O O 9/2 -3/2 3/2

O O O O 1,0 O O O O

** 0- - -0- ---Ó- - - - -Ó- - - - -6--, --1- ----6---0- ---6-----6

'l O
O~
O
O
O
O
O
O
O
O
O
O
O
O
O
O
O
O
O
O
O
O
1
O
O
O
O
1 O
01

O O
O -1

1
-0 J'

-1

[x x x: x O: O x x: O O]
i i i This is a minimal pro
"Wherexdenotes thenonzero col~mn, and"O thezerócolumn. Thiscomputation
is very easy to carry out by hand. The notation : is explained in Appendix A. The numerator m
There are four linearly depe~dent cblumns in TI"; three of them are primary discussed in Section í
dependent columns as indicated by thearrows. Note that in Theorein G-14', the numerator matrix
"linearly dependent columns will appear only in B-rows (see dual of Theorem to achieve pole or den
G-13). In the minimal design problem,dependent columns mayhowever introduce any undesil
appear in B- as well as A-columns as shown in this example. Corresponding numerator matrices ir
DESIGN OF COMPENSATORS: INPUT-OUTPUT FEEDBACK SYSTEMS 533

to the three primary dependent columns, we use the dual formula of (A-II)
to compute
o O O
-1 O O
O I 1
103
1 O O
--------
O 1 O
=3. We form T l as O O O
Igorithm to compute O O I
O O I
O O O

which yields immediately


O s O O 1 O
-1 O O -1 O O
O 1 s+1 O O 1
Y(s) = - - --------- with Y/le = --------
1 O s+3 1 O 1
I O O 1 O O
Y(s) has rank 3 for every s in e, hence it is column irreducible. It is also column
reduced, for Y/le is of ful1 colu~n rank. Hence Y(;) is a minimal polynomial
basis of the right nul1 space. The first and third columns of Y/le yield a Ylle of
O O rank 2; hence we have
O O
O O
1 -1
-1 O
-3/2 3/2
O s+ 1
O O
--------
-------- I s+ 3
O O

1 O
O -1

-~J
O
I and
O -1

This is a minimal proper solution. I

n. This computation
lÍned in Appendix A. The numerator matrix of a plant cannot be affected by state feedback, as
of them are primary . discussed in Section 7-3, nor by output feedback; hence the only W'á,/ toaffect
:ü in Theorem G-l.4', . the numerator matrix is by direct cancel1ation. In the design of compeü.sators .
·See dual of The'orem to achieve pole or denominator matrix assign~e~t, we have beeÍl careful ilot to
lumns may however introduce any undesirablepole-zero cancel1ation. This is possible becausethe
lple. Corresponding numerator matrices in these designs are not specified. In the model matching,
534 LINEAR TIME-INVARIANT COMPOSITE SYSTEMS

because the numerator matrix is also specified, undesirable pole-zero cancella­ Then it can be showr
tion may be unavoidable. Therefore in the model matching, the choice of Reference S85, whe
Gm(s) must be very careful. Otherwise, even if we can find a proper T(s) and D-l(S)N(s); the metl
implement it in a well-posed input-output configuration, the design is still rings. The method I
unacceptable. S51,andS218. Wefir:
[Lo Mo : L l
9-8 Concluding Remarks Then a solution exisl
solution of the equati<
In this chapter, we studied a number of topics: characterization of composite ical comparisons of '
systems by their transfer matrices, stability of feedback systems, and the design In addition to Eq\
of compensators in the unity feedback systems and the plant input-output
feedback systems to achieve arbitrary pole placements, asymptotic tracking,
disturbance rejection, and decoupling. We also studied the implementation of also arises in the desig
open-loop compensators by using plant input-output feedback systems. from the Diophantil1E
All the design problems in this chapter hinge essentially on the solution of See References S88, S
the Diophantine equation: asymptotic tracking a
the internal model is c
L(s)O(s)+ M(s)N(s)= F(s) (9-211 a)
bypasses the equation
or O(s)L(s)+ N(s)M(s)= F(s) (9-211b)
this texto
Several methods are available to solve this equation. First, we may apply a In this chapter we
sequence of polynomial elementary matrices to transform [O'(s) N'(s)]' into a plant input-output fe
Hermite row form or a upper triangular form as in (G-33): compensator is not ~
N(s)O j 1(s)Nc(S). In t
U 11 (s) U 12(S)J [O(S)J = [R(S)J assignable, and the n:
[ Uds) Uds) N(s) O Dc(s) in N(s)Oj l(s)D c
Then a solution exists in (9-211a) if and only if R(s) is a right divisor of F(s). A completely concel1ed :
general solution L(s) and M(s) can then be obtained as in Problem G-14o See Dc(s) in the plant inI
References S34, S139, and S1890 Another method is to find an observable feedback configuratio
realization {A, B, C} of O-l(s)N(s) as in (6-131). We then use the operator [lD L(s), M(s), and Oc(s) il
defined in (6-225) to compute ITv(F(s)). Ifwe use the method in Problem G-15 Dc(s) in the latter, and
or any other method to compute In the design, the
fracUonal formo Tw
[I(s) == D(s)Q J. (s) -:-- f 1 {,~')
ll1e Ílrst method, we
with briFl(S) <briO(s), then we have ITv(F(s» = F1(s). We then express Fl(s) sequence of polynomi
in terms o[ the basis in (6-223) as one as in (G-33). The
In this case, we must '
Fl(s)=L(s)E 1
reduced one. In the
Then a solution L(s) and M(s) exists in (9-211 b) if and only if A -l(s)B(s). We fom
ofA(s) and B(s), and n
ránk[B AB ... An-1B] = rank [B AB ... An-1B El] left to right. From the
tainaright-coprime fr
or, equivakntly, El lies in the space spanned by the column of
reducéd; in fact, it is a
[B AB . 0'0 An-1B]. The solution of El =[B ABo" A"-1B]P is the
Once a right-cop
represent"ation.of M(s) in (9-211 b)with respect to the basis L(s), that is, M(s) =
, step'is to co"mpute th<
L(s)P. Next we compute
necessary because we
N(s)M(s) = D(s)Q2(s) + F 2(S) search linearly indepe
CONCLUDING REMARKS 535

e pole-zero cance\la­ Then it can be shown that L(s) in (9-211 b) is given by L(s) = Ql(S) - Q2(S). See
ching, the choice of Reference S85, where the method is developed only for strictly proper
Id a proper T(s) and D-l(S)N(s); the method, however, has been extended to linear systems over
n, the design is still rings. The method presented in this chapter fo\lows the line of References 18,
S51, and S218. We first translate (9-211) into a set oflinear algebraic equations as
[Lo Mo:L 1 M 1 :-·<L/II MmJSm=[F o F 1 .. , FU+/IIJ~F

Then a solution exists in (9-211) if and only if F is the row space of S",. The
solution of the equation yields immediately the required L(s) and M(s). N umer­
ization of composite ical comparisons of these three approaches are not available at presento
;tems, and the design In addition to Equation (9-204), the equation
: plant input-output L(s)D(s) + N(s)M(s) = F(s) (9-212)
asym ptotic tracking,
le implementation of also arises in the design of multivariable systems. This equation is quite different
dback systems. from the Diophantine equation, and its solutions are much more complicated.
Iy on the solution of See References S88, S 139 and sno. The equation is essential in the study of
asymptotic tracking and disturbance rejection (without robust). In our design,
the internal model is chosen as c/J -1(s)lp. This results in a robust design and also
(9-211 a)
bypasses the equation. Consequently, the equation in (9-212) does not arise in
(9-211b) this texto
irst, we may apply a In this chapter we studied two feedback configurations: unity feedback and
[D'(s) N'(s)]' into a plant input-output feedback. In the former, the denominator matrix of the
compensator is not assignable, and the resulting overall transfer matrix is
N(s)Djl(S)N c(s). In the latter, the denominator matrix ofthe compensatorsis
assignable, and the resulting overall transfer matrix is N(s)Dj 1(s)Dc(S). The
Dc(s) in N(s)Dj 1(s)Dc(S) can be used to generate the desired numerator or be
~ht divisor of F(s). A completely concelled by O f(S). The reason that we have freedom in assigning
Problem G-14. See Oc(s) in the plant input-output feedback configuration but not in the unity
) find an observable feedback configuration is very simple. There are three sets of parameters
use the operator IT D L(s), M(s), and Dc(s) in the former, but only two sets of parameters N,(s) and
lOd in Problem G-15 Dc(s) in the latter, and two sets of parameters are needed to achieve the designo
In the design, the transfer matrix G(s) must be expressed in a coprime
fractional formo Two such procedures are developed in Appendix G. In
Lhe nrsi method, we íirst find a noncoprime righi fraciion anci then use el
fe then express F 1(s) sequence of polynomial elementary operations to reduce it to a right coprime
one as in (G-33). The D(s) obtained by this process may not be column reduced.
In this case, we must find a unimodular matrix to transform D(s) to a column
reduced one. In the second method, we first find a noncoprime left fraction
A - 1 (s)B(s). We form a generalized resultant from the coefficient matrices
of A(s) andB(s), and then search its linearly independent columns in order from
left to right. From the primary linearly-dependent columns, we can readily ob­
tain a right-coprime fraction: Furthermore,the resulting O(s) is always column
)y the column of reduced; in fact, it is also row reduced and -inthe polynomial echelon formo
A/- 1 BJP is the
Once a right-coprime fraction G(s) = N(s)D - 1(.s:) is obtained, the next
sL(s)~that ¡s, M(s) =
step is to compute the -row index of G(s). In the scalar case, this step is un­
necessary because we have.v = n =deg D(s). In the multivariable case, we
search linearty ilidependent rows of Sm in order from top to bottom. Once all
536 LINEAR T1ME-INVARIANT COMPOSITE SYSTEMS

N rows of the last block row ofS m become Iinearly dependent, the mis the row 9-5 Are the feedback sys
indexo The total number of Iinearly independent N rows in Sv should be equal
to deg det D(s). This can be used to check the correctness of the computation.
-J----
+ I 1 1_
With the computed v, the chosen F will be in the row space of Sv- 1> and the - t L- -
sr, + I J f

l~
solutions of (9-211) yield immediately the required compensators.
The search of linearly dependent rows or columns of a matrix is basically
an ill-conditioned problem, beca use a zero row can be made to be nonzero by
introducing a very small number. Consequently, no matter what method,
singular value decomposition or Householder transformations with pivoting, is Figure P9-5
used, difficulties remain. In om design, the order of rows cannot be changed.
Hence the singular value decomposition or Householder transformations with 9-6 Is the feed back systel
pivoting cannot be used. However we may use Householder transformations
without pivoting and back substitution. These two methods are numerically
stable and can be used in the designo We emphasize that the numerical stability
of an algorithm and the ill- or well-condition of a problem are two distinct
properties. Since the search of dependent rows is an ill-conditioned probiem,
even though we use only numerically stable methods, difficulties may still
arise in the computation. 8180 stable? asymptotic
Ll¡(s)Ll 2 (s)det (1 + (;ds)G 2 (.

Problems
9-7 Given the plant 9(05)
9-1 Consider system so that the poles of

~. 05+2
and . g2(S)=-· 9-8 Given the plant 9(05)
05+3 degree 1 so that the poles of
Is the tandem connection of 9¡(S) followed by 92(S) completely characterized by its overall proper compensator C(s) c
transfer function? Verify your result by showing the controllability and observability of -1, -2, -3, and -4.
its composite dynamical equation.
9-9 Consider the plant gi
9-2 1s the feedback system with 9¡(S), given in Problem 9-1, in the forward path and
pensator in the llnity feedb
92(05) in the feedback path completely characterized by its overall transfer fllnction?
com pensator of dcgrec 1 in
9-3 Consider
:t_ll

:~:j
f) l s !l·10 Consider the íeedb<\!
iúnctlOn"l Let N(s)=N¡(s)
~J. N¡(s)N¡(s)/D¡(s), wíth deg 1
compensator P(S)fDc(s) an!
05+2
Dc(s) be Hurwitz? lf Dc(s)
Are their paralIeI and tandem [(;¡(s) folIowed by (;2(05)] connections controllable? obser­ the system be acceptable?
vable? completely characterized by their transfer matrices? wilI be acceptable? See Prc

9-4 ls the paralIel connection of

I 1
A

G¡(s) =
05+1

t
O and
A

G 2 (s) =

l
05+2
O

s
controlIable and observable? Note that (; ¡(s) and (;2(S) have the same set of poles.
O
1.
+l
J
Figure P9~10
PROBLEMS 537

ldent, the m is the row 9-5 Are the feedback systems shown in Figure P9-S BIBO stable? asymptotically stable?
; in Sv should be equal
,s of the computation.
,pace of Sv _ 1, and the
+
Ti~"(S-+-11
I I I ¡
1.,-- s h­
_s_+_2_ I 5(S
I
+ 1)

·ensators.
fa matrix is basically
lade to be nonzero by
matter what method,
ltions with pivoting, is Figure P9-S
",s cannot be changed.
9-6 Is the feedback system shown in Figure 9-11 with
: transformations with
lolder transformations 1 I

l I
thods are numerically
the nurnerical stability
S2 -1
G¡(s)= _1_ G,(s) =
-
s +2
1
s +3
1 J
,blem are two distinct
-conditioned problem, s- 1 s +1 s ~3
;, difficulties may still
BIBO stable? asymptotically stable? Use both the coprime fraction formula and
+ (;1(S)(;2(S)). Which one is simpler to use for this problem?
6¡(s)6 2(s)det(I

9-7 Given the plant g(s) = (s - I)/s(s - 2). Find a compensator in the unity feedback
system so that the poles of the resulting system are - 1, - 2, and - 3.

+2 9-8 Given the plant g(s)=(s2-1)/(s2'---3s+I). Find a proper compensator C(s)of


+3 degree 1 so that the poles of the unity feedback system are - 1, - 2, and - 3. Find a strictly
haracterized by its overall proper compensator e(s) of degree 2 so that the poles of the unity feedback system are
bility and observability of -1, -2, -3, and -4.

9-9 Consider the plant given in Problem 9-8. Find a set of lwo poles so thal the com­
in the forward path and pensator in the L1nity feedback system is a constant. Find a set of three poles so that (he
1 transfer fLlnction'J compensator of degree 1 in the uni(y feedback syslem is improper.

s +1l
9-10 Consider the feedback system shown in Figure P9-1O. What is iíS overall transrer
function'! Let N(s) = N ¡(s)N 2(s), where 1'/2(S) can be cancelied. Snow ihat, ¡Or any 9f(.I') =
;-7- 2j'

N[ (s)Nf(s)/Df(S), with deg DfeS) - deg (N 1 (s)Nf (s)) ~ deg D(s)- deg N(s), there exist proper
compensator P(s)/De(s) and Ne(S)/De(s) to achieve the designo If Df(s) is Hurwitz, will
s +2 De(s) be HLlrwitz? lf De(s) is not Hurwitz and if the system is implemented as shown, will
:tions eontrollable? obser­ the system be acceptable? Can you find a different but equivalent implementation which
will be aceeptable? See Problem 9-32 and Reference S34.

. \ PIs) ~.

ID;W~

1e same set of poles. Figure P9-10


538 LINEAR T1ME-INVARIANT COMPOSITE SYSTEMS

9-11 Consider 9-18 Consider the sys!l


5+1
~ 5(5-1)
G(5) = 1
é
f J
52 - 1

Find a compensator so that the poles of the overall unity feedback system are - 1, - 2 ± j,
Find a compensator in 11
and the rest chosen as - 2.

9-12 Repeat Problem 9-11 for


~
52~J
[ 5 +1
G(5)= - ­ Is the resulting Gf(5) cycL
5(5 - 1)

9-13 Which of the following matrices are cyclic? 9-19 Consider (;(5) = N(.

~ ~
---.!_
5+ 1
5-2 1 l ~ 5 +2
(5-1X5+1)
Let N(5)O-1(5)=O-l(S)Ñ
M(5) are left coprime, so ¡

r
Gl(s)= 1 G 2(5) = 1 9-20 Lel D-l(S)N(s) be
nomial ~atrix ifand only
5+3 (5 - 1)2 5- 1
malrix M(s) such lhat M(~

~
G,(5)=
.
l5-1~;5+1) (5-1~1(5+1)J ~
2
r:
G 4 (5)= O
9-21 Consider lhe feedba
h'lck syslem is asymptol
where aH poles of r(s) He ¡
5-1 5-1 coprime. Show thal
9-14 Show that ir all elements of G(5) have no poles in common, then G(5) is cyclic. Note e(s)=(I +
that repeated po\es in individual g¡j(5) are, however, permitted.
has no cIosed righl-half-pla

9-15 Find the set of 2 x 1 vectors t such that 6«;(5» of6«;(5)t), where

5 +1
~
G(5) =
5-1
-2
=!'J
s
2
f
s- 1 s

9-16 Show that ifthe rational matrix C(5) is cyclic, so is (:(5) = C(5) + K for any constant K.

9:22 ~onsider lhe unity fe


9-17 Consider the system in Example 3 on page 484; that is, O-l(s)N(s) be coprime frac
nomial matrices such that
2+ 1
H(s) with aH poles inside lh
~
G(5) = .
.[ O
5
2
-1 52
5
-1
52 +5 + 1
J
52 - 1 stabilizes lhe Unity feedback
Try G(s) =(5 + 0/5(5 +2) an,
Find a compensator in the unity feedback system so that the poles of the resulting system
consists of the ro<its of 52 + s+ 1 and the resi fram - 1. Compár~ your result with the one 9-23 Giveng(5)=(5-1)j5(.~
in the text. feedback systein lo have gf(S
PROBLEMS 539

9-18 Consider the system in Example 2 on page 476, that is,

G(s)
A = l~ ~ °l [52 0J-I[1° °5 0J
lOs
=
1
°°­ s
:k system are -1, -2 ±j, Find a compensator in the unity feedback system so that the resulting denominator is

lS + 1)2
Dj(s) = [

Is the resulting Gj(5) cyclic? Compare your result with the one in the text.

9-19 Consider G(s) = N(S)O-I(S)M(s), where N(s), O(s), and M(s) are polynomial matrices.
Let N(s)O - I(S) = O - 1 (s)N(s), where O(s) and Ñ[s) are left coprime. Show that if O(s) and
M(s) are left coprime, so are O(s) and Ñ(s)M(s).

9-20 Let O-I(s)N(s) be a left coprime fraction. Show that M(s)D-l(S)N(s) is a poly­
nomial matrix if and only if O(s) is a right divisor of M(s), that is; there exists a polynomial
matrix M(s) such that M(s) = M(s)O(s).

9-21 Consider the feedback system shown in Figure P9-21. It is assumed that the feed­
o'lck system is asymptotically stable. Let G(s) = O -1(s)N(s) and r(s) = D,~ l(s)N,(s)ro,
where all poles of r(s) He inside the closed rigbt-half s plane and D,.(.1) and N..(s) are left
coprime. Show that
e(s) =(1 +G(sW lr(s) = (O(s) + N(s)t ID(s)O,:-l(s)N..(s)ro
1, then G(s) is cyclic. Note
has no closed right-half-plane poles if and only if O,.(s) is a right divisor of O(s).

), where
~(S) ~(s)

figure PS-2'¡
:(s) + K for any constant K.
9-22 Consider the unity feedback system shown in Figure 9-16. Let G(s) = N(S)O-I(S) =
0-1(s)N(s) be coprime fractions of the plant transfer matrix. Let X(s) and Y(s) be poly­
nomial matrices such that X(s)D(s) + Y(s)N(s) = I. Show that for any ralional matrix
H(s) with all poles inside the open left-haH s plane, the compensator

C(s)= [X(s)- H(s)Ñ{sW1[Y(s)+ H(s) 0(5)1

stabilizes the unity feedback system. lfH(s) is proper, will Cls) be proper'? [An5wer: No.
TryG(s) ==(s +1)/s(s +2) and H(s)=O.J

lles of the resulting system


re your result with the one 9-23 Given 9(S) = (s -1)/s(s - 2) and 9j(S) = (s - 1)/(2s 2 +45 + 3), design an input-output
feedback system to have g¡(5) as its overall transfer function.
540 LINEAR TlME-INVARIANT COMPOSITE SYSTEMS

9-24 Consider the plant transfer matrix 9-32 Show that the tra
N(s)[De(s)D(s) + Ne(s)N(s
A
G(s) =
[sI
is nonsingular and can t
ditions, will there exist ¡:
Verify that the row index of G(s) is 2. Design an input-output feedback system to have design?

(s +1)2 O J
~

DJ(s) = [ O (s +1)3

as its denominator matrix. Choose De(s) as diag {s +2, s + 2}. ~ ILS;


11- ~
9-25 Repeat Problem 9-24 to have ~
_ [s
DJ(s) =
+1
O
Figure P9-32

as its denominator matrix. 9-33 Consider G(s) = N


proper if and only ir D(s) i~
9-26 Repeat Problem 9-24 to have
S(S +2) 9-34 Fínd a mínimal pr
DJ(s) = [ s +2J

[ ,~
s +2 s +2
as its denominator matrix. Is the system decoupled? Are all compensators proper? Is 1
the system well posed? ls there any unstable pole-zero cancellation?
s+2
9-27 Consider the plant given in Problem 9-24. Design a well-posed input-output
feedback system to decouple it without involving any unstable pole-zero cancellations. 9-35 Let Y(s) be a n x J
í = 1,. 2, ... , p, and let )
¡J.¡,
9-28 Consider Corollary 9-24 with G(s) strictly proper. Show that if Dcill¡(s) = De¡O(s) Consider
and OJ/¡e = D'K' that is, column degrees and column-degree-coefficient matrices of OJ(s)
and O(s) are all equal, then the compensator Co(s) = D; 1 (s)L(s) is strictly proper. The
compensator C 1 (s) = D; 1 (s)M(s) is generally proper. Show that
9-29 Considcr Equations (9-70) and (9-74). Show that if 111::::: v- 1, for any real k¡, í =
1, 2,. ,eL, the following
o •

[D co Neo: '0'; Dc", Nc",]+[k 1 k 2 o •• k.] [ -80 A o :"': -8", A",] if and only ir V(s) is colu
Rererence S95 and can be
meets (9-74), where (;(=(111+ l)q-n and [-Ro Ao : ... : -B", A",] is a basis of the ininin12..l. 1fin" t~t ,~¡ =
(J.-dimensional kIt ¡"¡uii space o[S", computcQ as in Appendix G. ThíS ,s;:, ¡),u<c¡,o,CIL¡ ''''~L'G" show the equaiity, we she
of all solutions of (9-74).
relation Yd = Y/KCX, where:
9-30 Consider G(s) = N(s)D - I(S), T(s) = D r- 1 Nr\s), and
Go(s) = G(s)T(s) = N(s)[O,(s)D(sW 1 N,(s)

Show that if Dr(s)D(s) and N,(s) are left coprime, any feedback implementation of Go(s)
reduces es'sentially to an open-loop system.

9-31 Consider

G(s)=
A

. I
[O I J[S2 + 1
s +1 1
Design a decoupled system to track step reference inputs and reject plant entered distur­
bance of the type ero . .
PROBLEMS 541

9-32 Show that lhe transfer matrix of the feedback system in Figure P9-32 is GJ(s)=
N(s)[D,(s)D(s) + N,(s)N(sW 1 K(s). Let G(s) = N(s)D- '(.1')= Nl(S)N 2 (s)D- 1 (s), where N 2 (s)
is nonsingular and can be cancelled. Given GJ(s) = NI (s)D j 1 (s)NJ(s), under what con­
ditions, will there exist proper compensators D,- 1 (s)K(s) and D,-l(S)Nc(s) to achieve the
design?
'eedback system to have

Figure P9-32

9-33 Consider G(s) = N(s)D-l(S) with óciN(s)'::;ÓciD(S), for aH i. Is it true that G(s) is
proper ir and only ir D(s) is column reduced? Hnot, find a counterexample.

9-34 Find a minimal proper solution of

[ s~ ::~] 1 [1.1'+.1'2::+ 3)
2(,+1)
(s-I~;s+3)
J

compensators proper? Is T(s) =


ation?
s .1'(.1' - 1)
- O -­
.1'+ 2 (.1'+ 2)2 .1'+2
1 well-posed input-output
pole-zero cancellations. 9-35 Let Vis) be a 11 x p polynomial matrix or rull column rank with column degrees
jJ.¡, i = 1;2, ... , p, and lel x(s) = [Xl (s) X2(S) ... xp(s)]' be any p x I polynomial vector.
)w thal ir L\J> J(i) = bciD(s) Consider
Jefficient matrices of O¡is)
y(s) = Y(s)x(s)
.(5) is strictly proper. The
Show that

2': v - 1, rOl' any real k¡, i = degy(s)= max {¡L¡+degx¡(sj}


i:x,(s) FO
if and only if Y(s) is column reduced. This is called tbe prediclable degree properly in
A o : .. ' : - B", A",J
Reference S95 and can be used to show the degrees of a minimal polynomial basis to be
B", A",J is a basis of tbe 111 ininla.l. Hin!: Let d =:: lllax [p; + deg .':f{S) ~ ~~~h r->n ,..._1~? ri~~1 . . "le ha'J·~ !:t~g ~'/(,r;) :.::; J 'T,~
This is a paramelerizalion show the equality, we show the y,¡ in y(s)=YO+Y1S+'" +y,¡s'¡ is nonzero by using the
relation Y,¡ = Y,,,a, where -,¡(s) = aii -"¡ + ... and a = [a 1 a2 ... a p ]'.

,(s)

:k implementation of Go(s)

reject plant entered distur-


A-1 Gaussian E

Consider the n x m m:

A
First we assume a 11 =/=
Elementary Transformations the ith row, for i =o 2, ~

Consider a matrix A with e1ements in IR or iC. The following operations on A


where e¡1 = -ail/all.
are called the elementary operations or transformations: (1) multiplying a row or
matrixK I is the prodl
a column by a nonzero real or complex number, (2) interchanging two rows or
EJ. If a~ z is different 1
two columns, and (3) adding the product of one row or column and a number to
row of KIA and -a}z/,
another row or column. These transformations can be achieved by using the
elementary matrices, for n = 5, of the forms 1 O
O 1
1 O O O O 1 O O O O 1 O O O O KzKIA = O e 3Z
O 1 O O O O O O O 1 O 1 O O O
EI = O O 1 O O Ez = O O 1 O O EJ = O O 1 O O
O O O e O O O O 1 O O d O L O
where e¡Z = -atz/a~z a
O O O O 1 O 1 O O O O O O O 1 can be transformed int<
(A-1 )
x x x
where e and d are real or complex numbers and e =/=0. Elementary matrices are IX
.0 x x X
squ.o.re an.d nODsingular. Their inverses Pcre I
\) V /'o )\

1 O O O O 1 O O O O
O 1 O O O O 1 O O O O O O x
E-l-
1 - O O 1 O O E21 = Ez E-l-
J - O O 1 O O O O O O
O O O l/e O O -d O 1 O
O O O O 1 O, O O O 1 O O O O
(A-2)
where X denotes possib
They are again elementary matrices. The premultiplicationof E¡ on A operates
called the gaussian elim
on the rows of A, whereas the postmultiplication of E¡ on· A operates on thé matrices. 1
columns· of A. For example, EzA interchanges the second and flfth rows;
EJA adds the product of the second row and d to the fouáh row of A; whereas
AE J adds the product of thefourth column and d to the second column ofA. 1 AII diagonal elemenls in (A­
by llsillg additional column
542
GAUSSIAN ELIMINATION 543

A-1 Gaussian Elimination

Consider the n x m matrix


a lz a t3

A= ~Zl [a" a zz a Z3
a'-l
a Zm
(A-3)

anl a nz a n3 a nm
First we assume all ,p O. We add the product of the fmt row and (-a¡JÍall) to
s the ith row, for i = 2,3, ... , n; then we have
o O
1 O
°l
O
a
O
ll
:-~?- -aa- F --. ~ ~ -_aa-:
I 22 23 2m
T
­

~1J A=0O j ~1z a13


O 1 (A-4)
a1m
O O : a~z
1
a~3 a~m
Dwing operations on A whereeil = -ail/all, i=2, 3, ... ,n and aIj=aij +eilalj' Note that the n x n
1)multiplyinga rowor
matrixK I is the product of n -1 number of elementary matrices of the form of
rchanging two rows or E 3 . lf a1z is different fram zero, then the addition of the product of the second
)Iumn and a number to
row of KIA and -alz/a1z to the ith row, i = 3,4, ... , n, yields
. achieved by using the
1 O O O a ll al2 a l 3 ... Qlm
O 1 O O O a12 ab . .. a1m
'-----------2-­
1 O O O O KzK1A= O e 32 1 O KIA= O O • aZ 33
. .. a 3'11 (A-5)

··, a~3
I

O 1 O O O
E. 3 = O O 1 O O
O en 1 O 1 O O 1
1
a;m
O d O 1 O

where ei 2 = -aIz/aiz and at = a¿ +e¡zaL. Proceeding in this manner, matrix A


O O O O 1 can be transformed into a matrix whose elements below the diagonal are zero as
(A-1)

lementary matrices are


!
I~;0
x x
x x
O X
x x
x x
X. X
:l
X
1
fx
'0
)(

X.
x
;(
x
f..
x xl
j,
"
O O O O O O x x x x
or
1 O O O O O O x x x
O 1 O O O O O O O O O O O x x x
-d O 1 O O O O O x x
O O O 1 O O O O O O
(A-6)
(A-2)
where x denotes possible nonzero elements. This process of transformations is
ion ofEjon Aoperates cal1ed the·gaussian eliminaríon. The matrices in (A-6) are upper right triangular
on A operates on th<; matrices. 1
econd and fifth rows; .
Jrth row of A; whereas
1 AII diagonal elements in (A-6) are nOt nccessarily nonzcro. Il is possiblc lO makc lhern nonzcro
e second column of A.
byusing additional coturnn operations. Sce also Equation (G-28).
544 ELEMENTARY TRANSFORMATIONS

This process will fail if any of a¡¡, ah aj3' or other diagonal element is zero. with x'x = 1. where x
Even if all of them are nonzero, the process may still encounter some com­ (1 - 2xx')' = 1 - 2xx' =
putational difficu\ty. For example, if a¡¡ is very small in magnitude, then
(- a¡da¡ ¡) = eil will be a very large number. lf there are errors in representing H'H=
([ Ij, these errors will be greatly amplified by e¡ ¡ in the operations. F urthermore,
we ha ve H 1 = H'; hl
these errors may propagate and be reamplified at \ater stages. Consequently,
singular values of H "
the final resu lt may be overwhelmed by errors. Because of this phenomenon, the
ratio of the largest sin
gaussian elimination is said to be numerically unstable.
the use of Householde
To overcome this difficulty, before carrying out the elimination. we first
of a problem lsee Prol
search the element with the largest absolute value 2 in the first column of A,
An important pro
and then interchange this row with the f¡rst row by using a matrix of the form of
two vectors a and b 1
E 2 . We then carry out the elimination for the first column. Before carrying
exists a Householder t
out the elimination for the second column, we search the largest element in
x = la - b)flla - b11 2 , th
magnitude in the second column of K¡A excluding the first element, and bring
it to the a~2 position. This process of elimination is called the gaussian elimina­
lion with parlial piVOl ing. In this process, we have Ieijl.:; 1, and the errors will
not be amplified. Hence this process is more stable numerically than the
gaussian elimination. By using partial pivoting, we conclude that every matrix Since a'a = b'b by aSSl
A can be transformed into the form in (A-6) by using a sequence of elementary is invariant under trar
row operations.
The stability of the gaussian elimination can be further improved by using
complete pivoting. In this case, we must use elementary row operations as well
as elementary co\umn operations. We first search the largest e\ement in magni­
tude among all elements·of A. We then move the element to the f¡rst column by This establishes. the ·a
an elementary column operation and then to the first row by an elementary row K¡ and K 2 in (AA) ane
operation. We then carry out the elimination of the first column of the trans­ Consider the n x m
formed A. After this elimination, we repeat the process for the remaining of the first colllmn al
matrix. This process is called the gaussian elimination with complete pivoting. there exists a Househ·
This process is more stab\e numerically than the one with partial pivoting; how­ of the right-hand-side I
ever, it is more costly becallse the search ofthe largest element in magnitllde is a row of H ¡ A and repea
time-consuming process. Hence complete pivoting is not used as often as partial matrix. Proceeding i:
pivoting. According to Reference Sl38, from the point of view of overal1 per­ sequence of Househol(
formance which ineludes efficiency, accuracy. reliability. generality. and ease Di The numerical sta!:;
iíiUUilion or columns.
use, the gaussian elimination with partiai pivoting is satisfactory for most
general matrices. we permute the columi
apply the Householde
process, the diagonal e
*A-2 Householder Transformation order of magnitude.
pivoling.
The numerical stability of the elimination can be further improved by using the We note that Hom
Householder trarisformation. Let 1 be a unít matrix. A HOLlseholder trans­ K¡ and K 2 in (A-4)anc
formation is a ~quare matrix of the form 3 after the application of
H=I -'--2xX:

2 For conciseness: we call such elernent the largest elernent in rnagnitude.

'AIl argurnents still hold ir H is defined as H = [ - 2xx*, where x* is ¡he cornplex conjugate transpose 4Thesign rnay bechosen to b,
or x. In this case, wehave H - 1 = H* and H ís unitary. tion (see Problern A-S), we I
HOUSEHOLDER TRANSFORMAnON 545

agonal element is zero. with x'x = 1, where x is a column vector and x' is its transpose. Clearly H' =
encounter sorne com­ II - 2xx')' = 1- 2xx' = H; hence H is symmetric. Because, by using x'x = 1,
LlI in magnitude, then
H'H = (1 - 2xx')(I - 2xx') = 1 - 4x:x' + 4xx'xx' = 1
: errors in representing
rations. Furthermore, we have H-' =H'; hence H is orthogonal (see Appendix E). Because all the
stages. Consequently, singular values of H are 1 (Problem E-9), its condition number, defined as the
,fthis phenomenon, the ratio of the largest singular value and the smallest singular value, is 1. Hence
the use of Householder transformations will not impair the numerical property
e elimination, we first of a problem (see Problem E-lO).
the first colurnn of A, An important property of Householder transformations is that given any
a matrix of the form of two vectors a and b of equal euclidean norms, that is, Ila\l~ =a'a = b'b, there
umn. Before carrying exists a Householder transformation H such that Ha = b. 1ndeed, if we choose
the largest element in x =(a - b)flla - bll z, then we have
¡rst element, and bring
:d the gaussian elimina­
:;; 1, and the errors will
Ha = [1 _ 2(a - b)(a - b)'] a =a _ 2(a - b)[(a - b)'a]
Ila - bll~ a'a -a'b - b'a +b'b
numerically than the
:\ude that every matrix Since a'a = b'b by assumption and a'b = b'a for this is a scalar quantity which
;equence of elementary is invariant under transposition, we have

her improved by using 2(a'a - b'a)


Ha =a - ·-----(a -b)= b
row operations as well . 2(a'a - b'a)
~gest element in magni­
This establishes the assertion. Now we use this property to show that the
,t to the first column by .
,¡ by an e\ementary row
K, and K z in (A-4) and (A-S) can be chosen as Householder transformations.
Consider the n x m matrix A given in (A-3). First we compute the norm (J
st column of the trans­
of the first column a, of A. We choose b[ as [±(J O O .. , 0]'.4 Then
:ess for the remaining
with complete pivoting. there exists a Householder transformation H, such that H[A is in the form
ofthe right-hand-side matrix in (AA). Next we delete the first column and first
1 partía\ pivoting; how­
row of H,A and repeat the process for the first column of the remaining sub­
:ment in magnitude is a
matrix. Proceeding in this manner, the matrix A can be transformed by a
t used as orten as partial
of view of overall per­ sequence of Householder transformations into the form shown in (A-6).
The numerical stability of this process can be further improved by the per­
generality. and ease of
muiation of columns. First we compute the i10rms of aii coiurnns of A. NeXL
. satisfactory for most
we permute the column with the largest norm with the first column. We then
apply the Householder transformation to carry out the eliminatíon. By this
process, the diagonal elements of the resulting matrix will be in the decreasing
order of magnitude. We call this process Householder lransformations wilh
pivoting.
improved by using the We note that Householder transfor.mations are not triangular matrices as
A Householder trans- K¡ and K z in (A-4) and (A-S). Hence the rows of A are completely scrambled
after the application of Householder tfansformations.

Ide.

4TI1esign may be chosen lo beequal lO -sign (/11 10 reduce lhe roundoff errors. In lhe QR facloriza­
complex conjugale lranspose

lion (see Problem A-S), we muSl choose howeverlhe positive signo


546 ELEMENTARY TRANSFORMATlONS

If a matrix is transformed into the form in (A-6), the rank of the matrix is is no nonzero eIernen
equal to the number of nonzero rows in (A-6). This is a simple way of computing to the next row. If""
the rank of a matrix. 5

where K ~ KIl-¡K,,_
A-3 Row-Searching AIgorithm 6 1 on its diagonal. so i:
[fthe;th rowof Á is a
In our application, it is often required to search the linearly independent rows of on its previous rows.
matrix A in order from top to bottom. By this, we mean that we first check
whether or not the first row of A is nonzero, or equivaIently, linearly independent rb j ¡
by itseIf. We then check whether or not the second row is linearly independent
with bjj = Lis just th
of the [¡rst row. In the kth step, we check whether or not the kth row of A is
The matrix K can
Iinearly independent of its previous k - 1 rows. If a row is Iinearly dependent
11 - 1. This is comp[
onits previous rows, the row is to be eliminated from subsequent consideration.
computing any row o
Furthermore, we may want to find the coefficients of the linear combination of of
the dependent row.
Let the (i, j)th element of the n x m matrix A be denoted by aij' Let a ¡k
be any nonzero eIement in the first row af A. This element will be called the
pivot element ar simply the pivoto Let K¡ be of the form shown in (A-4) with
ei¡ = - aik/alk, i = 2, 3, ... , n. Then the kth coIumn, except the first element, of
K¡A = (a¡)) IS a zero column, where aL
=aij +ei¡a¡j' Let aL be any nonzero
element in the second row of K¡A. Let K z be of the form shown in (A-5)
with eiZ = - aMaL. Then the jth column, except the firsl two elements, of To compute thejth re
K z K1A = (alj) is a zero column, where alj =ai) + eizaL. In this process, if there jth row of K under i1

o O

e21 1 O

SThe computation or the rank of a matrix is a difficult problem. For example, the matrix ----
e 3t : e 32 : 1

AS=\: -; =:
O O O
=:
1
-'1
-\
-1
-1
¡
fjl 1
1
('j2

bj2
1
1

: ..
ejJ

~j: _ _ _
_

O

O o o
j[ l1en il is straightiorw
is c1early or rank 5 and nonsingular. However, we have
4 bjj =
2 -1 2 -4 1
1] [2-
Asx ~ As 2- 2 = 2- 4] =2- 4 [1] 1 ~ 2- 4 e
[
2- 3
2- 4
2- 4
2- 4
[

1
b jk =

Thus,"iLn in A n is"very large, there exists a nonzero x such that A.x = 2 1 -'e--+O and A. is nearly
singular. This irtformation cannot be detected from the determinant of A. (which is equal to 1
1
ror all nI, norfrom Its eigenvalues [which are all equa\ to (see Problem 2,23)],
nor from the form
of A.. However, if we compute the singular valuesof A,,(see the singular value decomposition in
Appendix E}, tÍlen the smallest singular value can be shown to behave as 2-'; for large 11 (see
Problem E-13), and the rank degeneracy orAn can therefore be detected. Thus the singular value
decomposition is considered the most reliable methodof compúting the rankof a matrix. 7 Ki cannol be chosen as He
6This algorithm is a simp[¡fied version of the one in Reference S2Z r()WSof A.
ROW-SEARCHING ALGORITHM 547

~rank of the matrix is is no nonzero element in a row, we assign K¡ as a unít matrix and then proceed
7
np1e wayof computing to the next row. If we carry this process to the last row of A, then we have
Kn-¡K n- z -" KzKIA~KA=A (A-7)

where K ~ KII _ ¡K II - z '" K 1 • Since every K¡ is a lower triangular matrix with


Ion its diagonal. so is K. The number of nonzero rows in A gives the rank of A.
Ifthejth row of Á is a zerO row vector, then thejth row of A is linearly dependent
'ly independent rows of on its previous rows. F urthermore, the coefficients of combination
~an that we fIrst check
,y, linearly independent (A-S)
is linearly independent with b jj = L is just thejth row of K.
IOt the kth row of A is The matrix K can be computed by direct multiplications of K¡, i = L 2..
'" is linearIy dependent f1 - 1. This is complicated. We introduce in the following a recurs.ive way of
Iseq uent consideration. computing any row of K. First we store the ith column of K in the ith column
: linear combination of of

:noted by aij' Let alk o O O


ment will be caBed the eZI 1 O O
m shown in (A-4) with F= e31 e 3Z I O (A-9)
;:pt the first element, of
~et aL
be any nonzero
enl enz e n3
~ form shown in (A-S)
first two e1ements, of To compute the jth row of K. we take the firstj rows of F and then arrange the
In this process, if there jth rúw of K undet it as

O O O
e Z! I O O

~xample, ¡he matrix


e 31 : e32 : 1 O
First.i rows of F (A-10)

ej l ' ejZ : ej3 l'j(j-l) O O

'- bj -z-:- b;;------b;j-_-I~ --b~;:


I }
O O ~ith row of K

Then it is straightforward lo verify tha t


b jj = 1.

'~e

=2' -ne ..... O and A" is nearly


nI of A. (which is equal lO I
j .

L bj~epk
¡ e1k+ I)kj
bjj ] ~(k+Z)k

ejk

k = j - 1, j - 2, . . . , 1
(A-11 )

p;k-t 1
em 2-23)], fiar from ¡he form
guIar value decomposilion in
have as 2-', for large n (see
ted. Thus ¡he singular value
: ¡he rank of a níalrix: 7 .K; eannOI be ehosen as Householder lransrormalíons beeause lhey will seramble ¡he order br the
T()WS of A.
548 ELEMENTARY TRANSfORMATIONS

We see that bjk is just the inner product of the vector on its right-hand side and
the vector above it as shown in (A-10)8 Hence the coefficients of combination
in (A-8) or, equivalently, thejth row of K can be readily computed by this simple
procedure.
In the application of this algorithm in this text, the information of the entire
K = K II - 1 . . . K 1 is never needed. We need only a few rows of K. Therefore
it is better to store K¡ in F as in (A-9) and write (A-7) as
The last row of A is
to write
(A-12)
1
Whenever a row of K is needed we then use the procedure in (A-lO) and (A-11)
to compute that row of K from F. -2 1
F~A= -4 -1
2 -1
Example 1
1 -2
Find the linearly independent rows of
Note that the ith colu

\-1 -1 (~: -2 1 row of A is a zero ro


two rows. Similarly

A~H
-1 4 -2 4
four rows.
-3 8 -6 6
To find the coeffil
1 -4 10 1 by using (A-11), from
L7 1 -2 10 4

We choose the (1, 3) element as the pivot and compute

r -~
K1A = -4 O 1
1
A=
~:~=: - ~ :~: -~ ~
3 1 O 2 2 ~Al Hence we have

l
2 O O 1
10001
3
6
-1
O
O 6
085
3

Except the first element, the first column of K 1 is the third column of A divided
or
where a¡ is the ith rov
by - 2. We next chaase the (2, 1) element of A\ as the piVOl i:"w.d compute

1 -1 -1 :1~ -2 1
O 1 :):~ 1 O 2 2
KzA 1 = O -1 1 A1 = O O O O O ~Az
O -1 O 1 O -2 O 4 1
O -2 O O O -2 O 4 1
Since the third row of A z is a zt:ro row, we set K 3 = 1 and proceed to the next row.

The pivot element of the fourth row of A z is chosen as shown. We compute Hence we have .

ar

8This presentation was suggested by Proressor J. W. Wang.


ROW-SEARCHING ALGORITHM 549

ts right-hand side and


1 -1 -1 ~A; -2 1
lcients of combination
,mputed by this simple O 1 (~) 1 O 2 2
K4 K 3 Az = K4Az = O O 1 Az = O O O O O ~A
'ormation of the entire O O O 1 O -2 O @ 1
~ows of K. Therefore O O 0-1 O O O O O
The last row of Á is a zero row, and the search is completed. We use (A-12)
(A-12 ) to write

1 -1 -1 2 -2
~e in (A-lO) and (A-ll) 1
-2 1 3 1 O 2 2
FtA= -4 -1 1 iA = O O O O O ~A
2 -1 O 1 O -2 O 4 1
1 -2 O -1 O O O O O
Note that the ith column of F is the ith column of K¡, i = 1,2,3,4. Since the third
row of Á is a zero row, the third row of A is linearly dependent on its previous
two rows. Similarly, the fifth row of A is linearly dependent on its previous
four rows.
To find the coefficients of combination for the third row of A, we compute,
by using (A-ll), from the first three rows of Fas

~]
O
O
-2 1 O
,- - -; 1- ",-- -_:. - - - : ....-1

2 2 '-211:-1} ~1~'
I 11"'- - " ...... I
--- ---------
2 2 f'...A¡ Hence we have
6 3 [-2 -1 O OJA=O
8 5 or -23¡ -az +a 3 =0 (A-"i3)
d column of A divided where a¡ is the ith row of A.
ivot and compute To find the coefficients al cambination for the fifth í"OW 'Ji A.. w" r;i)nCr.)l)f'~
-2 1 1
2 2 -2 1
O O ~Az -4 -1
4 1 2 -1 ;-0-; 1
, ,
4 1 1 -2 :_0_: -1
roceed to t he next row. -1 O r ---f:
:~-=-
-------­
own. We compute Hence we have
[1 - 1 O - liJA = O
or a¡ -3 z +03 3 -'a 4 +a 5 =0 (A-14)
550 ELEMENTARY TRANSfORMA TIONS

This algorithm of searching linearly independent rows of A in the order of To find the coefficien
the first row, the second row, and so forth will be referred to as the row-searching
algorithm. In this algorithm, if the ith row is linearly dependent, then the ith
row will not contribute to the linear combination for the jth linear dependent which implies
row for j > i. For example, since a 3 is linearly dependent, its coefficient is zero
in (A-I4); hence as is expressed as a linear combination of its previous linearly
independent rows. In general, in (A-8), if the ith row, with i <j, of A is linearly
dependent, then bji =0. Indeed, if the ith row is linearly dependent, then the and
ith row of A is a zero row, and the K¡ in (A-12) is a unit matrix. Consequently,
all elements below e¡i in the ith column of F are zero. That is, epi = O, for To find the coefficient
p = i + 1, i +2, ... ,j. Hence from (A-ll), we have bji =0. Because of this
property, the b jk computed in this algorithm are unique. Without bji=O, the
coefficients of combination may not be unique. For eX?'>lple, the addition of which im plies
(A-13) and (A-14) yields
t
-1 1JA=O

which does not have b S3 =0, and is a different combination. The property and
=0 if the ith row is dependent is essential in establishing Theorem G-14
b ji
in Appendix G. We see that the b ij a
A remark is in order regarding the numerical stability of the row-searching stable method (see R¡
algorithm. If the pivot element of each row is chosen as the leftmost nonzero formed iuto the forro
element, then this algorithm reduces essential1y to the gaussian elimination easily obtained. I n ~
(without any pivoting). Though the pivot is chosen as the largest element in linearly dependent ro
magnitude of each row, the pivot is no"t necessarily the larges't element in magni­ be identical to the 01
tude of its column. Hence the row-searching algorithm by choosing the largest this method of searcl
element in magnitude as the pivot is not exactly equivalent to the gaussian searching algorithm, 1
elimination with partial pivoting. Consequently, there is no guarantee that the The row searchinl
row-searching algorithm will be numerical1y stable. for the following two
In the search of linearly independent rows of A, it is essential not to alter the not complicate or ob~
order of rows. The column positions, however, can be arbitrarily altered. can be easily carricd o
Hence we may apply to the columns of A the gaussian elimination with partial problems by hand, or
pivoting or the Householder transformations to transform A into the form fidence in using a digi
a¡¡ O O O O' chosen for ihe conver
element with value +
aZ¡ azz O O O
zero elements as the p
AL¡L z '" ~ a 31 a 32 O O O ~A (A-15 )
is for pedagogical reas
a 41 a4Z a 43 O O use the method discus
a S¡ asz aS3 O O

where Li are elementary matrices or the Householder transformations. 9 It is * A-4 Hessenber


assumed that al l 4=0, azz 4=0; and a43 4=0. Clearly (A-15) implies that the first,
second, and fourth rows of A are linearly independentof their previous rows Every square matrix
and the third and fifth rows of A are linearly dependent on their previous rows. similarity transformal
values and eigenvectol
we shal1 use elemcnta
9 Householder transformations with pivoting {permutation of rows), however, are not permitted. a matrix into a speci
HESSENBERG fORM 551

s of A in the order of To find the coefficients of linear combination for the third row, we solve
o as the row-searching o OJA =0
:pendent, then the ith
: jth linear dependent which implies
, its coefficient is zero _ -a32
If its previous linearly b 32 ----­
a22
h i <j, of A is linearly
{ dependent, then the _ -(031 +b 32 a21)
and b 31 ­
atrix. Consequently, all
That is, epi=O, for To find the coefficients of linear combination for the fifth row, we solve
=0. Because of this
Without bji=O, the (A-16)
·..lple, the addition of which implies

lJA=O

ation. The property and


shing Theorem G-14
We see that the bij are obtained by back substitution. This is a numerically
of the row-searching stable method (see Reference S212). Hence we conclude that once A is trans­
the leftmost nonzero formed into the form in (A-lS), the coefficients of linear combination can be
gaussian elimination easily obtained. In solving these coefficients, the elements corresponding to
:he largest. element in linearly dependent rows are set to zero as in (A-16). In so doing, the result will
~est element in magni­ be identical to the one obtained by the row-searching algorithm. Although
y choosing the largest this method of searching linearly dependent rows is less direct than the row
alent to the gaussian searching algorithm, the method is stable numerically.
no guarantee that the The row searching algorithm, however, will be used exclusively in this text
for the following two reasons. First, the concept is very simple. Its use will
;ential nol to alter the not complicate or obscure the basic issue of the problem. Second, the method
le arbitrarily altered. can be easily earried out by hand for simple problems. After solving one or lwo
mination with partial problems by hand, one would understand the material better and gain a con­
)rm A into the form fidence in llsing a digital computer. For hand ea1clllation, the pivot should be
chosen rOl' [he convenienee 01 computation. For exampie, we may choose ene
element with value + 1 or the element whose column has the largest number of
zero elements as the pivoto In summary, the use of the row searching algorithm
~A (A-15)
is for pedagogical reasons. In actual digital eomputer computation, one should
use the method discussed in (A-l S) or other method which is numerically stable.

lnsformations. 9 It is *A-4 Hessenberg Form


I implies that the first,
.f their previous roWs Every square matrix can be transformed into aJordan canonical form by a .
[l their previous rows.
similarity transformation. However, this requites the computation of eigen-.
values and eigenvectors and is anumerically 1lI1stabh: problein~ In the followfng,
we shall use elementary matrices inthe similarity transformation to transform
vever, are not permitted. a matrix into a special form, called the Hessenberg formo This form is very
552 ELEMENTARY TRANSFORMATIONS

important in computer computation and its generalized form will be used in Problems
Section 5-8.
Let A be an n x n matrix. Consider A-1 Use the row searcl

1 :0 O O ...
__ J _ _ _ _ _ _ _ _ _ _ _ _ _ _O X X X X X

O: X X X X X

O: O X X X X
PIA~ I
P ll A= Use the círc!ed elements,
O: O X X X X coefficients of its linear Co
,
A-2 Repeat Problem A
O: O X X X X

The matrix P II is chosen to make the first column of A, except the first two
elements, zeros as shown. This can be achieved by using gaussian elimination
with partial pivoting or a Householder transformation. The inverse of PI has
the same form as PI' The postmultiplication of PIA by P¡I will not operate
on the first column of P¡A; hence the pattern of zeros in the first column of
PI A is preserved in P IAP¡I. Next we find a P 2 so that Verify that b ji = O if the it

10:0 O O X X X X X A-3 Is it possible to 01


Hessenberg form?
9__ ! _: _~ _9 ~_ x x X X X

O o:, O X X X X
A-4 Show that if Q¡. i =
PIAP¡I= O O
O O: X X X so is Q~Q,"Qm-l'" Qz
,
,
, A-S Let A be an 11 xm n
O O: O O X X X
that

Again the inverse of P 2 has the same form as P 2 and the postmultiplication of
P 2 P¡AP¡1 by Pi l will not operate on the first two columns of P 2 P¡AP¡I.
Hence the pattern ofzeros in the first two columns of P 2 P IAP¡1 is preserved where R is an upper trian
in P 2 P IAP¡I Pi l. Proceeding similarly, we can transform a matrix, by a is the first nI columns of <:
sequence of similarity transformations, into the form holder transformations a
() N.. f8.ctorizatíon 8.::1.0 t~(­
elements of JR may be diff¡
x x X X X X X
¡atter, they can be chosen .
X X X X X X X

O x x x x x x A-6 Let A be a square


matrix A~ RQ has the s¡
O O x x x x x (A-17)
are similar. The QR algo,
O O O x x x x values oC A, is based 00 tI­

O O O O x x x
Ü O O O O x x

an
This is called upper Hessenhergjorm and can be obtained by using numerically
stable methods. This form is very useftil in many computer computations.
PROBLEMS 553

I form will be used in Problems

A-1 Use the row searching algorith"l to find linearly dependenl rows of

x
x
x
x [~:r &~
s'j
4 1.5
x x Use the circled elements as pivot elements. Use the procedure in (A-9) to (A-12) to find the
x x coefficients of ¡ts linear combination.

x A-2 Repeat Problem A-I for the matrix


x
o
-~]

3 -1
\, except the first two I :)) 2
g gaussian elimination
The inverse of PI has
I P1
1
will not operate
in the first column of
[ i
20
3
I
6
1
O
3
O -3
3
7
O
-1

Verify that bji =0 if the ith row. with i <jo is linearly dependent.

x x x A-3 Is it possible to obtain a triangular form by using the procedure of obtaining a


Hessenberg form?
x x x
x x x A-4 Show that if Qi' i = L 2. . .. m. are unitary matrices. that ¡s. QtQi = Q¡Qt = I. then
x x x .. so isQ~QmQm-l'" Q2QL

A-S Let A be an n x m matrix of rank m. Show that there exists a unitary matrix Q such
x x x
that

: postmultiplication of or
,lumns ofP 2 P¡AP¡I.
2 P 1 AP ti is preserved where R is an upper triangular matrix with nonnegalive clements 011 the diagonal and Q¡
.sform a matrix, by a is the first m columns of Q*. This is called the QR !actorizacion of A. (Hinc: Use House­
holder transformations and Problem A-4. Note that the only difference between the
QR factorization and the I~t'::n~seholder tr8.nsform.8.~.~,C)'l. ~.~ th!::1) t'1~ sl~~"?.s ~.J ~r,:,; r:!~2.g.~·):n?~
elements of R may be differenl. In the former. the signs must be chosen as positive: in the
latter. they can be chosen as positive or negative to reduce the roundoff errors.)

A-6 Let A be a square matrix and let A = QR be its QR factorization. Show that the
matrix A~RQ has the same set of eigenvalues of A =QR. (Hinc: Show that A and A
(A-17)
are similar. The QR algorichm, which is the most reliable method of computing the eigen­
values of A, ís based on this property. See References SI81 and S200.)

d by using numerically
Iter computations.
The sum, product
o at any point) of an
nomials, exponential
entire real lineo
If a function is kn
determinable from an

B known. This can be


derivatives at t o are
(t 0 - eo, t o +e o)· Ne)
again, from (B-l), we
Analytic Functions both directions, we c~
analytic continuation.
of a Real Variable
Theorem 8-1
If a functionf is anal
arbitrarily smaH nonz
D.

Proof
If the function is iden
Let D be an open interval in the realline IR and letf(') be a function defined on (to, ( 1 ), then the func!
D; that is, to each point in D, a unique number is assigned to f The function By analytic continua'
fU may be real-valued or complex-valued.
A function fO of a real variable is said to be an element of dass C" on D
if its nth derivative, ¡<"lO, exists and is continuous for aH t in D. ero is the class
of functions having derivatives of aH orders.
A function of a real variable, f('), is said to be analytic on D iff is anelement
of ero and if for each t o in D there exists a positive real number eo such that, for
aH t in (to - eo, t o +eo), f(t) is representable by a Taylor series about the point
to:

A remark is in order at this point. If f is a function of a complex variable


and if f has continuous derivative, then it can be shown that f has continuous
second derivative, third derivative, ... , and in fact a Taylor-series expansiono
Therefore, a function of a complex variable may be defined as analytic if it has
a continuous derivatíVe. However, for functions of a real variable, even if a
function possesses derivatives of aH order, it may still not be analytic. For
example; the function
1/,2
for t-l=O
f(t)= { ~ . for t =0
is not analytic at t = O,' even though it is infinitely differentiable at t = O; see
Reference 9.

554
ANALYTIC FUNCTlONS OF A REAL VARIABLE 555

The sum, product, or quotient (provided the denominator is not equal to


O at any point) of analytic functions of a real variable is analytic. AII poly­
nomials, exponential functions, and sinusoidal functions are analytic in the
entire realline.
If a function is known to be analytic in D, then the function is completely
determinable from an arbitrary point in D ifall the derivatives at that point are
known. This can be argued as follows: Suppose that the value of ! and its
derivatives at t o are known, then by using (B-l), we can compute! over
(to.-8 0, t o +80)' Next we choose a point tI that is almost equal to t o +8 0 :
again, from (B-l), we can compute! over (to +80, t o +8 0 +8 1), Proceeding in
both directions, we can compute the function! over D. This process is called
analytic continuation.

Theorem 8-1
If a function! is analytic on D and if! is known to be identically zero on an
arbitrarily small nonzero interval in D, then the function! is identically zero on
D.

Proof

If the function is identically zero on an arbitrarily small nonzero interval, say


a function defined on (lo, tI), then the function and its derivatives are all equal to zero on (to, tI)'
~d to f The function By analytic continuation, the function can be shown to be identically zero.
Q.E.D.
lent of class C" on D
in D. CO is the class

:m D if! is an element
mber 80 such that, for
;eries about the point

(8-1 )

)f a complex variable
:hat f has continuous
{Ior-series expansiono
:d as analytic if it has
'al variable, even if a
IOt be analytic. For

entiable at t=O; see


where Ilu(t)II~(u(t),
Section 2-8.)

Proof
The solution of the sI

e Defüle
x(t 1

Minimum-Energy Control 1 x~
Then the assumption

-i
X=
lO
l1

Subtracting both side

Consider the n-dimensional linear time-varying state equation


which implies that
E: x= A(t)x + B(t)u
where x is the n x 1 state vector, u is the p x 1 input vector, and A and B are,
respectively, n x n and n x p matrices whose entries are continuous functions of
\L l!>(.

t defined over (- 0:<


ro). lfthe state equation is control1able at time to, then By using (2-94),we ca
given any initial state Xo at time to and any desired final state xl, there exists a
finite tI > to and an input u[lo,ld that will transfer the state Xo at time to to Xl at
time t 1> denoted as (x o, to) to (x 1> t d. Define
With the use of(C-l),
W(t o, t¡)~ l'¡ <l>(t o, ,)B(,)B*(,)lI>*(to, ,) d,
lO
Then the input ufro,ld defined by
Considcr now
Iilio(t) = (<<!l(t o, t)R(t))*W - l(rO, tI )r$(i o. tdx! - Xo] fOí aH t in [too t, J U>1 ¡
will transfer (xo, to) to (Xl, td. This is proved in Theorem 5-4. Now we show
that the input UfrO,I,) consumes the minimal amount of energy, among all the
u's that can transfer (xo, to) to (Xl, t¡). By sorne manipulatio

Theorem C-1

Let u{'o,ld be any control that transfers.(xo, to) to (x 1, t ¡), and let UO be the control
defined in (C-l) that accomplishes thesame transfer; then

1', Ilul(t)112 dtc.l·


'o
1,

lO
IluO(t)112 dt

I This appendix c10sely rollows Reference 69.

556
MINIMUM-ENERGY CONTROL 557

where IIU(t)II~«U(t), U(t»)112~(U*(t)U(t))1/2, the euclidean norm of U(t) (See


Section 2-8.)

Proof
The solution of the state equation E is

(C-2)

Define
x~<ll-I(tl' tO)x(tl)-x(tO)=<ll(t o, t¡)XI-XO

Then the assumptions that u l and UO transfer (X 0, to) to (xl, t¡) imply that

x=
i
"
lO
<ll(t o, r)B(r)ul(r) dr =

Subtracting both sides, we obtain


i'l
lO
<ll(t o, r)B(r)u°(r) dr

lation
which implies that

:tor, and A and B are, ( [ <I>(t o, r)B(r)(ul(r) -uO(r)) dr, W-l(t o, tl)X) =0
Jntinuous functions of
)llable at timeto, then By using (2-94), we can write this equation as

i"
state XI, there exists a
e Xo at time to to Xl at (u] (r) - uO(r), (4)(t o, r)B(r »*W -I(t o, ti }x) dr =o (C-3)
'o
With the use of(C-\), Equation (C-3) becomes
) dr 'l
i'o
(l.I l (r)-uO(r), uO(r» tlr =0 (C-4)

Consider now

m 5-4. Now we show


energy, among all the
LII ul 2
l1 elr
By sorne manipulation and using (C-4), we obtain

r'l Ilul(r)W elr = Jtor" llul(r} -uo(r} +uo(r}11 2 dr


10
nd let UO be the control
n = [' Ilul(r} - uO(r}11 2 elr + [' IluO(r)W dr

+2 [ (u 1 (r)-uO(r},uO(t»dr

= r'¡ Ilul(r) -
Jto
u O(r)112 dr + r"
Jro
IluO(r)W elr
558 MINIMUM-ENERGY CONTROL L

Since

is always nonnegative, we conclude that

Q.E.D.

We see that in transferring (xo, to) to (Xl, t¡), if minimizing

is used as a criterion, the control defined in (C-I) is optimal. Since, in many


instances,

is related to energy, the control U o is caBed the minimum-energy control.

Consider the linear ti

where x is the n x 1 st
vector; A, B, and e ¡;
The response of FE Ü

where Xo is the initial


We consider now
the input u changes v;
occur in sampled-dat
used to generate u.
sampler and a filter, e
u(t) = u
where T is a positive
0, T, 2T, ... , are called
constant inputs given "
the" behavior at samp
dynamical equation (
izing
Q.E.O.
D
Controllability after the
Introduction of Sampling
imal. Since, in many

-energy control.

Consider the linear time-invariant dynamical equation


FE: x=Ax +Bu ( 0-1a)
y=Cx (0-1b)

where x is the n x 1 state vector, u is the p x 1 input vector, y is the q x 1 output


vector; A, B, and e are n x n, n x p, and q x n constant matrices, respectively.
The response ofF E is given by

x(t)= eA1xo+ I eA(I-'lBu(r)dr (0-2a)

y(t) = Cx(t) ( D-2b)

where Xo is the initial state at t = o.


We consider now the case in which the input u is piecewise constant; that is,
the input u changes values only at discrete instants of time. Inputs of this type
occur in sampled-data systems or in systems in which digital computers are
used to generate u. A piecewise-constant funcHon is often generated by a
sampler and afilter, caBed zero-order hold, as shown in Figure 0-1. Let
u(t) = u(k) for kT 50t «k + l)T; k =0,1,2, ... (0-3)

whert:: T is a posiÜve constant, caBed the sampling periodo The discrete times

O, t; 2T, . .. , are caBed sampling instants. The behavior of FE with the piecewise­

constant inputs.given in (0-3) can be computed from (0-2). However, if only

. the behavior at sam pling instants O, T, 2 T, ... , is of interest, a discrete-time

dynamicalequatio~ can be written to give the responseof x(k)~x(kT) at

559
560 CONTROLLABILlTY AFTER THE INTRODUCTION Of SAMPLlNG

This pro blem is imp


u(e) .::::¡-
T~u ~(e)
Hold
kJ u(
See Problem 7-30.
A discrete-time e
u(e) u(k) for any initial state ~
" q uence {u(n)} of finit<
tions in statements 3 ¡

o
1"
tjil
o
1

T
1
I
2T
I
1
I
L-J
l.
I
are directly applicab
Let ;,¡(A) denote;
imaginary part and t:

Theorem 0-1

Assume that the dYI


Figure 0-1 sufficient condition f
(0-5) to be controllat
whenever Re [A¡(A) ­
For the single-inp
k = O, 1,2, . . .. From (0-2a), we have
(k+ I)T First we remark (
x(k+l)=e A (k+l)T xo + o eA[kT+T-tIBu(T)dT
J
}'iis an eigenvalue o
Al =T +jf3 and Az =.
=eAT[eAkTxo+ J: T
eA(kT-tlBU(T)dT]
Xz = e(t+ i(fJ + ZnajT))T =
tions of the theorem ¡
(k+ 1)T conditions or the the(
+
lkT
eA(kT+J·;-tlBu(T) dT

The term in brackets in (0-4) is equal to x(k); the input U(T) is constant in the
(0-4) values of A, then {Xi,

Proof of Theorem D
interval (kT, kT + T) and is equal to u(k); hence (0-4) becomes, after the change
of variable a = kT + T -T, We assume, without I
shown in Table 5-1.
x(k +1)=e ATx(k) + ( f eAada) Bu(k)

which is a discrete-time state equation. Therefore, ir the input is piecewise


constant over the same interva! T; and if only the respons"C ~t '-Íl'C s2xrpj¡·",,.
instants is of interest, the dynamical equatio:1 FE in (0-1) can be replaced by the
following discrete-time linear time-invariant dynamical equation
DFE: x(k + 1) = Áx(k) +Bu(k) (0-5a)
y(k) = Cx(k) (0-5b) where

where A=e AT (0-6)

B=(f e
At
dT) B~MB (0-7) A

C=C (0-8)

lfthe dynamical equation FE iscontrollable, it is ofinterest to study whether


the system remains controllable after the introduction of sampling or, corre­
spondingly, whether the discrete-time dynamical equation DF E is controllable.
CONTROLLABILlT'I"'!\FTER THE INTRODUCTlON OF SAMPLlNG 561

This problem is important in the design of dead-beat sampled-data systems.


See Problem 7-30.
A discrete-time dynamical equation DF E is defined to be controllable ir
for any ¡nitial state Xo and any x 1 in the state space, there exists an input se­
q uence {u(n)} offinite length that transfers X o to x l' The controllability condi­
tions in statements 3 and 4 ofTheorem 5-7 and in Theorem 5-21 and its corollary
are direct\y applicab\e to the discrete-time case w1thout any modification.
---¡ Let ),¡(A) denote an eigenvalue of A, and let "1m" and "Re" stand for the
h
I I
imaginary part and the real part, respectively.
I I l.
2T I I Theorem 0-1
L-J
Assume that the dynamical equation FE given in (D-1) is controllable. A
sufficient condition for the discrete-time dynamical equation DFE given in
(D-5) to be controllable is that 1m [,.li(A) - }./A)] i= 2nr:t./T for IX = ± 1, ± 2, ... ,
whenever Re [A¡(A)-A/A)] =0.
For the single-input case (p = 1), the condition is necessary as well. I

First we remark on the conditions of the theorem. Because of Á = eAT , ir


u(r) dr Ai is an eigenvalue of A, Xi ~ eA¡T is an eigenvalue of Á (Problem 2-32). Let
}'l =r +jf3 and Az =r +j(f3 +2nrt./T), for IX = ± 1, ±2, .... Then we have
Xz = é<+ ¡(P + Zna.IT))T = e«+ jPlT + jZna. = ei<+ jPlT = Xl' In other words, if the condi­
tions of the theorem are violated, different Ai may yield the same Xi- Hence the
conditions of the theorem ensure that if {Ai' i = 1, 2, ... , m} are distinct eigen­
. (0-4) values of A, then {Xi' i = 1,2, ... , m} are distinct eigenva\ues of Á.

u(r) is constant in the Proof of Theorem D-1


Jmes, after the change
We assume, without loss of generality, that A is in the Jordan canonical form
shown in Table 5-1. Then, from (0-6), (2-70), and (2-71), we have
:)

he input is piecewise - A-
I.
e
A

(1
,,1'

e
O
A12T
O
«] l I
anse at the sampling .A==q t-;:-::~

:;an be rep\aced by the


quation l O O eA"'~'(m) J T

(0-5a) T(II;j-l)
T
(0-5b) where (nij-1)!
T(II¡j- Z l
(0-6)
O
(nij ­ 2)!
Á-·6. A ijT -J.
(0-7) 1)= e - \"l T(lI i j - 3 l
O O (D-10)
(nij-3)!
(0-8)

~rest to study whether


If sampling or, corre­
O O J
and 1i = eA¡T Although A i ¡ is a Jordan block, Áijis not in aJordan formo Now
1 DFE is controllab\e
------ ----~
--.. .'~--

- ------ - - - - - - - - - - --- - - - - - - - - - - - - - - - - - - - _.. __ ._---------_._----._--_.. -.~------_.-

562 CONTROLLABILITY AFTER THE INTRODUCTION OF SAMPLING

we shall transform it into a lordan formo We compute condition of the the,


if T > O, M is nonsin
o x x x x
If Á is in the for[
O O x x x
b/ij, j = 1, 2, ... , r(i), i
O O O x x tions on the eigenval
Áij- X¡I =
Hence the necessary
O O O O x trollable is that the r,
O O O O O rU), are liriearly inde
the case if and only i
where x denotes possible nonzero element. Because of its special form, it is conclude that under
easy to verify that (Áij - XJ? has one more zero superdiagonal row than the discrete-time equ
(Á ij - XJ); and finally, similar to (2-64), (Á¡j - XJ)"i¡-l has only one nonzero time equation in (0­
element at the upper right comer and (Áij - XJ)"¡j is a zero matrix. Conse­ In the single-inpl
quently, e = [O O ... O 1]' is a generalized eigenvector of grade nij (see there will be two or r
Section 2-6). From this vector we can obtain a similarity transformation of Equation (0-14). H
the form lary 5-21, not con tro

x x x x O In the proof of T

O x x x O theorem.

O O x x O
Qij= (0-11 )
Theorem D-2
O O O x O If a continuous-tim~
O O O O 1 its discretized state ec
such that
This theorem is i
, by using an input de
II.¡ O O O
be controlled by an il
O Xi 1 O O
O O Xi O O Example
PijÁijQij = (D-12)
Consider the sample
O O O X¡ expansion, we have
O O O O X¡
where Pti~ Qij 1 and has exactly the same form a~ Qij' We see that the lordan (5 + 1)(5 + 1
form of Aij is the same as A ij with A i replaced by Ai' Define

(0-13)

then (O-Sa) can be transformed into

x(k + 1) = PÁP- Ix(k) + PMB~(k) (0-14) u(¡)1 ,Hold-;f;.-


~¡-L
'

with P Áp-l exactly in the form in Table 5-1 with A¡replaced by Xi- Now we
shall use this equation toestablish the theorem., ' '
First we note that, because of (0-9) and (0-10), the matrix M defined in
(0-7) is of block diagonal form, each blockis a triangularform of order n¡j
with all diagonal elements equal to (l-e-AiT)/A i if A;=I=O or T if A¡ = O. The Figure 0-2
CONTROLLABILITY AFTER THE INTRODUCTlON OF SAMPLING 563

condition of the theorem implies Ai T +- 2mx, if Re )'i = O or e - i.¡T +- 1. Hence


if T > O, M is nonsingular.
lf A is in the form in Table S-1 and if {A, B} is controllable, then the rows
b1ij , j = 1, 2, ... , r(i), in (S-71a) are linearly independent. Now under the condi­
tions on the eigenvalues stated in Theorem 0-1, Xi, i = 1, 2, ... , m, are distinct.
Hence the necessary and sufficient conditions for {PAP-¡, PMB} to be con­
trollable is that the rows of PMB corresponding to the rows of buj , 1= 1, 2, ... ,
r(i), are linearly independent. Because of the special forros of P and M, this 1S
the case if and on!y if b1ij,j = 1, 2, ... , rU) are linearly independent. Hence we
. its special form, it is conclude that under the conditions on the eigenvalues stated in Theorem 0-1,
lerdiagonal row than the discrete-time equation in (O-Sa) is controllable if and only if the continuous­
h.as only one nonzero time equation in (O-la) is controllable.
zero matrix. Conse­ In the single-input case, if the conditions on the eigenvalues are violated,
~ctor of grade nij (see there will be two or more Jordan blocks associated with the same eigenvalue in
rity transformation of Equation (0-14). Hence the single-input equation (0-14) is, following Corol­
lary S-21, not controllable. This establishes the theorem. Q.E. D.

In the proof of Theorem 0-1, we have essentially established the following


theorem.
(0-11 )
Theorem D-2
lf a continuous-time linear time-invariant state equation is not controllable,
its discretized state equation for any sampling period is not controllable. 1m

This theorem is intuitively obvious. lf a state equation is not controllable


by using an input defined for all time, its discretized equation certainly cannot
be controlled by an input defined only at discrete instants of time.

(0-12) Example
Consider the sampled-data system shown in Figure D-2. By partial fractiol1
expansion, we have

We see that the Jordan (s + 1)(s + 1 + 2i)(s + 1 - 2i) s + 1 (s + 1 + 2i) (s + 1 - 2i)


fine

(0-13)
3
2

T
s 1
(0-14)
~
8 y
(s + 1) (s + 1 + 2j) (s + \ - 2j) -3 -2 -\ 2 3
llaced by Xi' Now we

: matrix M defined in
ular form of order nij
O or T if Ai = O. The Figure 0-2
564 CONTROLLABILlTY AFTER THE INTRODUCTION OF SAMPLING

Consequently, an irreducible Jordan-form realization of the system S can be


found as
o
-1-2i o
O ][XI] +[1]
x? 1 u
O -1+2i X3 1
y = [2 -1 -I]x
By using (0-5) through (0-8), a discrete-time state equation can be computed as
1 -e -T
O _1_(I_e- T - 2iT )
e - T- 2iT
1 +2i u(k) (D-15) Si r
O
_1_(1 _e-T+2iT)
1-2i
We conclude from Theorem 0-1 that the discrete-time state equation (0-15)
is controllable if and only if
2nrx
T op-- =nrx A hermiáan form of
2 geneous polynomial ,
2nrx n
and Top--=- rx rx = ± 1, ±2, ...
~ 2
This fact can also be verified directly from (0-15) by using either the criterion or, in matrix form,
p[B AB A2 B] = 3 or Corollary 5-21.

Problems

0-1 Consíde:r the: continuolls-time: state e:qllatíon


where the mJs are an
~ 1\1; +l~
o Sinc~ every hermitio.f
- 1--- 2i O
O -1 +2;J Lo O
Show that its discretized state equation is always controllable ror any T including T = 27((44 where Mi is the com
for cx = ± 1, ±2, .... This shows that the conditions on the eigenvalues in Theorem 0-1
x*M
are not necessary ror a multiple-input discretized equation to be controllable:.
where M =~(Ml +M
0-2 Show that sufficient conditions for thé discrete-time dynamical equation in (0-5) to be can be written as x*lV
observable are that the dynamical equation iri (O-I)1s observable and lni [A¡(A)-A)A)] i=­ is called a hermitian tí
2ncx/T for cx'= ±1, ±2, ... , whenever Re [A-¡(A)-J,}A)]=O. For the single-output case
In thestudy of hel
(q = 1), the cOndit-ions are necessary as well.
product. Observe te
operatol" that maps t
lf the inner product (
the system S can be

r:I

T
can be computed as E
Hermitian Forms and
-T- 2iT)
u(k) (D-15) Singular Value Decomposition
- T+ 2iT)

state equation (D-15)

A hermiLian form of n comp\ex variables X¡,X2, ... , X Il is a real-valued homo­


geneous polynomial of the form
11

I
'¡,j~ ¡
mijxiXj

Ig either the criterion 01', in matrix form,

X2 -][:~: :~:
X Il • • :~'j'l[:~,]
" =x *M ¡x
6. (E-1 )

1~1"1 '~1"2 1~11l1l x"


where the mi/s are any complex numbers and Xi is the comp\ex conjugate of Xi'
Since every hermitian form is assumed lo be real-valued, we have
x*M¡x =(x*M¡x)* =x*Mix
IY T including T = 2ncx/4 where Mi is the complex conjugate transpose of M¡; hence
lValues in Theorem D-1
;on tro Hable. x*M¡x =x*(~(M¡ +Mj')x~x*Mx (E-2)

where M =~(M¡ + Mi). lt is clear that M = M*. Thus every hermitian form
11 equation in (D-S) to be
can be written as x*Mx with M = M*. A matrix M with the property M = M*
lnd 1m [},¡(A)-})A)] f
is ca\led a hermitian matrix.
r the single-output case
In the study of hermitian forms; it is convenient to use the notation of inner
product. .Observe that the hermitia'n matrix M can be considered as a linear
operator that maps the n-dimensi.ona\ Gomplex vector space (1[", C) into itself.
If the inner product of (C', C)is chosenas' .
(x, y)~x*y ( E-3)

565
566 HERMITIAN FORMS AND SINGULAR VALUE DECOMPOSITION

where X and y are any vectors in (iC", q, then the hermitian form can be written Now we show thal
as matrix can be chosen '
following theorem.
x*Mx = (X, Mx) = (M*x, X) = (Mx, X) (E-4 )

where, in the last step, we have used the fact that M* = M. Theorem E-3
The eigenvectors of a f:
Theorem E-1
are orthogonal.
All the eigenvalues of a hermitian matrix M are real.
Proof
Proof
Let e¡ and ej be the eig
Let il be any eigenvalue ofM and let e be an eigenvector ofM associated with il; ),¡ and }'j, respectively
that is, Me = ile. Consider
(e, Me) = (e, ile) = il(e, e) ( E-5)
and
Since (e, Me) is a real number and (e, e) is a positive real number, fram (E-S)
we conclude that il is a real number. Q.E.D. where we have used t
fram (E-7), we obtai
Theorem E-2 (e¡, e) =0.

The Jordan-form representation of a hermitian matrix M is a diagonal matrix.


Since every eigen'
Proof hermitian matrix M a
the eigenvectors assoc
Recall fram Section 2-6 that every square matrix which'maps (iC", iC) into itself normal. We consider)
has a Jordan-form representation. The basis vectors that give a Jordan-form the same eigenvalue.
representation consist of eigenvectors and generalized eigenvectors of the vectors associated wil
matrix. We show that if a matrix is hermitian, then there is no generalized orthonormal vectors f
eigenvector of grade k ~2; we show this by contradiction, Suppose there
exists a vector e such that (M - il¡1/e = O and (M - il¡l)k - le 1=0 for sorne Ll I = 1
eigenvalue il¡ of M. Consider now, for k ~2,
O= «(M -il¡1)k-2 e, (M -il¡1)ke ) = «(M -il¡1)k-I e, (M _il¡l)k-I e )
U2 =€
= Ii(M - il¡1)k - leW

which implies (M - il¡ll- le = O. This is a contradiction. Hence there is no


generalized eigenvector of grade k ~2. Consequently, there is no Jordan block
whose order is greater than one.' Hence the Jordan-form representation of a
hermitian matrix is a diagonal matrix. In other words, there exists a nonsingu­
lar matrix P such that PMP - 1 = M and M is a diagonal matrix with eigenval ues
um=e
on the diagonal. Q.E.D.

Two vectors x,. y are said to be orthogonal if and only if (x, y) = O. A vector The procedúre for de
x is said to be no.rmalized if and only if (x, x) ~ IIxl1 2 = 1. It is' clear that every Schmidt oi,thonormalizl
vector x can benormalized by choosing x = (l/llxll)x. A set of basis vectors that (q¡, qj'j :=0 for i i=
{ql, Q2, .. . "qll} issaid to be an orchonormal basis if and only if Fram Theorem E·
conclude thatfor any r.
i 1=} with'respect lo which t
( E-6)
i=} or equivalently, for any
HERMITIAN FORMS AND SINGULAR VALUE DECOMPOSI1'ION 567

1 form can be written Now we show that the basis of a lordan-form representation of a hermitian
matrix can be chosen as an orthonormal basis. This is derived in part from the
following theorem.
(E-4 )
Theorem E-3
The eigenvectors of a hermi tian matrix M corresponding to different eigenval ues
are orthogonaL

Proof

Let e¡ and ej be the eigenvectors of M corresponding to the distinct eigenvalues


M associated with A; A¡ and Aj, respectively; that is, Me¡ = A¡e¡ and Mej = Aje j. Consider

(ej, Me¡) = (ej, A¡e¡) =A¡(ej, el) ( E-7)


( E-5)
and (ej, Me¡) = (Mej, el) = (Ajej, el) = )'j (ej, el) ( E-S)
l number, [ram (E-S)
Q.E.D. where we have used the fact that the eigenvalues are reaL Subtracting (E-8)
+
from (E-7), we obtain (A¡ - Aj)(ej, el) = O. Since A¡ Áj, we conclude that
(e¡, e j ) =0. Q.E.D.

is a diagonal matrix.
Since every eigenvector can be normalized and since eigenvectors of a
hermitian matrix M associated with distinct eigenvalues are orthogonal, then
the eigenvectors associated with different eigenvalues can be made to be ortho­
laps (C", C)into'itself normaL We consider now the linearly independent eigenveCtors associated'with
.t give a Jordan-form the same eigenvalue. Let {e 1, ez, ... , em } be a set of linearly independent eigen­
eigenvectors of the vectors associated with the same eigenvalue. Now we shall obtain a set of
ere is no generalized orthonormal vectors from the set {el, ez, . .. , e m }. Let
¡ion. Suppose there
.¡I)k - le +0for some

Uz =ez - (ql, eZ)ql

.. Hence there is no
:re is no lordan block
n representation of a
ere exists a nonsingu­
,"-1
atrix with eigenvalues
Q.E.D.
um=e m- ¿ (qk,em)qk
k=l

, (x, y) = O. A vector The pracedure for defining q¡ is illustrated in Figure E-l. 1t is called the
1t is clear that every S,chmidt orthonormalization pr0cedure. By direct verification, it can be shown
\. set of basis vectors +
that (q¡, qj) =0 for i j.
nly if 'From Theorem E-3 and the Schmidt orthonormalization procedure we
condude that for any hermitian matrix there exists a set of orthonormal vectors
( E-6) with respectto which the hermitian matrix has adiagonal-form representation;
or equivalently,for any hermitian matrix M, there exists a nonsingular matrix Q
568 HERMITlAN FORMS AND SINGULAR VALUE DECOMPOSITION H


, If the rank of H is r. SI
3 A21 », z2 > . . . '''j'
; 2 > O a'ne
u ---­
---- - -
the orthonormal eigenv
Q = [ql qz
Then Theorem E-4 imp
ql

~< ql,e 2
where :E 2 = diag{ A¡, },~,
Figure E-1

and
whose columns are orthonormal, such that
which implies
!VI = Q - lMQ g PMP - l
where M is a diagonal matrix and P g Q - l. Let Q = [ q l qz qll].
where 2: = diag {Al' Az,
Because of the orthonormal assumption, we have

Q'Q Ji::;:] [ql q2 qll] = I Then (E-13) becomes Rj


normal. Let Hz be cho
l(qll)*
R*HQ=[
Hence Q-I =Q*, and p- i =P*. A matrix Q with the property Q* =Q-I
is called a unitary matrix. We summarize what we have achieved in the fol!ow­
Ciearly, (E-14) implies
ing.
is orthonormal, we h:
becomes
Theorem E-4
A hermitian matrix M can be transformed by a unitary matrix into a diagonal
matrix with real elements; or, equivalently, for any hermitian matrix M there
exists a nonsingular matrix P with the property P - l = p* such that This is stated as a theor
tVA = PMP*
Theorem E-5 (Singula
where !VI is a diagonal matrix with real eigenvalues ofM on the diagonal. I
Every m x n matrix H o
In the fol!owing, we shal! utilize Theorem E-4 to develop the singular value
decompositions for matrices. Let H be an m x n matrix. Then the matrix R*HQ'
H*H is a square matrix of order n. Clearly H*H is hermitian; hence its eigen­
whereR*R = RR* = 1m ,
values are al! real. BecauseH*His positive semidefinite; its eigenvalues are al!
non:negative (Theorem 8-19). Let il¡, i = 1, 2, ... ,n be the eigenvalues of H*H. Al ~A2~'" ~Ar > O.
The set.{íl¡ ~O, i = 1,2, ... , n} is calledthe singular values of H. lfH is hermitian,
its singular values are equal to the absolute values of the eigenvalues of H. Although :E is uniql
Fot convenience, we arrange {,q such that arenot necessarily uniql
the corresponding colun
( E-9) space spanned by the t
HERMITiAN FORMS AND SINGULAR VALUE DECOMPOSITION 569

If the rank of H ís r, so ís the rank of H*H (Problem E-3). Hence we have


Ai ~Al ~ '" X; >0 and A;+¡ =A;+z = ... =A~ =0. Let q¡, i = 1,2, .. " ,n,be
the orthonormal eigenvectors of H*H assocíated wíth }.f. Define
Q = [q¡ qz ... q,.: q,+ ¡ (E-10)

Then Theorem E-4 implies

Q*H*HQ=[~Z ~J (E-11)

where 1;z = díag{Ai, J,i, ... , An. Using Q = [Q¡ Qz], (E-lI) can be written as
Q!H*HQz =0 ( E-12)

and QtH*HQ¡ =1;2


which implies
1;-IQtH*HQ¡1;-¡ =1 ( E-13)
Q = [q¡ qz ... qll].
where 1; = diag {A¡, Az, ... , J.,.}. Define the m x r matrix R¡ by
Rl~HQ¡1;-¡ (E-14 )

Then (E-l3) becomes RtR¡ = 1 which implies that the columns of R¡ are ortho­
normal. Let R z be chosen so that R = [R¡ R z] is unítary. Consider

. [Rt] "] = [RtHQi RtHQz]


R*HQ= R! H [ Ql Qz R*HQ R*HQ
( E-15)
.he property Q* = Q-¡ z ¡ z z
: achíeved in the follow­
Clearly, (E-14) ímplies HQ¡ = R¡1; and (E-12) ímplies HQz =0. Because R
is orthonormal, we have RtR¡ = I and R!R ¡ = O. Conseq uently, (E-15)
becomes

, matrix into a diagonal


rmítían matríx M there
R*HQ=[~ ~J

~* such that This ís stated as a theorem.

Theorem E-5 (Singular value decomposition)


)ll the díagonal. I
Every m x n matrix H of rank r can be transformed into the form

R*HQ=[~~] H=R[~ ~JQ*


velop the singular value
Ltrix. Then the matrix or
mitian; hence its eigen­
e, íts eígenvalues are a11 where R*R = RR * = 1m , Q*Q = QQ* = 1m .and 1; = díag {A¡,A Z' . " . , A,} wíth
:he eigenvilJues of H*H. A¡ ~Az ¿ ... ¿A,>O. I

DI H. ·{fH iS hermítian,
,f the eigenvalues of H. Although L is uniquely determined by H, the unitary matrices R and Q
.are not necessarily unique. Indeed let Al be a multiple eigenvalue of H*H; then
the correspqnding columns ofQ may be chósen as any orthonormal basis for the
( E-9) spacespanned by the eigenvectors of H*H corresponding to Al. Hence Q
570 HERMITlAN FORMS AND SINGULAR VALUE DECOMPOSITlON

is not unique. Once Q is chosen, R¡ can be computed from (E-14). The choice E-9 Show that aH the ~
of R z again may not be unique so long as [R 1 RzJ is unitary. are equal to 1.
The singular value decomposition has found many applications in linear
systems. In Section 6-S, we use it to find an irreducible realization from a E-10 What are the eiger
Hankel matrix. It can also be used to find simplified or approximated models
of systems. See References S141, S161, and S171. The singular value de­
composition is also essential in the study of sensitivity and stability margin of
multivariable systems. See References S34 and S194. For computer pro­
grams, see Reference S82. El may arise in gaussian e
The elements of matrices in this appendix are permitted to assume real or tion with partial pivoting.
complex numbers. Certainly all results still apply if they are limited to real largest and smallest sinDlI
numbers. For a real matrix M we have M* = M', where the prime denotes speaking, a condition ~u:
computation.
the transpose. A real matrix with M = M' is called a symmetric matrix; a real
matrix with M- 1 = M' is called an orthogonal matrix. With these modifications
E-11 Show that the conl
in nomenclature, all theorems in this appendix apply directly to real matrices.

Problems is positive definite if and e


E-1 Ir Mis an n x n matrix with complex coefficients, verify that x*Mx is a real number
E-12 Show that Theorerr
for any x in C" if and only if M* = M. Ir M is an n x n matrix with real coefficients, is it
tive semidefinite. lf H i:
true that x'Mx is a real number for any x in IR" if and only if M' = M?
what are the differences be
E-2 Find an orthonormal set from the vectors
E-13 What arethesingul

mm UJ
Al =!

E-3 Show that if the rank of H is r, so are the ranks of H*H and HH* (Hint: Find a P
such that AP = [A O] and rank A = rank Á.)

E-4 Find the eigenvalues and singular values of the matrices

10 011 : fl O Il
H¡=l~ v"
h2=l~ ~J
(j

O -IJ O

E-5 Find the singular value decompositions of the matrices in Problem E-4.

E-G Show that

IIAlb =largestsingular value of A


= [)'max(A *Al] 1/2

E-7 In Theorem E-S, we have Q*Q = QQ*= 1... If we write Q = [Q¡ Qz], do we have
Ql*Q¡ = 1,. and Q¡Qf = In? .

E-S Are aH the. eigenvalues ofa unitary matrix (including orthogonal matdx) real?
Show that all the eigenvalues of a unitary ma:trix have magnitudes equalto 1.
PROBLEMS 571

om (E-14). The choice E-9 Show that all the singular values of a unitary matrix (including orthogonal matrix)
.nitary. are equal to L
applications in linear
ible realization from a E-lO What are the eigenvalues and singular values ofthe elementary matrices
. approximated models
['he singular value de­
and stabili ty margin of
El =[ ~ ~ ~]
-3 O 1
E 2 = 0.5
[0.8
1 O 0J
1 O
O 1
For eomputer pro­
El may arise in gaussian elimination without any pivoting: E 2 may arise in gaussian elimina­
tion with partial pivoting. lfthe condition number of a matrix is defined as the ratio ofthe
itted to assume real or
largest and smallest singular values, which matrix has a larger condition number'! Roughly
hey are limited to real speaking, a condition number gives the amplification factor of the relative errors in the
lere the prime denotes computation.
ymmetric matrix; a real
'ith these modificatiuns E-ll Show that the controllability grammian
irectly to real matrices.
W" = I eA'BB*e A', de

is positive definite if and only if {A, B} is controlIable.

hat x*Mx is a real number E-12 Show that Theorem E-S reduces to Theorem E·4 if H is square, helmitian, and posi­
with real coefficients, is it tive semidefinite. If H is square and hermitian (without being positive semidefinite),
'=M? what are the differences between Theorem E-S and Theorem EA?

E~13 What are the singular values of the following matrices?

A _

[1 -IJ
O 1
-1
1
-1]
-1
O 1

md HH* (Hint: Find a P

n Problem E-4

orthogonaI matrix) real'!


tdes eq uaI to 1.
clearly x is a right ei
the operator d on 1
obtain
d(XY
Hence ),¡ + J-lj is i
See Defini tion 2-12.

F Next we prove tha


that Y/k is an eigenvalu

On the Matrix Equation


or
AM+MB=N
We now show tha"
cornrnon. We prove

be the characteristic
Then we have

In this appendix we shall study the rnatrix equation AM + MB = N, where If Y/kI - A and B have
A, B, M, and N are n x n complex-valued matrices. We observe that all the
n x n complex-valued matrices, with the usual rules of multiplication and addi­
tion, form a linear space (Problern 2-10). Let us denote this space by (X, C). is nonsingular. Inde(
The dimension of (X, C) is n 2 • Consider the operator d defined by eigenvalues of ~(B)
If Y/kI- A and B ha'
d(M)~AM +MB ror aH M in X det ~(B) +0. Hence.
It is clear that the operator d rnaps (X, C) into itself and is a linear operator. From (F-1), we ca
The equation AM + MB = N is often caHed a Lyapunov matrix equation.

Theorem F-1
Let.si!: (X, C)->(;c C) b~ th.e open~üor denned by <s4{~~) = fi:J\!R ~- fA!1], f~,:- ~:.F r~,~: .'.
X. Let A¡, for i = 1,2, ... , 1.:s; n, be the distinct eigenvalues of A and let 11 j' for j = 1, The surnmation of the
2, ... , m ~ n, be the distinct eigenvalues of B. Then (A¡ +11) is an eigenvalue of with IX o = 1, yields
d. Conversely, let Y/k> k = 1, 2, . " . , p.:s; n2 , be the distinct eigenvalues of d,
then ror each k,
which, together with (

ror sorne i and sorne j.

Proof Since Ll(B) is nonsing


tion that M /=-0. Hen
We prove, first, that A¡ + I1j is an eigenvalue of d. Let x andy be nohzero nx 1 I
~ . . eigen vaJue. N ow the I
and Ix n vectors such that ror sorne i and for son
Ax =A¡X
572
ON THE MATRIX EQUATION 573

c1early x is a right eigenvector of A and y is a left eigenvector of B. Applying


the operator d on the n x n matrix xy that is c1early a nonzero matrix, we
obtain
d(xy) = Axy + xyB = },¡xy + Jijxy = (/'í + Jij)xy
Hence }'í + Jij is an eigenvalue of d for i = 1, 2, ... , 1, and j = 1, 2, ... , In.
See Definition 2-12.
Next \Ve prove that aH eigenvalues of d areofthe form }'í + Jlj. Let us assume
that r¡k is an eigenvalue of d. Then, by definition, there exists a M f. Osuch that

d (M) = AM + MB = r¡kM
or (r¡kI - A)M = MB (F-1 )

We now show that the matrices I1kI - A and B have at least one eigenvalue in
common. We prove this by contradiction. Let

be the characteristic polynomial oí I1kI - A, that is Ms) = det(sI -11kI + A).


Then we have
(F-2)

AM + MB = N, where lf I1kI - A and B have no common eigenvalue, then the matrix


le observe that aH the Ó(B)
ultiplication and addi­
te this space by (X, C). is nonsingular. lndeed, if Jiü i = 1, 2, ... , n, are eigen val ues of B, then Ó(¡.L¡) are
defined by eigenvalues of Ó(B) (Problem 2-32) and det Ó(B) = nÓ(Ji¡) (Problem 2-22).
lf I1kI - A and B have no common eigenvalue, then Ó(Ji¡) f.O, for aH i, and
X det Ó(B) f.O. Hence Ó(B) is nonsingular.
Id is a linear operator. From (F-l), we can develop the foHowing equalities
matrix equation.

(r¡,J - A)"M = MBn


AM +-ME for aH M in
)f A and let Jij, for j = 1, The summation ofthe products Of(I1kI- A)íM = MB í and IY.¡, for i =0,1,2, ... , n,
r- Ji) is an eigenvalue of with lY.o = 1, yields
inct eigenvalues of d,

which, together with (F-2), implies

O= MÓ(B) (F-3)

Since Ó(B) is nonsingular, (F~3) implies M =0. This contradicts the assump­
tion that Mi=- O. Hence the matrices 1] kI - A and B have at least one commori
and y be nonzero n x 1 eigenvalue. Now theeigenvalue of1]kI - A is ofthe form 17k - }e¡. Consequently,
for sorne iand forsome j,
or Q.E.D.
".!

574 ON THE MATRIX EQUATION

Corollary F-1 a Theorem F-2

Any matrix representation of the operator d is nonsingular if and only if Let s1: (X, C)->(X, 1[
A¡ +fJ-j=/=-O for aH i,j. M in 7... Let A¡, for i =
for j = 1, 2, ... , m ~
Proof eigenvalue of d. Co
values of .sd, then for
Since the linear operator .sd maps an n 2 -dimensional linear space into itself,
it has a matrix representation (Theorem 2-3). A matrix representation can be
easily obtained by writing the n 2 equations AM +MB =C in the form of
Aro = e, where ro is an n 2 x 1 column vector consisting of aH the n2 elements of for sorne i and sorne j
M. The coroHary follows directly from the fact that the determinant of A is the
product of its eigenvalues (Problem 2-22). Q.E. D. Corollary F-2a
Any matrix represenl
Corollary F-1 b
A¡fJ-j=/=-l for aH i,j.
Ir all the eigenvalues of A have negative real parts, then for any N there exists a
unique M that satisfies the matrix equation
Corollary F-2b
A*M +MA=-N (F-4)
lf aH the eigenvalues
Furthermore, the solution M can be expressed as exists a unique M tha

M = I'" eA"Ne At dt (F-5)

Furthermore, the solut


Proof
Since the eigenvalues of A * are the complex conjugates of the eigenvalues of A,
aH the eigenvalues of A and A* have negative real parts. Consequently,
A¡ + fJ-j=/=-O for aH i, j, which implies that the matrix representation of d(M) =
A*M +MA is nonsingular. Hence, for any N there exists a unique M satisfying
The procedure usel
d(M) = - N.
F-2. CoroHary F-2b (
Now we show that the M in (F-5) is the solution of (F-4). Because all the
three methods of solvi
eigenvalues of A have negative real parts, the integral in (F-5) converges.
CQusider ] By eoua.ting the .--;:.
obtain a set of Hne
A*M +MA= roo (A*eA'tNe AI +eA"NeA!A)dt= roo ~(eA'INeAI)dt consists of n 2 equat
1 Jo ~ the hermitian or Sy
A•
=e tNeA'[
It= 00

= - N
linear algebraic equ
1=0
a computing center
2. Compute the infinit
This establishes the coroHary. Q.E.D. have magnitudes Ir
infinite series can b,
Although the matrix M can be solved from Cf~5), the formula is not suitable
for computer computation. The solution of AM.+MB = N has been exten~' M,
sively studied. in ihe literature. The reader is referred to References S9, S 18; and
S107. . . . . .' . The computation m
We list in the rollowing the discrete-time version ofTheorem F-l. than a predeterrnim
ON THE MA TRIX EQUAnON 575

Theorem F-2

ngular ir and only ir Let stl: (x, C) -> (x, C) be the operator defined by d(M) = AMB - M for aH
M in X. Let )'i' for i = 1, 2, ... , /.::; n, be the distinct eigenvalues or A and let ¡lj'
for j = 1, 2, ... , m'::; n, be the distinct eigenvalues of B. Then (A¡!lj - 1) is an
eigenvalue of d. Conversely, let ¡lb k = 1, 2, ... ,p '::;n 2 , be the distinct eigen­
values of d, then for each k,
near space into itself,
representation can be {:=-S)
~ = e in the form of
f all the n2 elements of for some i and some j.
leterminant of A is the
Q.E.D. Corollary F-2a
Any matrix representation of the operator d is nonsingular if and only if
A¡!lr/= 1 for all i, j.
)r any N there exists a
Corollary F-2b
(F-4)
Ir all the eigenvalues of A have magnitudes less than 1, then for any N there
exists a unique M that satisfies the matrix equation

A*MA-M= -N (F-7)
(F-5)
Furthermore, the solution M can be expressed as
00

M = L (A*)kNAk
k~O
If the eigenvalues of A,
parts. Consequentiy, I
esentation of d(M) =
>a unique M satisfying The procedure used to prove Theorem F -1 can be used to establish Theorem
F-2. Corollary F-2b can be readily verified by direct substitution. We discuss
(F -4). Because all the three methods of solving (F-7) to conclude this appendix:
al in (F-5) converges.
L By equating toe corresponding elF.ment:o. 0f A*M.<I:\ - ¡"lIT ~.Dd·- i"{ WP; r:2 n
obtain a set of linear algebraic equations. Ir A is an n x n matrix, the set
consists of n 2 equations. The number of equations can be reduced by using
the hermitian or symmetry property of M (see Reference S16). The set of
linear algebraic equations can then be solved by using existing subroutines in
a computing center.
2. Compute the infinite power series in (F-8) directly. Ir all the eigenvalues of A
Q.E.D. have magnitudes much less than 1, the series will converge rapidly.. The
infinite series can ·be computed recursively as follows:
formula is not suitable
8 = N has been exten­ with M o =0
References S9, S18, and
The computation ma) be stopped when IIMk+l-' Mkll first becomes snialler
fheorem F-1. than a predetermined number.
576 ON THE MATRIX EQUATION

3. Define
Pk+1=Pl P1=A
M.k+ 1= PtMkP k + M k
By this process, the convergence in method 2 can be speeded up (see Refer­
ence S4).

Problems

F-1 Find a matrix representation of the operator $i'(M) = AM + MB with

A=[_~ -~J B=[~ ~J

and verify Theorem F-l.

F-2 Find the M to meet A*M +MA = - N with A given in Problem F-l and N = 1,

by setting up a set of linear algebraic equations.

F-3 Solve Problem F-2 by using (F-S). Which method, by solving algebraic equations

or by direct integration, is simpler?


lo this appendix, we ,
F-4 Establish, without using Corollary F-la, the uniqueness part of Corollary F-lb. sion to the polynomi
(H int: Use d(e A"(M 1 - M 2 )e A')jdt = O.) ioto two coprime poi
studied by the Freoch
F-5 Under what conditions can the solution ofAM +MB = - N be expressed as
system problems, ho\
M = t" eA'NeB1dt
the fractions of tran
lo the concepls of ce
F-6 Prove Theorem F-2 and its corollaries. hence its importance (
results in the scalar ca
F-7 Find a matrix representation of d(M) = AMB- M, with A and B given in Problem extended to the matrix
F-l and verify Theorem F-2.
for numerical comput
dix A will be constant
!o-s Salve (F-7) with A and N given in Problems F-l and F-2.

F-9 Transform A*M +MA = - N into a discrete-time Lyapunov equation by using the

G-1 Coprimene!
transformation
A-->(Ad - 1)(Ad + 1)-1 Let D(s) be a polyoomi
D(~

The polynomial D(s)


associated with the h
polyoomial is called n

1 In this appendix, capital le'


N
speeded up (see Refer­

G
+MB with
Polynomials and
Polynomial Matrices

Problem F-1 and N = 1,

Iving algebraic equations

In this appendix, we shall study the coprimeness of two polynomials, its exten­
part of Corollary F-1 b. sion to the polynomial matrix case and the factorization of a rational matrix
into two coprime polynomial matrices. The coprimeness of po.1ynomials was
studied by the French mathematician E. Bezout in 1764; 'its application to linear
N be expressed as system problems, however, was quite recent. The concept of coprimeness in
the fractions of transfer-function matrices is, roughly speaking, equivalent
to the concepts of controllability and observability in dynamical equations;
hence its importance cannot be overstated. In this appendix, the concepts and
results in the scalar case will be developed in such a way that they can be readily
\ and B given in Problem extended to the matrix case. The material will be presented in a manner suitable
for numerical computation. The row-searching algorithm discussed in Appen­
dix A will be constantly used.

IOV equation by using the G-1 Coprimeness of Polynomials

Let D(s) be a polynomial with real coefficients and indeterminate s expressed as 1


D(s) = Dns' +D._1s·- 1 + ... +D¡s +D o

The polynomial D(s) is said to be of degree n if D. =1=0. The coefficient Dn


associated with the highest power. of s is caBed the leading eoe,ffieient. A
polynomial is caBed monie if its leadingcoefficient is equal to 1.

'In this appendix, capital letterswithoutboldface are also used lo denote scalars.

577
578 POLYNOMIALS AND POLYNOMIAL MATRICES

Theorem G-1 5. Go to step 2.


Let D(s) and N(s) be two polynomials and let D(s) 1=0. Then there exist unique At the end, the result
polynomials Q(s) and R(s) such that In Equation (G-1:
by D(s), and D(s) is a.
N(s)=Q(s)D(s) +R(s)
a divisor of N(s), theI
and deg R(s) < deg D(s) (G-1 )
and N(s). Note that
D(s) and N(s), and is
Proof trivial common facto
If deg N(s) < deg D(s), then Q(s) =0 and R(s) = N(s). Let deg N(s) :::::deg D(s).
Then by direct division, we have the relation in (G-1) with Q(s) as the quotient Definition G-1
and R(s) as the remainder with deg R(s) < deg D(s). For example, if N(s) = A polynomial R(s) is :
2s 3 + 3s - 1 and D(s) = S2 + s - 2, then by long division, we have common divisor of D
2s -2 D(s) and N(s). Ir a gCI
S2 + s - 2)2s 3 + O + 3s - 1 then D(s) and N(s) an
2s 3 + 2s 2 - 4s
In other words, t\\
- 2s 2 + 7s-1 mon factors. Ir they
-2s2 - 2s +4 The gcd is unique
9s-5 for any nonzero numl
the gcd is uniq ue.
and N(s) = (2s - 2)D(s) +(9s - 5)
Given two polyno:
Now we show that Q(s) and R(s) are unique. Suppose there are other QI(S) called the euclidean a
and R 1(s) such that
N(s) =

N(s) =Q(s)D(s) +R(s)=Q¡(s)D(s) +R¡(s) D(s) =

which implies R¡(s)=

(Q(s) - Q ¡ (s))D(s) = R¡ (s) - R(s) (G-2)


R p - 2 (s) =
IrQ(s)-Q¡(s) 1=0, the degree of the left-hand-side polynomial of (G-2) is equal Rp_¡(s)=
to or larger than that of D(s), whereas the degree of the right-hand-side poly­
nomial of(G-2) is smaller than that of D(s). This is not possible. Hence we have This process will even

Q(s) =Q¡(s) and R¡(s) = R(s) From the last equ


This completes the proof of the theorem. Q.E.D. remainder, by Rp(s). '
(1 + QpQp+ t}Rp(s); he
ward, it can be ShOWI
The division procedure in this theorem can be put into a programmatic Rp(s). Hence Rp(s) is ;
format as follows: Now we claim th~
1. Q(s) =0. and Y;(s) such that R.
2. lf deg N(s) < deg D(s), stop. with X¡(s) = -Q¡(s) a
3. Let N m and D;, be the leading coefficients of N(s) and D(s) with m> n: equation shows that
N '.
N(s)+-N(s) --:-'D m s'"'-"O(s)
'Ir D" has avery small absol
" or this algorilhm. Hencc ¡
the gaussian elimination w
COPRIMENESS OF rOL YNOMIAlS 579

5. Go to step 2.
Jen there exist unique At the end, the resulting N(s) is the remainder R(S).2
In Equation (G-l) if R(s) =0, N(s) is said to be divisible (without remainder)
by D(s), and D(s) is afactor, or a divisor, of N(s). lf R(s) is a divisor of D(s) and
a divisor of N(s), then R(s) is called a common divisor or common factor of D(s)
(G-1 ) and N(s). Note that a nonzero constant is a common factor of every nonzero
D(s) and N(s), and is called a trivial common factor bf D(s) and N(s). A non­
trivial common factor wíll be a polynomial of degree 1 or higher.
:t deg N(s) ::?:deg D(s).
h Q(s) as the quotient Definitíon G-1
or example, if N(s) = A polynomial R(s) is a greatest cammon divisor (gcd) of D(s) and N(s) if R(s) is a
Ne have common divisor of D(s) and N(s), and is divisible by every common divisor of
D(s) and N(s). Ir a gcd of D(s) and N(s) is a nonzero constant (independent of s),
then D(s) and N(s) are said to be relatively prime or coprime. 1

In other words, two polynomials are coprime if they have only trivial com­
mon factors. Ir they have nontrivial common factors, they are not coprime.
The gcd is unique only up to a constant; that is, if R(s) is a gcd, then cR(s),
for any nonzero number e, is also a gcd. Ir we require a gcd to be monic, then
the gcd is unique.
Given two polynomials D(s) and N(s), by a sequence of long divisions, often
~ there are other Q¡(s) caBed the euclidean algorithm, we can write
N(s)=Q¡(s)D(s) +R¡(s) deg R¡ <deg D

1(s) D(S)=Q2(S)R¡(s) +R 2(s) deg R 2 < deg R¡

R ¡ (s) = Q3(s)R 2(s) + R 3(s) deg R 3 < deg R 2

(G-3)
(G-2)
R p_2(s)=Qp(s)R p_¡(s) + Rp(s)

lmial of (G-2) is equal Rp_1(s)=Qp+l(S)Rp(s) +0

ríght-hand-side poly­
;sible. Hence we have This process will eventually stop because the degree of R¡(s) decreases at each
step. \"1-/ e claim. that .R p(s) is ;3 gC(~ el .9(.-:;) aD.(~ /''1(.:.'>
From the last equation of (G-3) we see that Rp_¡(s) is divisible, without
Q.E.D. remainder, by Rp(s). The next to the last equation can be written as R p_2(S) =
(1 + QpQp + ¡)Rp(s); hence Rp- 2(S) is also divisible by Rp(s). Proceeding up­
ward, it can be shown that R p - 3 , . . . , R¡, D(s), and N(s) are all divisible by
into a programmatic Rp(s). Hence Rp(s) is a common divisor of D(s) and N(s).
Now we claim that for each R¡(s) in (G-3), there exist polynomials X¡(s)
and Y;(s) such that R¡(s) = X ¡(s)D(s) + Y;(s)N(s). ·This is clearly true for R 1(s)
with X ¡ (s) = - Q¡ (s) and Y¡ (s) = 1. The substitution of R ¡ (s) into the s~cond·
D(s) with m> n: equation shows that the claim holds for Rz(s). Proceeding downward, ·the

. .. .
'Ir D" has a very small absolute value, large errors may·arise on a digital eomputer implementation
or this algorithm. Henee thismethod may not be nllmerie~lIy stable. This situation is similar to
the gaussian elimination without any pivoting.
580 POLYNOMIALS AND POLYNOMIAL MATRICES

claim can be verified for every R í , i = 1, 2, ... ,p. Hence there exist polynomials Proof
X(s) and Y(s) such that 1. If D(s) and N(s) an
Ris) = X(s)D(s) + Y(s)N(s) (G-4) Hence the matrix
are not coprime, ti
This equation implies that every common divisor of D(s) and N(s) divides Ris). and the rank of [
Indeed if C(s) is a common factor, that is, D(s) = D(s)C(s) and N(s) = Ñ(s)C(s), has rank 1 at ever~
then we have Rp(s) = [X(s)D(s) + Y(s)Ñ(s)]C(s). Hence Rp(s) is divisible by its rank only at th
every common divisor of D(s) and N(s). Consequently, Ris) is a gcd. AH gcd 2. This follows direci
differ at most by a constant; hence every gcd can be expressed in the form of 3. If D(s) and N(s) ¡
(G-4). This is stated as a theorem. divisor, we obtain
coprime, there exi~
Theorem G-2 = 1. The substitu
Every gcd of the polynomials D(s) and N(s) is expressible in the form

R(s) = X(s)D(s) + Y(s)N(s)


which implies deg
where X(s) and Y(s) are polynomials.
We give a remar~
Although every gcd is expressible in the form of (G-4), the converse is not gcd of D(s) and N(s).
true. That is, a polynomial expressible as in (G-4) is not necessarily a gcd of which are roots of R(.
D(s) and N(s) (why?). s in iC. Hence if D(s)
rank 1 for almost al! 5
Theorem G-3 s in iC.
Consider two polynomials D(s) and N(s) with D(s) 1=- O. Then D(s) and N(s) are In the following, v
coprime if and only if any one of the foHowing conditions holds:
D(s) =
1. For every s in iC, the field of complex numbers, or for every reol of D(s), the N(s) =
2 x 1 matrix [D(S)] has rank 1.
N(s) and let
2. There exist lwo polynomials X(s) and Y(s) such that 3
X(s)D(s) + Y(s)N(s) = 1 (G-5)

3. There exist no polynomials A(s) and B(s) such that


N(s) B(s) with no assumption o
-- ---- (G-6)
D(s) A(s) Do DI Dz
or, equivalently, O Do Di

-B(s)D(s) +A(s)N(s)=[ -B(s) A(Sn[D(S)]=o (G-7) O O O .


N(s) S~ ------------­
- No NI N z

.and deg A(s) < deg D(s)


O No Ni

O O O
3i!' can be shown th~1 deg X(s) < deg N(s) and deg Y(s) < deg D(s)' (see Reference. SI25). This
properlY is nOl needed in lhis téxt.
COPRIMENESS Of POLYNOMIALS 581

:here exist polynomials Proof


1. Ir D(s) and N(s) are coprime, there is no s in iC such that D(s) = Oand N(s) =0.
(G-4) Hence the matrix [D(s) N(s)]' has rank 1 for every s in iC. lf D(s) and N(s)
are not coprime, there exists at least one s in iC such that D(s) = Oand N(s) = O
and N(s) divides R/s). and the rank of [D(s) N(s)]' is zero at that s. The matrix [D(s) N(s)]'
(s) and N(s) =Ñ(s)C(s), has rank 1 at every s except at the roots of D(s); hence it is necessary to check
e Rp(s) is divisible by its rank only at the roots of D(s).
R/s} is a gcd. AH gcd 2. This foHows directly from Equation (G-4).
pressed in the forrn of 3. Ir D(s) and N(s) are not coprime, by canceling their nontrivial common
divisor, we obtain A(s) and B(s) with deg A(s) < deg D(s). Ir D(s) and N(s) are
coprime, there exist polynomials X(s) and Y(s) such that X(s)D(s) + Y(s)N(s)
= L The substitution of N(s) = B(s)D(s)/A(s) into it yields
in the form
[X(s)A(s) + Y(s)B(s)]D(s) = A(s)
which implies deg A(s) 2>:deg D(s). This completes the proor. Q.E.D.
11
We give a remark concerning the rank of [D(s) N(s)]'. Let R(s) be the
4), the converse is not gcd of D(s) and N(s). Then the matrix [D(s) N(s)]' has a rank ofO at those s
ot necessarily a gcd of which are roots of R(s). The matrix however still has a rank of 1 at all other
s in iC. Hence if D(s) and N(s) are not coprime, the matrix [D(s) N(s)]' has
rank 1 for almost al! s in iC. If they are coprime, the matrix has rank 1 for al!
s in iC.
fhen D(s) and N(s) are In the following, we discuss a method of solving Equation (G-7). Let
IS holds:
D(s)=D o +DIs +Dzs z + +DlIs n (G-S)
every root of D(s), the N(s)=N o +NIs +Nzs z + +Nmsm (G-9)

and let

A(s)=A o +A¡s + +An_ls"-¡ (G-10)


(G-5)
B(s)=B o +B¡s + +Bm_¡sm-¡ (G-11 )

with no assumption of A II _ I f.O and Bm _ I f.O. Define


(G-6)

)(S)J =0
les)
(G-7) S ~
-
: '__ :_:__ :_:__ :::
No N¡ Nz
_::~: :~-_.
Nm
__
O
__ ::__: ,__ :::
O O
_:_.c_.__ :._ }
O O}
:w,
O No NI Nm - I Nm O O O O n
: rows
O 00 O O No NI Nm - I Nm
see Reference S125). This
( G-12)
582 POLYNOMIALS AND POLYNOMIAL MATRICES

It is a square matrix of order n + m and is called the Sylvester matrix or the r


resultant of D(s) and N(s). The substitution of(G-8) to (G-l1) into (G-7) and
equating the coefficients of Si, i =0,1, ... , n + m - 1, yields

[-B o -B¡ -B 2 ... -B m -¡ : A o A¡ A 2 '" An-¡JS=[ -B: AJS=O

l
(G-13)

This is a set of n +m linear algebraic homogeneous equations. We see that


the polynomial equation -B(s)D(s) + A(s)N(s) =0 has been transformed into
the equation in (G-13). Ifthe resultant S is nonsingular, then the only solution
of (G-13) is the trivial solution A(s) = B(s) = O. In other words, there exists no
A(s) of degree n -1 or less to meet N(s)!D(s) = B(s)/A(s); hence D(s) and N(s) are,
following Theorem G-3, coprime. Ir the resultant S is singular, a nontrivial
solution [-B AJ exists in (G-13). In other words, polynomials A(s) of
degree n -1 or less and B(s) exist such that N(s)!D(s) = B(s)!A(s), and D(s) and
N(s) are not coprime. Hence we have established the following corollary.

Corollary G-3
The polynomials D(s) and N(s) are coprime if and only if their Sylvester matrix Note that the matri~
S defined in (G-12) is nonsingular. 11 There are five nonzel
singular. Correspon,
six rows of F, by usir
Whether or not D(s) and N(s) are coprime hinges on the existence of a non­
trivial solution in (G-13). Mariy numerically stable meihods and "canned"
subroutines are available in the literature and computing centers to solve this
problem. However, we are interested in only the special solution which yields Using (G-I0), (G-ll),
the smallest degree in A(s). The row searching algorithm discussed in Appendix
A turns out to yield such a solution. This is illustrated by an example

Example 1 and
Consider the polynomials D(s) = - 2s 4 + 2s 3 - S2 - s +1 and N(S)=S3 +2s 2
-2s -1- 3. V;fe forrn thefr resultant~

1 -1 -1 2 -2 O O

O 1 -1 -1 2 -2 O
We note that the j
O O 1 -1 -1 2 -2
first linearly depender
S= ----------------------------- (G-14)
right-hand-side matri
3 -2 2 1 O O O

O -2 B(s) and ,1(s) of sma\'


3 2 1 O O

linearly dependent rO\


O O 3 -2 2 .1 O
(G-15). Thisis noi 1
O O O 3 -2 2. 1
(G-16) are coprime.
In the Sylvester 1m
We use the row searching algorithm discussed in Appendix A t() search the
in the ascending pOwe
linearly dependent rows bf s. The circled pivots are chosen as shown. The
descending power of 5
result is as follows:
COPRIMENESS Of POLYNOMIALS 583

'y/vester matrix or the 1


(G-ll) into (G-7) and
O 1
ls
O O 1
18=[ -B: AJ8=0 F:S= -3 -1 O 1 :8
(G-13) O -3 O -3 1

O O O -0.5 O 1

---------------------------------;..
luations. We see that O O 0.5 -1.5 -0.5 O
leen transformed into
then the only solution
(Ü -1 -1 2 -2
O O
words, there exists no
ence D(s) and N(s) are, O tI,
'"
-1 -1 2 -2 O
singular, a nontrivial O O 1 -1 -1 2 "'.. -=-- 2-'
- ~I
polynomials A(s) of O O 6 -4 4 (~~ O ~S (G-15)
~(s)/A(s), and D(s) and O O (~}D 17 -17 O O
lowing eorollary. O O O O O O O
O O O O O O O

their Sylvester matrix Note that the matrix F is the one defined in Equation (A-9) of Appendix A.
I There are five nonzero rows in S; henee the resultant S has a rank of 5 and is
singular. Corresponding to the first zero row ofS, we can obtain from the first
six rows of F, by using the recursive formula in (A-ll), the equation
he existence of a non­
ethods and "eanned" [1.5 0.5 O: -0.5 O 1 0]8 =0
g centers to solve this
solution whieh yields Using (G-I0), (G-ll), and (G-13), we have
discussed in Appendix
>y an example A(s) =S2 -0.5 B(s) = -0.5s -1.5
Nls) B(s) -0.5s -1.5 s +3
and ( G-16)
D(s) A{.S) s2-0.5 -2s 2 +1
2
1 and N(s) =S3 +2s
Hence D(s) ana N(s) are not coprime.
O
O We note that the B(s) and A(s) eomputed from [-B A]S =0 by using the
-2 first linear1y dependent row of S, which corresponds to the first zero row of the
(G-14 )
O right-hand-side matrix of (G-15), are coprime. Suppose not, then there exist
O B(s) andA(s) ofsmaller degrees to meet [-B A]8=0. This implies that a
linear1Y dependenirowofSwill appear before the appearance ofthe zero rows in
O
(G-15). Thisis nat possible. Hence we conclude that the B(s) and A(s) in
1 (G-,16) are cbprime.
:ndix A to searchthe In theSylvester matrix.in (G-12), the coefficients of D(s) and N(s) are arranged
losen as shown. The in the'ascending power of s. Clear1Y' we can also arrange the eoefficients in the
descendingpowef of s as
584 POLYNOMIALS AND POLYNOMIAL MATRICES

~l
DIl o o ... O It is a 2(k + 1) x (n + k
has two rows formed
and N(s). We note tI
~ ~ In
- - - - - - -:
m-l
~ ~ -~-o -~ _~n_ P~II~_I- - !J~1/----~ ~ ~ -~ ~o-J
- - - - - 1- -
right by one column.
Sk is a square matrix;
:: ::::: :: We search now lin
o o o o o Nm Nm- I NI No For convenience of di
(G-17) rows formed from Ni
linearly independent (
Using this Sylvester matrix, Corollary 0-3 still holds. However, if we use Dn ¡. O and the structü
(0-17) to compute A(s) and B(s), care must be taken in determining their degrees. elements aboye Dn in
To see this, we write is linearly independen
[-B A]S=[-B m _ 1 ..• -Bo:AI _1 ... Ao]S=O may not be linearly iI
Sk' we can readily see
In general, B m - I ¡. O. Hence the degree of B(s) is m -1, and consequently, previous rows, then al
the degree of A(s) is n - 1. However, if the degree of the gcd of D(s) and N(s) Hence the total numbe
is 2 or higher, then B(s) and A(s) have a common factor of form Sk. No such tonically as k increase~
problem will arise in using (0-12). Hence we shal1 arrange the coefficients of number of linearly in<
polynomials in ascending order throughout this appendix. how many more blod
Let v be the total
G-2 Reduction of Reducible Rational Functions words, the v N rows in
and al! N rows, not in
Consider a rational function N(s)jD(s) = N(s)D -I(S), where D(s) and N(s) on their previous row~
are polynomials.1f D(s) and N(s) are coprime, the rational fun"ction N(s)/D(s) dent; hence we have
is said to be irreducible. Otherwise, it is said to be reducible. In application, it
is often required to reduce a reducible rational function to an irreducible one.
Clearly the Sylvester matrix in (0-12) can be used for this reduction, as shown
in the example of the previous section. In this section, we shall modify the Since Sk is a 2(k + 1) x
procedure to improve its computability and to lay ground for the extension to A necessary condition
the matrix case. or k ::; n - 1. Hence w
Consider the two polynomials linearly independent 1
Let A(s) and B(s) b
( G-18~

and (G-19) and


N o assumption of N n ¡. O is imposed; hence the degree of N(s) can be smaller Then from the equatio
than n. We define, for k =0,1,2, ... ,
Do DI Dn - I Dn O O O
} 1 block row we can obtain, similar
No N1 Nn - I N Il O O O
Ó- - - D - - -6 - - - - - - - - D -_- - - D Ó- - - - - - - - - Ó­
- -
[ -B o
o 1 n I n

O No N1 Nn- I Nn O '"0 k + 1 block rows If Sk has a full row ral


the trivial. solution A(
O O O nontrivial solution wil
O O O degree among all A(s) 2
degree of A(s) to meet
( G-2ü) N rows in S •.
_. __ .------~--------~----~--
--~- - - --- -- - - ~ ~ - - ---------~ -- -- - ---- ---~-~~--

REDUCTION OF REDUCIBLE RATIONAL FUNCTIONS 585

~1
O It is a 2(k + 1) x (n + k + 1) matrix and consists of k + 1 block rows. Each block
has two rows formed from the coefficients, in the ascending power of s, of D(s)
and N(s). We note that each block is the shifting of its previous block to the
.,. DI DO
._---------­ right by one column. If k < n -1, Sk has more columns than rows; if k = n -1,
... O O Sk is a square matrix; if k > n - 1, then Sk has more rows than columns.
We search now linearly independent rows of Sk in order from top to bottom.
NI No J For convenience of discussion, the rows formed froin Di are calJed D rows; the
(G-17)
rows formed from Ni are called N rows. First we note that all D rows in Sk are
linearly independent of their previous rows. This follows from the assumption
However, if we use D" f.O and the structure of Sk' For example, if a new block is added to Sk> all
~rmining
their degrees. elements aboye D" in the last column of Sk+l are zeros; hence the new D row
is linearly independent of its previous rows. The new N row, however, may or
. Ao]S=O may not be linearly independent of its previous rows. From the structure of
Sk> we can readily see that once an N row becomes linearly dependent on its
-1, and consequently, previous rows, then all N rows in subsequent blocks wi.ll be linearly dependent.
e gcd of D(s) and N(s) Hence the total number of linearly independent N rows in Sk will increase mono­
. of forro Sk. No such tonically as k increases. However, once the number ceases to increase, the total
nge the coefficients of number of linearly independent N rows will remain to be the same no matter
ix. how many more block rows are added to Sk'
Let v be the total number of linear independent N rows in Soo' In other
:;tions words, the v N rows in SV-l are alllinearly independent of their previous rows,
and all N rows, not in the first v block rows of Sk> k;::: v, are linearly dependent
where D(s) and N(s) on their previous rows. Note that all (k + 1) D rows in Skare linearly indepen­
loal f~nction !v (s)/D(s) dent; hence we have
'ble. In application, it 2(k+1) fork':::;;v-1
rank S.= { (G-21 )
to an irreducible one. k (k +1) +v for k;:::v
is reduction, as shown
1, we shall modify the
Since Sk is a 2(k + 1) x (n +k + 1) matrix, if k.:::;; v - 1, then Sk has a full row rank.
A necessary condition for Sk to have a ful\ row rank is that 2(k + 1)':::;; n + k + 1
Id for the extension to
or k.:::;; n - 1. Hence we conclude that v':::;; n or, eq uivalently, the total number of
linearly independent N rows in Sk is at most equal to n.
Let A(s) and B(s) be two polynomials defined as
( G-18)
li(5; = Aa +- Al.) -l·' '"¡"·/~/{'·)
( G-19) (G-22)
and B(s)=B o +B 1 s+'" +Bkl
)f N(s) can be smaller Then from the equation
- B(s)D(s)+ A(s)N(s) =0 (G-23)

} l block row we can obtain, similar to (G-13),


[-13 o A o : -B 1 Al:"': -B k Ak]Sk=O (G-24)

k + l block rows IfS k has a full row rank, the only solution in (G-24) or, equivalently, (G-23) is
the trivial solutionA(s)=O; B(s) =0. Ask increases fromO,I, 2, ... ,the first
nontrivial solutionwill appear atk = v. Hence the v in (G-21) yields thesmallest
degree amongall A(s) and B(s) which satisfy (G-23). In other words, the smal\est
degree of A(s) to meet (G-23) is equal to the total number of linear independent
( G-20) N rows in Sv.
586 POLYNOMIALS AND POLYNOMIAL MATRICES

Theorem G-4 G-3 Polynomia


Consider two polynomials D(s) and N(s) with deg N(s)::s:deg D(s)=n. The
A matrix A(s) with
A(s) and B(s) solved from (G-24) by using the first linearly dependent row of Sk Similar to matrices \
are coprime, where Sk is defined as in (G-20). elementary operatior
1. M ultiplication of
Proof
2. Interchange any t,
Let v be the least integer such that the last N row ofS, is linearly dependent of its 3. Addition of the pr
previous rows. Since all D rows are linearly independent, this N row is the row or column.
first linearly dependent row in Sk, for k ;?: v. Corresponding to this dependent
These operations ca!
row, the solution of(G-24) yields an A(s)of degree v and B(s) such that N(s)/D(s) =
n = 5, of the form
B(s)/ A(s). Now if A(s) and B(s) are not coprime, there exist ,.4(s) of a degree
smaller than v and 8(s) such that 1 O O O O
N(s) _ B(s) _ 8(s)
O 100 O
D(s) - A(s) - ,.4(s)
El = O O 1 O O
O O O e O
This implies that a linearly dependent row appears before the last N row of
O O O O 1
S,. This is not possible. Hence A(s) and B(s) are coprime. Q.E. D.

We note that if k ;?:n, the resultant Sk has more rows than columns and sol u­ with e f= O and d(s) is
tions always exist in (G-24). For example, if k =n, then A(s) =D(s) and B(s) = elementary matrices ~
N(s) are solutions of (G-24) and (G-23); if k = n + 1, then A(s) =D(s)(s + e) and inverses are
B(s) = N(s)(s + e), for any real e, are solutions of (G-24). Clearly these solutions
1 O O O
are of no interest to uso
O 100
E¡-l = O O 1 O
Corollary G-4
O O O c- 1
The two polynomials D(s) and N(s) with deg N(s)::s: deg D(s) = n are coprime if O O O O
and only if the square matrix SIl- 1 of order 2n defined in (G-20) is nonsingular
or if and only ifthe total number of linear independent N rows in Sn-1 is equal
to n. III
They are again elem
operares on lÍle row~
This corollary follows directly from Theorem G-4, and its proof is left as an operates on the colum
exercise. From Theorem G-4, we see that the reduction of N(s)/D(s) hinges on and fifth row. E 3 A(s)
the search of the first linearly dependent row in Sk. The row searching algorithm fourth row of A(s); w]
in Appendix A is developed exactly for this purpose. It is illustrated in Example A(s) and d(s) to the S(
1 of Section G-1 and will not be repeated. operations and A(s)E i
The matrixSk has a special structure: Every block row is a shift ofits previous The set of polynor
block row. By using this shifting property, a very efficient method is developed plication is not a poly!
in Reference S140 to search tije first lin~arly dependent row of Sk' To fully tions, then the set beco
utilize the shifting property, theelimination :must be carried out from left to rational functions, the
right and the meihod is generally not numericálly stable. For hand calculation, developed for matrices
the method can definitely be used'in "place bf the ro\\' s~arching itlgorithm. equally applicable. 1
The result, however, may not be in an echelon form in the matrix case. ' field of rational functic
-~ - ---_ .. ------:---_._---~.- .. _.~-------~._ .. _-----------_.~-------,-------------_ .. _-~----_._-.-._----------_.-.~~ .. -- -.

POlYNOMIAl MATRICES 587

G-3 Polynomial Matrices

;).:::;;deg D(s) = n. The A matrix A(s) with polynomials as elements is called a polynomial matrix.
ly dependent row of Sk Similar to matrices with elements in IR 01' e, we may introduce the following
elementary operations on A(s):
1. Multiplication of a row 01' column by a nonzero real 01' complex number.
2. Interchange any two rows 01' two columns.
nearly dependent ofits 3. Addition of the product of one row 01' column and a polynomial to another
ent, this N row is the row 01' column.
jing to this dependent These operations can be carried out by using the elementary matrices, for
s)such that N(s)/D(s)= n = 5, of the form
exist A(s) of a degree
1 O O O O 1 O O O O I O O O O
O I O O O O O O O I O 1 O O O
EI = O O 1 O O Ez = O O 1 O O E3 = O O 1 O O
O O O e O O O O 1 O O des) O 1 O
fore the last N row of O O O O 1 O 1 O O O O O O O I
me. Q.E.D.
(G-25)

han columns and solu­ with e 1=0 and des) is a polynomial. We note that the determinants of these
A(s) =D(s) and B(s) = elementary matrices are nonzero constants and are independent of s. Their
nA(s)=D(s)(s+c) and inverses are
Clearly these solutions I O O
I O O O O O O
O 1 O O O, O 1 O O O
E-I I -
­ O O 1 O O E; I =E z E-l-
3 ­ O O I O O
O O O c- 1 O O -des) O 1 O
D(s) = n are coprime if O O O O 1 O O O O 1
I (G-20) is nonsingular
(G-26)
¡ rows in Sn -1 is equal
• They are again elementary matrices. The premultiplication. of 1E¡ DIi A(s)
operates Gil the rows 01' A(s), whereas túe postrnuiUpiicalioii 01 Ji<;¡ on A(s)
Id its proof is left as an operates on the columns of A(s). For example, EzA(s) interchanges the second
of N(s)/D(s) hinges on and fifth row. E 3 A(s) adds the product of the second row of A(s) and des) to the
)W searching algorithm
fourth row of A(s); whereas A(s)E 3 adds the product of the fourth column of
illustrated in Example A(s) and des) to the second column of A(s). We cal! EiA(s) elementary row
operations and A(s)E¡ elementary column operations.
is a shift of i ts previous The set of polynomials does not form a field beca use its inverse in multi­
t method is developed ' plication is not a polynomial. Ir we extend the set to include all rational func­
t row of Sk' To fully . tions, then the set becomes a field. Fo'r matrices with elements in 'thefield of
,nried out from left to rational functions, the concepts of linear independence, rank, and singulafity
For hand calcu¡'ation? ' developed for matrices with elements in the field of real 01' complex nUJ;nbers are
, searching algorithm. equally applicable. Hence,' ifwe c6nsiderp61ynomials as eiementsof the
le matrix case. field of rational functíons; then we may apply the concept of linear dependence
588 POLYNOMIALS AND POLYNOMIAL MATRICES

and rank to polynomial matrices. For example, the determinant of the poly­ k¡th
nomial matrix coturnn
J
... O al,k l
+2 s-I~J
al,k¡-l
S

[:
... O O O
[ s-1 S +2
O \1 O

is (s + 2)2 - (s - 1)2 = 6s + 3 which is not the zero element in the field of rational
O O O
functions. Hence, the matrix is nonsingular and has a full rank." The non­
singularity of the matrix in the field of rational functions does not imply that the O O O
matrix is nonsingular for all s in iC. For example, the matrix has rank 1, rather
than 2 at s = - 0.5.
Conversely, if the determinant of a polynomial matrix, which is a special
lf O O O

case of rational matrices, is equal to the zero element of lR(s), the field of rational The first r rows are
functions, then the polynomial matrix is singular. For example, the polynomial nOnzero element, a¡.ki
of ai,k; must be on th,
matrix
is, k¡ <k 2 < ... <k,.. ~
the degree of a¡,k;, tha
S +2
lf deg ai,k¡=O, then l
[ S2 +3s +2 these properties is sa
of (G-28) without th,
has a determinant of (s + 2)(S2 - 1) - (s - 1)(S2 + 3s + 2) = O. Hence the matrix just a upper right tria
is singular. Consequently, there exist rational functions a¡(s) and az(s) such lf A is a matrix v
that echelon (row)form. JI
row is 1 and all elen
(G-27) words, the leftmost J
column. By reversir
For this example, we may choose a¡(s) = 1 and a 2(s) = -1/(s + 1). obtain a different, bl
Let a(s) be the least common denominator of a ¡ (s) and a 2(s), and let IX ¡ (s) = appear in the remain(
a(sp¡(s) and &2(S) = a(s)cx2(S). Then (G-27) implies Every polynomia
ementary row operat

Triangularization Pn
where IX! (s) and 1X 2(S) are polynomials. Hence we conclude that polynomja 1
veetors are linearly dependent in the field of rational funetions if and only ir ;jtep 1. íVli(s)=Á\s) al
they ean be made dependent by using only polynomials as eoeffieients 4 (see nonzero.
Problem 2-9). Step 2. Ir all e1ement:
Let A(s) be a polynomial matrix with rank r in the field of rational funetions zero, go to step 6;
IR (s). We show that, by using exclusively elementary row operations, A(s) can Step 3. Search the ele
be transformed into the form and bring it to the I
element monie. Ce
S.tep 4. Compute j
Add the produet
M¡(s), i=2, 3, ...
4In DefinifiQn 2-2 ira field is replaced by a ring ~; (see Footnote 3 or chapter 2), then·(E{, 1Ki;l is
called·al1lod4l.e over.the ringo A fl X 1 or I x fl polynomial vector can be considered"as an demenl
~dditional column ope!
5 [[
or therational. vector space (W(s), ~(s)), or an elementar the modllle (IR "[s], ~[s]). A set or poly­
ioto a.diagonal matrix. ca
nomíal veciors is Hnearly independentover the field ~(s) ir and only ir the set ís linearly indeperident
In (G-28), we set, ror red\,
over thering ~[s]. See Relerence S34.
POLYNOMIAL MATRICES 589

terminant of the poly- klth kzth k 3 th k,th


column column column column
1
O O al.k, al,k 1 +1 al.k,-I al ,k, al,kz+ 1 al,k, Gl,k r + 1

O O O O O al,k 2 a2,kz+ 1 al.k,. a2,k,.+ 1


O O 0 O O O a3.k,

t_28l
a3,kr G3,k,.+ 1

in the field of rational


O O O O ... O O O
. full rank. The non­
ioes not imply that the O O O O O O O
a,..k r
O
G¡',krT 1

O
"'1
ltrix has rank 1, rather

rix, which is a special


O O O O O O O O O J
(s), the field of rational The first r rows are nonzero rows with polynomial elements. The left-most
ample, the polynomial nonzero element, a¡,k¡, of each row is a monic polynomial. The column position
of a¡,k¡ must be on the right hand side of the column position of Q¡ - \ ,k¡ _\' that
is, k 1 < k z < ... < k,.. 5 The degrees of all elements aboye a¡,k¡ are smaller than
the degree of a¡,k¡, that is, deg aj,k¡< deg a¡,k¡, for í = 2,3, ... , r;j = 1, 2, ... , í-1.
If deg a¡,k¡ = O, then aj,k¡ = O, for j = 1, 2, ... , í - 1. The matrix in (G-28) with
these properties is said to be in the Hermíte (row)form. A matrix in the form
of (G-28) without the property deg aj,k¡ < deg a¡,k¡ has no special name; it is
,O. Hence the matrix just a upper right triangular matrix.
s o:\(s) and 0:2(S) such If A is a matrix with real e\ements, the Hermite form is said to be in the
echelon (row)form. In this case, the leftmost nonzero e\ement of every nonzero
row is 1 and all elements aboye it are zero (becatise deg a¡,k¡ = O). In other
(G-27) words, the leftmost nonzero element 1 is the only nonzero element in that
column. By reversing the order of rows and the order of columns, we will
l/(s + 1). obtain a different, but equivalent echelon (row) formo This form will often
d o:z(s), and let C(I(S) = appear in the remainder of this appendix.
Every polynomial matrix can be triangularized by using exclusively el­
ementary row operations. This is presented as an algorithm.

Triangularization Procedure
lude that polynomial
nctions if and only if Step L M(s) = A(s) and oeieie lts cotUülllS ll'Oll1 leíl unlil Lfle úrsl COIUfIlll ¡S
ls as coefficients4 (see nonzero.
Step 2. If all elements, except the first element, of the first column of M(s) are
i of rational functions zero, go to step 6; otherwise, go to step 3.
'" operations, A(s) can Step 3. Search the element with the smallest degree in the first column of M(s)
and bring it to the (1, 1) position by the interchange of two rowS. Make the
element monic. Call the resulting matrix M I(S) = (mb(s))
Step 4. Compute m¡\(s)=qd(s)mll(s)+mh(s) with degmh(s)<degmll(s).
. Add the product of the first row of M 1 (s) and -Q¡I(S) to the ith row of
M 1(s), i = 2,3, ... ,n. Callthe resulting matrix Mz(s) = (m¡z¡(s)).
r chapter 2), then (.'r, IR ,) is
le cons'idered as an elemen(
~n[s], !R[s]). A sel of poly­
5 Ir addi;io~al column operations are employed, ~'e can always have k¡ = ¡ and lransfonu the matrix
into a diagonal matr.ix caUed the Smith formo The interested reader is referred lo References S34.
e set is linearly independent
In (G-28), we set, ror reducing lhe size of the matrix, k 3 =k , + l.
590 POLYNOMIALS AND POLYNOMIAL MATRICES

Step 5. M(s)~M2(s) and go to step 2. The product of the


Step 6. Delete the first row and the first column of M(s) and rename it M(s);
go to step 1. i

Steps 2 to 5 will reduce the degree of the (1, 1) element of M(s) by at least 1;
hence, after a finite number of iterations, we will go to step 6. We repeat the Hence, we have
process for the submatrices of M, and eventually we will transform A(s) into
the form in (G-28). This completes the triangularization of the matrix. ro -s-1:
In order to have the property deg aj.k, < deg ai,k¡, for j = 1, 2, ... , i - 1, we
need sorne additional row operations. Let aj,k¡ = qj(s)ai,k; + a],k; with deg a],k; <
deg ai,k¡. We add the product of the ith row and - qj(s) to the jth row, the
l~ ---;2~~ -~ i-:
resulting matrix will retain the form of (G-28) and has the property deg aj,k¡ < In this example,
deg ai,k;, j = 1, 2, ... , i - 1. Thus we have established the following theorem. stop at step 4. Dual
formed, by a seq uel
Theorem G-5 triangular matrix or I
Since the determi
Every polynomial matrix can be transformed into the Hermite row form in independent of s, so
(G-28) by a sequence of elementary row operations. 1 matrices are called UI~

Definition G-2
Example 1
A square polynomial
We give an example to illustrate this theorem. Consider is nonzero and indep(

A(')~[-~
-1
3dl
S2 +s - 2
S2 +25 -1
l -4 [-1s
-1
,'+,-2]
3s +1
52 +2s -1
~
[1O S3-,'-,+2
O
+s + 1
+S2
5 +1
] Theorem G-6
A square polynomial
nomial matrix.
[1
[~ [~ ~ +11 (G~9)
-s'-s+2 l -,' -s+2 ]
Proof
-J. O s+ 1 ~ 5 +1 ~

O S3 + 52 + 5 + 1 J O
Let M(s) be unimod
[Adj M(s)J/det M(s) i
In the first step, we interchange the first and second row. This is achieved by polynomial matrices.
the multiplicati.on of A(s) by the element8.ry matrix derroted by 1 listed 2.t th" eDr4 nomials. Sinee M(s)
of this paragraph. In the second step, we multiply the first row by - 1, add the The oníy way lor poly:
product of the first row and s to the second row and add the product of the first are both nonzero con:
row and - 1 to the third row. This is achieved by the matrix denoted by 2.
In the proof, we ha
In the third step, we interchange the second and third row. We then add the
unimodular matrix.
product of the second row and - (S2 + 1) to the third row in step 4. In the last
written as a product o
step, we add the product of the second row and s to the first row so .that the Unimodular matr
degree of a12(s) is smaller than that of all(s). The corresponding elementary tions. For every 5 in

matrices are . bers. In general, a nI

.of complex numbers f

5 4 3 2 1 which are the roots of

~] [~ ~ ~]O [- ~ ~ ~] [~O O~ ~]
5 O The ránk of a poly
1 1 or postmultiplied by
O _(S2 + 1) 1 O 1 -1 O 1 1 modular matrices.
POLYNOMIAL MATRICES 59!

The product of these five matrices is


;) and rename it M(s);

[~
I
( G-30)

t of M(s) by at least 1;
;tep 6. We repeat the Hence, we have
ill transform A(s) into
1 of the matrix. 3s + 1 ] I 1 2 l
r j = 1,2, ... , i - 1, we s2+ s -2 =1 __ ~ __ ~~I_-J (G-31)
+ a},k, with deg aJ.k; < S2 +2s -1 loo
!!l
(s) to the jth row, the
le property deg a¡,k; < In this example, if we are interested in only the triangularization, we may
he following theorem. stop at step 4. Dual to Theorem G-S, every polynomial matrix can be trans­
formed, by a sequence of elementary column operations, into a lower left
triangular matrix or the Hermite column form which is the transpose of (G-28).
Since the determinants of elementary matrices are nonzero constants and
Hermite row form in independent of s, so is the polynomial matrix in (G-30). Such polynomial
I matrices are called unimodular.

Definition G-2
A square polynomial matrix M(s) is called a unimodular matrix ifits determinant
is nonzero and independent of s. ii

2 Theorem G-6
-5 -S+2 ]

s3+ s2+ s +1
A square polynomial matrix is unimodular if and only if its inverse is a poly­
s+l nomial matrix.

•[i ~ + 1]
Proof
(G-29) Let M(s) be unimodular. Then det M(s) is a constant. Hence M - 1(.~) =
[Adj M(s)]/det M(s) is c1early a polynomial matrix. Let M(s) and M -l(S) be
polynomial matrices. Then, det M(s) = a(s) and det M -l(S) = bes) are poly­
'. This is achieved by
nomials. Since M(s)M - I (s) = 1, we have det M(s)' det M- 1 (s) = a(s)b{s) = J
:d by lUsted at the ene!
'st row by - 1, add the The oniy way for polynomials a(s) and bes) to meet a(s)b(s) = 1 is that a(s) and b(s)
are both nonzero constants. Hence, M(s) and M - l(S) are unimodular. Q.E. D.
the product of the first
matrix denoted by 2.
In the proof, we have shown that the inverse of a unimodular matrix is also a
ow. We then add the
unimodular matrix. It can be shown that every unimodular matrix can be
v in step 4. In the last
written as a product of elementary matrices in (G-2S).
e first row so that the Unimodular matrices are c1early nonsingular in the field of rational func­
'espónding elementary tions. For every s in e, they are also rionsingular in the field of complex num­
bers. In general, a nonsingular polynomial matrix is nonsingular in the field
of complex numbers for almost aIl S in iC. It becomes singular only at those s
2 1 which are the roots of its determinant. . .

,! ~][! HJ

The rank of a polynomial matrix in !R(s) will not challge if it is premultiplied


or postmultiplied by nonsingular polynomial matrices, in particular, uni­
modular matrices.
592 POLYNOMIALS AND POLYNOMIAL MATRICES

G-4 Coprimeness of Polynomial Matrices


where R(s) is an UPI
In this section, the concept of coprimeness for scalar polynomials will be extend­ (G-28). The form of
q rows of the right-r
ed to polynomial matrices. Since the multiplication of matrices does not com­
ract that the compc
mute in general, the situation here is more complicated.
Hence it has a rank 01.
Consider A(s) = B(s)C(s), where A(s), B(s), and C(s) are polynomial matrices
it has at least q linear!:
of appropriate orders. We call C(s) a right divisor of A(s) and A(s) a left multiple
row operations, it is (
of C(s). Similar!y we call B(s) a left divisor of A(s) and A(s) a right multiple of B(s).
Since U(s) is unimc
Consider two polynomial matrices N(s) and O(s). The square polynomial
matrix R(s) is called a common right divisor of N(s) and D(s) if there exist poly­ U 11
nominal matrices N(s) and D(s) such that [ U Z1
N(s) = Ñ(s)R(s) O(s) = D(s)R(s) (G-32) Then we have
In this definition, N(s) and O(s) are required to have the same number of columns.
Their numbers of rows, however, can be different.
and [
Definitíon G-3
Hence, R(s) is a como
A square polynomial matrix R(s) is a greatest common right divisor (gcrd) of
N(s) and O(s) if R(s) is a common right divisor of N(s) and O(s) and is a left
multiple of every common right divisor of N(s) and O(s). Ir a gcrd is a uni­
modular matrix, then N(s) and O(s) are said to be right coprime. I!l Let R¡(s) be any comr
and N(s) = Ñ(s)R 1 (s).
Dual to this definition, a square polynomial matrix Q(s) is called a greatest [U 11(S)0(S) + U ds)t\
. common left .divisor (gcld) of A(s) and B(s) if Q(s) is a common left divisor of A(s) we conclude that R(s)
and B(s) [that is, there exist polynomial matrices A(s) and B(s) such that A(s) =
Q(s)A(s), 8(s) = Q(s)B(s)] and Q(s) is a right multiple of every common left Example 2
divisor Ql(S) of A(s) and B(s) [that is, there exists a polynomial matrix W(s)
such that Q(s) = Ql(S)W(S)]. Find a gcrd of the po:
Now we shall extend Theorem G-2 to the matrix case.
<;
O(s) = "
[ -1
Theorem G-7
Consider the p x p and q x p polynomial matrices O(s) and N(s). Then they
llave a gcrd R(s) expressible In. th.e forrri
·
From (G-30), we have

O -s-1
R(s) = X(s)O(s) + Y(s)N(s)
where X(s) and Y(s) are p x p and p x q polynomial matrices, respectively.

Proof
l O
-( -. -1
-sy+~-+

Hence a gcrd of O(s) ,

We form the composite polynomial matrix [O'(s) N'(s)]', where the prime de­
notes the transpose. Then Theorem G-S implies that there exists aunimodular
matrix U(s) such that
which is not a unimód
p[U 11 (s) U 12(S)][D(S)] = [R(S)]}P (G-33)
q U Z1 (s) Uds)_ N(s) O}q Let W(s) be any P.
N(s). Then W(s)R(s)
U(s) premultiplication of t
COPRIMENESS Of POLYNOMIAL MATRICES 593

where R(s) is an upper triangular polynomial matrix of the form shown in


(0-28). The form of R(s) is immaterial here. What is important is that the last
omials wil1 be extend­ q rows of the right-hand-side matrix are zero rows. This fol1ows from the
atrices does not com­ fact that the composite matrix [D'(s) N'(s)]' is of dimension (q +p) x p.
Hence it has a rank of at most p (in the field of rational functions). Consequently,
: polynomial matrices it has at least q linearly dependent rows. Therefore, by a seq uence of elementary
md A(s) a left multiple row operations, it is always possible to achieve (0-33).
a right multiple ofB(s). Since Ves) is unimodular, its inverse is a polynomial matrix. Let
he square polynomial
)(s) if there exist poly­ VileS) UdS)]-¡ ~[Vll(S) Vds)]~V(S) (G-34)
[ v 2 ¡(s) Uds) - v 2 ¡(s) Vds) ­
(G-32) Then we have

le number of columns. D(S)] =


[ N(s)
[V
11 (s)
V 2¡(S)
V 12(S)][R(S)]
Vds) O
(G-35)

and D(s) = V ¡¡(s)R(s) N(s)=V 2 ¡(s)R(s)


Hence, R(s) is a common right divisor of D(s) and N(s). From (G-33) we have
right divisor (gcrd) of
and D(s) and is a left R(s)=Ull(s)D(s) +Vds)N(s) (G-36)
¡). lf a gcrd is a uni­
:oprime. I Let R¡(s) be any common right divisor of D(s) and N(s); that is, D(s) = D(s)R¡(s)
and N(s) = N(s)R ¡ (s). The substitution of these into (G-36) yields R(s) =
I(S) is called a greatest [U 11 (s)D(s) + Vds)N(s)]R¡(s); that is, R(s) is a left multiple of R¡(s). Hence,
non left divisor of A(s) we conc1ude that R(s) is a gcrd.. This establishes the theorem. Q.E.D.
:i B(s) such that A(s) =
)f every common left Example 2
.Iynomial matrix W(s)
Find a gcrd of the polynomial matrices

D(s)=[_~ N(s) = [-1 S2 +2s-1J

and N(s). Then they From (G-30), we have

J'l º j
'ices, respectively.
l 9_ ----:=-_1_ - - - - ~ --- -- ! --- IL~~:~] = [~~~)] =
;0
1
-s-1
S2 +s + 1
Hence a gcrd of D(s) and N(s) is
:
: - (S2
,
s
+ 1)
'1
__
O
L
X±- L
O

]', where the prime de­


ore exists a unimodular R(s) = [~
which is not a unimodular matrix. Hence D(s) and N(s) are not right coprime. I
p
(G-33)
q Let W(s) be any p x p unimodular matrix, and let R(s) be a gcrd'of D(s) and
N(s). Then W(s)R(s) isalso a gcrd of N(s) and O(s). This can be proved by
premultiplication of the unimodular matrix diag{W(s), I q } to (G-33). Hence
594 POLYNOMIALS AND POLYNOMIAL MATRICES

the gcrd of 0(.'1) and N(s) is not unique. For example, the polynomial matrix

W(S)=[i .'1:
1
n has rank p (in the
2. There exist polyr
is unimodular for any positive integer k; hence such that
R¡(S)=W(S)R(S)=[l+1
. Sk
lIJ[OI 2J=[l+1
.'1+1 i
k
2s k+S+3J
2s +s+l
This is cal1ed the 1
is also a gcrd of the 0(.'1) and N(s) in the example. We see that the degrees of 3. There exist no po
the elements of R¡(s) may be larger than those ofO(s) and N(s). This phenome­ such that B(s)D(s)
non can never arise in the scalar case.
Let R¡(s) and R 2 (s) be two different gcrds of 0(.'1) and N(s). Can they always -B(
be related by a unimodular matrix? The answer is affirmative if the matrix
[O'(s) N'(s)]' is of full column rank. and

Corollary G -7
Ir [0'(.'1) N'(s)]' is of ful1 column rank, in particular, if 0(.'1) is nonsingular, Proof
then aH gcrds of 0(.'1) and N(s) are nonsingular and are related by unimodular
matrices. l. For convenience \\

Proof
From (G-33), if [0'(.'1) N'(s)]' is of full column rank, so is [R'(s) O')'. Hence,
R(s) is nonsingular. Let R¡(s) be any gcrd ofO(s) and N(s). Then by definitlon, For every s in e, t
we have two polynomial matrices W ¡(s) and W 2(.'1) such that complex numbers.
same as the rank ,
R(s) = W ¡(s)R¡(s) R¡(s) = W 2(s)R(s). unimodular and h.
rank p for every s i
which imply is a polynomial o
following the nonsi
R(s) = W ¡(s)W2(s)R(s) (G-37)
such that det R(s)
Since R(s) is nonsingular, we have W ¡(s)W 2(.'1) = l. Hence, both W ¡(s) and Hence, weconclud.
W 2(s) are unimodular matrices. Consequently, RJ(s)=Wz(s)R(s) i5 also of [D'(s) N'(s)]' i5
nonsingular. Q.E.u. uJ. tút ·úrsL Slaltll1(
condition is met at
In application, we often have the condition that 0(.'1) is nonsingular. With check the rank onl
this condition, the condition in Corollary G-7 is always mel. In this case, the 2. To show the secon
gcrd of 0(.'1) and N(s) is unique in the sense that all gcrds can be obtained from
a single gcrd by premultiplying unimodular matrices.
Ir 0(.'1) and N(s) an
Theorem G-8 nomial matr.ix. T
Let O(s) and N(s) be p x p and q x p potynomial matrices, and let 0(.'1) be yields (G-38) with :
nonsingular. Then 0(.'1) and N(s) are right coptime if and only if any one of the Conversely, we
'following conditions holds: and N(s). Lét R(s
N(s) = N(s)R(s). T
1. For every s in e, or fofevery root of·thedeterminant of 0(.'1), the (p +q)x p
mátrix
COPRIMENESS OF POLYNOMIAL MATRICES 595

he polynomial matrix D(s)J


[ N(s)
has rank p (in the field of complex numbers).
2. There exist polynomial matrices X(s) and Y(s) of order p x p and p x q
such that
k
2s +s+ 3J X(s)D(s) +Y(s)N(s) = 1 (G-38)
2s k +s+ 1
This is cal1ed the Bezoui ideniiiy in References S34 and S125.
:ee that the degrees of 3. There exist no polynomial matrices B(s) and A(s) of order q x p and q x q
N(s). This phenome­ such that B(s)D(s) = A(s)N(s) or, equivalently,

lI(s). Can they always


lrmative if the matrix
- 8(s)D(s) + A(s)N(s) = [ - 8(s) A(s)] [~~:Q =0 (G-39)

and deg det A(s) < deg det D(s).

f D(s) is nonsingular, Proof


elated by unimodular
1. For convenience we rewrite (G-33) in the following

U 11 (s) Uds)J[D(S)] = [R(S)] (G-40)


[ U 11 (S) Uds) N(s) O
s [R'(s) O'J'. Hence,
). Then by definitiún, For every s in C, the unimodular matrix U(s) is .nonsingular in the field of
hat complex numbers. Hence, for every s in C, the rank of [D'(s) N'(s)]' is the
same as the rank of R(s). Ir D(s) and N(s) are right coprime, then R(s) is
;). unimodular and has rank p for every s in C; hence [D'(s) N'(s)J' also has
rank p for every s in C. Ir D(s) and N(s) are not right coprime, then det R(s)
is a polynomial of degree one or higher [Note that R(s) is nonsingular
(G-37)
following the nonsingularity of D(s)]. Therefore, there is at least one s in C
such that det R(s) =0 or, equivalently, the rank of R(s) is smaller than p.
:nce, both W 1(s) and Hence, we conc1ude that if D(s) and N(s) are not right coprime, then the rank
) = W z(s)R(s) is also of [D'(s) N'(s)]' is smaller than p for sorne s in e This completes the proof
Q.E.D. oi" ihe nrst statement. Since D(s) is noúsiúguiü( ay assúmp~¡0li, ,fié ,ilíü(
condition is met at every s in C except the roots of det D(s) =0. Hence we
is nonsingular. With check the rank only at the roots of det D(s) = O.
net. In this case, the 2. To show the second statement, we write the first p equations of (GAO) as
can be obtained from
U 11 (s)D(s) + U ds)N(s) = R(s) (G-41 )

Ir D(s)andN(s) are right coprime, R(s) is unimodular, and R -1(S) is a poly­


nomial. matrix. The premultiplication of R - I(S) on both sides of (G-41)
ices, and let D(s) be yields (G~38) with X(s) = R -1(S)U ll (s) and Y(s) = R -1(S)U 12 (s).
only if any one of the Cónversely, we show that (G-38) implies the right copdmeness 'of D(s)
and N(s). Let R(s) be a gcrd of D(s) and N(s); that is, D(s) = D(s)R(s) and
N(s)=Ñ(s)R(s). The substitution of these into (G-38) yie\ds
)f O(s), the (p +q) x P
[X(s)D(s) + Y(s)Ñ(s)] R(s) = 1
596 POLYNOMIALS AND POLYNOMIAL MATRICES

which implies The substitution o


R -1(S) = X(s)D(s) + Y(s)N(s) = polynomial matrix
It follows írom Theorem G-6 that R(s) is unimodular. Hence, D(s) and N(s)
are right coprime. Since V(s) is unimc
3. We write the bottom q equations of (G-40) as
U 21 (S)D(s) + Uds)N(s) = O This relation hold~
This is already in the form of (G-39) if we identify B(s) = - U 21 (s) and A(s) = U- 1(s). We shall
U 22 (s). Hence what remains to be proved is the following inequality: Theorem G-S. Ir
R(s) with deg det
deg det A(s) = deg det U 22 (s) < deg det D(s)
x det R(s), which ir
From (G-34) and (G-35), we have
D(s) = V 11 (s)R(s) (G-42)
Because of (G-47)
Hence, if D(s) is nonsingular, so are V II(S) and R(s). Using the identity D(s).
Conversely, if dI

[1-V 21 (S)V¡/(s)
O][V
1
11 (s)
V 21 (S)
VdS)]
Vds)
ment and conclude

In the following, w
= [V
OII(S) VI2 (S) ] (G-43) left coprime. Given t\
V 22(S) - V21(S)V ¡/(s)V ds) of rows, then, similar
formations so that
we can write the determinant of V(s) defined in (G-34) as
det V(s) = det V II(s)det [V ds) - V 21(S)V ¡/(s)V ds)] "1=0 (G-44)
q[Q(s
for all s in C. Taking the inverse of (G-43) yields q
Based on this, we have
V 11 (S) VdS)]-I[ 1 0J-l [V 11 (S) ~dS)]-1 (G-45)
[ V 21 (S) Vds) -V 21 (S)V¡/(s) 1 = O l..}. Theorem G-S'

where t.=V 22 (S)-V 21 (S)V¡II(S)Vds) is nonsingular following (G-44). Let A(s) and 13(s) be q
The inverse of a triangular matrix is again triangular and can be readily singular. Then A(s) an
computed. After computing the inverse of the right-hand-side rnatrix of ing condition.s holds:
(G-45), we [hen move the second inverse on the left-hand side of the equaiíty 1. For every s in e, or
in (G-45) to the right-hand side: matrix

V 11 (S) VdS)]-1 = [Vi}(S) -V¡/(S)VdS)/}.-IJ[ 1 0IJ


[ V (S) Vds) o /}.-1 -V 21 (S)V¡l(s)
21 has rank q in the fie
2. There exist polynon
that

where X denotes elements which are not needed in the rollowing. The 3. There exists no poi:
comparison of this equation with (G-34)yidds such that

Ui2(S) = /}.- 1 = [V 22(S) - V 21 (s)V ¡/(s)V 12(S)] - 1


COPRIMENESS OF POLYNOMIAL MATRICES 597

The substitution of this equation to (G-44) yields

matrix
det Ves) = det V 11 (s) (G-46)
det Uds)
Hence, D(s) and N(s)

Since Ves) is unimodular, we have deg det Ves) =0. Hence, (G-46) implies
deg det V 11 (s) = deg det Uds) (G-47)

This relation holds for any unimodu.lar matrix U(s) and its inverse Ves) =
= - U 21 (S) and A(s) = U- 1(s). We shal1 now use (G-47) and (G-42) to establish statement 3 of
ing inequality: Theorem G-S. Ir D(s) and N(s) are not right coprime, there exists a gcrd
R(s) with deg det R(s»O. From (G-42), we have det D(s)=det V 11 (S)
D(s) x det R(s), which implies
deg det D(s) > deg det V 11(S)
(G-42)
Because of (G-47) and A(s) = - Uds), we conclude deg det A(s) < deg det
Jsing the identity D(s).
Conversely, if deg det A(s) < deg det D(s), we may reverse the aboye argu­
ment and conclude that D(s) and N(s) are not right coprime. Q.E.D.

In the fol1owing, we develop a dual ofTheorem G-S for matrices which are
(G-43) left coprime. Given two polynomial matrices A(s) and B(s) ofthe same number
of rows, then, similar to (G-33), there exists a sequence of elementary trans­
formations so that
as
q p .
,-.J'--.. ,-.J'--..

5')V 12(S)] i=- O (G-44)

q[Q(s) O] = q{[A(s) B(sn[V 11 (s) V 12(s)J}q


~ V 21(S) Vds) }p
q P q p
Based on this, we have the fol1owing.
(G-45)
Theorem G-S'

tr fol1owing (G-44). Let A(s) and B(s) be q x q and q x p polynomial matrices and let A(s) be non­
singular. Then A(s) and B(s) are left coprime if and only if any ane of the fol1ow­
r and can be readily
-hand-side matrix of ing condítions holds:
Id side oftheequality 1. For every s in ic, or for every root of the determinant of A(s), the q x (q + p)
matrix

1
- V2l(s)V¡/(s)
0J
1 [A(s) B(s)] (G-48)

has rank q in the field of complex numbers.


2. There exist polynomial matrices X(s) and Ves) of order q x q and px q such
ili~ .

A(s)X(s) +B(s)V(s) = 1
the fol1owing. The 3. There exists no polyriomial matrices N(s) and D(s) of order q x parid p x p
such that
A(s)N(s) = B(s)D(s)
598 POLYNOMIALS AND POLYNOMIAL MATRICES

or, equivalently, singuar and deg det J


are right coprime.
- A(s)N(s) + B(s)D(s) = [A(s) B(s)] [- ~~;~] =0 (G-SO)

and deg det D(s) < deg det A(s). G-S Column- al
We give the fo\lowing coro\lary to conclude this section. Consider two polyno
singular, then the mat
Corollary G-S given a q x p rational
Let D(s) and N(s) be p x p and q x p polynomial matrices and let D(s) be non­
singular. Let U(s) be a unimodular matrix such that
or
U(S)[D(S)]~[U11(S) U dS)][D(S)] = [R(S)]
N(s) - U 21 (S) Uds) N(s) O where N(s), D(s), A(s:
Then we have polynomial matrices.

~
~ d~
1. U 22 (S) and U 21 (S) are left coprime.
r'

13

2. Uds) is nonsingular and N(s)D- 1(s) = - Ui}(s)U 21 (s). d 11 l2 dn 13 '


3. D(s) and N(s) are right coprime if and only if deg det D(s) =deg det Uds). n21
Proof
-
d 21
n 22
-
d 22
n23
--

d23 J
.

Since U(s) is unimodular, it has rank p +q for every s in C. This implies that
for every s in e, its submatrix [U 21 (S) U 22 (S)], aq x (p + q) polynomial matrix,
has rank q. Hence U 22 (S) and U 21 (S) are, fo\lowingTheorem G-S', left coprime.
We show the nonsingularity of U 22 (S) by contradiction. Suppose U 22 (S) is
not nonsingular, then there exists a 1 x q polynomial vector a(s), not identically
zero, such that
a(s)Uds)=0
which, together with U 21 (S)D(s) + Uds)N(s) =0, implies
a(s)U 21 (S)D(s) =0 where nij and d ij are r
um¡-
l . . 1 of fe.
11 "0
l-:<LI-.
1
'-' l ~"
:uT\'s'J"I ah.
(;(S). These fractions .
a(s)U 21 (S) =0 not right coprime and
A rational functiol
Hence we have a(s)[U 21 (S) Uds)] =0. This contradicts with the fact that 3-5). In terms of th{
[U 21 U 22 ] has rank q in the field of rational functions. Hence we conclude determined. For exa
that U 22 (s) is nonsingular. Consequently, from U 21 D + U 22 N =0, we have the degree of the nun
to that of its denomil
situation is more com
Part 3 ofthis theorem has been essentia\ly est~blished in the proof ofTheorem Given a polynomia
G-S. Indeed, if D(s) and N(s) are right ¿oprime; then R(s) in (G-42) is unimodular power of s in a\l entrie
and deg detD(s) =deg det V 11(S). This, together with (G-47), implies .deg det bciM(~
D(s) = deg det U~2(S), Conversely; if deg det D(s) = deg det Uds), we. may
briM(~
reverse the above argument to conclude degdet R(s) = O. SinceRes) is non­
COLUMN AND ROW-REDUCED POLYNOMIAL MATRICES 599

singuar and deg det R(s) =0, R(s) must be unimodular. Hence D(s) and N(s)
are right coprime. Q.E. D.
N(S)] =0 (G-50)
D(s)

G-S Column- and Row-Reduced Polynomial Matrices


on. Consider two polynomial matrices N(s) and D(s). Ir D(s) is square and non­
singular, then the matrix N(s)D- 1 (s) is generally a rational matrix. Conversely,
given a q x p rational matrix C(s), we can always factor G(s) as
s and let O(s) be non- C(s) = N(S)0-1(S) (G-51 )

or C(s) = A- 1 (s)B(s) (G-52)

where N(s), D(s), A(s), and B(s) are, respectively, q x p, p x p, q x q, and q x p


polynomial matrices. For example, we have

n1Z

l~
nlZ
d ll d 12 l3 =~el -
d ez
~e3 =[~ll J
""1 [" ""1
o). n1Z
d ñ
13
D(s) =deg det Uds). nZ1 n22 n23 nZ1 n22 n23 n21 nZZ nZ3
- - - -
d Z1 d22 d Z3 del d ez d e3

O
iC. This implies that
q) polynomial matrix,
rem G-8', left coprime. l"
x O
O
d ez
O
0l'
O .
dd
(.G-53)

)fi. Suppose Uzz(s) is

~l[dd Tlñ ñ ~l3J


:or a(s), not identically u ñ 1Z

lññd"1 d"1
ñ 22 n Z3 O
O
d,·z
ll

ñ Z1 ñ 22
12
n23
Z1 -
d,.z d,.z d,·z

where nij and d ij are polynomials, dei is the least common denominator of the
ith column of (;(s), and el,., is the least common denominator of the ¡th row of
=0 implies C(s). These fractions are easy to carry out; however, N(s) and D(s) are generally
not right coprime and A(s) and B(s) are generally not left coprime.
A rational function C(s) is called strictly proper if C((0) < 00 (see Definition
icts with the fact that 3-5). In terms of the elements of C(s), the properness of C(s) can be easily
. Hence we conc1ude determined. For example, the rational matrix C(s) is proper if and only if
-UzzN =0, we have the degree of the numerator of every element of C(s) is smaller than or equal
to that of its denominator. In terrns of the fractions in (G-51) or (G-52) the
situation is more complicated. 'We shall study this problem in this section.
1 the proofofTheorem Given a polynomial column or ro.w vector, its degree is definedas the highest
n (G-42) is unimodular power of s in all entries of the vector.. We define
G-47), implies degdet
b~J"l(s) = the degree of the ith column of M(s) .
:g det Uds), we may
o. Since R(s) is non- b,.¡M(s) = the degree of the ith row of M(s)
600 POLYNOMIALS AND POLYNOMIAL MATRICES

Definition G-4
and call b ei column degree and by¡ row degree. For exarnple, for
S3 +2s +1 A nonsingular p x p poIy
M(s)=[S+l

s-l S3

(G-54)

we have b el = 1, b e2 = 3, b e3 = 1, and brl = 3, b r2 = 3. [t is called row reduced iI

Theorem G-9
Ir (;(s) is a q x p proper (strictly proper) rationalmatrix and if (;(s) = N(s)O -1(S)
=A -1(s)B(s), then The rnatrix O(s) in (1
0< b el O(s) +bezO(s) = 2
bciN(s) ~bciD(s)
colurnn reduced but not
for i = 1,2, ... , p, and

brjB(s) ~brjA(s)
for j = 1, 2, ... ,q. is colurnn reduced but
rnatrix is always both co]
Proof Let beiM(s) =k ei. Th,
We write N(s) = (;(s)O(s). Let niis) be the ijth elernent of N(s). Then we have
p
where Hc(s) = diag{lei, i'
nij(s) = L
k= 1
gik(S)dds) i = 1,2, ... , q the column-degree coeiTic
colurnn of M(s) associatel
Note that, for every element in the jth column of N(s), the suÍnrnation is carried rernaining terms and its i­
over the jth column of O(s). Ir G(s) is proper, the degree of ni)s), i = 1,2, ... , q, the M(s) in (G-56) can be
is srnal1er than or equal to the highest degree in dkj(s), k = 1. 2, __ . ,p. Hence
we have
M(s)
j = 1, 2, ... , p
[n terrns of (G-57), we ha
The rest of the theorem can be sirnilarly proved. Q.E.D.
det M(s) = (det 1\
We showed in Theorem G-9 that if G(s) = N(s)KY- '(s) is strictly proper..
then column degrees of N(s) are smaller than the corresponding column degrees nence, we conclucie chat
of D(s). lt is natural to ask whether the converse is also true. In general, the coefficient matrix M',e is I
answer is negative, as can be seen from Similar to (G-57), we

S2
N(s) = [1 2] D(s)= [ s+l (G-55)
where H,.(s) = diag{sk r i , i'
row. MIli_ wil1 be called tt
where bciN(s) < beiO(s), i = 1; 2. However, we have
ficients of {he ¡th row c
N(S)O-l(S) = [-2S1 -1 M1r(s) contains the remail
For example, the matrix

whichis neither striétly proper norproper. . M(s) =\


[n order to resolve this difficulty, we need a new concept.
COLUMN AND ROW-REDUCED POLYNOMIAL MATRICES 601

mple, for Definition G-4


A nonsingular p x p polynomial matrix M(s) is called column reduced if
] (G-54) deg det M(s) = L
p

bciM(s)
i~1

It is called row reduced if


p

deg det M\s) = L b,.iM(S)


i= 1
and if (;(s) = N(s)D -1 (S)
The matrix D(s) in (G-55) is not column reduced because deg det D(s) =
)(s)J
o <bciD(s) +b c2 D(s) =2 +1 =3; neither is it row reduced. A matrix may be
column reduced but not row reduced or vice versa. For example, the matrix

l(S)J M(s) -
2
_[3s2 +2s 3 2s + 1J (G-56)
s +s - s
is column redueed but not row reduced (verify!)_ A diagonal polynomial
matrix is always both column and row reduced.
Let bc¡M(s) = k ci - Then the polynomial matrix M(s) can be written as
t of N(s)_ Then we have (G-57)

where Hc(s) = diag{sk ci, i = 1,2, _. _, p}. The constant matrixM hc will be called
., q
the column-degree coefficient matrix; its ith column is the coefficients of the ith
column of M(s) associated with i ci . The polynomia-I matrix M¡c(s) conta-ins the
.he summation is carried
remaining terms and its ith column has a degree smaller than k ci ' For example,
~e of nij(s), i = 1, 2, .. _, q,
the M(s) in (G-56) can be written as
, k = 1. 2, ... , p. Hence

M(s)=[~ ~I~2 ~J+L~3 . ~J


p
In terms of (G-57), we have
Q.E_D.
det M(s) = (det MhcJSLkci +terms with degrees smaller than "L,k ci
»-I(S) is strictly proper,
ponding column degrees Menee, we concíude that M(s) is coiumn-reauceli ij anc! omy ij its cOiumn-degree
Iso true. In general, the coefficient matrix M hc is nonsingular.
Similar to (G-57), we can also write M(s) as

M(s) = H,_(s)M hr + M1r(s) (G-58)


(G-55)
where R(s) =diag{i'-¡, i = 1, 2, .. _, p} and k"i = b,_¡M(s) is the degree of the ith
row. M h ,. will be called the row-degree coefficient matrix; its ith row is the coef­
ficients of the ith row of M(s) associated with Sk'·i. The polynomial matrix

~J Mi.cs)contains the remaining terms and its ith row has adegree smaller than k r ¡·
For ex.ample, the matrix in (G-56) can be written as

M s)
(
=[S2O (j J[3
S2
0J +[ 2s
1 O s- 3
mcept.
602 POL YNOMIALS AND POLYNOMIAL MATRICES

In terms of (G-58), we have that M(s) is row reduced if and only if its row-degree The column degrees a
coefficient matrix M h , is nonsingular.
With the concept, we can now generalize Theorem G-9 to the following.

Theorem G -1 O
Let N(s) and D(s) be q x p and p x p polynomial matrices, and let D(s) be column
reduced. Then the rational function N(s)D- 1(s) is proper (strictly proper) if and is singular. Hencl
and only if exist (Xl_ (X1_ and (X3 SUC

for i = 1, 2, ... ,p.

Proof We normalize the (Xi as


to be 1. In this exam¡
The necessity part has been established in Theorem G-9. We now show the
choose 0':1 = 1, and (Xl .
sufficient parto Following (G-57), we write
postmultiply the unim
D(s) = DheHe(s) + Dle(s) = [Dile + D lc (s)H e- 1(s)]H e(s)
N(s) = NheHc(s) + Nds) = [N lle + Nle(s)He-1(S)]He(s)
Ul(~
where 6eiDlc(S) < 6ci D(S) ~ J1.i, He(s) ~ diag {si", sl'" . .. , sl'P}, and 6 ei N1e(s) < J1.i.
Then we have
(;(s) ~ N(s)D- 1(s) = [N',e + Nte{s)He- 1(S)] [D he + D 1e (s)H e- 1(s)]-1 to M(s), we obtain

Clearly Nds)H e- 1 (s) and D1e(S)He- I(S) both approach zero as s-+ oo. Hence we
have M(s)U 1
lim (;(s) = N',eD,;;, 1
s-oo

where D he is nonsingular by the column reducedness assumption ofD(s). Now if where the degree of n
6ciN(S) '::;6 ei D(s), N'Ie is a nonzero matrix and (;(s) is proper. Ir 6ciN(S) < 6ciD(S), verified that M 1(s) is Cl
Nhc is a zero matrix and (;(s) is strictly proper. Q.E.D.
From theexample,
The Hermite-form polynomial matrix shown in (G-28) is column reduced. column degree can b,
Since every polynomial matrix can b~ transformed ¡nto the Hermite f0rm b)' ~'.
sequence of elementary row transformations (Theorem G-5), we conclude that nomial matrix can be r
every nonsingular polynomial matrix can be transformed to be column reduced the earlier statement a~
by a sequence of elementary row operations. It turns out that the same can be
achieved by a sequence of columns operations. This will be illustrated by an
Theorem G -11
example.
For every nonsingular
Example U(s) and V(s) such th
reduced.
Consider
+1 S 52 +2s +1 An algorithm is' av
M(s)= 2s - 2 ~2S2 +1 column- or rbw~reducf
[
-s 5s 2 -2s to the matrix case.
COLUMN AND ROW-REDUCED POLYNOM1AL MATRICES 603

Id onlJI if its row-degree The column degrees are 1, 2, and O. The column-degree coefficient matrix is

}-9 to the fol1owing. 1


-2
5
and let 0(5) be column
and is singular. Hence, M(s) is not column reduced. Since M llc is singular there
'per (strictly proper) if
exist (;(1- (;(2' and (;(3 such that

~)]

We normalize the (;(i associated with the column with the highest column degree
to be 1. In this example, the second column has the highest degree; hence, we
·9. We now show the
choose Cl. 2 = 1, and (;(1 and (;(3 can be computed as Cl. 1 = 3, (;(3 = -2. Now if we
postmultiply the unimodular matrix
(s)]Hc(s)

~ [~ ~] ~[~ ~]
Cl. 1 S 3s
1(s)J Oc (s)

Ud') 1 1
2
Cl. 3 S -2s 2

>1c(s)Hc-1(s)] -1 to M(s), we obtain

oas s-oo.
~J~M'(S)
Hence';'e 5s +1
[' +1
M(s)UI(s)= ~~2 -6s +1
-2s

nption ofD(s). Now if where the degree of the second column is reduced by one. It can be readily
er. IU)ci N(s) < bciD(s), verified that MI (s) is column reduced. Iil
Q.E.D.
From the example, we see that by a proper elementary column operation, the
28) is column reduced. column degree can be reduced, whereas the determinantal degree remains
the Hermite form by a unchanged. Hence, by a sequence of eleYrt~Dtary s01~~n~~ ~:J~r~l~:~:J~).8, 2. ;..~<>Jy.
3-5), we conclude that nomial matrix can be reduced to be column reduced. We summarize this with
1to be column reduced the earlier statement as a theorm.
It that the same can be
ill be illustrated by an
Theorem G -11
For every nonsingular polynomial matrix M(s), there exist unimodular matriCes
U(s) and V(s) such that M(s)U(s) and V(s)M(s) are column reduced or row
reduced. •

] An algorithm is availablein Reference S137 to transform a matrix.into a


column- or row~reduced olie. In the fol1owing, we shall extend Theorem G-'1
to the matrix case.
604 POLYNOMIALS AND POLYNOMIAL MATRICES

Theorem G-1 2
Thís theorem is du
of Theorem G-12 is c
Let D(s) and N(s) be p x p and q x p polynomial matrices and let D(s) be non­

compute Q(s) and R(


singular. Then there exist unique q x p polynomial matrices Q(s) and R(s)

available in Reference
such that
In the following, w
N(s) = Q(s)D(s) + R(s) Consider p x p polyn<
and R(s)D-I(s) is strictly proper A is a p x p constant m
1, i = 1, 2, ... , p. We
which can be replaced by, if D(s) is column reduced,
i = 1, 2, ... , p
Define
Proof
Consider the rational matrix (;(s) = N(s) D - I (s2. This rational matrix is not and
necessarily proper. If every element gij(s) of G(s) is decomposed as guCs) =
gijsP(S) +%(s), where gijsp(s) is a s~rictly proper rational function and %(s) Corollary G -12
is a polynomial, then we can write G(s) as
Let D(s) = sl- A, and
G(s) = N(s)D-I(s) = Gsp(s) +Q(s) (G-59) exist unique polynomi
where (;sp(s) is a strictly proper rational matrix and Q(s) is a polynomial matrix.
The postmultiplication of D(s) to (G-59) yields and
N(s) = Q(s)D(s) + R(s)
This corollary can
with R(s) =:' Gsp(s)l)(s) or Gsp(s) =R(s)D-1(s)
Q,(s) = N"s" - l + (NIlA
Since R(s) is equal to the difference of two polynomial matrices [R(s) = N(s)­
and is left as an exercis
Q(s)D(s)], it must be a polynomial matrix.
To show uniqueness, suppose there are other Q(s) and R(s) such that their colurnn and row '

N(s) = Q(s)D(s) +R(s) = Q(s)D(s) + R(s) (G-60)


G-6 Coprime Frr
and R(s)D - I(S) is strictly proper. Then Equation (G-60) implies
[R(s) - R(s)]D-l(s) = Q(s) - Q(s). Consider a q x p prop,
!S c:3.11eC ~J rfqh.t-('D.TJri'
Its right-hand side is a polynorfliaI rfiatri/~, ;¡vhefeas lts left-hanci siae is a stricltj A- 1(s)8(s) a leji-coprir.
proper rational matrix. This is possible only if Q(5) =Q(s) and R(s) = R(s). will also be called an iI
The column degree inequality follows directly from Theorem G-lO. Q.E.D. many fractions, sorne e
related by the followin
Theorem G-12'

Let A(s) and B(s) be q x q and q x p polynomial matrices. IfA(s) is nonsingular, Theorem G -1 3
there exist unique q x p polynomial matrices Q(s) and R(s) such that
.Consider a q x p prop
B(s) = A(s)Q(s) + R(s) C(s) = N(s)D - I (s). Th
and A -1(s)R(s) is striCtly proper .. p x p nonsingular poly
which can be replaced by, if A(s) is row reduced, N(s)

i = 1, 2, ... ,q • Ir the rraction N(s)D - I


- - - ---_._-----------­

COPRIME FACTORIZATIONS OF PROPER RATlONAL MATRICES 605

This theorem is dual to Theorem G-12, and its proof is omitted. The proof
of Theorem G-12 is constructive in nature, and its procedure can be used to
;es and let O(s) be non­
compute Q(s) and R(s). Ir O(s) is colurnn reduced, difTerent procedures are
matrices Q(s) and R(s)
available in References S34, S137, and S236 (see also Problem G-15).
In the following, we discuss two special cases of Theorems G-12 and G-12'.
Consider p x p polynomial matrices D(s) and N(s). Let D(s) = sI - A, where
o..
A is a p x p constant matrix. Clearly D(s) is nonsingular and bciD(s) = ¡D(s) =
1, i = 1, 2, ... ,p. We write N(s) as

p
Define
N,.(A) = N"A" + N"_IA"-1 + + NoI
rational matrix is not and N¡(A)=A"N" +A"-IN"_1 + +IN o
decomposed as 9¡is) =
nal function and q;j(s) Corollary G-12
Let O(s) = sl- A, and let N(s) be an arbitrary polynomial matrix. Then there
(G-59 ) exist unique polynornial matrices Qr(s) and Q¡(s) such that
is a polynomial rnatrix. N(s) = Q,.(s)(sl- A) + N..(A)
and N(s) = (sI - A)Q¡(s) + N¡(A) III

This corollary can be readily verified by using


)D-1(s)

Q,.(s)= N"S"-l +(N"A +N"_I)S"-2 + ... +(N"A"-l +N"_IA"-2 + ... +N 1)


matrices [R(s) = N(s)­
and is left as an exercise. Note that N..(A) and N¡(A) are constant matrices and
J.d R(s) such that their column and row degrees are all equal to zero.

R(s) (G-60)
G-6 Coprime Fractiolls of Proper Rational Matrices
O) implies
). Consider a q x p proper rational matrix (;(s). The fraction (;(s) = N(s)D - I(S)
is caIled a right-coprim€ fracUon jr N(s) and JIJJ(s) are right coorime; G(s) =
~ft-hand
side is a strictiy A -1(s)8(s) a lefl-coprimefraclion if A(s) and 8(s) are ieft coprime. Either one
= Q(s) and R(s) = R(s). will also be called an irreduciblefraclion. Given a (;(s), it is possible to obtain
~orem G-10. Q.E.D. many fractions, sorne are irreducible and sorne are noto However, they are all
related by the following theorem.

. If A(s) is nonsingular, Theorem G-13


R(s) such that
Consider a q xp proper rational matrix (;(s) with the right-coprirne fraction
(;(s) = N(s)D~ 1(s). Then for any other fraction G(s) = N(s)D - I(S) there exists a
p x p nonsingu.lar polynomial matrix T(s) such that

-.N(s) = N(s)T(s) . and D(s) = D(s)T(s)


q I If the fraction N(s)D -1(S) is also rightcoprirne, then T(s) is unimodular. •
606 POLYNOMIALS AND POLYNOMIAL MATRICES

Proof 6 compute R - l(S) and

Let 0- l(S) = Adj O(s)/det O(s), where Adj stands for the adjoint and det stands D
for ~he ~eterminant ofa matrix. Then, N(s)O- l(S) = Ñ(s)D- 1 (s) and O(S)O-I(S) Then we have
= O(s)O- l(S) = 1 imply
(;(s) = N(s)D- J

N(s) Adj O(s) det D(s) = Ñ(s) Adj D(s) det O(s) = Ñ(s)O- J
and O(s) Adj O(s) det D(s) = D(s) Adj D(s) det O(s) and N(s)D - l (s) is irre<
This procedure, how(
Let R(s) be a gcrd of N(s)and D(s), and let R(s) be a gcrd of Ñ(s) and D(s). Then,
nomial matrix and is r
it is clear that R(s) Adj D(s) det D(s) is a gcrd of the two left-hand-side poly­
we also compute Ves)
nomial matrices in (G-61) and R(s) Adj D(s) det D(s) is a gcrd of the right-hand­
side polynomial matrices in (G-6l). Because of the equalities in (G-61),
R(s) Adj D(s) det D(s) and R(s) Adj D(s) det D(s) are two different gcrds of N(s)
Adj D(s) det D(s) and D(s) Adj D(s) det D(s). We claim that the polynomial and N(s)D-l(S)=V Z1
matrix D(s) Adj D(s) det D(s) is nonsingular. lndeed, beca use of det D(s)D- l(S)
ríght coprime fraction
=det [D(s) Adj O(s)/det D(s)] = det [D(s) Adj D(s)]/(det D(s))P = 1, we have of R(s) can be avoided
det [D(s) Adj D(s) det D(s)] = [det D(s) det D(s)]P =/=0. Hence Corollary G-7 sarily column reduced
implies the existence of a unimodular matrix W(s) such that In this section, we
R(s) Adj D(s) det D(s) = W(s)R(s) Adj D(s) det D(s) fraction from a right
procedure is similar to
which implies
Consider the q x p
R(s)D- l(S) = W(s)R(s)D - l(S) N(s)D-I(s), where A(s)
and p x p polynomial
or D- l(S) = R - 1 (s)W(s)R(s)D- l(S) (G-62) written as B(s)D(s) = A
Since R(s) is unimodular following the irreducibility assumption of N(s)D - l(S),
the matrix

T(s) = R - 1 (s)W(s)R(s) (G-63) If we consider polYI


!R(s), then Equation (G
is a polynomial matrix The substitution of(G-63) into (G-62) yields 0- l(S) = sequently, aH 1 x (p +­
T(s)D - I(S) or D(s) = D(s)T(s). The substitution of 0- 1 (s) = T(s)D - l(S) into nomials) satisfying
Ñ(s)D - l(S) = N(s)D- l(S) yields immediately Ñ(s) = N(s)T(s). The nonsingular­
i,ty of T(s) folIo-viS frorú the l10nsingularities of T{(sj, -;:;V(sj, ana 1R(s).
If Ñ(s) and D(s) are right coprime, then R(s) is unimodular. Consequently,
the T(s) in (G-63) is also unimodular: This completes the proof ofthe theorem. is a linear space over !R(.
Q.E.D. Following Definition 2
to (p +- q) - rank [D'(:
From this theorem, we see that all irreducible fractions of a proper rational dimensional null space
matrix are related by unimodular matrices. Hence, the irreducible fraction is qualifies as a basis (Th
unique in the sense that all irreducible fractions can be generated from a single only the pblynomial so
irreducible fraction. pan of \l. A set of q v
Conside( a fraction (;(s) = N(s)D- l(S). lf it is not irre'ducible, we may in (\11" !RIs]),7 if every
use the proce'dure in (G-33) to compute the gcrd R(s) of N(s) and·D(s). We then
71lis in faó a. free module OVt
6For a differenl proof, see Problems G-12 and G-l3. chaplcr and Refercncc S34.
COPRIME FACTORIZATlONS OF PROPER RATIONAL MATRICES 607

compute R -I(S) and


D(s) = O(s)R -I(S) N(s) = N(s)R - I(S)
.adjoint and det stands
)O-I(s)and O(S)O-I(S) Then we have
G(s) = N(S)O-I(S) = N(s)R(s)[D(s)R(s)] -1 = N(s)R(s)R -I(S)O-I(S)
t 0(.1')

= N(s)D-I(s)

t 0(.1') (G-61 )
and N(s)D - 1(.1') is irreducible. This is one way to obtain an irreducible fraction.
This procedure, however, rcquires the computation of the inverse of a poly­
>fN(s)andD(s). Then, nomial matrix and is rather complicated. Ir in the process of generating (G-33),
'10 left-hand-side poly­ we also compute Ves) in (G-35), then we have
gcrd ofthe right-hand­
: equalities in (G-61), O(S)J = Vis) [R(S)] = [V II(S)J R(s)
I different gcrds of N(s)
[ N(s) O V 21(S)
m that the polynomial and N(S)0-I(S)=V 21 (s)V¡?(s). Since Ves) is unimodular, V21(s)V1II(s) is a
;ause of det O(s)O -I(S) ríght coprime fraction (why?). By this method, the computation of the inverse
let D(s»P = 1, we have of R(s) can be avoided. Note that V 11(.1') obtained in this process is not neces­
Hence Corollary G-7 sarily column reduced.
that In this section, we shall introduce a method of obtaining a left-coprime
det D(s) fraction from a right fraction, not necessarily coprime, and vice versa. The
procedure is similar to the scalar case discussed in Section G-2.
Consider the q x p proper rational matrix (;(s). Let G(s) = A -1(s)8(s) =
N(S)O-I(S), where A(s), B(s), N(s), and D(s) are, respectively, q x q, q x p, q x p,
and p xp polynomial matrices. Theequality A-I(s)B(s)=N(s)O-I(s) can be
(G-62)
written as 8(s)0(s) = A(s)N(s) or
Jmption of N(s)O- I(S),
[-B(s) A(S)][D(S)] =0 (G-64)
N(s)
(G-63) If we consider polynomials as elements of the field of real rational functions
lR(s), thcn Equation (G-64) is a homogencous linear algcbraic cC]ualion. Con­
(G-62) yields O-lis) = sequently, all 1 x (p + q) vectors x(s) with elements in lR(s) (including poly­
-I(S) =T(S)O-l(S) into nomials) satisfying
f(s). The nonsingular­ i[})(s11
;), and R(s). ;~(s) l_ N(síJ = J

)dular. Consequently,
e proof of the theorem. is a linear space over IR (s), denoted as (\1, !R(s». It is a subspace of (IR [J +q(s), !R(s».
Q.E.D. Following Definition 2-11, we call it the left null space. Its dimension is equal
to (p + q) - rank [0'(.1') N'(s)]' = p + q - P = q (Problem 2-51). In this q­
'ns of a proper rational dimensional null space (\1, lR(s», any set of q linearly independent vectors in W
: irreducible fraction is qualifies as a basis (Theorem 2-1). In our study we are however interested in
~enerated from a single only the polynomial solutions of (G-64). We use Wp to denote the polynomial
part of V A set of q vectors in Wp will be called a polynómial basis, or a basis
t irreducible, we may in (W p, IR [s]V if every vector in \1 p can be expressed as a unique combination
I/(s)and D(s). We then
71l is in rael a rrec module over lhe polynomial ring !REs] wilh dimension q. 5ee roolnole 4 arlhis
ehapter and Rererenee 534.
608 POLYNOMIALS AND POLYNOMIAL MATRICES

the first q linearly de¡:


of the q vectors by using only polynomials as coefficients. We note that every we define
basis of (W P' !R [s]) is a polynornial basis of (W, !R(s)); the converse however is not
true in general. 8 It turns out that A(s) and B(s) are left coprime if and only if the 1st block {
set of the q rows of [-8(5) A(s)] is a basis of (W p , !R[s]) (Problern G-16).
Arnong all bases in (W P' !R[s]), sorne have the additional property that the row
2nd block {
degrees of the basis vectors are srnallest possible. This type of basis is called a
minimafpofynomiaf basis. It will be shown that the set of q rows of [ - B(s) A(s)]
is a rninimal polynornial basis if and only if A(s) and B(s) are left coprime and
A(s) is row reduced. In the fol1owing, we discuss a method to find a minimal (k + 1)th block {
polynornial basis.
Instead of solving (G-64) directly, we shall translate it into a hornogeneous
The rows formed frOl
linear algebraic equation with real nurnbers as entries. Let
rnatrix Sk has k +1 b
Now it is assumed th::
D(s)=D o +DIs + +Ddsd

its linearly dependent


N(s) = No +NIs + +Ndsd

(G-65)
A(s)=A o +A 1 s + +Amsm Lemma G-1
and B(s) = Bo + BIs + +Bmsm
If (;(s) = N(s)D- 1 (s) i~

pendent of their previ

where Di, Ni, A i , and B i are p x p, q x p, q x q, q x p constant matrices. By


substituting (G-65) into (G-64) and equating the coefficient of Si to zero yield,
This lemma will
similar to (G-22) and (G-24), we obtain
N(s)D-I(s) be a rigl
N(s)D-I(s) is not pro
Do D 1 •.. Dd O O'" O ample, a linearly depe
No NI ... N d O O··· O
0- - -Ó~ -.-. ~ -i>~~ ~ -i>~ -Ó -.-. ~-Ó

[-80 A o : -B I A I :---: -Bm A m] O No ... N d ­ I N d O ... O =0


-
· - - -.- - - - - - - - - - - - - - - - - - - -. ­ even though we have
·· .. ..
Ir, instead ofSk, we
O O Do DI···· "Dd
O O ... No NI ..... Nd­
(G-66)

'We shail caD ¡he mairix formed trom lDl i and Ni the generafized resultant of D(s)
and N(s). Ir D(s) and N(s) are known, this equation can be used to solve B i
and A¡ and, consequently, B(s) and A(s). Conversely, if A(s) and B(s) are given,
a similar equation can be set up to solve for D(s) and N(s). In Equation (G-66), then the statement is al
there are q rows of unknown [ - B o A o ... - Bm A m ]. In order to have q consider
rows of nontrivial solutions, there must be, roughly speaking, q linearly de­
pendent rows in the resultant in (G-66). Since it is desirable to have m, the
degree of A(s) and B(s), as small as possible, we shal1 try to use, roughly speaking,
If we use Sk' then a lin<
Because af Lernrn
only in the N rowS. L
BThe discussion'is briefand Ihe·re~der needs nol be concerned because the subsequenL development (i + 1)th block raw. B
is independent or the discussion. for a. complete discussion, see Rderence S95, where
[D'(s) N'(sl]' is assumedto have afu\l rank. Our problem assumes D(s) to. be nonsingular and
the devel~pment can be simplified slightly.
COPRIME FACTORIZATlONS OF PROPER RATlONAL MATRICES 609

the first q linearly dependent rows in the resultant in (G-66). In order to do so,
ltSo We note that every we define
converse however is not
oprime if and only if the 1st block { Do D] ... Do O O' .. O

!R[s]) (Problem G-16). No NI'" No O O' .. O


} f0~number of
il property that the row
.0- --O; -. ~ ~ -ñ ~ ~ ~ -D~ - -0- -.-. -. - O­ dependent rows)
type of basis is called a 2nd block { ~ __ ~_o_ ~ ~ ~ _f"!'!.-_I__f"!o__~ __. ~ ~ _~_ } f 1
qrowsof[ -B(s) A(s)]
:(s) are left coprime and (k + l)th block 0- -0- - -o ~ ~ -ñ~ -- -i>~ ~ -.-.-. -D~J

-'-0

:thod to find a minimal {O O ... No NI' ..... No } fk

: it into a homogeneous
The rows formed from Di will be called D rows; those from Ni, N rows. The
Let
matrix Sk has k + 1 block rows; each block row has p D rows and q N rows.
Now it is assumed that the row-searching algorithm has been applied to Sk and
its linearly dependent rows in order from top to bottom have been identified.
(G-65)
Lemma G-1
Ir G(s) = N(s)D- 1 (s) is proper, aH D rows in Sb k =0, 1, ... ,are linearly inde­
pendent of their previous rows. !l
:onstant matrices. By
:ient of Si to zero yield,
This lemma will be proved later. This lemma does not require that
N(s)D - 1 (s) be a right-coprime fraction nor D(s) be column reducedo lf
N(s)D -1 (s) is not proper, then the statement is not true in general. For ex­
O' O O" O ample, a linearly dependent row will appear in the D rows of Sk formed for
O O O
~I:::~s ~Tl=[_:
O"

~-----------

-1

-1
Do O
No O
O
O =0
[:2 S-:l]

even though we have bc¡N(s) ~bciD(s), where b ci denotes the column degree.
Ir, instead ofSk , we arrange Di and Ni in the descending power of s as
DI ······Do
NI ..... Nd­
¡

Do DO-l ...
No N O - 1 •••
Do
No
O
O
. ..

..o O
0l
(G-66) A --------------------------­

S" ==! ®!Di TI""'í'l: lf'\',C ~ ~


ralized resultant of D(s)
an be used to solve Bi
A.(s) and B(5) are given,
l?---~!
.
...
- - - -'-'-'- - ~_1_
..
..
~_o_ - ~ '- '- - ?J
-
. .

)0 In Equation (G-66), then the statement is again not valid even if N(s) D - I(S) is proper. For example,
'm]. In order to have q consider
:>eaking, q linearly de­
:sirable to have ni, the
. use, roughly speaking,
[1 1][~ .:Tl
.. If we use Sb then ~ linearly dependent row 'w\1i appear in a D row.
.. Because of Lemma G-l, the linearly dependent rows of Sk will appear
orily in the N rows. Let r¡ be the number .of linear1y dependent N rows in the
the subsequenldevi:lopmenl (i + 1)th block r6w. Because of the structure of Sk' we have
see Rererence S95, where
. D(s) 10 be nonsingularafld (G-G9)
610 POLYNOMIALS AND POLYNOMIAL MATRICES

Let V be the least integer such that r v = q or, equivalently, matrices

(G-70)
are left eoprime, ane
This implies that as k increases, the total number of linearly independent N formo
rows in Sk will increase monotonically. However, once the number ceases to
increase, no matter how many more block rows are added, the number of Proof

linearly independent N rows will remain the same. Define In order not to be o'

n=(q-ro)+(q-rl)+'" +(cl-r v -¡) (G-71 )

lt is the total number of linearly independent N rows in Sk for k;::o: v - 1. lt


turns out that n is the degree of (;(s) or the dimension of any irreducible realiza­
tion of (;(s).
The number oflinearly dependent N rows in SI' is equal to ro + rl + ... + r",
which is c1early larger than q. However, there are only q primary linearly
dependent rows in S". A dependent row is called primary if the corresponding
row in the previous block is independent of its previous rows. For example, all
the ro dependent rows in the first block of SI' are primary linearly dependent
rows. However, the rl dependent rows in the second block of SI' are not all
primary because ro of the corresponding rows in the first block have already
appeared as linearly dependent rows. Hence, in the second block of SI" there
are only rl -ro primary linearly dependent rows. Similarly, there are rz -rl
primary linearly dependent rows in the third block of Sv.· Proceeding in this
manner, we conclude that the number of primary linearly dependent rows in
SI' is equal to

rO+(rl-rO)+(rz-r¡}+'" +(rv-r,._¡}=r,.=q
where x denotes non;
Consider Equation (G-66) with m replaced by v: independent of their r
From (G-74), wc hay·
(G-72)
study the structure of
These A¡ and lB¡ are to 1]<:': obtained by using the row-searching algo,ithm. ,~
other words, they are the q rows of l(in iK§", computed as in (A-7), correspon­
ding to the q primary dependent rows of SI"

Theorem G-14
The primary depel
Consider a q x p proper rational matrix (;(s) factored as (;(s) = N(s)D- 1 (s). shown. Correspondi
We form Sk and search its linearly dependent rows by using the row-searching (G-75) assume the for
algarithm. 9 Let [ - B o Aa ,.. -B,. Av] be the q rows of I< in KS v=Sv
SO
corresponding to the q primary de.pendent rows af SI" Then the palynomial
'-[--(~l?.Al-a-?-z -a~f\ :
-B o a~1 a~2 a~3:
"On a digital compllter complltation, thÚ; algorithm ShOllld be replaced by a nllmerically stable a~1 a~z a~3:
method:' SecAppendix A. (fthe fOW searching aigorithm is not empioyed, theresllltis generally
not in the polynomial echelon formo
COPRIME FACTORIZATIONS OF PROPER RATIONAL MATRICES 611

matrices
"
A(s) = L Ais i
and B(s) = L Bis i (G-73)
i=O i=O
(G-70)
are left coprime, and A(s) is in a canonical form called the polynomial echelon
nearly independent N formo
the number ceases to
idded, the number of Proof
me In order not to be overwhelmed by notations, we assume p = q = 3 and
(G-71 )
-D
x
in Sk for k 2: v -1. It
my irreducible realiza­
x },o~o
x
al to 1'0+1'1 + ... +r., D
x
~
lly q primary linearly
T if the corresponding X
},' 1
ows. For example, all KS3=~ O (G-74 )
uy linearly dependent -- ­
D
)Iock of S" are not all

~
x
:st block have already
},' 1
x
:ond block of S., there
larly, there are r2 -r1 O
¡v' ProCeeding in this D

~3
rly dependent rows in O ~

O },' ~

=r,,=q O
where x denotes nonzero row and O zero rows. Since all D rows are linearly
independent of their previous rows, they are not written out explicitly in (G-74).
From (G-74), we have ro = O, 1'1 = 1, r2 = 1, r3 = 3 = q, and v = 3. In order to
18 v =0 (G-72)
study the structure of A(s), we write (G-72) as
rrching algorithm. in
as in (A-7), correspon­

(G-75)

The primary dependent rows of S3 in (G-74) are indicated by the arrows as


as (;(s) = N(s)D- 1 (s). shown. Corresponding to these primary dependent rows, the B i and Ai in
;ing the row-searching (G-7.5) ass~me the form
rows of K in KS v =Sv
SO Si S2 S3
Then the polynomial .".
,
aO. .·a? 2 a?3:
. r
.A

(f~ ,
.\~~
,

~
aL al -- O O O , O O

~80

1\ . ·12 ,,
a21 ti~2 a~~:
O'
-, -8 1 aL ai2 ..
O , -8 2 aL ai2 O : -B 3 :~Ü O
,
:ed by a numerically slable
a~l a° 32 aO33 ,
I aL a12 O ,, aL a~2 O , O ct~
ioyeci, lhe resu il is generally
(G-76)
612 POLYNOMIALS AND POLYNOMIAL MATRICES

This is in an echelon formo The column positions ofthe three :1) are determined and the v defined in (
by the primary linearly dependent rows. For example, the last row of the
second block of S3 is a primary dependent row; hence the last column of the
second block in (G-76) has the element O). lf (G-76) is obtained by the row We shaIl cal! v the re
searching algorithm, then (G-76) has the foIlowing properties:
v3::; .... then (G- 76'
The it
1. AIl elements (including - BJ on the right-hand side of:D are zeros.
2. AIl elements, except element ce ,
of the columns corresponding to primary Hence we have
linearly dependent rows of S" are zeros.
{set of row degrees o
3. AIl elements ofthe column corresponding to nonprimary linearly dependent
rows of S,. are zeros. Since A(s) is row redl
q
Property 3 is the same as saying that aIl columns which are on the right-hand deg det A(s) = I: v,
side and occupy the same positions in each block as those columns with the j;:;: 1

((:elements are zero columns. For example, in (G-76), the sixth column ofthe
third and fourth blocks are zero columns because they are on the right-hand
where deg det stands
side of the sixth column, with element:))' of the second block. We note that
we are ready to show
the aboye three properties may overlap. For example, the rightmost column
then there exists a po
of (G-76) is a zero column following 1 alone 01' foIlowing 2 alone. Because of
these properties, A(s) becomes

a?l +alls a?2 +al2s


i - - - - i - -2 - - 3'
, - <5 - - - and
A(s) = [ :~:1_~~3~~~~3~~ _~~_: a32+ai2s+a~2s2 (G-77)
,. <5 - - - -1- - - - -2- -2- - - 3' Because of deg det Ql
a~l +a1ls+a~ls2 ·a32 +a32s+a32s +s '
1- 1

The elements encircled by dotted Hnes will be called pivot ·elements. Their
positions are determined by the elements:I: in (G-76). We note that every row This implies that the
and every column has only one pivot element. Because of Property 1, the than n. This is not p(
degree of a pivot element is larger than the degree of every right-hand-side pletes the proof of thi
element in the same row and is larger than 01' equal to the degree of every left­
hand-side element in the same roW. Because of properties 2 and 3, the degree Proof of Lemma G -'
of a pivot element is larger than the degree of every other element in the same
lf the linearly depem}(
column. A polynomial matrix with these properties is said to be in the row we can see that the 1'0\
polynomial echelonform 01' Popov form. 10 It is clear that a polynomial matrix Shl.ce A\{s) is rov! red 1!'
in the echeleon form is column reduced and row reduced. aependen t row ol' §,. si
We discuss now the row degrees of A(s). We note that in each block row of in the column of -B i
S" there are q N rows, and the ith N row, i = 1, 2, ... , q, appears v + 1 times in be larger than the con
S". Define, for i = 1, 2, ... , q, tion that (;(s) is propt
Vi ~ number of linearly independent ith N row in S" (G-78) in the N rows.

They will be caBed the row indices of G(s) = N(s)D-l(s). For the example in Example 1
(G-74), we have VI =3,1.'2=3, and 1.'3=1.· Clearly·the n defined in (G-71) is
. also given by Consider
(G-79) 1
s +l s-
\OThe nu'~erical"malrices [-B o Ao ... -,-B. A •.]and [A o Al ... A,.] oblained by G(s) = 7

[ s- 2
Ihe row searching algorilhm \vi\! always be in lhe echelon formo Hence Ihe corresponding poly­
nomial rri~lrix [ - B(s) A(s)) and A(s) are said lo be in ¡lJi': polynomia\ echelon formo S2 -1 s­
- - - - ---------- ----

COPRIME FACTORiZATlONS OF PROPER RATIONAL MATRICES 613

lree :1) are determined . and the v defined in (G-70) is


e, the last row of the v = max {Vi, i = 1, 2, ... , q} (G-80)
the last col umn of the
_s obtained by the row We shall call v the row index of C(s). Ir we rename Vi as '\\ such that VI '::;"2'::;
. " , then (G-76) and (G-77) imply that
V3 '::;
erties:
The ith row degree of A(s) = Vi i = 1,2, ... , q.
r:~e are zeros.
responding to primary Hence we have
{set of row degrees of A(s)} = {Vi' i = 1, 2, " . " , q} = {l'¡, i = 1, 2, ... , q} (G-81)
ary linearly dependent

Since A(s) is row reduced, we have


q
are on the right-hand
deg det A(s) = I Vi = n = total nurnber of linear independent N rows in S"
lOse columns with the
i ~ 1
he sixth colurnn of the
(G-82)
are on the right-hand

l block. We note that


where deg det stands for the degree of the deterrninant. With this background,
the rightmost column
we are ready to show that A(s) and B(s) are left coprime. Suppose they are not,
g 2 alone. Because of
then there exists a polynomial matrix Q(s) such that
deg det Q(s) > O
A(s) = Q(s)A(s) B(s) = Q(s)B(s) (G-83)
and A -1(s)B(s) = A - 1(s)B(s) (G-S4)
(G-77)
Because of deg det Q(s) > O, we have
deg det A(s) < deg det A(s) = n
>ivot elements. Their
Ve note that every row This implies that the nurnber of linearly independent N rows in Sv is smaller
Ise of Property 1, the than n. This is not possible. Hence A(s) and B(s) are left coprime. This com­
every right-hand-side pletes the proof of this theorem. Q.E. D.
le degree of every left­
¡es 2 and 3, the degree Proof of Lemma G-1
~r element in the sarne
Ir the linearly dependent rows ofS. appear only in the N rows, then frorn (G-76)
said to be in the row
we can see that the row degrees of B(s) are smaller than or equal to those of A(s).
t a polynomial matrix
Since A(s) is row reduced, the computed A -l{s)B(s) is proper. Now ir a linearly
1.
dependent row of §. shows up in a D row, then í~Í) element in ,G-(6) wili appear
tt in each block row of
in the colurnn of - B i . Now because of property 1, the row degree of B(s) will
appears v + 1 times in
be larger than the corresponding row degree of A(s). This violates the assump­
tion that C(s) is proper. Hence all linearly dependent rows of Sv must appear
row inS. (G-78) in the N rows. Q.E.D.
. For the example in
Example 1
n defined in (G-71) is
Consider"

,~~ S~IJ=[S-1
(G-79) 1"
s-1
S2 -1""

\" ... A,] oblaincd by


C(s)"~ s. +1 -"
S2 2 S2
lJ"[S2-
2 O
1 O
s-1
J-I
Ice the correspondingpoly­ [
¡al echclon for 111. S2 -1 s- d s- 1
614 POLYNOMIALS AND POLYNOMIAL MATRICES

The fraetion N(s)D - ~(s) is earried out by taking the least eommon denominator Note that the pivots
of eaeh eolumn of G(s) as the eorresponding diagonal e\ement of D(s). Thís /"2 = 2 = q, we have l'

fraetion happens to be right eoprime. In general, a fraetion obtained by this the arrows shown.
process will not be coprime. the formula in (A-l1
We form S2 and apply the row-searehing algorithm:
1 [~.5 =~.5:
O
-1 1 Note that the solutic
O 2 O 1: ,
- -Ó- - -6 -- í ---i :- -1- - - - - - - - - - - - - - ~ - - - - - - - - - - - - - - - - - - ­
, ,
8(5) = _ [~.5
O O O O: O 1
and
FtS 2 = O O 1 O: O O
O O O O ;-1 O 0.5 1: A(s) =[-0.5
--------------,---------- J
O O O 1: O O -1 0;--(---------------­ -1
O O O O: O 1 0.5 O: O 1 and G(5) =A- 1 (s)B(s
O O O 1 :-1 -1 O O: O O 1 linearly independent
the polynomial eehel
O O O O: O -2 -1 0:-1 O O
:~D O O O 1 O Combining Theo'
O ~~~D O 1 O O eorollary reduces to I
-1 1 {)) O O O
Corollary G-14
O 2 O 0(1)0.
- - - - - - - - ---= ¡ -- -6 --0- --6- -~:ü --ó-----­ Consider a proper
right eoprime if and
O -1 O ~I~ O O
** linearly independent
-1 (Ü 1 O O O
O 2 O O 1 O In the following \\
--·--------·----~1---6--0---ó--(6·-0
properties of (;(5) anc
O -\ O :T~ O O Hon. First, we note
altered without affee
-1 1 1 O O O
and (G-84) by ehoos
O 2 O O o J
::-=- -(~ O O O 1 O 2, ... ,q} rather than
O ::~ I: O 1 O O lish the main resulto
O - - Ó- - -(j) - - -( - -~ i --- -0­
Lemma G-2
O O O 2 ,X: O
o 3 O O :I~ O The m x m polynom
the fie\d of rationa\ fl
O -\ O 1 O O
Ó- - - :l~ ---0- ---O . --0- -. -O
O
O
O
O
O
O
O
O
O
00
O
:I~ O
v,~;

[
TO

.

O O O O O· (i~ O O
0- ---O . --ó -- -O ---0- . - ·0--- O
O .~
is offull row rank (i
O O O O O O O O
COPRIME FACTORIZATIONS OF PROPER RATIONAL MATRICES 615

common denominator Note that the pivots are chosen for the convenience of hand calculation. Since
:::lement of D(s). This r2 = 2 = q, we have v = 2. The primary dependent rows of S2 are indicated by
ction obtained by this the arrows shown. Corresponding to these primary dependent rows, we use
the formula in (A-U) to compute

[°1.5 -2.5: -0.5 -1:-1


-1 :-1 °:-1
Note that the solution is in the echelon formo Hence, we have

-2.5 -J - -l--11 0J s-[0 °-J 2.5


[~.5
° ° [s-0.5
B(s) = _ S2 = -J
-1 -1 s-1 s +1
and

_[-0.5 -IJ [0.5 lJ [0 0J s 2_[0.5S-0.5 :~~_:.JJ'


A(s ) -
- 1 ° ° ° 1° + s+ - .---_.
:_.~.2_ -=- )) °
and G(s) = A -l(s)B(s). Note that deg det A(s) is equal to the total number of
1
línearly independent N rows in S2, A(s) and B(s) are left coprime, and A(s) is in
°° 1
O 1
the polynomial echelon formo lilI

Combining Theorems G-8 and G-14, we have the following corollary. The
corollary reduces to Corollary G-4 for the scalar case.

Corollary G-14
Considero a proper rational matrix C(s). The fraction G(s) = N(s)D-i(s) is
right coprime if and only if deg det D(s) = n, where n is the total number of
línearly independent N rows in S" -1 or Sk for k 2. v - 1. lii1

In the following we show that the row indices defined in (G-78) are invariant
properties of G(s) and are independent ofthe N(s) and D(s) used in the computa­
tion. First, we note that the order of the rows of the matrix in (G-76) can be
altered without affecting Equation (G-75). This can also be seen from (G-83)
and (G-84) by choosing Q(s) as an elementary matrix which interchanges the
rov:,¡ positions of ftl(S) and 1B(s). 1-If~nce '.,:VhéÜ is irnportal1t 1~:; thé set : ",';, ; :::";. ':.
2, ... , q} rather than the individual Vi' We need the following lemma to estab­
lísh the main resulto

Lemma G-2
The m x m polynomial matrix T(s)=T o +Tis + ... +Tjsj is nonsingular (in
the field of rational functions) if and only if the numerical matrix

O
...

!
T1 Tj
[ T, ...

V.~
To T j - i Tj
.0 : ] k+ 1 block 'Ows (G-85)

O O To Ti ... T j

O -E-­
is of full row rank (in the field of complex numbers) for k = 0, 1,2, ....

616 POLYNOMIALS AND POLYNOMIAL MATRICES

Proof yields the number of.


consider (G-74). Sin
IfT(s) is singular, there exists a 1 x m nonzero polynomia\ vector we have one N row wit
cx(s) = CX +cxs + ... +CXkS k
o with row index 3. In (
the row indices are Ul
such that indices are also intrim
cx(s)T(s)=0 D(s) used in the comp
This equation implies As implied by (G-8
[CX o CX 1 .•• cxk]Vk= O Hence we condude tr
In other words. if G(~
Hence, if T(s) is singular, Vk does not have a full row rank for every k. coprime and A(s) is ro
By reversing the aboye argument, we can show that if Vk does not have a As discussed follo\!
flll.l row rank for every k, then T(s) is singular. This completes the proof of the
lemma. Q.E.D.

With this lemma, we are ready to establish the following theorem. form a q-dimensional
G-14 are left coprime:
Theorem G -1 5 nomial basis of the nu1
Let G(s) = A -1(s)B(s) be a left-coprime fraction and A(s) be row reduced. rows of Sk are seard
Then the row degrees of A(s) are intrinsic properties of G(s) and are independent [ - B(s) A(s)] are the
ofthe N(s) and D(s) used in Theorem G-14 in the computation. that 6r ¡B(s) .::s;6,.¡A(s) aj
degrees of [ - B(s) A
Proof Since the row degrees
degrees is unique (The<
Because of (G-81), it is sufficient to show that the row indices defined in (G-78) minimal polynomial bao
are independent of the N(s) and D(s) used in the computation. We recall that A(s) is nol row reduced
every right fraction N(s)D -1(S) of G(s) can be obtained from a single right­ have deg det Á(s) = de~
coprime fraction Ñ(s)D -1(S) by the relationship

D(S)J = [?(S)J T(s) (G-86)


[ N(s) N(s)
Thus the row degrees o
where T(s) is a p x p nonsingular polynomial matrix (Theorem G-13). If we [ - B(s) ~~(s)J is o. rnin
write 1'(5) as T(s) = T o +ir ¡s + ... +- Tjsj, then (G·gfi) i'Tp!ip,~. t::.·c ;. . (\ ,
c.
i.:OpfiiIiC diíú h\S) ¡S roy

Sk =SkV d+k (G-87) additional nice properl


quence of the row sea!
where Sk and Sk are defined similarly as in (G-67), d is the degree of D(s) as in obtained by using the r
(G-65) and Vd +k is defined as in (G-85). Since Vd +k has a full row rank for every basis and in a canonice
k, (G-87) implies We discss briefly 1
rank Sk = rank Sk k =0,1,2,. . . (G-88)

Consequently, I'¡ ='¡, i =0,1,2, ... , where F¡are defiried as in (G-67). Hence
i =0, 1, 2, ... , vare independent of N(s) and D(s) used. in the computation.
I'¡,
Allsolutions y(s) of [A
Now we shall show that the set {Vi, i = 1, 2, ... ,q} is uniquely determinable [f D(s) and N(s) are rig.
of
from {r¡, i = 0, 1,2, ... , v} .We óbserve Ihat /'¡ gives the number !V rows with is a 'polynomial basis o
row index i or smaller. Define r _1 = O. Then . the set is a minimal po
/ =0, 1, ... , v (G-89) coluriln searching algo
COPRIME FACTORIZATIONS OF PROPER RATIONAl MATRICES 617

yields the number of N rows with row index exactly equal to i. For example,
consider (G-74). Since rO-I"-I=O, rl-ro=l, 1"2-rl=0, and r 3 -r2=2,
.al vector we have one N row with row index 1, no N row with row index 2, and two N rows
with row index 3. In other words, the row indices are {1, 3, 3}. This shows that
the row indices are uniquely determined by rj, i =0, 1, ... v. Hence the row
indices are also intrinsic properties of (;(.1') and are independent of the N(s) and
D(s) used in the computation. Q.E. D.

As implied by (G-81), the set of row indices is equal to the set of row degrees.
Hence we conc\ude that the row degrees are also intrinsic properties of (;(.1').
In other words, if (;(.1') is factored as A -1(s)B(s), where A(s) and B(s) are left
.k for every k. coprime and A(s) is row reduced, then the set of row degrees of A(s) is unique.
: if V k does not have a As discussed following Equation (G-64), all solutions x(s) of
lpletes the proa f of the
Q.E.D.
X(s{~~;~J =0
ving theorem. form a q-dimensionalleft null space. The A(s) and B(s) computed in Theorem
G-14 are left coprime; hence the set of the q rows of [ - B(s) A(s)] is a poly­
nomial basis of the null space (Problem G-16). Because the linearly dependent
A(s) be row reduced. rows of Sk are searched in order from top to bottom, the row degrees of
:.1') and are independent [ - B(s) A(s)] are the smallest possible. The properness of N(s)D - 1(.1') ensures
Jtation. that b,.¡B(s)'::;b,.¡A(s) and that all pivot elements appear in A(s); thus the row
degrees of [ - B(s) A(s)] are equal to those of A(s) and A(s) is row reduced.
Since the row degrees of A(s) are smallest possible and since the set. of row
degrees is unique (Theorem G-15), the set of q rows of [ - B(s) A(s)] is indeed a
dices definedin (G-78) minima! po!ynomia! basis. We show that if A(s) and B(s) are left coprime but
tation. We recall that A(s) is not row reduced, then [ - B(s) A(s)] is not a minimal basis. Indeed, we
d from a single right­ have deg det A(s) = deg det A(s) and

L b.. ¡A(s) > degdet A(s) =degdet A(s) = I (),.¡A(s)


( G-86)
Thus the row degrees of [ - B(s) Á(s)] are not minimum. This establishes that
'heorem G-13). If we [ - B(s) A(s)] is a minimal polynomial basis if and only ir A(s) and 8(.1') are !eft
pUes, rOl k = 0, 3., 2.} ... " coprime and A(s) is row reduceo. 'l'ile basis obtained in Theüeem 0-14 has une
(G-87) additional nice property; it is in the polynomial echelon form. This is a conse­
quence of the row searching algorithm. In conc\usion, the solution of (G-66)
he degree of D(s) as in obtained by using the row searching algorithm is a polynomial basis, a minimal
full row rank for every basis and in a canonical formo
We disc ..ss briefly the dual case of Equation (G-64). Consider
(G-88)

~d as in (G-67). Hence
[A(s) B(s)] [ - ~~;J = O
:d in the computation. All solutions y(s) of [A(s) B(s)]y(s)= O forro a p-dimensional right null space.
uniq uely determinable If D(s) and N(s) are right coprime, the set of the p columns of [ - N'(s) D'(s)]'
Iuniber of N rows with is a polynomial basis of thenull space. rr, in addition, D(s) is column reduced,
the set is a minimal polynomial basis. If the solution is obtained by using the
(G-89) column searching algorithm (dual to the row searching algorithm), then tlle
618 POLYNOMIALS AND POLYNOMIAL MATRICES

minima! polynomial basis is in an echelon formo These results are dual to


Theorems G-14 and G-15 and are stated as theorems.

Theorem G-14'
Consider a q x p proper rational matrix G(s) factored as G(5) = A -1(s)8(s) with
G-2 The polynomials [
A(s) = I~=o A¡Si and B(s) = I~=o B¡Si. We form
if the square matrix S of
Aa Bo O O , :O O1 matrix S" _, of arder Li.
Al B1 Aa Bo : . .. : O O statements.
, .
,, : G-3 Transform the mal
Am _ 1 Bm ­ I : Am ­ z Bm ­ Z : Aa Bo k + 1 block columns (each
T k = Am B", Bm ­ , :, Al B 1 block has q +p columns)
O O
Ám­
Am
O
I

B",
O
I

: Az B z r
lo
,: into the Hermite row for
,.
O O O O :A
, m Bm

and search linearly dependent columns in order from Idt to right. Let r¡ be the
number of linearly dependent B columns in the (i + 1)th block, and let jJ. be the
i
r
least integer such that Ji = p. Then D(s) = Di and N(s) = L;=o Nis , L;=o G-4 Find a gcrd of the I
solved from
D(

G-5 Are the following p


S +1
a. D,(s)= [
(s-lXs +2)
by using the column-searching algorithm corresponding to the p primary linearly b. D 2(s)=D I(s) N2(~
dependent columns ofT Ji' are right coprime. Furthermore, D(5) is in the col umn c. D 3 (s)= D,(s) N3(~

polynomial echelon form, that is, the degree of every pivot element of D(s) is d. D 4 (s) = D,(s) N 4 ()
larger than the degree of every other element in the same row, larger than the
degree of every lov/er el~D1en.t in ~Jv~ same columt~. ;",;,;:~,¿, hre tne Vau's l j-'~j\S.
degree of every upper element in the same column. We call jJ. the column index in Problem O-S? Are th
of G(s). notes the transpose.)

Theorem G -, 5' G-7. ls the matrix


Let G(s) = N(s)D-1(s) be a right-coprime fraetion and D(s) be column reduced.
Then tl).e eolumn degreesof D(s) are intrinsic properties of G(s) and the set of
the column degrees is unique. .. I

eolumn reduced? [f not,


Problems V(s)M(s) are column redu

G·' Apply the row searching algorithm to Skin(G-20) to reduce the following rationál
G-8 ls the M(s) in Probl
functions to irreducible ones:
·-------..- - - - - _ . _ - ­

PROBLEMS 619

se results are dual to N(s) 5'+5 2 _5+2


- = --;::--=--­
D(s) 2s'-s2+ s +1
N(s) 2s' +25 2 - s -1
D(s) 25 4 +5 2 -1
(;(s) = A-1(s)B(s) with
G-2 The polynomials D(s) or degree n and N(s) or degree In < n are coprime ir and only
ir the square matrix S of order n + m in (G-12) is nonsingular or if and only ir the square
matrix S" - [ of order 2n in (G-20) is nonsíngular. Show the equivalence of these t'NO
statements.

G-3 Transform the matrix


~ block columns (each
ck has q + P columns) o (s + 1)2 _S2 +s+ 1]
O -s-1 s-1
S2 +s+1 s
into the Hermite row formo

to right. Let r¡ be the s+1


,)lock, and let J1 be the O
"1'
and N(s) = L.i;O N ¡s,i
G-4 Find a gcrd of the fol1owing two matrices by using (G-33):
52 +2s
O(s)= [ 15 2 -5
5+3
3s-2
J N(s)=[s 1]

G-5 Are the following pairs of polynomial matrices right coprime?


S +1
a. 01(S)= [ N¡(s) =[s +2 s +1]
(s-I)(s+2)
1Ithe p primary linearly b.0 2 (s)=D\(s) N 2 (s)=[s-1 s + 1]
~, D(s) is in the column c.O,(s)=O¡(s) N,(s)=[s+1 s - 1]
vot element of D(s) is d.0 4 (s)=D¡(s) N 4 (s)=[s s]
e row, larger than the
r ¡han or eq ua] lO 1he G-o Are the pairs l&i\s), J¡\sjj, ¡ = 1, 2, .\ 4, ieil copniúe ir" Á¡\s) = ;Dii(.),/ anu 1Ü>¡(.)j = l·~il':;i
:all fJ. the column ¡ne/ex in Problem G-5? Are they left coprime if A¡(s) = O¡(s) and B¡(s) = N;(s)? (The prime de­
notes the transpose.)

G-7. Is the matrix


s) be column reduced.
of (;(s) and the set. of 2s + l'

• O

5-1

column 'reduced? Ir not, find unimodular matrices U(s) and V(s) such that M(s)U(s) and
V(s)M(s) are column reduced. .'

:e the following ratiqnal


G-B Is the M(s) in Problem G-7 row reduced? Ir not, transform it to a row-reduced one.
-------~----------~-­
._~---_._~~----- ------~--------------~-- ----------;­

620 POLYNOMIALS AND POLYNOMIAL MATRICES

and N(s), show that for ;


G-9 Find, by applying the row searching algorithm to Sk in (G-67), left coprime fractions
tine equation ifand onl
from the following three different right fractions: matrix F(s) such that F[
A [.1'3 +.1'2 +s + 1 .1'2 + s J[S4 +.1'2
G(s) =
.1'2 + 1 2.1' .1'2 + 1
are soll1tions of the Dio
=[.1' 71
~I:2 -s+l
o J-' matrix T(s). the matrice~

=fS2 +.1'
bs+1
o
-s +1
J-' and

Are the results the same? Which right fraction, if there is any, is right coprime? are solutions of the Dior

G-10 Find a right fraction from the left fraetion G-15 Consider

G(S)=[S2~1 s2~J-r~1 ~:~J \vith ()ciR(S) < (\;D(s) = 11,


1f D(s) is column reducec
(soIUtiOn: [~ s~lIs~l s2~lTl) N(s)W'(s)~r
=(
G-11 Let (;(.1) = A -'(s)B(s) = N(s)O- '(s) be two eoprime fractions. Show that there where Nis) is a polynon
exists a unimodular matrix of the form of Np(s). Define H(s) = (
V 11 (s) V ds)J
[ B(s) Np(s) = N(s
A(s)
Q(s)~Q(s
such that D(s)~DI;;:

[V 11(.1) V12(S)J[
B(s) A(s) _ -
O(S)J=['J
N(s) O
or V 12(.1) V I I (S)J[ - N(S)J = [IJ
[ A(s) B(s) 0(.1') O
where the ith row of Qk i
compllted recursively as
(Hint: Ir O and N are right coprime, we have Q

[~:: ~::I -~J=[~J


Q

and U H and U 22 are left coprime and Vil V 21 = ND-'. Using the dual ofTheorem G-13, whcre D[~O irl>/l. f
there exists a unimodular matrix M such that MV 21 = B and MV 22 = A.) R(s) = N(s) - Q(s)O(s). S

Silow that ji N(.I')lJ}-l(.I') = I\J{s)D ¡(sj are [wo right-coprime í"ractions, lhea ¡he
:':;;-")2 G·"hi Consicier G(s) = ¡.
matrix V(s)=D-'(s)O(s) is unimodular. (Hint: The equation XN+YO=I implies rational fUl1ction solutior
XÑiJ-¡ D +YDD-' O =(XÑ +YD)V = 1 which implies U-' to be a polynomial matrix.)

G-~3 Prove Theorem G-13 by using Problem G-12.


and let Wp denote the poi
G-14 Consider the Diophantine polynomial equation nomial basis of W if ever
basis by using only pe
0c(s)O(s) + Nc(s)N(s) = F(s) of [ - B(s) A(s)] is a poI
[Hint: (=» Let [ - B(s)
where O,., O, N" N, and F are, respectively, p x p, p x p, p x q, q x p, and p x" p polynomial
matrices. Let " " polynomialmatrix T(s) SI

U 11(.1) Udsf\[O(s)J~[R(S)J
[ U2\(s) udst N(s) O Show that T(s) is unimodu
be any vector in Wp • Sho'
where the leftmost matrix is unimodular and R(s) is a gcrdof 0(.1) and N(s). Given O(s)
PROBLEMS 621

and N(s), show that for any F(s), there exist solutions D/s) and Nc(s) to meet the Diophan­
·67), left coprime fractions
tine equation if and only if R(s) is a right divisor of F(s), that is, there exists a polynomia1
matrix F(s) such that F(s) = F(s)R(s). Show also that
3 J-l D~(s) = F(s)U11(s)
2 +25
are solutions of the Diophantine equation. Finally, sho\\' that, for any p x p polynomial
matrix T(s), the matrices

Dc(s) = D~(s) +T(s)U} 1(S)


and Nc(s) = N~(s) +T(s)ll}}(s)

s right coprime? are solu tions of the Diophantine equation (see Reference S 139).

G-15 Consider
N(s)=Q(s)O(s) +R(s)
-lJ
-2 with bc¡R(s) < bciO(s) = /l¡. Let H(s) = diag {S·', SI", ... , Sl'p} and Jet 0(5) = OhcH(S) + D1c(s).
Ir D(s) is column reduced, then D{oc is nonsinguJar. We rewrite the equation as
-1 )
N(s)H-I(S)~ Np(s) + N,.(s) = Q(S)(OhcH(S) + D1c(s))H - I(S) + R(s)W I(S)
=Q(s)D"c(I +Dl~IOlcls)Wl(S)) +R(s)H-1(s)
:tions. Show that there where N p(s) is a polynomial matrix and N,(s) is strictly proper. Let Vi be the row degrees
of Np(s). Define H(s) = diag {s", S''', ... ,s'p} and /l = max {Jl i }, v = max (v,}. We write

Np(s) = Ñ(s)[N po + Np1s- 1 + ... + Np"s-']


Q(s)~Q(S)Ohc~H(s)[Qo+ QIS-I + ... + Q,05-''J
D(S)~0/;;,iOlcls)H-l(S)~0IS-I+02S-2+ .,. +O.s-·

where the ilh row of Qk is a zero row if k> Vi. Show that the nonzero rows of Qk can be
compuled recursively as
Qo = N po
k- 1

Qk = N pk - I Q,D k- 1 k=I,2, ... "


1=0

he dual ofTheorem G-13, where Ol~ O if / > JI. From Q(s), we can compute Q(s) = Q(s)DI;;' 1 and lhen compute
0 22 = A.) R(s) = N(s) - Q(s)O(s). See Reference S137.

)rime fractions, then the G-';<'; COi1siüc( G(SI = h. '(s)J8;(s) = r<J(s)íLv '(S). Lel W (¡eIlOl\;; ¡he leí, ¡luir ~fl<i¡;" \.)\ ,d,
n XN+YD=I imp1ies ralional function solutions of
be a polynomial matrix.)
x(s)[O(S)]=0
N(s)
and let Wp denote the polynomial part of W. A set of polynomial vectors is called a poly­
nomial basis of W if every vector in Wp can be expressed as a unique combination of the
basis by usingonly polynomials as coefficients. Show thal the set of lhe rows
of [ - 8(s) A(s)l is a polynomial basis of W if and on\y if A(s) and 8(s) are left coprime.
[Hint: (~) Let [ ....:. B(s) Á(s)J be left coprime. 'If [ - 8(s) A(s)J is a basis, there exists a
( p, and p x p polynomial
po\ynomia\ malrix TM S1Ich that .
[- B(s) Á(s)].=T(s)[ ~ 8(.1') A(s)J
Show that T(s) is unimodular and hence A(s)andB(s) are left coprime.( <:= )Let [ - b(s) ¡¡(s)J
be any veclor in V p. Show that ¡¡(s) ¡lo O, and thenappend (q - 1) vectors in VI' lo it to form
s) and N(s). Given Oís)
622 POLYNOMIALS AND POLYNOMIAL MATRICES

[ - B(s) A(s)] with A(s) nonsingular and


- - [O(S)J
=0
[-B(s) A(s)]
N(s)
Then use Theorem G-13 to show the existence of a unique polynomial matrix T(s) such
¡hat [ - B(s) A(s)] = T(s)[ - B(s) A(s)] and [- bis) a(s)] = t(s)[ - B(s) A(s)].]

G-17 Let Mp(s) be a polynomial matrix and let Vis) be a unimodular matrix such that
Mp(s)U(s) is column reduced. Show that M; 1 (s) is proper if and only if 6 d V(s).:s;
6 d (l\1 p (s)V(s)), for al! i.

In this appendix we ~
functions. We discw
case.

Definition H-1
A number Je (real or (
function g(s) if Ig(Je)1 =

If a proper ration:
trivial common facto¡
that every root of the
numerator of q(s) is ~
mOL oi the acnomina·
not a pole of

although it is a root o
Consider a proper
tion

Ir wc write g(s) = N(s)


common factor), then
'namial matrix T(s) such
[ - 8(s) A(s)].]

adular matrix such that


r and anly ir Dei U(s) s H
Pales and Zeras

In this appendix we shall introduce the concepts of pole and zero for transfer
functions. We discuss first the single-variable case and then the multivariable
case.

Definitíon H-1
A number A (real 01' complex) is said to be a pole of a proper rational transfer
function g(s) if ig(A)1 = oc. It is said to be a zero of g(s) if g(A) = O. ¡¡j

Ir a proper rational transfer function is irreducible (that is, there is no non­


trivial common factor between its numerator and denominator), then it is clear
that every root of the denominator of 9(S) is a pole of 9(S) and every root of the
numerator of g(05) is a zero of 9(5). Without this irreducibility assumption, "­
root oi tile Cienorülnawi oi Y(S} iílay not Oí:; <l pole oí' Y(s). t'ur el..dnipie, - 1 iS
not a pole of
A S +1
g(s) = 052 +3s +2

although it is a root of S2 + 3s + 2.
Consider a proper rational function g(s) with the following irreducible realiza­
tion
X '7Ax +bu y =cx +eu
If we write g(s) = N(s)/ D(s) and N(s) an.O D(s) are coprime (have no nontrivial
" commoti facto'i), then we have
D(s) =k det (s 1 - A)
623
624 POLES AND ZEROS

with sorne constant k. If D(s) is monic (the coefficient associated with the highest not a pole of 9(s), th,
power of s is equal to 1), then k = 1. We discuss in the following the implica­ output is to apply e;'
tions of poles and zeros.
Theorem H-2
Theorem H-1 Consider a system w
Consider a single-variable system with proper transfer function 9(S) and an tion {A, b, e, e}. Ir
irreducible realization {A, b, e, e}. Then a number 1, is a pole of 9(S) if and only is not a pole of g(s), t
if there exists an initial state X o such that the zero-input response at the output and the input u = e;·r
of the system is equal to
To prove this the
y(t) = re A' for all t ~O
(sI -A)-l(S­
for sorne nonzero constant r.
for any 1, that is not ;
the partial fraction e;
Proof
The zero-input response of the system is given by
(s ­
y(s) =e(sI -A(I X (O)
The identity in (H-l)
Ir ), is a pole of 9(S), then it is an eigenvalue of A. Let v be an eigenvector (A-I,I) and (sI-A)
of A associated with A; that is, Av = AV. Then it can be verified (Problem 2-32)
that v is an eigenvector of (si - A( 1 associated with eigenvalue (s - ,,1.)-1. Proof of Theorem H
Hence we have
The response of the
Y(s) = c(sl - A)-I V=cv(s - ,,1.)- 1 as derived in (4-25),
ar y(t)=eve AC forallt~O
y(s) = c(:
What remains lo be shown is that the constant r =cv is different from zero. =c(.
The realization {A, c} is observable by assumption; hence the matrix [sI' -A' :
cT has a full rank at every s in C. Consequently, for every nonzero vector, in The substitution of (f
particular, the vector v, we have
y(s) =c(sl- AY­

[Al -AJc
vf.O
=e(sI - AY­
Since (Al - A)v = O, we must have r = el' f.O; otherwise, it would have violated
the aboye condition. This compíetes the proof of í:he necessity of [he theorem.
To show the converse, we show that if y = re A', t ~ O, then A is a pole of 9(S). y(s) = g(A)(s - A:
Ir y(t) = re A', then we have
or y(t) = g(A )e A'
y(s) = c(sl - A)-I X (O) = r(s - ,,1.)-1
This completes the pr
or det (s~ _ A) e [Adj (sI - A)]x(O) = s ~A In this theorern,
Otherwise, lhe theore
or (s - A)c [Adj (sI - A)Jx(O) = r det (si - A) ing that the impedan
which implies det(Ai - A) = O. HenceA is an eigenvalue of A and, consequently, a positive real functi(
a pole o[g(s). This completes the proóf orthe theorem. Q.E.D. give a physicannterpl
g(s), that is,gv,) = O, t .
This theorem states that if 1, is a poIe of g(s), the mode eAC can be generated identically zero even :
at the output by an initial state without the application of any input. If A is blocked by the systerr
POLES AND ZEROS 625

,ciated with the highest not a pole of 9(S), then this is not possible; the only way to generate e}" at the
following the implica­ output is to apply e;" at the input.

Theorem H-2
Consider a system with proper transfer function 9(S) and an irreducible realiza­
tion {A, b, c, e}. Ir the input u(t) is of the form e"', where }" real or complex,
function 9(S) and an
is not a pole of 9(S), then the output due to the initial state x(O) = - (A - Al) - 1 b
pole of 9(S) if and only
and the input u = e}" is equal to y(t) = g(J,)eJ.' for t ~ O. I:ii
response at the output
To prave this theorem, we need the following identity:
(sl-Ar1(s-A)-1 =(AI-A)-1(S-A)-1 +(sI-A)-1(A-,U)-1 (H-1)

for any A that is not an eigenvalue of A. Note the similarity of this identity to
the partial fraction expansion
1 1 1 1 1
---- =--'--+--'-­
(s-a)(s-A) (A.-a) (s-),) (s-a) (a-J,)
The identity in (H-l) can be readily verified by post- and premultiplication of
et v be an eigenvector (A - ),1) and (si - A). (Problem H-1.)
'erified (Problem 2-32)
eigenvalue (s - A) - 1. Proot ot Theorem H-2
The response of the system due to the initial state x(O) and the input u(t) is,
as derived in (4-25),
Y(s) = c(sI - A)-I X (O) +c(sl - A) - 1bu(s) +eu(s)
is different fram zero. =c(sI - A)-1 X(0) +c(sI - A)-lb(s - A)-1 + e(s - ),)-1 ( H-2)
e the matrix [sI' -A' :
ery nonzero vector, in The substitution of (H-l) into (H-2) yields

y(s)=c(sI - A)-I X (O) +c(AI-Ar lb(s _A)-l


+c(sl- A)-I(A - JcI)-lb + e(s-Ar 1
=c(sl- A)-I[X(O) + (A -AI)-lb] + [c(AI- A)-lb+e](s- A)-1
t would have violated
cessity of the theorem.
then A is a pole of 9(S). Y(s) = g(A)(S - A)-1
or y(t) = g(A)eJ.' for t ~ O
This completes the proof. Q.E.D.

In this theorem, the assumption that J, is not a pole of g(s) is essential.


Otherwise, the theorem.does not hold. This theorem is very useful in establish­
-A) ing that the impedance of a linear time-invariant, lumped, passive network is
,f A and, consequently, a positive realfuncti'on (see Reference 31). This theorem can also be used to
Q.E.D. give a physical interpretation ofthe zeros of a -transfer function. If), is a zero of
. 9(s), that is, g(A) = 0, then' for a certain initial state, the output of the system is
e e;" can be generated identicallyzero even if the input e;" is applied. In otherwords, the input e;" is
of any input. Ir A is blocked by the system. .
626 POLES AND ZEROS

Carollary H-Z (Transmission-blocking property) then pG(s) = p, wher


the field of rational f
Cansider a system with proper transfer function 9(s) and an irreducible realiza­
p x 1 rational vector
tian {A, b, e, e}. Ir}, is a zero of 9(s), then the response at the output due to the
initial state x(O) = - (A - íll) - 1b and the input u(t) = eA' is identically zero. 111
Ir we multiply the 1('
We shall now extend the concepts of poles and zeros to the multivariable
become
case. Consider a q x p proper rational matrix (;(s) with the following coprime
fraction
(;(s) = O¡-I(s)N¡(s) = N,.(s)O,:- I(S) where M(s) and Pes) 2
where DI> N¡, N,., and O,. are, respectively, q x q, q x p, q x p, and p x p poly­
nomial matrices. Furthermore, N/(s) and O¡(s) are left coprime; N..(s) and O..(s)
are right coprime. Let for al1 possible inpu
smal1er than q. Simi
X=Ax +Bu y=Cx +Eu pes) to the system, th,
be an irreducible realization of (;(s). Then we have
det(sl - A) = k 1 det 01(S) = k 2 det O,(s) Hence the number of
rank, there are no red
where k 1 and k 2 are constants. In view of this relationship, we may give the
Consider C(s) wit
following definition.

Definition H-Z
Ir (;(s) is of full rank
A number íl, real or complex, is said to be a pote of a proper rational matrix This implies that for
(;(s) if and only if it is a root of det O(s) = O, where O(s) is the denominator a rank eq ual to min (
matrix of any right- or left-coprime fraction of C(s). 111
Definition H-3
Similar to Theorem H-l, we have the following theorem to characterize the
<;;;onsider a q x p pr
poles of (;(s). G(s) = D¡- 1 (s)N¡(s). j

Theorem H-3 fuI! rank (in Ihe field


is said lo be a transr.
Consider a multivariable system with proper transfer matrix (;(s) and an ir­ in <C, the field of com
-ec1ucible
1 -..... _ n~;¡ 1;7"t'''-[:
_ ........ _A_"-'C ,o.U ... J j}, 18\'"
t--, ., ~1 ~.".'.,"".,-.,.;,.,.-,.»__ .'." ...1..'.-.,)·.h<eó.··
'L.,...} ~f' ,-,,_ .Z '.".,_ :-'.. .·.~,10-..c f':.! ..\ .•

~_'.1 •. "~'. ""\'J/

only if there exists an initial state X o such that the zero-input response al the Example 1
output of the system is egual to
Consider the left eopr
y(t) = reAl
s
o

l
for some nonzero vector r. I
~ s +2
The proof of this theorem is identical to the one of Theorem H-l and will G 1 (s) = O
not be repeated. We note that every pole of C(s) must be a pole of sorne element
of(;(s),and every pole of every element of (;(s) must be a pole of (;(s). This
factfol1ows from the faet that det O(s) is equal to the least eommon denominator This N¡(s) has rank 2 I
ofaIi Iílinors of C(s) (see Definition 6-1and Theorem 6-2).
We shall now extend the concept of zeros to the inultivariable case. First,
we assume that the q x p proper rational matrix (;(s) has a full rank in the field 'If G(5) does not llave a fuI
of rational fundions. By this, we mean that if q'~ p, then p(;(s) = q; if q> p, ;. in C'is atransmission zen
POLES AND ZEROS 627

then p(;(s) = p, where p denotes the rank. If (;(s) does not have a full rank in
the field of rational functions, then there exists a 1 x q rational vector M(s) or a
an irreducible realiza­ p x 1 rational vector P(s) such that
the output due to the
is identically zero. 11 M(s)(;(s) =0 or (;(s)P(s) =0 ( H-3)

If we multiply the least common denominator of M(s) or P(s) to (H-3), they


s to the multivariable become
the fol1owing coprime
M(s)(;(s) =0 or (;(s)P(s) = O

where M(s) and P(s) are polynomial vectors. Since y(s) = G(s)u(s), we have

q x p, and p x p poly­ M(s)y(s) = M(s)G(s)u(s) =0


prime; N..(s)and D,.(s) for al1 possible inputs. This implies that the number of effective outputs is
smal1er than q. Similarly, if G(s)P(s) =0, and if we connect a precompensator
P(s) to the system, then for all possible inputs u(s) we have
G(s)P(s)u(s) =0

:s) Hence the number of effective inputs is smaller than p. Hence, if G(s) has a full
rank, there are no redundant or noneffective input and output terminals.
;hip, we may give the Consider (;(s) with the fol1owing coprime fractions:
(;(s) = D¡-1(S)NtCs) = N,.(s)D,:-1(S) (H-4)

If (;(s) is of full rank in the field of rational functions, so are NtCs) and Nr(s)..
roper' rational matrix This implies that for almost every A in ic, the q x p complex matrix NtCA) has
s) is the denominator a rank equal to min (p, q) in the field of complex numbers.

• Definition H-3
~m to characterize the Consider a q x jJ proper rational function C(s) with the coprime fraction
C(s) = D¡-1(S)N¡(s). It is assumed 1 that C(s) and, consequently, N¡(s) have a
full rank (in the field of rational function). Then a number )" real or complex,
is said to be a transmission zero of (;(s) if and only if rank N¡(A.) < min (p, q),
latrix G(.~) ~nd an ir­ in ic, the field of complex numbers. ~
E~ pole af {~\s) ir and
input response at the Example 1
Consider the left coprime fraction
s
• o +11
l
~ s +2 s O
s+2 =[s+2 0J-1[S
G 1 (s)= O 1 ·0 O s+1
'heorem H-l and wíll s + 1
S2

l poleofsomeelement 7 S
a pole of (;(s). This
:ommon denominator This N¡(s) has rank 2 ror eYery s in iC; hence G ¡(s) has no transmission zero. I
).
:ivariable case. First,
a full rank in the field 'If (;(s) does no! have a full rank, lhen pN(A) < min(p, q) for every A in e. In olher words. every
en pG(s)=q; ir q > p, i.in e is a lransmission zero of (;(s). This is a degenera le case and will not be considered.
628 POLES AND ZEROS

Example 2 Case 1: q > P

Consider the Ieft coprime factorization In this case, we have p


in ce for alI i. in ce ex,
zera, then pN¡(i,) <¡;
vector k such that

and consequently,
This N ¡(s) has rank 1 at s = Oand s = - 2. Hence Oand - 2 are two transmission
zeras of (;(s). Note that O and - 2 are also poIes of (;2(S). ~ y(r
This property is si
From ExampIe 2, we see that (;(s) may not be weIl defined at its transmission H-2. Note that if 1,
zeras. Hence we cannot use p(;(},) < min(p, q) to define its transmission zera. pN(}.)=p and N(J,) k ,
We note that N¡(s) and Nr(s) in (HA) are both q x p poIynomiaI matrices. If X
is not a poIe of (;(s), then it is cIear that p N¡(X) < min(p, q) if and onIy if p Nr(X) <
min (P, q). In fact, it is shown in Reference S34 by using the Smith-McMillan
form that there exist unimodular matrices V(s) and U(s) such that Case 11: q < P

N¡(s) = U(s)Nh)V(s) (H-5) In this case, we have p


we have pN¡(,1,) sq <,
Hence the transmission zeros of (;(s) can be defined by using N¡(s) or Nr(s). We N¡(X); = O and
note that if N¡(s) and Nr(s) are square, the transmission zeros of (;(s) are the roots
of det N¡(s) or det Nr(s), where det stands for the determinant. y((
Fram the aboye two examples, we see that the transmission zeros of (;(s) Since this equation he
exhibits some phenomena which do not exist.in the scalar case. A transmission transmission zero of (
zera may appear as a pole of the same (;(s). Even though elements of (;(s) sider this case separat
have zeras, (;(s) may not have any transmission zera. The transmission zeros If q < p, then A is
of a square (;(s) may be different fram the zeras of det G(s). In spite of these nonzera 1 x q constan
differences, the transmission praperties of the transmission zeras of (;(s) are
quite analogous to the scalar case. We shall establish these in the foIlowing.
If, in addition, ), is ne
Theorem H-4
is also a nonzero vectl
Consider a multivariable system with q x p praper rational matrix (;(s) and an "S'··r
irreducible realization [¡}~, R, C, lE}. ifthe input au(t) is üf thé [OiTú ~K(/'\ ·v:j~·i.é{t;'L
is real or complex and is not a pole of (;(s) and k is an arbitrary p x 1 constant
vector, then theoutputdue to this inputand the initial state x(O) = -(A -,1,1)-1 Bk which is, because of (1
is equal to fy(t) = Ofor t ;::0:0, exist
y(t) = G(A)ke At fort;::o:O (H-6) be used to characteriz.
I
The praof of this theorem is identical to the one of Theorem ,H-2 and will Corollary H-4 (Trans
not be repeated.
The substitution of (;(..1.) = D¡- 1 (A)N¡(A) into (H-6) yields Consider a multivaria
is assumed that G(s) J
y(t) = D¡-I(A)N¡(A)ke At (H-7) andthe irreducible rea
where N¡(s) is a q x p polynomial matrix offu\l rank. In thdollowing, we discuss 'zera of(;(s), then ther
output of the systern
separate1y the case q ;::o:p and the case q < p. We use p to denote the rank.
POLES AND ZEROS 629

Case 1: q"2:.p
In thiscase, we have pN¡(s) = p in ~(s) by assumption. Hence we have pN¡(}.) = P
in iC for all 1, in iC except the transmission zeros of C(s). Ir l. is a transmission
zero, then pN¡(A) <p. Consequently, there exists a p x 1 nonzero constant
vector k such that
NLC).)k=O
and consequently,
2 are two transmission
s). I for all t"2:.0 ( H-8)

This property is similar to the transmission-blocking property of Corollary


ned at its transmission H-2. Note that if ), is neither a transmission zero nor a pole of (;(s), then
its transmission zero.
nomial matrices. Ir X
+
pN(..1.) = p and N(X)k O, for all constant k, which implies

ifand only if pN.(A) < y(t) = D¡- l(X)N¡(),)k/'/ +O


~ the Smith-McMillan
uch that Case 11: q < p

( H-5) In this case, we have p N¡(s) = q in ~(s) by assumption. ln this case, for any), in ic,
we have pN¡(X)::::;q <p in iC. Consequently, there exists a nonzero k such that
ing N¡(s) or N,(s). We N¡(X"- = O and
os of C(s) are the roots
nant. forall t"2:.0 ( H-9)
;mission zeros of G(s) Since this equation holds for every X in iC, it cannot be used to characterize the
.case. A trans'mission transmission zero of C(s) as in the case of q "2:.p. For this reason, we must con­
>ugh elements of C(s) sider this case separately from the case q "2:. p.
'he transmission zeros Ir q < p, then X is a transmission zero of C(s) if and only if there exists a
](5), In spite of these nonzero 1 x q constant vector h such that
;ion zeros of G(s) are
e in the following. hN¡(A) =0 ( H-10)

If, in addition, X is not a pole of G(s), the 1 x q vector f defined by f = hD¡{..1.)


is also a nonzero vector. Consider
lal matrix (;(s) and an
f~l{t) == fG{.J..)~~,(!A( -::-: i~H}J;{ ~ '?~l\- ¡ f J.)h-\!/ )\i!{ '1"l:
the formke At , where;i. A
= hN¡(X)ke ' ( H-11 )
bitrary p x 1 constant
x(O) = -(A-XI)-lBk which is, because of (H-lO), identically zero for any k. This property, namely
fy(t) =0 for t "2:.0, exists only if X is a transmission zero of G(s). Hence this can
( H-6) be used to characterize the transmission zero of G(s) for the case q < p.
I
rheorern H-2 and will Coroltaiy H _4 (Transmission-blocking property)

ds Consider a multivariable system with q x p proper rational matrix C(s). It


is assumedthat G(s) has a full rank, the coprime fraction G(s) = D¡-l(s)N¡(s)
( H-7) and the irreducible realization{A, B, C, E}. (1) If q "2:.p, and if X is a transmission
e following, we discuss zero of G(s), then there exists a nonzero p x l' constant vector k such that the
¡denote the rank. output of the system due to the initial state x(O) = - (A ~ AI)-l Bk and the
630 POLES AND ZEROS

input u(t) = ke J•t is identically equal to zero. (2) Ir q < p and if }, is a transmission zero of C(s) if and 01

zero but not a pole of G(s), then for the input u(t) = ke.l.t, where k is an arbitrary
p x 1 constant vector, there exists a nonzero 1 x q vector f such that the output
y(t) due to u(t) = ke.l. t and the initial state x(O) = - (A - i,I)- 1 Bk has the property

fy(t) = fG(}.)ke.l. t = O for all t? O Iiil

This t heorem fol


This corollary is just a restatement of (H-8) and (H-IO). We see that this In establishing (H-l
corollary is quite similar to the one for the scalar case. Because of this trans­ Hence the proof is al
mission-blocking property, the zeros defined in Definition H-3 are called the this restriction, we ~
transmission zeros. q ?p, and if A is a tr
The transmission zero can also be defined by using dynamical equations. nonzero k such that t:
Consider a system with a coprime fraction G(s) = D1-l(S)N1(s) and an irreducible
dynamical equation {A, B, C, E}. Consider the system matrix x(O) = -1

Sl-A are identically zero.


S(s)~ [ -c ( H-12)

Ir A, B, C, and E are, respectively, n x n, n x p, q x n, and q x p matrices, then The combination of (


Ses) is a (n+q) x(n+p) matrix. Though elements ofS(s) are polynomials, we
shall consider them as elements of the field of rational functions. Because of

0J[SI-A BJ [Sl-A B J Since [ - x'(O) k'J' is


Iq -C E - O C(sl-A)-lB+E Now consider the
zero, for any·k there eJ
for t ?O, or fC(A)k = (

=[
SI-A B
O C(s) = O
J
[SI-A B
D¡-l(S)Nt(s)
J holds for every k. He

we have or f[C(Ai
Define f 1 =fC(AI - A;
Sl-A
P [ -C G(s) J
B =p(sI-A.)+pG(s) A

and
which can be combiné
where p denotes the rank in the field of rational functions. Ir C(s) is of full
rank, that is, pG(s) = min(p, q), then

sl- 1\
p [ -C :J
= n +min(p, q) (H-14 ) Since [f 1 f] is a nom
the link between Coro
Now if s is replaced by an element in e, the field of complex numbers, say s = A, In the following, ti
then SeA) is a matrix with elements in e and its rank must be computed in e will be removed. The
be skipped in the first

Theorem H-5 Theorem H-6

Consider a q x p proper rational matrix G(s) with full rank and with an irredu­ Consider a q x p pro¡::
cible realization {A, B, C, E}. Ir Ais not a pole of G(s), theri Áisa transmission dimensional irreducib
POLES AND ZEROS 631

d if A is a transmission zero of G(s) if and only if


{here k is an arbitrary
AI-A
r such that the output pS(I,)=p
[ : ] < n +min(p, q) (H-15 )
- 1Bk has the property -C
i

This theorem foilows directly from Definition H-3 and Equation (H-13).
lO). We see that this
In establishing (H-13), the nonsingularity of sI - A is implicitly assumed.
Beca use of this trans­
Hence the proof is applicable only if }, is not a pole of G(s). Before removing
Jn H -3 are ca\!ed the
this restriction, we shall reestablish Theorem H-S from Corollary H-4. If
q ~p, and if A is a transmission zero, Corollary H-4 states that there exists a
dynamical equations.
nonzero k such that the output due to u(t) = ke At and the initial condition
I(S) and an irreducible
Jatrix x(O)=-(A-AI)-IBk Or (A-AI)x(O)+Bk=O (H-16)

are identically zero. Since y(t)=Cx(t) +Eu(t)= Cx(t) +Eke A', we have
( H-12)
y(O) = Cx{O) + Ek = O ( H-17)
d q x p matrices, then The combination of (H-16) and (H-17) yields
l') are polynomials, we
lunctions. Because of [A~-CA :1 -:(O)] =0

lB ] Since [-x'(O) k'J' is a nonzero vector, we have pS{l) < n +p.


I,)-IB +E
Now consider the case q < p. Corollary H-4 states that if Ais a transmission
zero, for any k there exists a nonzero 1 x q vector f such that fy{t) = fG(A~eA' = O.
for t ~ O, or fG(A)k = O. From (H-ll), we can see that, for the same f, fG(A)k = O
holds for every k. Hence we have

fG(A) =0
or f[C(AI - A)-IB +E] =fC(AI- A)-IB HE=O
Define f l = fC(AI - A) - 1 Then we have
-A) +pG(s)
fl(AI - A) = fC
¡(s) (H-13)
and flB +fE. =0
which can be combined as
ons. lf (;(5) is of fuil
f][AI-A 8J=0 ( H-18)
[fl -C E
(H-14)
Since [f1 fJ is a nonzero vector, (H-18) im plies pS(A) < n +q. This completes
the link between Corollary H-4 and Theorem H-S.
¡ex numbers, say s == A, In the fo\!owing~ the restriction that A is not a pole of G(s) in Theorem H-S
1st be computed in C. will be removed. The proof reHes heavily on the result of Section 6-6 an'd may'
be skippedin the first reading.

Theorem H-6

Ik and with an irredu­ Consider a q x p proper rational matrix G(s) with a fu\! rank and with an n­
Ilen A is a transmission dimensional irreducible realizatíon {A, B, C, E}. Then A is a transmission
632 POLES AND ZEROS

zero of (;(s) if and only if SI-A

pS(A) = P [
AI-A
-c :J < n +min(p, q)
p[ -C
and the theorem follow
2
Proof The concept of tral
possible ways of definin
Let E = (;( co). We factor (;(s) as
duce one more definitic
(;(s)=E +D- 1 (s)Ñ(s)=D- 1 (s)[D(s)E +Ñ(S)]~D-l(S)N(s) (H-19)
1 Definition H-4
where D- (s)N(s) is strictly proper and left coprime and D(s) is row reduced.
We then apply the procedure in Section 6-6 to find an irreducible {Ao, Bo, Co} Consider a q x p prope
such that, similar to (6-133) and (6-130), divisor (gcd) of the nur
are called the blocking :
Co(sI - A o)- 1 = D -1(s)L(s) (H-20)
and N(s) = L(s)8 0 (H-21) Example 3
where L(s) is defined as in (6-129). Furthermore, we have {D(s), L(s)} left Consider
coprime and {sI -A o, Co} right coprime. Hence, Theorems G-S and G-S'
imply that there exist polynomial matrices X(s), - Ves), X(s), and Ves) such that
X(s)(sI - A o) - Y(s)C o = In
L(s)X(s) + D(s)V(s) = I q
which, together with D(s )C o = L(s )(s 1 - A o), can be written in matrix form as The gcd of the three nur
zero of (;(5)_ 1t can be ~
~(S)J =[I
~(S)
[L(s) Y(S)][SI - A o
D(s) - Co Ves)

O
X(s)X(s) +Y(S)Y(S)J (i-1-22)
Iq
- 2 (Problem H-5).

From this example,


The right-hand-side matrix is clearly unimodular; hence the left-hand-side of blocking zero are nc
matrices are also unimodular (Why? Note that the left-hand-side matrices are blocking zero can neve
polynomial matrices). We compute zero can also be a polt
zero of (;(s) will appear
[L(s)
~(S) V(s)J[SI - Ao
D(s) - Co :OJ=[¿1
~(s)Bo +Y(S)EJ
L(s)B o + D(s)E transmission zero. Th(
Ir (;(s) is a scalar
fl. ~{.(S)~G -l-\Y(s)}f,~
=Lo N(s) J Similar to transmi~
blocking the transmissi
In the last step, we have used N(s) = D(s)E + Ñ(s) = D(s)E + L(s)8 0 in (H-19) component of the inpUl
and (H-21). Since the leftmost matrix in (H-23) is unimodular, we have, for
every s in e,

P [SI -Ao BOJ= [In X(s)Bo +Y(S)EJ­


E p O N(s) -n +pN(s) where <p(s) andf¡(s) are
-Co
in the output y¡(t), then
Since {A, B, e,·
E} and {Ao, Bo, Co, E} are both irreducible realizationsof is blocked from y¡{t).
C(s), they are sti-ictly system equivalent. Hence we have, for every s in e,
Theorem H-7

'This proof \vas provided by Professor L. S. Chang of China University of Sciences and Technology,
Consider a system with
HoFei. form shown in (H~25).
--~-- _ .._._._-_. . ._. . ._--- - - - - - - - - - - - _ . _ - - . _...- - - - ­

rOLES AND ZEROS 633

sI -A
p [ -C
BJ
E
= p [SI -Aa
-Ca
BoJ
E =11 +pN(s) (H-24 )

l(p, q)
and the theorem follows from the definition of transmission zero. Q.E.D.

The concept of transmission zeros introduced aboye is only one of many


possible ways of defining zeros for rational matrices. In the following, we intro­
duce one more definition.
]~D-l(S)N(s) (H-19)
Definition H-4
nd D(s) is row reduced.
irreducible {Aa, B o, Ca} Consider a q x p proper rational matrix G(s). Let {3(s) be the greatest common
divisor (gcd) of the numerators of all elements of G(s). Then the roots of {3(s)
are called the blocking zeros of G(s).
( H-20)
(H-21 )
Example 3
'e have {D(s), L(s)} left
'heorems G-S and G-S'
X(s), and Y(s) such that
Consider
s +1
s +2

(s +2)(s + 1)

j
S2 +2s +2

ritten in matrix form as The gcd of the three numerators of (;3(S) is s + l. Hence -1 is the only blocking
zero ofG(s).. lt can be shown that the transmission zeros of G(s) are O, -;-1, and
X(s) +Y(S)Y(S)] - 2 (Problem H-S). l!iI
1 (H-22)
q

From this example, we see that the definition of transmission zero and that
lence the left-hand-side
of blocking zero are not equivalent. lf every element of G(s) is irreducible, a
t-hand-side matrices are
blocking zero can never be a pole of G(s). On the other hand, a transmission
zero can also be a poleo If we factor G(s) = N,.(s)D,:-l(S), then every blocking
)8 0 +Y(S)E] zero of (;(s) will appear in every element of N,(s); hence every blocking zero is él
)8 0 + D(s)E transmission zero. The converse is, of course, not necessarily true.
If G(s) is a scalar transfer function, then there is no difference between
)lB\o +- Y(s)ll.!'.l ·~.ransmjsslon. 7,f;"(OS .3.n.et blor:k~.ng ~~r~s.
N(s) J Similar to transmission zeros, blocking zeros also have the property of
blocking the transmission of certain mode. Let u¡{s), i = 1, 2, ... , p, be the ith
)(s)E +L(s)B o in (H-19)
component of the input vector u(s). Let
limodular, we have, [01'
i= 1,2, ... ,p ( H-25)

EJ = n +p N(s) where cP(s) and f¡(s) are polynomials. Let s -lX be a root of cP(s). lf e~t appears
in the output y¡(t), then S-lX is said to appear as a mode of y¡(t). Otherwise, it
educible realizations of is blocked from y¡(t). .
ve, rOl' every s in iC,
Theorem H- 7 .

ty of Sciences and Technology, Consider a system with proper transfer matrix G(s). Letthe input u(s) be of the

form shown in (H-2S). Then, for any initial state, no root of cP(s) will appear as

634 POlES AND ZEROS

a mode at any output terminal if and only if every root of <jJ(s) is a blocking zero Problems
of G(s).
H-1 Prove the identity
Proof
H-2 Consider the dyna
Ir every root of <jJ(s) is a blocking zero of G(s), then G(s) contains no root of
<jJ(s) as a pole. Consequently, for any initial state, the zero-input response of
the system will not contain any mode due to the roots of <jJ(s). Let y¡(s) be the
ith component of the output vector Y<s) and let G(s) = (g ij(s». Then the zero­ Find an initial state so tI"
state response of the system is given by
. p H-3 Consider the egua
y¡(s) = L gij(s)u}s) ( H-26) = é' and ti
y(t) due to ult)
j= 1

Let f3(s) be the gcd of the numerators of all gij(s). Then we can write 9ij(S) as H-4 Prove Theorems 1-:

A nij(s) H-S What are the poles


%(s) = f3(s) dij(s)

and (H-26) becomes, by using (H-25),

y¡(s) = ±
j= 1
f3(s) nij(s)h(s)
<jJ(s) d¡}s)
Now if every root of <jJ(s) is a blocking zero, then <jJ(s) will divide f3(s). Hence
no root of <jJ(s) will appear as a mode of y¡(t).
Now suppose the root, s - IX, of <jJ(s) is not a blocking zero. Then there
exists at least one nkl(s) which does not contain 's - d. as a factor; otherwise,
s - ex would be a blocking zero. We choose /;(s) = O for j = 1, 2, ... ,1 - 1,
1+ 1, ... , p andfi(s) = 1, then H-G Let N(s)D-1(s) = t'
set of transmission zeros
Yk(S) = f3(s) nk!(s)
<jJ(s) dkl(S)
H-7 Show thal
and s - IX will appear as a mode of Yk(t). This completes the proof ofthis theorem A-
Q.E.D. rank [ _

Le[ ¡;¡(s) = k/(s - ,t). H), is a lransmission zero oí G(s), lflen e"; Wii! no¡ apr ea ,'
under any state feedback a
as a mode at the output vector only for a certain k. Ir A. is a blocking zero of conclusion from the facts
G(s), then eA! will not appear at the output vector for any k. These concepts of transfer-function matrix a
zeros will be used in the study of tracking control problems in Chapter 9. function matrix?
The transmission zeros of a proper transfer matrix G(s) are defined from a
coprime fraction of G(s) in Definition H-3. We see from Theorem H-6 that H-S What are the blocki
they can also be defined from an irreducible realization of G(s) by using the
system matrix in (H-15). They can again be defined fro.m the Smith-McMillan
form of G(s), see References S34 and S185. F or discussions of v~rious definitioils
of zeros, see References S152 and S186; for their computation, see References
sn and S148. Ir all transmission zeros lie inside the open'left half s-plane,
G(s) is said to be minimum phase. See Reference S67.. The presehtatión of this
appendix follows closely References S78 and S94.
POLES AND ZEROS 635

f cP(s) is a blocking zero Problems

H-1 Prove the identity in Equation (H-l).

H-2 Consider the dynamical equation


(s) contains no root of
zero-input response of x= [ -1O -2lJ x+ [O'
1
Ju y=[2 lJx
)f cP(s). Let Yi(S) be the
:g¡j(s)). Then the zero­ Find an initial state so that its zero-input response is y(c) = 5e- 1 for aH t zO.

H-3 Consider the equation in Problem H-2. Find the initial state so that the response
( H-26) y(t) due to u(t) = e31 and the initia\ state is of the form e 31 for t zO.

H-4 Prove Theorems H-3 and H-4.


n we can write gij(S) as
H-S What are the poles and transmission zeros of the transfer matrices:

G1(s)=
.1'-1
O

l
1
.1'+10
2(s-W
.1'+1
(s - 1)2
J .1'+1
.1'+2
(s +2)(s+l~
.1'2 +2.1' +2
l
~ 1rT::~
O 1 s -IJ
.viii divide f3(s). Hence G 3 (s) =
A [
.1'2
.1'-1 O s +1
cing zero. Then there
as a [actor; otherwise,
OJ- 1[1 .1'- lJ
.1'-1 2 .1'2
) for j=l, 2, ... ,/-1,
H-G Let N(s)D - 1(.1') = Ñ(s)D -1 (s) be two right-coprime fractions of (;(.1'). Show that the
set of transmission zeros defined fram N(s) and the one fram Ñ(s) are the same.

H-7 Show that


le proo[ of this theorem
Q.E.D. rank [A -Al BJ
=rank
[T(A +BK-AI)r 1

-C E -(C+EK)r l

then e1.¡ wiil not appear


under any state feedback and any equivalence transformation? Can you arrive at the same
A is a blocking zero of conclusion fram the facts that a state feedback does not affect the numerator matrix of a
y k. These concepts of transfer-function matrix and that an eq uivalence transformation does not affect a transfer­
blems in Chapter 9. function matrix?
G(s) are defined fram a
om Theorem H-6 that H-S What are the blocking zeras of the transfer matrices in Problem H-5?
n of G(s) by using. the
TI the Smith-McMilIan
1S of various definiiions
utation, see Re[erences
. open left half s-plane,
fhe presentation o[ this
n. Chang, S. S.
McGraw-Hill.
t3, Chen. C. T.. "c
Automatic Col1t
14. --o "Linear il
vol. 15, pp. 127
References 15, --o "Represel
Tra/1s. Automat
16. --o "Stabilit y
vol. 56. pp. 821
17. --o "A note (
AC-13, pp. 597·
18. - - , "Design ,
vol. 57. pp. 46­
19. - - , "Design o
Preprints 1970.
There are two lists of references. The first one was used in the preparation of 20. - - , and Deso
the original edition and the second one, the present edition. The lists are not systems," IEEE
necessarily exhaustive nor do they indicate the original sources.
21. - - , and - - ,
IEEE Ti·am. Au
1. Anderson, B. D. O., and D. G. Luenberger, "Design of multivariable
feedback systems," hoc. lEE (London), vol. 114, 1967, pp. 295-399, 1967. 22. -.-, - - , and
and observabilil
2. - - , R. W. Newcomb, R. E. Kalman, and D. C. Youla, "Equivalence of
Corztrol, vol. AC
linear time-invariant dynamical systems," 1. Franklin Inst., vol. 281,
pp. 371-378, 1966. 23. Chi, H. H., an,
simulation," Pn
3. Athans, M., and P. L.Falb, Optimal Control. New York: McGraw-Hill,
1966. 24. Coddington, E.
Equatiol1s. Ne\
4. Bass, R. W .. and I. Gura, "High order system dcsign via state-spacc con­
siderations," Preprints 1965 JACC, pp. 311-318. 25. DeBon, F., and
with inaccessib
5. Bellman, R., Introduction to Matrix Analysis. New York: t",lcGraw-Hil1.
1960.
26. Dertouzos, M. 1
6. Brand, L., "The companion matrix and its properties," Am. Math block-diagram ~
Monthly, vol. 71, pp. 629-634,1964; vol. 75, pp. 146-152,1968.
1969.
7 Brockett, R. W., "Poles, zeros, and feedback: State space interpretation," 27. DeRusso, P. M.
IEEE Trans. Automatic Control, vol. AC-lO, pp. 129-135, 1965. New York: WiJ¡
8. - - , and M. Mesarovic, "The reproducibility of multivariable systems," 28. Desoer, C. A., '
J. Math. Anal. AppI:, vól. U,pp. 548-563, 1965. . Ti·(ll1S. Circuit T
9. Buck, R. c., Advarú:ed Calculus. New York: McGraw-Hill, 1956. 29. - , EECS 22:
10. Bucy, R. S., "Canonical forms lor multivariable systems," IEEE Trans. 1968.
Automatic Control, vol. AC-lJ, pp. 567-569,1.968'­ 30. - - , and C. 1
11. Chang, A., "Ari algebraic characterization of controllability:' JEEE systems," IEEE
Ihms. Automatic Control, vol. AC 10, pp. 112-113, 1965. 31. --o and E. S.
636 1969.
REFERENCES 637

12. Chang, S. S. L., Synthesis ol Optimal Control Systems. New York:


McGraw-Hill, 1961.
13. Chen, C. T., "Output controllability of composite systems," IEEE Twns.
Automatic Control, vol. AC-12, p. 201, 1967.
14. - - , "Linear independence of analytic functions," SIAM J. Appl. Math ..
vol. 15, pp. 1272-1274, 1967.
15. - - , "Representation of linear time-invariant composite systems." ! EEE
Trans. Automatic Control. vol. AC-13, pp. 277-283. 1968.
16. - - , "Stability of linear multivariable feedback systems," Proc. IEEE,
vol. 56, pp. 821-828,1968.
17. - - , "A note on pole assignment." IEEE Trans. Automatic Control, vol.
AC-13, pp. 597-598, 1968.
18. - - , "Design of feedback control systems," Proc. Natl. Electron. Cont:
vol. 57, pp. 46-51,1969.
19. - - , "Design of pole-placement compensators for multivariable systems,"
Preprints I970 J ACC.
1 in the preparation of 20. - - , and Desoer, C. A., "Controllability and observability of composite
ion. The lists are not systems," IEEE 7/-ans. Automatic Control, vol. AC-12, pp. 402-409, 1967.
ources.
21. - - , and - - , "A proof of controllability of Jordan form state equations,"
IEEE 7/-ans. Automatic Control, vol. AC-13, pp. 195-196, 1968.
:sign of multivariable
67, pp. 295.-399, 1967. 22. - - , - - , and A. Niederlinski, "Simplified condition for controllability
and observability of linear time-invariant systems," IEEE 7/-ans. A utomatic
'oula, "Equivalence of
Control, vol. AC-ll, pp. 613-614,1966.
tnklin Inst., vol. 281,
23. Chi, H. H_, and C. T. Chen, "A sensitivity study of analog computer
simulation," Proc. Allerton Conf, pp. 845-854, 1969.
'York: McGraw-Hil1,
24. Coddington, E. A., and N. Levinson, Theory of Ordinary Di[(erential
EquatiolJs. New York: McGraw-HilL 1955.
;n via state-space con­
25. Dellon, F., and P. E. Sarachik, "Optimal control of unstable linear plants
with inaccessible states," IEEE 7/-al'ls. Automatic Control, vol. AC-13,
, York: McGraw-Hili,
pp. 491 -495 19~8
26. Dertouzos, M. L., M. E. Kaliski, and K. P. Polzen, "On-line simulation of
operties," Am. M ath
block-diagram systems," IEEE 7/-ans. Computers, vol. C-18, pp. 333-342,
6-152,1968.
1969.
space interpretation,"
27. DeRusso, P. M., R. J. Roy, and C. M. Close, State Variablesfor Engineers.
1-135,1965.
New York: Wiley, 1965.
IUltivariable systems,"
28. Desoer, C. A., "A general formulation of the Nyquist criterion," IEEE
7/-ans. Circuit Theory, vol. CT-12, pp. 230-234, 1965.
'aw-Hill; 1956.
29. --,EECS 222 Lecture Notes, University ofCalifornia, Berkeley, Fall
ystems/' IEEE Trons. 1968'.
30. _."-.,and' C. T. Chen, "Controllability and observability of feedback
ontrollability," IEEE systems," IEEE 7/·ons. Automatic Control, vol. AC-12,pp. 474-475,1967.
1965.
31. --,and E. S. Kuh, Basic Circuit Theory. New York: McGraw-Hill,
1969.
638 REFERENCES

32. - - , and P. Varaiya, "The minimal realization of a nonanticipative canonical (phasl


impulse response matrix," SI A M J. Appl. Math., vol. 15, pp. 754-763, 1967. AC-9, pp. 312-3
33. - - , and M. Y. Wu, "Stability of linear time-invariant systems," IEEE 52. Joseph, P. D., a
1I-ans. Circuit Theory, vol. CT-15, pp. 245-250, 1968. Applications and
34. - - , and - - , "Stability of multiple-Ioop feedback linear time-invariant 53. J ury, E. L, Sampl
systems," 1. Math. Anal. Appl., vol. 23, pp. 121-129,'1968. 54. Kalman, R. E., "
35. Duffin, R. J., and D. Hazony, "The degree of a rational matrix function," lems," y,·ans. AS
1. SIAM, vol. 11, pp. 645-658,1963. 55. _._, "Gn '[he !
36. Falb, P. L., and W. A. Wolovich, "Decoupling in the design of multivariable Congl'. Autom. e
control systems," IEEE 1I-ans. Automatic Control, vol. AC-12, pp. 651-659, 56. - - , "Contribut
1967, vol. 5, pp. 102-1
37. Ferguson, J. D., and Z. V. Rekasius, "Optimallinear control systems with 57. _._, "Canoniea
incomplete state measurements," IEEE 1I'ans. Automatic Control, vol. Acad, Sci. U.S., v
AC-14, pp. 135-140,1969.
58. - - , "On the sta
38. Friedman, B., PrincipIes and Techniques of Applied Mathematics. New Theory, vol. CT-'
York: Wiley, 1956.
59. - - , "Lyapunov
39. Gantmacher, F. R., The Theory of Matrices, vols. 1 and 2. New York: Proc, N atl. Acad.
Chelsea, 1959.
60. - - , "Mathema
40. Gilbert, E. G., "Controllability and observability in multivariable control Control, vol. 1, PI
systems," SIAM J. Control, vol. 1, pp. 128-151,1963.
61.--, "When is a
41. - - , "The decoupling of multivariable systems by state feedback," SI AM vol. 86, pp. 51-6C
J. Control, vol. 7, pp. 50-63, 1969.
62. - - , "Irreducib1
42. Gueguen, C. J., and E. Toumire, "Comments on 'Irreducible Jordan form SI A M 1., vol. 13,
realization of a rational matrix,'" IEEE 1I-ans. Automatic Control, vol.
63. - - , "Toward a
AC-15,1970.
Proc. 4th 1BM Se
43. Hadley, G., Linear AIgebra. Reading, Mass.: Addison-Wesley, 1961.
64. - - , "On structL
44. Halmos, P. R., Finite Dimensional Vector Spaces, 2d ed. Princeton, NJ.: reprint of paper
Van Nostrand, 1950. Automatic Contr
45. Herstein, I. N., Topics in AIgebra. Waltham, Mass.: Blaisdell, 1964. 65" ------, and J. E. 1
46. Heymann, M., "'Comments cOn pole assignrnenl in muill-itlpul COíl­ seco no metnod' (
trollable linear systems,'" IEEE 1I-ans. Automatic Control, vol. AC-13, 1960.
pp. 748-749, 1968. 66. - - , and R. S. Bu

47. Ho, B. L., and R. E. Kalman, "Effeetive construction oflinear state variable 1I-ans. ASME, ser

models from inputfoutput data," Proc. Third Allerton Conf, pp. 449-459, 67. - - , and T. S. En

1965. N ASA, Washingt,

48. Ho, Y. c., "What constitutes a controllable system," 1RE Trans. Automatic 68. - - , P. L. Falb, ¡;
Control, vol. AC-7, p. 76, 1962. . New York: MeGI.
49. Hsu, C. H., and C. T. Chen, "A proof of stability ofmultivariable feedback 69. ---, Y. C. Ho, al
systems," Proc: IEEE, vol. 56, pp. 2061-2062, 1968. systems," Contrib .
50. Jacob, J. P., and E. Polak, "On the inverse ofthe operator (-) =A(J +8(')," 70. 'Kapian, W.,Opel
Am. Math. Monthly, vol. 73, pp. 388-390, 1966. Addison-Westey,
51. Johnson, C. D., and W. M. Wonham, "A note on the transformation to 71. Kreindler, A, and
• _·••.• __
~_L.'_. __
~ _ . ~ •. ~ _. __ ~ . _ _.,---,--._. ._. - ' - - - - _ ~ • ~_~~~ __ . ~ _ . _ , ~ _

REFERENCES 639

of a nonanticipative canonical (phase-variable) form," IEEE 1J·ans. Automatic Control, vol.

. 15, pp. 754-763,1967. AC-9, pp. 312-313,1964.

3.riant systems," IEEE 52. Joseph, P. D., and 1. T. Tou, "On linear control theory," AlEE 1J·ans.
3. Applications and Industry, vol. 80, pt. Ir, pp. 193-196,1961.
k linear time-invariant
53. Jury, E. l., Sampled-Data Control Systems. New York: Wiley, 1958.
.1968.
54. Kalman, R. E., "A new approach to linear filtering and prediction prob­
onal matrix function,"
lems," Trans. AS ME, ser. O, vol. 82, pp. 35-45, 1960.
55. - - , "On the general theory of control systems," Proc. First Intern.
design ofmultivariable Congr. Autom. Control, Butterworth, London, pp. 481-493, 1960.
)1. AC-12, pp. 651-659, 56. - - , "Contribution to the theory of optimal control," Bol. Soco Mat. Mex.,
vol. 5, pp. 102-119, 1960.
.r control systems with 57. - - , "Canonical structure of linear dynamical systems," Proc. N atl.
utomatic Control, vol. Acad. Sci. U.S., vol. 48, no. 4, pp. 596-600, 1962.
58. - - , "On the stability of linear time-varying systems," 1RE 1J'ans. Circuit
'd Mathematics. New Theory, vol. CT-9, pp. 420-423, 1962.
59. - - , "Lyapunov functions for the problems of Lur'e in automatic control,"
1 and 2. New York: Proc. Natl. Acad. Sci. U.S., vol. 49, pp. 201-205, 1962.
60. - - , "Mathematical description of linear dynamical system," SI AM J.
1mu ltivariable control Control, vol. 1, pp. 152-192,1963.
53.
61. - - , "When is a linear control system optimal?" Trans. ASM E, ser. O,
state feedback," SIAM
vol. 86, pp. 51-60, 1964.
62. - - , "Irreducible realizations and the degree of a rational matrix,"
-red ucible J ordan form SIAM J., vol. 13, pp. 520-544,1965.
.utomatic Control, vol. 63. - - , "Toward a theory of difficulty of computation in optimal control,"
Proc. 4th IBM Sci. Computo Symp., pp. 25-43, 1964.
son-Wesley,1961.
64. - - , "On structural properties of linear constant multivariable systems,"
d ed. Princeton, N.J.: l'eprint of papel' 6A, Thil'd Congl'ess of the lnternational Federation of
Automatic Control, 1966.
.: Blaisdell, 1964. M. - - , and J. E. Bertram, "Control system a.nalysis and design vía the
t in muíti-input con­ 'second method' ofLyapunov," "li'ans. ,,;S!Vi E, S¡;;I. S', mi. 82, pp. ~./ i .:;~:,
; Control, vol. AC-13, 1960.
66. - - , and R. S. Bucy, "New results in linear filtering and prediction theory,"
1 of linear
state variable Trans. ASME, ser. O, vol. 83, pp. 95-108,1961.
ton Conf, pp. 449-459, 67. - - , and T. S. Engler, A User's Manualfor Automatic Synthesis Programs,

NASA, Washington, O.e., 1966.

, IRETrans. Automatic 68. - - , P.L. Falb, and M. A. Arbib, Topics in MathematÜ';al System Theory.
New York: McGraw-Hill, 1969.
multivariable feedback 69. - - , Y. e. Ho; and K. S. Narendra, "Co~tiollability of linear dynamical

systems,'; Contrib. Differential Equations, voL l~pp. 189:"213, 1961: .

erator(')=A(') +B(')," 70. Kaplan, W" Operational Methods for Linear Systems. Reading, Mass,:

Addison-Wesley, 1962.

the transformation to 71. Kreindler, A., and P. E. Sarachik, "On the concepts of conlrollability and
640 REFERENCES

observability of linear systems," IEEE Trans. Automatic Control, vol. the 'second me
AC-9, pp. 129-136, 1964. pt. 4, pp. 694-í
72. Kuh, E. S., and R. A. Rohrer, "The state-variable approach to network 91. Polak, E., "An

analysis," Proc. IEEE, vol. 53, pp. 672-686, 1965. system to state

577-579,1966.

73. - - , and - - , Theory of Linear Active Networks. San Francisco:


Holden-Day,1967. 92. Pontryagin, L.
Addison-Wesle
74. Kumar, S., "Computer-aided design of multivariabie systems," M.S.
thesis, State University ofNew York, Stony Brook, 1969. 93. Rekasisu, Z. V.

variable feedba

75. Kuo, F. F., and J. F. Kaiser, System Analysis by Digital Computer. New
York: Wiley, 1966. 94. Sandberg,1. W.

vol. 42, pp. 355

76. LaSalle, l, and S. Lefschetz, Stability of Liapul1ov's Direct Method. New


York: Academic, 1961. 95. --o "On the
tions," Bell S YSi
77. Lefschetz, S:, Differential Equations: Geometric Theory. New York:
Interscience, 1957. 96. Schwartz, L., Ti
78. Luenberger, D. G., "Observing the state of a linear system," IEEE 7I·ans. 97. Schwarz, R. J.,

Military Electronics, vol. MIL-8, pp. 74-80, 1964. Hill, 1965.

79. - - , "Observers for multivariable systems," IEEE Trans. Automatic 98. Silverman, L. 1\
Control, vol. AC-ll, pp. 190-197, 1966. Ph. D. dissertati
80. - - , "Canonical forms for linear multivariable systems," IEEE Trans. 99. - - , "Transfo
Automatic Control, vol. AC-12, pp. 290-293,1967. variable) form,"
1966.
81. Mantey, P.. E.,. "Eigenvalue sensitivity and state-variable selection,"
1EEE 1i·ans. Automatic Control, vol. AC-13, pp. 263-269, 1968. 100. - - , "Stable re:
Conv. Record, VI
82. Mayne, D. Q., "Computational procedure for the minimal realization of
transfer-function matrices," Proc. 1EE (London), vol. 115, pp. 1363-1368, 101. - - , "Synthesi
1968. passive realizati
1968.
83. McMillan, B., "Introduction to formal realizability theory," Bell System
Tech. 1., vol. 31, pp. 217-279,541-600,1952. 102. - - , and B. O.
01 linear system:
84. Morgan, B. S., Jr., "The synthesis of linear multivariable systems by state
variable feedback," Proc. 1964 lACC, pp. 468-472. 103. - - , and H. E
; •. ~.,.., 1­
", "":', .
85. Narendra, K. S., and C. P. Neuman, "Stability oí a class 01 ulfíeremial
equations with a single monotone linearity," SI AM 1. Control, vol. 4, 104. Truxal, J. G., e
pp. 295-308, 1966. 105. Weiss, L., "The
86. Nering, E. D., Linear Algebraand Matrix Theory. NewYork: Wiley, 1963. observability," j
87. Newcomb, R. W., Active Integrated Circuit Synthesis. Englewood Cliffs, 106. - - , "On the st
N.J.: Prentice-Hall, 1968. Princeton Con!
659-680, 1968.
88. Ogata, K., State Space Analysis of Control Systems. Englewood Cliffs,
N.J.: Prentice-Hal1, 1967. 107. ---, "Lectures
590, University ,
89. Panda, S. P., and C. T. Chen, "Irreducible Jordan form realization of a
rational matrix," IEEE Trans. Automatic Control, voL AC-14, pp. 66-69; 108. ----:-:-, and P. L. I
1969. parametrized elf
90. Parks, P. c., "A new proof of the Routh-Hurwitz stability criterion using vol. 3, pp. 67-7':;
REFERENCES 641

utomatic Control, vol. the 'second method' of Lyapunov," Proc. Cambridge Phil. Soc., vol. 58,
pt. 4, pp. 694-720, 1962.
~ approach to network 91. Polak, E., "An algorithm for reducing a linear, time-invariant differential
system to state form," IEEE 7/·ans. Automatic Control, vol. AC-11, pp.
577-579,1966.
'orks. San Francisco:
92. Pontryagin, L. S., Ordinary Differential Equations. Reading, Mass.:
Addison-Wesley, 1962.
lfiable systems," M.S.
(,1969. 93. Rekasisu, Z. V., "Oecoupling of multivariable systems by means of state
variable feedback," Proc. Third Allerton Conf, pp. 439-447, 1965.
)igital Computer. New
94. Sandberg, I. W., "Linear multiloop feedback systems," Bell System Tech. J.,
vol. 42, pp. 355-382, 1963.
s Direct Method. New
95. - - , "On the Lz-boundness of solutions of nonlinear functional equa­
tions," Bell System Tech. 1., vol. 43, pp. 1581-1599, 1964.
Theory. New York:
96. Schwartz, L., Théorie des distributions. Paris: Hermann & Cie, 1951, 1957.
Lr system," IEEE Trans. 97. Schwarz, R. J., and B. Friedland, Linear Systems. New York: McGraw­
Hill, 1965.
EEE 7/·ans. Automatic 98. Silverman, L. M., "Structural properties of time-variable linear systems,"
Ph.O. dissertation, Oept. of Elec. Eng., Columbia University, 1966.
systems," IEEE Trans. 99. - - , 'Transformation of time-variable systems to canonical (phase­
variable) form," IEEE 7/·ans. Automatic Control, vol. AC-ll, pp. 300-303,
1966.
ate-variable selection,"
53-269, 1968. 100. - - , "Stable realization of impulse response matrices," 1967 IEEE I ntern.
Conv. Record, vol. 15, pt. 5, pp. 32-37.
: minimal realization of
'01. 115, pp. 1363-1368, un. --, "Synthesis of impulse response matrices by internally stable and
passive realizations," IEEE Trans. Circuit Theory, vol. CT-15, pp. 238·-245,
1968.
ity theory," Bell System
102. --, and B. O. O. Anderson, "Controllability, observability and stability
of linear systems," SIAM J. Control, vol. 6, pp. 121-129,1968.
ariable systems by state
2. 103. - - , and H. E. Meadows, "Controllability and observability in time­
)f a class of differentiaí
AM J. Control, vol. 4, 104. Truxal, J. G., Control System Synthesis. New York: McGraw-Hill, 1955.
105. Weiss, L., 'The concepts of differential controllability and differential
New York: Wiley, 1963. observability," J. Math. Anal. Appl., vol. 10, pp. 442-449, 1965.
esis. Englewood Cliffs, 106. - - , "On the structure theory of linear differential systems," Proc. Second
Princeton Conf, pp. 243-249, 1968; also SIAM 1. Control, vol. 6, pp.
?ms. Englewood Cliffs,
659-680, 1968.
107. --, "Lectures on controllability and observability," Tech. Note BN­
Lllform realization of a 590, University of Maryland, Jan. 1969:· .
, vol. AC-14, pp. 66-69, 108. --,and P. L. Falp, "Oolezal's theOl:em, linear algebra with continuously
parametriZed elements,and time-varying systems," Math. System TheorJ',
stability criterion using vol. 3, pp. 67-75,1969.
642 REFERENCES

109. - - , and R. E. Kalman, "Contributions to linear system theory," 1m. J. of interconnec


Eng. Soc., vol. 3, pp. 141-171,1965. vol. AC-28, pp
110. Wolovich, N. A., "One state estimation of observable systems," Preprints S9. Armstrong, E.
1968 J ACC, pp. 210-220. control. New

111. Wonham, W. M., "On pole assignment in multi-input controllable linear SlO. Astrom, K. J.,
systems," IEEE Trans. Automatic Control, vol. AC-12, pp. 660-665, 1967. Academic, 19T
112. - - , and A. S. Morse, "Decoupling and pole~assignment in linear multi­ SIL -.- , "Algebn
variable systems-a geometric approach," SI AM J. COl7trol, vol. 8, pp. Polytechnica S.
1-18,1970. S12. Athans, M., M
113. Wylie, e R., Jr., Aduanced Engineering Mathematics. New York: Networks ami
McGraw-Hill,1951. McGraw-Hill,
114. Youla, D. e, 'The synthesis of linear dynamical systems from prescribed S13. Balestrino, A.,
weighting patterns," SI AM 1. Appl. Math., vol. 14, pp. 527-549, 1966. variable systen
matic Control, .
115. - - , and Plinio Tissi, "n-port synthesis via reactance extraction-part I,"
1966 IEEE Intern. Conu. Record, vol. 14, pt. 7, pp. 183-208. S14. Bar-ness, Y., a
sampling," 1ni.
116. Zadeh, L. A., and C. A. Desoer, Linear System Theory. New York:
S15. Barnett, S., fI;.
McGraw-Hill, 1963.
Reinhold, 1971
117. Zames, G., "On the input-output stability of time-varying nonlinear
feedback systems," pts. I and n, IEEE y,·ans. Automatic Control, vol. S16. - - , "Regular
AC-ll, pp. 228-238 and 465~476, 1966. Proc. Cambrid[;
817. Barry, P. E., ":

unpublished ne

Supplementary References
S18. Bartels, R. H., ~
SI. Anderson, B. D. O., "Internal and external stability of linear time­ equation AX +
varying systems," SIAM 1. Control and Optimization, vol. 20, pp. 408-413, 826,1972.
1982. S19. Bengtsson, G.,
S2. - - , and M. R. Gevers, "On multivariable pole-zero cancelIations and dornain approa
the stability of feedback systems," IEEE Trans. Circuits and Syslems, S20. Bhattacharyya,
vol. CAS-28, pp. 830-833, 1981. Automatic COl1l
53, - - , and Jury, ]E. I., "Genera1.ized BezolJtian 2nd Syl,'/F.ste;· 'i'8t,ir:?,;;: ,­
multivariable linear control," JEEE Trans. Automatic Control, vol. AC-21, of rnultivariabl,
pp. 551-556,1976. Atherton, Ed.),
S4. - - , and J. B. Moore, Optimal Filtering. Englewood Cliffs, NJ.: S22. Birkoff, G. and
Prentice-HalI, 1979. York: Macmill
SS. Antsaklis, P. J., and J. B. Pearson, "Stabilization and regulation in S23. Bitmead, R. R.,
linear multivariable systems," IEEE Trans. Automatic Control, vol. common diviso
AC-23, pp. 928-930, 1978. Trans. Automat
S6. Aplevich, J. D., "Direct computation of canonical forms oflinear systems S24. Bongiorno, J. J.
by elementary matrix operations," IEEE' y,"ans. Automatic Control, system," I nI. J.
vol. AC-19, pp. 124-126, 1974. . pp. 183-190, 19
S7. - - , "Tableau methods for analysis' an,d design of linear systems," S25. - - , and - - o
Automatica, vol. 15, pp. 419-429, 1979. back systems in
S8. Araki, M., and M. Saeki, "A quantitative condition for the well-posedness Contl'ol, vol. AC
SUPPLEMENTARY REFERENCES 643

;ystem theory," Int. 1. of interconnected dynamical systems," IEEE Trans. Automatic Control,
vol. AC-28, pp. 569-577, 1983.
)le systems," Preprints S9. Armstrong, E. S., ORACLS: A design system for linear multivariable
control. New York: Dekker, 1980.
)Ut controllable linear S10. Astrom, K. 1., Introduction to Stochastic Control Theory. N ew York:
12, pp. 660-665, 1967. Academic, 1970.
nment in linear multi­ SI L - - , "Algebraic system theory as a tool for regular design," ACTA
Polytechnica Scandinavica, Ma 31, pp. 52-65, 1979.
J. Control, vol. 8, pp.
SU. Athans, M., M. L. Destouzos, R. N. Spann, and S. 1. Mason, Systems,
emaUcs. New York: Networks and Computational Multivariable Methods. New York:
McGraw-Hill,1974.
,tems from prescribed S 13. Balestrino, A., and G. Celentano, "Pole assignment in linear multi­
pp. 527-549,1966. variable systems using observers of reduced order," IEEE ll-cms. Auto­
matic Control, vol. AC-24, pp. 144-146, 1979.
:e extraction-part 1,"
83-208. S14. Bar-ness, Y., and G. Langhotz, "Preservation of controllability under
sampling," Int. J. Control, vol. 22, pp. 39-47, 1975.
Theory. New York:
S15. Barnett, S., Matrices in Control Theory. London: Van Nostrand
Reinhold,1971.
ne-varying nonlinear
S16. - - , "Regular greatest common divisor of two polynomial matrices,"
ltomatic Control, vol.
Proc. Cambridge Philos. Soc., 72, pp. 161-165,1972.
SI7.. Barry, P. E., "State feedback, pole placement and transient response,"
unpublished notes.
S18. Bartels, R. H., and G. H. Stewart, "Algorithm 432, solution of the matrix
bility of linear time­ equation AX +XB =C," Commun. Ass. Comput Mach., vol. 15, pp. 820­
vol. 20, pp. 408-413,
'1,
826,1972.
S19. Bengtsson, G., "Output regulation and internal models-a frequency
:ero cancellations and domain approach," Automatica, vol. 13, pp. 333-345, July 1977.
Circuits and Systems, S20. Bhattacharyya, S. P., "The Structure of Robust Observers," 1EEE 1l-cms.
Automatic Control, vol. AC-21 (4), pp. 581-588, 1976.
Syl'vesteT rD.atrices in.
ic Control, vol. AC-21, of multivariable systems," in Multivariable Technological Systems (D. P.
Atherton, Ed.), New York: Pergamon, 1979.
glewood Cliffs, N.J.: S22. Birkoff, G. and S. MacLane, A Survey of Modern Algebra, 3d ed. New
York: Macmillan, 1965.
)TI and regulation in S23. Bitmead,R. R., S. Y. Kung, B. D. O. Anderson, and T. Kailath, "Greatest
romatic Control, vol. common divisors via generalizedSylvester and Bezout matrices," IEEE
Trans. Automatic Control, vol. AC-23, pp. 1043-1047,1978.
xms oflinear systems S24. Bongiorrio, J. J.; andD. C. Youla, "On observers in multivariable control
. Automatic Control, system," 1m. 1. Control, vol. 8, pp. 221-243, 1968; its discussion; ·Yo.1.12,
pp. 183-190, 1970.
n of linear systems," S25. - - , and - - , "On the design ofsingle-loop single-input-oütput feed­
back systems in thecomplex-frequency domain," 1EEE 1)·ans. Automatic
for the well-posedness Control, vol. AC-22, pp. 416-423,1977.
644 REFERENCES

842. - - , "Minimí;
826. Brasch, F. M., le, and 1. B. Pearson, "Pole assignment using dynamic
puters. vol. C-2
compensator," IEEE 1/·ans. Automatic Control, vol. AC-15 (1), pp. 34-43,
1970. S43. - - "An alg
Ins1., vol. 297, ¡
827. Brockett, R. W., Finite-dimensional Linear Systems. New York: Wiley,
1970. 844. - - , "lnertia
Appl., vol. 49, ~
828. Bruni, c., A. Isidori, and A. Ruberts, "A method of realization based on
the moments of the impulse response matrix," J EEE Tt·ans. Automatic 845. - - , "Synthes
Control, vol. AC-14, pp. 203-204, 1969. vol. 32, pp. 112
829. Brunovsky, P., "A classification of linear control1able systems," Kyber­ 546. - - , Analysis G
netika, vol. 3, pp. 173-188, 1970. Pond Woods, 1
830. Budin, M. A., "Minimal realization of discrete systems from input­ 547. - - , One-dime
output observations," IEEE TI·ans. Automatic Control, vol. AC-16, pp. 1979.
395-401,1971. 848. - - , "Identifi(
831. Byrnes, C. 1., and P. K. Stevens, "Pole placement by static and dynamic Prac. 1982 IF A
output feedback," Proc. 21st IEEE Con! Decision and Control, pp. 130­ 849. - - , "A contril

133, 1982. systems," Pro{;.

832. Cal1ier, F. M. and C. A. Desoer, "An algebra of transfer functions for 850. - - , "Feedbac

distributed linear time-invariant systems" IEEE Trans. Círcuits and Report, SUNY

Systems, vol. CAS-25, pp. 651-662, 1978. lts simplification and clarifica­ 851. - - , and C. H.

tion, vol. CAS-27, pp. 320-323, 1980. systems," Prepr

833. '--, and - - , "Stabilization, tracking and disturbance rejection in 852. - - , and Y. S. f
multivariable convolution systems," 1978 IEEE Con! ón ¡jeCisi~n and from finite inp
Control, San Diego, Calif., 1979. AC-24, pp. 748­
834. - - , and - - , Multivariable Feedback Systems. New York: Springer­ 853. - - , and D. P
Verlag, 1982. IEEE TI'ans. Au
835. - - , and C. D. Nahum, "Necessary and sufficient conditions for the 854. - - , and S.
complete controllability and observability of systems in series using the 5ystem ami Con
coprime factorization of a rational matrix," IEEE 7¡·ans. Circuits and 855. Cheng, L., and j
Systems, CAS-22, pp. 90-95, 1975. variable linear 1
S3~ . Chang, F. C.~ ttThe inverse of gEn~n:1.Hz~d 'O/~1.:-~c!e:'~lT1Gndé /r.L:~·~:·L~ ~~·d·;)L~6~­
the partíal fraction expansion," IEEE Trans. Automatic Control, vol. 19, S56. - - , and - ­
pp. 151-152, 1974. Trans. Automati,
837. Chang, T. S., and C. T Chen, "On the Routh-Hurwitz criterion," IEEE 857. Cheng, V. H. l
Trans. Automatic Control, vol. AC-19, pp. 250-251,1974. transfer functior
838. Chen, C. T., I ntroduction to Linear System Theory. New York: Halt, IEEE Trans. Au.
Rinehart and Winston, 1970. 858. Csaki, F. G., "S
839. -.- , "Irreducibility af dynamical equation realizations of sets of dif­ matrices," IEEE
ferential equations," IEEE Trans. Automatic Control, vol. AC-15, p. 131, 859. D' Alessandro, O
1970. : lordan form rea
S40, -.- ., "A new look at transfer-function design," Proc. IEEE, vol. 59, pp. 1971.
1580--:1585, Nov. 1971. . . 860. Daly, K. C.,"T
841. - - , "A generalizatian of the inertia theorem," SIAM J. Appl. Matiz., elementary sim
vol. 25, no. 2, pp. 158-161,1973. pp. 1-15, 1976.
SUPPLEMENTARY REFERENCES 645

;nment using dynamic 842. - - , "Minimization of linear sequential machines," IEEE 71·a/1s. Com­
1. AC-15 (l), pp. 34-43, puters, vol. C-23, pp. 93-95,1974.
843. --o "An algorithm for Jordan form transformation," J. Fra/1klin
tS. New York: Wiley, Inst., vol. 297, pp. 449-455, 1974.
S44. - - , "Inertia Theorem for general matrix equations," 1. Math. Anal.
)f realization based on Appl., vol. 49, pp. 207-210,1975.
EEE 71·ans. Automatie 845. --o "Synthesis of linear sequential machines," In[ormation Control.
vol. 32, pp. 112-127, 1976.
able systems," Kyber­ S46. - - , Analysis amI Synthesis 01linear control systems. Stony Brook, N.Y.:
Pond Woods, 1978.
systems from input­ S47. - - , One-dimensional Digital Signal Proeessing. New York: Dekker,
lntrol, vol. AC-16, pp. 1979.
848. - - , "Identification of linear time-invariant multivariable systems,"
by static and dynamic Proe. 1982 1FA C Symposium on Ident Uleea ion, J une 1982.
and Control, pp. 130­ 849. - - , "A contribution to the design of linear time-invariant multivariable
systems," Proe. Am. Automatie Control Conf., June, 1982.
transfer fu nctions for 850. - - , "Feedback implementations of open-Ioop compensators." Tech.
~ Trans. Cireuits and Report, SUNY, Stony Brook, N.Y., 1983.
ification and c\arifica­
851. - - , and C. H. Hsu, "Design of dynamic compensator for multivariable
systems," Preprint J A CC, pp. 893-900, 1971.
,turbance rejection in S52. - - , and y.. S. Kao, "Identification of two-dimensional transfer functions
Con! dn Deeision and from finite input-output data," IEEE 71·ans. Automatie Control, vol.
AC-24, pp. 748-752,1979.
New York: Springer­ S53. - - , and D. P. Mital, "A simplified irreducible realization algorithm,
IEEE 71·ans. Automatie Control, vol. AC-17, pp. 535-537,1972.
:nt conditions for the 854. - - , and S. Y. Zhang, "Multivariable multipurpose controllers."
:ms in series using the System and Control le tt as, vol. 3, pp. 73-76, 1983.
'E Trans. Cireuits and 855. Cheng, L., and J. B. Pearson, Jr., "Frequency domain synthesis of multi­
variable linear regulators," IEEE Trans. Automatie Control, vol. AC-23,
:londc iTiali:ix Lhrougá
matie Control, vol. 19, 856. - - , and - - , "Synthesis of linear multivariable regulators," IEEE
71-ans. Automatie Control, vol. AC-26, pp. 194-202, 1981.
'witz criterion," IEEE 857. Cheng, V. H. L., and C. A. Desoer, "Limitations on the closed-Ioop
, 1974. transfer function due to right-half plane transmission zeros of the plant,"
'y. New York: Holt, IEEE Trans. Automatie Control, vol. AC-25, pp. 1218-1220, 1980.
858. Csaki, F. G., "Sorne noteson the inversion of confluent Vandermonde
zations of sets of dif­ matrices," IEEE Trans. AutomatieControl, vol. AC-20, pp. 154-157, 1973.
'01, vol. AC-15, p. 131, 859. D'Alessandro, O. P., anci S: .Guili;ínelli, "A direct procedure for irreducible
'Jordan form realization," Ricnerehe Di Automatiea, vol. 2, pp. 193-198,
·oe. IEEE, vol. 59, pp. 1971.
860. Daly, K. c., "The computationof Llienberger canonical forms using
'IAM 1. Appl. Math., elementary similarity transformations," Int. 1. Systems Sei., vol. 7,
pp. 1-15, 1976.
646 REFERENCES

576. - - , and M.
561. Oamen, A. A. H., and A. K. Hajdasinski, "Practical tests with different stable plant, ~
approximate realizations based on the singular value decomposition of AC-26, pp. 4O~
the Hankel matrix," Proc. 6th 1FAC Symposium on Identification,
Washington, O.C, 1982. 577. - - , R. W. L
design: The fra
862. Oatta, K. B., "An algorithm to compute canonical forms in multivariable 1EEE 7)·ans. A
control systems," IEEE Trans. Automatic Control, vol. AC-22, pp. 129­
S78. - - , and J. D
132, 1977.
and their dyn;
863. - - , "Mínimal realization in companion forms," J. Frank./in Inst., vol. vol. CAS-21 (1
309, pp. 103-123, 1980.
S79. - - , and M..
864. Davison, E. J., "The output control of linear time-invariant multivariable New York: Ac
systems with unmeasurable arbitrary disturbances," IEEE Y¡·ans. Auto­
580. - , and Y..
matic Control, vol. AC-17 (5), pp. 621-630, 1972; its correction, vol.
AC-20, p. 824, 1975. compensator,"
1976.
865. - - , "A generalization of the output control of linear multivariable
systems with unmeasurable arbitrary disturbances," IEEE Y¡·ans. Auto­ 881. - - , and Y. 1
problem: A sel
matic Control, vol. AC-20 (6), pp. 788-792, 1975.
vol 16 (C T. L,
866. - - , "The robust control of a servomechanism problem for linear time
invariant multivariable systems," IEEE Trans. Automatic Control, vol. 882. Oongarra, J. J.,

U ser's Cuide. P

AC-21 (1), pp. 25-34, 1976; addendum, vol. AC-22, p. 283,1977.


867. - - , "Sorne properties of minimum phase systems and 'squared-down' 883. Emami-NaeinL

systems," IEEE Trans. Automatic Control, vol. AC-28, pp. 221-222, 1983. variable robust

Control, pp. 89]


868. - - , and A. Goldenberg, "Robust control of a general servomechanism
problem: The servo compensator," Autoinati~a, vol. 11 (5), pp. 461-471, 884. - - , and P. V
1975. transfer functio
55, 1982.
569. - - , and S. H. Wang, "Properties of linear time-invariant multivariable
885. Emre, E., ''rhe
systems subject to arbitrary output and state feedback/' IEEE Trans.
Automatic Control, vol. AC-18, pp. 24-32, 1973; its correction, vol. dynamic feedb~
AC-18, p. 563, 1973. 620,1980.
870. ~-, and - - , "Properties and calculation of transmission zeros oC 886. - - , "Pole-zer l

linear multivariable systems," Automatica, vol. 10, pp. 643-658, 1974. 1EEE Con! De<

using output feedback," IEEE Trans. Automatic Control, vol. AC-20, tion for relatiVt
pp. 516-518,1975. Control, vol. AC
872. - - , and - - , "An algorithm for the calculation of transmission 888. - - , and - ­
zeros of (C, A, B, O) using high-gain output feedback," IEEE Trans. solutions, "SI A
Automatic Control, vol. AC-23, pp. 738-741, 1978. 889. Eykhoff, P., SY5
873. Oesoer, C. A., Notesfor a Second Course on Linear Systems. Ne:w York: 890. Fadeev, O. K.,
Van Nostrand, 1970. Algebra. San F
874. - - , F. M. Callier, and W. S. Chan; '~Robustness of stability conditions 591. Fahmy, M. M.,
for linear time-invaríant feedback systems," 1EEE Y¡~ans. Automatic multivariable s:
Control, vol. AC-22, pp.. 586-590, 1977. . pp. 690-,-693, 19:
875..- - , and W. S.Chan, "The feedback interconnection of lumped linear 892. Fairman, F. W.
time-invariantsystems," J. Frank.lin Inst., vol. 300, pp. 335-351, 1975~
SUPPLEMENTARY REFERENCES 647

876. - - , and M. J. Chen, "Oesign of multivariable feedback systems with


;al tests with different
stable plant, and its extension," IEEE T/·ans. Automatic Control, vol.
l1ue decomposition of
AC-26, pp. 408-415, pp. 526-527,1981.
um on Identification,
877. - - , R. W. Liu, John Murray, and Richard Sacks, "Feedback system
design: The fractional representation approach to analysis and synthesis,"
forms in multivariable 1EEE Trans. Automatic Control, vol. AC-25, pp. 399-412, 1980.
. vol. AC-22, pp. 129­
878. - - , and J. O. Schulman, "Zeros and poles of matrix transfer functions
and their dynamic interpretation," IEEE Trans.Circuits and Systems,
J. Franklin Inst., vol. vol. CAS-21 (1), pp. 1-8, 1974.
879. - - , and M. Vidyasagar, Feedback Systems: Input-Output Properties.
lVariant multivariable New York: Academic, 1975.
,," IEEE u-ans. Auto­
2; its correction, vol. 880. - - , and Y. W. Wang, "On the minimum order of a robust servo­
compensator," IEEE Trans. Automatic Control, vol. AC-23 (1), pp. 70-73,
1976.
f linear multivariable
,," IEEE Trans. Auto­ 881. - - , and Y. T. Wang, "Linear time-invariant robust servomechanism
problem: A self-contained exposition," in Control and Dynamic Systems,
vol 16 (c. T. Leondes, Ed.). New York: Academic, pp. 81-129,1980.
:oblem for linear time
utomatic Control, vol. 882. Oongarra, J. J., C. B. Moler, J. R. Bunch and G. W. Stewart, LINPACK
2, p. 283, 1977. User's Guide. Philadelphia: SIAM, 1979.
IS and 'squared-down'
883. Emami-Naeini, A., and G. F. Franklin, "Zero assignment in the multi­
28, pp. 221-222,1983. variable robust servomechanism," Proc. 21st IEEE Con! on Decision and
Control, pp.891~893, 1982.
leral servomechanism
11. 11 (5), pp. 461-471, 884. - - , and P. VanDooren, "On computation of transmission zeros and
transfer functions," Proc. 21st IEEE Con! Decision and Control, pp. 51­
55,1982.
lvariantmultivariable
:dback," IEEE Trans. 885. Emre, E., "The polynomial equation QQc +RP c =<1> with application to
i; its correction, vol. dynamic feedback," SIAM 1. Control & Optimization, vol. 18, pp. 611­
620,1980.
:ransmission zeros oí 886. - - , "Pole-zero cancel1ations in dynamic feedback systems," Proc. 21st
pp. 643-658, 1974. 1EEE Con! Decision and Control, pp. 422-423, 1982.
multivariable systems
Control, vol. AC-20, tion for relatively prime polynomial matrices," IEEE Trans. Automatic
Control, vol. AC-22, pp. 239-242, 1977.
.tion of transmission 888. - - , and - - , "The equation XR +QY=<1>: A characterization of
dback," IEEE Trans. solutions, "SIAM 1. Control & Optimization, vol. 19, pp. 32-38, 1981.
889. Eykhoff, P., System Identification. New York: Wiley, 1974.
Systems. New· York: 890. Fadeev, O. K., and V. N. Faddeeva.,Computational Methods in Linear
Algebra. San Francisco: Freeman, 1963.
)f stability coriditions 891. Fahmy, M. M., and J. O'Reilly, "On eigenstructure assignment in linear
EE 1I·ans.. Automatic multivariable systems," IEEE Trans. Automatic Control, vol. AC-27,
.. pp. 690-693," 1982.
tion of lumped linear
892. Fairman, F. W., and R. O.Gupta, "Design of multifunctional reduced
pp. 335-351, 1975.
648 REFERENCES

order observers," Int. 1. Systems Scí., vol. 11, pp. 1083-1094, 1980. SI09. Gopinath, B., ,
Bell Syst. Tech.
S93. Fernando, K. V.~ and H. Nicholson, "Singular perturbatíonal model
reduction of balanced systems," IEEE Trans. Automatic Control, vol. SIlO. Gupta, R. D., !
AC-29, pp. 466-468, 1982. the design of si
vol. CAS-28. Pi
S94. Ferreira, P. G., and S. P. Bhattacharyya, "On blocking zeros," IEEE
8111. Ham, Y. P., an
Trans. Automatic Control, vol. AC-22 (2), pp. 258-259, 1977.
Liapunov theo
S95. F orney, G. D., J f., "Minimal bases of rational vector spaces with applica­ 733- 734,1981­
tions to multivariable linear systems," SIAM J. Control. vol. 13, pp. 493­
SH2. Hayton, G. E.,
520,1975.
32, pp. 567-57~
S96. Fortmann, T. E., and K. L. Hitz, An lntroduction to Linear Control SI13. Heymann, M.,
Systems. New York: Dekker, 1977.
D ynamical Syst
S97. - - , and D. Williamson, "Design of low-order observers for linear S114. - - , 'The po
feedback control laws," IEEE Trans. Automatic Control, vol. AC-17, pp. Control, vol. A(
301:-308,1972.
S115. Hikita, H., "De~
S98. Francis, B. A., "The multivariable servomechanism problem from the to output tracki
input-output viewpoint," 1EEE y,·ans. Automatic Control, vol. AC-22 (3),
S116. Householder, J
pp. 322-328, 1977.
Waltham Mass.
S99. - - , and W. M. Wonham, "The role of transmission zeros in linear S117. Hsia, T. e., SYSI
multivariable regulators," lnt. J. Control, vol. 22 (5), pp. 657-681, 1975.
S118. Hsu, e. H., and
SI00. - - , and - - , "The internal model principie of control theory," Auto­ matrices," lnt. J
matica" vol. 12,pp. 457-465, 1976.
SU9. Ikeda, M., H. 1
S101. Franklin, G. F., and e. R. Johnson, "A condition for full zero assignment linear time-vary
in linear control systems," IEEE Trans. Automatic Control, vol. AC-26, Vol. 13, pp. 304·
pp. 521-523,1981. S120. Inouye, Y., "N
S102. Fuhrmann, P. A., "AIgebraic system theory: An analyst's point of view," discrete-time sy~
J. Franklin Inst., vol. 301, pp. 521-540, 1976. S121. Johnson, e. D..
8103. Garbow, B. S., J. M. Boyle, J. J. Dongarra, and e. B. Moler, "Matrix Regulator and
eigensystem routines-Eispack guide extensions," Lecture Notes in Control, vol. AC
Computer Sciences, vol. 51. New York: Springer-Verlag, 1977.
§¡~. Gentín, Y. V., ano S. Y. Kung, "A two-variabíe approach to the moaei amI Dynamic Sys
reduction problem with Hankel norm-criterion," IEEE y,·ans. Circuits pp. 287-489, 19~
Systems, vol. CAS-28, pp. 912-924, 1981.
S 123. Jury, E. 1., lnner
SI05. Gibson, J. A., and T. T. Ha, "Further to the preservation of controllability 1974.
under sampling," lnt. J. Control, vol. 31, pp. 1013-1026, 1980.
S 124. Kagstrom, B., al
SI06. Gohberg, 1. e., and L. E. Lerer, "Resultants of matrix polynomials," of the Jordan no
Bull. Amer. Math. Soc., vol. 82, pp. 565-567, 1976. M ath. Software, .
S107. Golub, G. H., S. Nash ande. Van Loan, "A Hessenberg-Scheer method
S125. Kailath, T., Linel
for the problem AX + XB ==C," 1EEEY,·ans. Automatic Control, vQ!; AC­
24, pp. 909-913, 1979. . S126. Kalman, R. E.. "(
forros," ACTA, J
8108. - - , and J. H. Wilkinson, "Ill-conditioned eigensystems aI1d the com­
putation oftheJordan canonical form," SIAM Rev., vol. 18, pp. 578-619. S127. - - , "On the COI
1976. SfAMJ. Controi
"--"'-'-'_."."--- _._--- - -- "--'---'----~~~ ~ ~.. _-_ __..-.­
.•

SUPJ>LEMENTARY REFERENCES 649

Sl09. Gopinath, B., "On the control of linear multiple input-output systems,"
083-1094, 1980.
Bell Syst. Tech. 1., vol. 50, pp. 1063-1081,1971.
perturbational model
SIlO. Gupta, R. D., F. W. Fairman, and T Hinamoto, "A direct procedure for
,utomatic Control, vol.
the design of single functional observers," IEEE 7i-ans. Circuits Systems,
vol. CAS-28, pp. 294·-300, 1981.
)Iocking zeros," IEEE SIll. Harn, Y. P., and e T Chen, "A proof of a discrete stability test via the
-259, 1977.
Liapunov theorem," IEEE 7i·ans. Automatíc Control, vol. AC-26, pp.
or spaces with applica­
733-734, 1981.
)I1trol, vol. 13, pp. 493­ S1l2. Hayton, G. E., "The generalized resultant matrix," ¡ni. J. Control, vo1.
32, pp. 567-579, 1980.
~ion to Linear Control S1l3. Heymann, M., Structure and Realization Problems in the Theory oJ
Dynamical Systems. New York: Springer-Verlag, 1975.
:r observers for linear Sl14. - - , "The pole shifting theorem revisited," IEEE 7i-ans. Automatíc
:ontrol, vol. AC-17, pp. Control, vol. AC-24, pp. 479-480, 1979.
SllS. Hikita, H., "Design of exact model matching systems and its applications
ism problem from the to output tracking problems," Int. J. Control, vol. 34, pp. 1095-1111, 1981.
Control, vol. AC-22 (3), S1l6. Householder, A. S., The Theory of Matríces ín Numerical Analysis.
Waltham Mass.: Blaisdell, 1964.
nission zeros in linear S1l7. Hsia, Te, System Identificatíon. Boston: Lexington, 1977.
(5), pp. 657-681,1975. Sl18. Hsu, e H., and e T. Chen, "Conversion of cyclicity for transfer function
control theory," Auto- matrices," Int. J. Control, vol. 16, no. 3, pp. 451-463,1972.
S1l9. Ikeda, M., H. Malda, and S. Kodaina, "Estimation and Feedback in
or full zero assignment linear time-varying systems: A deterministic theory," SIAM 1. Control,
;c Control, vol. AC-26, Vol. 13, pp. 304-326, 1975.
S120. Inouye, Y., "Notes on controllability and constructability of linear
nalyst's point of view," discrete-time systems," I nt. 1. Control, vol. 35, pp. 1081-1084, 1982.
S121. Johnson, e D., "Accommodation of External Disturbances in Linear
e B. Moler, "Matrix Regulator and Servomechanism Problems," IEEE 7i-ans. Automatic
1S," Lecture Notes in Control, vol. AC-16 (6), pp. 635-644 (1971).
[-Verlag, 1977.
SJ.2!..? --., 4'Theory oí disturbance-aCCOIDJIJ.odadng controners~)) ~n ~~(}ntr(}r
lpproach to the model
and Dynamíc Systems, vol. 12 (e T Leondes, ed.). New York: Academic,
, IEEE 1i-ans. Circuíts
pp. 287-489, 1976.
S 123. Jury, E. L, Inners and Stability of Dynamic Systems. New York: Wiley,
lation of controllability
1974.
--1026,1980.
S 124. Kagstrom, B., and A. Rule, "An algorithm for numerical computation
, matrix polynomials,"
of the Jordan normal form of a complex matrix [F2]," ACM Trans.on
Math. Software, vol. 6, pp. 437-443, 1980.
;enberg-Scheer method
SUS. Kailath, T., Linear Syilems. EnglewoodCliffs,N.J.: Prentice-Hall, 1980.

matie.Control, vol. AC­


S 126. Kalman, R. E., "On partial realizations; transfer functions, and canonical

forms," ACTA', Polytechníca Scandinavíca, Mil. 3i, pp. '9-32,1979.

lsystems and the com­

vol. 18,.pp. 578-619,

'V.,
S 127. - - , "On the computation of the reachablejobservable canonical form,';

SIAM J. Control & Optímizatíon, vol. 20, pp. 408-413,1982.


650 REFERENCES

S128. Kamen, E. W., "New results in realization theory for linear time-varying 8146. Lal, M., H. Sí
analytic systems," IEEE Trans. Automatic Control, vol. AC-24, pp. 866­ algorithm for
878,1979. IEEE y,-ans., \
S 129. Kimura, H., "Pole assignment by gain output feedback," IEEE y,·ans. S147. Laub, A. J., "
Automatic Control, vol. AC-20, pp. 509-516,1975. MIT Rep. ESI
S 130. - - , "Pole assignment by gain output feedback," IEEE y,·ans. Automatic 8148. - - , and B. t
Control, vol. AC-22, pp. 458-463, 1977. techniques, Al:
S131. - - , "Geometric structure of observers for linear feedback control S149. Líu, R., and 1.
laws," IEEE Trans. Automatic Control, vol. AC22, pp. 846-854, 1977. fiability of in~
S132. - - , "On pole assignment by output feedback," Int. J. Control, vol. 28, vol. AC-22, pp.
pp. 11-22, 1978. S150. Luenberger, O
S133. Klein, G., and B. e Moore, "Eigenvalue-generalized eigenvector assign­ matic Control, .
ment with state feedback," IEEE y,·ans. Automatic Control, vol. AC-22, S151. MacFarlane, A
pp. 140-141, 1977. Proc. 2 1st lEE.
SI34. Klema, V. e, and A. J. Laub, "The singular value decomposition: Its S152. MacFarlane, A
computation and sorne application," IEEE Trans. Automatic Control, variable systen
vol. AC-25, pp. 164-176, 1980. variable theory
S135. Krishnamurthi, V., "Implications of Routh stability criteria," IEEE 8153. Mansour, M.,
Trans. Automatica Control, vol. AC-25, pp. 554-555, 1980. Lyapunov met
8136. Krishnarao, 1. S., and e T. Chen, "Properness of feedback transfer 707- 708, 1982.
function matrices," Int. J. Control, vol. 39, pp. 57-61, 1984. S154. Maronlas, J., a
S 137. - - , and - - , "Two polynomial matrix operations," IEEE Trans. control system
Automatic Control, vol. AC-29, No. 3, March, 1984. Int. J. Systems:
vol. 10, pp. 33­
S138. Kronsjo, L., AIgorithms - their complexity and efficiency. New York:
Wiley,1979. 8155. Mayne. O. Q.,
rea lization alg<
S139. Kucera, V., Discrete Linear Control- The Polynomial Equation Approach. pp. 306-307. 19
Chichester: Wiley, 1979.
S156. Miminis, G. S.,
SI4(). Kung, S., "Multivariable and multidimensional systems: Analysis and time-invariant l
Oesign," Ph.O. dissertation, Stanford University, 1977.
S157. --o and --­
8141. - - , "A new iow-order approximation a!go!"ithr:0 "/1'" "/-.,h·,,
decomposition," Proc. 12th Amer. Asilomar Con! Circuits, Systems alld pp. 62-67, 1982
Computers, 1978. S158. Mital, O. T., a
S142. - - , and T. Kailath, "Fast projection methods for minimal design time-invariant s
problems in linear system theory," Automatica, vol. 16, pp. 399-403, 1980. 8159. Moler, e B., an
S143. - - , - - , and M. Morf, "A generalized resultant matrix for polynomial the exponential
matrices," Proc. IEEE Corif. on Decision and Control, pp. 892-895, 1976. S16O. Moore, B. e, '
S 144. - - , - - , and - - , "Fast and stable algorithms for minimal design variable system
problems," Proc. Fourth IFAC Int. Symposium on Multivariable Tech. Trans. Aútomati
Systems (D. P. Altherton, ed). Candon: Pergamon, pp. 97-104,1977. vol. AC-22:pp. ¡
1977. '.
S145. Kuo, Y. L., "On the im;ducible Jordan form realization and the degree
of a rational matrix," IEEE Trans. Circuir Theory, vol. CT-17, pp. 322­ S161. - - , "Singular
332,1970. ConI Decision e
SUPPLEMENTARY REFERENCES 651

for linear time-varying 8146. Lal, M., H. Singh, and K. A. Khan, "A simplified minimal realization
r, vol. AC-24, pp. 866­ algorithm for symmetric impulse response matrix using moments,"
IEEE T¡·ans., vol. AC-18, pp. 683-684, 1973.
edback," IEEE Trans. S147. Laub, A. 1., "Linear multivariable control: numerical considerations,"
MIT Rep. ESL-P-833, July 1978.
r EEE Trans. Automatic S148. - - , and B. C. Moore, "Calculation of transmission zeros using QZ
techniques, Automatica, vol. 14, pp. 557-566, 1978.
lear feedback control S149. Liu, R., and L. C. Suen, "Minimal dimensional realízation and identi­
~2, pp. 846-854, 1977. fiability of input-output sequences," IEEE T¡·ans. Automatic Control,
I nI. J. Control, vol. 28, vol. AC-22, pp. 227-232,1977.
SISO. Luenberger, D. G., "An introduction to observers," IEEE T¡·ans. Auto­
l.ed eigenvector assign­ matic Control, vol. AC-16, pp. 596-603, 1971.
ic Control, vol. AC-22, S151. MacFarlane, A. G. J., "Gains, phase and angles in multivariable systems,"
Proc. 21st IEEE Con! on Decision and Control, pp. 944-947, 1982.
ue decomposition: Its S152. MacFarlane, A. G., and N. Karcanias, "Poles and zeros of linear multi­
IS. Automatic Control, variable systems: A survey of the algebraic geometric and complex­
variable theory," /nt. J. Control, vol. 24, pp. 33-74, 1976.
bili ty critería," IEEE S153. Mansour, M., "A note on the stability of linear discrete systems and
;5, 1980.
Lyapunov method," / EEE Trans. Automatic Control, vol. AC-27, pp.
; of feedback transfer
707- 708, 1982.
-61, 1984.
S154. Maronlas, 1., and S. Barnett, "Canonical forms of time-invariant linear
'ations," IEEE Trans.
control systems: a survey with extension. PartI, single-input cases,"
Int. J. Systems Sci., vol. 9, pp. 497-514,1978; Part n, multivariable case,
t

vol. 10, pp. 33-50, 1979.


1Jiciency. New York:

S155. Mayne, D. Q., "An elementary derivation of Rosenbrock's minimal


realization algorithm," IEEE Trans. Automatic Control, vol. AC-18,
ial Equation Approach. pp. 306-307, 1973.
S156. Miminis, G. S., and C. e Paige, "An algorithm for pole assignmenl o[
;ystems: Analysis and time-invariant linear systems," Int. J. Control, vo 1. 35, pp. 341-354, 1982.
1977. S157. - - , and - - , "An algorithm for pole-assignment of timc-invariant
m Vi3 singular value Inuiti-input linear systems,' Froc. ¡¡st IEEE -C'onI uecision ami LoniJOi,
Circuits, Systems and pp. 62-67, 1982.
SI58. Mital, D. T., and C. T. Chen, "Canonical form realization of linear

) for minimal design time-invariant systems," Int. J. Control, vol. 18, pp. 881-887, 1973.

16, pp. 399-403, 1980. S159. Moler, C. B., and C. F. Van Loan, "Nineteen dubious ways to compute

matrix for polynomial the exponential of a matrix," SIAM Rev., vol. 20, pp. 801-836, 1978.

'01, pp. 892-895, 1976. S160. Moore, B. c., "On the ftexibility offered by state feedback il'! multi­

lS for mínimal desígn variable ·systems beyond closed. loop eigenvalue assignment," IEEE

n Multivariable Tech. Trans. Automatic Control, vol. AC-21, pp. 689-692, 1976: its {;omments,

, pp. 97-104, 1977. voL AC-22, pp. 888-889 by Porterand D'Azzo, p. 889 by Gonrishankar,

1977.' .
zation and the degree
, vol. CT-17, pp. 322­ S161. - - , "Singular value analysis of linear systems," in Proc. 17th / EEE
C0171 Decision Control, pp. 66-73,Jan. 1979.
652 REFERENCES

S162. - - , "Principal component analysis in linear systems: Controllability, S179. Porter, B., and
observability and model reduction," IEEE 7)·ans. Automatic Control, time tracking
vol. AC-26, pp. 17-32,1981. 1974.
S163. Morse, A. S., "Minimal solutions to transfer function equations," IEEE S180. Pugh, A. c.,
7)·ans. Automatic Control, vol. AC-21, pp. 131-133,1976. pp. 315-324,
SI64. - - , W. A. Wolovich and B. D. O. Anderson, "Generic pole assignment: 8181. Ralson, A., a
Preliminary results," IEEE Trans. Automatic Control, vol. AC-28, pp. New York: IV
503-506, 1983. S182. Rice, L Matl
S165. - - , and W. M. Wonham, "Decoupling and pole assignment com­ McGraw-Hill
pensation," SIAM 1. Control, vol. 8, pp. 317-337,1970. S183. Rissanen, J,
SI66. Owens, D. R, Multivariableand Optimal Systems. New York: Academic, Control, vol. S
1981. 8184. Roman, J. R.
S167. Pace, 1. S., and S. Barnett, "Comparison of algorithms of calculations of observers for
g.c.d. of polynomials," Int. J. Systems Sci., vol. 4, pp. 211-226, 1973. Trans. AutomG
S168. Pace, 1. S., and S. Barnett, "Efficient algorithms for linear system calcula­ S185. Rosenbrock, I
tions Part 1-Smith form and common divisor of polynomial matrices," Wiley- Intersci
Int. J. Systems Sci., vol. 5, pp. 68-72, 1974. S186. - - , 'The zeI
S169. Paige, C. c., "Properties of numerical algorithms related to computing 8187. - - , "The tn
controllability," IEEE 7)-ans. Automatic Control, vol. AC-26, pp. 130-138, vol. 25, pp. 11
1981. S188. - - , "Comm.
S170. Patel, R. Y., "Computation ofmatrix fraction descriptions of linear time­ A survey of t
invariant systems," IEEE Trans. Automatic Control, vol. AC-26, pp. 148­ Int. J. Control
161,1981. 8189. - - , and G.
S17!. - - , "Computation of numerical-order state-space realizations and Int.1. Control.
observability indices using orthogonal transformation," Int. J. Control, 8190. - - , and A. J
vol. 33, pp. 227-246,1981. 25, pp. 389-39
SIn. - - , and N. Munro, Multivariable System Theoryand Design. New S191. Rozsa, P., and .
York: Pergamon, 1982. of a rational 1
S173. Pearson, J. B., R. W. Shields, and P. W. States, Jr., "Robust solutions to comments by ,
linear multivariable control problems," IEEE Trans. AuforYw(ic "~')!'i ¡'I}i 326. 1978.
vol. AC-19, pp. 508-517, 1974. SI92. Saeks, R., and
S174. - - , and P. W. Staats, Jr., "Robust controllers for linear regulators," disturbance re
IEEE Trans. Automatic Control, vol. AC-19 (3), pp. 231-234,1974. Ae-26, pp. 203
S175. Pernebo, L., "An algebraic theory for the design of controllers for linear S193. Sain, M. K., '
multivariable system. Part 1: Structure matrices and feedforward designo Academic, 198
Part 11: Feedback realizations and feedbackdesign, IEEE 7)-ans. Auto­ S194. Safonov, M. G
matic Control, vol. AC-26, pp: 171-172, pp. 1~3-194, 1981. of multivariab
S176. - - , and L. M. Silverman,:'Model reduction via balanced state space . matrix," IEEE
representation," 1EEE Trans. Automatic Control, vol. AC-2l, pp. 382-387, S195. Schumacher, J
1982. . 1EEE 7)·ans. Al
SI77.Popov, Y. M., "Invariant description oflinear time-in~aiiantcontrol1able SI%. Scott, R.W., al
systeITIs," SI AM 1. Control, vol. 10, pp. 252-264, 1972. feedback transf
S178. - - , Hyperstability 01 Control Systems. Berlin: Springer-Verlag, 1973. vol. AC-21, pp.
SUPPLEMENTARY REFERENCES 653

stems: Control1ability,
S179. Porter, B., and A. Bradshaw, "Design of linear multivariable continuous­
lS. Automat ic Control,
time tracking systems," Int. J. Syste111s Sci., vo\. 5 (12), pp. 1155-1164,
1974.
tion equations," IEEE
S180. Pugh, A. c., "Transmission and system zeros," Int. 1. Contro!' vol. 26,
" 1976.
pp. 315-324, 1979.
:neric pole assignment:
S181\ Ralson, A., and P. Rabinowitz, A First Cow'se in Numerical Analysis.
,ntrol, vo\. AC-28, pp. New York: McGraw-Hil1, 1978.
S182. Rice, J., Matrix Computation and Mathematical Software. New York:
pole assignment com­ McGraw-Hil1,1981.
1970. S183. Rissanen, J., "Recursive identification of linear systems," SI AM J.
New York: Academic, Control, vol. 9, pp. 420-430,1971.
SI84. Roman, J. R., and T. E. Bullock, "Design of minimum order stab1e
thms of ca1culations of observers for linear functions of the state via realization theory," IEEE
>p. 211-226, 1973. Trans. Automatic Control, vol. AC-20, pp. 613-622,1975.
r linear system calcula­ S185. Rosenbrock, H. H., State-Space and Multivariable Theory. New York:
po1ynomial matrices," Wiley-Interscience, 1970.
S186. - - , "The zeros ofa system," Int. J. Control, vol. 18, pp. 297-299,1973.
; related to computing S187. - - , "The transformation of strict system equiva1ues," Int. J. Control,
)1. AC-26, pp. 130-138, vol. 25, pp. 11-19,1977.
S188. - - , "Comments on 'Poles and zeros of linear multivariable systems:
riptions of linear tíme­ A survey of the a1gebraic, geometric and complex variable theory,'"
1/, voL AC-26, pp. 148­ Int. 1. Corarol, vol. 26, pp. 157-161, 1977.
S189. - - , and G. E. Hayton, "The general prob1em of po1e placement,"
pace realizations and Int.1. Control, vol. 27, pp, 837-852,1978.
lUon," Int. J. Control, S190. --', and A. J. 1. van der Weiden, "lnverse System," Int. J. Control, vol.
25, pp. 389-392, 1977.
)ry and Design. New S191. Rozsa, P., and N. K. Sinha, "Efficient algorithm for irreducible realizatían
of a rational matrix," Int. J. Control, vol. 20, pp. 739-748, 1974. lts
, "Robust solutions to comments by Y. S. Kao and C. T. Chen, Int. 1. Control, vol. 28, pp. 325­
ns. Automatic Control, 326,1978.
SI92. Saeks, R., and 1. Murray, "Feedback system design: The tracking and
for linear regulators," disturbance rejectíon prob1ems," IEEE Ti·ans. Automatic Control, vol.
),231-234,1974. AC-26, pp. 203-217, 1981.
f eontrollers for linear S193. Sain, M. K., Introduction to Algebraic System Theory. New York:
nd feedforward designo Academic, 1981.
~n, IEEE Trans: Auto­ 8194. Safonov, M. G., A. 1. Laub, and G. L. Hartmann, "Feedback properties
94, 1981.
of multivariable systems. The role and use of the return diffetenee
l balanced state space
matrix," IEEE Trans., vol. AC-26, pp. 47-65, 1981.
,\. AC-21,pp. 382-387,
8195. Schumacher, J. M., "Compensator synthesis using (C, A, B)~p'airs,"
IEEE Trans. Automatic Control, vol. AC-25; pp. 1133-1137, 1980' .
·invariant co ntrollable
S196. Seott, R. W., and B. D. O. Anderson, "Comments on 'Conditíons fora
972.
feedback transfer matrix to be proper,''' IEEE Ti·ans. Automatic Control,
Jringer-Verlag, 1973.
voI.AC-21, pp. 632-634,1976.
654 REFERENCES

8197. Shieh, L. S., and Y. T. Tsay, "Transformations of a class of multivariable 8216. W olovich, W
control systems to block companion forms," IEEE Trans. Automatic IEEE Ti'ans. A
Control, vol. AC-27, pp. 199-203,1982. 8217. - - , "Multiv
8198. Silverman, L., "Realization of linear dynamical systems," IEEE Ti·ans. IEEE Ti'al1s. 1
Automatic Control, vol. AC-16, pp. 554-567, 1971. 8218. - - , Linear ~
8199. Smith, M. e, "Matrix fractions and strict system equivalence," Int. 1. 8219. - - , "Outpu
Control, vol. 34, pp. 869-884,1981. vol. AC-20, pr
8200. Stewart, G. W., Introduction to Matrix Computations. New York: S22{t - - , "Sket pr
Academic,1973. vol. AC-23, Pf
8201. Tsui, C. e, and e T. Chen, "An algorithm for companion form realiza­ 8221. - - , "On the
tion," Int. J. Control, vol. 38, pp. 769-777,1983. Ti·ans. Automa
8202. - - , "Computational aspects in linear system theory," Ph.D. Disserta­ 8222. - - , "Multípl
tion, SUNY, Stony Brook, 1983. Automatic Ca;
8203. Van Dooren, P. M., "The generalized eigenstructure problem in linear 8223. - - , P. Ants
system theory," IEEE Trans. Automatic Control, vol. AC-26, pp. 111-129, minimal and
1981. Control, vol. A
8204. Varga, A., and V. Sima, "Numerically stable algorithm for transfer 8224. - - , and P. I
function matrix evaluation," Int. J. Control, vol. 33, pp. 1123-1133, 1981. variable syster
8205. Vidyasagar, M., "Conditions for a feedback transfer function matrix to 465,1979.
be proper," IEEE Trans. Automatic Control, vol. AC-20, pp. 570-571, 8225. Wonham, W.
1975. systems?" SI ~l
8206. - - , Nonlinear Systems Analysis. Englewood Cliffs, N.J.: Prentice-Hall, 8226. - - , "Towan
1978. System, Man ú
8207. - - , "On the well-posedness of large-scale interconnected systems," S227. - - , Linear
IEEE Trans. Automatic Control, vol. AC-25, pp. 413-421, 1980. New York: Sp
8208. Wang, J. W., and e T. Chen, "On the computation of the characteristic 8228. - - , and J. B.
polynomial of a matrix," IEEE Trans. Automatic Control, voL AC-27, multivariable s
pp. 449-551, 1982. 8229. Wu, M. Y., "A
8209. Wang, S. H., Design of Linear Multivariable Systems, Memo. No. ERL Trans. Automal
Ivn09, University of ü3.liforrüa, Berkeley, 1971 J~iÍ(jl. --,'':'ome (il

8210. - - , and E. J. Davison, "A minimization algorithm for the design of 1EEE Trans. A
linear multivariable systems," IEEE Trans. Automatic Control, vol.
8231. - - , "SolutiOI
AC-18, pp. 220-225, 1973.
J. Control, vol.
8211. - - , and - - , "A new invertability criterion for linear multivariable
8232. - - , "A succe
systems," IEEE Trans. Automatic Control, vol. AC-18, pp. 538-539, 1973.
time varying sy
8212. Wilkinson, J. H., The Algebraic Eigenvalue Problem. London: Oxford
University Press, 1965. 8233. Wu, y, e, ane
dynamíc systen
8213. - - , and .C. Reinsch (Eds.), Handbook for Automatic Computation, 504, 1980.
Linear Algebra, Vol. 11. New York:' Springer-Verlag, 1971.
. . 8234. Yeung, K. S., "
8214. Willems; J. c., The Analysis of FeedbackSystems. Cambridge: MIT
Trans. Automat.
Press, 1971.
8215. Willems, J. L., "Designof state observerS for linear discrete-time systems," 8235. Youla, D. e. J
Int. J. Systems Sci., vol. 11, pp. 139-147,1980. design of optirr
SUPPLEMENTARY REFERENCES 655

lclass of multivariable S216. Wolovich, W. A., "On determining the zeros of state-space systems,"
SEE Trans. Automatie IEEE 7]-ans. Automatie Control, vol. AC-18 (5), pp. 542-544, 1973.
S217. - - , "Multivariable system synthesis with step disturbances rejection,"
;ystems," IEEE Trans. IEEE 7]-ans. Automatie Control, vol. AC-19 (1), pp. 127-130,1974.
5218. --o Linear Multivariable Systems. New York: Springer-Verlag, 1974.
n equivalence," Int. J. S219. - - , "Output feedback decoupling," IEEE 7]-ans. Automatie COl1trol,
vo!. AC-20, pp. 148-149,1975.
ltations. New York: S220. - - , "Sket prime polynomial matrices," IEEE Trans. Automatie Control,
vol. AC-23, pp. 880-887, 1978.
mpanion form realiza­ S221. - - , "On the stabilization of closed-Ioop stabilizable systems," IEEE
7]-ans. Automatie Control, vol. AC-23, pp. 1103-1104, 1978.
eory," Ph.D. Disserta­ S222. - - , "Multipurpose controllers for multivariable systems," IEEE y,-ans.
Automatic Control, vol. AC-26, pp. 162-170, 1981.
:ure problem in linear S223. --, P. Antsaklis, and H. Elliott, "On the stability of solutions to
.1. AC-26, pp. 111-129, minimal and nonminimal design problems," IEEE Trans. Automatie
Control, vol. AC-22, pp. 88-94, 1977.
.1gorithm for transfer S224. - - , and P. Ferreira, "Output regulation and tracking in linear multi­
, pp. 1123-1133, 1981. variable systems," IEEE 7]-ans. Automatie Control, vol. AC-24, pp. 460­
fer function matrix to 465, 1979.
AC-20, pp. 570-571, S225. Wonham, W. M., "Tracking and regulation in linear multivariable
systems," SIAM J. Control, vol. 11, pp. 424-437,1973.
Is, N.J.: Prentice-Hall, S226. - - , "Towards an abstract internal model principIe," IEEE y,-ans.
System, Man and Cybernetics, vol. SMC-6 (11), pp. 735-740, 1976.
:rconnected systems," 8227.. --, Linear Multivariable Control: A Geornetrie Approaeh, 2d ed.
.3-421, 1980.
New York: Springer-Verlag, 1979.
,n of the characteristic
S228. - - , and J. B. Pearson, "Regulation and internal stabilization in linear
. Control, vol. AC-27,
multivariable systems," SIAM J Control, vol. 12, pp. 5--8,1974.
S229. Wu, M. Y., "A note on stability of linear time-varying systems," IEEE
?ms, Memo. No. ERL
Trans. Automatie Control, vol. AC-19, p. 162,1974.
§23®. - - , ·...S ome new resuits in ¡inear Lime-varying syslems" (lecn. conesp.),
thm for the design of
IEEE y,-ans. Automatie Control, vol. AC-20, pp. 159-161, 1975.
tomatie Control, vol.

S231. ---, "Solutions of certain classes of linear time varying systems," Int.
J. Control, vol. 31, pp. 11-20,1980.
r linear multivariable
S232. - - , "A successive decomposition method for the solution of linear
18, pp. 538-539, 1973.
time varying systems," Int. J. Control, vol. 33, pp. 181-186, 1981.
~m. London: Oxford
S233. Wu, Y. c., and Z. V. Rekasius, "Deterministic identification of linear
dynamic systems," IEEE Trans. A~tomatie Control, vol. AC-25~pp. 501­
tomatie Computation,
504,1980.
,lag, 1971.
S234.. Yeung, K. S., ~'A necessary condUíón for H urwitz polynorIlials;'" IEEE ..
!s. Cambridge: MIT
Trans. Automatie Control, vol. AC-27, pp. 251-252, 1982,

liscrete-time systems," S235. Youla, D. c., 1. J. Bongiorno, and H. A. labr, "Modern Wiener-Hopf
design of optima] controllers- Part 1 and Part n." IEEE y,-uns. Auto­
656 REFERENCES

matic Control, vol. AC-21, pp. 3-15, pp. 319-338, 1976.


8236. Zhang, S. Y, and C. T. Chen, "An algorithm for the division of two poly­
nomial matrices," IEEE Trans. Automatic Control, vol. AC-28, pp. 238­
240,1983.
8237. - - , and - - , "Design of unity feedback systems for arbitrary denomi­
nator matrix," IEEE Trans. Automatic Control, vol. AC-28, pp. 518-521,
1983.
8238. - - , and - - , "Design of robust asymptotic tracking and disturbance
rejection," Proc. 22nd IEEE Cotif. Decision and Control, Oec. 1983.
8239. IEEE Trans. Automatic Control 1956-1980 Cumulative Index, vol. AC­
26, no. 4, August 1981.

Additivity, 73,88
Adjoint equation, 163,
Algorithm
Euclidean, 579
Feedback implement
Leverrier,67
pole assignment, 476
row searching, 550
state estimator, 359, :
state feedback, 337, 3
triangularization, 58S
Analytic continuation, :
Analytic funclion, 78, 5:
Asymptolic stabilily, 4C

Asym pto tic tracking p


495

Basis, 14
change of, 18, 22
minimal polynomial,
orthonormal, 17,90
polynomial, 607, 621
Be·zout identity, 5.95
Black box,· 72 .
Blocking zer6; 499, 633
Bounded-input bc>undé
stabíHty, 385, 394
Canonical dccomposilic
1976.

le division oftwo poly­

1, vol. AC-28, pp. 238­

; for arbitrary denomi­


1. AC-28, pp. 518-521,

cking and disturbance !ndex


'antrol, Dec. 1983.
tlative Index, vol. AC-

Additivity, 73, 88
Canonical-form dynamical equation

Adjoinl eqllation, 163, 166, 195


controllable, 243, 284, 327, 346

AIgorilhm
Jordan, 150,209

ElIclidean, 579
observable, 241, 286, 329

Feedback implementation, 518, 520


Causality, 76

Levenier, 67
Cayley-Hamilton theorem, 49, 67

poIe assignment, 476


Characteristic polynomial

row searching, 550


of constant matrix, 33, 65

stale estimator, 359, 364


of ralional matrix, 235

state feedback, 337, 347


Characterization, complete, 433

trianglllarizalion, 589
Cholesky decomposilion, 413

Analytic conlinllalion, 554


Colllmn degree coeftkient matrix, 601

Analytic fllnctíon, 78, 554


Companion matrix, 26, 64, 346, 349

Asymplotic stability, 403, 409


block, 223

unifGnn!y, 403

Asymptolic trackíng problem, 340, 488,


eigenveclors of, 64, 65

495
Compensator equation, 460

Composite system, 433

Basis, 14

Controllability, 176

change of, 18,22

dilTerential, 180

minimal polynomial, 608

grammian, 184,322

orthonormal, 17, 90

index, 188

polynomial, 607,621

índices, 190

Bezout identity, 595

instantaneous, 181

Siack box, 72

matríx, 184, 200

Blockíng zero, 499, 633

output, 214

Bounded~input bOllnded-olllpllt" (BlBO)

output function, 216

. stabÜiiy, 385, 394 .

Popov-Belevitch-Halltlls test, 231

Canonical decomposition lheorem, 203


unifonn, 182

657
658 INDEX

Coprimeness, 579
(Dynamical equation)
(Function)

left,592
solution of, 134
integrable, 388

right,592
linear, 19

Convolution integral, 81
of matrix, 45

Cyclic matrix, 68, 230, 342


Eigenvalue,33
Function reproduc¡bil¡t~

assignment, 336
Function space, 10, 391

computation of, 61, 553


Functional estimator, 3é

Dead beat control, 382


geometric multiplicity of, 41,66
Fundamental cutset, \O:

Decoupling,372
index of, 47, 66
Fundamental loop, \02

static, 502
multiplicity or, 41, 66
Fundamental rnatrjx, 13

Degree,577 Eigenvector,33

column,600 assignment of, 351

row,600 generalized, 39
Gaussian elimination, 61

Degree of rational malrix, 235


grade of, 38
with complete pivolin,

computation of, 315


Elementary matrix, 542, 587
wíth partíal pivoling, (

Delta function, 74
Elementary operation, 542, 587
Generalized eigenvector,

Description of systern
Equilibrium state, 401
chain of, 39

differentia1 operator, 292


Equivalence, 146,200,296
grade of, 38

external, 3, 71
in the sense of Lyapunov (i.s.L.), 153
Gram determinant, 172,

input-outpul, 71, 72
strict system, 294
Grammian, controlIabili

internal, 3, 71
transrormation, 146, 151,200
observability, 197

polynomial matrix, 125, 287


zero-inpu 1, 148

state-variable, 71, 83
zero-state, 148

Diophantine equation, 460


Estimator, state, 281

Direct sum, 110


exp At, 55, 56
Hankel matrix, 245, 257,

Discrete-time, systern, 121,301


computation of, 143
Hermite form, 589

dynarnical equation, 123,226


Laplace transform of, 56
Hermitian matrix, 565

identification of, 300


properties or, 55
nonnegative definite, 4

Discretization of continuous-time equa­


Exponential stability, 333
positive definite, 413

tion, 560
positive semidefinite, 4

Disturbance signal, 488


Hessenberg form, 221,55
additive,494
Feedback syslem, 324
Homogeneíty, 73,88
rejection of, 488,495
constanl oulput, 377,445
Householder transformat
waveform strudured, 494
inpllt-outpllt, 432
wilh pivoting, 545

Divisor, 579
outpul, 343, 377, 379,474
Hurwitz polynomial, 395

HUf\N!í.Z sta b!1it'l !~;:::1 ~9~


cornmon, 579
plant inpul-output, 432

greatest comí11on, 579, :>92


state, 324, 341, 344, 379

Iefl,592
unity, 432, 457

right, 592
wcll-posedness or, 118

Identification, determínisl

Domain, of function, 19
Feedback gain malrix, 341, 348

Impulse function, 74

of linear operator, 27
Fíeld,8

Impulse response, 75

Duality, theorern of, 195


of complex numbers, 9

matrix,76

Dynamical equation, 71, 87


of rational functions; 9

Inner product, 59

computer simulat.ion of, 91


of real numbers, 9

Input-output pair, 87

equivalent, 146
Floquet, theory of, 154

Internal ruodel principie, '

input-normal, 322
Fraction,599

Inverse problem, 217, 529

internally balanced, 222


coprime, 605

Irreducible dynamical eql

lordan-form, 150,209
irreducible, 605

irreducible polynomíal.1

of RLC network, .101


Frobenius form, 64

tion,291

output~normal, 322
block,223

Irreducible realizatíon, 20

reducible, 207
Function, analytic 554

INDEX 659

(Function)
J ordan block, 38, 49

ation)

integrable, 388
Jordan-canonical-form matrix, 37,41,53,

14

línear, 19
60,209

of matrix, 45

Function reproducibility, 216

Function space, 10, 391


L p sta bility, 390

36

Functíonal estímator, 369


Laplace transform, 56, 81

)f, 61, 553

Fundamental cutset, 102


Lcverrier algorithm, 67

ltiplicity of, 41, 66

Fundamental loop, 102


Linear algebraic equations, 26

Fundamental matrix, 136


Linear independence

',41,66

of time functíons, 170

of vectors, 12, 13, 40, 41

351

)
Linear mapping, 20

Gaussian elimination, 61, 543

Linear operator, 20

with complete pivoting, 544

matrix rcpresentation of, 21

'ix, 542, 587


with partial pivoling, 61, 544

change of basis, 22

alion, 542,587
Generalized eigenvector, 38

nullity, 29, 31

~, 401
chain of, 39

null space of, 29

,200,296
grade of, 38

range space of, 27

Lyapunov (i.s.L.), 153


Gram determinant, 172, 184

Linear space,9

94
Grammian, controllability, 184

basis of, 14

1, 146, 151,200
observability, 197

dimension of, 13

Linear system, 73, 87

Linear time-invariant system, 80, 89

281

Hankel matrix, 245, 257, 265


Linearity, 20,73,87

Hermite form, 589"


Link, 102

f,143

Hermitian matrix, 565


Lyapunov equation, 572

)rm of, 56

nonnegative definite, 413


Lyapunov function, 415

positive definite, 413


Lyapunov theorem, 414, 421

lit y, 333

positive semidellnite, 413


Lyapunov transformation, 153,428

Hessenberg form, 221, 552

324
Homogeneity, 73, 88

Householder transformation, 61,544


Markov paral1lelerS, 245, 301

t, 377, 445

with pivoting, 545


Mason's gain formula, 158

32

Hurwitz polynomial, 395


Matrix, 7, 24

7,379,474

Hurwitz stabilitv tesl, 39P.


·;Cf~\i('., 6P., 230. 3¿l2

put,432

echelon form, 307, 589

344, 379

elementary, 542, 587

function of, 51

of,118

Identification, deterministic, 302


fundamental, 136

ltrix, 341,348

Impulse funclion, 74
Hermite, 589

Impulse response, 75
hermitian,412

Ibers,9

matrix,76
nonsingular, 28, 32

:ions,9

Inner product, 59
norm of, 57

,9
Input-output pair, 87
orthogonal, 220, 570

,154
Internal model principIe, 490
polynomial, 587

Inverse problem, 217, 529


polynomial of, 45

Irreducible dynamical equalion, 207


" principal minar, 413

Irreducible polynomial matrix dcscri p­ leading,413

tion,291
rank or, 27

Irreducible realization, 208, 233


singular value of, 568

554

660 INDEX

(Matrix)
Parameter perturbation, 491 Realization, 155
stale transition, 137
Parameterization, 540 input-normal, 322
symmetric, 412
Periodically varying system, 153 internally balanced, 2".
Sylvester, 582
Physica\ system, 1 irreducible, 208
system, 293, 439
Pole, 441, 623 minimal-dimensional,
trace of, 67,129
Pole-zero cancellation, 437,443,444 of time-varying diffe
unitary, 220
unstable, 444, 459 252

Minimal design problem, 529 Pole-zero excess inequality, 459, 487 output-normal, 322

Minimal-dimensional realization, 233 Polynomial matrix, .587 partíal, 302

Minimal polynomial, 46 column reduced, 601


reducible, 207

Minimum-energy control, 178,558 echelon form, 306, 612


Reachability, 176,227
Mode,145 Hermite form, 589
Regu\ator problem, 340
Model,2 Popov form, 612
Relaxedness, 77,79, 161
matching, 529 row reduced, 601
Representa lion
simplified,271
unimodular, 591
of linear operator, 21
Module, 588, 607
Popov-Belevich-Hautus test, 231 of vector, 15
Monic polynomial, 46, 577
Positive definite matrix, 171, 182, 184 Response, 72
Moment,301
Positive semidefinite matrix, 413 impulse, 75, 126
Predictable degree properly, 541 steady-state, 494, SOl
Primary dependenl row, 247, 257, 262, step, 107, 129
Nilpotent matrix, 69
272, 306, 610 lransient, 494, 50 I
Nonanticipalive system, 76
Principie minor, 413 zero-input, 89, 140
Nonlinear system, 73, 87
leading, 413 zero-state, 89, 140
Norm, of matrix, 57
Proper rational matrix, 82 ReSl1ltanl, 582
Euclidean, 57
characteristic polynomial of, 235 generalized, 608
induct:d,57
column index of, 468,618 . Riccati equation, 68
of vector, 57
cyclic, 469 Ring,9, 588
Normal trce, 101
degree of, 235, 315 Robust design, 491, SOl
Nullity, 29
row index of, 466, 613 Routh-Hurwitz criterion,
Null space, 29, 284
Pseudoinverse, 270 Roulh table, 397
left, 29, 69
Psel1dostate, 288, 293, 295 Row degree coefficient m~
right,29
Pulse funclion, 74 Row searching algorithm,

QR algorithm, 553 Sampled transfer funclion


Observability, 192, 197
QR faclorizatioD, 553
diírerential, 196 ~;<-iÚliJlillb iü:::'ldlHS, .JjSJ
Ql1adratic form, 411.
grammian, 197, 322
Sampling period, 121,559
index,199
scalar,8
indices, 199, 264
Sca\ar producl, 59
Range
instantaneous, 197
Schmidt orthonormalizat¡
of function, 19
matrix, 198
Schwarz inequality, 59, 69
of linear operator, 27
uniform, 197
Separation property, 367
Rank, of matrix, 27, 28, 32
Observer (see State estimator) computation of, 546 Sequence, input, 121
Open-loop control system, 324 of product of. matrices, 31 output, 121
Orthogonal matrix, 570 Rational functian, 82 persisten tIy exciting, 30' .
Orthogonality of vectors, 566 irreducible, 584 weighling, 121
Orthonormal set, 17,566 reducible, 584 z transform of, 122
Output controllability, 214 Rational mati-íx, 82 Servocompensator, 504
Output equation, 87, 89 proper,82 Shifting property; 80, 89
Output function controllability, 216 strictly proper,82 Similar matrices, 23
INDEX 661

Realizatíon, 155
Similarity transformation, 23, 60

Jrbation, 491

input-normal, 322
Simulation ofdynamical equation, 91

!n,540
internally balanced, 271, 322
analog computer, 92

'ying system, 153

irreducible, 208
digital computer, 93

,1

minimal-dimensional, 233
Singular value decomposition, 61,269,569
of time-varyíng differential equation,
Solution space, 135

lIation,437, 443,444
252
Stability, asymptotic, 403, 409

.459

output-normal,322
bounded-input bounded-output, 385

; ineq uality, 459,487

partíal, 302
exponential, 405, 409

:rix, 587

reducible, 207
global, 403

:ed,601
Reachability, 176,227
in the sense of Lyapunov (i.s.L.), 402

306,612

Regulator problem, 340


total, 407, 411

,589

Relaxedness, 77, 79, 161


uniform, 402, 408

512

Representation
Stabilization, 339,490
601

of linear operator, 2 I
Statc, 83 -86

i91

of vector, 15
definitíon of, 83

-Hautus test, 231

Response, 72
equilibrium,401

matrix, 171,182,184

impulse, 75, 126


initial,89

'Inite matrix, 4 I3

steady-state, 494, 501


pseudo-, 288

'ee property, 541

step, 107, 129


variable, 86

lent row, 247, 257, 262,

transient, 494, 501


vector, 86

O
zero-input,89, 140
State equation, 87, 89, 134

413

zero-state, 89, 140


State estímalor, 354

Resultanl,582 algorithm of, 359

matrix,82

generalized, 608
asymptolic, 356

polynomial of, 235

Riccati equation, 68
full-dimensiollal,355

of,468,618

Ring, 9, 588
fUllctional, 369

Robusl design, 491, 501


open-Ioop,355

315

Routh-Hurwitz criterion, 398,417


reduced-dimensiollal,361

l66,613
Routh table, 397
State feedback, 334

70

Row degree coefficient matrix., 601


Slate space, 90

,293,295

Row searching algorithm, 247, 257,272


State transition matrix, 137

Subspace, 11,27

Superposition principie, 73

Sampled transfer functíon, 122


Sylvester's inequality, 31

53

matrix, 123
Svivestcr nl.rJlrix; SR2­
,553

S~nllplii1g insiarn.s, 550


Syslem, ¿

4í2

Sampling períod, 121,559


causal, 70

scalar,8
continuous-time,121

Scalar product, 59
discrete-time, 121

Schmídt orthonormalization, 567


linear, 73, 87

Schwarz inequality, 59, 69


matrix,293

tor, 27

Separation property, 367


multivariable, 71

27,28,32

Sequence, input, 121


.reduced, 323

f,546

output, 121
relaxed,·385

latriees, 31

persistently exciting, 304, 308


single-variable,7I

\,82

weightíng, 121
time-invariant, 80, 89

r.
z transform of, 122
zero-memory, 72

Servocompensator, 504

32

Shifting property, 80, 89


Time invariance, 80, 89

Similar matrices, 23
Time varying, 80,89
82

662 INDEX

Total stability (T-stability), 407, 411


Vector, 9

Trace of matrix, 67, 129


normalized, 566

Transfer function, 82
representation of, 15

matrix, 82, 90
Vector space, 9

sampled, 122
basis of, 14

Transformation, 143
complex,10

elementary, 542, 587


dimension of, 13

equivalence, 146, 151,200


ralional, 10

Householder, 220, 544


real, la

Lyapunov, 153
see also Linear space

orthogonal, 220,222

Transmission blocking property, 626, 629

Transmission zero, 443, 496


Well posedness, 118

Tree,102
Weighling sequence, 121

branch,102

normal, 102

Truncalion operator, 125

Zero,623

blocking, 499,633

Uniform controllability, 182


input-decoupling,292

Uniform observability, 197


output-decoupling, 292

Unimodular matrix, 591


transmission, 443, 627

Uniqueness of solution of differential


Zero-input response, 89, 140

equation, 87, 89
Zero-memory system, 72

Unitary matrix, 269, 568


Zero-state equivalence, 148

Unit-time-delay system, 77, 82, 86


Zero-state response, 89, 140

z-transfer function, 122

Vandermonde determinant, 64
z transform, 122

66
1 or, 1S

13

r space

118

once, 121

633

ng, 292

.ling,292

143,627

me, 89, 140

,tem,72

ilence, 148

lse, 89, 140

-n,122

l
:j

1
1

------_._._---'"_._--"----~._-~-----------_._-------~----~----_._---~--_._--------------~---------------------------~--'"._-----

You might also like