STATE SPACE
and
LINEAR SYSTEMS
..
BY
The importance of state space analysis is recognized in fields where the time behavior
of any physical process is of interest. The concept of state is comparatively recent, but the
methods used have been known to mathematicians for many years. As engineering, physics,
medicine, economics, and business become more cognizant of the insight that the state space
approach offers, its popularity increases.
This book was written not only for upper division and graduate students, but for prac
ticing professionals as well. It is an attempt to bridge the gap between theory and practical
use of the state space approach to the analysis and design of dynamical. systems. The book
is meant to encourage the use of state space as a tool for analysis and design, in proper
relation with other such tools. The state space approach is more general than the "classical"
Laplace and Fourier transform theory. Consequently, state space theory IS applicable to all
systems that can be analyzed by integral transforms in time, and is applicable to many
systems for which transform theory breaks down. Furthermore, state space theory gives
a somewhat different insight into the time behavior of linear systems, and is worth studying
for this aspect alone.
In particular, the state space approach is useful because: (1) linear systems with time
varying parameters can be analyzed in essentially the same manner as timeinvariant linear
systems, (2) problems formulated by state space methods can easily be programmed on a
computer, (3) highorder linear systems can be analyzed, (4) multiple inputmultiple output
systems can be treated almost as easily as single inputsingle output linear systems, and
(5) state space theory is the foundation for further studies in such areas as nonlinear
systems, stochastic systems, and optimal control. These are five of the most important
advantages obtained from the generalization and rigorousness that state space brings to
the classical transform theory.
Because state space theory describes the time behavior of physical systems in a mathe
matical manner, the reader is assumed to have some knowledge of differential equations and
of Laplace transform theory. Some classical control theory is needed for Chapter 8 only_
No knowledge of m:atrices or complex variables is prerequisite.
The book may appear to contain too many theorems to be comprehensible and/or useful
to the nonmathematician. But the theorems have been stated and proven in a manner
suited to show the range of application of the ideas and their logical interdependence.
Space that might otherwise have been devoted to solved problems has been used instead
to present the physical motivation of the proofs. Consequently I give my strongest recom
mendation that the reader seek to understand the physical ideas underlying the proofs rather
than to merely memorize the theorenls. Since the emphasis is on applications, the book
might not be rigorous enough for the pure mathematician, but I feel that enough informa
tion has been provided so that he can tidy up the statements and proofs himself._
The book has a number of novel features. Chapter 1 gives the fundamental ideas of
state from an informal, physical viewpoint, and also gives a correct statement of linearity.
Chapter 2 shows how to write transfer functions and ordinary differential equations in
matrix notation, thus motivating the material on matrices to follow. Chapter 3 develops
the important concepts of range space and null space in detail, for later application. Also
exterior products (Grassmann algebra) are developed, which give insight into determinants,
and which considerably shorten a number of later proofs. Chapter 4 shows how to actually
solve for the Jordan form, rather than just proving its existence. Also a detailed treatment
of pseudoinverses is given. Chapter 5 gives techniques for computation of transition
matrices for highorder timeinvariant systems, and contrasts this with a detailed develop
ment of transition matrices for timevarying systems. Chapter 6 starts with giving physical
insight into controllability and observability of simple systems, and progresses to the point
of giving algebraic criteria for timevarying systems. Chapter 7 shows how to reduce a
system to its essential parameters. Chapter 8 is perhaps the most novel. Techniques from
classical control theory are extended to timevarying, multiple inputmultiple output linear
systems using state space formulation. This gives practical methods for control system
design, as well as analysis. Furthermore, the pole placement and observer theory developed
can serve as an introduction to linear optimal control and to Kalman filtering. Chapter 9
considers asymptotic stability of linear systems, and the usual restriction of uniformity is
dispensed with. Chapter 10 gives motivation for the quadratic optimal control problem,
with special emphasis on the practical timeinvariant problem and its associated computa
tional techniques. Since Chapters 6, 8, and 9 precede, relations with controllability, pole
placement, and stability properties can be explored.
The book has come from a set of notes developed for engineering course 122B at UCLA,
originally dating from 1966. It was given to the publisher in June 1969. Unfortunately,
the pUblication delay has dated some of the material. Fortunately, it also enabled a number
of errors to be weeded out.
Now I would like to apologize because I have not included references, historical develop
ment, and credit to the originators of each idea. This was simply impossible to do because
of the outline nature of the book.
I would like to express my appreciation to those who helped me write this book. Chapter
1 was written with a great deal of help from A. V. Balakrishnan. L. M. Silverman helped
with Chapter 7 and P.K.C. Wang with Chapter 8. Interspersed throughout the book is
material from a course given by R. E. Kalman during the spring of 1962 at Caltech. J. J.
DiStefano, R. C. Erdmann, N. Levan, and K. Yao have used the notes as a text in UCLA
course 122B and have given me suggestions. I have had discussions with R. E. Mortensen,
M. M. Sholar, A. R. Stubberud, D. R. Vaughan, and many other colleagues. Improvements
in the final draft were made through the help of the control group under the direction of
J. Ackermann at the DFVLR in Oberpfaffenhofen, West Germany, especially by G. GrUbel
and R. Sharma. Also, I want to thank those UCLA students, too numerous to mention, that
have served as guinea pigs and have caught many errors of mine. Ruthie Alperin was very
efficient as usual while typing the text. David Beckwith, Henry Hayden, and Daniel Schau~
helped publish the book in its present form. Finally, I want to express my appreciation of
my wife Merideth and my children Erik and Kristin for their understanding during the
long hours of involvement with the book.
DONALD M. WIBERG
University of California, Los Angeles
June 1971
CONTENTS
Page
Chapter 1 MEANING OF STATE ••••• t ....................... ,. .................. . 1
Introduction to State. State of an Abstract Object. Trajectories in State
Space. Dynamical Systems. Linearity and Time Invariance. S~Tstems Con
sidered. Linearization of Nonlinear Systems.
Page
Chapter 7 CANONICAL FORMS OF THE STATE EQUATION ........... 147
Introduction to Canonical Forms. Jordan Form for TimeInvariant Systems.
Real Jordan Form. Controllable and Observable Forms for TimeVarying
Systems. Canonical Forms for TimeVarying Systems.
/
Chapter 1
Meaning of State
1.1 INTRODUCTION TO STATE
To introduce the subject, let's take an informal, physical approach to the idea of state.
(An exact mathematical approach is taken in more advanced texts.) First, we make a
distinction between physical and abstract objects. A physical object is an object perceived
by our senses whose time behavior we wish to describe, and its abstraction is the mathe
matical relationships that give some expression for its behavior. This distinction is made
because, in making an abstraction, it is possible to lose some of the relationships that make
the abstraction behave similar to the physical object. Also, not all mathematical relation
ships can be realized by a physical object.
The concept of state relates to those physical objects whose behavior can change with
time, and to which a stimulus can be applied and the response observed. To predict the
future behavior of the physical object under any input, a series of experiments could be
performed by applying stimuli, or inputs, and observing the responses, or outputs. From
these experiments we could obtain a listing of these inputs and their corresponding observed
outputs, i.e. a list of inputoutput pairs. An inputoutput pair is an ordered pair of real
time functions defined for all t == to, where to is the time the input is first applied. Of
course segments of these input time functions must be consistent and we must agree upon
what kind of functions to consider, but in this introduction we shall not go into these
mathematical details.
Definition 1.1: The state of a physical obiect is any property of the object which relates
input to output such that knowledge of the input time function for t == to
and state at time t = to completely determines a unique output for t == to.
Example 1.1.
Consider a black box, Fig. 11, contain
ing a switch to one of two voltage dividers.
Intuitively, the state of the box is the posi
tion of the switch, which agrees with Defi
nition 1.1. This can be ascertained by the
experiment of applying a voltage V to the
input terminal. Natural laws (Ohm's law)
dictate that if the switch is in the lower
position A, the output voltage is V/2, and
if the switch is in the upper position B, the
output voltage is V/4. Then the state A
determines the ihputoutput pair to be
(V, V /2), and the state B corresponds to
(V, V/4). Fig. 11
1
2 MEANING OF STATE [CHAP.!
Definition 1.2: An abstract object is the totality of inputoutput pairs that describe the
behavior of a physical object.
Instead of a specific list of input time functions and their correspondIng output time
functions, the abstract object is usually characterized as a class of all time functions that
obey a set of mathematical equations. This is in accord with the scientific method of
hypothesizing an equation and then checking to see that the physical object behaves in a
manner similar to that predicted by the equation. Hence we can often summarize the
abstract object by using the mathematical equations representing physical laws.
The mathematical relations which summarize an abstract object must be oriented,
in that m of the time functions that obey the relations must be designated inputs (denoted
by the vector u, having m elements 'Ui) and k of the time functions must be designated
outputs (denoted by the vector y, having k elements Yi). This need has nothing to do with
causality, in that the outputs are not "caused" by the inputs.
Definition 1.3: The state of an abstract object is a collection of numbers which together
with the input u(t) for all t ~ to uniquely determines the output y(t)
for all t ~ to.
In essence the state parametrizes the listing of inputoutput pairs. The state is the
answer to the question "Given u(t) for t ~ to and the mathematical relationsh~ps of the
abstract object, what additional information is needed to completely specify y(t) for t ~ to?"
Example 1.2.
A physical object is the resistorcapacitor network
shown in Fig. 12. An experiment is performed by applying
a voltage u(t), the input, and measuring a voltage yet), the R
ww
output. Note that another experiment could be to apply
yet) and measure u(t), so that these choices are determined
by the experiment.
The list of all inputoutput pairs for this example is
I
u(t) c I r
yet)
the class of all functions u(t). yet) which satisfy the mathe
matical relationship
RCdy/dt + y = u (1.1)
This relationship explicitly exhibits the list of inputoutput pairs. For any input time function u(r) for
T ~ to. the output time function yet) is uniquely determined by y(to), a number at time to· Note the
distinction between time functions and numbers. Thus the set of numbers y(to) parametrizes all input
output pairs, and therefore is the state of the abstract object described by (1.1). Correspondingly, a
choice of state of the RC network is the output voltage at time to·
Example 1.3.
The physical object shown in Fig. 13 is two RC networks in series. The pertinent equation is
R2C2 d2 y/dt 2 + 2.5RC dy/dt + y = u (1.3)
WN 'V'NI/'
r
u(t)
I R
Ic
2R
li C yet)
with a solution
y(t)
(1.4)
Here the set of numbers y(to) and ~~ (to) parametrizes the inputoutput pairs, and may be chosen as state.
Physically, the voltage and its derivative across the smaller capacitor at time to correspond to the state.
Definition 1.4: A state variable, denoted by the vector x(t), is the time function whose
value at any specified time is the state of the abstract object at that time.
Note this difference in going from a set of numbers to a time function. The state can
be a set consisting of an infinity of numbers (e.g. Problems 1.1 and 1.2), in which case the
state variable is an infinite collection of time functions. However, in most cases considered
in this book, the state is a set of n numbers and correspondingly x(t) is an nvector function
of time.
Definition 1.5: The state space, denoted by ~, is the set of all x(t).
Example 1.4.
The state variable in Example 1.2 is x(t) = y(t}, whereas in Example 1.1 the state variable remains
either A or B for all time.
Example 1.5.
y(t»)
The state variable in Example 1.3 is x(t) = dy .
(
dt (t)
The state representation is not unique. There can be many different ways of expressing
the relationship of input to output.
Example 1.6.
In Example 1.3, instead of the voltage and its derivative across the smaller capacitor, the state could
be the voltage and its derivative across the larger capacitor, or the state could be the voltages across
both capacitors.
There can exist inputs that do not influence the state, and, conversely, there can exist
outputs that are not influenced by the state. These cases are called uncontrollable and
unobservable, respectively, about which much more will be said in Chapter 6.
Example 1.7.
In Example 1.1, the physical object is state uncontrollable. No input can make the switch change
positions. However, the switch position is observable. If the wire to the output were broken, it would be
unobservable. A state that is both unobservable and uncontrollable makes no physical sense, since it can
not be detected by experiment. Examples l.2 and 1.3 are both controllable and observable.
One more point to note is that we consider here only deterministic abstract objects.
The problem of obtaining the state of an abstract object in which random processes are
inputs, etc., is beyond the scope of this book. Consequently, all statements in the whole
book are intended only for deterministic processes.
4 MEANING· OF STATE [CHAP. 1
Example L9.
In Example 1.3, note that equation (14) depends on t, u(r), x(to) and to, where x(t o) is the vector with
components y(to) and dy/dt(to). Therefore the trajectories + depend on these quantities.
Suppose now u(t) = 0 and Re = 1. Let Xl = yet) and Xz = dy/dt. Then dx/dt = X2 and d 2y/dt 2 =
dX2/dt. Therefore dt = dXl/x2 and so d2y/dt 2 = x2dxZ/dxl' Substituting these relationships into (1.3)
gives
~_+~~~~~+_Xl
Example 1.10.
The oriented mathematical equation yet) = ju(t) cannot give an abstract object, because either the
input or the output must be imaginary.
If a mathematical relationship always determines a real output y(t) existing for all
t ~ to given any real input u(t) for all time t, then we can form an abstract object. Note
that by supposing an input u(t) for all past times as well as future times, we can form an
abstract object from the equation for a delayor y(t) = u(t  T). [See Problem 1.1.J
CHAP. 1] MEANING OF STATE 5
However, we can also form an abstract object from the equation for a predictor
y(t) = u(t + T). If we are to restrict ourselves to mathematical relations that can be
mechanized, we must specifically rule out such relations whose present oufputs depend
on future values of the input.
Given that we have a dynamical system relating y(t) to u(t), we would like to construct
a set of mathematical relations defining a state x(t). We shall assume that a state space
description can be found for the dynamical system of interest ,satisfying the following
conditions (although such a construction may take considerable th~ught):
Condition 1: A real, unique output y(t) = fI(t, +(t; to, Xo, U{T)), u(t)) exists for all t> to
given the state Xo at time to and a real input U(T) for 7' ~ to.
Condition 2: A unique trajectory q,{t; to, Xo, u(.)) exists for all t > to given the state at
time to and a real input for all t ~ to.
Condition 1 gives the functional relationship y(t) = 7j(t, x(t), u(t)) between initial state
and future input such that a unique output is determined. Therefore, with a proper state
space description, it is not necessary to know inputs prior to to, but only the state at time to.
The state at the initial time completely summarizes all the past history of the input.
Example 1.11.
In Example 1.2., it does not matter how the voltage across the capacitor was obtained in the past.
All that is needed to determine the unique future output is the state and the future input.
ConditiDn 2 insures that the state at a future time is uniquely determined. Therefore
knowledge of the state at any time, not necessarily to, uniquely determines the output. For
a given u(t), one and only one trajectory passes through each point in state space and exists
for all finite t ~ to. As can be verified in Fig. 14, one consequence of this is that the state
trajectories do not cross one another. Also, notice that condition 2 does not require the
state to be real, even though the input and output must be real.
Example 1.12.
The relation dy/dt = u(t) is obviously a dynamical system. A state space description dx/dt = ju(t)
with output y(t) = jx(t) can be constructed satisfying conditions 15, yet the state is imaginary.
6 MEANING OF STATE [CHAP. 1
Condition 3 merely requires the state space description to be consistent, in that the
starting point of the trajectory should correspond to the initial state.. Condition 4 says
that the input uCr) takes the system from a state x(to) to a state x(t),' and if X(tl) is on
that trajectory, then the corresponding segment of the input will take the system from
X(tl) to x(t). Finally, condition 5 has been added to assure causality of the inputoutput
relationship resulting from the state space description to correspond with the causality of
the original dynamical system.
Example 1.13.
We can construct a state space description of equation (1.1) of Example 1.2 by defining a state
x(t) = yet). Then condition 1 is satisfied as seen by examination of the solution, equation (1.2). Clearly
the trajectory ¢(t; to, xo, U(T)) exists and is unique given a specified to. Xo and U(T), so condition 2 is satisfied.
Also, conditions 3 and 5 are satisfied. To check condition 4, given x(to) = y(t o) and U(T) over to:::: T:::: t,
then
x(t) =: (1.8)
where
x(t l ) = x(to)eCtotl)/RC + RlC f tl eCTotl)/RC U(TO) dTo
to
(1.9)
Substitution of (1.9) into (1.8) gives the previously obtained (1.2). Therefore the dynamical system (1.1) has
a state space description satisfying conditions 15.
Henceforth, instead of "dynamical system with a state space description" we will simply
say "system" and the rest will be understood.
are the corresponding outputs YI(t) and Y2(t) to the states Xl(t) and X2(t) with inputs Ul(T) and UZ(T).
Since any magnitude of voltage is permitted in this idealized system, any state X3(t) = aXI(t) + (3X2(t),
any input U3(t) = aUl(t) + /3uz(t), and any output Y3(t) = aYI(t) + j3Y2(t) will appear in the list of input
output pairs that form the abstract object. Therefore part (1) of Definition 1.7 is satisfied. Furthermore,
let's look at the response generated by X3(tO) and U3(T).
[ctXI(t )
O
+ /3x2(tO)]e<toO/RC + RIC it to
eCTtJ/RC [aUI(T) + ,BU2(r)] dT
aYl(t) + j3Yz(t) = Y3(t)
Since Y3(t) = xs(t), both the future output and state correspond to X3(t O) and uaCt) and the system is linear.
CHAP. 1] MEANING OF STATE 7
Example 1.15.
Consider the system of Example 1.1. For some a: and (3 there is no state equal to a:A + /3B, where A and
B are the switch positions. Consequently the system violates condition (1) of Definition 1.7 and is not linear.
Example 1.16.
Given the system dx/dt = 0, y = ucosx. Then Yl(t) = Ul(t) COSX1(t O) and Y2(t) = U2(t) COSX2(tO)'
The state X3(t) = x3(tO) = aX 1(t O) + /3x2(t O) and is linear, but the output
yet) = [a: Ul(t) + !3u2(t)] cos [a:x 1(tO) + /3X2(t O)] ¥= aYl(t) + /3Y2(t)
except in special cases like Xl(t O) = X2(t O) = 0, so the system is not linear.
If a system is linear, then superposition holds for nonzero u(t) with x(to) = 0 and also
for nonzero x(to) with u(t) = 0 but not both together. In Example 1.14, with zero initial
voltage on the capacitor, the response to a biased ac voltage input (constant + sin rot) could
be calculated as the response to a constant voltage input plus the response to an unbiased
ac voltage input. Also, note from. Example 1.16 that even if superposition does hold for
nonzero u(t) with x(to) = 0 and for nonzero x(to) with u(t) = 0, the system may still not
be linear.
Definition 1.8: A system is timeinvariant if the time axis can be translated and an equiva
lent system results.
One test for timeinvariance is to compare the original output with the shifted output.
First, shift the input time function by T seconds. Starting from the same initial state
Xo at tinle to + T, does y(t + T) of the shifted system equal y(t} of the original system?
Example 1.17.
Given the nonlinear differential equation
dx
dt
and the appearance of the last term on the right gives a different system. Therefore this is a time
varying nonlinear system. Equations with explicit functions of t as coefficients multiplying the state will
usually be timevarying.
Example 1.19.
A timevarying linear difference system of order n with one input and one output is described by
the equation
y(k + n) + a:l(k) y(k + n  1) + ... + lXn(k) y(k) = i3o(k) u(k + n) + ... + i3n(k) u(k) (1.11)
The values of a(k) depend on the step (k) of the process, in a way analogous to which the aCt) depend on
t in the previous example.
Example 1.20.
To reduce the second order nonlinear differential equation d2 y/dt 2  2y 3 + u dy/dt = 0 to two first
order nonlinear differential equations, define y = Yl and dy/dt = Y2' Then
dYl/dt Y2
Suppose a solution can be found (perhaps by computer) to equations (1.12) for some
initial conditions Y1(tO), Y2(t O), ... , Yn(t o) and some input w(t). Denote this solution as the
trajectory 1>{t; ~v(t), Yl(t O), ••• , Yn(t O), to). Suppose now that the initial conditions are changed:
dy d n  1y
y(to) = Yl(tO) + Xl(tO), dt (to) = yz(to) + X2{tO), ... , dt n 1 (to) = Yn(t O) + Xn(tO)
where X1(tO), X2(tO) and Xn{t o) are small. Furthermore, suppose the input is changed slightly
to u(t) = w(t) + v{t) where v(t) is small. To satisfy the differential equations,
d(cf>l + x 1)/dt = 11 (4)1 + Xl' cf>2 + X2 , ••• , cf>~ + X n , W + v, t)
CHAP. 1] MEANING OF STATE 9
If 11' [2' ... , In can be expanded about <P1' <P z' ••• , <Pn and w using Taylor's theorem for sev
eral variables, then neglecting higher order terms we obtain
d<Pl + dXl a/1 + all X2 ah + a/l v
dt dt 11 (<PI' <PZ' ••• , <P n' W, t) + Xl
aYI aY2 + + aYn Xn au
d<pz + dX2 a[z alz . + aYn
a/z + al2
dt dt = I Z (<p1'<P 2 , ···,<pn,w,t) + Xl
aYl + Xz
ayz +
"
Xn au v
•••• ,. ................ l1li •••••••••• " ,. " • • • • .. • • •• " .. II II .......................... II. •••••
where each a//aYi is the partial derivative of li(Y l ' Yz' ... , Yn' u, t) with respect to Yp evalu
ated at Y1 = <PI' Yz = <P2' ••• , Y n = <Pn and U = w. Now, since each <Pi satisfies the original
equation, then d<p/dt = Ii can be canceled to leave
which is, in general, a timevarying linear differential equation, so that the nonlinear equa
tion has been linearized. Note this procedure is valid only for Xl, X2, ••• , Xn and v small
enough so that the higher order terms in the Taylor's series can be neglected. The matrix
of al/aYj evaluated at Yj = <Pi is called the Jacobian matrix of the vector f(y, u, t).
Example 1.21.
Consider the system of Example 1.20 with initial conditions y(to):= 1 and y(to) := 1 at to:= 1.
If the particular input w(t) := 0, we obtain the trajectories ¢l(t) :::: t 1 and ¢2(t) :::: t z. Since 11 == Yz,
then all/aYl:= 0, al1layz:::: 1 and aldau:::: O. Since f 2 := 2yi UYz, then alzlaYl:= 6yi, afz/BY2 :::: u and
al~au :::: Y2' Hence for initial conditions y(to):= 1 + XI(tO), y(to):= 1 + X2(t O) and inputs u:::: v(t),
we obtain
This linear equation gives the solution yet) == Xl(t) + t 1 and dy/dt:= X2  t 2 for the original nonlinear
equation, and is valid as long as Xl' X2 and 'V are small.
Example 1.22.
Given the system dy/dt == ky  y2 + u. Taking u(t) := 0, we can find two constant solutions ¢(t):::: 0
and 1f;(t) := k. The equation for small motions x(t) about ¢(t) := 0 is dx/dt == It:x + u so that y(t) ~ x(t),
and the equation for small motions x(t) about 1f;(t) :::: k is dx/dt:= kx + u so that y(t) = k + x(t).
10 MEANING OF STATE [CHAP. 1
Solved Problems
1.1. Given a delay line whose output is a voltage input delayed T seconds. What is the
physical object, the abstract object, the state variable and the state space? Also,
is it controllable, observable and a dynamical system with a state space description?
The physical object is the delay line for an input u(t) and an output y(t) = u(t ~ T). This
equation is the abstract object. Given an input time function u(t) for all t, the output yet) is
defined for t ~ to, so it is a dynamical system. To completely specify the output given only u( t)
for t. ~ to. the voltages already inside the delay line must be known. Therefore, the state at
timetIJ is x(to) = u[t o T , to) ~ where the notation u[t2 , tl) means the time function u(r) for 7 in the
interval t2 === 7 < t 1 • For E> 0 as small as we please,· U[toT, to) can be considered as the un
countably infinite set of numbers
{u(to  T), 'u(to  T + e), •.. , u(t o  e)}
In this sense we can consider the state as consisting of an infinity of numbers. Then the state
variable is the infinite set of time functions
x(t) = U[tT, t) == {u(t  T), u(t  T + e), ... , u(t  e)}
The state space is the space of all time functions T seconds long, perhaps limited by the
breakdown voltage of the delay line.
An input u(t) for to === t < to + T will completely determine the state T seconds later, and~
any state will be observed in the output after T seconds, so the system is both observable and con
trollable. Finally, x(t) is uniquely made up of x(t  7) shifted r seconds plus the input over T sec
onds, so that the mathematical relation yet) == u(t  T) gives a system with a state space
deseri ption.
i
n=l
where en == I y(r, to) sin n"r dr. All that is needed to determine the output is y(r, to), so that
o .
y(r, t) is a choice for the state at any time t. Since y(r, t) must be known for almost all r III the
interval 0,::::: r == 1, the state can be considered as an infinity of numbers similar to the case of
Problem 1.1.
1.3. Given the mathematical equation (dy/dt)2 = y2 + Qu. Is this a dynamical system
with a state space description?
A real output exists for all t === to for any u(t), so it is a dynamical system. The equation
can be written as dy/dt == s(t)y, where set) is a member of a set of time functions that take on
the value +1 or 1 at any fixed time t. Hence knowledge of y(t o) and set) for t ~ to uniquely
specify yet) and they are the state.
The former equation can be integrated immediately to xz(t) = X2(tO), a straight line in the phase
plane. The latter equation is solved by multiplying by dx/2 to obtain x2 dx2i2 + Xl dXl/2 ::= O.
' can b
ThIS e 'm t egrat e d to 0 btam
. x 2 (t) + Xl2 (t) = Xz2 (to) + Xl2 (to). The result is an equation of circles
2
in the phase plane. The straight lines lie to the left of the line X2  Xl ::= 1 and the circles to
the right, as plotted in Fig. 15.
2.~~+__+~,
3.~ __~____   + __ _~
~4~====~====~~====~~~~_t~Xl
6.~_. ____~
Note Xl increases for x2 > 0 (positive velocity) and decreases for x2 < 0, gIvmg the motion
of the system in the direction of the arrows as t increases. For instance, starting at the initial
conditions xl(t O) and xz(to) corresponding to the point numbered 1, the system moves along the
outer trajectory to the point 6. Sim'ilarly point 2 moves to point 5. However, starting at either
point 3 or point 4, the system goes to point 7 where the system motion in the next instant is not
determined. At point 7 the output y(t) does not exist for future times, so that this is not a
dynamical system.
1.5. Given the electronic device diagrammed in Fig. 16, with a voltage input u(t) and a
voltage output y(t). The resistors R have constant values. For to ~ t < t l , the
switch S is open; and for t === t l , S is closed. Is this system linear?
s
u(t) R y(t)
Fig. 16
Referring to Definition 1.7, it becomes apparent the first thing to do is find the state. No other
information is necessary to determine the output given the input, so there is no state, i.e. the
dimension of the state space is zero. This problem is somewhat analogous to Example 1.1, except
that the position of the switch is specified at each instant of time.
To see if the system is linear, sincex(t) = 0 for all time, we only need assume two inputs Ul(t)
and U2(t). Then Yl(t) = uI(t)/2 for to ~ t < tl and Yl(t) = uI(t)/3 for t:=, t l • Similarly Y2(t);:::
u2(t)/2 or uz(t)/3. Now assume an input o.:ul(t) + {3U2(t). The output is [o.:ul(t) + {3u2(t)]J2 for
to === t < tl and Ia::Ul(t) + {3u2(t)]/3 for t> tl' Substituting Yl(t) and Y2(t), the output is aYl(t) + {1Y2(t),
shOWing that superposition does hold and that the system is linear. The switch S can be consid
ered a timevarying resistor whose resistance is infinite for t < tl and zero for t::=: t l . Therefore
Fig. 16 depicts a timevarying linear device.
12 MEANING OF STATE [CHAP.!
1.6. Given the electronic device of Problem 1.5 (Fig. 16), with a voltage input u(t) and
a voltage output y(t). The resistors R have constant values. However, now the posi
tion of the switch S depends on y(t). Whenever y(t) is positive, the switch S is open;
and whenever y(t) is negative, the switch S is closed. Is this system linear?
Again there is no state, and only superposition for zero state but nonzero input need be
investigated. The inputoutput relationship is now
yet) = [5u(t) + u(t) sgn u(t)]l12
where sgn u = +1 if u is positive and 1 if u is negative. Given two inputs Ul and U2 with
resultant outputs Yl and Yz respectively, an output y with an input u3 = aul + {3~ is expressible as
y = [5( aul + (3U2) + (aul + fJ u 2) sgn (aul + !3u z)]l12
To be linear, ay + j3Y2 must be equal y which would be true only if
aUl sgn ul + j3u2 sgn U2 = (aUl + (3u2) sgn (aul + !3U2)
This equality holds only in special cases~ such as sgn Ul = sgn u2 = sgn (au1 + j3u2), so that the
system is not linear.
Is this timevarying?
~
This abstract object is that of Example 1.2, with
RC = 1 in equation (1.1). By the same procedure
used in Example 1.16, it can also be shown time
invariant. However, it can also be shown timeinvari
ant by the test given after Definition 1.8. The input t
time function u{t) is shifted by T, to become (t). u to t
~
Then as can be seen in Fig. 17,
~ (t) = u(t  T)
I~(t)
17
~
to+T
Let g = T  T:
to
e~+TuuW dg
to +T t+ T
t
~
Evaluating Yat CT = t +T gives Iy(t)
y (t + T) = xoe to  t
+ ft eetuW dg t
to to+T t+T
which is identical with the output y(t). Shifted System
Fig. 17
CHAP. I] MEANING OF STATE 13
Supplementary Problems
1.8. Given the springruass system shown in Fig. 18. What is the
physical object, the abstract object, and the state variable? "
1.9. Given the hereditary system yet) = it"" K(t, T) U(r) dT where
K(t, T) is some singlevalued continuously differentiable function
of t and T. What is the state variable? Is the system linear?
Is the system timevarying? Fig. 18
1.10. Given the discrete time system x(n + 1) = x(n) + u(n), the series of inputs u(O), u(I), ... , u(k),
and the state at step 3, xeS). Find the state variable x(m) at any step m =: O.
1.11. An abstract object is characterized by yet) =u(t) for to ~ t < t l , and by dy/dt = du/dt for t =: t l .
lt is given that this abstract'object will permit discontinuities in yet) at t l . What is the dimension
of the state space for to === t < tl and for t =: tl?
1.12. Verify the solution (1.2) of equation (1.1), and then verify the solution (1.3) of equation (1.4).
Finally, verify the solution Xl + 2xz = C(2xl + xZ)4 of Example 1.9.
1.13. Draw the trajectories in two dimensional state space of the system y+ y = O.
1.14. Given the circuit diagram of Fig. 19(a), where the nonlinear device NL has the voltagecurrent
relation shown in Fig. 19(b). A mathematical equation is formed using i = Cv and v = let) = f(Ov):
v= (1IC)fl(v)
where 1 1 is the inverse function. Also, veto) is taken to be the initial voltage on the capacitors.
Is this mathematical relation a dynamical system with a state space description?
C NL
~ vet)
Ca)
~ (b)
i
Fig. 19
1.20. Show equation (1.10) is timeinvariant if the coefficients £ri and Pj for i = 1,2, "'J nand
j = 0, 1, ... , n, are not functions of time.
1.22. What is the state space description for the anticipatory system yet) = u(t + T) in which only con
tion 5 for dynamical systems is violated?
1.24. What is the state space description for the diiferentiator y = du/dt?
1.25. Is the equation y = j(t)n a dynamical system given values of jet) for to === t === tl only?
k1 3k
1.10. x(k) = x(3) + i=3
~ u(i) for k = 4,5, . .. and x(k) = x(3)  ~ u(3  i)
i=l
for k = 1,2,3. Note we
need not "tie" ourselves to an "initial" condition because anyone of the values xCi) will be the
state for i = 0, 1, ... n.
1.11. The dimension of the state space is zero for to::::: t < tb and onedimensional for t:=: t 1• Because
the state space is timedependent in general, it must be a family of sets for each time t. Usually
it is possible to consider a single set of inputoutput pairs over all t, i.e. the state space is time
illVariant. Abstract obj ects possessing this property are called uniform abstract objects. This
problem illustrates a nonuniform abstract object.
1.12. Plugging the solutions into the equations will verify them.
1.14. It is a dynamical system, because vet) is real and defined for all t ~ to. However, care must be
taken in giving a state space description, because jl(V) is not single valued. The state space
description must include a means of determining· which· of the lines 12, 23 or 34 a particular
voltage corresponds to.
1.22. No additional knowledge is needed other than the input, so that state is zero dimensional and the
state space description is yet) = u(t + T). It is not a dynamical system because it is not realizable
physically if u(t) is unknown in advance for all t::=" to. However, its state space description
violates only condition 5, so that other equations besides dynamical systems can have a state
space description if the requirement of causality is waived.
1.23. Yes. yet) = etxo + ft e(Tt) u(r) dT, and the contribution of the initial condition Xo depends
to
on when the system is started. If Xo = 0, the system is equivalent to dx/dt = x + u where
y = x, which is timeinvariant.
1.24. If we define
du/dt = lim [u(t+£)u(t)]I€
£10
so that y(to) is defined, then the state space is zero dimensional and knowledge of u(t) determines
yet) for all t ::=" to. Other definitions of du/dt may require knowledge of u(t) for to  € ~ t,.o:: to,
which would be the state in that case.
1.25. Obviously yet) is not defined for t > tl> so that as stated the equation is not a dynamical system.
However, if the behavior of engineering interest lies between to and t l , merely append y = Ou for
t ::=" tl to the equation and a dynamical system results.
Chapter 2
)        _ y(t)
Definition 2.2: A scalar is a diagrammatical abstract object having one input u(t) and one
output y(t) such that the input is scaled up or down by the time function a(t)
as indicated in Fig. 22. The output obeys the relationship y(t) = a(t) u(t).
Definition 2.3: An integrator is a diagrammatical abstract object having one input u(t),
one output y(t), and perhaps an initial condition y(to) which may be shown
or not, as in Fig. 23. The output obeys the relationship
16
CHAP. 2] METHODS FOR OBTAINING THE STATE EQUATIONS 17
Definition 2.4: A delayor is a diagrammatical abstract object having one input u(k), one
output y(k), and perhaps an initial condition y(l) which may be shown or
not, as in Fig. 24. The output obeys the relationship
y(i + l + 1) = u(i + l) for j = 0,1,2, ...
with initial condition y(to). An interconnection for this system is shown in Fig. 25.
Fig. 25
Since 0:. is a constant function of time, the integrator and scalo"r can be interchanged if the initial condi
tion is adjusted accordingly, as shown in Fig. 26.
Fig. 26
This interchange could not be done if 0:. were a general function of time. In certain special cases it
is possible to use integration by parts to accomplish this interchange. If o:.(t) = t, then (2.1) can be inte
grated to
y(t} = y(t o) + ft 'T[Y('T) + u('T)] dr
to
18 METHODS FOR OBTAINING THE STATE EQUATIONS [CHAP. 2
Fig. 2·7
Integrators are used in continuous time systems, delayors in discrete time (sampled
data) systems. Discrete time diagrams can be drawn by considering the analogous con
tinuous time system, and vice versa. For timeinvariant systems, the diagrams are almost
identical, but the situation is not so easy for timevarying systems.
Example 2.2.
Given the discrete time system
y(k + l + 1) :::: ay(k + 1) + au(k + l) (2.2)
with initial condition y(l). The analogous continuous time systems is equation (2.1), where d/dt takes the
place of a unit advance in time. This is more evident by taking the Laplace transform of (2.1),
aY(S) + aU(S)
and the £ transform of (2.2),
zY(z)  zy(l) :::: aY(z) + aU(z)
Hence from Fig. 25 the diagram for (2.2) can be drawn immediately as in Fig. 28.
Fig. 28
If the initial condition of the integrator or delayor is arbitrary, the output of that
integrator or delayor can be taken to be a· state variable.
Example 2.3.
The state variable for (2.1) is yet), the output of the integrator. To verify this, the solution to equa
tion (2.1) is
yet) = y(to) e(Y.Ctto) + 0'. f t e(Y.CtT) U(T) dT
to
Note y(t o) is the state at to, so the state variable is yet).
Example 2.4.
The state variable for equation (2.2) is y(k + l), the output of the delayor. This can be verified in a
manner similar to the previous example.
CHAP.2J METHODS FOR OBTAINING THE STATE EQUATIONS 19
Example 2.5.
From Fig. 27, the state is the output of the second integrator only, because the initial condition of the
first integrator is specified to be zero. This is true because Fig. 27 and Fig. 25 are equivalent systems.
Example 2.6.
A summer or a scalor has no state associated with it, because the output is completely determined by
the input.
from which the flow diagram shown in Fig. 29 can be drawn starting with the output y
at the right and working to the left.
u(t)  +.....,...
.......
I~y(t)
Xn  1 anIY + Xn + f3 n  Iu
X10 anY + f3nU
20 METHODS FOR OBTAINING THE STATE EQUATIONS [CHAP. 2
Using the first equation in (2.5) to eliminate y, the differential equations for the state vari
ables can be written in the canonical matrix form
Xl a
1 1 0 0 Xl /3 1  a 1/3 o
X2 a
2 0 1 0 X2 /3 2  a 2/3 0
d
dt
.......... ,. ........... " III
+ u (2.6)
 an  1 0 0 1 X n 1 f3 n  1  an 1 f3 o
Xn an 0 0 0 Xn /3 n  a rJ30
We will call this the first canonical form. Note the Is above the diagonal and the a'S down
the first column of the n x n matrix. Also, the output can be written in terms of the state
vector
y = (I 0 '" 0 0) (2.7)
Note this form. can be written down directly from the original equation (2.3).
Another useful form can be obtained by turning the first canonical flow diagram "back
wards." This change is accomplished by reversing all arrows and integrators, interchang
ing summers and connection points, and interchanging input and output. This is a heuristic
method of deriving a specific form that will be developed further in Chapter 7.
y(t)
u(t)
Here the output of each integrator has been relabeled. The equations for the state
variables are now
Xl 0 1 0 0 Xl 0
X2 0 0 1 0 X2 0
d
dt = • • • • .. • • • .. • • .. ~,. .. • • • • • • • .. .. ill • • •
+ u (2.9)
0 0 0 1 0
Xn an  1 an  2
a 1 Xn 1
and y (2.10)
This will be called the second canonical form, or phasevariable canonical form. Here the
Is are above the diagonal but the a'S go across the bottom row of the n x n matrix. By
eliminating the state variables x, the general inputoutput relation (2.8) can be verifie~d.
The phasevariable canonical form can also be written down upon inspection of the original
differential equation (2.3).
y (2.11)
y (2.12)
Consider first the case where the denominator polynomial factors into distinct poles Ai,
i = 1,2, ... , n. Distinct means Ai 01= Aj for i 01= j, that is, no repeated roots. Because most
practical systems are stable, the Ai usually have negative real parts.
pn + a 1pn1 + ... + O::n1P + an = (p  A 1 )(P  A2) ' •• (p  An) (2.13)
The partial fraction expansion (2.14) gives a very simple flow diagram, shown in Fig.
211 follo,ving.
22 METHODS FOR OBTAINING THE STATE EQUATIONS [CHAP. 2
Note that because p. and A. can be complex numbers, the statesx.t are complex;'valued
~ ~
(2.16)
Xn = AnXn+U
Consider now the general case. For simplicity, only one multiple root (actually one
Jordan block, see Section 4.4, page 73) will be considered, because the results are easily
extended to the general case. Then the denominator in (2.12) factors to
(2.17)
instead of (2.13). Here there are v identical roots. Performing the partial fraction expan
sion for this case gives
PIU P2U Pvu PiJ+I U Pnu
y = f3 oU + (PAl)V + (PAl)Vl + ... + PAI
+ PAv+I
+ ... + PAn'
(218)
where f(p) is the polynomial fraction in P from (2.12). This gives the flow diagram shown
in Fig. 212 following.
CHAP. 2] METHODS FOR OBTAINING THE STATE EQUATIONS 23
u(t}"""'"~..,...f y(t)
XV1 A1X v  l + Xv
Xv = ).IX v+ U (2.20)
XV+1 Av+lXv +l +U
Xn AnXn+U
y = (3ou + PIX! + P2X2 + ... + PnX n
24 METHODS FOR OBTAINING THE 'STATE EQUATIONS (CHAp. 2
The matrix differential equations associated with this Jordan· form are
Xl Al 1 0 0 0 0 0 Xl 0
X2 0 Al 1 0 0 0 0 X2 0
..... W'.,. .... II .. lit ........... W' ..... 0#'"" ..... " ..
d X v 1 0 0 0 Al 1 0 0 XI/1 0
dt Xv
= 0 0 0 0 Al 0 0 Xv + 1
u (2.21)
Xv+l 0 0 0 0 o AII+1 0 XII + 1 1
ill ••••• ill ............ " ill .....................
Xn 0 0 0 0 0 0 An Xn 1
In the n X n matrix, there is a diagonal row of ones above each Al on the diagonal, and then
the other A's follow on the diagonal.
Example 2.7.
Derive the Jordan form of the differential system
jj + 2y + y == u+ u (2.23)
Equation (2.23) can be written as y :::: (~: 1~)2 u whose partial fraction expansion gives
o 1
y (p + 1)2 U + P + 1u
Figure 213 is then the flow diagram.
yet)
u(t)+40/
+
Fig. 213
Because the scalor following Xl is zero, this state is unobservable. The matrix state equations in
Jordan form are
METHODS FOR OBTAINING THE STATE EQUATIONS 25
= (1.10)
Example 2.8.
Consider the second order equation
d 2y dy d 2u du
dt 2 + Il'l(t) dt + 1l'2(t)y = f3o(t) dt2 + f31(t) dt + f32(t)U (2.26)
Then by (2.24)
Y == Xl + yo(t)u (2.27)
and differentiating,
(2.28)
Differentiating again,
(2.90)
+ u (2.38)
0 0 0 1 Ynl(t)
xn an(t) an_ 1 (t) an_2 (t) a 1(t) xn Yn(t)
y (1 0 ... 0)
Specifically, if the system has only one input u and one output y, the differential equa
tions for the system are
dx/dt = A(t)x + b(t)u
Y = ct(t)x + d(t)u
and similarly for discrete time systems. Here c(t) is taken to be a column vector, and
ct(t) denotes the complex conjugate transpose of the column vector. Hence ct(t) is a row
vector, and ct(t)x is a scalar. Also, since u, y and d(t) are not boldface, they are scalars.
Since these state equations are matrix equations, to analyze their properties a knowledge
of matrix analysis is needed before progressing further.
Solved Problems
2.1. Find the matrix state equations in the first canonical form for the linear time
invariant differential equation
y + 5y + 6y = it + u (2.41)
with initial conditions y(O) = Yo, y(O) = yo. Also find the initial conditions on the
state variables.
Using p = dldt, equation (2 .."1) can be written as p2y + 5py + 6y = pu + u, Dividing by p2
and rearranging,
1 1
Y = (u5y) + ~(u6y)
p p2_
The flow diagram of Fig. 2~14 can be drawn starting from the output at the right.
u~~~~~
.>t.... Y
Fig. 2..14
Next, the outputs of the integrators are labeled the state variables Xl and x2 as shown. Now
an equation can be formed using the summer on the left:
X2 = 6y +u
Similarly, an equation can be formed using the summer on the right:
Xl = X2  5y +u
Also, the output equation is y = Xl' Substitution of this back into the previous equations gives
Xl 5X1 + X2 + u
(2.42)
28 METHODS FOR OBTAINING THE STATE EQUATIONS [CHAP. 2
The initial conditions on the state variables must be related to Yo and Yo, the given output initial
conditions. The output equation is xl(t) = yet), so that Xl(O) = yeO) = Yo. Also; substituting
yet) = Xl (t) into (2.42) and setting t == 0 gives
These relationships for the initial conditions can also be obtained by referring to the flow diagram
at time t = O.
2.2. Find the matrix state equations in the second canonical form for the equation (2.41)
of Problem 2.1, and the initial conditions on the state variables.
The flow diagram (Fig. 214) of the previous problem is turned "backwards" to get the flow
diagram of Fig. 215.
y......I
.....+u
Fig. 215
The outputs of the integrators are labeled Xl and x2 as shown. These state variables are dif
ferent from those in Problem 2.1, but are also denoted Xl and Xz to keep the state vector x(t) nota
tion, as is conventional. Then looking at the summers gives the equations
y = Xl + X2 (2.43)
and y (1 1)(:~)
The initial conditions are found using (2.43),
2.3. Find the matrix state equations in Jordan canonical form for equation (2.41) of Prob
lem 2.1, and the initial conditions on the state variables.
The transfer function is
y
+
+
u.. J~y
Fig. 216
The state equations can then be written from the equalities at each summer:
Yo = 2xz(O)  Xl (0)
Yo = 2X2(0)  0: 1(0)
The state equation is used to eliminate Xl(O) and X2(0):
y (1 1)(:~)
Find the differential equation relating the input to the output.
In operator notation, the state equations are
PX l X2
PX2  6X l  5x2 +U
2.5. Given the feedback system of Fig. 217 find a state space representation of this
closed loop system.
K
R(s)+.t G(s) =   14~_ C(s)
8+1
1
H(s) = 8 +3
Fig. 217
The transfer function diagram is almost in flow diagram form already. U sing the Jordan
canonical form for the plant G(s) and the feedback H(s) separately gives the flow diagram of
Fig. 218.
r 1
I I
r ( t )  ........ 1  +   9    _ c(t)
+ I
I
I
IL __________________ ! ~
Fig. 218
Note G(s) in Jordan form is enclosed by the dashed lines. Similarly the part for H(s) was drawn,
and then the transfer function diagram is used to connect the parts. From the flow diagram,
d
dt
(Xl)X2 =
(1 1) (Xl) + (1)
K 3 X2 0 ret), c(t) = (K O)(~~)
CHAP. 2] METHODS FOR OBTAINING THE STATE EQUATIONS 31
, 2.6. Given the linear, timeinvariant, multiple inputmultiple output, discretetime system
y/n + 2) + (l:lyl(n + 1) + a2Yl(n) + Y2YZ(n + 1) + YSY2(n) == (31u1(n) + 8Iu 2(n)
y 2(n + 1) + Y Y2(n) + (Xsyl(n + 1) + YI(n) ==
1 (X4 (32u/n) + 82u2(n)
Put this in the form
x(n+ 1) Ax(n) + Bu(n)
y(n) Cx(n) + Du(n)
Starting from the right, the flow diagram can be drawn as shown in Fig. 219.
Fig. 219
Any more than three delayors with arbitrary initial conditions are not needed because a
fourth such delayor would result in an unobservable or uncontrollable state. From this diagram
the state equations are found to be
32 METHODS FOR OBTAINING THE STATE EQUATIONS [CHAP. 2
2.7. Write the matrix state equations for a general timevarying second order discrete time
equation, i.e. find matrices A(n), B(n), C(n), D(n) such that
x(n + 1) = A(n) x(n) + B(n) u(n)
y(n) = C(n) x(n) + D(n) u(n) (2.48)
given the discrete time equation
y(n + 2) + a 1(n) y(n + 1) + a 2 (n) y(n) f3 o(n) u(n + 2) + f3 1(n) u(n + 1) + {32(n) u(n) (2.49)
Analogously with the continuous time equations. try
xl(n + 1) = x2(n) + "h(n) u(n) (2.50)
Substituting (2.52). (2.53). (2.54) into (2.49) and equating coefficients of u(n), u(n + 1) and u(n + 2)
gives
Yo(n) J3o(n  2)
2.8. Given the timevarying second order continuous time equation with zero input,
y + O!I(t)y + a 2(t)y = 0 (2.55)
Substituting (2.57), (2.58) and (2.59) into (2.55) and equating coefficients of Xl and Xz gives the
equations
Yl  alYl + (0:2  tl1hl = 0
Y2 + (a1  2C1!z)yz + (CI!~  a2 a 1 + Cl!2 ' a2)YZ + 2·h  a211 = 0
In this case, Yl(t) may be taken to be zero, and any nontrivial yz(t) satisfying
12 + (a1  2a2)YZ + (a~  Cl!Z a l + IX2  0:2h2 = 0 (2.60)
will give the desired canonical form.
This problem illustrates the utility of the given time~varying form (2.38). It may always be
found by differentiating known functions of time. Other forms usually involve the. solution of
equations such as (2.60), which may be quite difficult, or require differentiation of the ai(t). Addition
of an input contributes even more difficulty. However, in a later chapter forms analogous to the
first canonical form will be given.
Supplementary Problems
2.9. Given the discrete time equation yen + 2) + 3y(n + 1) + 2y(n) = u(n + 1) + 3u(n), find the matrix
state equations in (i) the first canonical form, (ii) the second canonical form, (iii) the Jordan canon
ical form.
2.10. Find a matrix state equation for the multiple inputmultiple output system
YI + atih + a2YI + Y3112 + Y4YZ ::::: (h u 1 + SlUZ
yz + 11112 + 1zY2 "+ a3111 + Cl!4Y1 ::::: f32 u 1 + S2 U Z
2.11. Write the matrix state equations for the system of Fig. 220, using Fig. 220 directly.
u
Xz + Xl
~~y
p + p
2.13. The procedure of turning "backwards" the flow diagram of the first canonical form to obtain the
second canonical form was never· justified. Do this by verifying that equations (2.8) satisfy the
original inputoutput relationship (2.3). Why can't the timevarying flow diagram corresponding
to equations (2.24) be turned backwards to get another canonical form?
Motor Armature: Ri + L ~; = V  Kf :
Inertial Load:
Obtain a matrix state equation relating the input voltage V to the output shaft angle 8 using a
state vector
2.17. The equations describing the time behavior of the neutrons in a nuclear reactor are
6
Prompt Neutrons: In (p(t)  {l)n :I ""iCi
+ i=l
Delayed Neutrons:
6
where f3 = :I f3i and p(t) is the timevarying reactivity, perhaps induced by control rod motion.
i=1
Write the matrix state equations.
2.18. Assume the simplified equations of motion for a missile are given by
Lateral Translation: z + K ¢ + K a + Kaf3
1 2 = 0
Rotation: if; + K 4a + K5f3 = 0
2.19. 'Consider the timevarying electrical network of Fig. 222. The il
voltages across the inductors and the current in the capacitors
can be expressed by the relations
ea  el :t (L1i l ) =
di l
L ldt +
. dL I
'/,1&
d del dGI
il  i2 dt (Glel) Glert + cI&
o o
d~ . dL2
el  eb :t (L 2i 2 ) = L2dt + t2dt
It is more convenient to take as states the inductor fluxes
PI = itto
(e a  el) dt + PI(tO)
d:x
dt
A(t)x + B(t)u
where the state vector x, input vector u, and output vector yare
(i)
x=GJ u (::) y
=G)
(ii)
x=G) u
(::) y =G)
2.20. Given the quadratic Hamiltonian H = iqTVq + ipTTp where q is a vector of n generalized co
ordinates, p is a vector of corresponding conjugate momenta, V and Tare n X n matrices corre
sponding to the kinetic and potential energy, and the superscript T on a vector denotes transpose.
Write a set of matrix state equations to describe the state.
(iii) x(n + 1) = (~ _~) x(n) + (~) u(n); yen) (1 2) x(n)
36 METHODS FOR OBTAINING THE STATE EQUATIONS [CHAP. 2
~) (~' ~l)U
1 Y3
. 0:2 0 Y4
C"
x +
a3 0 [1 1 x
a4 0 Y2 0 f32
°2
(~ ~) x
0 0
y
0 1
.
X
(; 0:1
0
0
0:4
{12
°2
(~ ~)x
0 0
Y
0 1
2.11.
:t G;) "'2
0 "'3 X3 "'3
U
Y (1 oo)G:)
2.12. For Xl = v) X2 = V, ::li3 = V, X4 = ·v: x5 =w, X6=W the initial conditions result immediately and
01 0 0 0 0 0
00 1 0 0 0 0
0 0 0 1 0 0 0
x
0 500 105 6 0 0
x + 500
u
0 0 0 0 0 1 0
100 1 1 0 100 2 0
2.13. The timevarying flow diagram cannot be turned backwards to obtain another canonical form be
cause the order of multiplication by a timevarying coefficient and an integration cannot be inter
changed.
2.14. (i)
u~~~__,
I    . . y
Fig. 223
CHAP. 2] METHODS FOR OBTAINING THE STATE EQUATIONS 37
(ii) The values of Yo, Y1 and Y2 are given by equations (2.33), (2.34) and (2.35).
(iii) X1(0) = Yo  Yo(O) u(O) x2(0) = Yo  (fo(O) + Y1(0» u(O)  Yo(O) U(O)
2.16.
d
dt ( i)
fJ
dfJ/dt
:::: fJ (0 1 0) ( ;
de/dt
)
2.17. !£
dt (~1).:
Os
o
2.18. (i)
(~ ~)
1
c =
(ii)
Ka 1
K2 K S
(iii)
o
K4
(~
1 0
c == 'Kex 0
2.19. (i)
(ii) A 0 o1 )
.. 1
LL 1
2 2
c D
::::
2.20. The equations of motion are
Using the state vector x :::: (q1 qn PI ... Pn)T, the matrix state equations are
X :::: ( 0
V
T)o x
Chapter 3
Definition 3.2: A row of a matrix is the set of all elements through which one horizontal
line can be drawn.
Definition 3.3: A column of a matrix is the set of all elements through which one vertical
line can be drawn.
Example 3.2.
The rows of the matrix of Example 3.1 are (0 2 j) and (1i x 2 sin t). The columns of this matrix are
Definition 3.6: A scalar, denoted by a letter that is not in boldface type, is a 1 x 1 matrix.
In other words, it is one element. When part of a matrix A, the notation
atj means the particular element in the ith row and jth column.
Definition 3.7: A vector, denoted by a lower case boldfaced letter, such as a, or with its
contents displayed in braces, such as {at}, is a matrix with only one row or
only one column. Usually a denotes a column vector, and aT a row vector.
38
CHAP. 3] ELEMENTARY MATRIX THEORY 39
Definition 3.8: The diagonal of a square matrix is the set of all elements aij of the matrix
in which i = j. In other words, it is the set of all elements of a square
matrix through which can pass a diagonal line drawn from the upper left
hand corner to the lower right hand corner.
Example 3.3.
Given the matrix
The diagonal is the set of elements through which the solid line is drawn, b 11 , b22 , b33 , and not those sets
determined by the dashed lines.
Definition 3.9: The trace of a square matrix A, denoted tr A, is the sum of all elements
on the diagonal of A.
n
trA ~ au
i=l
Definition 3:13: To perform matTix multiplication, the element Cij of the product matrix C
is found by taking the dot product of the ith row of the left matrix
A and the jth column of the right matrix B, where C = AB, so that
n
Cij = ~ aikb kj •
k=l
Note that this definition requires the left matrix (A) to have the same number of columns
as the right matrix (B) has rows. In this case the matrices A and B are said to be compatible.
It is undefined in other cases, excepting when one of the matrices is 1 x 1, i.e. a scalar.
In this case each of the elements is multiplied by the scalar, e.g. aB = {ab ii } for all i and j.
Example 3.4.
The vector equation y = Ax, when y and x are 2 X 1 matrices, i.e. column vectors, is
40 ELEMENTARY MATRIX THEORY [CHAP. 3
2
where Yi = ~ aikxk for i = 1 and 2. But suppose x = Bz, so that Then
k=l
2 ( 2 ) 2
Yi = ~ aik ~ bkz = ~ cz
k=l 3= 1 J 1 j=1 ~J j
Example 3.4 can be extended to show (AB)C = A(BC), i.e. matrix multiplication is asso.;.
ciative. But, in general, matrix multiplication is not commutative, AB =F BA. Also, there
is no matrix division.
Example 3.5.
To show AB # BA, consider
AB
(~ ~)(~ ~) D
BA (~. ~)(~ ~) C
BF
(~ ~)(~ ~) C
Suppose we have the vector equation Ax = Bx, where A and Bare n X n matrices. It
can be concluded that A = B only if x is an arbitrary nvector. For, if x is arbitrary, we
may choose x successively as el, e2, ... , en and find that the column vectors al = h1,
a2 = h2, ... , an = b n • Here ej are unit vectors, defined after Definition 3.17, page 41.
Definition 3.14: To partition a matrix, draw a vertical and/or horizontal line between two
rows or columns and consider the subset of elements formed as individual
matrices, called submatrices.
As long as the submatrices are compatible, i.e. have the correct order so that addition
and multiplication are possible, the submatrices can be treated as elements in the basic
operations.
Example 3.6.
A 3 X 3 matrix A can be partitioned into a 2 X 2 matrix All, a 1 X 2 matrix A21 , a 2 X 1 matrix A 12 ,
and a 1 X 1 matrix A22 •
A
( ~::+~:: )
A similarly partitioned 3 X 3 matrix B adds as
A + B
and multiplies as
AB
Facility with partitioned matrix operations will often save time and give insight.
CHAP.3J ELEMENTARY MATRIX THEORY 41
Example 3.7.
To find the determinant of a 3 X 3 matrix, all possible permntations of 1,2,3 mllst be found. Per
forming one transposition at a time, the following table can be formed.
P PI' P2' Pa
0 1, 2, 3
1 3, 2, 1
2 2, 3, 1
3 2, 1, 3
4 3, 1, 2
5 1, 3, 2
This table is not unique in that for P = 1, possible entries are also 1,3,2 and 2,1,3. However, these
entries can result only from an odd p, so that the sign of each product in a determinant is unique. Since
there are 8! = 6 terms, all possible permutations are given in the table. Notice at each step only two
numbers are interchanged. Using the table and (3.1) gives
det A = (I)Oallu22u33 + (1)1a12a21a33
+ (1)2a12u23aSI + (1)3u13a22a31 + (1)4a18u21a32 + (1)5ullu23a32
Theorem 3wl: Let A be an n X n matrix. Then det (AT) = det (A).
Proof is given in Problem 3.3, page 59.
Theorem 3.2: Given two n X n matrices A and B. Then det (AB) = (det A) (det B).
Proof of this theorem is most easily given using exterior products, defined in Section
3.13! page 56. The proof is presented in Problem 3.15, page 65.
AE
~) ==
Using Theorem 3.2 on the product AE or EA, it can be found that (i) interchange of two
rows or columns changes the sign of a determinant, i.e. det (AE) = detA, (ii) multiplication
of a row or column by a scalar a multiplies the determinant by a, i.e. det (AE) = a det A,
and (iii) adding a scalar times a row to another row does not change the value of the de
terminant, i.e. det (AE) = det A.
CHAP.3J ELEMENTARY MATRIX THEORY 43
Taking the value of a in (ii) to be zero, it can be seen that a matrix containing a row
or column of zeros has a zero determinant. Furthermore, if two rows or columns are
identical or multiples of one another, then use of (iii) will give a zero row or column, so that
the determinant is zero.
Each elementary matrix E always has _an inverse El, found by undoing the row or
column operation of I. Of course an exception is a = 0 in (ii).
Example 3.9.
The inverse of E = G~) is El = G~). The inverse of E = (~ ~) is El = (~ n.
Definition 3.27: The determinant of the matrix formed by deleting the ith row and the jth
column of the matrix A is the minor of the element aij, denoted detMij. The
cofactor Co = (l)Hi detMij.
Example 3.10.
The minor of a22 of a 3 X 3 matrix A is det M22 = alla33  a13aSl' The cofactor c22 = (1)4 det M22 =
detM22 •
Proof of this theorem is presented as part of the proof of Theorem 3.21, page 57.
Example 3.11.
The Laplace expansion of a 3 X 3 matrix A about the second column is
det A = a12c12 + a22cZ2 + aS2c32
 a I 2(a21 a 33  a23 a 31) + a22(a ll aaa  a13 a 31)  a32(alla23  a13 a 21)
Corollary 3.4: The determinant of a triangular n x n matrix equals the product of the
diagonal elements.
Proof is by induction. The corollary is obviously true for n = 1. For arbitrary n, the
Laplace expansion about the nth row (or column) of an n x n upper (or lower) triangular
matrix gives detA = annCnn • By assumption, Cnn = ana22' • ·anl,nl, proving the corollary.
Explanation of the induction method: First the hypothesis is shown true for n = no,
where no is a fixed number. (no = 1 in the foregoing proof.) Then assume the hypothesis
is true for an arbitrary n and show it is true for n + 1. Let n = no, for which it is known
true, so that it is true for no + 1. Then let n = no + 1, so that it is true for no + 2, etc.
In this manner the hypothesis is shown true for all n ~ no.
Corollary 3.5: The determinant of a diagonal matrix equals the product of the diagonal
elements.
Definition 3.29: The adjugate matrix of A is adj A = {cu} T, the transpose of the matrix
of cofactors ofA.
The adjugate is sometimes called "adjoint", but this term is saved for Definition 5.2.
Then (3.2) can be written in matrix notation as
A adj A = I det A (3.3)
Theorem 3.6: Cramer's rule. Given an n X n (square) matrix A such that detA =1= O.
Then
AI de! A adj A
Example 3.12.
The inverse of a 2 X 2 matrix A. is
Another and usually faster means of obtaining the inverse of nonsingular matrix A is
to use elementary row operations to reduce A in the partitioned matrix A II to the unit
matrix. To reduce A to a unit matrix, interchange rows until an =1= O. Denote the inter
change by E 1 • Divide the first row by an, denoting this row operation by E 2 • Then
E2E1A has a one in the upper left hand corner. Multiply the first row by Ui1 and subtract
it from the ith row for i = 2,3, ... , n, denoting this operation by Ea. The first column of
EaE2EIA is then the unit vector el. Next, interchange rows E3E2EIA until the element in
the second row and column is nonzero~ Then divide the second row by this element and
subtract from all other rows until the unit vector e2 is obtained. Continue in this manner
until Em' .. E1A = I. Then Em'" El = AI, and operating on I by the same row operations
will prod Uc€) A 1. Furthermore, det A 1 = det E1 det E 2 • •• det Em from Theorem 3.2.
Example 3.13.
To find the inverse of (_~ ~), adjoin the unit matrix to obtain
(~ ~ I ~ ~)
CHAP. 3] ELEMENTARY MATRIX THEORY 45
It turns out the first column is already ell and all that is necessary to reduce this matrix to I is to add the
second roW to the first (det Es = 1).
(~ o
1
I 11)
1 0
The matrix to the right of the partition line is the inverse of the original matrix, which has a determinant
equal to [(l)(l)(l)} ~l = 1.
Example 3.14.
(a) The matrix equation
has no solution because no ~ exists that satisfies the two equations written out as
~ 0
2~ 1
(~)
is satisfied for any ~.
To find the necessary and sufficient conditions for existence and uniqueness of solutions
of matrix equations, it is necessary to extend some geometrical ideas. These ideas are
apparent for vectors of 2 and 3 dimensions, and can be extended to include vectors having
an arbitrary number of elements.
Consider the vector (2 3). Since the elements are real, they can be represented as points
in a plane. Let (t 1 t 2 ) = (2 3). Then this vector can be represented as the point in the
t l , t2 plane shown in Fig. 31.
• (2 3)
~~+~~~~~~.~.~.~~
If the real vector were (1 2 3), it could be represented as a point in (~1 ~2 g3) space
by drawing the ~3 axis out of the page. Higher dimensions, such as (1 2 3 4), are harder
to draw but can be imagined. In fact some vectors have an infinite number of elements.
This can be included in the discussion, as can the case where the elements of the vector are
other than real numbers.
Definition 3.31: Let CUn be the set of all vect01'S with ncomponents. Let al and a2 be vectors
having n components, i.e~ al and a2 are in CUn. This is denoted al E CUn,
a2 E CUn. Given arbitrary scalars (31 and !32, it is seen that ((31al + !32a2) E <Vn,
i.e. an arbitrary linear combination of al and a2 is in G(}n.
'VI is an infinite line and is an infinite
G[}2
plane. To represent diagrammatically these
and, in general, rvn for any n, one uses the area
enclosed by a closed curve. Let 1l be a set of
vectors in G[}n. This can be represented as
shown in Fig. 32. Fig.32. A Set of Vectors 11 in "Un
Definition 3.32: A set of vectors 11 in rvn is closed under addition if, given any al E 11 and
any a2 E 11, then (at + a2) E 11.
Example 3.15.
(a) Given 11 is the set of all 3vectors whose elements are integers. This subset of 'V3 is closed under addi
tion because the sum of any two 3vectors whose elements are integers is also a 3vector whose elements
are integers.
(b) Given 'U is the set of all 2vectors whose first element is unity. This set is not closed under addition
because the sum of two vectors in 1l must give a vector whose first element is two.
Definition 3.33: A set of vectors U in G[}n is closed under scalar multiplication if, given any
vector a E'U and any arbitrary scalar /3, then (3a E 'U. The scalar f3 can be
a real or complex number.
Example 3.16.
Given 11 is the set of all 3vectors whose second and third elements are zero. Any scalar times any
vector in 'U gives another vector in 1l, so U is closed under scalar multiplication.
Definition 3.34: A set of nvectors 'U in ti()n that contains at least one vector is called a vector
space if it is (1) closed under addition and (2) closed under scalar multi
plication.
If a E 1{, where 'U is a vector space, then Oa::::: 0 E'U because 'U is closed under scalar
multiplication. Hence the zero vector is in every vector space.
Given 81, a2, ... , an, then the set of all linear combinations of the 3i is a vector space
(linear manifold).
1l
3.7 BASES
Definition 3.35: A vector space 'U in G(Jn is spanned by the vectors 31, a2, ... , ak (k need not
equal n) if (1) al E 11, a2 E 11, ... , ak E 11 and (2) every vector in 'U is a
linear combination of the ai, 32, . . . , ak.
· CHAP. 3] ELEMENTARY MATRIX THEORY 47
Example 3.17.
Given a vector space 'U in 'Va to be the set· of all 3vectors whose third element is zero. Then (1 2 0),
(1 1 0) and (0 1 0) span 'U because any vector in 'U can be represented as (a (1 0), and
Definition 3.36: Vectors a1, a2, .•. , ak E G()n are linearly dependent if there exist scalars
!31, !32, .•• , /3k not all zero such that /31a1 + (32a2 + ~ .. + j3kak O. =
Example 3.18.
The three vectors of Example 3.17 are linearly dependent because
Note that any set of vectors that contains the zero vector is a linearly dependent set.
Definition 3.37: A set of vectors are linearly independent if they are not linearly dependent.
Theorem 3.7: If and only if the column vectors a1, a2, •.• , an of a square matrix A are
linearly dependent, then det A = O.
Proof: If the column vectors of A are linearly dependent, from Definition 3.36 for some
13 1, i32 , ••• , f3 n not all zero we get 0 = f3 1a 1 + {32a2 + ... + f3nan, Denote one nonzero f3 as f3 i ,
Then
1 0 o
o 1 o
o 0 o
o 0 1
Since a matrix with a zero column has a zero determinant, use of the product rule of
Theorem 3.2 gives det A det E = O. Because det E = f3 i ¥ 0, then det A = O.
Next, suppose the column vectors of A are linearly independent. Then so are the
column vectors of EA, for any elementary row operation E. Proceeding stepwise as on
page 44, we find E 1 , ••• , Em such that Em'" ElA = I. (Each step can be carried out since
the column under consideration is not a linear combination of the preceding columns.)
Hence, (detEm)" . (detEI)(detA) = 1, so that detA ¥ O.
Using this theorem it is possible to determine if al, a2, ... , ak, k ~ n, are linearly de
pendent. Calculate all the k x k determinants formed by deleting all combinations of n k
rows. If and only if all determinants are zero, the set of vectors is linearly dependent.
Example 3.19.
Consider (~) and ( ; ) . Deleting the bottom row gives det G:) = O. Deleting the top row
gives det (! ~) = 12. Hence the vectors are linearly independent. There is no need to check the
Definition 3~38: A set of nvectors ai, a2, ... , ak form a basis for '11 if (1) they span ,'11 and
(2) they are linearly independent.
48 ELEMENTARY MATRIX THEORY [CHAP. 3
Example 3.2U.
Any two of the three vectors given in Examples 3.17 and 3.18 form a basis of the given vector space,
since (1) they span 1..( as shown and (2) any two are linearly independent. To verify this for (1 2 0) and
(1· 1 0), set
Example 3.21.
Any three non coplanar vectors in threedimensional Euclidean space form a basis of C0 3 (not necessarily
the orthogonal vectors). However, note that this definition has been abstracted to include vector spaces
that can be subspaces of Euclidean space. Since conditions on the solution of algebraic equations are the
goal of this section, it is best to avoid strictly geometric concepts and remain with the more abstract ideas
represented by the definitions.
Consider 'U to be any infinite plane in threedimensional Euclidean space 'V s. Any two noncolinear
vectors in this plane form a basis for 11.
Theorem 3.8: If al, a2, ... ,ak are a basis of 'U, then every vector in 'U is expressible
uniquely as a linear combination of a1, a2, ... , ak.
The key word here is uniquely. The proof is given in Problem 3.6.
To express any vector in 'U uniquely, a basis is needed. Suppose we are given a set of
vectors that span 'U. The next theorem is used in constructing a basis from this set.
Theorem 3.9: Given nonzero vectors a1, a2, ... , am E "Un. The set is linearly dependent if
and only if some ak, for 1 < k ~ m, is a linear combination of al, a2, ... , akl.
Proof of this is given in Problem 3.7. This theorem states the given vectors need only
be considered in order for the determination of linear dependency. We need not check
and see that each ak is linearly independent from the remaining vectors.
Example 3.22.
Given the vectors (1 1 0), (2 2 0) and (1 0 0). They are linearly dependent because (2 2 0) =
2(1 1 0). We need not check whether (1 0 0) can be formed as a linear combination of the first
two vectors.
To construct a basis from a given set of vectors al, a2, ... , am that span a vector space
11, test to see if a2 is a linear combination of al. If it is, delete it from the set. Then test
if 83 is a linear combination of al and a2, or only a1 if a2 has been deleted. N ext test
84, etc., and in this manner delete all linearly dependent vectors from the set in order. The
remaining vectors in the set form a basis of 'U.
Theorem 3.10: Given a vector space Ii with a basis a1, a2, ... , am and with another basis
bl, b2 , ••• , hI. Then m· = l.
Proof: Note al, hi, h2, ... , hI are a linearly dependent set of vectors. Using Theorem
3.9 delete the vector hk that is linearly dependent on al, hI, ... , h k 1. Then a1, hi, ... , b k 'l,
hk+ 1, ••• , bl still span 'U. Next note a2, a1, hI, ... , hk~l, b k +h .•• , hI are a linearly depen
dent set. Delete another bvector such that the set still spans 'U. Continuing in this manner
gives aI, •.. , 82, a1 span 'U. If l < m, there is an 81 + 1 that is a linear combination of aI, ... ,
82, al. But the hypothesis states the avectors are a basis, so they an must be linearly
independent, hence l ~ m. Interchanging the b and avectors in the argument gives m ~ l,
proving the theorem.
CHAP. 3] ELEMENTARY MATRIX THEORY 49
Since all bases in a vector space 11 contain the same number of vectors, we can give
the following definition.
Definition 3.39: A vector space 11 has dimension n if and only if a basis of 11 consists of n
vectors.
Note that this extends the intuitive definition of dimension to a subspace of 'V m •
o Ax
If all ~i are zero, x = 0 is the trivial solution, which can be obtained in all cases. To obtain
a nontrivial solution, some of the ~. must be nonzero, which means the ai must be linearly
~
dependent, by definition. Consider the set of all solutions of Ax = O. Is this a vector space?
(1) Does the set contain at least one element? '
Yes, because x =0 is always one solution.
(2) Are solutions closed under addition?
Yes, because ifAz = 0 and Ay = 0, then the sum x = z + y is a solution of
Ax=O.
(3) Are solutions closed under scalar multiplication?
Yes, because if Ax = 0, then j3x is a solution of A({3x) = O.
So the set of all solutions of Ax = 0 is a vector space.
Definition 3.40: The vector space of all solutions of Ax = 0 is called the null space of A.
Definition 3.41: The dimension of the null space of A is called the nullity of A.
Corollary 3.12: If A is an n x n matrix with n linearly independent columns, the null space
has dimension zero. Hence the solution x = 0 is unique.
Theorem 3.13: The dimension of the vector space spanned by the row vectors of a matrix
is equal to the dimension of the vector space spanned by the column vectors.
See Problem 3.9 for the proof.
Example 3.23. 1 2
Given the matrix ( 3). It has one independent row vector and t,herefore must have only one
246 '
independent column vector.
Definition 3.42: The vector space of all y such that Ax = y for some x is called the range
space of A.
It is left to the reader to verify that the range space is a vector space.
50 ELEMENTARY MATRIX THEORY [CHAP. 3
Example 3.24.
The range space and the null space may have other vectors in common in addition to the zero vector.
Consider
A = (~ ~) b = (~) c = (;)
Then Ab = 0, so b is in the null space; and Ac = b, so b is also in the range space.
Definition 3.43: The rank of the m x n matrix A is the dimension of the range space of A.
Theorem 3.14: The rank of A equals the maximum number of linearly independent column
vectors of A, i.e. the range space has dimension r.
The proof is given in Problem 3.10. Note the dimension of the range space plus the
dimension of the null space = n for an m x n matrix. This provides a means of determining
the rank of A. Determinants can be used to check the linear dependency of the row or
column vectors.
(2) x + y = y + x,
(3) (x + y) + z = x + (y + z),
(4) for each x and y in 'U there is a unique z in 1l such that x + z = y,
(5) aX is in 'U,
The vectors of Definition 3.7 (nvectors) and the vector space of Definition 3.34 satisfy
this definition. Sometimes a and (3 are restricted to be real (linear vector spaces over the
field of real numbers) but for generality they are taken to be complex here.
CHAP. 3] ELEMENTARY MATRIX THEORY 51
Example 3.25.
The set 11 of time functions that are linear combinations of sin t, sin 2t, sin St, ... is.a linear vector
space.
Example 3.26.
The set of all solutions to dx/dt = A(t, x is a linear vector space, but the set of all solutions to dx/dt =
A(t) x + B(t) u for fixed u(t) does not satisfy (1) or (5) of Definition 3.44, and is not a linear vector space.
Example 3.27.
The set of all complex valued discrete time functions x(nT) for n = 0,1, . .. is a linear vector space,
as is the set of all complex valued continuous time functions x(t).
All the concepts of linear independence, basis, null space, range space, etc., extend im
mediately to a general linear vector space.
Example 3.28.
The functions sin t, sin 2t, sin St, .. ' form a basis for the linear vector space 11 of Example 3.25, and
so the dimension of 1l is countably infinite.
Definition 3.45: A metric, denoted p(a, b), is any scalar function of two vectors a E 14 and
b E 14 with the properties
(1) p(a, b) ~ 0 (distance isalways positive),
(2) p(a, b) = 0 if and only if a =b (zero distance if and only if the
points coincide),
(3) p(a, b) = p(b, a) (distance from a to b is the same as distance
from b to a),
(4) p(a, b) + p(b, c) ~ p(a, c) (triangle inequality).
Example 3.29.
(a) An example of a metric for nvectors a and b is
p(a, b) =
(ab)t(ab)
+
J1I2
[1 (a  b)t (a  b)
(b) For two real continuous scalar time functions x(t) and yet) for to:::: t === tv one metric is
Definition 3.46: The norm, denoted Iiall, of a vector a is a metric from the origin to the
vector a E 11, with the additional property that the "yardstick" is not a
function of a. In other words a norm satisfies requirements (1)(4) of a
metric (Definition 3.45), with b understood to be zero, and has the addi
tional requirement
(5) Ilaall = lailiali.
52 ELEMENTARY MATRIX THEORY [CHAP. 3'
Definition 3.47: The inner product, denoted (x, y), of any two vectors a and b in 11 is a
complex scalar function of a and b such that given any complex numbers
a and {3,
(1) (a, a) ;:".. 0,
(2) (a, a) = 0 if and only if a = 0,
(3) (a, b)* = (h, a),
(4) (aa + ph, c) = a*(a, c) + ,8*(b, c).
An inner product is sometimes known as a scalar or dot product.
Note (a, a) is real, (a, ab) = a(a, b) and (a, 0) = o.
Example 3.33. n
Theorem 3.17: For any inner product (a, b) the Schwarz inequality [(a, b)]2::::::: (a, a)(b, b)
holds, and furthermore the equality holds if and only if a = ab or a or b
. is the zero vector. .
Proof: If a or b is the zero vector, the equality holds, so take b =F O. Then for any
scalar j3,
0::::::: (a + ph, a + fih) = (a, a)
+ ,8*(b, a) + j3(a, h) + 1J312(b, b)
where the equality holds· if and only if a + Ph = O.. Setting j3 = (b, a)/(b, b) and re
arranging gives the Schwarz inequality.
CHAP. 3] ELEMENTARY MATRIX THEORY 53
Example 3.36.
Using the inner product of Example 3.34,
Definition 3.49: The natural norm, denoted Ila112' of a vector a is IIal12 = V(a, a).
Definition 3.50: An orthonormal set of vectors aI, a2, ... ,ale is a set for which the inner
product
j}
{~
if i #
if i = j
where 8ij is the Kronecker delta.
Example 3.37.
An orthonormal set of basis vectors in 'V n is the set of unit vectors ell e21 ••. I en, where ei is defined as
o
o
= 1 _ ith position
o
o
Given any set of k vectors, how can an orthonormal set of basis vectors be formed from
them? To illustrate the procedure, suppose there are only two vectors, al and a2. First
choose either one and make its length unity:
becomes
Because hI must have length unity, Yl = IIa11121, Now the second vector must be broken up
into its components:
54 ELEMENTARY_ MATRIX THEORY [CHAP. 3
aj  i~ (bi , aj)bi
j1
82  (b l ,az)b1 = ~(~)
Y2
 ~(i
1
O)~(~)~(~)
Y2 V2
{2
1
1 0
_1
2V2
(!)
2
Theorem 3.19: Any finite dimensional linear vector space in which an inner product exists
has an orthonormal basis.
Proof: Use the GramSchmit orthonormalization procedure on the set of all vectors
in 'U, or any set that spans 'U, to construct the orthonormal basis.
Definition 3.51: Given a set of basis vectors h1, h2, ... , hn for G()n, a reciprocal basis is a set
of vectors r1, r2, ... , rn such that the inner product
i, j = 1, 2, ... , n (3.5)
Note that an orthonormal basis is its own reciprocal basis. This is what makes "break
ing a vector into its components" easy for an orthonormal basis, and indicates how to go
about it for a nonorthonormal basis in CVn.
, Example 3.40.
Consider the set 'U of time functions that are linear combinations of sin nt, for n = 1,2, . ... Then
00
any x(t) in 'U can be expressed as x(t) = ~ ~n sin nt. The operation of integrating with respect to
time is a linear operator, because n=l
Example 3.41.
The operation of rotating a vector in CU 2 by an angle ¢ is a linear operator of 'V 2 onto CU 2 • Rotation of a
vector aa + f3b is the same as rotation of both a and b first and then adding a times the rotated a plus fi
times the rotated b.
Theorem 3.20: Given a vector space 1,11 with a basis b 1 , b 2 , ••• , bn, ... , and a vector space
112 with a basis C1, C2, ••• , Cm, • • •• Then the linear operator L whose
domain is 'U 1 and whose range is in 112 can be represented by a matrix
{YjJ whose ith column consists of the components of L(b i ) relative to the
basis Cl, C2, ••• , Cm, • • • •
Proof will be given for n dimensional 'U 1 and nt dimensional 1l2 • Consider any x in 'U 1•
Then x = :t
i=1
pib i • Furthermore since L(b i ) is a vector in 'U 2 , determine Yji such that
L(x) t
i=l
PiL(b)
n
Hence the jth component of L(x) relative to the basis {c j } equals ~ Yji,811 i.e. the matrix
{Yji} times a vector whose components are the Pi. i=l
Example 3.42.
The matrix representation of the rotation by an angle ¢ of Example 3.41 can be found as follows.
Consider the basis e 1 = (1 0) and e2 = (0 1) of CU 2• Then any x = (fi1 f32) = file1 + f32 e2' Rotating
e 1 clockwise by an angle ¢ gives L(e1) =
(cos 4»e1  (sin ¢)e2. and similarly L(e2) =
(cos ¢)e2 + (sin 1»ev
so that 1'11 = cos 1>, 1'21 = sin ¢, Y12 = sin 1>. Y22 = cos ¢. Therefore rotation by an angle 1> can be
represented by
L
COS 1>
( sin 1>
sin
cos ¢
¢)
56 ELEMENTARY MATRIX THEORY [CHAP. 3
Example 3.43.
An elementary row or column operation on a matrix A is represented by the elementary matrix E as
given in the remarks following Definition 3.26.
The null space, range space, rank, etc., of linear transformation are obvious extensions
of the definitions for its matrix representation.
Definition 3.53: Given mvectors ai in 11, where 11 has dimension n. An exterior product
cf>P = 3i /\ 3j /\ ... /\ 3k for p = 0,1,2, ... , n is a vector in an abstract vector
space, denoted /\1YU, such that for any complex numbers a and (3,
(aai + (3aj) /\ ak /\ ' •. /\ at = a(ai /\ 3k /\ ... /\ al) + (3(aj /\ ak /\ ••• /\ aL) (8.6)
ai/\'" 1\3j/\" '/\ak/\'" /\ar = 3i/\" ' / \ ak/\" '/\3j/\'" /\31 (8.7)
3i /\ aj /\ ... /\ 3k # 0 if ai, 3j, ••• , 3k are linearly independent.
Equation (8.6) says ~p is multilinear, and (3.7) says pP is alternating.
Example 3.44.
The case p= 0 and p = 1 are degenerate, in that 1\ o'U is the space of all complex numbers and
1\ 111 = 1J, the original vector space of mvectors having dimension n. The first nontrivial example is then
1\2'U. Then equations (3.6) and (3.7) become the biliriearity property
(3.8)
By (3.10) we see that Hi /\ aj is linear in either 3i or 3j (bilinear) but not both because in general
(aai) /\ (a3) # a(3i /\ 3j).
Note that setting 3) = 3i in (3.9) gives Furthermore if and only if 3j is a linear com
bina tion of ai' ~ /\ 3j = O.
n n
Let b 1• b 2 , ••• , b n be a basis of 'U. Then 3 i ::::::: ~
/:=1
a'kbk and 3j = :I
1=1
I'tbl> so that
':CHAP.3] ELEMENTARY MATRIX .THEORY 57
,'Since hk /\ hk ::::: 0 and hk /\ b t ::::: b[ /\ b k for k> l, this sum can be rearranged to
11. 1.,..1
ai /\ aj = ~ ~ (llk'Yl lll'Yk)b k /\h l (3.11)
1=1 k=l
Because ai /\ aj is an arbitrary vector in /\2'11, and (3.11) is a linear combination of hie /\ hb then the vectors
hk /\ h z for 1"";;: k < l ".,;;: n form a basis for /\2'11. Summing over all possible k and l satisfying this rela
tion shows that the dimension of /\211 is n(n  1)/2 ::::: (;).
Similar to the case /\ 2U, if any ai is a linear combination of the other a's, the exterior
product is zero and otherwise not. Furthermore the exterior products bi /\ h j /\ ••• /\ bk for
1 ~ i < j < ... < k ~ n form a basis for/\ P'U, so that the dimension of /\ Pl.{ is
n!
(np)ip!
(8.12)
Definition 3.54: The determinant of a linear operator L whose domain and range is the
vector space U with a basis hi, h2' ... , bn is defined as
L(h~) /\ L(h2) /\ ... /\ L(bn)
detL hi /\ b 2 /\ ••• /\ bn
(3.13)
This definition is consistent with the Definition 3,53 for exterior products, and so
cpP /\ ",q is itself an exterior product.
Also, if 1n =n then a1/\ ••• /\ an2 /\ a n 1 is an nvector since ~nl has dimension n from
(3.12), and must equal some vector in 'U since u and /\ n'lU must coincide.
:Theorem 3.21. (Laplace expansion). Given an 11, x n matrix A with column vectors ai.
Then det A = a1 /\ 82 /\ ... /\ an = a[(a2 /\ ... /\ an).
C':,y Proof: Let ei be the ith unit vector in 'Vnand Ej be the ith unit vector in 'Vnl, i.e. ei has
',''1,'b.components and Ej has n 1 components. Then ai = alie1 + a2ie2 + ... + aniCn so that
58 ELEMENTARY MATRIX THEORY [CHAP. 3
The first exterior product in the sum on the right can be written
el /\ a2 /\ ... /\ an = el /\ (al2e1 + a22e2 + ... + an2en) /\ ••• /\ (alnel + ... + annen )
Using the multilinearity property gives
el /\ a2 /\ . • . /\ an
0)
a22
(0 )
aZn
el /\
( :
anz
/\ . . . /\ :
ann
Performing exactly the same multilinear operations on the left hand side as on the right
hand side gives
e, 1\ [a22 (:J + ... + an,(:'J} ... [a2n (:.) + ... + ann(:JJ
1\
so that a2/\ .•. /\ an = (Cll C21 ••• Cnl)T and Theorem 3.21 is nothing more than a state
ment of the Laplace expansion of the determinant about the first column. The use of column
interchanges generalizes the proof of the Laplace expansion about any column, and use of
det A = det AT provides the proof of expansion about any row.
Solved Problems
3.L Multiply the following matrices, where aI, 32, hI and b 2 are column vectors with n
elements.
(i) (1, 2) (!) (iii) (:D (b'lb2 ) (v)
(~ ~J(j)
(ii) (~) (3 4) (iv) (a,la2) ~D (vi) (a, Ia2) (~ ~)
U sing the rule for multiplication from Definition 3.13, and realizing that multiplication of a
k X n matrix times an n X m matrix results in a k X m matrix, we have
(ii)
1 X 3)
( (2 X 3)
(1 X4») = (iv) (vi)
(2 X 4)
3J ELEMENTARY MATRIX THEORY
59
Find the determinant of A (a) by direct computation, (b) by using elementary row
and column operations, and (c) by Laplace's expansion, where
0 0
2 0
A
(1 D
::=
3 1
0 0
(a) To facilitate direct computation, form the table
Since a determinant is all possible combinations of products of elements where only one element
is taken from each row and colUmn, the individual terms in the sum are the same. Therefore the
only question is the sign of each product. Consider a typical term from a 3 X 3 matrix: aala12u23,
i.e, P1 =3, P2 = =
1, Ps 2. Permute the elements through a12a31a23 to Ct12a2Sa3i> so that the row
numbers are in natural 1, 2, 3 order instead of the column numbers. From this, it can be concluded
in general that it takes exactly the same number of permutations to undo PI' P2;' . " Pn to 1,2, ... , n
as it does to permute 1,2, ... , n to obtain Pl' P2 • .. " Pn • Therefore p must be the same for each
product term in the series, and so the determinants are equal.
60 ELEMENTARY MATRIX THEORY [CHAP. 3
For n = 2, det V = O2  01' which agrees with the hypothesis. By induction if the hypothesis
can be shown true for n given it is true for n 1, then the hypothesis holds for n :=: 2. Note each
term of det V will contain one and only one element from the nth column, so that
3.5. Show' det(! ~) det A det C, where A and Care n x nand m x m matrices
respecti vely.
Either det A=::O or det A :/= O. If det A = 0, then the column vectors of A are linearly
dependent. Hence the column vectors of (:) are linearly dependent, so that
det(! ~) = 0
and the hypothesis holds.
If det A oF 0, then AI exists and
The rightmost matrix is an upper triangular matrix, so its determinant is the product of the
diagonal elements which is unity. Furthermore, repeated use of the Laplace expansion about the
diagonal elements of I gives
Use of the product rule of Theorem 3.2 then gives the proof.
CHAP.8J ELEMENTARY MATRIX THEORY 61
3.6. Show that if at, a2, ... , ak are a basis of 'U, then every vector in 'U is expressible
uniquely as a linear combination of aI, a2, ... , a".
Let x be an arbitrary vector in 'U. Because x is in 'U, and 'U is spanned by the basis vectors
aI' a2' ... , ak by definition, the question is one of uniqueness. If there are two or more linear
combinations of a l1 a2' ... , ale that represent x, they can be written as
k
x = ~ f3iai
i=1
and k
x ~ aiai
i=l
Because the basis consists of linearly independent vectors, the only way this can be an equality is
for f3i = ai' Therefore all representations are the same, and the theorem is proved.
Note that both properties of a basis were used here. If a set of vectors did not span 'U, all
vectors would not be expressible as a linear combination of the set. If the set did span 'U hut
were linearly dependent, a representation of other vectors would not be unique.
3.7. Given the set of nonzero vectors aI, a2, ... , am in G()n. Show that the set is linearly
dependent if and only if some ak, for 1 < k ,.:::: m, is a linear combination of aI, a2,
... , akl.
where the f3i are not all zero since ak is nonzero. Then
where not all the f3i are zero. Find that nonzero 13k such that all f3i = 0 for i> k. Then the
linear combination is
3.8. Show that if an In. x n n1atrix A has n columns with at most r linearly independent
columns, then the null space of A has dimension n  r.
Because A is m X n, then ai E CUm' X E "Un Renumber the ai so that the first r are the inde
pendent ones. The rest of the column vectors can be written as
(3.16)
beca use a r + I' .•. , an are linearly dependent and can therefore be expressed as a linear combination
of the linearly independent column vectors. Construct the n  r vectors Xl' X2 • ••. ,x n  r such that
62 ELEMENTARY MATRIX THEORY [CHAP. 3
0 0 1
Note that Ax1 = 0 by the first equation of (3.16), Ax2 = 0 by the second equation of (3.16), etc.,
so these are n  r solutions.
Now it will be shown that these vectors are a basis for the null space of A, First, they must
be linearly independent because of the different positions of the 1 in the bottom part of the vectors.
To show they are a basis, then, it must be shown that all solutions can be expressed as a linear
combination of the "4. i.e. it must be shown the Xj span the null space of A. Consider an arbitrary
solution x of Ax = O. Then
~n 0 0 1 0
nr
Or, in vector notation, x = ~ ~r+ixt + S
i=l
where s is a remainder if the xi do not span the null space of A. If s = 0, then the "4 do span
the null space of A. Check that the last n  r elements of s are zero by writing the vector equality
as a set of scalar equations.
nr
Multiply both sides of the equation by A. Ax ~ ~r+iAxi + As
i=l
3.9. Show that the dimension of the vector space spanned by the row vectors of a matrix
is equal to the dimension of the vector space spanned by the column vectors.
Without loss of generality let the first rcolumn vectors 3 i be linearly independent and let s
of the row vectors ai be linearly independent. Partition A as follows:
1 aLr+l atn
. , .... , .......... , . , ... .
~
I aLr+l
..... ·1·'·'·········,·······,·
Xr 1 ar. r + 1 • • . am
=
Xr + l i a r +1, r + 1 ar + L n
..... 'I·······················
Xm 1 am,rt 1
3] ELEMENTARY MATRIX THEORY 63
Yl ' .,
~~"""
Yr Yr+ I ' • • Yn)
( '~~~ , , .,' •• ,' , • ~~r' • , ~~:.~~ ~' , , ',' .. ,' • , ;~~
r+l
SOthat Xi = (ail ai2 ," and
air)
~
= (ali a2j yr a r + l,j)' Since the Xi are Tvectors, ~
i=l
bixi = 0
for some nonzero bi , Let the vector b T = (b l bz .,' br + 1) so that
1+1
o ~ biXi
i=l
Since the last n  l' column vectors a i are linearly dependent, ai = ~ Cl'ijaj for i = 1'+ 1, ' , "n.
r r 3=1
Then Yi = ~ a··y· so that bTYi = ~ aijbTYj = 0 for i = r+ 1, ' , "n, Hence
3=1 ~J J ]=1
o= (b TYl b TY2 '" bTYt bTYr+ 1 . . . bTYrJ = blal + b2a 2 + ... + br + 1 arT 1
Therefore l' + 1 of the row vectors ai are linearly dependent, so that s::::: '/", Now consider AT.
The same argument leads to r:::::: 8, so that r = 8,
Show that the rank of the m X n matrix A equals the maximum number of linearly
independent column vectors of A.
Let there be r linearly independent column vectors, and without loss of generality let them be
r
aI' a2 • ' , ., aT' The ai = ~ aijaj for i = r + 1, . , " n, Any Y in the range space of A can be ex
j=1
pressed in terms of an arbitrary x as
n
y = Ax = ~ aixi
i=l
This shows the ai for i = 1, ...• r span the range space of A, and since they are linearly inde
pendent they are a basis, so the rank of A = r.
For an m x n matrix A, give necessary and sufficient conditions for the existence and
uniqueness of the solutions of the matrix equations Ax = 0, Ax = b and AX = B
in terms of the column vectors of A.
For Ax = 0, the solution x = 0 always exists, A necessary and sufficient condition for
uniqueness of the solution x = 0 is that the column vectors a 1.,." an are linearly independent.
To'show the necessity: If a i are dependent, by definition of linearly dependent some nonzero ~i
n
exist such that ~ ai~[ = O. Then there exists another solution x = (~l , .. ~n) # O. To show
i=l n
sufficiency: If the a[ are independent, only zero ~i exist such that ~ ai~i = Ax = 0,
i=l
n
For Ax = b, rewrite as b = ~ aixi' Then from Problem 3.10 a necessary and sufficient
i=l
condition for existence of solution is· that b lie in the range space of A, i.e. the space spanned by
the column vectors. To find conditions on the uniqueness of solutions, write one solution as
n n n
(1']1 '1]2 ' • , 7J n ) and another as (~1 g2 ' ., ~?l)' Then b= ~ a i 7Ji = ~ ai~i so that 0 = ~ (1,11  ~i)ai'
i=1 i=1 i=l
The solution is unique if and only if aI' ... , an are linearly independent,
64 ELEMENTARY ·MATRIX THEORY [CHAP. 3
Whether or not b = 0, necessary and sufficient conditions for existence and uniqueness of
solution to Ax = b are that b lie in the vector space spanned by the column vectors of A· and that
the column vectors are linearly independent.
To show rank A = rankATA, note both A and ATA have n columns, Then consider any vector
y in the null space of A, i.e. Ay O. Then ATAy = =
0, so that y is also in the null space of ATA.
Now consider any vector z in the null space of ATA, i.e. ATAz = O. Then zTATAz = llAzll~ = 0, so
that Az = 0, i.e. z is also in the nun spaca of A. Therefore the null space of A is equal to the
null space of ATA, and has some dimension k. Use of Theorem 3.11 gives rank A = nk =
rank ATA. Substitution of AT for A in this expression gives rank AT = rank AAT.
3.13. Given an m x n matrix A and an n X k matrix B, show that rankAB ==== rank A,
rankAB ==== rankB. Also show that if B is nonsingular, rank AB = rank A, and
that if A is nonsingular, rankAB = rankB.
Let rank A = r, so that A has r linearly indepedendent column vectors a 1, ••. , a".. Then
r n
ai = ~ (lijaj _ for i = r + 1, ... , n.
j=l
Therefore AB = ~ aib: where
i=l
bi are the row vectors
of B, using partitioned multiplication. Hence
AB = ~
i=l
aib'[ +. i
~=r+l
~
]=1
a'ijajb; =
i=1
~ ai (bi + :i akib~)
k=r+l
so that all the column· vectors of AB are made up of a linear combination of the r independent
column vectors of A, and therefore rank AB ~ r.
Furthermore, use of Theorem 3.15 gives rank B = rank BT. Then use of the first part of this
problem with BT substituted for A and AT for B gives rankBTAT === rank B. Again, Theorem 3.15
gives rank AB = rank BTAT, so that rank AB === rank B.
If A is nonsingular, rank B = rank A l(AB) ~ rank AB, using AI for A and AB for B in
the first result of this problem. But since rank AB === rank B, then rank AB = rank B if A1
exists. Similar reasoning can be used to prove the relnaining statement.
3.14. Given n vectors Xl, X2, ••• , Xn in a generalized vector space possessing a scalar product.
Define the Gram matrix G as the matrix whose elements gij = (Xi, Xj). Prove that
detG = 0 if and only if Xl,X2, •• • ,Xn are linearly dependent. Note that G is a matrix
whose elements are scalars, and that Xi might be a function of time.
Suppose det G = O. Then from Theorem 3.7, /31g1 + fJ2g2 + ... + [:3ngn = 0, where g is a column
n n
vector of G. Then 0 = ~ fJiU'lj = ~ [:3i(Xi, x).
i=1 i=1
3.15. Given two linear transformations Ll and L2J both of whose domain and range are
in 11. Show det(LIL2) = (detLl)(detL2) so that as a particular case detAB =
det BA = det A det B.
Let U have a basis hi. b 2 , ••• , bn' Using exterior products, from (3.13),
L 1L 2 (b 1) L 1L 2 (b 2 ) /\ ••• /\ L 1L 2 (h n )
/\
hI /\ b2 /\ • • • /\ b n
If L2 is singular, the vectors L 2 (b i ) = Ci are linearly dependent, and so are the vectors L 1L 2 (b i ) =
£1 CCi)' Then det (L 1L 2 ) = 0 = det L 2 • If L2 is nonsingular, the vectors Ci are linearly independent
and form a basis of 'U. Then Cl/\ C2 /\ ••• /\ Cn ¥= 0, so that
Cl /\ c2 /\ • • • /\ en
Note use of this permits the definition of a projection matrix P = 1  abT(aTb}l such that
det P = 0, Pa = 0, PTb = 0, p2 = P, and the transformation y = Px leave:;; only the hyperplane
bTx pointwise invariant, i.e. bTy = bTpx = bT[I  abT(aTb)l]x :::; (b T  bT)x == O.
66 ELEMENTARY MATRIX THEORY [CHAP. 3
Supplementary Problems
3.17. Prove that an upper triangular matrix added to or multiplied by an upper triangular matrix results
in an upper triangulal' matrix. 
3.18. Using the formula given in Definition 3.13 for operations with elements, multiply the following
matrices (: ~)(: ~ s~ t)
Next, partition in any compatible manner and verify the validity of partitioned multiplication.
3.19. Transpose ( ~ 2
j
sin t
)
'
and then take the complex conjugate.
3.21. Prove all skewsymmetric matrices have all their diagonal elements equal to zero.
3.23. Prove that matrix addition and multiplication are associative and distributive, and that matrix
addition is commutative.
3.24. Find a nonzero matrix which, when multiplied by the matrix B of Example 3.5, page 40, results
in the zero matrix. Hence conclude AB = 0 does not necessarily mean A = 0 or B = O.
3.25. How many times does one particular element appear in the sum for the determinant of an n X n
matrix'!
(~ ~) I (~ ~) (~ 1 ~) 1
4
3.30.
3.31.
Given A =
G D· 1
5
Find AI.
1
3.33. Let a nonsingular matrix A be partitioned into Alh A 12 , A21 and A22 such that Au and A22  A21 Ail Al2
have inverses. Show that
AI =
(a) (b)
0'0'2122) (~1)
~2
= (0)
0
(0)
:~:) G:)
0'32
= (~)
0
Using algebraic manipulations on the scalar equations ai1~1 + ~2~2 = 0, find the conditions under
which no solutions exist and the conditions under which many solutions exist, and thus verify
Theorem 3.11 and the results of Problem 3.11.
3.36. Let x be in the null space of A and y be in the range space of AT, Show xTy == O.
3.38. For A = (_~ !). show that (a) an arbitrary vector z = (::) can be expressed as the
sum of two vectors, z = x + y. where x is in the range space of A and y is in the null space of the
transpose of A, and (b) this is true for any matrix A.
3.39. Given n X k matrices A and B and an m X n matrix X such that XA = XB. Under what conditions
can we conclude A = B?
3.40. Given x, yin '"'(..(, where b 1, b 2 , •• " b n are an orthonormal basis of 'U. Show that
n
(x, y) ~ (x, bi)(bi • y)
i=l
3.41. Given real vectors x and y such that IIxl12 = Ily112' Show ex + y) is orthogonal to (x  y).
3.44. Given the three vectors 81 = ev'IT9 4 3), a 2 = (v'li9 1 7) and a3 = eVi19 10 5), use
the GramSchmit procedure on al' a 2 and as in the order given to find a set of orthonormal basis
vectors.
3.45. Show that the exterior product 4P = 31/\ ... /\ a p satisfies Definition 3.44, i.e. that is an element
in a generalized vector space /\P'ti, the space of all linear combinations of pfold exterior products.
3.46. Show (aIel + a2e2 + aSe3) /\ ({he1 + f32e2 + f3a e s) = (a2f33  O!sf32)e1 + (alf33  0!3f31)e2 + (a1f32  a2f31)ea.
illustrating that 3/\ b is the cross product in 'Va.
3.47. Given vectors Xl. X2' •• "Xn and an n X n matrix A such that Yl' Y2' ... , Yn are linearly independent,
where Yi = Axi . Prove that XII x2' ... ,xn are linearly independent.
n!
3.48. Prove that the dimension of /\P'U =
(n  p)! p!
3.19.
0
2
'l7")
x2
(
j sin t
3.24. ( _22:
I" :
I" ) for any a and fJ
3.S1. No
3.34. No
3.38. (a) x = (_~) a and y =:: (~) () where a and () are arbitrary, and since x and yare independ
ent .they span G[J2'
3.44. b 1 == (v'i19 4 3)/12, b2 == (0 3 4)/5 and a3 is coplanar with al and a 2 so that only b 1 and b 2
are required.
Matrix Analysis
4.1 EIGENVALUES AND EIGENVECTORS
Definition 4.1: An eigenvalue of the n x n (square) matrix A is one of those scalars A that
permit a nontrivial (x # 0) solution to the equation
Ax = AX (4.1)
Example 4.1.
Find the eigenvalues of (~ :). The eigenvalue equation is
Then
{G :)  A (~ n}(::) (~) or
Definition 4.2: The characteristic polynomial of A is det (A  AI). Note the characteristic
polynOll1ial is an nth order polynomial. Then there are n eigenvalues
Al, A21 ••. , An that are the roots of this polynomial, although some might be
repeated roots.
Definition 4.4: Associated with each eigenvalue Ai of the n x n matTix A there is a nonzero
solution vector Xi of the eigenvalue equation AXi = AiXi. This solution
vector is called an eigenvector.
Example 4.2 ••
In the previous example, the eigenvector assocated with the eigenvalue 1 is found as follows.
or
(~)
Then 2xl + 4X2 = 0 and Xl + 2X2 = 0, from which Xl = :2X2' Thus the eigenvector Xl is Xl = (~) X2
where the scalar x2 can be any number.
Note that eigenvectors have arbitrary length. This is true because for any scalar a',
the equation Ax = Ax has a solution vector aX since A(ax) = aAx = aAx = A(ll'x).
69
70 MATRIX ANALYSIS [CHAP. 4
Definition 4.5: An eigenvector is normalized to unity if its length is unity, i.e. Ilxll = l.
Sometimes it is easier to normalize x such that one of the elements is unity.
Example 4.3.
The eigenvector normalized to unit length belonging to the eigenvalue 1 in the previous example is
Xl 1(2) h
= _r;;:
v5 1
I
Izmg Its fi rst e ement to umty gIves
, w ereas norma 1'" .. Xl = ( 1)
1/2
.
Example 4.4.
Written out, the state equations (4.2) are
dXI(t)/dt = allxI(t) + al2x 2(t) + ... + alnXn(t)
dX2(t)/dt = a2l x l(t) + a22x 2(t) + ... + a2nx n(t)
Now define a new variable, an nvector y(t), by the one to one relationship
y(t) = Ml.x(t) (4.3)
It is required that M be an n X n nonsingular constant coefficient matrix so that the solution
x can be determined from the solution for the new variable y(t). Putting x(t) = My(t)
into the system equation gives
M dy(t)/dt = AMy(t)
Multiplying on the left by Ml gives
dy(t)/dt = M1AMy(t) (4.4)
( ~ ~~ :'.'... ~.)(~:~~~)
Yl(t) )
dy(t) !!:. Y2(t) = .A.y( t)
a;r = dt : .. ..
(
Yn(t) o 0 ... An Yn(t)
CHAP. 4] MATRIX ANALYSIS 71
Writing this equation _out gives dyJdt = AiYi for i = 1,2, ... , n. The solution can be ex
pressed simply as Yi(t) = Yi(O)e Ait • Therefore if an M such that M1AM =.A. can be found,
solution of dx/dt = Ax becomes easy. Although not always, such an M can usually be found.
In cases where it cannot, a T can always be found where T1AT is aln10st diagonal.
Physically it must be the case that not all differential equations can be reduced to this
simple form. Some differential equations have as solutions teA;t, and there is no way to get
this solution from the simple form.
The transformation M is constructed upon solution of the eigenvalue problem for all
the eigenvectors Xi, i = 1,2, ... , n. Because AXi = A1Xi for i = 1,2, ... , n, the equations
can be "stacked up" using the rules of multiplication of partitioned matrices:
(AXI I AX21 ... I AXn)
(AlXl I A2 x 21 ... IAnXn)
o 0 ... An
Theorem 4.1: If the eigenvalues of an n X n matrix are distinct, then the eigenvectors are
linearly independent.
Note that if the eigenvectors are linearly independent, M is nonsingular.
Proof: The proof is by contradiction. Let A have distinct eigenvalues. Let Xl, XQ, ••• , Xu
be the eigenvectors of A, with Xl, XQ, ••• ,XI,; independent and Xk+ 1, ••• , Xu dependent. Then
k
~ {3 l..J
..::;.;
x.! for j = k+1, k+2, .. . ,n where not all (3 tJ.. = O. Since x.J is an eigenvector,
i=1
k
\Xj = Aj ~ f3ijXi
i=1
for j = k+1, .. . ,n
k Ii:
But the Xi, i = 1,2, ... , k, were assumed to be linearly independent. Because not all f3 ij
are zero, some Ai = Aj. This contradicts the assumption that A had distinct eigenvalues,
and so all the eigenvectors of A must be linearly independent.
Then the eigenvalues of 8 are found as the roots of det (8  AI) = O. But
det (S  AI) det (TIAT  AI)
det (TIAT  ATlIT)
det [Tl(A  AI)T]
Using the product rule for determinants,
det (S  AI) = det Tl det (A  AI) det T
Since detTl = (det T)l from Problem 3.12, det (S  AI) = det (A  AI). Therefore we
have proved
Corollary 4.3: All similar matrices have the same traces and determinants.
Proof of this corollary is given in Probleln 4.1.
A useful fact to note here "~lso is that all triangular matrices B display eigenvalues on
the diagonal, because the detetminant of the triangular matrix (B  AI) is th~ _product of
its diagonal elements.
Proof: By Theorem 4.2, the diagonal matrix A must have the eigenvalues of A appear
ing on the diagonal. If AT = TA, by partitioned matrix multiplication it is required that
Ati = Ait, where t are the column vectors of T. Therefore it is required that T have the
eigenvectors of A as its column vectors, and Tl exists if and only if its column vectors are
linearly independent.
It has already been shown that when the eigenvalues are distinct, T is nonsingular.
So consider what happens when the eigenvalues are not distinct. Theorem 4.4 says that
the only way we can obtain a diagonal matrix is to find n linearly independent eigenvectors.
Then there are two cases:
Case 1. For each root that is repeated k times, the space of eigenvectors belonging to
that root is kdirnensional. In this case the matrix can still be reduced to a diagonal form.
~ ~ ~).
Example 4.5.
Given the matrix A = ( Then det (A  AI) = i\.(1  i\.)2 and the eigenvalues are
1 0 0
0, 1 and 1. For the zero eigenvalue, solution of Ax = 0 gives x = (0 1 1). For the unity eigenvalue,
the eigenvalue problem is
(
~ ~ ~) (:~)
1 0 0 X3
(1) ( : ; )
X3
where xl and X2 are arbitrary. Hence' any two linearly independent vectors in the space spanned by
(0 1 0) and (1 0 1) will do. The transformation matrix is then
T M = C~ ~D
and TIAT = A., where A has 0, 1 and 1 on the diagonal in that order.
Note that the occurrence of distinct eigenvalues falls into Case 1. Every distinct eigen
value must have at least one eigenvector associated with it, and since there are n distinct
eigenvalues there are n eigenvectors. By Theorem 4.1 these are linearly independent.
Case 2. The conditions of Case 1 do not hold. Then the matrix cannot be reduced
to a diagonal form by a similarity transformation.
Example 4.6.
Given the matrix A = (~ ~). Since A is triangular, the eigenvalues are displayed as 1 and 1.
Then the eigenvalue problem is
which gives the set of equations X2 = 0, 0 = O. All eigenvectors belonging to 1 have the form (Xl O)T.
Two linearly independent eigenvectors are simply not available to form M.
J = (4.7)
74 MATRIX ANALYSIS [CHAP. 4
Each Lji(Ai) is an upper triangular square matrix, called a Jordan block, on the diagonal
of the Jordan form J. Several Lji(Ai) can be associated with each value of Ai, and may differ
in dimension from one another. A general Lji(Ai) looks like
Ai 1 0 0
0 Ai 1 0
Lji(Ai) 0 0 Ai 0 (4.8)
.................
0 0 0 Ai
where Ai are on the diagonal and ones occur in all places just above the diagonal.
o 0 0 A2
diagonal in a Jordan block, wherever a zero above the diagonal occurs in J there must occur a boundary
between two Jordan blocks. Therefore this J contains three Jordan blocks,
There is one and only one linearly independent eigenvector associated with each Jordan
block and vice versa. This leads to the calculation procedure for the other column vectors
tt of T called generalized eigenvectors associated wit~ each Jordan block Lji(Ai):
AXi AiXi
Atl Ait1 + Xi
(4.9)
Note the number of tl equals the number of ones in the associated Lji(Ai). Then
A(Xi I t1 It21 ... Itt I ... ) (AiXi I Ait1 + Xi I Ait2 + t1 I ... I Ait! + tl1 I •.. )
= (Xi Ih It21 ... Itl I ... )Lji(Ai)
This procedure for calculating the tl works very well as long as Xi is determined to within
a multiplicative constant, because then each tz is determined to within a mUltiplicative
constant. However, difficulty is encountered whenever there is more than one Jordan
block associated with a single value of an eigenvalue. Considerable background in linear
algebra is required to find a construction procedure for the t! in this case, which arises so
seldom in practice that the general case will not be pursued here. If this case arises, a
trial and error procedure along the lines of the next example can be used.
Example 4.8.
Find the transformation matrix T that reduces the matrix A to Jordan form, where
A G~ D.
CHAP. 4] MATRIX ANALYSIS 75
The characteristic equation is (2  ;\)(3  ;\)(1 ;\) + (2 ;\) = O. A factor 2 ;\ can be removed,
and the remaining equation can be arranged so that the characteristic equation becomes (2 ;\)3 = O.
Solving for the eigenvectors belonging to the eigenvalue 2 results in
What combination should be tried to start the procedure described by equations (4,.9)? Trying the general
expression gives
Then
7"2 + 73 = f3
7"2  73 [3
=
These equations are satisfied if a f3 This gives the correct x =
a(1 1 l)T. Normalizing x by setting
a = 1 gives t = ('7'1 '7'2 1  72)T. The transformation matl'ix is completed by any other linearly inde
pendent choice of x, say (0 1 _1)T, and any choice of '7'1 and 7'2 such that t is linearly independent of the
choices of all x, say '7'1 = 0 and '7'2 = 1. This gives AT = TJ, or '0
(~ ! ~)( ~ ~~)
o 1 1 1 0 1
(~~ ~)(~ ~ ~)
1 0 1 0 0 2
Example 4.9.
Some typical quadr~tic forms are
.w.l 7~i
.w.2 = 3~i  2~1~2 + ~~ + 5~1~3  7~2~1
Theorem 4.5: All quadratic forms E<. can be expressed as the inner product (x, Qx) and
vice versa, where Q is an n X n Hermitian matrix, i.e. Qt = Q.
Let Q = {qij} = i{aij+ajJ. Then qij = qw so Q is real and symmetric, and ~=XTQX.
Next, (x, Qx) to ~ (the problem is to prove the coefficients are real):
n n
(x, Qx) LL
= i=lj=l qi"~i~'
J J
and
Then n n
(x, Qx) = lex, Qx) + t(x, Qtx) = L ~ (qt·J + q:)~i~'
! i=lj=l J J
n n
So (x, Qx) = LL
i=1 j=1
Re (qi')~i~'
J J
=~ and the coefficients are real.
Theorem 406: The eigenvalues of an n X n Hermitian matrix Q = Qt are real, and the
eigenvectors belonging to distinct eigenvalues are orthogonal.
The most important case of real symmetric Q is included in Theorem 4.6 because the
set of real symmetric matrices is included in the set of Hermitian matrices.
Proof: The eigenvalue problems for specific Ai and Aj are
(4.11)
Since Q is Hermitian,
QtXj = A.jXj
Theorem 4.7: Even if the eigenvalues are not distinct, a set of n orthonormal eigenvectors
can be found for an n X n normal matrix N.
The proof is left to the solved problems. Note both Hermitian and real symmetric
matrices are normal so that Theorem 4.6 is a special case of this theorem.
Corollary 4.8: A Hermitian (or real symmetric) matrix Q can always be reduced to a
diagonal matrix by a unitary transformation, where UIQV =.A and
VI = vt.
Proof: Since Q is Hermitian, it is also normal. Then by Theorem 4.7 there are n
orthonormal eigenvectors and they are all independent. By Theorem 4.4 this is a necessary
and sufficient condition for diagonaIization. To show a transformation matrix is unitary,
construct V with the orthonormal eigenvectors as column vectors. Then
xt1
xt2
CHAP. 4] MATRIX ANALYSIS 77
But xTx
j = (Xi'X) = 8ij because they are orthonormal. Then utu = I. Since the column
vectors of U are linearly independent, U 1 exists, so multiplying on the right by Ul gives
ut = Ul, which was to be proven.
Therefore if a quadratic form fL = xtQx is given, rotating coordinates by defining
x = Uy gives ~ == ytutQUy = ytAy. In other words, ~ can be expressed as
£t = Al1Ylj2 + A21Y21 2 + ... + An IYnl 2
where the Ai are the real eigenvalues of Q. Note ~ is always positive if the eigenvalues
of Q are positive, unless y, and hence x, is identically the zero vector. Then the square root of
~ is a norm of the x vector because an inner product can be defined as (x, Y)Q = xtQy.
Definition 4.8: An 11, X n Hermitian matrix Q is positive definite if its associated quadratic
form ~ is always positive except when x is identically the zero vector.
Then Q is positive definite if and only if all its eigenvalues are> O.
Theorem 4.9: A unique positive definite Hermitian matrix R exists such that RR = Q,
where Q is a Hermitian positive definite matrix. R is called the square root
ofQ.
P1'oof: Let U be the unitary matrix that diagonalizes Q. Then Q = UAU t . Since
Au is a positive diagonal element of A, defineAl/2 as the diagonal matrix of positive A1/ 2.
Q = UAlIZAl/2Ut = UAl/2UtU.A 1 / 2 U t
Now let R = UA1!2Ut and it is symmetric, real and positive definite because its eigenvalues
are positive. Uniqueness is proved in Problem 4.5.
One way to check if a Hermitian matrix is positive definite (or nonnegative definite)
is to see if its eigenvalues are all positive (or nonnegative). Another way to check is to use
Sylvester's criterion.
Definition 4.10: The mth leading principal minor, denoted detQm, of the n x n Hermitian
matrix Q is the determinant of the matrix Qm formed by deleting the last
n  m rows and columns of Q.
Theorem 4.10: A Hermitian matrix Q is positive definite (or nonnegative definite) if and
only if all the leading principal minors of Q are positive (or nonnegative) .
....
A proof is given in Problem 4.6.
Example 4.11.
Given Q = {qij}' Then Q is positive definite if and only if
o < det Ql = q11: 0 < det Q 2 = det (q~l q12); .•. , 0 < det Q n = det Q
q12 QZ2
If < is replaced by ~. Q is nonnegative definite.
Rearrangement of the elements of Q sometimes leads to simpler algebraic inequalities.
78 MATRIX ANALYSIS [CHAP. 4
Example 4.12.
The quadratic form
Q. =
is positive definite if ql1 > 0 and qnq22  qi2 > O. But Q. can be written another way:
GI 
~ qll;1 2+ 2 q12;1;2 + 2 _ (;2;1) (q22
q22;2 . q12) (~2)
q12 qu ~1
which is positive definite if q22> 0 and qllq22  qi2 > O. The conclusion is that Q is positive definite if
det Q > 0 and either q22 or qu can be shown greater than zero.
Theorem 4.12: ]jAII 2  pmax' where pZmax is the maximum eigenvalue of AtA, and further
more
o ~ Pmin ~
IIAxllz ~ Pmax
Ilxllz
To calculate IIA1Jz' find the maximum eigenvalue of AtA.
AtAgi = P;gi
Since (x, AtAx) = (Ax, Ax) = [[Ax]!; ~ 0, then AtA is nonnegative definite and pf ~ O.
Since At A is Hermitian, PT is real and the gj can be chosen orthonormal. Express any x
n
in CV n as x = ~ eigi . Then
i=1 n n
I]AxI]~ ~ ]1~iAgill:
i=1
~ gfp;
i=1
Ilxll; n = n
~ II~igill~
i=1
~ ~f
i=1
n n n
Since (pnmin ~ ~f ~ ~ ~TP7 ~ (pT)max ~ ~L taking square roots gives the inequality of
i=l 1=1 i=1
the theorem. Note that IIAxliz = Pmaxllxl12 when x is the eigenvector gi belonging to (P~)lnaK'
Definition 4.12: Given an analytic function f(a) of a scalar a, the function of an n X n 1natrix
A is f(A) = i
k=O
tkAk/k!.
Example 4.14.
Some functions of a matrix A are
cosA = (cosO)I + (sinO)A + (cosOjA2/2 + ... + (1)mA2m/(2'm)! +
eAt = (eO)I + {eO)At + (e O)A2 t 2/2 + ... + Aktk/k! + ...
80 MATRIX ANALYSIS [CHAP~4
f(A) = k=O
00
"'1:, fkAk/k! = t
k=O
fk,TJkT 1/k!
Theorem 4.14: If A = TATl, where.A. is the diagonal matrix of eigenvalues Ai, then
Therefore
f(A) ==
/(A') I(A2) 0) (4.14)
(
o f(An}
Also, let T = (Xl Ix21 ... IXn) and Tl = (rl Ir21 . .. Irn)t, where ri is the reciprocal basis
vector. Then from (4.13) and (4.14),
f{A)
The theorem follows upon partitioned multiplication of these last two matrices.
The square root function /(0:) == a 1l2 is not analytic upon substitution of any Ai. There
fore the square root R of the positive definite matrix Q had to be adapted by always taking
the positive square root of Xi for uniqueness in Theorem 4.9.
Definition 4.13: Let f(a) == a in Theorem 4.14 to get the spectral representation of
A == TAT~l:
A =
CHAP. A] MATRIX ANALYSIS 81
Note that this is valid only for those A that can be diagonalized by a similarity trans
formation.
To calculate I(A) = T/(J)TI, we must find I(J). From equation (4.7),
(Ll1(Al~..
Jk =
o
0
Lmn(An)
)k =
_ ( L~l(Al~..
0
k 0
Lmn(An)
)
where the last equality follows from partitioned matrix multiplication. Then
I(J) ~ IkJk/k! =
/(Lll) 0) (4.15)
k=O ( o '. I(Lmn) ,
Hence, it is only necessary to find I(L) and use (4.15). By calculation it can be found that
for an l x l matrix L,
k ) AkUIJ
( ll
k ) Ak CZ 2 J (4.16)
( l 2
o o
where (mn) = (nm)!m!'
n! the number of combinations of n elements taken m at a time.
all = t
k,.=ll
Ik(
l
~)Ak(tl}/k! =
1
1:
k=ll
I k Ak CIl)/[(l1)!(kl+1)lJ (4.17)
but
f
k=ll
I k Ak  a  1 )/(kl+1)! (4.18)
The series converge since I(A} is analytic, so that by comparing (4.17) and (.4.18),
1 dlI/(A}
(l I) ! dA l  1
Therefore
/()") dl/dA [(ll)!J_ldl_ll/dAI_l)
I(L) =
(
0.... !~~)........... .[~l.~ .2} .!].~1.~1~~~/.~>".1~~
.. (4.19)
° 0 ... I(A}
From (4.19) and (4.15) and Theorem 4.13, I(A) can be computed.
Another almosLequivalent ,method that can be used comes from the CayleyHamilton
theorem.
Example 4.15.
Given A Then
Example 4.16.
From Example 4.15, the given A matrix satisfies 0 = A2  6A + 51. Then A2 can be expressed in
terms of A and I by
A2 = SA  51 (4.21)
Also A3 can be found by multiplying (4.21) by A and then using (4.21) again:
Similarly any power of A can be found by this method, including AI if it exists, because (4.21) can be
multiplied by AI to obtain
AI = (61A)/5
This method avoids the calculation of T and Tl at the expense of solving
n
Example 4.17.
For the A given in Example 4.15, cos A = riA + Y21. Here A has eigenvalues }.1 = 1, }.2 = 5. From
cos A = YtA + Y21 we obtain
cos }.1 Y1}.1 + "'12
cos
2
1( 11) + co~ (1 11)
1 1 2 1
Use of complex variable theory gives a very neat representation of f(A) and leads to
other computational procedures.
Theorem 4.17: If f(a} is analytic in a region containing the eigenvalues >"i of A, then
Proof: Si~ce f(A) = Tf(J)Tl = T [2~j .f [(s) (sl  J)l as]Tl, it suffices to show
that f(L) = 27ij f /(s) (s I ~ L)l ds. Since
sI  L
(
s ~~ s.~~~
o
.. .
0
.. ::: .... ..)
...
~
s)..,
then
(S  >..)11 (8  >..)12 1)
(81  L)l (s 2y)' ( ... ......~ (~~. ~).'~~ .. : : : .... ~ .~ ~..
o o . .. (8  >..)11
The upper right hand term all of 21. £: /(8) (sI  L)l ds is
7iJ J'
all 27ij
1 f f(s}ds
(8 + >..}1
Because all the eigenvalues are within the contour, use of the Cauchy integral formula then
gives
1 d1 1f(>..)
(Zl)! d>..ll
which is identical with equation (4.19).
Example 4.18.
Using Theorem 4.17,
cos A = 2~i f cos s(81  A)l ds
Then
cos A
_ 2 1:

cos
21rj:r (s  1)(8  5)
8 (8  3 2) 2 s  3
ds
84 MATRIX ANALYSIS [CHAP. 4
4.8 PSEUDOINVERSE
When the determinant of an n X n matrix is zero, or even when the matrix is not square,
there exists a way to obtain a solution "as close as possible" to the equation' Ax = y. In
this section we let A be an m X n real matrix and first examine the properties of the real
symmetric square matrices ATA and AAT, which are n X nand m x m respectively.
Proof: From Problem 3.12, there are exactly r nonzero eigenvalues of ATA and AAT.
Then ATAg.t = p~g,
t t
for i = 1,2, . .. , rand ATAg;• = 0 for i = r + 1, ... , n. Define an
=
mvector hi as hi == Ag/Pi for i 1,2, ... , r. Then
AA'I'bi = AATAgJpi = A(pfg)/Pi = pfhi
Furthermore,
hih; = gfATAg/ PiPj = Pjgfg/Pi = 8ij
Since for each i there is one normalized eigenvector, hi can be taken equal to fi and (2) is
proven. Furthermore, since there are r of the p?,t
these must be the eigenvalues of AAT and
(1) is proven. Also, we can find fi for i = r+ 1, ... , m such that AATfi= 0 and are ortho
normal. Hence the fi are an orthonormal b~sis for 'Vm and the gi are an orthonormal basis
for CV n •
To prove (3), ATAgi = 0 for i= r+l, ... ,m. Then IIAgill~ = gfATAgi = 0, so that
Agi = O. Similarly, since AATfi = 0 for i == r+ 1, ... , m, then ATfi = 0 and (5) is proven.
Finally, to prove (4),
(ATAg)/p~ P1?g./p·1
~
'CHAP. 4] MATRIX ANALYSIS 85
Example 4.19.
6 0 4 0
Let A  (6 1 0 6 ). Then m = r = 2, n = 4.
72 6 24 36)
AAT = (52 86) ATA = 6 1 0 6
86 73 ( 24 0 16 0
36 6 0 36
f1 = 0.6 0.8}T
f2 = 0.8 O.6)T
From the above, propositions (1), (2), (3), (4) and (5) of Theorem 4.19 can be verified directly. Computationally
it is easiest to find the eigenvalues p2 and p2 and eigenvectors fl and f2 of AAT, and then obtain the gj
from propositions (4) and (3). 1 2
l'
Proof; The m x' n matrix A is a mapping from CVn to CVm , Also, gl' g2' ... , gn and
£1' £2' .•. ,fm form orthonormal bases for CV n and CV m respectively. For arbitrary x in G() n'
n
X = ~ ~,g.
i=l 1 1
where ~i = gfx (4.22)
Note that the representation of Theorem 4.20 holds even when A is rectangular, and has
no spectral representation.
Example 4.20.
Using the A matrix and the results of Example 4.19,
6
(6
° 4 0)
1 0 6
= 10 (0.6) ~0.84 0.08 0.24 0.48) + 5 ( 0.8) (0.24 0.12 0.64 0.72)
0.8 0.6
Definition 4.14: The pseudoinverse, denoted A I, of the m x n real matrix A is the n x rn
r
real matrix A 1 = ~ p:lg j fJ.
i=l ~
Example 4.21.
Again, for the A matrix of Example 4.19,
=
0.1
o,84)
0.08 (0.6 0.8)
0.24
+ 0.2
( 0.24)
0.12 (0.8 a.6)
0.64
= (~~~~!! ~:~:~:)
0.1168 0.0576
(
0.48 0.72 0.0864 0.1248
86 MATRIX ANALYSIS [CHAP. 4
Theorem 4.21: Given an m X n real matrix A and an arbitrary mvector y, consider the
equation Ax = y. Define Xo = A  Iy . Then IIAx  Yl12 ~ IIAxo  YI12 and
for those z # Xo such that jlAz  Yl12 : : : II Ax o y112, then Ilzi 12 > Ilxo 112.
In other words, if no solution to Ax::::: y exists, Xo gives the closest possible solution.
If the solution to Ax = y is not unique, Xo gives the solution with the minimum norm.
P1~oof; Using the notation of Theorems 4.19 and 4.20, an arbitrary mvector y and an
arbitrary nvector x can be written as
m n
where rh::::: fry and ~i = gJx. Then use of properties (2) and (3) of Theorem 4.19 gives
n r m
Ax  y = ~ ~iAgi ~ (~iPi  'I])fi + ~ 'l]ifi (4.24)
i=l i=l i=r+l
To minimize IIAx  Yl12 the best we can do is choose gi = 7]/Pi for i::::: 1,2, ... , r. Then
those vectors z in G()n that minimize IIAx  ylk can be expressed as
r
z :::::
~
i=l
giTJ/Pi +
where ~i for i = r+l, ... , n is arbitrary. But
The z with minimum norm must have ~i::::: 0 for i::::: r + 1, ... , n. Then using 7]i ::::: fT y from
T T
(4.23) gives z with a minimum norm ::::: ~ p:l7]i ig = ~ p:lgif[y = A  I y ::::: Xo.
i=l 1 i=l 1
Example 4.22.
Solve the equations 6x! + 4X3 = 1 and 6xl + x 2 + 6X4 = 10.
This can be written Ax = y, where y = (1 10)T and A is the matrix of Example 4.19. Since the
rank of A is 2 and these are four columns, the solution is not unique. The solution Xo with the minimum
norm is
0.4728)
0.1936
0.4592
(
1.1616
Proof; For a solution to exist, 'lJi = 0 for i = r + 1, ... , m in equation (4.24), and
~i for i = 1,2, ... , r. Then any solution x can be written as
::::: 7]/Pi
r n
X = ~ gi"fJ/Pi ~
+ i=r+l gi'i (4.25)
i=l
where ti for i::::: r + 1, ... , n are arbitrary scalars. Denote in G()n an arbitrary vector
n
Z = ""
.tt::..; g.'.,I].
where f:.
 I
= g!'z.
~
Note from Definition 4.14 and Theorem 4.20,
i=l
T T
= ~ ~
~.tt::..;
plg.f!'f
i t1kkk
gTp (4.26)
k=l i=l
CHAP. 4] MATRIX ANALYSIS 87
n
Furthermore, since the gi are orthonormal basis vectors for 'V n , 1 = ~ gig;. Hence
i=l
n 7l.
From equation (4.23), TJi = fry so that substitution of (4.27) into (4.25) and use of Definition
4.14 for the pseudo inverse gives
x = AIy + (IA1A)z
Some further properties of the pseudoin verse are:
1. If A is nonsingular, A  I = A 1.
2. A IA =1= AA 1 in general.
3. AAIA=A
Solved Problems
4.1. Show that all similar matrices have the same determinants and traces.
To show this, we show that the determinant of a matrix equals the product of its eigenvalues
and that the trace of a matrix equals the sum of its eigenvalues, and then use Theorem 4.2.
Factoring the characteristic polynomial gives
det (A  AI) = (AI  },,)(A2  A)' .. (An  A) = AIA2" 'A n + ... + (Al + A2 + ... + An )(A)nl + (A)n
Setting A = 0 gives det A = Al A2' .. An. Furthermore,
det (A  AI) (al  Aed /\ (a2 l\e2) /\ ... /\ (an  Ae n )
+ (A) [el /\ a2 /\ ... /\ an. + al /\ ez /\ ... /\ an + ... + a1/\ a 2 /\ .•.
al /\ a2 /\ ... /\ an /\ en]
+ '" + ( A)nl [al /\ e2 /\ ... /\ en + el /\ a2 /\ ... /\ en + ... + el /\ e2 /\ ... /\ an]
+ (A) ne l/\ e2 /\ ... /\ en
Comparing coefficients of A again gives AIA2'" An = a 1 /\ a2/\ ..• /\ an' and also
Al + A2 + ... + An = al /\ ez /\ .•. /\ en + el /\ 32 /\ ••• /\ en + '" + el /\ e2 /\ ... /\ an
However,
al /\ e2 /\ ... /\ en :;:: (anel + a21e 2 + ... + an1en) /\ ~2 /\ ••• /\ en
(a) A
(! =: =~)
3 4 1
(c) A
( ~~o ~ ~)
01
(b) A
( ~o ~ =~)
13
(d) A
( ~o ~ 01
~)
(
8Ai 8 2 )
4
3
3  Ai
4
2
1  Ai
Xi =
(0)
0
0
giving Xl = G)' G)' G) x, = X3 =
where the third element has been normalized to one in X2 and Xs and to two in Xl' Then
G 2)( 8
D G DG D
3 2
4 1
3
2
3
2
1
=
3
2
1
0
2
0
G~ =DGJ G)
and only one vector, (a 0 O)T where a is arbitrary.
=
Therefore it can be concluded there is only
one Jordan block Lll(l), so
C~ D
1
J Lu(I) == 1
0
Solving,
11)
gives tl ==
G 2 4
1 2
(f3 2a a)T where f3 is arbitrary.
tl ::::
_Finally, from
G) == xl
11)
2 4 t2 :::::
1 2
we find t2 == (y 2f3  a {3  a)T where y is arbitrary. Choosing a, {3 and y to be 1,0 and 0
respectively gives a nonsingular (Xl I tl l,t2 ), so that
(~o
~ ~)(~ ~ =~)(~ ~ ~)
1 2 0 1 3 0 1 1
== (~ : ~)
0 0 1.
(
L (1)
I 0
0) (1~ ~1_~0)
L 2 (1) =
2o 1) 0 tl
o 0
from which f3 = 0 80 that Xl = (a 0 O)T and tl = (y 8 28  a)T where y and {} are also
arbitrary. From Problem 4.41 we can always take a = 1 and y, 8, etc. = 0, but in general
(d) The A matrix is already in Jordan form. Any nonsingular matrix T will transform it, since
A == I and Tl(I}T == I. This can also be seen from the eigenvalue problem
(A  AI)x = (I  (1)I)x = Ox = 0
so that any 3vector X is an eigenvector. The space of eigenvectors belonging to 1 is three
dimensional, so there are three Jordan blocks LU(A) = 1, L 1Z (A) = 1 and LlS(A) = 1
on the diagonal for A == 1.
90 MATRIX ANALYSIS [CHAP. 4
4.3. Show that a general normal matrix N (Le. NNt = NtN), not necessarily with distinct
eigenvalues, can be diagonalized by a similarity transformation U such that ut = Ul.
The proof is by induction. First, it is true for a 1 X 1 matrix, because it is already a diagonal
matrix and U = I. Now assume it is true for a k 1 X k 1 matrix and prove it is true for a
t l
k Xk matrix; i.e. assume that for U kl  Ukl
Let
Form T with the first column vector equal to the eigenvector Xl belonging to AI' an eigenvalue of Nk;.
Then form k  1 other orthonormal vectors x2' X3,' •• , Xk from 'V k using the Gram~Schmit process,
and make T = (xl I x21 ... I Xk)' Note TtT = I. Then,
AIXll n1 x2 nr Xk)
( ;; }Xl IX, I ... I xkl
(
~~~2.'... ~~~~ ........ ~.;.~k.
AIXkl ntx2 ntxk
and
Therefore
AIo a12
a22 ..•
a1k)
a2k
(A'at2t 0
a;z '"
0)
a~k
(Aiaf2 0
a;2 ••.
0a~2 ) (At at2 0 a12 •.• a2k
alk
)
==
( ••• oil ill " ............ oil .. oil ~
.... : .............. " • .. • • " .... II .... " ...... ill .. .. • .. " ... ill •• " .. " .. " .. .
o ak2 •.• akk aIk a;k ••• a~k a~k a;k ••• akk 0 ak2 •.• akk
Equating the first element of the matrix product on the left with the first element of the product
on the right gives
where
and Ak~l is normal. Since A k  l is k 1 X k  1 and normal, by the inductive hypothesis there
t
exists a U kl = Ul t
kl such that Uk  1 Ak  l Uk  1 = D, where D is a diagonal matrix.
CHAP. 4] MATRIX ANALYSIS 91
Therefore the matrix TS k diagonalizes N k , and by Theorem 4.2, D has the other eigenvalues
of N k on the diagonaJ.
A2, A3, ••• , Ak
I = StSk = StISk
To show the converse, start with AB = BA. Assume A has distinct eigenvalues. Then
Axi = AiXi, so that AB~ = BAxi = AiB~. Hence if Xi is an eigenvector of A, so is B~. For
distinct eigenvalues, the eivenvectors are proportional, so that BXi = Pi~ where Pi is a constant of
proportionality. But then Pi is also an eigenvalue of H, and ~ is an eigenvector of B. By normal~
izing the ~ so that Xi tXi = 1, U = (Xl I ... Ixn) simultaneously diagonalizes A and B.
If neither A nor B have distinct eigenvalues, the proof is slightly more complicated. Let A
be an eigenvalue of A having multiplicity m. For nondistinct eigenvalues, all eigenvectors of
A belong in the m dimensional null space of A  AI spanned by orthonormal xi' X2'" •• , "m' There
m
fore B~ = .~ cijXj'
t=l
where the constants cij can be determined by Cij = "T BXi' Then for
C = {cij} and X = (Xl Ix21 ... Ix n ), C = XtBX = XtBtX = ct, so C is an m X m Hermitian
matrix. Then C = UmDmU! where Dm and Um' are m X m diagonal and unitary matrices respec
tively, Now A(XUm ) = ;\(XUm ) since linear combinations of eigenvectors are still eigenvectors,
and Dm = Ut,XtBXUm . Therefore the set of m column vectors of XU m together with all other
normalized eigenvectors of A can diagonalize both A and B. Finally, (XUm)t(XUm ) = (UlxtxUm ) =
(UkImUm ) = 17/l> so that the column vectors XUm are orthonormal.
4.5. Show the positive definite Hermitian square root R, such that R2 = Q, is unique.
Since R is Hermitian, UA i ut = R where U is orthonormal. Also, R2 and R commute, so that
both Rand Q can be simultaneously reduced to diagonal form by Problem 4.4, and Q = UDUt.
Therefore D = .Ai. Suppose another matrix 8 2 = Q such that S = V A2 vt. By similar reason
ing, D = Ai. Since a number> 0 has a unique positive square root, .A2 =.A i and V and U are
matrices of orthonormal eigenvectors. The normalized eigenvectors corresponding to distinct
eigenvalues are unique. For any nondistinct eigenvalue with orthonormal eigenvectors Xl' x2" .. , x n ,
AXXt
92 MATRIX ANALYSIS [CHAP. 4
and for any other linear combination of orthonormal eigenvectors, Yl' Y2' ... ; Ym ,
where Tt = T 1 Then
m m'
Hence Rand S are equal even though U and V may differ slightly when Q has nondistinct eigen
values.
4.6. Prove Sylvester's theorem: A Hermitian matrix Q is positive definite if and only
if all principal minors det Qm > O.
If Q is positive definite, ~ == (x, Qx) ==== 0 for any x. Let Xm be the vector of the first m
elements of x. For those XO whose last m  n elements are zero, (x m• Qmxm) = (xO, QXO) ==== O.
Therefore Qm is positive definite, and all its eigenvalues are positive. From Problem 4.1, the
determinant of any matrix equals the product of its eigenvalues, so det'!m > O.
If det'lm > 0 for m = 1,2, ... , n, we proceed by induction. For n = 1, det Q = Al > O.
Assume now that if det Q 1 > 0, ... , det Qnl > 0, then Qnl is positive definite and must
possess an inverse. Partition Qn as
+qnnq:~\q)(: I Q;:l
q
<k = : )(_Q_:...,...1 ) (4.28)
We are also given det Qn > O. Then use of Problem 3.5 gives det'tn = (qnn  qtQ;~lq) detQnl'
so that qnn  qtQ;~lq > O. Hence
> o
for any vector (xtl 1 x~). Then for any vectors y defined by
substitution into (x, Qx) and use of (4.28) will give (y, QnY) > O.
11 ~ Akll
n=k+l
:::
11
~
=k+l
11Allk
CHAP. 4] MATRIX ANALYSIS 93
by properties (3) and (4) of Theorem 4.11. Using property (6) and IIAII < 1 gives S = lim Sk'
k+«!
Note JIAkt 1 11 ~ IIAll k + 1 so that lim Ak+l = O. Since (IA)Sk == !Ak+l, taking limits as
k_eo
k  co gives {I  A)S == I. Since S exists, it is (1  A)I. This is called a contraction mapping,
because lI(I  A)xlj :=: (1IIAli)llxll :=: Ilxll·
1 1 2)
4.8. Find the spectral representation of A
(o 1 1 .2
0 1
.
3
The spectral representation of A == ~ Ai~rit. The matrix A has eigenvalues AI::::: 1,
i=1
11.2 ::::: 1  j and A3=1 + j, and eigenvectors Xl = (1 1 0.5)T, x2::::: (j 1 O)T and xs::::: (i 1 O)T.
The reciprocal basis ri can be found as
1 j _j)1 ( 0 0
4 )
1 1 1 ::::: 0.5 i 1 2 2i
( 0.5 0 0 i 1 2+2j
Then
A (1) (~.5 )(0 0 2) + (1 j)(f) (0.5j 0.5 1 j) + (1 + j)(f) (0.5j 0.5 1 + j)
4.9. Show that the relations AA lA = A, A lAA 1 = A I, (AA 1)T = AA  I and
(A IA)T = A IA define a unique matrix A I, that can also be expressed as in De
finition 4.15.
r r
Represent A ~ Pifig[ and AI = ~ plgkf~. Then
i=1 k=1 k
AA1A
r r r
~ ~ ~ PiP~lpkfig'[ gjfJfkg~
i=l j=1 k=1 ,
r
AA IA = ~ Pifig'[ A
i=l
r
AIA = ~ gigf == (AIA)T
i=l
r
Similarly (AA 1)T == (~ fifT) T == AA I.
i=1
To show uniqueness, assume two solutions X and Y satisfy the four relations. Then
The following chain of equalities can be established by using the equation number above the
equals sign as the justification for that step.
X ~ XAX ~ ATXTX ~ ATyTATXTX ~ ATyTXAX ~ AT¥TX ~ YAX
~ YAYAX ~ YAyxTAT b yyTATXTAT ~ yyTAT b YAY ~ Y
Therefore the four relations given form a definition for the pseudoinverse that is equivalent to
Definition 4.15.
4.10. The outcome of y of a certain experiment is thought to depend linearly upon a para
meter x, such that y= aX + (3. The experiment is repeated three times, during
which x assumes values Xl = 1, X2 = 1 and X3 = 0, and the corresponding out
comes Y1 = 2, and Y2 == 2 and Y3 = 3. If the linear relation is true,
2 a(l)+ (3
2 a(l) + P
3 a(O) + f3
However, experimental uncertainties are such that the relations are not quite satisfied
in each case, so a and f3 are to be chosen such that
3
L (Yi  aXi  f3)2
i=l
is minimum. Explain why the pseudoinverse can be used to select the best a and f3,
and then calculate the best a and f3 using the pseudoinverse.
The equations can be written in the form y = Ax as
3
Defining Xo = AI y , by Theorem 4.21, Ily  Axolli = ~ (Yi  aOxi  f30)2 is minimized. Since
ATA = (~ :), then p, = V2 and P2 = va. an;' g, = (1 0) and g2 = (0 1). Since
f i = Ag/Pi' then f1 = (1 1 0)/"/2 and £2 = (1 1 1)/V3. Now AI can be calculated from
Definition 4.15 to be
t i
(i 1 !
0)
so that the best ao =2 and f30 = 1.
n
Note this procedure can be applied to ~ (Yi  axf  (3xi  y)2, etc.
i=1
4.11. Show that if an n X n matrix C is nonsinguIar, then a matrix B exists such that
C == eB" or B = In C.
Reduce C to Jordan form, so that C = TJTI. Then B = In C = Tin JT1, so that the
problem is to find In L(A) where L(X) is an l x l Jordan block, because
CHAP. 4] MATRIX ANALYSIS 95
Since A. ¥= 0 because Cl exists, In L(A.) exists and can be calculated, so that In J and hence In C
can be found.
Note B may be complex, because in the 1 X 1 case, In (1) = hr. Also, in the converse case
where B is given, C is always nonsingular for arbitrary B because Cl = e B •
Supplementary Problems
4.12. Why does at least one nonzero eigenvector belong to each distinct eigenvalue?
4.14. Suppose all the eigenvalues of A are zero. Can we conclude that A = O?
4.15. Prove by induction that the generalized eigenvector tt of equation (4.9) lies in the null space of
(A  AiI)l+l.
4.17. Let Xl and x2 be eigenvectors of a matrix A corresponding to the eigenvalues of A.! and A.2' where
A.l # A.2' Show that ax! + j3x2 is not an eigenvector of A if a # 0 and j3 # O.
4.18. U sing a similarity transformation to a diagonal matrix, solve the set of difference equations
where
(~)
4.19. Show that all the eigenvalues of the unitary matrix U, where utu = I, have an absolute value
of one.
4.20. Find the unitary matrix U and the diagonal matrix .A. such that
utC{ ~ t)u = A
4.22. Given a 3 X 3 real matrix P oft 0 such that p2:;::: O. Find its Jordan form.
I;Q
4.28. Show that the coefficient an of I in the CayleyHamilton theorem An + alAnl + ... + anI:;::: 0
is zero if and only if A is singular.
4.30. Let the matrix A have distinct eigenvalues A1 and A2' Does A3(A  A1I){A  A2I) == O?
4.31. Given a real vector x:::::: (Xl X2 ••• xn)T and a scalar a. Define the vector
grad x 0:: :;::: (aa/aX1 aa/fJX2 .•• fJa/fJxn)T
Show grad x xTQx = 2Qx if Q is symmetric, and evaluate grad x x TAx for a nonsymmetric A.
n t
4.32. Show that I:;::: ~ Xjri'
i=l
4.38. Prove that the listed properties 118 of the pseudoinverse are true.
4.40. Given a real m X n matrix A. Starting with the eigenvalues and eigenvectors of the real sym
metric (n + m) X (n + m) matrix ( : \ !T), derive the conclusions of Theorems 4.19 and 4.20.
CHAP. 4] MATRIX ANALYSIS 97
4.41. Show that the T matrix of J = TIAT is arbitrary to within n constants if A is an n X n matrix
in which each Jordan block has distinct eigenvalues. Specifically, show T = ToK where To is
fixed and
Kl 0 ... 0)
K =
(
.0." .~~ .. :::....~.
° ° ". Km
where K j is an l X l matrix corresponding to the jth Jordan block L j of the form
4.42. Show that for any partitioned matrix (A I 0), (A 10)1 (!I) .
4.43. Show that Ixt Axl === IIAI1211xlii .
4.44. Show that another definition of the pseudoinverse is AI = lim (ATA + d')lAT, where P is any
positive definite symmetric matrix that commutes with ATA. €+O
4.14. No
4.16. Yes
~}
0
COl)
4.20. A
u 1
0
U ~ V2 0
1 01
0
G ~.8).
0 1
4.21. T 0.6
0.8 0.6
J ;:::::
G D 1
0
0
1 0
4.22. J =
GD 0
0
or J =
G 0
0
98 MATRIX ANALYSIS [CHAP. 4
4.28. If AI exists, AI = a;l[An1 + a 1An2 + ... ]. If an = AIA2" .An = 0, then at least one eigen
value is zero and A is singular.
4.29. Yes
4.30. Yes
4.33. No
4.34. eAt
4e t + 6e 2t  est
Set  ge 2t + eSt
3e t + 4e2t  eSt
6e t  6e2t + eat
e t + 2e2t 
2et  3e2t + eSt
eSt)
(
4e t + 8e 2t + eat 8e t + 2e2t + eSt et + e2t + eSt
4.35.
sin wt)
cos CJJt
4.36. 1(4
25 :3
00)
4.37. AI = 5~ GD
4.40. There are r nonzero positive eigenvalues Pi and r nonzero negative eigenvalues Pi' Corresponding
to the eigenvalues Pi are the eigenvectors (g~) and to Pi are (_g~). Spectral representation of
(! I~T) then gives the desired result. f, f,
Definition 5.1: The transition matrix, denoted ~(t, to), is the n x n matrix such that
x(t) = ,pet; Xo, to) = (t(t, to)xo.
This is true for any to, i.e. x(t) = cI»(t, T) XCi) for 7' > t as well as ,=== t. Substitution of
x(t)= 4t(t, to)xo for arbitrary Xo in the zeroinput linear state equation dx/dt = A(t)x
gives the matrix equation for (t(t, to),
Bq,(t, to)/Bt = A(t) ",(t, to) (5.1)
Since for any Xo, Xo = x(t o) = (t(to, to)xo,
the initial condition on q,(t, to) is
q,(to, to) = I (5.2)
Notice that if the transition matrix can be found, we have the solution to a timevarying
linear differential equation. Also, analogous to the continuous time case, the discrete
time transition matrix obeys
cJt(k+l,m) = A(k)4»(k,m) (5.3)
99
100 SOLUTIONS TO THE LINEAR STATE EQUATION [CHAP. 5
In the continuous time case, cpl(t, to) always exists. However, in rather unusual cir
cumstances, A( k) may be singular for some k, so there is no guarantee. that the inverses
in equations (5.10) and (5.11) exist.
Proof of Theorem 5.1: Because we have a linear zeroinput dynamical system, the
transition relations (1.6) and (1.7) become lfJ(t, to) x(to) = 4l(t, tl) X(tl) and X(tl) = q;(tl, to) x(to).
Combining these relations gives 4t(t, to) x(to) = cp(t, t 1 ) 4t(tl, to) x(to). Since x(to) is an arbi
trary initial condition, the transition property is proven. Setting t2 = to in equation (5.5)
and using (5.2) gives cIJ(to, t 1) cIJ(tl, to) = I, so that if det cp(to, t 1 ) # 0 the inversion property
is proven. Furthermore let 8(t) = cp(t, 0) and set tl = °
in equation (5.5) so that
cIJ(tz, to) = 8(t2) 4t(O, to). Use of (5.6) gives cp(O,to) = cpl(to, 0) = gl(tO) so that the separation
property is proven.
To prove the determinant property, partition rp into its row vectors 4»1' 4»21 •.. , +n' Then
det cIJ = +1 /\ +2 /\ ... /\ +n
and
d(det cp)/dt d"t.
""1
/ dt /\ '1"2
A. /\ • • • /\..1..
"t"n
+ ..I.. /\
"t"I"1'2
d..l.. / dt /\ ... /\..1..
't'n
Because this is a linear, timevarying dynamical system, each element aik(t) is continuous
and singlevalued, so that this uniquely represents d+/dt for each t.
n
+1/\ ... /\ d+/dt /\ ... /\ +n +1/\ ... /\ L k=l
aik + k /\ ••• /\ 'Pn = aii +
1 /\'" /\ +i /\ ... /\ 9>n
Then from equation (5.13),
d( det q,)/ dt ==: an +
1 /\(1)2 /\ ••• /\ +n + a22+1 /\ +2 /\ •.• /\ +n + ... + ann+! /\+2/\ ••• /\ +n
[tr A(t)] det cp
Separating variables gives
d(det cIJ)/det cp = tr A(t) dt
Integrating and taking antilogarithms results in
ftt [tr A(T)] dr
det cp(t, to) = ye 0
where y is the constant of integration. Setting t = to gives det 4t(to, to) = det I == 1 = y,
so that the determinant property is proven. Since efW = 0 if and only if j(t) = 00, the
inverse of q,(t, to) always exists because the elements of A(t) are bounded.
The proof of the properties for the discrete time transition matrix is quite similar,
and the reader is referred to the supplementary problems.
CHAP. 5] SOLUTIONS TO TH~ LINEAR STATE EQUATION 101
00
Proof: The Maclaurin series for eAt =L Aktk/k!, which is uniformly convergent
k=O 00
1. Series method:
(5.16)
2. Eigenvalue method:
(5.17)
and, if the eigenvalues are distinct,
3. CayleyHamilton: nl
eAt L yJt)Ai (5.18)
i=O
nl
eLnO'l)t 0 o o o
o e L21 (Al)t o o o
ill ............................................. ill ill ill .... ill ill. ill.
o o eL;r(Al)t 0 o (5.19)
o o o eL12{A2)t o
o o o o
102 SOLUTIONS TO THE LINEAR STATE EQUATION [CHAP. 5
(5.20)
then
e"it teAit t~1 e"?it/(l 1) ')
4. Resolvent matrix:
1:. 1 {R(s)} (5.22)
where R(s) = (sI  A)I.
The hard part of this method is computing the inverse of (sI  A), since it is a poly
nomial in s. For matrices with many zero elements, substitution and elimination is
about the quickest method. For the general case up to about third order, Cramer's
rule can be used. Somewhat higher order systems can be handled from the flow diagram
of the Laplace transformed system. The elements rij (s) of R(s) are the response of the
ith state (integrator) to a unit impulse input of the jth state (integrator). For higher
order systems, Leverrier's algorithm might be faster.
Theorem 5.3: Leverrier's algorithm. Define the n x n real matrices F 1, F 2 , ••• , Fn and
scalars Bl, B2 , ••• , Bn, as follows:
trAFdl
trAF2/2
en = tr AFnln
Then
snl F1 + sn2F2 + ... + sFn 1 + Fn
(sI A)1 = sn + (hs n 1 + ... + en1S + en
(5.23)
Also, AFn + enI = 0, to check the method. Proof is given in Problem 5.4.
Having R(s), a matrix partial fraction expansion can be performed. First, factor
detR(s) as
detR(s) = sn + fh8 n 1 + ... + On1S + en = (s  1.1)(8  1.2)' .. (s  An) (5.24)
where the Ai are the eigenvalues of A and the poles of the system. Next, expand R(s) in
matrix partial fractions. If the eigenvalues are distinct, this has the form
1 1 1
R(8) R1
A1
S 
+ R"
S  1.2 ~
+ '" + \Rn
S  I\n
(5.25)
 1 d 
m i
I
(n~  i)! ds
'   sAmRs
m i  [( ) ()] S=X
(5.27)
Then eAt is easily found by taking the inverse Laplace transform. In the case of distinct
roots, equation (5.25) becomes
eAt = d'ltRl + eA2tR2 + ... + eAntRn (5.28)
3. CayleyHan1ilton: 111
~ Yj(k)Ai (5.32)
i=O
nl
where the Yi(k) are evaluated from Jk = ~ Yr(1c)Ji where from equation (4.15),
i=l
L~l(Al) 0 o o o
o L;'l(Al) o o o
o o o o
and if Lji(Ai) is l x l as in equation {5.20},
A~ kA~I 1l)[(l1) 1(1c  l + 1) 1] _1)
(k 1Ar+
(
~.....A.~ .... : : : ...(~.! ~: ~:~'! ~ (~~ ~~!. (.k. ~ .l.~~: :J.: (5.34)
o 0 ... Ai
4. Resolvent matrix:
(5.35)
where R(z) = (zI  A)I.
Since R(z) is exactly the same form as R(8) except with z for s, the inversion procedures
given previously are exactly the same.
104 SOLUTIONS TO THE LINEAR STATE EQUATION [CHAP. 5
The series method is useful if Ak = 0 for some k = k o• Then the series truncates at
ko  1. Because the eigenvalue problem Ax = Ax can be multiplied by A kl to obtain
0= AkX = AA k1X = Akx, then A = O. Therefore the series method is useful only for
systems with ko poles only at the origin. Otherwise it suffers from slow convergence,
roundoff, and difficulties in recognizing the resulting infinite series.
The eigenvalue method is not very fast because each eigenvector must be computed.
However, at the 1968 Joint Automatic Control Conference it was the general consensus
that this was the only method that anyone had any experience with that could compute
eAt up to twentieth order.
The CayleyHamilton method is very similar to the eigenvalue method, and usually
involves a few more multiplications.
The resolvent matrix method is usually simplest for systems of less than tenth order.
This is the extension to matrix form of the usual Laplace transform techniques for single
inputsingle output that has worked so successfully in the past. For very high order
systems, Leverrier's algorithm involves very high powers of A, which makes the spread of
the eigenvalues very large unless A is scaled properly. However, it involves no matrix
inversions, and gives a means of checking the amount of roundoff in thatAFn + onI should
equal o. In the case of distinct roots, R. = xir:t so that the eigenvectors can easily be
obtained. Perhaps a combination of both Leverrier's algorithm and the eigenvalue method
might be useful for very high order systems.
which does not simplify unless dT(t}/dt = 0 or some very fortunate combination of elements.
We may conclude that knowledge of the timevarying eigenvalues of a timevarying
system usually does not help.
The behavior of a timevarying system depends on the behavior of the coefficients of
the A(t) matrix.
Example 5.2.
Given the timevarying scalar system dE/dt = ~ sgn (t  t 1) where sgn is the signum function, so that
sgn (t  t 1) = 1 for t < tl and sgn (t  t 1 ) = +1 for t > t 1• This has a solution ~(t) = ~(to)e(tto)
for t < tl and ~(t) = ~(tl)e(ttl)for t> t l . For times t < t 1, the system appears stable,J~~t actually
the solution grows without bound as t ~ co. We shall see in Chapter 9 that the concept of staBility must
be carefully defined for a timevarying system.
Also, the phenomenon of finite escape time can arise in a timevarying linear system,
whereas this is impossible in a timeinvariant linear system.
CHAP. 5] SOLUTIONS TO THE LINEAR STATE EQUATION 105
Example 5.3.
Consider the timevarying scalar system
Then
Proof: Separating variables in the original equation, d~/~ = a(t)dt. Integrating and
taking antilogarithms gives ~(t) = foef:u(71) d71 •
Theorem 5.5: If A(t) A(T) = A(T) A(t) for all t, T, the timevarying linear differential sys
tem dx/dt == A(t)x has the transition matrIx 4'(t, T)
. = est T
A(7) d'TI
•
This is a severe requirement on A(t), and is usually met only on final examinations.
Taking derivatives,
a ftr A(l1) d1'l (5.37)
e
at
But from equation (5.36),
A(t)e f;A(71) d1)
Example 5.4.
~). t2,,2 fl..,. + t) ,we
Given A(t) ==: ( : Then from A(t) A(T) = G(t, T);;::: ( 0 1 see immediately
Fig.5~2
can be solved by assuming y ==: tAo Then a scalar polynomial results for A, analogous to the'
characteristic equation. If there are multiplicities of order m in the solution of this poly
nomial, y = (1n t)mlt"" is a solution for i = 0, 1,2, ... , ml.
CHAP. 5] SOLUTIONS TO THE LINEAR STATE EQUATION 107
A number of "classical" second order linear equations have closed form solution in the
sense that the properties of the solutions have been investigated.
Bessel's equation:
(5.38)
Hermite equation:
y 2ty + 2ay = 0
Laguerre equation:
ty + (1 t)y + ay = 0
with solution Ln(t), or
ty + (k + 1 t)y + (a  k)y = 0
with solution dkLn(t)/dt k •
Hypergeometric equation:
Ordinary:
t(1 t)ii + [y  (a + fi + l)t]y  afiy o
Confluent:
tii + (yt)iJ  ay =0
Mathieu equation:
ii + (a + fi cos t)y = 0 (5.89)
or, with T = cos t,
2
The solutions and details on their behavior are available in standard texts on engineering
mathematics and mathematical physics.
Also available in the linear timevarying case are a number of methods to give q.(t, T)
as an infinite series. Picard iteration, PeanoBaker integration, perturbation techniques,
etc., can be used, and sometimes give quite rapid convergence. However, even only three
or four terms in a series representation greatly complicate any sort of design procedure,
so discussion of these series techniques is left to standard texts. Use of a digital or analog
computer is recommended for those cases in which a closed form solution is not readily
found.
Theorem 5.7: (Floquet). Given the dynamical linear timevarying system dx/dt = A(t)x,
where A(t) = A(t + 1Il). Then
q.(t, T) == P(t, T)e RCt  T
)
It was shown in Example 3.25 that the solutions to dx/dt = A(t)x for any initial con
dition form a generalized vector space. The column vectors 4>i(t, T) of 4t(t, r) span this vector
space, and since det q,(t, T) 01= 0, the +Jt, T) are a basis. But equation (5.40) states that the
Pi(t + w, 'f) are solutions to dx/dt = A(t)x, so that
1'1
Then
c = q,(r, t) q,(t+w, T)
Note that Cl exists, since it is the product of two nonsingular matrices. Therefore by
Problem 4.11 the logarithm of C exists and will be written in the form
C = eWR (5.42)
If pet, r) can be any matrix, it is merely a change of variables to write
q,(t, T) = pet, r)eR(t'I") (5.43)
But from equations (5.43), (5.41) and (5.42),
P(t+w, r) = 4t(t+w, r)eR(t+~T) == q,(t, r)eWReR(t+IiJ'I")
q,(t, T)eRCt'I") = pet, T)
Example 5.6._
Consider the Mathieu equation d2 x/dt 2 + (0: + f3 cos t)x = 0 (5.39). Its periodic solutions are called
Mathieu functions, which exist only for certain combinations of  IX . and f3. The values of IX and 13
for which these periodic solutions exist are given by the curves in Fig. 5.3 below. These curves then
form the boundary for regions of stability.
CHAP. 5] SOLUTIONS TO THE LINEAR STATE EQUATION 109
Fig. 53
Whether the regions are stable or unstable can be determined by considering the point f3 = 0 and
a: < 0 in region 1. This is known to be unstable, so the whole region 1 is unstable. Since the curves
are stability boundaries, regions 2 and 6 are stable. Similarly all the odd numbered regions are unstable
and all the even numbered regions are stable. The line f3 =0, a:::::: 0 represents a degenerate case,
which agrees with physical intuition.
It is interesting to note from the example above that an originally unstable system
might be stabilized by the introduction of a periodicallyvarying parameter, and vice versa.
Another use of Floquet theory is in simulation of cJt(t, T). Only c)(t,7") for one period w
need be calculated numerically and then Floquet's theorem can be used to generate the
solution over the whole time span.
Proof: Since the equation dx/dt = A(t)x has a solution x(t) = lIt(t, to), in accordance
with the method of yariation of parameters, we change variables to k(t) where
x(t) = cIJ(t, to) k(t) (5.45)
110 SOLUTIONS TO THE LINEAR STATE EQUATION [CHAP. 5
Integrating from to to t,
k(t) = k(to) + rt +(to, 1') B(T) U(T) dT
Jto (5.46)
Since equation (5.45) evaluated at t = to gives x(to) = k(t o), use of (5.45) in (5.46) yields
and
y(t) = CeA<tto)x(to) + _
J
r CeA<tT)Bu(T) dT + Du(t)
to
t
(5.48)
kl
x(k) 4»(k, m) x(m) + ~ iIJ(k, i + 1) B(j) uU) (5.51)
j=m
This is very similar to the corresponding equation (5.44) for differential equations
except the integral is replaced by a sum.
Often difference equations result from periodic sampling and holding inputs to dif
ferential systems.
net) X u(tk )
Hold
dx/dt = A(t)x + B(t)u yet)
o 'T J y = C(t)x D(t)u + T
Fig. 54
In Fig. 54 the output of the hold element is u(k) = U(tk) for tk ~ t < tk+ 1, where
= T for all k. Use of (5.44) at time t = tk+1 and to = tk gives
~k+l  tk
x(k + 1) (5.52)
where the subscript 8 refers to the difference equations of the sampled system. For time
invariant differential systems, As = eAT, Bs = iT eACTr)B dT, C S :::::: C and Ds = D. Since
in this case As is a matrix exponential, it is nonsingular no matter what A is (see the com
ment after Problem 4.11).
Although equation (5.50) is always a representation for iIJ(k, m), its behavior is not
usually displayed. Techniques corresponding to the differential case can be used to show
this behavior. For instance, Floquet's theorem becomes cp(k, m) = P(k, m)Rkm, where
P(k, m) = P(k + 00, m) if A(k) = (k + Also (0).
has solutions of the form )..k/k!. Piecewise timeinvariant, classical second order linear,
and series solutions also have a corresponding discrete time form.
112 SOLUTIONS TO THE LINEAR STATE EQUATION [CHAP. 5
This suggests that y(t) can be written as a matrix generalization of the superposition
integral,
y(t) = (5.54)
where H(t, T) is the impulse response matrix, i.e. hij(t,T) is the response of the ith output at
time t due to an impulse at the ith input at time T. Comparison of equations (5.53) and
(5.54) gives
t===T
H(t, T) = t<T
(5.55)
In the timeinvariant case the Laplace transformation of H(t, 0) gives the transfer function
matrix
~{H(t,O)} = C(sIA)lB +D (5.56)
where
C(k) 4J(k, m. + 1) B(m) k>m
H(k,m) = D(k) k= m (5.58)
{ o k<m
DeJinition 5.2: The adioint, denoted La, of a linear operator L is defined by the relation
(p, Lx) = (LaP, x) for all x and p (5.60)
We are concerned with the system dx/dt = A(t)x. Defining L = A(t)  Id/dt, this· becomes
Lx = O. Using the scalar product (p,x) =
equation (5.60) using integration by parts.
it1 to
ptxdt, the adjoint system is found from
(p,Lx) =
i tl
ptA(t)x dt 
it! dx
pt dt dt
to to
Since Lx = 0 for all x, then (p,Lx) = 0 for all xand p. Using (5.60) it can be con
cluded LaP = 0, so that the adjoint 8ystem is defined by the relation
dp/dt = At(t)p (5.61)
Denote the transition matrix of this adjoint system as "M(t, to), i.e.,
o~(t, to)/ot = At(t) ~(t, to) with ~(to, to) = I (5.62)
Theorem 5.10: Given the system dx/dt = A(t)x + B(t)u and its adjoint system dp/dt =
At(t)p. Then tl
The column vectors I/Ii(t, to) Ofik(t, to) are the reciprocal basis to the column vectors tpi(t, to)
of +(t, to). Also, if u(t) = 0, then pt(t) x(t) = scalar constant for any t.
From the transition relation, x(t) = .(t, to) x(to) and pet) = 'i'(t, to) p(to) so that
pt(to) Ix(to) = pt(to) 'i't(t, to) eIt(t, to) x(to)
for any p(to) and x(to). Therefore (5.64) must hold.
The adjoint system transition matrix gives another way to express the forced solution
of (5.44):
x(t)
The variable of integration T is the second argument of (J(t, 7"), which sometimes poses simula
tion difficulties. Since eIt(to, T) = .l(T, to) = 'i't(T, to), this become~
x(t) q,(t, to) [x(to) + J: ,.t(T, to) OtT) U(T) dTJ (5.65)
in which the variable of integration 7" is the first argument of "!V(T, to).
The adjoint often can be used conveniently when a final value is given and the system
motion backwards in time must be found.
Exa~:~:n5.7~x/dt = (3: ~:)x. Use the adjoint system to find the set of states (",(1) "2(1)) that
permit the system to pass through the point x1(2) = 1.
Since pt(2)x(2) = pt(l}x{l). if we choose pt(2) = (10), then (10)x(2) = xl(2) = 1 = pt(l)x(l) =
Pl(l) xl(l) + P2(!) x 2(1). But pel) = .(1,2) p(2), so that
pt(l) = 0.2(3e1.5 + 2e6 2e6  2e1.5)
The set of states xC!) that gives xl(2) = 1 is determined by
1 = (O.6e1.5 + OA( 6)xl(1) + (OAe 6  0.4e1.5)x2(1)
Solved Problems
5.1. Given 4»(t, to), find A(t).
Using (5.2) in (5.1) at time to = t, A(t) = a~{t, to)/at evaluated at to = t. This is a quick
check on any solution.
(./2 0)
G D C~
0 0
eAt o) o + ...
= 1 + 4t 4t + 0 6t 2 8t2
0 t o 0 2t2 _2t2
=
C +!'2+ ...
t
1 2t
0
+ 4t2/2  2t(1 ~ 2t) + ...
tel  2t + 4t2/2 + ... )
4t(1 2t +o4t2/2 + ... )
1  2t + 4t2/2 + 2t(1  2t) + ...
)
(b) The eigenvalues of A are 1, 2 and 2, with corresponding eigenvectors (1 0 O)T and (0 2 l)T.
The generalized eigenvector corresponding to 2 is (0 1 l)T, so that
(~ ~
o 1
~)
0
= (~ ~ ~)(~ ~ ~)(~ ~ ~)
0 1 1 0 0 2 0 1 2
Using equation (5.17),
(c) Again, the eigenvalues of A are calculated to be 1, 2 and 2. To find the Yi (t) in equation
(5.18),
eJt =
C~'
0
e 2t
0
t'~2)
e2t
== roO
0
1
0
0) (1
~ + Y1 ~ 2o
o
1
2
0) + Y2 C 0)
0
0
o4
o
4
4
which gives the equations
e t = Yo  Y1 + Y2
e 2t Yo  2Y1 + 4Y2
te 2t = Y1  4Y2
o
12 16
0)
4 4
Summing these ll'latrices again gives the answer obtained in (a).
1 2 4
Res) o 8 +2 (s + 2)2 (s + 2)2
1 _1_+_2_
o (8 + 2)2 s+2 (8+2)2
R(s)
= .! C D .!2G D (.:2 G D
1 :
0
0
0
+
0
1
0
+ P
0
2
1
Addition will give R(s) as in (d1).
(dS) The flow diagram of the Laplace transformed system is shown in Fig. 5·5.
Fig. 55
Therefore,
1
0 0
s+l
8 4
R(s) 0
(s + 2)2 (8 + 2)2
1 8+4
0 (8 + 2)2
(s + 2)2
Again a partial fraction expansion can be performed to obtain the previous result.
F1 ::::: I 01 ::::: 5
F2 :::::
(1 0;)+51
04
o1
82 ::::: 8
Fa:::::
(4 ~ 8
0
1 4
~) +81 83 4
~G
0 0 0
RCs) =
1
0 D+·G
S3 + 582
1
1
+ 88
D+ G
+ 4
0
1 D
A partial fraction expansion of this gives the previous result.
CHAP.5J SOLUTIONS TO THE LINEAR STATE EQUATION 117
5.3. Using (a) the eigenvalue method and then (b) the resolvent matrix, find the transition
matrix for
(a) The eigenvalues a.re 1 and 2. with eigenvectors (1 O)T and (1 1)T respectively. The re
ciprocal basis is (1 1) and (0 1). Using the spectral representation equation (5.31),
,2"1
z 1
z+ z 1 z
z+ z)
+ 2. (l)k  (2».:)
(2)k
( z
o z+2
An intermediate result must be proven before proceeding, namely that tr F(s) == d¢/ds. In the
proof of Theorem 3.21, it was shown the cofactor cli of a general matrix B can be represented as
elj ::::: ej /\ b2 /\ ••• /\ bn and similarly it can be shown cij == b l /\ ... /\ hil A ej /\ bi+ 1/\ ••• /\ b n • so
that letting B == sl  A and using tr F(s) == c11(s) + c22(8) + ... + enn(s) gives
tr F(s) == el /\ (se2  a2) /\ ... 1\ (sen  an) + (se l  al) /\ e2/\ ... /\ (sel  an)
+ ... + (sel  al) /\ ... 1\ (8en l  anI) /\ en
But
and
drp/ds == el/\ (se2  a2) /\ ... /\ (sen  an) + .. , + (sel  al) /\ ... /\ en == tr F(s)
and so the intermediate result is established.
Substituting the definitions for ¢(s) and F(s) into the intermediate result,
trF(s) == tr (snlF l + sn2F2 + ... +sFn  1 +Fn)
== snl trFl + sn2 trF2 + .. , + s trFn 1 + trFn
drp/ds ::::: nsnl + (n 1)8 1 S n  2 + ... + 8n  l
for k = 1,2, .. . ,nI. These are half of the relationships required. The other half are obtained
by taking _their trace
A(t) A(T) =
It can be shown similar to Problem 4.4 that two n X n matrices Band C with n independent eigen
vectors can be simultaneously diagonalized by a nonsingular matrix T if and only if B commutes
with C. Identifying A(t) and A(T) for fixed t and T with Band C, then A(t) = T(t)A(t)Tl(t) and
A(T) = T(T)A(T)Tl(T) means that T(t) = T(T) for all t and T. This implies the matrix of eigen
vectors T is constant, so dT/dt = O. Referring to Example 5.1, when dT(t)/dt = 0 then a general
solution exists for this special case of A(t)A(T) = A(T)A(t).
For the given timevarying system, A(t) has the eigenvalues Al = a + je t and i\.2 = a  jet.
Since A(t) commutes with A(r), the eigenvectors are constant, (j l)T and (j 1)T. Consequently
T = (
1
j
~) A(t) =
(
a 01et T'l ::;:: 1
2
(1 1)
j 1
so that
sin (etl e t2»)
cos (et1  e t2 )
X ea(tito)
COS (eto  e t1 ) sin (eto  e t1 »)
( sin (eto  e t1 ) cos (eto  e t1 )
To verify 4t l (t l • to) ::;:: Iilo(to, til, calculate IiloI(t I ; to) by Cramer's rule:
4t I (t t) ::;::
ea(tot1)
.
(COS. (eto  et1) sin (eto  etl»)
1, 0 cos2 (eto  e:t1) + 8m2 (eto  e tI) sm (eto  et1) cos (eto  et1)
Since cos2 (eto  etl) + sin2 (eto  et1) ::;:: 1 and sin (eto  et1) ::::: sin (et1 eto), this
equals 4t(to, til.
8(t) ::;::
sin (1 e t »)
cos (1 e t )
Then
8 I (t) :::: e at COS .(1 e t ) sin (1 e t »)
( sin (1 e t ) cos (1 e t )
Since we have
cos (eto  e t1 ) = cos (eto 1) cos (1 e t1)  sin (e to  1) sin (1  etl)
and a similar formula for the sine, multiplication of B(tI ) and 8 I (to) will verify this.
Finally, det !ft(t, 7") ::;:: e2a (t't) and trA(t)::;:: 2a, so that integration shows that the determinant
property holds.
dx/dt = (0 1)x
1  (~+ a)/t 2a  lit
Writing out the matrix equations, we find dxl/dt::;:: x2 and so
d 2x l ldt2 + (2a + lit) dxlldt + [1 + (0:2 + a}/tJxl ::::: 0
Multiplying by the integrating factor teat, which was found by much trial and error with the
equation, gives
t(e at d 2x l /dt 2 + 2/Xe at dx1/dt + a2eatxl) + (eat dx1/dt + o:eatxl) + teatXI 0
which can be rewritten as
td2(eatXl)/dt2 + d(eatxI)ldt + teatxI ::;:: 0
This has the same form as Bessel's equation (5.38), so the solution is
xI(t) cleatJO(t) + c 2e at Y O(t)
x2(t) = dx1ldt ::;:: o:Xt(t) + eat[cIJI(t) + c2dYo/dt]
120 SOLUTIONS TO THE LINEAR STATE EQUATION [CHAP. 5
eaT C~2(T)X~T~Xl(T»)
80 that x(t) = FG(t) Gl(T) Flx (T), where
F = (1 a ~)
Yo(t) )
dYo/dt
and so ~(t, T)= FG(t) Gl(T)FI. This is true only for t and T> 0 or t and T < 0', because
at the point t = 0, the elements of the original A(t) matrix blow up. This accounts for Y 0(0) = 00.
Admittedly this problem was contrived, and in practice a man~made system would only
accidentally have this form. However, Bessel's equation often occurs in nature, and knowledge
that as t ~ co, J 0 (t) ,..., V2lrrt cos (t  '/T/4) and Yo (t) ,.., V2/'/Tt  sin (t  7r/4) gives great insight
into the behavior of the system.
5.7. Given the timevarying difference equation x(n + 1) = A(n) x(n), where A(n) = Ao
if n is even and A(n) = Al if n is odd. Find the fundamental matrix, analyze by
Floquet theory, and give the conditions for stability if Ao and Al are nonsingular.
From equation (5.50),
AIAOA1' .. AIAo if k is even and m is even
AoAIAO' .. AIAO if k is odd and m is even
.z,(k,m)
=
1 Al AoAI ... AoAl
AoAIAo' .. AOA!
if k is odd and m is odd
if k is even and
For m even, .z,(k, m) P(k, m) (A 1A o)(km)/2, where P(k, m) I if k is even and P(k, m) = =
m is odd
(AoAlI)1I2 if k is odd. For m odd, (J(k, m) = P(k, m)(AoAI)Ckm) 12, where P(k; m) = I if k is
odd and P(k,m) = (AIAOI)1I2 if k is even. For instability, the eigenvalues of R = (AIAo)l/2
must be outside the unit circle. Since the eigenvalues of B2 are the squares of the eigenvalues of B,
it is enough to find the eigenvalues of A1Ao. This agrees with the stability ~nalysis of the equation
x(n + 2) = A1AoX(n).
y =(1 O)x
5.9. In the system of Problem 5.8, let u(t) = eCa  2)t, y(to) = Yo and (dy/dt)(to) = ayo.
Find the complete response.
From equations (5.44) and Problem 5.5,
yet) = eaCtto) cos (eto  et)yo + eat st e7" sin (e r  e t ) dT
to
CHAP. 5] SOLUTIONS TO THE LINEAR STATE EQUATION 121
5~10. Given a step input U(8) = 6/8 into a system with a transfer function
8+1
H(8) = 82 + 58 + 6
Find the output y(t) assuming zero initial conditions.
The easiest way to do this is by using classical techniques.
6(8 + 1) = 1: + _3_ _ _ 4_
..e{y(t)} = U(8) H(8) =
+ 582 + 68
s3 8 8+2 8+3
Taking the inverse Laplace transform determines y = 1 + 3e 2t  4e 3t•
Doing this by state space techniques shows· how it corresponds with the classical techniques.
From Problem 2.3 the state equations are
This integral is usually very complicated to solve analytically, although it is easy for a computer.
Instead, we shall use the transfer function matrix of equation (5.56).
2 1 8+1
= == R(s)
8+3 8+2 82 + 58 + 6
This is indeed our original transfer function, and the integral (5.68) is a convolution and its
Laplace transform is
..e{y(t)}
5.11. Using the adjoint matrix, synthesize a form of control law for use in guidance.
We desire to guide a vehicle to a final state x(tt), which is known. From (5.44),
Choose u(t) = U(t)c where U(t) is a pre specified matrix of time functions that are easily mech
anized, such as polynomials in t. The vector c is constant, except that at intervals of time it is
recomputed as knowledge of x(t) becomes better. Then c can be computed as
However, this involves finding C)(tb t) as the transition matrix of dxldt == A(t)x with x(t) as the
initial condition going to x (tt). Therefore C)(tj, t) would have to be computed at each recomputa
tion of c, starting with the best estimates of x(t). To avoid this, the adjoint transition matrix
'i'(r, tf ) can be found starting with the final time tf' and be stored and used for all recomputations
of c because, from equation (5.64), 'l't(r, tf) = .p(tf , r) and c is found from
c = [f: ~t(T. tf) B(.) U(.) dTJ' [~t(t. 1,) x(t)  x(tf )]
Supplementary Problems
5.12. Prove equations (5.9), (5.10), (5.11), and (5.12).
5.14. Prove that AeAt == eAtA and then find the conditions on A and B such that eAe B == eBeA.
5.15. Verify that 4t(t,7") = eACt  r ) and cp(k, m) == Akm satisfy the properties of a transition matrix
given in equations (5.5)(5.12).
5.18. Calculate the transition matrix .p(t,O) for dxldt == (~ ~)x using (a) reduction to Jordan
5.20. Find eAt using the resolvent matrix and Leverrier's algorithm, where A == (~ ~ ~).
o 03
5.21. Find eAt using the CayleyHamilton method, where A is the matrix given in Problem 5.20.
CHAP. 5] SOLUTIONS TO THE LINEAR STATE EQUATION 123
A (~ ~ ~)
1 1 2
5.23. Use the resolvent matrix and Cramer's rule to find eAt for A as given in Problem 5.22.
5.24. Use the resolvent matrix and Cramer's rule to find Ak for A as given in Problem 5.22.
5.25. Find eAt by using the Maclaurin series, CayleyHamilton and resolvent matrix methods when
A ~ (~~).
5.26. Find the fundamental, or transition, matrix for the system
dx
at (~ u
o
using the matrix Laplace transform method.
Compute yet) using the transition matrix if u is a unit step function. Compare this with the
solution obtained by finding the 1 X 1 transfer function matrix for the input to the output.
Compute yen) using the transition matrix if u is the series of ones 1,1,1, ... ,1, ....
1
5.29. (a) Calculate q:,(t, to) for the system dx/dt == (
0 21)X using Laplace transforms.
5.30. How does the spectral representation for eAt extend to the case where the eigenvalues of A are
not distinct?
ni
5.31. In the CayleyHamilton method of finding eAt, show that the equation eJt ~ Yi(t)Ai can always
i=O
be solved for the Yi(t). For simplicity, consider only the case of distinct eigenvalues.
5.32. Show that the column vectors of *'(t, r) span the vector space of solutions to dx/dt == A(t)x.
5.33. Show A(t) A(T) == A(T) A(t) when A(t) == a:(t)C, where C is a constant n X n matrix and a(t) is a
scalar function of t. Also, find the conditions on ai/t) such that A(t) A(r) == A(r) A(t) for a
2 X 2 A(t) matrix.
5.35. Prove Floquet's theorem for discrete time systems, ~(k, m) = P(k, m)Rkm where P(k, m) =
P(k + w, m)if A(k) = A(k + w).
q,(t, to) and verify it satisfies Floquet's result q,(t, to) = pet, to)eRCtto) where P is periodic and
R is constant. Also find the fundamental matrix of the adjoint system.
5.37. The linear system shown in Fig. 56 is excited by a square wave set) with period 2 and amplitude
= 1. The system equation is ii + [,8 + a sgn (sin 7Tt)]y = O.
Is(t)1
Fig. 56
It is found experimentally that the relationship between a and ,8 that permits a periodic
solution can be plotted as shown in Fig. 57.
~~~~~~_rr_fi
o 2 3 4 5 6
Fig. 57
Find the equation involving a and ,8 so that these lines could be obtained analytically. (Do not
attempt to solve the equations.) Also give the general form of solution for all a and ,8 and mark
the regions of stability and instability on the diagram.
5.38. Given the sampled data system of Fig. 58 where S is a sampler that transmits the value of e(t)
once a second to the hold circuit. Find the state space representation at the sampling instants of
the closed loop system. Use Problem 5.10.
jj
ret) 0
+
.(t) Hold
H 82
8+1
+ 5s + 6
oy(t)
Fig. 58
5.39. Find tt( t, T) and the forced response for the system t2;j + t~ + n :::: p(t) with 'I1(t o) = no and
~(to) = ~o.
CHAP. 5] SOLUTIONS TO THE LINEAR STATE EQUATION 125
(~:) = (~ ~)(::) or y = Cx + Du
Find the transfer functions .,e{Yl}/.,e{ul} and .,e{Y2}/.e{u2} using the relation
"c{y}/.,e{u} = [C(ls  A)IB + DJ
5.41. The steady state response xss(t) of an asymptotically stable linear differential system satisfies the
equation
dxs/dt = A(t)xss + B(t) u(t)
but does not satisfy the initial condition xss(to) = Xo and has no reference to the initial time to.
~(t,T) is known.
(a) Verify, by substitution into the given equation, that
xss(t) = ft~(t'T)B(T)U(T)dT
where f t is the indefinite integral evaluated at T = t. Hint: For an arbitrary vector function
f{t, T),
d
dt
fset) f(t, T) dT = let, {J(t»
dj3
dt 
do:.
f(t, o:.(t» Cit +
f~(t) a
aff(t, T) dt"
a(t) a(t)
(b) Suppose ACt) = ACt + T), B(t) = B(t + T), u(t) = u(t + T), and the system is stable. Find an
expression for a periodic xss(t) = xss(t + T) in the form
5.42. Check that h(t) = et+a(tT) sin (e T  e t ) satisfies the system equation of Problem 5.8 with
u(t)= .sCt  7').
2
5.43. Find in closed form the response of the system (1 _ t 2) d y _ ! dy u to the input u(t) =
dt 2 t dt
tyt 2 ..... 1 with zero initial conditions.
5.44. Consider the scalar system dy/dt = (1 + t)y + (1 + t)u. If the initial condition is yeO) = 10.0,
find the sign and magnitude of the impulse in u(t) required at t 1.0 to make y(2) 1.0. = =
0 3/t2)
5.45. Given the system dx/dt = ( 1 2/t x. Find the relationship between Xl(t) and X2(t) such
5.46. Show that if an n X n nonsingular matrix solution T(t) to the equation dT/dt = A(t)T  TD(t}
is known explicitly, where D(t) is an n X n diagonal matrix, then an explicit solution to dx/dt =
A(t)x is also known.
126 SOLUTIONS TO THE LINEAR STATE EQUATION [CHAP. 5
5.16. A
(2 2)
2 2
t
5.18.
(
e
o
01)
eAt = 0.5. 2t G 1
1
o
o
0) +
o O.5e 4t
( 11
1 1
o o 1  0 0
5.21. eAt (6e2t  8e 3t + 3e 4t)I + 0.5(7 e  2t  12e 3t + 5e 4t )A + O.5( e  2t  2e 3t + e 4t)A2
5.22. eAt
5.23. (sI  A)l F(s)/s(s  1)2 D/s + B/(s  1) + e/(s _1)2 where
D = (~ ~ =~) e
C~ ~ D B = (~ ~ ~)
1 11 1 1 0
5.24. Ak
( k
k1
k 1
k
2k +
2k;1
1)
1 1
5.27. yet) ![ll 2e (tt o)/2  e (tt o)] U(t  to)
5.30. Let the generalized eigenvector tt of A have a reciprocal basis vector s1' etc. Then
eAt = eAtxlrt + eAttlst + teAtxlsT + ...
5.31. A Vandermonde matrix in the eigenvalues results, which is then always invertible.
CHAP. 5] SOLUTIONS TO THE LINEAR STATE EQUATION 127
where T = cos to  cos t, so 4»(t, to) = pet, to) and R O. Also ~t(t, to)
4»(to, t).
41'(2,1) =
COSO (sin 0)/0)
4»(1,0) =
cosy (sin Y)/Y)
( s sin IS cos 8 ( I' sin I' cosy
For periodicity of the envelope zCt + 471/8) = zCt), eigenvalues '}. of e 2R = e:!:.:Jo.
det ('}./  e2R) = '}.2 ~ '}.2 cos (J + 1 = '}.2  '}. tr e2R + det e 2R
det cp(2, 1) det 41(1,0) = 1
2 cos 8 = 2 cos I' cos e;  (1'/8 + S/y) sin y sin S
The solution is of the form .pCt, T) = PCt, T)eR(tT) and the given curves form the stability boundaries
between unstable regions and periodic regions.
Reference: B. Van Der Pol and M. J. O. Strutt, On the Stability of the Solutions of Mathieu's
Equation, Philosophical Magazine, 7th series, vol. V, JanuaryJune 1928, pp. 1838.
5.38. x(k + 1) =
1 + e 2 )/2
( (1 e 3 )/3
e 2 
(5e3  2)/3
1) x(k) +
(3  3e 2 ) r(k)
2  2e 3 6
5.39. 'I](t) '1]0 cos (In t/to) + ~oto sin (In tlto)  ft sin (1n rlt) p(r)/T dT
to
1  10e 4
5.44. u(t) = 2e 5/2 oCt  1.0)
which is equivalent to starting from the zero state and going to a final state (tt) = x
x(t1)  tIt(tt, to) x(to). Therefore if we can show the linear system can go from 0 to any
x(t 1), then it can go from any x(to) to any X(tl)'
The concept of observability will turn out to be the dual of controllability.
Definition 6.2: A state x(t) at some given t of a system is observable if knowledge of the
input u(r) and output y(.) over a finite time segment to < or === t com
pletely determines x(t).
If all states x(t) are observable, the system is called completely observable. If observa
bility depends on to, the state is said to be observable at to. If the state can be determined
for 'T in any arbitrarily small time segment independent of to, it is totally observable.
Finally, we may talk about observability when U(i) = 0, and give similar definitions for
zeroinput observable.
To determine complete observability for linear systems, it is necessary and sufficient to
see if the initial state x(to) of the zeroinput system can be completely determined from
y(r), because knowledge of x(to) and U(i) permits x(t) to be calculated from the complete
solution equation (5.44).
128
CHAP. 6] CONTROLLABILITY AND OBSERVABILITY 129
Example 6.1.
Consider the timevarying system
%t (::) ( ao f3O)(Xl)X2 +
(eat)
eSt U
From the flow diagram of Fig. 61 it can be seen that u(t) passes through scalors that are never zero.
uo+
Fig. 61
xl(tl)eatl = st!
to
U(T) dT = x2(tl)ef3tl
Only those xz(t 1) = x 1(t 1)e tt C,6O:) can be reached at tlo so that X2{t1) is fixed after Xl(t 1) is chosen. There
fore the system is not controllable.
Theorem 6.1: Given a scalar input u(t) to the timeinvariant system dx/dt = Ax + bu,
where A has distinct eigenvalues Ai. Then the system is totally controllable
if and only if the vector f = M1b has no zero elements. M is the modal
matrix with eigenvectors of A as its column vectors.
Proof; Only if part: Change variables by x = Mz. Then the system becomes
dz/ dt = Az + fu, where.A. is the diagonal matrix of distinct eigenvalues Ai_ A flow diagram
of this system is shown in Fig. 62 below.
If any element ii of the f vector is zero, the element of the state vector Zi is disconnected
from the control. Consequently any x made up of a linear combination of the z's involving
Zt will be uncontrollable. Therefore if the system is totally controllable, then all elements
it must be nonzero.
130 CONTROLLABILITY AND OBSERVABILITY [CHAP. 6
U oto
Fig. 62
If part: Now we assume all Ii are nonzero, and from the remarks following Definition
6.1 we need investigate only whether the transformed system can be transferred to an
arbitrary Z(t1) from z(to) = 0, where t1 can be arbitrarily close to to. To do this, note
i = 1,2, .. . ,n (6.1)
It is true, but yet unproven, that if the Ii are nonzero, many different u(.) can transfer
o to Z(tl).
Now we construct a particular U{r) that will always do the job. Prescribe U(T) as
n
U(T) == ~ I'keA~(1'tl) (6.2)
k=l
where the JLk are constants to be chosen. Substituting the construction (6.2) into equation
(6.1) gives
n
== ~ f. JLk( e Ak (tl1'), e At (tC 1'J) for all i (6.3)
k=l t
Proof: Only if part: The system is assumed controllable. The flow diagram for one
l x l Jordan block Lji(Ai) of the transformed system dzldt = Jz + fu is shown in Fig. 63.
The control u is connected to Zl, Z2, ••• , Z!l and Zl only if it is nonzero. It does not matter
if 11,/2, ... and fl1 are zero or not, so that the controllable system requires condition (2)
to hold. Furthermore suppose condition (1) did not hold. Then the bottom rows of two
different Jordan blocks with the same eigenvalues [Lvi(Ai) and L1ji(Ai)] could be written as
dzv/dt AiZv + /,,11
~~.~~
Fig. 63
Consider the particular state having one element equal to fTJzv  fvzn. Then
d(fT/zv f"zTJ)ldt = fT/(AiZv + fvu)  /,,(AiZT/ + /TJu)
= Ai (fTJ Zv  !"Z1j)
Therefore /1/Z,,(t)  /vZTJ(t) = [fTJzv(O)  /vzTJ(O)]d,!t and is independent of the control. We have
found a particular state that is not controllable, so if the system is controllable, condition
(1) must hold.
If part: Again, a control can be constructed in similar manner to equation (6.2) to show
the system is totally controllable if conditions (1) and (2) of Theorem 6.2 hold.
Example 6.2.
To illustrate why condition (1) of Theorem 6.2 is important, consider the system
(~ ~)(:~) + G) u
y = (1 3)(:~)
=
Then .,e{X1} = (8.,e{u} + xlO)/(s  2) and .e{X2} (.e{u} + x2o)/(s  2) so that y{t) =
(XIO  3x20)e2t re
gardless of the action of the control. The input is physically connected to the state and the state is
physically connected to the output, but the output cannot be controlled.
For discretetime systems, an analogous theorem holds.
132 CONTROLLABILITY AND OBSERVABILITY [CHAP. 6
Theorem 6.3: Given a scalar inputu(m) to the timeinvariant system x(m + 1) = Ax + bu(m),
where A is arbitrary. Then the system is completely controllable if and
only if conditions (1) and (2) of Theorem 6.2 hold.
Proof: Only if part: This is analogous to the only if part of Theorem 6.2, in that the
flow diagram shows the control is disconnected from at least one element of the state vector
if condition (2) does not hold, and a particular state vector with an element equal to
f7J zv  fv z7J is uncontrollable if condition (1) does not hold.
If part: Consider the transformed systems z(m + 1) = Jz(m) + fu(m), and for simplicity
assume distinct roots so that J = 4. Then for zero initial condition,
For an nth order system, the desired state can be reached on the nth step because
Zl(n)//I ,.\nl
1
.\n2
1 1 u(O)
Z2(n)//2 ,.\nl
2
,.\n2
2 1 u(1)
(6.5)
•••• ill ....... ill •• ill .... ,. ..
Zn(n)/fn .\nl
n
,.\n2
n 1 u(nl)
Note that a Vandermonde matrix with distinct elements results, and so it is nonsingular.
Therefore we can solve (6.5) for a control sequence u(O),u(1), .. • ,u(nl) to bring the system
to a desired state in n steps if the conditions of the theorem hold.
For discretetime systems with a scalar input, it takes at least n steps to transfer to an
arbitrary desired state. The corresponding control can be found from equation (6.5), called
dead beat control. Since it takes n steps, only complete controllability was stated in the
theorem. We could (but will not) change the definition of total controllability to say that
in the case of discretetime systems, transfer in n steps is total control.
The phenomenon of hidden oscillations in sampled data systems deserves some mention
here. Given a periodic function, such as sin rot, if we sample it at a multiple of its period
it will be undetectable. Referring to Fig. 64, it is impossible to tell from the sample points
whether the dashed straight line or the sine wave is being sampled. This has nothing to
do with controllability or observability, because it represents a failure of the abstract object
(the difference equation) to represent a physical object. In this case, a differentialdiffer
ence equation can be used to represent behavior between sampling instants.
Fig. 64
Theorem 664: Given a scalar output y(t) to the timeinvariant system dx/dt = Ax + Bu,
y = ctx + dTu, where A has distinct eigenvalues Ai. Then the system is
totally observable if and only if the vector gt = ctM has no zero elements.
M is the modal matrix with eigenvectors of A as its column vectors.
Proof: From the remarks following Defini
tion 6.2 we need to see if x(to) can be recon
structed from measurement of y(T) over to < T === t
in the case where U(T) = O. We do this by chang
ing variables as  x = Mz. Then the system be
comes dz/dt = Az and y = ctMz = gtz. The flow
diagram for this system is given in Fig. 6.5.
Each Zi(t) = zi(to)ehiCttO) can be determined
by taking in measurements of y(t) at times Tk =
to + (tl  to)k/n for k = 1,2, ... , n and solving
the set of equations y
11
Y(i k) L g,zi(tO)e"ikCtlto)/n
= i=l t
Theorem 6.7: For the single inputsingle output system dx/dt = Ax + bu, Y = ctx + du,
the transfer function ct(sI ~ A)lb has poles that are canceled by zeros if
and only if some states are uncontrollable and/or unobservable. A similar
statement holds for discretetime systems.
Prool; First, note that the Jordan flow diagram (see Section 2.4) to represent the
transfer function cannot be drawn with repeated eigenvalues associated with different
Jordan blocks. (Try it! The elements belonging to a particular eigenvalue must be com
bined.) Furthermore, if the J matrix has repeated eigenvalues associated with different
Jordan blocks, an immediate cancellation occurs. This can be shown by considering the
bottom rows of two Jordan blocks with identical eigenvalues A.
d
dt
(Zl)
Z2
Then for zero intial conditions ..({Zl} = (s~ A)lb1..({u} and ..({Z2} = (s  A)lb 2..({u}, so that
..({y} [Cl(S  A)lb 1 + C2(S  A)lb 2 + d]..c{u}
Combining terms gives
..c{y}
UOO~"I
I
I
· ·G+Qf~ y
I
I
Fig. 66
Comparing this figure with Fig. 212 shows that fiu i = Pi' the residue of Ai' If and only
if P'i = 0, a cancellation occurs, and Pi = 0 if and only if Ii and/or Ui = 0, which occurs when
the system is uncontrollable and/or unobservable.
Note that it is the uncontrollable and/or unobservable element of the state vector that
is canceled from the transfer function.
Proof; To reach an arbitrary x(tt} from the zero initial state, we must find a control
u(.) such that
x(tI) (6.6)
n
Use of Theorem 4.16 gives eACtl r ) L
i=l
·'/i(.)Ani so that substitution into (6.6) gives
since stl eACt1T)B8(,.  to) dT = eACtlto)B and the defining relation for 8Ck ) is
f_:
to
Using the inversion property of transition nlatrices and the definition of Q, equation (6.8)
can be rewritten
From Problem 3.11, page 63, a solution for the ILi always exists if rankQ = n. Hence SODle
(perhaps not unique) always exists such that the control (6.7) will drive the system to
ILi
X(t1).
The construction for the control (6.7) gives some insight as to why completely control
lable stationary linear systems can be transferred to any desired state as quickly as possible.
No restrictions are put on the magnitude or shape of u(.). If the magnitude of the control
is bounded, the set of states to which the system can· be transferred by t1 are called the
reachable states at t l , which has the dual concept in observabiIity as recoverable states at
t l . Any further discussion of this point is beyond the scope of this text.
136 CONTROLLABILITY AND OBSERVABILITY [CHAP. 6
A proof can be given involving a construction for a bounded u(t) similar to equation
(6.2), instead of the unbounded u(t) of (6.7). However, as tl ...,.. to, any control must become
unbounded to introduce a jump from x(t o) to x(t l ).
The dual theorem to the one just proven is
Theorems 6.8 and 6.9 are also true (replacing totally by completely) for the discretetime
system x(m + 1) = Ax(m) + Bu(m), y(m) = Cx(m) + Du(m). Since the proof of Theorem
6.9 is quite similar to that of Theorem 6.8 in the continuoustime case, we give a proof for
the discretetime case. It is sufficient to see if the initial state x(l) can be reconstructed
from knowledge of y(m) for l ~ m < 00, in the case where u(m) = O. From the state equa
tion,
y(l) = Cx(l)
y(l + 1) ::;: Cx(l + 1) = CAx(l)
u Y
L..._f 1 '_.....
Fig. 6~7
If b(t) = 0 for to == t < to + at, the system is uncontrollable in the time inter~al [to, to + at). If
b(t) = 1 for to + at:::;:: t :::;:: t l , the system is totally controllable in the time interval to + a. t == t ~ t I •
However, the system is not totally controllable, but is completely controllable over the time interval
to== t """ tl to reach a desired final state x(t l )·
CHAP. 6] CONTROLLABILITY AND OBSERVABILITY 137
Now suppose b(t) = 0 for t1 < t === t 2 • and we wish to reach a final state x(t2 ). The state x(t2 ) can
be reached by controlling the system such that the state at time t1 is x(t1) = e(t2 h)x(t2 ). Then with zero
input the free system will coast to the desired statex(t2 ) = <I>(t2 • t 1 ) x(t l ). Therefore if the system is totally
controllable for any time interval te === t === te + fl.t, it is completely controllable for all t:=, te'
For the timevarying system, a criterion analogous to Theorem 6.8 can be formed.
Theorem 6.10: The timevarying system dx/dt = A(t)x + B(t)u is totally controllable if
and only if the nlatrix Q(t) has rank n for times t everywhere dense in
[to, t l ], where Q(t) = (Q1 I Q21 ... [Qn), in which Ql = B(t) and Qk+l =
A(t)Qk + dQk/dt for k = 1,2, ... , nl. Here A(t) and B(t) are assumed
piecewise differentiable at least n  2 and n 1 tinles, respectively.
The phrase "for times t everywhere dense in [to, t l ]" essentially means that there can exist
only isolated points in t in which rank Q < n. Because this concept occurs frequently, we
shall abbreviate it to "Q(t) has rank n(e.d.)".
Proof: First we assume rank Q = n in an interval containing time 'fJ such that to < 'fJ < tl
and show the system is totally controllable.
Construct u( T) as n
U(T) ~ ILk8Ck1)( T  "l) (6.9)
k=l
Jrto I
t1 k 1
n d 
q.(tl' ,) B(,) U(T) dT k~l d,kl [q,(tl, ,) B(T)] '/"=7) ILk (6.10)
But
(6.11)
Note 0 = dl/dt = d(q.q,l)/dt = Acpq,l +cpdcpl/dt so that d~l/dt = cplA and equation
(6.11) becomes
Similarly,
dk  1
dT k  1 [CP(tl' T) B(T)] (6.12)
A solution always exists for the ILk because rank Q = nand cp is nonsingular.
Now assume the system is totally controllable and show rank Q = n. From Problem
6.8, page 142, there exists no constant nvector z =1= 0 such that, for times t everywhere
dense in to";:::: t ..<::: t l ,
ztq.(to, t) B(t) = 0
By differentiating k times with respect to t and using equation (6.12), this becomes
ztq.(to, t) Qk(t) = O. Since q.(to, t) is always nonsingular, there is no nvector yt =
ztq.(to, t) # 0 such that
o= (ytQl J ytQ2 J •.. J ytQn) = ytQ = Y1ql + Y2q2 + . . . + Ynqn
where qi are the row vectors of Q. Since the n row vectors are then linearly independent,
the rank of Q(t) is n(e.d.).
138 CONTROLLABILITY AND OBSERV ABILITY [CHAP. 6
Theorem 6.11: The timevarying system dx/dt = A(t)x + B(t)u, y;:::: C(t)x + D(t)u is totally
observable if and only if the matrix P(t) has rank n(e.d.), where PT(t) =
(Pi I pr I ... I PJ) in which PI = C(t) and P k + I = PkA(t) + dPk/dt for
k = 1,2, ... , nl. Again A(t) and B(t) are assumed piecewise differentiable
at least n  2 and n 1 times, respectively.
Again, the proof is somewhat similar to Theorem 6.10 and will not be given. The situ
ation is somewhat different for the discretetime case, because generalizing the proof follow
ing Theorem 6.9 leads to the criterion rank P = n, where for this case
PI = C(l), P2 = C(l + 1) A(l), ... , Pk = C(l + k  1) A(l + k  2)· .. A(l)
The situation changes somewhat when only complete controllability is required. Since
any system that is totally controllable must be completely controllable, if rank Q(t) has
rank n for some t> to [not rank n(e.d)] then x(t o) can be transferred to X(tl) for tl ~ t.
On the other hand, systems for which rank Q(t) < n for all t might be completely control
lable (but not totally controllable).
Example 6.4.
Given the system
_(ao O)(Xl)
(J X2
+ (fl(t»)u
h(t)
where
2krr ::: t < (2k + 1)rr
= k = 0, ±1, ±2, ...
(2k + l)rr ::: t < 2(k + 1)7T
Q(t)
At each instant of time, one row of Q(t) is zero, so rank Q(t) :::: 1 for all t. However,
If t  to > 7l', the rows of .;.(to, t) B(t) are linearly independent time functions, and from Problem 6.30,
page 145, the system is completely controllable for it  to> 7T. The system is not totally controllable because
for every to. if t2  to < 11, either 11(T) or 12(T) is zero for to::: T::: i 2·
However, for systems with analytic A(t) and B(t), it can be shown that complete con
trollability implies total controllability. Therefore rank Q = n is necessary and sufficient
for complete controllability also. (Note fl(t) and f2(t) are not analytic in Example 6.4.)
For complete controllability in a nonanalytic system with rank Q(t) < n, the rank of
q,(t, 1') B(1') must be_ found.
6.6 DUALITY
In this chapter we have repeatedly used the same kind of proof for observabiIity as was
used for controllability. Kalman first remarked on this duality between observing a
dynamic system and controlling it. He notes the determinant of the W matrix of Problem
6.8, page 142, is analogous to Shannon's definition of information. The dual of the optimal
control problem of Chapter 10 is the Kalman filter. This duality is manifested by the
following two systems:
CHAP. 6] CONTROLLABILITY AND OBSERVABILITY 139
Solved Problems
6.1. Given the system
dx
dt =
(1
o 21) + (0)
lOx 0 u, y (1 1 l)x
1 4 3 1
find the controllable and uncontrollable states and then find the observable and un
observable states.
Following the standard procedure for transforming a matrix to Jordan form gives A = TJTl
as
G~ D
Then f == T 1b == (0 1 O)T and gT
= C~ ~
== cTT = (0 1 1).
D( ~ : DC~ ~ 0
The flow diagram of the Jordan system is
shown in Fig. 68.
Zl
1
o
+ 82
+
Z2
+
o++{ 1 1
U Y
82
+
za
o 1
Fig. 68
The element zl of the state vector z is controllable (through Z2) but unobservable. The element
z2 is both controllable and observable. The element za is uncontrollable but observable.
Note Theorem 6.1 is inapplicable.
140 CONTROLLABILITY AND OBSERVABILITY [CHAP. 6
6.2. Find the elements of the state vector z that are both controllable and observable for
the system of Problem 6.1.
Taking the Laplace transform of the system with zero initial conditions gives the transfer
function. Using the same symbols for original and transformed variables, we have
SXl Xl + 2X2 ~ Xs (6.13)
SX2 (6.14)
(6.15)
(6.16)
From (6.15), Xl = 4x2 + (s  3)xs  u. Putting this in (6.13) gives (48  6)x2 = (8  l)u  (8  2)2 XS •
Substituting this in (6.14), (s  1)(8  2)2XS = (8 1)2U. Then from (6.16),
y
 [(s(8 _1)2
2)3
_ 0 + ~Ju
2)2
(s 


(81)(82)
(8 1)(8  2)2
u
Thus the transfer function k(s) = (8  2)1, and from Theorem 6.7 the only observable and con
trollable element of the state vector is z2 as defined in Problem 6.1.
6.3. Given the system of Problem 6.1. Is it totally observable and totally controllable?
Forming the P and Q matrices of Theorems 6.9 and 6.8 gives
P Q
Then rank P = 2, the dimension of the observable state space; and rank Q = 2, the dimension
of the controllable state space. Hence the system is neither controllable nor observable.
and that u(t) = e t and y(t) =2  (de t • Find X1(t) and X2(t). Find X1(O) and X2(O).
What happens when = O? Q'
Since y = xl' then XI(t) = 2  cdet. Also, dX1/dt = aX2, so differentiating the output gives
x2(t)== e t + tet. Then xl(D) = 2 and x 2(O) = 1. When a = 0, this procedure does not work
because dXl/dt = o. There is no way to find X2(t), because x2 is unobservable as can be verified from
Theorem 6.5. (For a = 0, the system is in Jordan form.)
6.5. The normalized equations of vertical motion y(r, 0, t) for a circular drumhead being
struck by a force u(t) at a point r = ro, 0 = 00 are
iJ2y
at2 V 2y + 27ir 8(r  ro) 8(0  00) u(t) (6.17)
where y(rl, 0, t) = 0 at r = rl, the edge of the drum. Can this force excite all the
modes of vibration of the drum?
The solution for the mode shapes is
00 oc
y(r, 0, t) ::::: ~ ~ I n (K m r/rl)[X2n,m (t) cos 2n1T09 + X2n +l,m (t) sin 2n'lTO]
m=l n=O
CHAP. 6] CONTROLLABILITY AND OBSERVABILITY 141
where ICm is the mth zero of the nth order Bessel function In(r). Substituting the motion of the first
harmonicm == 1, n == 1 into equation (6.17) gives
d2 x 2 ldt 2 ;\X21 + y cos 2;;8ou
d2x3ldt2 ;\x31 + y sin 2;;oou
where ;\ == K~ + (2;;)2 and y == ToJ1(lClrO!rl). Using the controllability criteria, it can be found
that one particular state that is not influenced by u(t) is the first harmonic rotated so that its node
line is at angle 8 0 , This is illustrated in Fig. 69. A noncolinear point of application of another
force is needed to excite, or damp out, this particular uncontrollable mode.
r~~r + Y
Fig. 69
A
G~D and B
G b)
For b = 0 it is equivalent to scalar control, and by condition (I) of Theorem 6.2 the system is
uncontrollable. For case (2),
0 1 0 2
Q
G0
1
1
1
0
1
1
1 D
The first three columns are linearly independent, so Q has rank 3 and the system is controllable.
For case (3),
1 1 1 2
Q :::: o 1 o 1
o 1 o 1
The bottom two rows are identical and so the system is uncontrollable.
142 CONTROLLABILITY AND OBSERVABILITY [CHAP. 6
~~ Gg)x + (~)u
The Q(t) matrix of Theorem 6.10 is
Q(t)
et(l t»)
e t
Then det Q(t) = et(e t + 2t  2). Since et + 2t = 2 only at one instant of time, rank Q(t) = 2(e.d.).
6.8. Show that the timevarying system dx/dt = A(t)x + B(t)u is totally controllable if
and only if the matrix W(t, ,) is postive definite for every, and every t> T, where
Note that this criterion depends on cJt(t, T) and is not as useful as Theorems 6.10 and
6.11. Also note positive definite W is equivalent to linear independence of the rows
of 4>(7,7]) Bh) for T ~ 1) ~ t.
If W(t,7") is positive definite, WI exists. Then choose
u(7") = Bt(r) 4>t(t 1, r) WI(t o, t l ) x(t I )
so that the system is totally controllable if W(t,7") is positive definite. Now suppose the system is
totally controllable and showW(t, '7") is positive definite. First note for any constant vector k,
Therefore the problem is to show W is nonsingular if the system is totally controllable. Suppose
W is singular, to obtain a contradiction. Then there exists a constant nvector z # 0 such that
(z, Wz) = O. Define a continuous, mvector function of time f(t);:::: Bt(t) q;t(to, t)z. But
ftlllf(r)II~ dT
to
stl to
ztcJlo(t o' t) B(t) Bt(t) cJlot(to, t)z dt
t!
o 
S to
ztcJlo(to, t) B(t) u(t) dt (6.18)
In particular, since the system is assumed totally controllable, take u(t) to be the control that
transfers 0 to cJlo(tl' to)z # O. Then
t!
z
J to
4J(to, t) B(t) u(t) dt
Substituting this into equation (6.18) gives 0 = ztz which is impossible for any nonzero z.
Therefore no nonzero z exists for which (z, Wz) ~ 0, so that W must be positive definite.
CHAP. 6] CONTROLLABILITY AND OBSERVABILITY 143
Supplementary Problems
6.9. Consider the bilinear scalar system d~/dt = u(t) ~(t). It is linear in the initial state and in the
control, but not both, so that it is not a linear system and the theorems of this chapter do not apply.
The flow diagram is shown in Fig. 610. Is this system completely controllable according to Defini
tion 6.1?
~(t)
u(t)~
Fig. 61.0
dx
dt (
~ ~ ~)
2 0 2
x + (~) u,
1
y = (2 1 O)x
Determine which states are observable and which are controllable, and check your work by deriving
the transfer function.
6.12. Six identical frictionless gears with inertia I are mounted on shafts as shown in Fig. 611, with
a center crossbar keeping the outer two pairs diametrically opposite each other. A torque u(t)
is the input and a torque yet) is the output. Using the angular position of the two outer gearshafts
as two of the elements in a state vector, show that the system is state uncontrollable but totally
output controllable.
1Ik:~>L..4_ Y (t) +
u(t)
6.13. Given the electronic circuit of Fig. 612 where u(t) can be any voltage (function of time). Under
what conditions on R, Ll and L2 can both i 1(t 1 ) and i 2(t 1 ) be arbitrarily prescribed for tl > to.
given that i 1(to) and i 2 (tO) can be any numbers?
6.15. Find some other construction than equation (6.2) that will transfer a zero initial condition to an
arbitrary z(t 1).
6.18. What are the conditions similar to Theorem 6'.2 for which a twoinput system is totally controllable?
dx
= y = (0 1 l)x
dt
Classify the elements of the state vector z corresponding to the Jordan form into observable/not
observable, controllable/not controllable.
6.21. Using the criterion Q:::: (b I Ab I ... I An1b), develop the result of Theorem 6.1.
sequence u(O), u(l) required to drive the state to the origin in two sample periods (i.e., x(2) = O}.
(a) Is the system observable? (b) Given the observation sequence y(1) = 8, y(2) = 14, find the
initial state x(O).
6.24. Prove Theorem 6.8, constructing a bounded U(T) instead of equation (6.7). Hint: See Problem 6.8.
6.25. Given the multiple inputmultiple output timeinvariant system dx/dt = Ax + Bu, y = Cx + Du,
where y is a kvector and u is an mvector. Find a criterion matrix somewhat similar to the Q
matrix of Theorem 6.8 that assures complete output controllability.
y(3)
u '_,
, Fig. 613
CHAP.6J CONTROLLABILITY AND OBSERV ABILITY 145
(b) If the overall interconnected system in part (a) is observable, show that Sa is observable.
(Note that u(i) and y(i) are vectors.)
(~
dx
= y = (e t e 2t)x
dt
Is this system totally controllable and observable?
6.29. Prove a controllability theorem similar to Theorem 6.10 for the discrete timevarying case.
6.30. Similar to Problem 6.8, show that the timevarying system dx/dt = A(t)x + B(t)u is completely
controllable for t1 > t if and only if the matrix Wet, T) is positive definite for every T and some
finite t > T. Also show this is equivalent to linear independence of the rows of 4t(T, '1]) BC?]} for
some finite 7J > T.
6.31. Prove that the linear timevarying system dxldt = A(t)x, y = C(t)x is totally observable if and
only if M(tll to) is positive definite for all t1 > to, where
6.10. The states belonging to the eigenvalue 2 are unobservable and those belonging to 1 are uncon
trollable. The transfer function is 3(8 + 3)1, showing only the states belonging to 3 are both
controllable and observable.
6.11. One state is observable and controllable, the other is neither observable nor controllable.
Q" = (11)
1 1
and h(s) = 8
2
+1
6.14. MB ¥= 0 and Za ¥= 0
6.18. Two Jordan blocks with the same eigenvalue can be controlled if 111/22  112/21 ¥= 0, the!'s being
the coefficients of Ul and Uz in the last row of the Jordan blocks.
146 CONTROLLABILITY AND OBSERVABILITY [CHAP. 6
6.20. The flow diagram is shown in Fig. 614 where zl and z2 are controllable, Z2 and Za are observable.
u_ ~y
Fig. 614
147
148 CANONICAL FORMS OF THE STATE EQUATION [CHAP. 7
If Re means "real part of" and 1m means "imaginary part of", this is
d (Rez + jImz) o )(Rez + jlmZ)
dt Rez  jImz ReL  jImL Rez  jlmz
ReL Rez  ImL Imz + jReL Imz + jlmL Rez)
( ReL Rez  ImL Imz  jReL Imz  jlmL Rez
By equating real and imaginary parts, the system can be rewritten in the "real" Jordan
form as
ReL ImL)(Rez)
( ImL ReL Imz
Theorem 7.1: If the system (7.1) has a controllability matrix QX(t), and an equivalence
transformation x(t) = T(t) z(t) is made, where T(t) is nonsingular and dif
ferentiable, then the controllability matrix of the transformed system
QZ(t) = Tl(t) QX(t) _ and rank Qz(t) == rank QZ(t).
Proof: The transformed system is dz/dt = Az(t)z + B%(t)u, where AZ = Tl(AxT  dT/dt)
and BZ = TIBx. Since QX = (Qi IQ~ I ... I Q~) and QZ is similarly partitioned, we need to
show Q~:::: Tl Q~ for k = 1,2, ... , n using induction. First Q~ = HZ = TIBx = TIQ:.
Then assuming Q~l = TIQ~_l'
Q% = AzQ~_l + dQ~_/dt
= Tl(AXT  dT/dt)(TIQ~_l) + dTl/dt Q~l + TldQ~_l/dt
Tl(~AxQ%_l + dQ%_tldt) = T1Q%
for k = 2,3, ... , n. Now QZ(t) = Tl(t) QX(t) and since T(t) is nonsingular for any t,
rank QZ(t) = rank QX(t) for all t.
CHAP.7J CANONICAL FORMS OF THE STATE EQUATION 149
Theorem 7.2: If the system (7.1) has an observability matrix PX(t), then an equivalence
transformation x(t) = T(t) z(t) gives PZ(t) = PX(t) T(t) and rank PZ(t) =
. rank PX(t).
The main problem is to keep Tc(t) differentiable and nonsingular everywhere, i.e. for all
values of t. Also, we will find Q:<: such that it hasn  nl zero rows.
·Proof: First assume (7.1) can be transformed by x(t) = Tc(t) z(t) to the form of (7.2).
Using induction, Q~ = (~:) = (<!J} and if Q: = Cf) then for i = 1,2, ... ,
Therefore
(Qo
.
Z1
1 .' '.'
... 0
QZI
n1 Qn1+
o
Z
l 1 •••
... 0
0:<:1).
~
where F(t) is the nl X m(nnl) matrix manufactured by the iteration process (7.4) for
i = nl, nl + 1, ... , nl. Since QZl has rank nl(e.d.), QZ must also. Use of Theorem 7.1 and
the nonsingularity of Tc shows QX has ranknt(e.d.). Furthermore, let Tc(t) (Tl(t) IT2(t»), =
so that .
Tc(t) Qz(t)
(T1QZl TtF)
Since Tl(t) is an n X nl diiferentiablematrix with ranknl everywhere and QZl is an nl x 1nnl
matrix with rank nl(e.d.), the only if part of ,the theorem has been proven:
150 CANONICAL FORMS OF THE STATE EQUATION [CHAP. 7
where Vz(t) is any set of n  nl differentiable columns making Vet) = (VI Vz) nonsingular.
But what is the system corresponding to the controllability matrix on the right? From
Theorem 6.10, (~,) = BZ. Also,
and
Therefore 0 = A~l(t)S(t), and since Set) has ranknl(e.d.) _and A~l(t) is continuous, by
Problem 3.11, A~l(t) = O. Therefore the transformationV(t) = (VI(t) Vz(t» is the required
equivalence transformation.
The dual relationship for observability is x(t) =TO(t) z(t) that transforms (7.1) into
We can extend this procedure to find states WI, Wz, Ws and W4 that are controllable and
observable, uncontrollable and observable, controllable and unobservable, and uncontrollable
and unobservable respectively. The system (7.1) is transformed into
d (WI)
dt Wa ( A~
A~
o
A~
)(WI) + (B~)
Ws B~ u
has a controllability matrix (~:: ~::) of rank n, + n3 ~ n( e.d.) and where the subsystem
d
dt
(Wl)
W2
= y = (7.8)
has an observability matrix (~:: ~:: ) of rank n, + n2 "" n(e.d. ). Hence these subsystems
are totally controllable and totally observable, respectively. Clearly if such a system as
(7.7) can be found, the states WI, W2, Ws, W4 will be as desired because the flow diagram of
(7.7) shows W2 and W4 disconnected from the control and Ws and W4 disconnected from the
output.
Here Hi( t) is a kn x ~ matrix with rank ni( e.d.), Si( t) is an ~ X mn matrix with rank ni( e.d.),
Ui(t) is an ni x n differentiable matrix with rank ~ everywhere, Vi(t) is an n x ~ differentiable
matrix with rank ni everywhere, and Ui(t) Vj(t) = 8ijI~. Furthermore, the rank of Hi and Si
must be such that the controllability and observability matrices of (7.8) have the correct
rank.
Proof: First assume (7.1) can be transformed to (7.7). By reasoning similar to (7.4),
(
~l QI3
o
FI2
0
P I2
P 22
0
0
pz
Q3I Q33 F32 Ga2 0
o o 0 G42 0
so that QX and px have rank ni +na (e.d.) and ni + n2 (e.d.), respectively. Let
Then
QX VI (Qll QI3 F lZ F I4) + Va(Q31 Q33 Fa2 F 34)
and
px (Pi1 piz Gi3 Gi4)TUl + (Pil pJ2 Gi3 G~)TU2
4. Factor px  RIUl = R2U2 and QX  VlSl = Vasa to obtain R2, U 2 , Va and S3.
5. Find the reciprocal basis vectors of U 2 to form V2.
6. Find V4 as any set of n4 differentiable columns making T(t) nonsingular~
Exam~]e 7.1·
GIven P x =
(Sin t 1.)
sin t 1 . Obviously rank PX = 1 and it is factored by inspection, but suppose we
try to mechanize the procedure for the general case by attempting elementary column operations analogous
to Example 3.13, page 44. Then
sin t
( sin t
l)((Sin t)i (sin
1 0
t)2)
(sin t)l
(11 00)
The matrix on the right is perfect for pz, but the transformation matrix is not differentiable at t = i1r for
i = ...  1, 0, 1, ' , ,
However, if a(t) and f3(t) are analytic functions with no common zeros, let
E(t) = a(t)
( f3(t)
f3(t»)
a(t)
Then (a(t) P(t»E(t) = (a 2(t)
+ f32(t) 0) and E(t) is always nonsingularand differentiable.
This gives us a means of attempting a factorization even if not all the elements are analytic.
If a(t) and P(t) are analytic but have common zeros Cl, tz, ... , Ck, ••• , the matrix E(t)
cali be fixed up as
E(t) (;~Z ~l~D Jl (1 tll:k)"Yk(l:k)
where Pk is the order of their common zero Ck and Yk(C k) is a convergence factor.
Example 7.2.
Let a:(t) = t2 and /3(t) = t S , Their common zero is tl =0 with order 2 = Pl' Choose Yt(fl} = rio
Then
E(t)
Theorem 7.6: The system (7.9) can be transformed to (7.10) by an equivalence trans
formation if and only if Qx(t), the controllability matrix of (7.9), is differ
entiable and has rank n everywhere.
Note this implies (7.9) must be more than totally controllable to be put in the form (7.10)
in that rank Q = n everywhere, not n(e.d.). However, using the methods of the previous
section, the totally controllable states can be found to form a subsystem than can be put
in canonical form.
o 0 ... 1 o
This is true because b x = Qxb w = qf and QXl(AxQxdQx/dt) =AW so that AXq:+dqf/dt =
qf+l for i = 1,2, ... , nl. Also AWen = QXl(AXq~  dq~/dt). Setting w = Kz, where
and
o 1 qn2,2 qn1,2
1 qll qn2,1 Qnl, 1
where qii = (l)i an for 1 === i === n; and for 1 === k < i === n,
Theorem 7.7: The system (7.9) can be transformed to (7.12) by an equivalence trans
formation if and only if QX(t), the controllability matrix of (7.9), is differ
entiable and has rank n everywhere.
Example 7.3.
Suppose we have a secondorder system. Then
and
By equating the two expressions it is possible to find, recursively, a2 = a1 and at:= 0:2  dal/dt.
It appears that the conditions of Theorems 7.6 and 7.7 can be relaxed if the hz(t) of
equation (7.10) or (7.12) is left a general vector function of time instead of en. No results
are available at present for this case.
Note that if we are given (7.9), defining y = Zl in (7.12) permits us to find a corre
sponding scalar nth order equation dnyldtn = alY + a2 dy/dt + ... + andnlyldtn1 + u.
For the case where u is a vector instead of a scalar in (7.9) a possible approach is to set
all elements in u except one equal to zero, and if the resulting QX has rank n everywhere
then the methods developed previously are applicable. If this is not possible or desirable,
the form
dw
(ft ==
(o~~ ~~...........~.
..
0 ... hi
(7.13)
may be obtained, where the fi are in general nonzero nvector functions of time, and
CHAP. 7] CANONICAL FORMS OF THE STATE EQUATION 155
A~ (~ ~ =
o
.. .. ::: ...
0 ...
:.~~.')
a~
This form is obtained by calculatingm QX matrices for each of the m columns of the B
matrix, Le. treating the systems as m singleinput systems. Then from any l of these
singleinput QX matrices, choose the first nl columns of the first QX matrix, the first ?t2
columns of the second QX matrix, ... , and the first nz columns of the lth QX matrix such that
these columns are independent and differentiable and nl + n2 + ... + nz = n. In this order
these columns form the T matrix. The proof is similar to that of Theorem 7.6.
To transform from this form to a form analogous to the first canonical form (7.10),
use a constant matrix similar to the K matrix used in the proof of Theorem 7.6. However,
as yet a general theory of reduction of timevarying linear systems has not been developed.
Solved Problems
7.1. Transform
1
x(m+1) 3
1
to Jordan form.
From Example 4.8, page 74, and Problem 4.41, page 97,
J = G 1
2
o
To ~ (
1
~
o
1
o 1
~), I)
Then
(~ 1
~ ~)(~0 ~0 3~)
0 1
:: (=~1 2~ ~)
8
Substitution of x::::; Tz in the system equation then gives the canonical form
1
z(m+ 1) :::: 2
o
156 CANONICAL FORMS OF THE STATE EQUATION [CHAP. 7
7.2. Transform
From Problem 4.8, page 93, and Problem 4.41, page 97,
= 1 ~ j)(a o
~ )z
~.5 ~ ~
x 13
(
0 o {3*
dz/dt 1j
o
0) + (2/a)
o z (1  j)/f3 u
o 1 +j (1 + j)/f3*
Then a = 2 and 13 = 1  j. Putting this in real form. give$
Observe that Q = (b IAb I A 2b) in accord with using the form of Theorem 6.10 in the
time~invariantcase, and not Q = (b 1Ab 1A 2b) which is the form of Theorem 6.8. Performing
elementary row operations to make the bottom row of QZ = 0 gives
px = ( ~ : 0= DG ~ D ( ~ ;
where the factorization is made using elementary column transformations. Using the trans
formation x = TOz,
dzldt =
(1 3 0) (1)
0 2
2 2
0
1
z + 0
2
u y (1 2 O)z
Then
pZV1 = (~
1
2
4
2)('') (0
1
5
1'21
1'31
= = gl gives VI . = C~V31) 1'31
1'31
and
~
1
=
D
U1Qa: = (un U12 ud ( 0 (1 1 1) 8 1 gives U1 (1 U13  1 U13)
2 0
Note UI V1 = 1 for any values of UtB and 1'31' The usual procedure would be to pick a value
for each, but for instructional purposes we retain U13 and val arbitrary. Then
and
~ ~)1
T =
G~ D G o 1
158 CANONICAL FORMS OF THE STATE EQUATION [CHAP; 7
y = (1 1 O)z
(~
5 162t
2 2 6t[40)
0 0
0 0
(~
t2 0
px =
t2 + 2t 0 St21 t )
t2  4t+ 2 0 St + 2t2  12t  1
S
t2t2+ 2t )
81 (2 2 2 2)
( t2  4t +2
t2 + 6t  6
(0 1 0 0)
U2 = (0 0 0 1)
24t2
 1
15t  18
)
Then vi = (0 0 0 1) and Vr =(0 0 1 0) and T:::::: (VI I V2 1Vsl v4). It is interesting to note
that the equivalence transformation
T=O~~D
will also put the system in the form (7.7).
CHAP. 7] CANONICAL FORMS OF THE STATE EQUATION 159
dx/dt =
( ~
sin t
t) x + ( cost t ) u
t cost
E(t)QX :::: t
( cos t
cos
t
t)( t
cos t
t cos t)
cos2 t
=
T = El(t)
t2
1
+ cos2 t
(t
cos t
ctos t)
From part (i) of Problem 7.3, the system can be put in the form
1 0
z(m+ 1) 2 1
(
o 0
T
160 CANONICAL FORMS OF THE STATE EQUATION [CHAP. 7
(1o 1)1(
2
1 0)(1)2
2 1
Then T ( 1 10)
2
from which we obtain the phasevariable canonical form
Zp(m+ 1)
By chance, this also happens to be in the form of the first canonical form.
(i) into first canonical form, (ii) into second canonical form, and (iii) find a scalar
secondorder equation with the given state space representation.
To obtain the first canonical form,
where
1 ( 8  4t  2t2 + (t 1) sin t  cos t )
6 + 2t  sin t 6  6t  2t2  (t + 6) sin t  3 cos t + sin2 t
T ==
2t + sin
3
t) (0
1
1
a1(t)
)1 (
2t  2a1  sin t
al +3
from which
where y == Zpl.
1 21) (0 1)
dx/dt
(14 3 1 1
010 x+ 01 u
To obtain the form (7.13), we calculate the two QX matrices resulting from each of the columns
of the Bx matrix separately:
CHAP.7J CANONICAL FORMS OF THE STATE EQUATION 161
QX (u1 only)
Note both of these matrices are singular. However, choosing T as the first two columns of
QX(U1 only) and the first column of Q% (U2 only) gives
so that
(~
4
dw/dt 4
o
Also, T could have been chosen as the first column of Q% (u1 only) and the first two columns of
Q:t (u2 only).
To transform to a form analogous to the first canonical form, let w ==Kz where
Then
K ;:::
G D
1
0
(~ Dz + G ~)u
0
dz/dt = 4
4
Supplementary Problems
(i) to a controllable system, (ii) to an observable system, (iii) to an observable and controllable
system.
7.14. Show that the factorization requirement on Theorem 7.3 can be dropped if T(t) can be nonsingular
and differentiable for times t everywhere dense in [to, tiJ.
162 CANONICAL FORMS OF THE STATE EQUATION [CHAP. 7
(~
6 2
dx/dt
2
8
4
8
4
6
!)
6
2
x y (1 1 1 l)x
7.17. Check that the transformation of Problem 7.5, page 159, puts the system in the form of equation
(7.2), page 149, by calculating AZ and B2.
7.19. Develop the transformation to a form similar to equation (7.11) for timevarying discretetime
systems.
7.21. Given the timeinvariant system dx/dt = Ax + enu where the system is in phasevariable canonical
form as given by equation (7.12). Let z = Tx where z is in the Jordan canonical form dz/dt =
.A.z + bu and.A. is a diagonal matrix. Show that T is the Vandermonde matrix of eigenvalues.
7.22. Verify the relationship for the qik in terms of the a" following equation (7.12) for a thirdorder
system.
7.23. Solve for the at in terms of the Cl'i for i = 1,2,:3 (thirdorder system) in a manner analogous to
Example 7.3, page 154.
11 4
dx
7.24. Transform the system
dt = 1/2 1
(
27/2 6
to phasevariable canonical form.
7.26. Using the results of Section 7.5, find the transformation x = Tw that puts the system
1
dx/dt 2
3
3
into the form dw/dt 1
1
1.27. Obtain explicit formulas to go to phasevariable canonical form directly in the case of timeinvariant
systems.
CHAP.7J CANONICAL FORMS OF THE STATE EQUATION 163
7.28. Use the duality principle to :find a transformation that puts the system dx/dt == A(t)x and y::::: C(t)x
into the form
dz y
::::: (0 0 '" 0 l)z
dt
7.29. Prove that IjTl1 < 00 for the transformation to phasevariable canonical form.
C~ 2D
1
7.9. T 0 where f3 is any number # O.
0
7.10. T :::::
(: 8)12
7.11. There is one controllable and observable state, one controllable and unobservable state, and one
uncontrollable and observable state.
7.15. No. QZ::::: V1(1 0) but VI does not have rank one everywhere.
puts the system into the form. of equation (7.6), for any r4i that make T nonsingular. Also, Jordan
form can be used hut is more difficult algebraically.
3/2 1 1)
7.24. Tl
( 5/2 12
1 1 1
7.25. A
7.26. T = G~ D
7.28. This form is obtained using the same transformation that puts the system dw/dt::::: At(t)w + ct(t)u
into phasevariable canonical form.
7.29. The elements of Qzl are a linear combination of the elements of QZ, which are always finite as
determined by the recursion relation.
Chapter 8
y  (3 2)x
x(t)
u(t)====>i 1"'" y(t)
Fig. 81
Also, flow diagrams of transfer functions (block diagrams) can be drawn in a similar
manner for timeinvariant systems. We denote the Laplace transform of x(t) as ..c{x}, etc.
164
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 165
Example 8.2.
The block diagram of the system considered in Example 8.1 is shown in Fig. 82.
.e{u}...J loC{Y}
Fig. 82
Vector block diagram manipulations are similar to the scalar case, and are as useful
to the system designer. Keeping the system representation in matrix form is often helpful,
especially when analyzing multiple inputmultiple output devices.
where alP ABC(t, T)/at = [A(t)  B(t)C(t)]cI» ABC (t, T) and lP ABC(t, t) = I. Further, we shall
be concerned only with inputs d(t) that do not drive Ily(t)11 to infinity before t = co, so that
we obtain a steady state as t tends to infinity.
d(t)==~ 1:==:===> y( t)
+ Zero output is
asymptotically stable
Fig. 84. Unity Feedback System with Asymptotically Stable Zero Output
166 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
Theorem 8.1: For the system of Fig. 84, lim e(t) = 0 if and only if d = C(t)w + g
toQ
where dw/dt == A(t)w + B(t)g for all t:::=:, to in which g(t) is any function
such that lim g(t) = 0 and A, B, C are unique up to a transformation
t::¢
onw.
Proof: Consider two arbitrary functions f(t) and h(t) whose limits may not exist as t
tends to co. If lim [f(t)  h(t)] = 0, then f(t)  h(t) = r(t) for all t, where r(t) is an
t_'"'
arbitrary function such that lim r(t) = O. From this, if 0 = lim e(t) = lim [d(t)  y(t)],
then for all t, t_oo toQ too
Jrtot C(t) q, ABC(t, or) B(T) d(T) dT + g(t) + C(t).p ABC (t, to) w(to)
=
where the change of variables g(t) = r(t) + C(t)q,ABC(t, to)[x(to)  w(to)] is onetoone for
arbitrary constant w(to) because lim C(t)cI»ABC (t, to) = O. This Volterra integral equation
too
for d(t) is equivalent to the differential equations dw/dt = [A(t)  B(t) C(t)]w + B(t)d and
d = C(t)w + g. Substituting the latter equation into the former gives the set of equations
that generate any d(t) such that lim e(t) = O.
t_oo
Conversely, from Fig. 84, dx/dt = A(t)x + B(t)e = [A(t)  B(t) C(t)]x + B(t)d. Assuming
d = C(t)w + g and subtracting dw/dt = A(t)w + B(t)g gives
d(xw)/dt = [A(t)  B(t) C(t)](xw)
Then lim e lim (d  y) = lim [g  C(t)(x  w»)
tco too tco
From the last part of the proof we see that e(t) = g(t)  C(t) CPABe(t, to)[x(to)  w(to)]
regardless of what the function g(t) is. Therefore, the system dw/dt = A(t)w + B(t)g with
d = C(t)w + g and the system dx/dt = [A(t)  B(t) C(t)]x + B(t)d with e = d  C(t)x are
inverse systems. Another way to see this is that in the timeinvariant case we have the
transfer function matrix of the open loop system H(s) = C(sI  A)lB relating e to y. Then
for zero initial conditions, ..e{d} = [H(s) + I]..c{g} and .,e{e} = [H(s) + I]l..c{d} so that
.,e{g} = .,e{e}. Consequently the case where g(t) is a constant vector forms a sort of bound
ary between functions that grow with time and those that decay. Of course this neglects
those functions (like sin t) that oscillate, for which we can also use Theorem 8.1.
Furthermore, the effect of nonzero initial conditions w(to) can be incorporated into g(t).
Since we are interested in only the output characteristics of the plant, we need concern our
selves only with observable states. Also, because uncontrollable but observable states of the
plant must tend to zero by the assumed asymptotic stability of the closed loop system, we
need concern ourselves only with states that are both observable and controllable. Use of
equation (6.9) shows that the response due to any Wi(tO) is identical to the response due to
an input made up of delta functions and derivatives of delta functions. These are certainly
included in the class of all g(t) such that lim g(t)
t""
O. =
Since the case g(t) = constant forms a sort of boundary between increasing and de
creasing functions, and since we can incorporate initial conditions into this class, we may
take g(t) as the unit vectors to give an indication of the kind of input the system can follow
with zero error. In other words, consider inputs
for i = 1,2, .. . ,m
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 167
C(t)
J to rt 4tA(t,T)B(T)(e Ie 1... lem)dT
1 2 =
_ Inputs of this form give unity error, and probably inputs that go to infinity any little bit
slower than this will give zero error.
Example 8.3.
Consider the system of Example 8.1 in which e(t) = U2(t) and there is no input u 1(t). The zero input,
where lim get) = 0, generate the class of inputs d(t) that the system can follow with zero error.
too
Solving this system of equations gives
For get) = 0, w~ see that the system can follow arbitrary steps and ramps with zero error, which is in
agreement with the classical conclusion that the system is of type 2. Also, evaluating
ito
[3(t  T) + 2J dT = 1.5t2 + 2t
shows the system will follow t2 with constant error and will probably follow with zero error any function
t 2 r: for any e > O. Thisis in fact the case, as can be found by taking get) = te.
Now if we consider the system of Example 8.1 in which e(t) ~ Ul(t) and there is no input u2(t),
then the closed loop system is ~~::::: [( ~ ~)  (~) (3 2)] X. The output of this system is
y ::::: O.5x2(O) + [3Xl(O) + 1.5x2 (O)]e 6t which does not tend to zero asymptotically so that Theorem 8.1 cannot
be used.
Definition 8.1: The system of Fig. 84 is called a typel system (l = 1,2, ... ) when
lim e(t) = 0 for the inputs dt = (t  to)llU(t  to)ei for all i = 1,2, ... , m.
t 00
In the definition, U(t  to) is the unit step function starting at t = to and ei is the ith unit
vector. All systems that do not satisfy Definition 8.1 will be called typeO systems.
Use of Theorem 8.1 involves calculation of the transition matrix and integration of the
superposition integral. For classical scalar typel systems the utility of Definition 8.1 is
that the designer can simply observe the power of s in the denominator of the plant transfer
function and know exactly what kind of input the closed loop system will follow. The fol
lowing theorem is the extension of this, but is applicable only to timeinvariant systems with
the plant transfer function matrix H(s) = C(sI  A)lB.
Theorem 8.2: The timeinvariant system of Fig. 84 is of type l ~ 1 if and only if
H(s) = sIR(s) + P(s) where R(s) and P(s) are any matrices such that
lim SR1(S) = 0 and lilim 81 1P(s)11 < 00.
so 80
168 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
Proof: From Theorem 8.1, the system is of type 1 if and only if ..e{(d1 1 d 2 ! ... I dm)} =
(ll) IsII = [H(s) +1]G(s) where .,e{gi}, the columns of G(s), are the Laplace transforms
of any functions gi(t) such that 0 = lim gi(t) = lim S~{gi} where S.,e{gi} is analytic for
Re s ~ O. teo 80
First, assume H(s) = sIR(s) + P(s) where lim SRl(S) = 0 so that nl(s) exists in a
neighborhood of s = O. Choose 80
... JI (~+~
sa 1~)
8 2
..e{d} 7> ~{y}
+ ~
1 + 128 + 382 0
82

Fig. 85
The system sho~ in Fig. 85 has a plant transfer function matrix H(s) that can be written in the form
H(s) =
_1
82
(681 + 1 9 1)
0 + ( 0
128 1 +3 1 )
0
= s2R(s) + pes)
in which
II;~ sP(s) I I (
0
 lim 00
 SfoO 12 + 38
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 169
and where
= l' (0
1m s
so 1
# 0
Since lim SRl(S) has a nonzero element, the system is not of type 2 as appears to be the case upon
80
first inspection. Rewrite H(s) in the form (where R(s) and P(s) are different)
H(s)
!
s
(6S81++ 1 _~1) + (03 0
2
9s
12
1) sIR(s) + P(s)
lim sRl(s)
so
= lim
so
8( 0 1 +\28)
9s  6
;;:::;
(~ ~)
8
1 + 128
Since the closed loop system has poles at 1, 1, 0.5, and 0.5, the zero output is asymptotically stable,
Therefore the system is of type 1.
To find the error constant matrix of a typel system, we use block diagram manipula
tions on Fig. 84 to get ~{e} = [1+H(s)]l.,e{d}. If it exists, then
lim e(t) = lim s[1 + Sl R(s) + P(S)]l~{ d}
t+oo s+O
for any l> O. Then an error constant matrix table can be formed for timeinvariant sys
tems of Fig. 84.
Steady State Error Constant Matrices
System Type Step Input Ramp Input Parabolic Input
0 lim [I + H(S)]l * *
s+o
1 0 lim Rl(8) *
SIO
2 0 0 lim &1(8)
sO
In the table * means the system cannot follow all such inputs.
Example 8.5.
The type1 system of Example 8.4 has an error constant matrix ;~ Rl(s) = (~ _~). Thus if
the input were (t  to)U(t  t o)e2J the steady state output would be [(t  to)U(t  to) + 6]e2' The system
can follow with zero steady state error an input of the form (t  to)U(t  t o)el'
Then the transfer function from y to u is H(s) = C(sI  A)IB + D, with poles determined
by det (sI  A) = O. Note for the multiple input and output case these are the poles of the
whole system. The eigenvalues of A determine the time behavior of all the outputs.
170 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP; 8
We shall consider the case where equations (8.1) represent the closed loop system.
Suppose that the characteristic equation det (sI  A) = 0 is linear in some parameter K so
that it can be written as
SU + (8 1 + o/lK)snl + (8 2 + o/2K)sn2 + ... + (8 n 1+ o/n_1K)s + 8n + o/nK = 0
d(t) ==::::::::::\
+
Fig. 86
Fig. 87
Example 8.7.
The feedback system of Fig. 88 has an unknown (8 + 3) sinha y
parameter a. Find the effect of variations in a upon 82 + 38 + sinh a
the closed loop roots.'
Let sinh a: = K. The usual procedure is to set
the open loop transfer function equal to 1 to find
the closed loop poles of a unity feedback system. Fig. 88
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 171
(8 + 8)"
1 = 82 + 3s + Ie
This can be rearranged to form the characteristic equation of the closed loop system, 82 + 38 + +
Ie
(8 + 3)IC ::.: O. Further rearrangement gives the standard root locus form under K variation.
8+4
1
K 8(8+ 3)
This happens to give the same root locus as in the previous example for sinh IX === O.
Then GUw)H(jw) = ctUwI  A)lb + d. Usually a choice of state variable x can be found
such that the gain or parameter variation K of interest can be incorporated into the c vector
only. Digital computer computation of {iwI  A)lb as w varies can be most easily done by
iterative techniques, such as GaussSeidel. Each succeeding evaluation of (jwi+lI  A)lb
can be started with the initial condition (jwl A)lb, which usually gives fast convergence.
Example 8.8.
Given the system shown in Fig. 810. The state space form of this is, in phasevariable canonical
form for the transfer function from e to 'I), '
dx
'U = (K O)x
dt
Ie
Then ct(j6lI  A)lb giving the polar plot of Fig. 811.
jw(jw + 1) •
ImGH
.e{e}
ReGH
+
.c{v} 1
8+1
Fig. 8~lO
About the only advantage of this over standard techniques is that it is easily mechanized
for a computer program. For multiple·loop or multipleinput systems, matrix block diagram
manipulations give such a computer routine even more flexibility.
RELATIONS WITH CLASSICAL TECHNIQUES (CHAP. 8
Example 8.9.
Given the 2 input  2 output system with block diagram
shown in Fig. 812.
. Then dx/dt :::: Ax + hIe! +b2 e2 and VI == and V2, == elx +
ctx. The loop connecting VI and el can be closed, so that +
61 ::: VI :::ct
x. Then ..e{v}
.c{V2} = el (81  A  bi et) lb2.,e{ e2}
50 that we can ask the computer to give us the polar plot of
cd (jwI  A  bIer )lb2 • Fig. 8.. 12
Each ai(t) is a time function to be chosen. This gives the closed loop system
o i 0 0
o 0 1 0
dz
== z
dt
o 1 o o
a 1 (t) a2(t) ocs(t) an(t)
n V
Then Zl(t) obeys z~n) + an(t)zi  + '" + a2(t)zl + a1(t)Zl = 0 and each Zi+l(t) for
i = 1,2, ... , n  1 is the ith derivative of Zl(t). Since the a/t) are to be chosen, the corre
sponding closed loop transition matrix q,z(t, to) can be shaped accordingly. Note, however,
thatx(t) =T(t)4tAt, to)zo so that shaping of the transition matrix iPz(t, to) must be done
keeping in mind, the effect of T(t).
This minor complication disappears when dealing with timeinvariant systems. Then
T(t) is constant, and furthermore each ai(t) is constant. In this case the time behavior of
x(t) and z(t) is essentially the same, in that both AX and AZ have the same characteristic
equation An + an ",n1 + ... + a 2 A+ a l = O.. For the closed loop system to have poles at the
desired values YI' 12' •.. , Yn, comparison of coefficients of the A.'s in (AYl)(A.a2)· •• (A.an) = 0
determines the desired value of each Q t••
Example 8.10;
~~ (~ ~ ~)" + \1(~)
Given the system
== u
2 3 t
It is desired to have a timeinvariant closed loop system with poles at 0, ~1 and 2. Then the desired system
will have a characteristic equation A3 + 3A2 + 2A := O. Therefore we choose u:= (2 1 (t + 3»)x, so
kT == (2 1 (t + 3»).
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 173
For multipleinput systems, the system is transformed to the form of equation (7.13),
except that thesul?system dw/dt = A~Wi + bf'ui must be in phasevariable canonical form
(7.12) and for i ¥= i, the A{j(t) must be all zeros except for the bottom row. Procedures
similar to those used in Chapter 7 can usually attain this form, although general conditions
are not presently available. If this form can be attained, each control is chosen as
ui = kt (t)Wi  e!Aijwj for j ¥= i to "place the poles" of Aii(t) and to subtract off the coupling
terms~
Why bother to transform to canonical form when trial and error can determine k?
Example 8.11.
Place the poles of the system of Problem 7.8 at PI' pz and Pa. We calculate
A1
det
[(
_~_
This is
A3  (k 13 + k21 + k22 + kZ3 + 5)A2 + [ku + 2k 13 + 3kz1 + 4k22 + 5k23 + k 13(k 21 + k 2Z )  kZ3(kn + k 1Z ) + 8]A
 ku  k13  2k21  4k22  6k 23  k n (k 22 + k 23 ) + k 12 (k 21 + k Z3 ) + k 13 (k 21  k 2Z )  4
It would take much trial and error to choose the k's to match
(A  Pl)(A  PZ)(A  Pa) = A3  (PI + pz + Pa)A2 + (PIP2 + PzPa + PIPa)A  PIPzPa
Trial and error is usually no good, because the algebra is nonlinear and increases greatly
with the order of the system. Also, Theorem 7.7 tells us when it is possible to "place the,
poles", namely when Q(t) has rank n everywhere. Transfor!llation to canonical form seems
the best method, as it can be programmed on a computer.
State feedback pole placement has a number of possible defects: (1) The solution appears
after transformation to canonical form, with no opportunity for obtaining an engineering
feeling for the system. (2) The compensation is in the feedback loop, and experience has
shown that cascade compensation is usually better. (3) All the state variables must be
available for measurement. (4) The closed loop system may be quite sensitive to small
variation in plant parameters. Despite these defects state feedback pole placement may
lead to a very good system. Furthermore, it can be used for very highorder and/or time
varying systems for which any compensation may be quite difficult to find. Perhaps the
best approach is to try it and then test the system, especially for sensitivity.
Example 8.12.
Suppose that the system of Example 8.10 had t  e instead of t in the lower right hand corner of the
A(t) matrix, where e is a small positive constant. Then the closed loop system has a characteristic equation
AS + 3AZ + 2A  e = 0, which has an unstable root. Therefore this system is extremely sensitive.
To reconstruct all the states at all times, we assume the physical system to be observed
is totally observable. For simplicity, at first only singleoutput systems will be considered.
We wish to estimate the state of dx/dt = A(t)x + B(t)u, where the output y = et(t)x. The
state, as usual, is denoted x(t), and here we denote the estimate of the state as x(t).
First, consider an observer system of dimension n. The observer system is constructed as
dx/dt == A(t)x + k(t)[et(t)x  y] + B(t)u (8.3)
where k(t) is an nvector to be chosen. Then the observer system can be incorporated into
the flow diagram as shown in Fig. 813.
Since the initial state x(to.), "where to is the time the observer system is started, is not
known, we choose x(to) == O. Then we can investigate the conditions under which x(t) tends
to x(t). Define the error e(t) == x(t)  (t). Thenx
de/dt = dx/dt  di/dt = [A(t) + k(t)ct(t)]e (8.4)
Similar to the method of the previous Section 8.6, k(t) can be chosen to "place the poles"
of the error equation (8 ..4). By duality, the closed loop transition matrix 1j"(t, to) of the
adjoint equation dp/dt = At(t)p  c(t)v is shaped using v == kt(t)p. Then. the transition
matrix fIJ(t, to) of equation (8 ..4.) is found as fIJ(t, to) = iJt(to' t), from equation (5.64). For
timeinvariant systems, it is simpler to consider dw/dt = Atw + cv rather than the
adjoint. This is because the matrix At + ckt and the matrix A + ke t have the same
eigenvalues. This is easily proved by noting that if A is an eigenvalue of At + ck t , its com
plex conj ugate A* is also. Then A* satisfies the characteristic equation det (A *1  At  ckt) = O.
Taking the complex conjugate of this equation and realizing the determinant is invariant
under matrix transposition completes the proof. Hence the poles of equations (8.3) and (8.4)
can be placed where desired. Consequently the error e(t) can be made to decay "as quickly
as desired, and the state of the observer system tends to the state of the physical system.
However, as is indicated in Problem 8.3, we do not want to make the error tend to zero
too quickly in a practical system.
Example 8.13.
Given the physical system
dx y = (1 l)x
dt
Construct an observer system such that the error decays with poles at ::2 and 3.
(3 0)
2 2 x +
A (1)
0 Y +0 u
(1)
N ow we consider an observer system of dimension less than n. In the case of a single~
output system we only need to estimate n  1 elements of the state vector because the known
output and the n 1 estimated elements will usually give an estimate of the nth element of
the state vector. In general for a system having k independent outputs we shall construct
an observer system of dimension n  k.
We choose P(t) to be certain n  k differentiable rows such that the n x n matrix
P(t)
C(t)
)1 = (H(t) I G(t)) exists at all times where H has n  k columns. The estimate x
(
is constructed as
x(t) = H(t)w + G(t)y or, equivalently, (P(t) ) x=( w) (8.5)
C(t) y
Analogous to equation (8.3), we require
P d~/dt = P[Ax + L(Cx  y) + Bu}
where L(t) is an n x k matrix to be found. (It turns out we only need to find PL, not L.)
This is equivalent to constructing the following system to generate w, from equation (8.5),
dwldt = (dPldt) x+ Pdx/dt = Fw  PLy + PBu (8.6)
(:r 1 (1
PI  P2 1
P2)
PI
(H I G)
176 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
Therefore
A
X = 1 (1)
Pi  P2 1
w + 1 (0)
Pi  P2 1
Y
and
F == (dP/dt + PA)H = (P, P.l (2 1)( 1) =
PI  P2 2 2 1
3PI + 4P2
PI  P2
so that (PI  P2)dw/dt = (BPI + 4p2)W  (pi  2p~)y  PI(Pl P2)U is the first~order observer. A bad
choice of PI/P2 with 1 < PI/P2 < 4/3 gives an unstable observer and makes the error blow up.
The question is, can we place the poles of F by proper selection of P in a manner similar
to that of the n.dimensional observer? One method is to use trial and error, which is some
times more rapid for loworder, timeinvariant systems. However, to show that the poles
of F can be placed arbitrarily, we use the transformation x = Tz to obtain the canonical
form
y (~~ ~~ .. ........... ~ ) Z
o 0 ." ct
where the subsystem dzi/dt = Aiizi + Biu and Yi = ct Zi is in the dual phase variable canon
ical form
i = 1,2, ... , l
(0 o l)Zi (8.7)
in which B, is defined from TI B = (~:) and n, is the dimension of the ith subsystem.
As per the remarks following Example 8.10, the conditions under which this form can
always be obtained are not known at present for the timevarying case, and. an algorithm
is not available for the timeinvariant multipleoutput case.
However, assuming the subsystem (8.7) can be obtained, we construct" the observer
equation (8.6) for the subsystem (8.7) by the choice of Pi = (I Ik i ) where k(t) is an (14 1)
vector that will set the poles of the observer. We assume ki(t) is differentiable. Then
0 0 0 ki1(t)
1 0 0 ki2(t)
Fi = 0 1 0 ki3(t)
.........................
o 0 1 k i• ni  (t) 1
Then
'_
A
X =
 '(A)
T.z = T i,
A
Zl
•
where
A
Zi=
Example 8.15.
Again, consider the system of Example 8.13.
To construct an observer system with a pole at 2, use the transformation x = Tz where (Tt)l =
dz
dt (02)
1 4
z + (4) u
1
y  (0 1)z
A
Z
(H+)(;)
where kl = 2 sets the pole of the observer at 2. Then F = 2 and PL = 2 so that the observer
system is dw/dt = 2w + 2y  2u. Therefore
T~
A
X = =
and the error PT1x  w = 2Xl + x2  w decays with a time constant of 1/2. This gives the block diagram
of Fig. 815.
u 8+4 Y
82 + 48 + 2 t~
1
8+2
Fig. 815
178 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
and a feedback control u(t) that has been formed to place the poles of the closed loop sys
tem as
u = W(t)x (8.11)
G J=:==:.J
u
Theorem 8.3: (Algebraic Separation). For the system (8.9) with observer (8.10) and feed
back control (8.11), the characteristic equation of the closed loop system can
be factored as det (AI  A  BW) det (AI  F).
This means we can set the poles of the closed loop system by choosing W using the pole place
ment techniques of Section 8.6 .and by choosing P using the techniques of Section 8.7.
Proof: The equations governing the closed loop system are obtained by substituting
equation (8.11) into equations (8.9) and (8.10):
Note that the bottom equation de/dt = Fe generates an input WHe to the closed loop
of observer system dx/dt = (A + BW)x. Use of Problem 3.5 then shows the characteristic
equation factors as hypothesized. Furthermore, the observer dynamics are in general
observable at the output (through coupling with x) but are uncontrollable by d and hence
cancel out of the closed loop transfer function.
Example 8.16.
For the system of Example 8.13, construct a onedimensional observer system with a pole at 2 to
generate a feedback that places both the system poles at 1.
We employ the algebraic separation theorem to separately consider the system pole placement and
the observer pole placement. To place the pole of
dx
dt y == (1 l)x
y
8 +4 I    _ _ _
82 +48+2
1 +
8+2
Fig.S~17
Note that the control is still essentiaIIy in the feedback loop and that no reasons were given as to
why plant poles at 1 and observer pole at 2 were selected. However, the procedure works for high
order, multiple input multiple output, timevarying systems.
In this section we consider these topics in a very cursory manner" mainly to show the
relationship with controllability and observability. Consequently we consider only: small
percentage changes in parameter variations, small noise compared, withthe'signal, and
nonlinearities that are almost linear. Under these assumptions we will show how each
effect produces an unwanted input into a linear system and then how to minimize this un
wanted input.
First we consider the effect of parameter variations. Let the subscriptN refer to the
nominal values and the subscript a refer to actual values. Then the nominal system',(the
system with zero parameter variations) can be represented by
dXN/dt = AN(t)xN + B(t)u
YN = C(t)XN + D(t)u (8.12)
These equations determine XN(t) and YN(t), so these quantities are assumed known. If some
of the elements of AN drift to some actual Aa (keeping B, C and D fixed only for simplicity),
then
dXa/dt Aa(t)Xa + B(t)u
Ya = C(t)Xa + D(t)u (8.13)
Then let ax = Xa  XN, SA = Aa  AN, 8y = Ya  YN, subtract
equations (8.12) from (8.13),
and neglect the product of small quantities BA Bx. Warning: That SA 8x is truly small
at all times must be verified by simulation. If this is so, then
d(8x)/dt == AN(t) 8x + BA(t) XN
8y == C(t) 8x (8.14)
In these equations AN(t), C(t) and XN(t) are known and 8A(t), theyariation of the parameters
of the A(t) matrix, is the input to drive the unwanted signal 8x.
For the case of noise disturbances d(t), the nominal system remains equations (8.i2) but
the actual system is
dKa/dt = AN(t)Xa + B(t)u + J(t)d
Ya = C(t)Xa + D(t)u + K(t)d (8.15)
XN dN d
d * +
+
Fig. 818
+ Fig. 8. . 19
3d
CHAP. 8J RELATIONS WITH CLASSICAL TECHNIQUES 181
Here the unwanted signal is 8d which is generated by the nominal XN. This can be
incorporated into a block diagram containing linear elements, and the effect of the non
linearity can be evaluated in a manner similar to that used in deriving equations (8.16).
d(8x)/dt = AN(t) ax + j(t)8d
8y = C(t) 8x + k(t) Sd (8.17)
Now observability and controllability theory can be applied to equations (8.14), (8.16)
and (8.17). We conclude that, if possible, we will choose C(t), AN(t) and the corresponding
input matrices B(t), D(t) or J(t), K(t) such that the unwanted signal is unobservable with
respect to the output 8y(t), or at least the elements of the state vector associated with
the dominant poles are uncontrollable with respect to the unwanted signal. If this is im
possible, the system gain with respect to the unwanted signal should be made as low as
possible.
Example 8.17.
Consider the system d:x/dt = (~ _ ~ ) x + G) u. The nominal value of the parameter ~ is zero
and the nominal input u{t) is a· unit step function.
We ean eliminate the effects of a variation upon the output if c is chosen such that the output observability
matrix (ctb ctAb) == O. This results in a choice ct = (0 y) where y is any number.
Furthermore, all the analysis and synthesis techniques developed in this chapter can be
used to analyze and design systems to reduce sensitivity and nonlinear effects and to reject
noise. This may be done by using the error constant matrix table, root locus, Nyquist,
and/or pole placement techniques on the equations (8.14), (8.16) and (8.17).
182 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
Solved Problems
dx
dt
[Ci1 _~) . (t 12)(1 l)]X
t+1 t+1
which has a double pole at 1. Therefore the zero output of the closed loop system is asymp
totically stable and the remarks following Theorem 8.1 are valid. The class of all d(t) that the
system can follow with lim e(t) = 0 is given by
t+ OQ
dw
dt
Cl1 t+1
1t ) w
+ (t~.2).
t+1
g
T
1
+1
(t +1  eT 
eT  t
t
t(1 eT  t
1 + te'Tt
»)
Then
d(t)
(Notice this system is unobservable.) For constant error let get) == K, an arbitrary constant. Then
K(t+1)
fto
t '7"+2
(r+1)2dT K(t + 1HIn (t + 1)  In (to + 1) + (to + 1)1  (t + 1)1]
and the system follows with error K all functions that are asymptotic to this. Since the system is
reasonably well behaved, we can assume that the system will follow all functions going to infinity
slower than Ie(t + 1) In (t + 1).
:t (::) Z3
(~ ~ ~)(::) + (~ ~)(~:)
0 0 3 Za 1 0
Place the poles of the closed loop system at 4, 5 and 6.
Transform the system to phase variable canonical form using the results of Problem 7.21.
4 9 0 0 le3
Then
1/2)
z
(
leI
0
0
1
0
1
K2
0
~
le  1
3
)(! 1
5/2
:4
3/2 1/2
1 x
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 183
and
dx
dt ( 0 1 0) (1 1 1)(Q'/(l
o 0
6 11
1
6
x + 1 2 3
1 4 9
1C2
lCa
Kl)( )
K2
0
::
dx
dt =
11
~ ~) 6
x + (~ ~ +~ 1/2Q'
1
~;::)(::)
4
To have the closed loop poles at 4, 5 and 6 we desire a charateristic polynomial AS + 15A.2 + 74A + 120.
Therefore we choose ul =
114xl  63x 2  9xa and U2 OXI + OX2 + OXa. =
In the case a = 0, the state Zt is uncontrollable with respect to Ul and, from Theorem 7.7,
cannot be put into phase variable canonical form with respect to u 1 alone. Hence we must use
U2 to control Zl. and ean assure a pole at 4 by choosing u2 = 3z 1 • Then we have the singleinput
system
can be transformed to phase variable canonical form, and ul:;:::: 12z2 + 6zs'
The above procedure can be generalized to give a means of obtaining multipleinput pole
placement.
8.3. Given the system d 2 y/dt2 = O. Construct the observer system such that it has a double
pole at yo Then find the error e'= x  x as a function of y, if the output really is
y(t) + 'J](t), where 'Y)(t) is noise.
The observer system has the form
d~
dt
=
(~ 1)A + (kl)z
o X k A
[(1 O)x  y  1]]
The characteristic equation of the closed loop system is A(A  k 1)  k2 = O. A double pole at y has
the characteristic equation A2 + 2YA + y2 = O. Hence set kl = 2y and k 2 :;:::: _y2. Then the
equation for the error is
Note the noise drives the error and prevents it from reaching zero.
The transfer function is found from
.e{1]} (2YS+ y2)
.e {(::)} 82 + 2ys + y2 y 2s
As y ~ 0::>, then el ~ 1] and ez ~ drJldt. If 1](t) = 170 cos IUt, then 1U110. the amplitude of d11/dt,
may be large even though '110 is small, because the noise may be of very high frequency. We con
clude that it is not a good idea to set the observer system gains too high, so that the observer system
can filter out some of the noise.
184 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
Design a feedback controller given dominant closed loop poles in the z plane at
(1 ± 1)/2.
We shall construct an observer to generate the estimate of the state from which we can con
struct the desired control. The desired closed loop characteristic· equation is ",2  X+ 1/2 = o.
Hence we choose u = 3~/2  2~2. To generate ~l and ~2 we choose a firstorder observer with a
pole at 0.05 so that it will hardly affect the response due to the dominant poles and yet will be large
enough to filter highfrequency noise. The transformation of variables
P = (1 0.05) and
which we shall need later. Next we let the actual eigenvalues Xf = Xf + SXL. Substituting this
into the eigenvalue equation for the actual Aa gives
Neglecting the last qua~tity on the right since it is of second order and using equation (8.18) then
leads to
v! 8Awi (8.20)
VTWi
Ad. = (
1 + a2 ct.) 1 0)
( 2 2 +
(ai30 0ct.)
2 2
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 185
1
(~) V1 = (~)
2 W2
(~) V2 (~)
Using equation (8.20),
).(21)
2
'1 + (1 0) ( a0 aO·
For larger values of a note we can use root locus under parameter variation to obtain exact values.
However, the root locus is difficult computationally for very high~order systems, whereas the
procedure just described has been applied to a 51storder system.
8.6. Given the scalar nonlinear system of Fig. 820 with input a sin t. Should K> 1
be increased or decreased to minimize the effect of the nonlinearity?
a sin t y a sin t
+ +
The nominal linear system is shown in Fig. 8~21. The steady state value of eN = (a:: sin t)/(K  1).
We approximate this as the input to the unwanted signal d 2 (Sy)/dt2 == e~, which gives the steady
state value of ay::;::: a3 (27 sin t  sin 3t)/[36(K 1)3]. This approximation that d 2 (oy)/dt2 = e~
instead of e!
must be verified by simulation. It turns out this a good approximation for lal ~ 1,
and we can conclude that for lal <" 1 we increase K to make ay/y become smaller.
+ +
8V
Fig. 8·22
186 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
The main point is to realize this "multiple loop" system is really a multiple inputmultiple
output system. The control system shown has constrained SPcr to be an inner loop. A means
of improving response is to rewrite the system in matrix form as shown in Fig. 823. This opens
up the possibility of "cross coupling" the feedback, such as having BPcr depend on 8P as well as
ST. Furthermore it is evident that the system is of type 1 and can follow a step input with zero
steady state error.
>
Fig. 823
8.8. Compensate the system (S+Pl)I(S+P Z)1 to have poles at "'1 and "'2' with an
observer pole at 7ro, by using the algebraic separation theorem, and discuss the effect
of noise 7J at the output.
The state space representation of the plant is
(~:) = (~r(:) = ( ~1
£lZ Vo  : ,  p.)(:)
The flow diagram of the closed loop system is shown in Fig. 824.
A
Xt =Y
Fig. 824
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 187
Note the noise 'tJ is fed through a firstorder system and gain elements to form ?.t. If the noise
level is high, it is better to use a secondorder observer or a Kalman filter because then the noise
goes through no gain elements directly to the control, but instead is processed through first and
secondorder systems. If there is no noise whatsoever, flow diagram manipulations can be used to
show the closed loop system is equivalent to one compensated by a lead network, i.e. the above flow
diagram with 'I] = 0 can be rearranged as in Fig. 825.
d 1 y
+
Fig. 825
Note the observer dynamics have cancelled out, and the closed loop system remains second
order. This corresponds to the conclusion the observer dynamics are uncontrollable by d. However
any initial condition in the observer will produce an effect on y, as can be seen from the first flow
diagram.
Supplementary Problems
8.9. Given the matrix block diagram of Fig. 826. Show that this reduces to Fig. 827 when the
indicated inverse exists.
d y
+
y
=d====::;:j1 G(I+HG)l ;>
8.10. Given the matrix block diagram of Fig. 828. Reduce the block diagram to obtain HI isolated in a
single feedback loop.
Fig. 828
188 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
8.12. Given the timevarying system d2y/dt 2 + aCt) dy/dt + f3(t)y ::::: u. Find a feedback control u such
that the closed loop system behaves like d 2z/dt 2 + 8(t) dz/dt + ¢(t)z ::::: O.
Construct a firstorder observer system with a pole at 3 and then find a feedback control
u =: kl~l + k 2 '052 that places both the closed loop system poles at 2.
8.17. Given the system dx/dt = A(t)x + B(t)u where y = C(t)x + D(t)u. What is the form of the
observer system when D(t) ~ O? What is the algebraic separation theorem when D(t) ¥= O?
where x(O) = o. As a measure of the sensitivity of the system, assuming jo:(t) 1 ~ 1, find ay(t)
as an integral involving aCt) and u(t). Hint: It is easiest to use Laplace transforms.
8.19. Given the nominal system dXN/dt = AN(t)xN + BN(t)UN and YN == CN(t)XN + DN(t)UN' What is the
equation for 8x corresponding to equation (8.1.0 when the parameters of the system become Aa(t).
Ba(t), Ca(t), Da(t) and ua(t)? .
1 + t + fet) U(t»)
8.20. Given the system dx/dt == ( f(t) t2 x. Choose u(t) such that at least one state
will be insensitive to small variations in f(t), given the nominal solution XN(t).
8.21. Given that the input d(t) is generated by the scalar system dw/dt = a(t)W and d(t) = [yet) + p(t)Jw,
under what conditions on pet) can the system dUJ/dt = c.:(t)x + f3(t)e with y = y(t)x follow any such
d(t) with zero error?
CHAP. 8] RELATIONS WITH CLASSICAL TECHNIQUES 189
8.22. Given the general nonunity feedback timeinvariant system of Fig. 830. Under what conditions
can lim e(t) = O? Set F = I and H;::; I and derive Theorem 8.2 from these conditions.
tco
+
.({y}
.e{d}.."l J....,/' .e{e}
Fig. 830
d(t) = ft C(t)
to
IPABC Ct, '7") B(r) d(r) dT + get) + C(t) q.ABC(t, to) w(to)
8.25. Given the system of Fig. 831 where the constant matrix K has been introduced as compensation.
Show that
(a) The type number of the system cannot change if K is nonsingular.
(b) The system is of type zero if K is singular.
d(t)
+
)t

<
K 
...,.. H(s) ">
Fig. 831
8.26. Design a system that will follow d(t) = sin t with zero steady state error.
d y
Fig. 832
190 RELATIONS WITH CLASSICAL TECHNIQUES [CHAP. 8
8.11. The closed loop system output y = (~)2 [(2 f  1) Xl(t O) + (t  to) X 2 (t O) ] tends to zero, so
Theorem 8.1 applies. Then t 0
d(t) = w,(t O) + ~ [1  C:)'] w2(t O) + Jot L[1  (f)'] y(.) d, '"" In t for t > some t,
so that the system can follow steps but not ramps with zero error.
8.13.
d~
dt G~ D~ G) + [(0 1 l)~YI
8.14. dw/dt w + (0 1)y,
.A
X
8.15.
1
d(t)+~ 1   __  _ yet)
82 + 38 +2
Fig. 833
8.17. Subtract D(t)u from 11 entering the observer and this reduces to the given formulation .
8.20. d(8x)/dt
1 + t.+ t(t) tt(t») Sx + (XN1) ot
( j{t) t2 XN2
8.21. pet) = e
 itto a(1)) d1) [
e(t) + ft y(t)e I
to T
tCl (1)) d1)
fJ(T) fJ(T) dT
]
+ y(t)/C where e(t) is any function such
8.22. lim [F(s)  G(s)(I + G{s) H(8»1]8~{d} = 0 and is analytic for s =::: O.
s+o
Definition 9.1: For every Xo and every to, if there exists a constant K depending on Xo and
to such that Ilx(t)11 ~ K for all t ~ to, then the response x(t) is bounded.
Even this simple definition has difficulties. The troubie is that we must specify the response
to what. The trajectory of x(t) = </J(t; U(T),X(tO), to) depends implicitly on three quantities:
U(T), x(to) and to. By considering only U(T) = 0 in this section, we have eliminated one
difficulty for a while. But can the boundedness of the response depend on x(to)?
Example 9.1.
Consider the scalar zeroinput nonlinear equation dx/dt = x + x 2 with initial condition x(O) = Xo.
The solution to the linearized equation dx/dt = x is obviously bounded for all xo' However, the solution
to the nonlinear equation is
x(t)
For all negative values of xo, this is well behaved. For values of Xo > 1, the denominator vanishes at a
time tl = In Xo  In (xo 1), so that lim x(t) = 00,
tt 1
It can be concluded that boundedness depends upon the initial conditions for nonlinear
equations in general.
Theorem 9.1: The boundedness of the response x(t) of the linear system dx/dt = A(t)x
is independent of the initial condition Xo.
191
STABILITY OF LINEAR SYSTEMS [CHAP. 9
192
Proof: Ilx(t)11 = 11q,(t; 0, Xo, to)11 = 114J(t, to)xoll ::=: 114t(t, to)llllxoll
Since Ilxoll is a constant, if Ilx(t)11 becomes unbounded as t ~ it is solely
00, due to cp(t, to).
Now we shall consider other, different types of stability. First, note that x = 0 is a
steady state solution (an equilibrium state) to the zeroinput linear system dx/dt = A(t)x.
We shall define a region of state space by Ilxll < 10, and see if there exists a small region of
nonzero perturbations surrounding the equilibrium state x = 0 that give rise to a trajectory
which remains within Ilx!1 < E. If this is true for all E> 0, no matter how small, then we
have
Definition 9.2: The equilibrium state x = 0 of dx/dt = A(t)x is stable in the sense of
Liapunov (for short, stable Ls.L.) if for any to and every real number E> 0,
there is some 0 > 0, as small as we please, depending on to and E such
that if Ilxoll < 0 then Ilx(t)11 < E for all t > to.
This definition is also valid for nonlinear s!ystems with an equilibrium state x = O.
It is the most common definition of stability, and in the literature "stable Ls.L." is often
shortened to "stable". States that are not stable Ls.L. will be called unstable. Note stability
Ls.L. is a local condition, in that 0 can be as small as we please. Finally, since x = 0 is an
obvious choice of equilibrium state for a linear system, when speaking about linear systems
we shall not be precise but instead will say the system is stable when we mean the zero state
is stable.
Example 9.2.
Consider the nonlinear system of Example 9.1. If Xo === 1, then
Ix(t)1 ::::
11 + (e t 1)(1 xo)1
In Definition 9.2 we can set 0 = € > 0 if e:::: 1, and if € > 1 we set 0:= 1 to show the zero state of
Example 9.1 is stable i.s.L. Hence the zero state is stable Ls.L. even though the response can become
unbounded for some Xo' (This situation corresponds to Fig. 92(b) of Problem 9.1.) Of course if the
response became unbounded for all Xo # 0, the zero state would be considered unstable. Another point to
note is that in the application of Definition 9.2, in the range where e is small there results the choice of
a correspondingly small 8.
Example 9.3.
Given the Van der Pol equation
Theorem 9.2: The transition matrix of a linear system is bounded as jl4J(t, to)jj < K(tO) for
all t === to if and only if the equilibrium state x = 0 of dx/dt = A(t)x
is stable Ls.L.
Note Ilx(t)11 is bounded if Ilcp(t, to)11 is bounded.
CHAP. 9] STABILITY OF LINEAR SYSTEMS 193
Proof: First assun1e 1)I'Jl(t, to))1 < K(tO) where K is a constant depending only on to.
If we are given any € > 0, then we can always find 8 = dK(tO) such that if Ilxoll < 8 then
E = K(to)8 > 111'Jl(t, to)llllxoll == 114I(t, to)xoll = Ilx(t)jj. From Definition 9.2 we conclude stability
i.s.L.
Next we assume stability Ls.L. Let us suppose cp(t, to) is not bounded, so that