LyapunovRiccati

download LyapunovRiccati

of 34

Transcript of LyapunovRiccati

  • 8/13/2019 LyapunovRiccati

    1/34

    Lyapunov Operator

    Let A F nn be given, and dene a linear operator LA : C nn C nn asLA (X ) := A X + XA

    Suppose A is diagonalizable (what follows can be generalized even if this isnot possible - the qualitative results still hold). Let { 1, 2, . . . , n} be theeigenvalues of A. In general, these are complex numbers. Let V C nn bethe matrix of eigenvectors for A , so

    V = [v1 v2 vn ] C nnis invertible, and for each i

    A vi = ivi

    Let wi C n be dened so that

    V 1 =

    wT 1wT 2...

    wT n

    Since V 1V = I n , we have that

    wT i v j = ij :=0 if i = j1 if i = j

    For each 1 i, j n deneX ij := viv j C nn

    130

  • 8/13/2019 LyapunovRiccati

    2/34

    Eigenvalues Lyapunov Operator

    Lemma 63 The set {X ij }i=1 ,.,n ; j =1 ,.,n is a linearly independent set, and this set is a full set of eigenvectors for the linear operator LA. Moreover, the eigenvalues of LA is the set of complex numbers i + j i=1 ,.,n ; j =1 ,.,nProof: Suppose { ij }i=1 ,.,n ; j =1 ,.,n are scalars and

    n

    j =1

    n

    i=1 ij X ij = 0nn (7.19)

    Premultiply this by wT l and postmultiply by wk

    0 = wT ln

    j =1

    n

    i=1 ij X ij wk

    =n

    j =1

    n

    i=1 ij wT l viv j wk

    =n

    j =1

    n

    i=1 ij li kj

    = lk

    which holds for any 1 k, l n . Hence, all of the s are 0, and the set{X ij }i=1 ,.,n ; j =1 ,.,n is indeed a linearly independent set. Also, by denitionof LA, we have

    LA (X ij ) = A X ij + X ij A= A viv j + viv j A= iviv j + vi j v j= i + j viv j= i + j X ij

    So, X ij is indeed an eigenvector of LA with eigenvalue i + j .

    131

  • 8/13/2019 LyapunovRiccati

    3/34

  • 8/13/2019 LyapunovRiccati

    4/34

    Lyapunov Operator Integral Formula

    If A is stable (all eigenvalues have negative real parts) then i + j = 0for all combinations. Hence LA is an invertible linear operator. We canexplicitly solve the equation

    LA (X ) = Qfor X using an integral formula.

    Theorem 64 Let A F nn be given, and suppose that A is stable. The LAis invertible, and for any Q F nn , the unique X F nn solving LA (X ) =Q is given by

    X = 0

    eA Qe A d

    Proof: For each t 0, deneS (t) := t0 eA Qe A d

    Since A is stable, the integrand is made up of decaying exponentials, andS (t) has a well-dened limit as t , which we have already denotedX := lim

    tS (t). For any t 0,

    A S (t) + S (t)A = t0 A eA Qe A + eA Qe A A d = t0

    dd

    eA Qe A d

    = eA tQe At QTaking limits on both sides gives

    A X + XA = Qas desired.

    133

  • 8/13/2019 LyapunovRiccati

    5/34

    Lyapunov Operator More

    A few other useful facts

    Theorem 65 Suppose A, Q F nn are given. Assume that A is stable and Q = Q 0. Then the pair (A, Q ) is observable if and only if

    X := 0 eA Qe A d > 0Proof: Suppose not. Then there is an xo F n , xo = 0 such that

    Qe At xo = 0

    for all t 0. Consequently xoeA tQe At xo = 0 for all t 0. Integrat-ing gives

    0 = 0 xoeA tQe At xo dt= xo 0 eA tQe At dt xo= xoXx o

    Since xo = 0, X is not positive denite.

    Suppose that X is not positive denite (note by virtue of the deni-tion X is at least positive semidenite). Then there exists xo F n ,

    xo = 0 such that xoXx o = 0. Using the integral form for X (sinceA is stable by assumption) and the fact tha Q 0 gives 0 Q

    12 eA xo d = 0

    The integrand is nonnegative and continuous, hence it must be 0for all 0. Hence QeA xo = 0 for all 0. Since xo = 0, thisimplies that ( A, Q ) is not observable.

    134

  • 8/13/2019 LyapunovRiccati

    6/34

    Lyapunov Operator More

    Theorem 66 Let (A, C ) be detectable and suppose X is any solution to A X + XA = C C (there is no apriori assumption that LA is an invertible operator, hence there could be multiple solutions). Then X 0 if and only if A is stable.

    Proof Suppose not, then there is a v Cn , v = 0, C , Re( ) 0with Av = v (and v A = v ). Since (A, C ) is detectable and theeigenvalue in the closed right-half plane Cv = 0. Note that

    Cv 2 = v C Cv= v (A X + XA ) v= + v Xv= 2Re ( ) v Xv

    Since Cv > 0 and Re( ) 0, we must actually have Re ( ) > 0and v Xv < 0. Hence X is not positive semidenite.

    Since A is stable, there is only one solution to the equation A X +XA = C C , moreover it is given by the integral formula,

    X = 0 eA C Ce A d which is clearly positive semidenite.

    135

  • 8/13/2019 LyapunovRiccati

    7/34

    Lyap Ordering

    Finally, some simple results about the ordering of Lyapunov solutions.

    Let A be stable, so LA is guaranteed to be invertible. Use L1A (Q) todenote the unique solution X to the equation A X + XA = Q .Lemma 67 If Q1 = Q1 Q2 = Q2, then

    L1A (Q1) L1A (Q2)

    Lemma 68 If Q1 = Q1 > Q 2 = Q2, then

    L1A (

    Q1) >

    L1A (

    Q2)

    Proofs Let X 1 and X 2 be the two solutions, obviously

    A X 1 + X 1A = Q1A X 2 + X 2A = Q2

    A (X 1 X 2) + ( X 1 X 2) A = (Q1 Q2)Since A is stable, the integral formula applies, and

    X 1 X 2 = 0 eA (Q1 Q2) eA d Obviously, if Q1 Q2 0, then X 1 X 2 0, which proves Lemma 5. If Q1 Q2 > 0, then the pair ( A, Q 1 Q2) is observable, and by Theorem3, X 1 X 2 > 0, proving Lemma 6.

    Warning: The converses of Lemma 5 and Lemma 6 are NOT true.

    136

  • 8/13/2019 LyapunovRiccati

    8/34

    Algebraic Riccati Equation

    The algebraic Riccati equation (A.R.E.) is

    AT X + XA + XRX Q = 0 (7.20)where A, R, Q Rnn are given matrices, with R = RT , and Q = QT , andsolutions X C nn are sought. We will primarily be interested in solutionsX of (7.20), which satisfy

    A + RX is Hurwitz

    and are called stabilizing solutions of the Riccati equation.

    Associated with the data A, R , and Q, we dene the Hamiltonian matrix

    H R2n

    2n

    byH :=

    A R

    Q AT (7.21)

    137

  • 8/13/2019 LyapunovRiccati

    9/34

    Invariant Subspace Denition

    Suppose A Cnn . Then (via matrix-vector multiplication), we view A asa linear map from C n C n .

    A subspace V C n is said to be A-invariant if for every vector v V , thevector Av V too.

    In terms of matrices, suppose that the columns of W C nm are a basis for

    V . In other words, they are linearly independent, and span the space

    V , so

    V = {W : C m}

    Then V being A-invariant is equivalent to the existence of a matrix S C mmsuch that

    AW = W S

    The matrix S is the matrix representation of A restricted to V , referred tothe basis W . Clearly, since W has linearly independent columns, the matrixS is unique.

    If (Q, ) are the Schur decomposition of A, then for any k with 1 k n ,the rst k columns of Q span a k-dimensional invariant subspace of A, as

    seen belowA [Q1 Q2] = [Q1 Q2]

    11 120 22

    Clearly, AQ1 = Q111.

    138

  • 8/13/2019 LyapunovRiccati

    10/34

    Eigenvectors in Invariant subspaces

    Note that since S Cmm has at least one eigenvector, say = 0 m , withassociated eigenvalue , it follows that W V is an eigenvector of A,

    A (W ) = AW = W S = W = (W )

    Hence, every nontrivial invariant subspace of A contains an eigen-vector of A.

    139

  • 8/13/2019 LyapunovRiccati

    11/34

    Stable Invariant Subspace Denition

    If S is Hurwitz, then V is a stable, A-invariant subspace.Note that while the matrix S depends on the particular choice of the basis for

    V (ie., the specic columns of W ), the eigenvalues of S are only dependenton the subspace V , not the basis choice.To see this, suppose that the columns of W C nm span the same space asthe columns of W .

    Obviously, there is a matrix T Cmm such that W = W T . The linearindependence of the columns of W (and W ) imply that T is invertible. Also,

    A W = AW T 1 = W ST 1 = W TST 1 S

    Clearly, the eigenvalues of S and S are the same.

    The linear operator A|V is the transformation from V V dened as A,simply restricted to V .Since V is A-invariant, this is well-dened. Clearly, S is the matrix repre-sentation of this operator, using the columns of W as the basis choice.

    140

  • 8/13/2019 LyapunovRiccati

    12/34

    AT X + XA + XRX Q = 0 Invariant Subspace The rst theorem gives a way of constructing solutions to (7.20) in terms of invariant subspaces of H .

    Theorem 69 Let V C2n

    be a n-dimensional invariant subspace of H ( H is a linear map from C2n to C2n ), and let X 1 Cnn , X 2 Cnn be two complex matrices such that

    V = spanX 1X 2

    If X 1 is invertible, then X := X 2X 11 is a solution to the algebraic Riccati equation, and spec (A + RX ) = spec (H |V ).Proof: Since V is H invariant, there is a matrix C nn with

    A R

    Q AT X 1X 2

    =X 1X 2

    (7.22)

    which givesAX 1 + RX 2 = X 1 (7.23)

    andQX

    1 AT X

    2 = X

    2 (7.24)Pre and post multiplying equation (7.23) by X 2X 11 and X 11 respec-tively gives

    X 2X 11 A + X 2X 11 R X 2X 11 = X 2X 11 (7.25)and postmultiplying 7.24 by X 11 gives

    Q AT X 2X 11 = X 2X 11 (7.26)Combining (7.25) and (7.26) yields

    AT (X 2X 11 ) + ( X 2X 11 )A + ( X 2X 11 )R (X 2X 11 ) Q = 0so X 2X 11 is indeed a solution. Equation (7.23) implies that A+ R X 2X 11 =X 1X 11 , and hence they have the same eigenvalues. But, by denition, is a matrix representation of the map H |V , so the theorem is proved.

    141

  • 8/13/2019 LyapunovRiccati

    13/34

    AT X + XA + XRX Q = 0 Invariant Subspace The next theorem shows that the specic choice of matrices X 1 and X 2 whichspan V is not important.

    Theorem 70 Let V , X 1, and X 2 be as above, and suppose there are two other matrices X 1, X 2 Cnn such that the columns of

    X 1X 2

    also span

    V . Then X 1 is invertible if and only if X 1 is invertible. Consequently, bothX 2X 11 and X 2 X 11 are solutions to (7.20), and in fact they are equal.

    Remark: Hence, the question of whether or not X 1 is invertible is a property

    of the subspace, and not of the particular basis choice for the subspace.Therefore, we say that an n -dimensional subspace V has the invertibility property if in any basis, the top portion is invertible. In the literature,this is often called the complimentary property .

    Proof: Since both sets of columns span the same n-dimensional subspace,there is an invertible K C nn such that

    X 1X 2 =

    X 1X 2 K

    Obviously, X 1 is invertible if and only if X 1 is invertible, and X 2X 11 = X 2K X 1K

    1= X 2 X 11 . .

    142

  • 8/13/2019 LyapunovRiccati

    14/34

    AT X + XA + XRX Q = 0 Invariant Subspace As we would hope, the converse of theorem 69 is true.

    Theorem 71 If X Cn

    n

    is a solution to the A.R.E., then there existmatrices X 1, X 2 C nn , with X 1 invertible, such that X = X 2X 11 and the columns of

    X 1X 2

    span an n-dimensional invariant subspace of H .

    Proof: Dene := A + RX Cnn . Multiplying this by X gives X =XA + XRX = Q AT X , the second equality coming from the fact thatX is a solution to the Riccati equation. We write these two relations as

    A R

    Q AT I nX

    =I nX

    Hence, the columns of I nX

    span an n-dimensional invariant subspace

    of H , and dening X 1 := I n , and X 2 := X completes the proof. .

    Hence, there is a one-to-one correspondence between n-dimensional invari-ants subspaces of H with the invertibility property, and solutions of theRiccati equation.

    143

  • 8/13/2019 LyapunovRiccati

    15/34

    AT X + XA + XRX Q = 0 Eigenvalue distribution of H In the previous section, we showed that any solution X of the Riccati equa-tion is intimately related to an invariant subspace, V , of H , and the matrixA + RX is similiar (ie. related by a similarity transformation) to the re-

    striction of H on V , H |V . Hence, the eigenvalues of A + RX will alwaysbe a subset of the eigenvalues of H . Recall the particular structure of theHamiltonian matrix H ,

    H :=A R

    Q AT Since H is real, we know that the eigenvalues of H are symmetric about thereal axis, including multiplicities. The added structure of H gives additionalsymmetry in the eigenvalues as follows.

    Lemma 72 The eigenvalues of H are symmetric about the origin (and hence also about the imaginary axis).

    Proof: Dene J R 2n2n by

    J =0 I I 0

    Note that JHJ 1 = H T . Also, JH = H T J = ( JH )T . Let p ( ) bethe characteristic polynomial of H . Then

    p( ) := det ( I H )= det I JHJ 1= det I + H T

    = det ( I + H )= ( 1)2ndet(I H )= p( )

    (7.27)

    Hence the roots of p are symmetric about the origin, and the symmetryabout the imaginary axis follows.

    144

  • 8/13/2019 LyapunovRiccati

    16/34

    AT X + XA + XRX Q = 0 Hermitian Solutions

    Lemma 73 If H has no eigenvalues on the imaginary axis, then H has neigenvalues in the open-right-half plane, and n eigenvalues in the open-left-

    half plane.

    Theorem 74 Suppose V is a n-dimensional invariant subspace of H , and X 1, X 2 Cnn form a matrix whose colums span V . Let Cnn be a matrix representation of H

    |V . If i + j

    = 0 for all eigenvalues i , j

    spec (), then X 1 X 2 = X 2 X 1.Proof: Using the matrix J from the proof of Lemma 72 we have that

    [X 1 X 2 ] JH X 1X 2

    is Hermitian, and equals ( X 1 X 2 + X 2 X 1) . Dene W := X 1 X 2 +X

    2X

    1. Then W

    = W

    , and W

    + W

    = 0. By the eigenvalueassumption on , the linear map L :C nn C nn dened byL (X ) = X + X

    is invertible, so W = 0.

    Corollary 75 Under the same assumptions as in theorem 74, if X 1 is in-vertible, then X 2X 11 is Hermitian.

    145

  • 8/13/2019 LyapunovRiccati

    17/34

    AT X + XA + XRX Q = 0 Real Solutions Real (rather than complex) solutions arise when a symmetry condition isimposed on the invariant subspace.

    Theorem 76 Let V be a n-dimensional invariant subspace of H with the invertibility property, and let X 1, X 2 Cnn form a 2n n matrix whose columns span V . Then V is conjugate symmetric ( {v : v V} = V ) if and only if X 2X 11 R nn .

    Proof: Dene X := X 2X 11 . By assumption, X R nn , and

    spanI

    nX = V

    so V is conjugate symmetric. Since V is conjugate symmetric, there is an invertible matrix K

    C nn such that

    X 1X 2

    =X 1X 2

    K

    Therefore, ( X 2X 11 ) = X 2X 11 = X 2K (X 1K )1 = X 2X 11 as de-sired.

    146

  • 8/13/2019 LyapunovRiccati

    18/34

    AT X + XA + XRX Q = 0 Stabilizing Solutions Recall that for any square matrix M , there is a largest stable invariantsubspace. That is, there is a stable invariant subspace V S that contains everyother stable invariant subspace of M . This largest stable invariant subspace

    is simply the span of all of the eigenvectors and generalized eigenvectorsassociated with the open-left-half plane eigenvalues. If the Schur form hasthe eigenvalues (on diagonal of ) ordered, then it is spanned by the rstset of columns of Q

    Theorem 77 There is at most one solution X to the Riccati equation suchthat A + RX is Hurwitz, and if it does exist, then X is real, and symmetric.

    Proof: If X is such a stabilizing solution, then by theorem 71 it can beconstructed from some n-dimensional invariant subspace of H , V . If A + RX is stable, then we must also have spec (H |V ). Recall that thespectrum of H is symmetric about the imaginary axis, so H has atmost n eigenvalues in C, and the only possible choice for V is thestable eigenspace of H . By theorem 70, every solution (if there are any)constructed from this subspace will be the same, hence there is at most1 stabilizing solution.

    Associated with the stable eigenspace, any as in equation (7.22) willbe stable. Hence, if the stable eigenspace has the invertibility property,from corollary 75 we have X := X 2X 11 is Hermitian. Finally, since H is real, the stable eigenspace of H is indeed conjugate symmetric, so theassociated solution X is in fact real, and symmetric.

    147

  • 8/13/2019 LyapunovRiccati

    19/34

    AT X + XA + XRX Q = 0 domS RicIf H is (7.21) has no imaginary axis eigenvalues, then H has a unique n-dimensional stable invariant subspace V S C2n , and since H is real, thereexist matrices X 1, X 2 R nn such that

    spanX 1X 2

    = V S .

    More concretely, there exists a matrix R nn such that

    H X 1X 2

    =X 1X 2

    , spec () C

    This leads to the following denition

    Denition 78 (Denition of domS Ric) Consider H R2n2n as dened in (7.21). If H has no imaginary axis eigenvalues, and the matrices X 1, X 2R nn for which

    spanX 1

    X 2

    is the stable, n-dimensional invariant subspace of H satisfy det(X 1) = 0 ,then H domS Ric. For H domS Ric, dene RicS (H ) := X 2X 11 , which is the unique matrix X R nn such that A + RX is stable, and

    AT X + XA + XRX Q = 0 .Moreover, X = X T .

    148

  • 8/13/2019 LyapunovRiccati

    20/34

    AT X + XA + XRX Q = 0 Useful Lemma

    Theorem 79 Suppose H (as in (7.21)) does not have any imaginary axis eigenvalues, R is either positive semidenite, or negative semidenite, and

    (A, R ) is a stabilizable pair. Then H domS Ric.

    Proof: LetX 1X 2

    form a basis for the stable, n-dimensional invariant sub-

    space of H (this exists, by virtue of the eigenvalue assumption aboutH ). Then

    A R

    Q

    AT

    X 1X 2

    =X 1X 2

    (7.28)

    where Cnn is a stable matrix. Since is stable, it must be thatX 1 X 2 = X 2 X 1. We simply need to show X 1 is invertible. First we provethat Ker X 1 is -invariant, that is, the implication

    x KerX 1 x KerX 1Let x KerX 1. The top row of (7.28) gives RX 2 = X 1 AX 1.Premultiply this by x X 2 to get

    x X 2 RX 2x = x X 2 (X 1 AX 1) x= x X 2 X 1x= x X 1 X 2x since X 2 X 1 = X 1 X 2= 0

    .

    Since R 0 or R 0, it follows that RX 2x = 0. Now, using (7.28)again gives X 1x = 0, as desired. Hence KerX 1 is a subspace that is

    -invariant. Now, if KerX 1 = {0}, then there is a vector v = 0 and C , Re ( ) < 0 such thatX 1v = 0

    v = vRX 2v = 0

    149

  • 8/13/2019 LyapunovRiccati

    21/34

    We also have QX 1 AT X 2 = X 2, from the second row of (7.28).Mutiplying by v gives

    v X 2 A I R = 0Since Re < 0, and (A, R ) is assumed stabilizable, it must be thatX 2v = 0. But X 1v = 0, and

    X 1X 2

    is full-column rank. Hence, v =

    0, and KerX 1 is indeed the trivial set {0}. This implies that X 1 isinvertible, and hence H domS Ric. .

    150

  • 8/13/2019 LyapunovRiccati

    22/34

    AT X + XA + XRX Q = 0 Quadratic Regulator The next theorem is a main result of LQR theory.

    Theorem 80 Let A,B,C be given, with (A, B ) stabilizable, and (A, C )

    detectable. Dene H :=

    A BB T C T C AT

    Then H domS Ric. Moreover, 0 X := Ric S (H ), and

    KerX

    C

    CA...

    CA n1

    Proof: Use stabilizablity of (A, B ) and detectability of ( A, C ) to show thatH has no imaginary axis eigenvalues. Specically, suppose v1, v2 Cnand R such that

    A BB T

    C T C

    AT

    v1v2

    = jv1v2

    Premultiply top and bottom by v2 and v1, combine, and conclude v1 =v2 = 0. Show that if ( A, B ) is stabilizable, then A, BB T is also sta-bilizable. At this point , apply Theorem 79 to conclude H domS Ric.Then, rearrange the Riccati equation into a Lyapunov equation anduse Theorem 64 to show that X 0. Finally, show that Ker X KerC ,and that Ker X is A-invariant. That shows that Ker X KerCA j forany j 0. .

    151

  • 8/13/2019 LyapunovRiccati

    23/34

    Derivatives Linear Maps

    Recall that the derivative of f at x is dened to be the number lx such that

    lim 0

    f (x + ) f (x) lx

    = 0

    Suppose f : R n Rm is differentiable. Then f (x) is the unique matrixLx R mn which satises

    lim 0n

    1

    (f (x + ) f (x) Lx ) = 0Hence, while f : Rn Rm , the value of f at a specic location x, f (x),is a linear operator from R n

    R m . Hence the function f is a map from

    R n L (R n , R m).Similarly, if f : H nn H nn is differentiable, then the derivative of f at xis the unique linear operator LX : H nn H nn such that satisfying

    lim 0n n

    1

    (f (X + ) f (X ) LX ) = 0

    152

  • 8/13/2019 LyapunovRiccati

    24/34

    Derivative of Riccati Lyapunov

    f (X ) := XDX A X XA C

    Note,

    f (X + ) = ( X + ) D (X + ) A (X + ) (X + ) A C = XDX A X XA C + DX + XD A A + D = f (X ) ( A DX ) (A DX ) + D

    For any W Cnn , let LW : Hnn Hnn be the Lyapunov operator

    LW (X ) := W X + XW . Using this notation,f (X + ) f (X ) LA+ DX () = D

    Therefore

    lim 0n n

    1

    [f (X + ) f (X ) LA+ DX ()] = lim 0n n1

    [ D ] = 0

    This means that the Lyaponov operator LA+ DX is the derivative of thematrix-valued functionXDX A X XA C

    at X .

    153

  • 8/13/2019 LyapunovRiccati

    25/34

    Iterative Algorithms Roots

    Suppose f : R R is differentiable. An iterative algorithm to nd a rootof f (x) = 0 is

    xk+1 = xk f (xk)f (xk)

    Suppose f : R n R n is differentiable. Then to rst orderf (x + ) = f (x) + Lx

    An iterative algorithm to nd a root of f (x) = 0 n is obtained by settingthe rst-order approximation expression to zero, and solving for , yieldingthe update law (with xk+1 = xk + )

    xk+1 = xk L1xk [f (xk)]Implicitly, it is written as

    Lxk [xk+1 xk] = f (xk) (7.29)

    For the Riccati equation

    f (X ) := XDX

    A X

    XA

    C

    we have derived that the derivative of f at X involves the Lyapunov operator.The iteration in (7.29) appears as

    LA+ DX k (X k+1 X k) = X kDX k + A X k + X kA + C In detail, this is

    (A DX k) (X k+1 X k)(X k+1 X k) (A DX k) = X kDX k+ A X k+ X kA+ C which simplies (check) to

    (A DX k) X k+1 + X k+1 (A DX k) = X kDX k C

    Of course, questions about the well-posedness and convergence of the itera-tion must be addressed in detail.

    154

  • 8/13/2019 LyapunovRiccati

    26/34

  • 8/13/2019 LyapunovRiccati

    27/34

    Identities

    For any Hermitian matrices Y and Y , it is trivial to verify

    Y (A DY ) + ( A DY ) Y + Y DY = Y (A D Y ) + ( A D Y ) Y + Y DY Y Y D Y Y

    (7.31)

    Let X denote any Hermitian solution (if one exists) to the ARE,

    XDX A X XA C = 0 (7.32)This (the existence of a hermitian solution) is the departure point for allthat will eventually follow.

    For any hermitian matrix Z

    C = XDX + A X + XA= X (A DX ) + ( A DX ) X + XDX = X (A DZ ) + ( A DZ ) X + ZDZ (X Z ) D (X Z )

    (7.33)This exploits that X solves the original Riccati equation, and uses the iden-tity (7.31), with Y = X, Y = Z .

    156

  • 8/13/2019 LyapunovRiccati

    28/34

    Iteration

    Take any C C (which could be C , for instance). Also, let X 0 = X 0 0be chosen so that A DX 0 is Hurwitz. Check: Why is this possible?Iterate (if possible, which we will show is...) as

    X +1 (A DX ) + ( A DX ) X +1 = X DX C (7.34)Note that by the Hurwitz assumption on A DX 0, at least the rst iterationis possible ( = 0). Also, recall that the iteration is analogous to a rst-orderalgorithm for root nding, xn+1 := xn f (xn )f (xn ) .Use Z := X in equation (7.33), and subtract from the iteration equation

    (7.34), leaving(X +1 X ) (A DX ) + ( A DX ) (X +1 X )

    = (X X ) D (X X ) C C (7.35)

    which is valid whenever the iteration equation is valid.

    157

  • 8/13/2019 LyapunovRiccati

    29/34

    Main Induction

    Now for the induction: assume X = X and A DX is Hurwitz. Notethat this is true for = 0.

    Since A DX is Hurwitz, there is a unique solution X +1 to equation (7.34).Moreover, since ADX is Hurwitz, and the right-hand-side of (7.35) is 0,it follows that X +1 X 0 (note that X 0 X is not necessarily positivesemidenite).

    Rewrite (7.31) with Y = X +1 and Y = X .

    X +1 (A DX +1 ) + ( A DX +1 ) X +1 + X +1 DX +1= X +1 (A

    DX ) + ( A

    DX ) X +1 + X DX

    C (X +1

    X ) D (X +1

    X )

    Rewriting,

    C = X +1 (A DX +1 ) + ( A DX +1 ) X +1+ X +1 DX +1 + ( X +1 X ) D (X +1 X )

    Writing (7.33) with Z = X +1 gives

    C = X (A DX +1 ) + ( A DX +1 ) X + X +1 DX +1 (X X +1 ) D (X X +1 )Subtract these, yielding

    C C = ( X +1 X ) (A DX +1 ) + ( A DX +1 ) (X +1 X )+ ( X +1 X ) D (X +1 X ) + ( X X +1 ) D (X X +1 )

    (7.36)

    Now suppose that Re ( ) 0, and v Cn such that ( A DX +1 ) v = v .Pre and post-multiply (7.36) by v and v respectively. This gives

    v C C v 0

    = 2Re( )

    0v (X +1 X ) v 0

    +D 1/ 2 (X +1 X )D 1/ 2 (X X +1 )

    v2

    0Hence, both sides are in fact equal to zero, and therefore

    v (X +1 X ) D (X +1 X ) = 0 .158

  • 8/13/2019 LyapunovRiccati

    30/34

    Since D 0, it follows that D (X +1 X ) v = 0, which means DX +1 v =DX v. Hence

    v = ( A DX +1 ) v = ( A DX ) vBut A DX is Hurwitz, and Re( ) 0, so v = 0n as desired. This meansA DX +1 is Hurwitz.Summarizing:

    Suppose there exists a hermitian X = X solvingXDX A X XA C = 0

    Let X 0 = X 0 0 be chosen so A DX 0 is Hurwitz Let C be a Hermitian matrix with C C . Iteratively dene (if possible) a sequence {X } =1 via

    X +1 (A DX ) + ( A DX ) X +1 = X DX C The following is true:

    Given the choice of X 0 and C , the sequence

    {X

    } =1

    is well-dened,unique, and has X = X .

    For each 0, A DX is Hurwitz For each 1, X X .

    It remains to show that for 1, X +1 X (although it is not necessarilytrue that X 1 X 0).

    159

  • 8/13/2019 LyapunovRiccati

    31/34

    Decreasing Proof

    For 1, apply (7.31) with Y = X and Y = X 1, and substitute theiteration (shifted back one step) equation (7.34) giving

    X (A DX ) + ( A DX ) X + X DX = C (X X 1) D (X X 1)The iteration equation is

    X +1 (A DX ) + ( A DX ) X +1 = X DX C Subtracting leaves

    (X

    X +1 ) (A

    DX ) + ( A

    DX ) (X

    X +1 )

    = ( X X 1) D (X X 1)The fact that A DX is Hurwitz and D 0 implies that X X +1 0.Hence, we have shown that for all 1,

    X X +1 X

    160

  • 8/13/2019 LyapunovRiccati

    32/34

    Monotonic Matrix Sequences Convergence

    Every nonincreasing sequence of real numbers which is bounded below hasa limit. Specically, if {xk}k=0 is a sequence of real numbers, and there is areal number such that xk

    and xk+1

    xk for all k, then there is a real

    number x such thatlim

    kxk = x

    A similar result holds for symmetric matrices: Specically, if {X k}k=0 is asequence of real, symmetric matrices, and there is a real symmetric matrix such that X k and X k+1 X k for all k, then there is a real symmetricmatrix X such that

    limk

    X k = X

    This is proven by rst considering the sequence eT i X kei k=0 , which showsthat all of the diagonal entries of the sequence have limits. Then look at(ei + e j )T X k(ei + e j ) k=0 to conclude convergence of off-diagonal terms.

    161

  • 8/13/2019 LyapunovRiccati

    33/34

    Limit of Sequence Maximal Solution

    The sequence {X } =0 , generated by (7.34), has a limitlim

    X =: X +

    Remember that C was any hermitian matrix C .

    Facts:

    Since each A DX is Hurwitz, in passing to the limit we may geteigenvalues with zero real-parts, hence

    maxi

    Re i (A DX + ) 0

    Since each X = X , it follows that X + = X + . Since each X X (where X is any hermitian solution to the equality

    XDX A X XA C = 0, equation (7.32)), it follows by passing tothe limit that X + X .

    The iteration equation isX +1 (A

    DX ) + ( A

    DX ) X +1 =

    X DX

    C

    Taking limits yields

    X + A D X + + A D X + X + = X + D X + C which has a few cancellations, leaving

    X + A + A X + X + D X + + C = 0

    Finally, if X

    + denotes the above limit when C

    = C

    , it follows that theabove ideas hold, so X + X , for all solutions X . And really nally, since X + is a hermitian solution to (7.32) with C , the

    properties of X + imply that X + X + .

    162

  • 8/13/2019 LyapunovRiccati

    34/34

    Analogies with Scalar Case dx2 2ax c = 0

    Matrix Statement Scalar SpecializationD = D

    0 d R , with d

    0

    (A, D ) stabilizable a < 0 or d > 0existence of Hermitian solution a2 + cd 0Maximal solution X + x+ = a+

    a 2 + cdd for d > 0

    Maximal solution X + x+ = c2a for d = 0A DX + a2 + cd for d > 0A DX + a for d = 0A

    DX 0 Hurwitz x0 > ad for d > 0

    A DX 0 Hurwitz a < 0 for d = 0A DX Hurwitz x > ad for d > 0A DX Hurwitz a < 0 for d = 0

    If d = 0, then the graph of 2ax c (with a < 0) is a straight line with apositive slope. There is one (and only one) real solution at x = c2a .If d > 0, then the graph of dx2

    2ax

    c is a upward directed parabola.

    The minimum occurs at ad . There are either 0, 1, or 2 real solutions todx 2 2ax c = 0.

    If a2 + cd < 0, then there are no solutions (the minimum is positive) If a2 + cd = 0, then there is one solution, at x = ad , which is where the

    minimum occurs.

    If a2 + cd > 0, then there are two solutions, at x

    = a

    d a 2 + cd

    d and

    x+ = ad + a2 + cdd which are equidistant from the minimum

    ad .