Kuster GE T 2013
Kuster GE T 2013
Kuster GE T 2013
George E Kuster
Master of Science
in
Mathematics
December 7, 2012
Blacksburg, Virginia
George E Kuster
(ABSTRACT)
There is much interest in the design of feedback controllers for linear systems that minimize
the H∞ norm of a specific closed-loop transfer function. The H∞ optimization problem
initiated by Zames (1981), [23], has received a lot of interest since its formulation. In
H∞ control theory one uses the H∞ norm of a stable transfer function as a performance
measure. One typically uses approaches in either the frequency domain or a state space
formulation to tackle this problem. Frequency domain approaches use operator theory, J-
spectral factorization or polynomial methods while in the state space approach one uses ideas
similar to LQ theory and differential games. One of the key computational issues in the design
of H∞ optimal controllers is the determination of the optimal H∞ norm. That is, determining
the infimum of r for which the H∞ norm of the associated transfer function matrix is less
than r. Doyle et al, [7], presented a state space characterization for the sub-optimal H∞
control problem. This characterization requires that the unique stabilizing solutions to two
Algebraic Riccati Equations are positive semi definite as well as satisfying a spectral radius
coupling condition. In this work, we describe an algorithm by Lin et al, [16], used to calculate
the H∞ norm for the state feedback and output feedback control problems. This algorithm
only relies on standard assumptions and divides the problem into three sub-problems. The
first two sub-problems rely on algorithms for the state feedback problem formulated in the
frequency domain as well as a characterization of the optimal value in terms of the singularity
of the upper-half of a matrix created by the stacked basis vectors of the invariant sub-space
of the associated Hamiltonian matrix. This characterization is verified through a bisection or
secant method. The third sub-problem relies on the geometric nature of the spectral radius of
the product of the two solutions to the Algebraic Riccati Equations associated with the first
two sub-problems. Doyle makes an intuitive argument that the spectral radius condition
will fail before the conditions involving the Algebraic Riccati Equations fail. We present
numerical results where we demonstrate that the Algebraic Riccati Equation conditions fail
before the spectral radius condition fails.
Dedication
To my grandparents.
iii
Acknowledgments
Completing this project would have been an impossible task without the support of numer-
ous others along the way. It would be an gross understatement to say that without the
guidance and enthusiasm of my advisor, Dr. Lizette Zietsman this thesis would not have
come to fruition. Her insight, patience and ability to keep me motivated are several reasons
I remained working steadily.
I would like to thank my fellow graduate students who served as sound boards when I needed
to piece together difficult topics. Specifically Alan Lattimer and Steven Boyce played indis-
pensable roles in helping me synthesize the many topics encountered in this research.
Dr. Vasil Skenderi was the most influential undergraduate mathematics professor I had the
pleasure of taking a course with. Unknowingly, he sparked my interest in research that in-
stilled a desire to attend graduate school.
Above all however, I am grateful to my wife Lianne for her unwavering support, encourage-
ment and understanding throughout this entire process.
George Kuster
iv
Contents
1 Introduction 1
2 Background 3
2.1 State Space Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.2 State Space Solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.3 Transfer Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.4 Controllability and Observability . . . . . . . . . . . . . . . . . . . . . . . . 8
2.5 Stabilizability and Detectability . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.6 H∞ norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
5 Conclusions 35
5.1 Conclusion on State Space Formulation . . . . . . . . . . . . . . . . . . . . . 35
5.2 Extending to Mathematics Education . . . . . . . . . . . . . . . . . . . . . . 36
Bibliography 38
v
List of Figures
vi
List of Tables
vii
Chapter 1
Introduction
One of the many challenges in control theory is the design of controllers that result in
sufficient performance of a plant under various types of inputs and disturbances. H∞ control
theory was designed to reduce modeling errors and unknown disturbances in a system, while
providing quantifiable optimization of large scale multivariable problems. In application most
solutions to H∞ control problems are actually sub-optimal controllers. That is, controllers
which achieve a predetermined bound on the H∞ norm of an associated transfer function. In
this formulation of the problem, the solution is robust for a predetermined bound. However,
if one can find the H∞ norm, one can answer the question as to the existence of a controller
for a certain range of perturbations to the system.
George Zames, [23] introduced the formal H-infinity control theory by formulating the
probloem of sensitivity reduction as an optimization problem with an operator norm, in
particular, the H-infinity norm [15]. Where H∞ is the space of all bounded analytic matrix
valued functions in the open right half complex plane. This formulation was based entirely
in the frequency domain. Zames suggested that using the H-infinity norm as a performance
measure would better satisfy the demands of applications in comparison to the popular Lin-
ear Quadratic Gaussian control design [25]. This brought about the design of H-infinity
optimal control, where one aims to find a controller which stabilizes a system while minimiz-
ing the effects of disturbances in the system. This norm is now used to numerically evaluate
the controller’s sensitivity, robustness and performance of the closed loop feedback system.
Shortly afterwards John Doyle developed tools for testing robust stability in an effort to
gauge model uncertainties and achieve the H-infinity control objective of minimizing the
effect of these perturbations on the regulated outputs [18]. In 1984 Doyle, using state space
methods, offered the first solution to a general multivariable H-infinity optimal control prob-
lem [6].
While these frequency domain formulations dealt with the robustness issues, the tools devel-
oped were both notedly complicated and provided limited insight concerning the nature of
1
George Kuster Chapter 1. Introduction 2
the stabilizing solutions. State space formulations were initially developed by Doyle, Francis
and Glover in the mid 1980’s, but were only successful to a degree, that is the solutions to
these early formulations had solutions of high order, see [8] and [11].
In the late 1980’s work by Khargonekar, Petersen, Rotea and Zhou, [13], [14], and [20], pre-
sented the solution to the H-infinity state feedback problem in terms of a constant feedback
gain matrix, and provided a formula for finding this gain matrix in terms of the solution to
a Riccati equation. Furthermore, they established a relationship between linear quadratic
game theory and H-infinity optimal control and gave the solution to the state-feedback
problem by solving an algebraic Riccati equation.
In 1989 Doyle, Glover, Khargonekar and Francis, [7], presented the first general state space
solution to the H-infinity control problem in their now classic paper State space solutions
to standard H2 and H-infinity control problems. Results in this paper, give necessary and
sufficient conditions for the existence of an admissible controller in terms of solutions of
algebraic Riccati equations and a coupling condition involving a spectral radius condition.
Thus formulation resulted in algorithms that uses concave criteria as well as Newton-like
methods, see [9], [10], and [21].
This study focuses on a collection of algorithms used for H∞ -control purposes, in particular
for state feedback, [3], [4], and output feedback problems. The latter though the algorithm
presented by Lin, Wang and Xu [16].
The outline of the remainder is as follows. Chapter 2 provides the necessary background
such as controllability, observability, detectabilty, stabilizability as well as other topics such
as transfer functions and introduces the H∞ norm. Chapter 3 discusses two algorithms based
on frequency domain tools to calculate the H∞ norm of a transfer function associated with
the state feedback case. The first algorithm is the bisection algorithm presented by [3] and
the second algorithm is a two step algorithm originally presented by [4].
The main discussion is found in Chapter 4 where we discuss and implement the algorithm
Lin, Wang and Xu [16] proposed to calculate the H∞ norm of the state and output feedback
control problems. The algorithm uses a state space formulation while appealing to frequency
domain tools to find the infimum of r for which the H∞ norm of the associated transfer
function matrix is less than r. The problem of finding the H∞ norm is broken into three
sub-problems. The first two rely heavily on the algorithms discussed in Chapter 3 and the
third relies on the geometric nature of the spectral radius of the product of the two solutions
to the Algebraic Riccati Equations associated with the first two sub-problems. Concluding
remarks are found in Chapter 5.
Chapter 2
Background
Example 2.1.1. To better illustrate the state space formulation, consider the simple example
where two masses are connected by a spring and damper. The control force u is applied to a
mass M1 , which directly effects the position, z of second mass M2 . The input of this system
is the force, and the output is the change of position of the second mass M2 . Here we would
like to control the displacement of M2 .
3
George Kuster Chapter 2. Background 4
Damper
ss
!
M1
M2
Spring
!
If the control u is known, the system (2.1.1) can be solved explicitly. Before we discuss this
in more detail, we review a few basic concepts and follow the discussion in [5].
Definition 2.2.1. Given an n × n matrix A, the matrix exponential eAt is defined as:
∞
At
X (At)k
e = .
k=0
k!
1. eA0 = A0 = I.
∞ ∞ ∞
d At X d Ak tk X Ak tk−1 X Ak tk
e = = = A = AeAt
2. dt k=0
dt k! k=1
(k − 1)! k=0
k! .
The inverse Laplace transform of fˆ, denoted by L−1 (fˆ), is the function f whose Laplace
transform is fˆ.
To determine the solution to (2.2.2)–(2.2.3), take the Laplace transform of (2.2.2) and solve
for x̂. This results in
x̂(s) = (sI − A)−1 x0 + (sI − A)−1 B û(s). (2.2.4)
Taking the inverse Laplace of (2.2.4) gives
Rt
Substituting (2.2.5) into (2.2.3) yields y(t) = CeA(t−t0 ) x0 + t0
CeA(t−s) Bu(s)ds + Du(t).
Definition 2.2.3. The matrix eA(t−t0 ) is called the state transition matrix.
In the event that the control u is not predetermined, various control strategies are used to
determine the control that would result in a desired outcome. In this thesis our focus will
be on H∞ control which is described in Section ???
ŷ(s)
Substituting (2.3.2) into (2.3.3), and setting G(s) = Equation (2.3.3) becomes
û(s)
ŷ(s) = G(s)û(s)
or
ŷ(s)
G(s) = . (2.3.4)
û(s)
George Kuster Chapter 2. Background 7
The function G(s) reflects the ratio between the system output ŷ(s) and the system input
û(s) and is called the matrix transfer function or simply the transfer matrix. Entry
(i, j) in the matrix represents the transfer function from the j th input to the ith output.
This transfer function represents properties of the system and is not representative of the
magnitude or nature of the inputs. It also does not serve to provide any information about
the structure of the system.
One main advantage to using transfer functions is they allow for taking a complicated system
in a time domain and represent it as a function of a single variable in frequency domain.
Unfortunately, this approach only works for single input single output systems. This means,
for multiple input-multiple output systems, there are multiple transfer functions used to
represent each input-output relationship. It should be noted that transfer functions are not
unique, that is many systems can share the same transfer function. They are unrelated to
the systems magnitude or the nature of the input.
Definition 2.3.1. The points p at which the transfer function G(p) = ∞ are called the poles
of the system.
If G(∞) is a constant matrix the transfer function is called proper and if G(∞) = 0, the
transfer function is called strictly proper.
Let us use the spring mass damper system from Example 2.1.1 to illustrate how one would
find the transfer function of a given system described by a state space representation.
Example 2.3.1.
Simplification leads to
s+2
G(s) = .
s4 + 3s3 + 6s2
Definition 2.3.2. For the continuous time state space representation given by (2.3.1), the
matrix
G(iω) = C(iωI − A)−1 B + D (2.3.5)
is called the frequency response matrix, where ω ∈ R is the frequency.
George Kuster Chapter 2. Background 8
is called controllable if the system can be driven to any final state x1 = x(t1 ), in finite time
t1 , from any initial state x(0) by selecting the control u(t), 0 ≤ t ≤ t1 accordingly.
The controllability of the system (2.1.1), is referred to as the controllability of the pair
(A, B). The following theorem provides verifiable conditions as to whether or not a system
is controllable.
Theorem 2.4.1. Let A ∈ Rn×n and B ∈ Rn×m (m ≤ n) then the following statements are
equivalent:
(ii) The n × nm controllability matrix CM = [B, AB, A2 B, . . . , An−1 B] has full rank.
Observability and controllability of a system are dual problems and observability is defined
as follows:
Definition 2.4.2. The continuous time system (2.1.1) is said to be observable if there
exists a t1 > 0 such that the initial state x(t0 ) can be uniquely determined from knowledge of
u(t) and y(t) for all t, where 0 ≤ t ≤ t1 .
The observability of the system (2.1.1) is referred to as the observability of the pair (A, C).
The duality means that (A, C) is observable if (A,T C T ) is controllable. Due to the duality of
observability and controllability, observability has similar characterizations as controllability
in Theorem 2.4.1.
George Kuster Chapter 2. Background 9
Controllability and observability play important roles in the existence of positive definite
and positive semidefinite solutions to Lyapunov equations.
Definition 2.4.3. Let A be a stable matrix, then the matrix:
Z t1
T
CG = eAt BB T eA t dt (2.4.1)
0
The following is an example illustrating the computation of both the observability and con-
trollability Grammians.
George Kuster Chapter 2. Background 10
where
−4 −8 −12 1
A = 0 −8 −4 , B = 1 , C = 1 1 1 , D = 0.
0 0 0 1
The matrix A is stable and using the Matlab command lyap to solve the Lyapunov equa-
tions in Theorems 2.4.3 and 2.4.4 we arrive at
0.1458 0.0208 0.0208
CG = 0.0208 0.0833 0.0833 = OG .
0.0208 0.0833 0.0833
We note that this matrix is singular, and thus we can conclude that (A, B) is not controllable
and (A, C) is not observable.
Definition 2.5.1. The equilibrium state of the system (2.5.1), is the vector xe satisfying
Axe = 0 (2.5.2)
1. Stable if for each > 0 there is a δ > 0 such that ||x(t) − xe || < for all t ≥ 0 if
||x(t0 ) − xe || < δ.
2. Asymptotically stable if the equilibrium state is stable and δ > 0 exists such that
||x(t0 ) − xe || < δ implies lim x(t) = xe .
t→∞
George Kuster Chapter 2. Background 11
This means that all solutions which start sufficiently close to an asymptotically stable equi-
librium point will not only remain near that point, but will also approach equilibrium as
time increases. Thus if xe = 0 and A is nonsingular, the uncontrolled system above is
asymptotically stable if x(t) → 0 as t → ∞.
The eigenvalues if the matrix A in (2.5.1) characterizes the stability of the system.
Theorem 2.5.1. System (2.5.1) is said to be asymptotically stable if and only if all of the
eigenvalues of the matrix A have strictly negative real parts.
Definition 2.5.3. The matrix A is called a stable matrix if all of the eigenvalues of A
have negative real parts.
As in the case of controllability, the solution to a Lyapunov equation are also characterized
in terms of the stability of (2.5.1).
Theorem 2.5.2. The system:
ẋ(t) = Ax(t),
is asymptotically stable if and only if, for any symmetric positive definite matrix M, there
exists a unique solution to the equation
XA + AT X = −M.
We consider the problem of finding a controller u which drives this system to a desired state.
That is, find a feedback matrix K such that the feedback matrix (A − BK) is stable. Not
every system can be driven to a specific state, so the importance of this section is that is
presents necessary and sufficient conditions for when a stabilizing feedback matrix exists.
Also suppose we have full information from the state, x(t), that is, at any time t information
about any variable from the state is available. Using linear control, the controller u(t) is of
the form,
u(t) = −Kx(t), (2.5.3)
where K is the feedback matrix.
ẋ(t) = (A − BK)x(t),
(2.5.4)
y(t) = (C − DK)x(t).
Recall that the uncontrolled system (2.5.1) is stable when the matrix A is stable. Thus
stabilizing the system (2.5.4) reduces to finding a matrix K such that (A − BK) is stable. If
George Kuster Chapter 2. Background 12
such a K exists, it is called a stabilizing feedback matrix and we the system is stabilizable.
The pair of matrices (A,B) is known as the stabilizable pair.
The following theorem provides necessary and sufficient conditions for determining if a sta-
bilizing feedback matrix exists for a particular system.
Theorem 2.5.3. Just as in the case of stabilizability, the following statements are equivalent:
3. For all λ and x 6= 0 such that x∗ A = λx∗ and Re(λ) ≥ 0 one have x∗ B 6= 0.
Corollary 2.5.1. If the pair (A,B) is controllable, then the pair is also stabilizable.
Definition 2.5.4. A system is said to be detectable if there exists a matrix L such that that
given the matrix pair (A, C), A − LC is stable.
1. (A, C) is detectable.
λI − A
2. The matrix has full rank for all Re(λ) ≥ 0.
C
4. (AT , C T ) is stabilizable.
It can be shown that is a system is controllable then the closed loop eigenvalues can be as-
signed arbitrarily (eigenvalue assignment problem). The issue with assigning the eigenvalues
arbitrarily is that there is no methodology to finding an optimal controller such that the
design characteristics are met as efficiently as possible.
This led to the development of numerous other optimal control strategies such as Linear
Quadratic optimal control. The objective of the linear quadratic control is to find a control
u which minimizes some predetermined quadratic cost function. Quadratic optimal control
is beyond the scope of this study. However, note that [7] presents a parallel development of
H∞ control and H2 optimal control which is known as Linear Quadratic Gaussian control.
George Kuster Chapter 2. Background 13
2.6 H∞ norm
First note that the transfer function of any system modeled by linear time invariant ordinary
differential equations is rational and has real valued coefficients as noted by [17] and [1]. Thus,
transfer functions of any such system have poles contained entirely in the left half plane and
are analytic in the right half plane. A space of such functions is called a Hardy space. We
shall denote this space as H∞ . For any multivariable system, the transfer function G(iω) is
a matrix. The singular values of a matrix A, σj (A) are definted as
σj (A) = λj (AAT ),
Then we have
||G(iω)|| = max σi (G(iω)) = σmax (G(iω)).
i
Definition 2.6.1. The H∞ (induced) norm of the transfer function G(s), denoted as ||G||∞
is defined as
||G||∞ = sup σmax (G(iω)), (2.6.1)
ω∈R
where sup denotes the supremum or least upper bound over all real valued frequencies ω.
ω∈R
This provides another way to characterize the stability of a system, namely if the systems
transfer function G(s) ∈ H∞ then the system is stable. As described in [15] for the case
when G is rational, G ∈ H∞ if and only if G has no poles in the closed right half plane.
Chapter 3
Recall the transfer function G(s) defined from coefficient matrices of the system (3.1.1)
G(s) = C(si − A)−1 B + D.
Theorem 3.1.1. Let G(s) be a stable transfer function and let r > 0. Then ||G||∞ < r if
and only if σmax (D) < r and Mr has no purely imaginary eigenvalues.
Theorem 3.1.1 relates the eigenvalues of Mr to the singular values of G(s). This theorem
implies that for r values larger than r∗ = ||G(s)||∞ , Mr will have no purely imaginary
eigenvalues. Further, this implies that if r < r∗ , Mr will have at least one purely imaginary
eigenvalue.
14
George Kuster Chapter 3. Algorithms: Frequency Domain 15
Definition 3.2.1. The Hankel singular values are the square roots of the eigenvalues of the
matrix CG OG , where CG and OG are the controllability and observability grammians, defined
by Equations 2.4.3 and 2.4.4 respectively.
Hankel singular values, denoted σHi following convention, are ordered in descending order
such that σH1 is the largest singular value.
The method for calculating rlb and rub from Step 1 (3.2.2), is known as the Enns-Glover
formula. We note that the authors suggest alternative formulas for computing the boundary
values for the bisection algorithm which can be computed with the following formulas:
n p o
rlb = max σmax (D), Trace(OG CG )/n ,
p (3.2.1)
rub = σmax (D) + 2 nTrace(CG OG )
Algorithm 3.2.1. Given system (3.1.1) with coefficient matrices A,B,C,D where A is stable
and error tolerance > 0.
Step 1: Compute the bounds for the bisection algorithm, rlb and rub , where
P σH1 }
rlb = max{σmax (D),
(3.2.2)
rub = σmax (D) + 2 nj=1 σHi
Step 2: Set r = (rub + rlb )/2
If 2(rub − rlb ) < , end.
Step 3: Compute Mr .
Step 4: Check for purely imaginary eigenvalues of Mr .
If Mr has a purely imaginary eigenvalue, then set rlb = r.
Else set rub = r.
By Theorem 3.1.1 if we find that Mr has no purely imaginary eigenvalues, we can take
r to be the new upper bound for the algorithm. Similarly if we find that Mr does have
purely imaginary eigenvalues, we can take r as the new lower bound. In order to ensure
George Kuster Chapter 3. Algorithms: Frequency Domain 16
the computed value for r∗ is truly an accurate approximation, we stop the iterations when
2(rub − rlb ) < .
It should be pointed out that computing the starting bounds for this algorithm requires
solving two Lyapunov equations (2.4.3) and (2.4.4), multiplying two n × n matrices and
computing the eigenvalues of this product. For large matrices these computations are com-
putationally expense and are impractical. Also, this algorithm relies on very accurate com-
putations of eigenvalues, which requires very specific eigenvalue solvers, see for example [2].
We must first compute OG and CG . Using the lyap command in Matlab , we attain:
0.1250 0 0.1250 0.1250 0 0
OG = 0 0.0625 0 , CG = 0 0 0 .
0.1250 0 0.1250 0 0 0
−4.0000 −8.0000 12.0000 15.6735 0 0
0 −8.0000 0 0 0 0
0 0 −16.0000 0 0 0
Mr =
−1.0000 −1.0000 −1.0000 4.0000 0 0
−1.0000 −1.0000 −1.0000 8.0000 8.0000 0
−1.0000 −1.0000 −1.0000 −12.0000 0 16.0000
The eigenvalues of Mr are {8.0000, 16.0000, −0.5714, 0.5714, −8.0000, −16.0000} Since none
of the eigenvalues are purely imaginary we set r = rub and continue the iterations.
After 20 iterations we find that rlb = 0.2499 and rub = 0.2500, thus the error bound is
satisfied and the process is terminated. We arrive at ||G(s)||∞ ≈ .2500.
George Kuster Chapter 3. Algorithms: Frequency Domain 17
As a consequence of this theorem, the imaginary eigenvalues of Mr are exactly the frequen-
cies where some singular value of G(ωi) = r. That is ωi is an eigenvalue of Mr if and only
if r is a singular value of G(ωi) where ω ∈ R. In Figure 3.1 ωi represents the frequency, r
represents the iterations and the mi represent the midpoint between each ωi and ωi+1 . Here
we see Theorem 3.3.1 allows for relating the purely imaginary eigenvalues back to multiple
values of r. Specifically each mi relates back to a value for r. In Figure 3.1 we can see that
the next iteration of r is taken to be the maximum of the singular value of G(s) for each mi .
The algorithm by Bruinsma and Steinbuch builds on the ideas of the bisection algorithm,
but uses Theorem 3.3.1 to search for ||G(s)||∞ using multiple values of ω. The bisection algo-
rithm only searches at one frequency per iteration, where as the two step algorithm searches
multiple frequencies per iteration. This is the main advantage to using this algorithm. The
authors claim this algorithm converges quadratically (they also claim ri < ri+1 , though no
proof of this could be determined).
!!!!!!
!!
!
!(
!!
!"
(!"
)
!!!!
)
!!
!!
!!
!!
!!
!!
!!
Figure 3.1: Relationship between singular values and eigenvalues
George Kuster Chapter 3. Algorithms: Frequency Domain 18
If G(s) has poles which are strictly real valued, then λi is the pole which maximizes | λ1i |.
As is the case with the bisection algorithm, this method is for systems where A is stable.
Algorithm 3.3.1. Given system (3.1.1) with coefficient matrices A,B,C,D where A is stable
one can find the H∞ norm using the following steps.
Step 1.1: Compute the transfer function G(s) of the given system.
Step 1.2: Find the poles of G(s).
If the poles are strictly real valued, set ωp = max|λi |.
Else set ωp = λi , where λi maximizes (3.3.2).
Step 1.3: Compute rlb using (3.3.1).
Step 2.1: Compute r = (1 + 2)rlb .
Step 2.2: Compute Mr from (3.1.2).
Sort all of the purely imaginary eigenvalues of Mr , and label them in descending
order ω1 , . . . , ωk .
If Mr has no purely imaginary eigenvalues, set rub = r and proceed to Step 3.
Step 2.3: For i = 1 to k − 1
(i) Compute mi = 21 (ωi + ωi+1 ).
(ii) Compute σmax (G(mi i)) = svdi .
Set rlb = max(svdi ).
Step 3: Compute kGk∞ = 21 (rlb + rub ).
Since the H∞ norm is defined as the maximum singular value along the imaginary axis, we
can sample along the imaginary axis to find the lower bound for the H∞ norm. To ensure
kGk∞ can be calculated using the two step algorithm, one must carefully choose a lower
bound for the algorithm.
After computing the eigenvalues of Mr , one must determine if r is an upper or lower bound,
this is accomplished using Theorem 3.1.1, similar to the bisection algorithm. However, if one
finds purely imaginary eigenvalues ω1 , ω2 , . . . , ωn , by Theorem 3.3.1 we know r is a singular
value for each G(ωi ). This implies there exists frequencies m1 , . . . , mn−1 , each mi between
its respective ωi , ωi+1 such that for some mi , G(ωi) < G(mi ). By computing σmax (G(mi )
for all i, the algorithm searches over a much larger area than the bisection algorithm. After
finding σmax (G(mi ) for all i, we set this value equal to rlb , and start the next iteration.
George Kuster Chapter 3. Algorithms: Frequency Domain 19
The poles of G(s) are the eigenvalues of A. We find the eigenvalues of A to be {−3.3589 +
0.2858i, −0.3628 + 0.9534i, −1.2784 − 0.2392i}. Thus A is stable.
Next we find that −0.3628 + 0.9534i, maximizes (3.3.2) and using (3.3.1) we
calculate rlb = 0.9907.
Using the above value for rlb we find,
−1 + i −1 −1 1.0188 0 0
−1 −2 + i −1 0 0 0
−1 −1 −2 − i 0 0 0
Mr = −1
.
−1 −1 1+i 1 1
−1 −1 −1 1 2+i 1
−1 −1 −1 1 1 2−i
−3.1927 + 0.2164i,
3.1927 + 0.2164i,
−1.3224 − 0.1563i,
1.3224 − 0.1563i,
−0.0000 + 0.8597i,
0.0000 + 1.0201i
We find and order the purely imaginary eigenvalues, w1 = −0.0000 + 0.8597i and
w2 = 0.0000 + 1.0201i. Finding the average of these two gives m1 = 0.0000 + 0.9399i.
The maximum singular value of G(m1 i) = 1.0122 = rlb .
We then start the next iteration since there are strictly imaginary eigenvalues of Mr . After
a total of four iterations we arrive at kGk∞ = 1.0121.
Remark: We note one major issue which arrises in implementing this algorithm is the need
for an accurate eigenvalue solver since the depends on the decision if the eigenvalues are
purely imaginary or not.
Chapter 4
Doyle et al [7] presented a state space characterization for the suboptimal H∞ control prob-
lem. This characterization includes the requirement of the unique stabilizing solutions to two
Algebraic Riccati Equations to be positive semidefinite as well as a spectral radius coupling
condition.
Theorem 4.1.1. Let (A, B) be stabilizable and (A, Q) be detectable, then the Hamiltonian
matrix H, has n eigenvalues with negative real parts, no eigenvalues on the imaginary axis
and n eigenvalues with positive real parts. Further the associated Riccati equation has a
20
George Kuster Chapter 4. Algorithms: State Space Approach 21
unique stabilizing (symmetric) solution X, and the eigenvalues of A − BK are the stable
eigenvalues of H.
we can set X = T2 T1−1 . Then X is uniquely determined by H and we say that H ∈ dom(Ric).
If H ∈ dom(Ric) then H has no purely imaginary eigenvalues and the matrix T1 is non-
singular. That is Ric : H → X and Ric(H) = X. One can verify a given H exists in
dom(Ric) with the following lemma from Doyle et al [7].
1. X is symmetric;
3. A − SX is stable.
− r12 BB T
A
H(r) = . (4.1.2)
−C T C −AT
XA + AT X + Q − XSX = 0,
1
where S = r2
BB T and Q = −C T C.
George Kuster Chapter 4. Algorithms: State Space Approach 22
Recall that the Hamiltonian matrix Mr , is connected to ||G||∞ through Theorem 3.1.1.
Thus, H(r) has no purely imaginary eigenvalues when ||G||∞ < r. Theorem 4.1.1 connects
the Hamiltonian matrix H(r) with its associated ARE and provides conditions under which
H(r) ∈ dom(Ric), or has no purely imaginary eigenvalues. Theorem 4.1.1 further connects
||G||∞ with the existence of a stabilizing solution X. Specifically it lays the groundwork for
determining necessary and sufficient conditions for the existence of an admissible controller
in a state space formulation.
That is if (A, B) is stabilizable and (A, Q) is detectable, then H(r) ∈ dom(Ric), hence there
exists matrices T1 and T2 such that,
T1
χ− (H) = Im ,
T2
Output Feedback
We now consider the output feedback case. A finite-dimensional time-invariant linear continuous-
time dynamical system may be described using the following system of first-order ordinary
differential equations:
ẋ(t) = Ax(t) + B1 w(t) + B2 u(t)
z(t) = C1 x(t) + D12 u(t) (4.1.3)
y(t) = C2 x(t) + D21 w(t).
Here x(t), y(t), z(t) are the state, sensed output and controlled output vectors respectively
and u(t), w(t) are the control and disturbance inputs respectively.
Assume throughout the discussion that follows that the matrices A, B1 , B2 , C1 , C2 , D12 , D21
are real and satisfy the following assumptions:
The two Algebraic Riccati Equations associated with the system are:
1
XA + AT X − X(B2 B2T − B1 B1T )X + C1T C1 = 0, (4.1.6)
r2
and its dual,
1 T
AY + Y T A − Y (C2T C2 − C C1 )Y + B1 B1T = 0. (4.1.7)
r2 1
The four assumptions (A1)—(A4) are standard for control applications. They were originally
formulated for H2 control and since H∞ control was developed in parallel with H2 control
these assumptions followed [7]. Assumption (A1) is necessary for the existence of a stabilizing
controller in the form u = Ky. Assumption (A2) is sufficient for the closed loop system to
be both stabilizable and detectable. If the closed loop system is not stabilizable, then it is
certainly not controllable and therefore the H∞ optimal control problem has no solution.
Assumptions (A3) and (A4) imply there is no overlap between the cost of the control input
and the state. In other words the cost penalty applied to z contains a non-singular and
normalized cost penalty on the control input u. Assumptions (A3) and (A4) can be relaxed
but they are made in an effort to obtain cleaner formulas.
Recall the Hamiltonian matrix H(r) given by (4.1.2), associated with system (4.1.1). It
is important to note the difference between H(r) and H∞ (r) given by (4.1.4). Specifically
the ability to make r12 B1 B1T arbitrarily large or small no longer guarantees a sign definite
matrix in the [1, 2] block of H∞ (r). As a consequence one is not guaranteed the existence
of a stabilizing solution to the ARE (4.1.6) even when H∞ (r) ∈ dom(Ric). In other words
H∞ (r) ∈ dom(Ric) is still a necessary condition for the existence of a stabilizing solution,
but it is now no longer a sufficient condition.
Doyle et al [7] shows that system (4.1.3) which satisfies (A1)—(A4), has an admissible
controller K, such that ||Tzw ||∞ < r if and only if the following necessary and sufficient
conditions are met:
where Tzw denotes the transfer function from the disturbance w to the controlled output z,
such that
Tzw = C1 (sI − A + B2 K)−1 B1 + D12.
Lin, Wang, Xu [16] present an algorithm to find ||Tzw || using a state space approach. The
algorithm breaks the three conditions (C1)—(C3) into three seperate sub-problems.
1. Finding the smallest r value for which (C1) holds rx , is performed first. In doing so, rx
is distinguished by the singularity of the T1 matrix created by the basis vectors of the
stable invariant subspace of H∞ (r). Using a reformulation (in terms of singularities) of
(C1) a secant method is applied (rather than a Newton type method) directly utilizing
the stable invariant subspace of H∞ (r).
2. The same technique is then applied to J∞ (r), to find ry , the smallest value for r such
that (C2) holds.
3. The geometric nature (strictly decreasing) of ρ(X(r)Y (r)) is then exploited to find r∗ ,
the value for which all three conditions (C1)—(C3), hold.
To find rx the smallest value such that condition (C1) holds, we must first find the interval
such that H∞ (r) has no purely imaginary eigenvalues. It is a well known result that there
is only one such interval, (rlb , ∞) [10], [12], [21]. Furthermore it is a well known result that
rlb = ||GZ ||∞ , where
GZ (s) = B1T (sI − (A − ZC1T C1 )T )−1 C1T ,
with Z = Ric(H2 ) and
AT −C1 C1T
H2 = .
−B2 B2T −A
Thus to find rlb we can apply either the bisection or two step method to the transfer function
G(iω) = B1T (iωI − (A − ZC1T C1 )T )−1 C1T . (4.1.8)
Alternatively a bisection approach could be used directly on H∞ (r). Upon calculating the
eigenvalues of H∞ (r), one could verify Re(λi ) 6= 0 and im(λi ) = 0 for 1 ≤ i ≤ n. If
Λ(H∞ (r̃)) ∩ C0 = ∅ then r̃ < rlb and if Λ(H∞ (r)) ∩ C0 6= ∅ then r̃ < rlb .
To satisfy the second condition of (C1), X(r) ≥ 0 we note that as a consequence of the
existence of only one interval (rlb , ∞) such that Λ(H∞ (r)) ∩ C0 = ∅ then once rlb is found
H∞ (r) has no purely imaginary eigenvalue for all r > rlb . Then one can find matrices T1 and
T2 which correspond to the stable eigenvalues of H∞ (r) where
T1 (r)
χ− (H∞ )(r) = Im ,
T2 (r)
and if T1 (r) is non-singular set X(r) = T2 (r)T1 (r)−1 . When such an X exists and is pos-
itive semi-definite for some r̃ then it positive semi-definite for all r > r̃, and is monotonic
decreasing by the following theorem from [12].
George Kuster Chapter 4. Algorithms: State Space Approach 25
Theorem 4.1.1. If a non-zero X(r̃) = Ric(H∞ (r̃)) exists and is positive semidefinite for
some r̃ > 0, then X(r) = Ric(H∞ (r)) exists and is positive semidefinite for all r ∈ [r̃, ∞)
and obeys the partial order 0 ≤ X(r2 ) ≤ X(r1 ) for r̃ ≤ r1 ≤ r2 < ∞.
Assume 0 < rx < ∞, Theorem 4.1.1 implies the existence of a least r such that X(r) ≥ 0,
where A ≥ 0 denotes the matrix A is positive semi-definite. As a result, one could implement
a bisection algorithm to find such an r noting if X(r) ≥ 0 then r > rx and if X(r) 0 then
r < rx . The following theorem from Lin et al (1998), is the keystone to this, as it provides
the means in which one can implement a bisection or secant method to search for this r. In
fact the following theorem is a reformulation of (C1) from [7].
Theorem 4.1.2. For r > rlb , let X(r) = Ric(H∞ (r)) whenever H∞ (r) ∈ dom(Ric). Suppose
that X(r2 ) ≥ 0 and X(r1 ) 0 such that r2 > r1 > rlb . Then the following are true:
(i) There exists at least one r ∈ (r1 , r2 ) such that T1 (r) is singular.
(ii) Let
rs = sup{r|T1 (r) is singular}. (4.1.9)
Then X(r) ≥ 0 for r > rs and X(r) 0 for rlb < r < rs , when X(r) exists.
Proof(outline)
(i) Assume that T1 (r) is nonsingular for all r ∈ (r1 , r2 ).
• By the monotonicity result in Theorem 4.1.1, we have 0 ≤ X(r) ≤ X(r̃), which implies
the smallest eigenvalue of X(r) is zero and further this applies to all r ≥ r̃.
• However this contradicts the fact that X(r2 ) ≥ 0, thus there exists an r ∈ (r1 , r2 ) such
that T1 (r) is singular.
(ii) Suppose there exists α > rs where X(α) 0, then by Theorem 4.1.1, we have α < r2 .
Thus from (i) there must exist β ∈ (α, r2 ) with X(β) singular. This contradicts the definition
of rs , thus X(r) ≥ 0 for all r > rs .
On the other hand, suppose there exists γ ∈ (rlb , rs ) with X(γ) ≥ 0, then by Theorem 4.1.1
we have X(r) exists for all r ≥ γ. However this is a contradiction because X(rs ) does not
exist. Thus X(r) 0 for r ∈ (rlb , rs ).
Assume that rx < ∞, and rlb < rs .
Then there exists at least one r ∈ (rlb , ∞) such that T1 (r) in,
T1 (r)
χ− (H∞ (r)) = Im ,
T2 (r)
George Kuster Chapter 4. Algorithms: State Space Approach 26
Thus the problem of finding the smallest value for r such that (C1) holds is now reformulated
into finding rs , the smallest value such that X(r) ≥ 0 for r > rs where rs ≥ rlb > 0.
Note rlb ≤ rs and H∞ (r) ∩ C0 = ∅ for all r > rlb , then by assumptions (A1) and (A2)
and Theorems 4.1.1 and 4.1.2, the matrices T1 and T2 are defined, T1 is non singular and
T2 (r)T1−1 (r) = X(r) ≥ 0 for all r > rs . Thus rs can be approximated with a bisection
strategy, checking the positive semi-definiteness of the matrix T2 (r)T1−1 (r). Explicitly, if
T2 (r)T1−1 (r) ≥ 0 then r > rs and if T2 (r)T1−1 (r) 0 then r < rs .
To increase the rate of convergence,one could apply a variant secant method to search for rs
as well. By Theorem 4.1.2, rs is the largest value such that T1 (r) is singular, then our goal
is to find the largest r such that f (r) = 0, where
However singular value plots are non-negative functions and by Theorem 4.1.2 f (r) may
have multiple zeros. For this reason the secant method is not guaranteed to converge using
f (r).
Thus we actually apply the secant method to the new smooth function
−σmin (T1 (r)) if r ≤ rs
f (r) = ,
σmin (T1 (r)) if r ≥ rs
that is, f along with its reflection over the r axis for r < rs . One issue which arrises from
this new function is now one must deduce which case to use, that is f or −f . To make this
determination we appeal once more to Theorem 4.1.2, and check if X(r) = T2 (r)T1−1 (r) ≥ 0
by computing the eigenvalues of the matrix T1T (r)T2 (r) which is congruent to T2 (r)T1−1 (r)
and avoids the inverse calculation for T1 (r).
This leads to an algorithm for the state feedback case. The following is an outline of the
algorithm.
Algorithm 4.1.1. This algorithm computes rx , the smallest value for r such that condition
(C1) holds.
Step 1:
Attain values for r− and r+ , using equations 4.1.10 - 4.1.12.
Compute ΛH∞ (r̃), the eigenvalues of H∞ (r̃), where r̃ = 21 (r− + r+ )
George Kuster Chapter 4. Algorithms: State Space Approach 27
!(!! )
r1 rnew
r2
rs
−!(!! )
Step 2:
Determine the next iterations interval [r− , r+ ].
Step 2.1
If the matrix H∞ (r̃) has purely imaginary eigenvalues then set r− = r̃ and move to Step 3.
Step 2.2
If the matrix H∞ (r̃) has no purely imaginary eigenvalues then compute the eigen-matrix
[T1 T2 ]T and set case = 1.
If T2 T1−1 is positive semidefinite, then set r+ = r̃.
If T2 T1−1 is not positive semidefinite, then set r− = r̃.
If T2 T1−1 is not positive semi definite set r+ large and return to Step 1.
We will now explore the complexity of each of the steps above, one at a time.
Step 1
To begin the algorithm one must find an initial interval on which to search for rx , the smallest
value for r for which (C1) holds. In practical cases assuming B2 is not singular, to compute
the upper bound of this interval we can set
σmax (B1 )
r+ = , (4.1.10)
σ0 (B2 )
where σmax is the largest singular value and σmin is the smallest singular value.
Much like the previous algorithms, one could use the Hankel singular values for the upper
and lower bounds as well, where
r− = σH1 , (4.1.11)
X
r+ = 2 σHi . (4.1.12)
i
Note However depending on the system it might be more efficient to set r− = 0 use 4.1.10
for r+ .
Note, if one computes rlb using the two step method or the bisection method on the system
response matrix GZ (iω), 4.1.8 then Step 2.1 below can be avoided since ||GZ ||∞ = rlb , thus
establishing the interval on which ΛH∞ (r) ∩ C0 = ∅ before starting this algorithm.
Step 2
In this step we determine the next upper bound or lower bound of the interval by checking
if the value for r̃ is above or below rx . This involves testing values of r to verify they satisfy
George Kuster Chapter 4. Algorithms: State Space Approach 29
condition (C1), ΛH∞ (r) ∩ C0 6= ∅ and Ric(H) ≥ 0. We must first find rlb , then we seek rs .
We note that rs may or not be the same value as rlb .
Since there exists only one such interval, if ΛH∞ (r) ∩ C0 6= ∅ then r̃ < rlb and we set r̃ = r− .
Doing so requires accurately computing the eigenvalues of a 2n × 2n Hamiltonian matrix.
Error in these calculations could severely effect the choice of r− since it must be determined
if Re(λi ) < T ol and im(λi ) > T ol for all 0 ≤ i ≤ n.
Since there can be values for r such that the Hamiltonian matrix H∞ (r) may have no eigen-
values on C0 , but X(r) 0, we must test the positive semi-definite property of T2 T1−1 (r̃).
Here we use the singular values of T1 (r̃) to determine if the value for r̃ is greater than or less
than rx . Theorem 4.1.1 implies that one can impliment a bisection algorithm in the search
for rx and Theorem 4.1.2 provides the means to implement this search, namely searching
over the singular values of T1 (r̃).
There are two major components to this step. First one must compute the 2n×n eigenmatrix
T1 (r̃)
,
T2 (r̃)
Step 3
In step three of the algorithm we update the next iterate of r̃. Here we use either the bisec-
tion method or secant method. We determine which method to use based on whether or not
George Kuster Chapter 4. Algorithms: State Space Approach 30
(r+ − r̃) is less than some predetermined tolerance, tol. If the difference is less than tol, we
use the secant method and if the difference is larger than tol we use the bisection method to
speed up the iterative process.
The bisection algorithm is standard, that is if (r+ − r− ) > tol set r̃ = 21 (r− + r+ ) and proceed
to Step 4 to check the stop criteria.
The secant method can be applied to both Step 2.1 (case = 0) and Step 2.2 (case =1),
however it is applied to different functions for the for the two different cases since the goal
for each case is different. For the Step 2.1, that is Λ(H∞ (r̃)) ∩ C0 6= ∅, where we let r̃ = r+
such that,
(r+ − r− ) min |λi (H∞ (r− )) − λj (H∞ (r− ))|
i6=j,i,j∈I0
r̃ = . (4.1.13)
min |R(λk (H∞ (r+ )))|+) min |(λi (H∞ (r− )) − λj (H∞ (r− ))|
1≤k≤2n i6=j,i,j∈I0
Step 4
There are two types of stop criteria used in this algorithm for each case. The first stop
criterion deals with the distance between the previous and the next iteration of r, ∆r̃. The
George Kuster Chapter 4. Algorithms: State Space Approach 31
second stop criterion deals with the magnitude of the functions 4.1.14 and 4.1.15. The two
stop criteria are necessary since ∆r̃ → 0 does not directly imply r → rx . Thus we combine
the convergence of ∆r̃ with the value of equations 4.1.14 or 4.1.15 accordingly.
Suppose the r̂ = r̃ + ∆r̃ is the updated iterate of r̃. For both cases the stop criteria ∆r < tol
was used, were tol = 10−6 .
For the function convergence in case=0, we are concerned with the eigenvalues of the Hamil-
tonian matrix, H∞ (r̃). The secant method uses equation 4.1.14. Since we have g(r) → 0 as
r → rlb then we want |g(r)| < tol and move to Step 5.
For case =1 we are concerned with the singular vales of T1 . For this case we use equation
4.1.15 for the secant method. Since we want to find the smallest value of r such that T1 is
singular we want f (r) = 0, thus we stop when |f (r)| < tol and move to Step 5.
Step 5
It is important to note that the vast majority of the computations in the algorithm rely
on the various matrices being well conditioned. The computation of T2 (r)T1 (r)−1 , is not
trivial and relies on the numerical computations being accurate. This accuracy depends on
the condition of the matrices T2 (r) and T1 (r). Further the secant method applied to 4.1.15
may have converged to an incorrect value of r. Thus we must double check the algorithm
converged to the correct value and also verify the condition of T1 (r) since it must be inverted.
In doing so the actual value for rx may not be attained, but rather a more conservative value
will found.
To ensure convergence of the algorithm to rx = rs we must verify if T2 (r̃)T1−1 (r̃) ≥ 0. If this
holds then we check if T1 (r̃) is well conditioned. If the matrix is well conditioned we stop
and accept r̃ = rx .
On the other hand if T1 (r̃) is ill-conditioned, we must find the smallest perturbation of r̃
such that T1 (r) is well conditioned.
If T2 (r̃)T1−1 (r̃) 0 then we know r̃ < rx thus set r− = r̃ and r+ sufficiently large and return
to Step 1.
It is for this reason that in implementing this algorithm, it would be immensely beneficial to
accurately calculate rlb before hand. The bisection algorithm presented earlier would provide
an approximation for rlb which could then be used as the lower bound in Algorithm 4.1.1.
It should be noted that (C1) and (C2) are dual problems of each other thus the above algo-
rithm is also used to find the value of ry , that is the smallest value of r such that condition
(C2) holds, by replacing A, B1 , B2 and C1 with AT , C1T , C2T and B1T respectively.
From this we see that rm is a lower bound for r∗ , the value for which (C1)—(C3) hold si-
multaneously.
To find r∗ we take advantage of the geometrical natures of ρ(r) and r2 . Specifically r2 is and
increasing function of r and ρ(r) is decreasing function of r [12]. One important advantage
of using this geometrical nature is that an upper bound can be easily found.
In implementing a bisection algorithm to approximate the value of r such that (C3) holds,
we note that ρ(r) > r2 for all r < r∗ and ρ(r) < r2 for all r > r∗ . Thus if ρ(rm ) ≤ rm
2
, then
∗ 2
r = rm and we stop. If howeverp ρ(rm ) > rm , by considering the nature of the two curves
we have r∗ > rm and further ρ(rm ) must be an upper bound for r∗ thus we set r+ = rm .
To numerically implement this, we use the standard bisection or secant methods on the
function
f (r) = ρ(r) − r2 .
We stop the iterations when −f (r) < tol, that is when r2 − ρ(r) is positive and sufficiently
small. The following algorithm can be used to compute r∗ , the smallest value of r such that
conditions (C1)—(C3) hold simultaneously.
Algorithm 4.1.2.
Step 1: Find rx and ry using Algorithm 4.1.1.
Step 2: Set r̃ = rm .
Step 3: Compute ρ(r̃).
If ρ(r̃) ≤ r2 , set r∗ = r̃ and stop. p
If ρ(r̃) > r2 , then set r− = r̃ and r+ = ρ(r̃).
Step 4: Update r̃.
If r+ − r− ≥ tol, use the bisection method.
If r+ − r− < tol use the secant method applied to the function ρ(r) − r2 .
Step 5: Stop Criteria
If −f (r̃) < tol holds then set r∗ = r̃ and stop
Else return to Step 4.
We note that the number of iterations for the examples is much higher than those found
in [16]. This is due to their use much more efficient and accurate numerical solvers, in com-
parison to the use of basic MATLAB commands here. It is noted however that although
the tools used here could be considered inferior, the algorithm still converged to the correct
values.
Example 1
Consider the system,
ẋ(t) = Ax(t) + B1 w(t) + B2 u(t),
z(t) = C1 x(t) + D12 u(t),
y(t) = C2 x(t) + D21 w(t),
This shows the importance of using accurate solvers quite remarkedly. Even though rlb is
approximately equal to rx and ry , in both cases the algorithm required a significant number
of iterations to converge.
Example 2
Consider system 4.1.3 with the following coefficient matrices,
−1 1
A= , B1 = 0 I , B2 = I,
0 4
I I
C1 = , C2 = I, D12 = , D21 = 0 I .
0 0
Here we see the value of rlbx need not equal rx . This also shows that the r value with
satisfies (C1) and (C2) is a lower bound for r∗ , but need not be close to r∗ .
George Kuster Chapter 4. Algorithms: State Space Approach 34
The following example was presented as a counter example to the arguement in [7] that (C3)
would fail before (C1) or (C2).
Example 3
Now consider the system with the following coefficient matrices,
− 12 − 21
1 3
− 75 7
− 10
5 2
A= 3 , B2 = , C2 = 7 ,
0 10
− 32 0 − 12 − 10
T 1 0 0 0 T 0 0 1 0
B1 = C1 = , D12 = D21 = .
0 1 0 0 0 0 0 1
Conclusions
35
George Kuster Chapter 5. Conclusions 36
is an impression that researching areas in applied mathematics isn’t beneficial for one who
will ultimately study mathematics education. After this experience, I realize this couldn’t
be further from the truth. Not only does completing a Masters Thesis play a key role in
preparing a student for independent research, but further it instills a true appreciation for
what it means to learn and understand a concept. Further, I have gained an appreciation
for my own learning process which is certainly beneficial to understanding how others learn.
Constructivism in Research
Research at the Masters level is a far more independent venture than one typically gets to
experience as an undergraduate. The advisor’s role is to ensure the student stays on track
and their work is appropriate for the task presented. The advisor does not explain or lecture
the student topics concerning the material past the point of redirection or filling gaps in the
student’s understanding. Hopefully, input from the student is met with interest, apprecia-
tion and consideration. That is, the student’s input is valued. The use of more guidance
and less lecture results in the student forming much more personal knowledge concerning
the subject, compared to an assimilation of the lecturer’s knowledge.
Deeper Understanding
Often times courses are structured in a linear fashion. That is, the course has been designed
to present topics in a predetermined order. This order may or may not be consistent with
the order in which topics would naturally fall and as a consequence may lack motivation.
For illustration purposes only, it is customary to teach integration before infinite series in
many university calculus sequences. Though it may be more natural to learn sequence and
series before integration. While actively participating in research one must learn topics as
they arise, not as they are presented. This provides a significant increase in motivation to
learn these topics
Overall completing this Masters Thesis had a large impact on not only my current under-
standing of control theory, but also on my future work as a mathematics education researcher.
Bibliography
[1] T. Başar and P. Bernhard. H-infinity optimal control and related minimax design prob-
lems: a dynamic game approach. Birkhäuser Boston, 2008.
[2] P. Benner, V. Mehrmann, and H. Xu. A numerically stable, structure preserving method
for computing the eigenvalues of real hamiltonian or symplectic pencils. Numerische
Mathematik, 78(3):329–358, 1998.
[3] S. Boyd, V. Balakrishnan, and P. Kabamba. A bisection method for computing the H∞
norm of a transfer matrix and related problems. Mathematics of Control, Signals, and
Systems (MCSS), 2(3):207–219, 1989.
[4] NA Bruinsma and M. Steinbuch. A fast algorithm to compute the H∞ norm of a transfer
function matrix. Systems & Control Letters, 14(4):287–293, 1990.
[5] B. Datta. Numerical methods for linear control systems. Academic Press, 2003.
[6] J.C. Doyle et al. Advances in multivariable control. In Lecture Notes at ONR/Honeywell
Workshop. Minneapolis, MN, 1984.
[7] J.C. Doyle, K. Glover, P.P. Khargonekar, and B.A. Francis. State-space solutions to
standard H2 and H∞ control problems. Automatic Control, IEEE Transactions on,
34(8):831–847, 1989.
[8] B.A. Francis and J.C. Doyle. Linear control theory with an H∞ optimality criterion.
SIAM Journal on Control and Optimization, 25(4):815–844, 1987.
[9] P. Gahinet. On the game riccati equations arising in H∞ control problem. SIAM Journal
on Control and Optimization, 32(3):635–647, 1994.
[10] P.M. Gahinet and P. Pandey. Fast and numerically robust algorithm for computing the
H∞ optimum. In Decision and Control, 1991., Proceedings of the 30th IEEE Conference
on, pages 200–205. IEEE, 1991.
[11] K. Glover. All optimal Hankel-norm approximations of linear multivariable systems and
their L∞ error bounds. International Journal of Control, 39(6):1115–1193, 1984.
[12] G. Hewer. Existence theorems for positive semidefinite and sign indefinite stabilizing
solutions of H∞ riccati equations. SIAM journal on control and optimization, 31(1):16–
29, 1993.
38
George Kuster Bibliography 39
[13] P.P. Khargonekar, I.R. Petersen, and M.A. Rotea. H∞ -optimal control with state-
feedback. Automatic Control, IEEE Transactions on, 33(8):786–788, 1988.
[14] P.P. Khargonekar, I.R. Petersen, and K. Zhou. Robust stabilization of uncertain linear
systems: quadratic stabilizability and H∞ control theory. Automatic Control, IEEE
Transactions on, 35(3):356–361, 1990.
[15] D.J.N. Limebeer and M. Green. Linear robust control. Prentice-Hall, Englewood Clis,
NJ, 1995.
[16] W.W. Lin, C.S. Wang, and Q.F. Xu. On the computation of the optimal H∞ norms
for two feedback control problems. Linear algebra and its applications, 287(1):223–255,
1999.
[17] K.A. Morris. Introduction to feedback control. Academic Press, Inc., 2000.
[18] A. Packard, M.K.H. Fan, and J. Doyle. A power method for the structured singular
value. In Decision and Control, 1988., Proceedings of the 27th IEEE Conference on,
pages 2132–2137. IEEE, 1988.
[19] P. Pandey, C. Kenney, A. Packard, and AJ Laub. A gradient method for computing the
optimal H∞ -norm. Automatic Control, IEEE Transactions on, 36(7):887–890, 1991.
[21] C. Scherer. H∞ -control by state-feedback and fast algorithms for the computation of
optimalH∞ -norms. Automatic Control, IEEE Transactions on, 35(10):1090–1099, 1990.
[22] D. Tall and S. Vinner. Concept image and concept definition in mathematics with partic-
ular reference to limits and continuity. Educational studies in mathematics, 12(2):151–
169, 1981.
[23] G. Zames. Feedback and optimal sensitivity: Model reference transformations, multi-
plicative seminorms, and approximate inverses. Automatic Control, IEEE Transactions
on, 26(2):301–320, 1981.
[24] M. Zandieh. A theoretical framework for analyzing student understanding of the con-
cept of derivative. Research in Collegiate Mathematics Education. IV. CBMS Issues in
Mathematics Education, pages 103–127, 2000.
[25] K. Zhou, J.C. Doyle, K. Glover, et al. Robust and optimal control, volume 40. Prentice
Hall Upper Saddle River, NJ, 1996.