International Journal of Mathematical Analysis
Vol. 12, 2018, no. 10, 439 - 468
HIKARI Ltd, www.m-hikari.com
https://doi.org/10.12988/ijma.2018.8856
Critical Cones for Regular Controls
with Inequality Constraints
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
IIMAS, Universidad Nacional Autónoma de México
Apartado Postal 20-126, CDMX 01000, México
Copyright c 2018 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth. This
article is distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is
properly cited.
Abstract
Based on the notions of normality and regularity for constrained
problems in optimization, a conjecture on second order necessary conditions related to different critical cones is posed for a wide range of
optimal control problems involving equality and inequality constraints.
Several properties of the sets and functions delimiting the problem are
derived and, for a specific case where the surmise has been proved correct, some fundamental questions derived from this result are completely
solved.
Mathematics Subject Classification: 49K15
Keywords: Nonlinear programming, optimal control, second order necessary conditions, normality, regularity
1
Introduction: three main questions
When dealing with the theory of second order necessary conditions for constrained problems in optimization, one usually faces three fundamental questions which can be summarized in these words: who (that is, which function
satisfies the conditions), when (under what assumptions), and where (on which
set do they hold). For constrained problems in the finite dimensional case, they
have been answered in a satisfactory way in terms of Lagrange multipliers but,
440
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
even for such a well-known theory, the functions, the assumptions, and the
sets where the conditions hold, may vary from text to text.
In this paper we shall be concerned with these same questions in the context of optimal control. Though second order conditions for such problems can
be found in the literature, some fundamental questions still remain open, and
our aim is to give answers to some of them. For the optimal control problem
we shall deal with, the control functions are restricted to satisfy equality and
inequality constraints for all points t in a fixed compact time interval. As
pointed out in [12], these constraints make the problem much more complex
than the mathematical programming problem, or even the isoperimetric problem (see [6, 7]), in part because one deals with infinitely many constraints,
one for each t. In some cases like in [21, 25], the main results on second order
conditions are not proved and, quoting [21], “the derivation of the conditions
is very special and difficult.”
There is an extensive literature explaining the importance of deriving second order conditions for optimal control problems, both from a theoretical
and an applied point of view. Particular attention deserves the research developed in [1–4] where one of the main features is that the necessary conditions
obtained make sense, and are derived, without a priori assumptions of normality, and second order conditions are expressed in terms of the maximum
of a quadratic form over certain multipliers, for all the elements of the critical
cone of the original set of constraints. Similarly, in [9–11, 24, 26], all essential
references in the subject, second order necessary conditions in Pontryagin form
are derived for problems with pure state and mixed (control-state) constraints,
and varying the set of multipliers (those satisfying Pontryagin’s principle, or
the larger set of Lagrange multipliers) where the maximum of the quadratic
forms on the set of critical variations is taken. For a more complete study of
the subject, see also [16, 18, 20, 22, 27–30] and references therein.
Our approach, based on the notions of normal and regular controls, is rather
different to the one found in those references. To begin with, one finds in those
references results from an abstract optimization theory on Banach spaces applied to the control problem posed over L∞ -controls, a technique which does
not work in our setting, where the fixed endpoint Lagrange control problem
studied is posed over piecewise C 1 trajectories and piecewise continuous controls (the same setting can be found, for example, in [18, 22, 24]). Secondly,
we shall be able to establish, for certain cases, the nonnegativity of a certain
quadratic form under weaker assumptions and on larger sets than those appearing in those texts (see [22, 28], where a similar set of critical directions
for our problem is proposed). Finally, the nature of our necessary conditions
differs from the previous ones in that the nonnegativity of the second variation holds for the same multipliers without the need to take a maximum over
different sets of multipliers.
441
Critical cones for regular controls
This paper is organized as follows. To clearly situate the contributions of
the article, we provide in Section 2 a summary of the main features, in terms
of the notions of regularity and normality, that occur in the finite dimensional
case. This is crucial in our exposition, not only because some of the results
given are used later, but also due to the way it is presented and for comparison
reasons with the infinite dimensional case. In Section 3 we pose the optimal
control problem we shall deal with and follow a way of reasoning similar to
that of the previous section, leading to a natural conjecture on second order
conditions, as well as to a new fundamental result related to a characterization
of a certain set of critical directions. A particular case, covering a wide range
of problems and recently derived in [27], is studied in Section 4, where we solve
some important surmises and questions related to the theory of second order
necessary conditions.
2
The finite dimensional case
In this section we shall deal with constrained minimum problems in finite
dimensional spaces. We shall state, in a succinct and clear way, first and second
order necessary conditions for optimality in terms of Lagrange multiplier rules.
The approach we follow is based on the notions of regularity and normality.
For a full account of these ideas we refer to [18, 19].
Suppose we are given a set S ⊂ Rn and a function f mapping Rn to R,
and consider the problem, which we label P(S), of minimizing f on S.
The notion of tangent cone given below is due to Hestenes [18] and, as
shown in [17], it is equivalent to the one introduced by Bouligand (1932), also
known as the contingent cone to S at x0 . Other authors such as Bazaraa,
Goode, Nashed, Kurcyusz, Rockafellar, Saks, Rogak and many more, have
given various equivalent definitions of such a cone (see [5, 17]).
Definition 2.1 We shall say that a sequence {xq } ⊂ Rn converges to x0 in
the direction h if h is a unit vector, xq 6= x0 , and
lim |xq − x0 | = 0,
q→∞
lim
q→∞
xq − x0
= h.
|xq − x0 |
The tangent cone of S at x0 , denoted by TS (x0 ), is the (closed) cone determined
by the unit vectors h for which there exists a sequence {xq } in S converging
to x0 in the direction h.
Equivalently (see [19]), TS (x0 ) is the set of all h ∈ Rn for which there exist
a sequence {xq } in S and a sequence {tq } of positive numbers such that
lim tq = 0,
q→∞
lim
q→∞
xq − x0
= h.
tq
442
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
Clearly, if {xq } converges to x0 in the direction h and f has a differential
at x0 , then
f (xq ) − f (x0 )
lim
= f ′ (x0 ; h).
q→∞
|xq − x0 |
If f has a second differential at x0 , then
f (xq ) − f (x0 ) − f ′ (x0 ; xq − x0 )
1
= f ′′ (x0 ; h).
2
q→∞
|xq − x0 |
2
lim
From these facts, necessary conditions for P(S) follow straightforwardly.
Theorem 2.2 Suppose x0 solves P(S) locally. If f has a differential at x0 ,
then f ′ (x0 ; h) ≥ 0 for all h ∈ TS (x0 ). If f has a second differential at x0 and
f ′ (x0 ) = 0, then f ′′ (x0 ; h) ≥ 0 for all h ∈ TS (x0 ).
Consider now problem P(S) with S defined by inequality and equality
constraints. We are given functions f, gi : Rn → R (i ∈ A ∪ B) and
S = {x ∈ Rn | gα (x) ≤ 0 (α ∈ A), gβ (x) = 0 (β ∈ B)}
where A = {1, . . . , p}, B = {p + 1, . . . , m}. The cases p = 0 and p = m are
to be given the obvious interpretations. Assume f and gi are continuous on a
neighborhood of x0 ∈ S and possess second differentials at x0 .
Definition 2.3 Define the set of active indices at x0 by
I(x0 ) := {α ∈ A | gα (x0 ) = 0}
and the set of vectors satisfying the tangential constraints at x0 by
RS (x0 ) := {h ∈ Rn | gα′ (x0 ; h) ≤ 0 (α ∈ I(x0 )), gβ′ (x0 ; h) = 0 (β ∈ B)}.
Note that TS (x0 ) ⊂ RS (x0 ). We shall say that x0 is a regular point of S if
TS (x0 ) = RS (x0 ).
The first order Lagrange multiplier rule is a consequence of Theorem 2.2
and some basic results on linear functionals. Let us define the set E of extremals
as the set of all (x0 , λ) ∈ S × Rm such that
i. λα ≥ 0 and λα gα (x0 ) = 0 (α ∈ A).
ii. If F (x) := f (x) + hλ, g(x)i then F ′ (x0 ) = 0.
Theorem 2.4 Suppose x0 solves P(S) locally. If x0 is a regular point of S,
then ∃λ ∈ Rm such that (x0 , λ) ∈ E.
The second order Lagrange multiplier rule follows by a simple application
of Theorem 2.2.
443
Critical cones for regular controls
Theorem 2.5 Suppose (x0 , λ) ∈ E, F (x) = f (x) + hλ, g(x)i, and S1 =
{x ∈ S | F (x) = f (x)}. If x0 solves P(S) locally, then F ′′ (x0 ; h) ≥ 0 for all
h ∈ TS1 (x0 ).
Note that
S1 = {x ∈ S | gα (x) = 0 (α ∈ Γ)}
where Γ = {α ∈ A | λα > 0}, and
RS1 (x0 ) = {h ∈ RS (x0 ) | gα′ (x0 ; h) = 0 (α ∈ Γ)}
= {h ∈ RS (x0 ) | f ′ (x0 ; h) = 0}.
In general, it may be difficult to test for regularity and one usually requires some criteria that implies that condition. A simple criterion is that of
normality, defined in [19] as follows.
Definition 2.6 We shall say that x0 ∈ S is normal relative to S if λ = 0 is
the only solution to
i. λα ≥ 0 and λα gα (x0 ) = 0 (α ∈ A).
P
′
ii. m
1 λi gi (x0 ) = 0.
To be precise, we should say in this definition that x0 ∈ S is normal relative
to the functions defining the set S and not relative to the set S defined by those
functions. In other words, if λ = 0 is the only solution to the above system,
then x0 ∈ S is normal relative to the function g and the sets A and B of
indices for inequality and equality constraints respectively. However, following
the definition introduced in [18, 19], we shall use the above definition, and no
confusion should arise if S is replaced with a different set defined also in terms
of inequality and equality constraints.
In [19], it is proved that x0 is normal relative to S if and only if the set
′
{gβ (x0 ) | β ∈ B} is linearly independent and, if p > 0, ∃h such that
gα′ (x0 ; h) < 0 (α ∈ I(x0 )),
gβ′ (x0 ; h) = 0 (β ∈ B).
This characterization is used in [19] to prove the following, crucial result.
Theorem 2.7 If x0 is a normal point of S then x0 is a regular point of S.
The following extended rule, the Fritz John necessary optimality condition,
yields in a natural way the above definition of normality (see [23] for a simple
proof based on the theory of augmentability).
Theorem 2.8 Suppose x0 solves P(S) locally. Then ∃λ0 ≥ 0 and λ ∈ Rm ,
not both zero, such that
i. λα ≥ 0 and λα gα (x0 ) = 0 (α ∈ A).
ii. If F0 (x) := λ0 f (x) + hλ, g(x)i then F0′ (x0 ) = 0.
444
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
Clearly, in this theorem, if x0 is also a normal point of S then λ0 > 0 and
the multipliers can be chosen so that λ0 = 1. The real numbers λ1 , . . . , λm
satisfying the first order Lagrange multiplier rule (or the Karush-Kuhn-Tucker
conditions) are called the Kuhn-Tucker or Lagrange multipliers, and the function F is the standard Lagrangian. In general, if x0 affords a local minimum
to f on S, those conditions may not hold at x0 , and some additional assumptions should be imposed to guarantee the existence of Lagrange multipliers.
Assumptions of this nature are usually referred to as constraint qualifications
(see [17]) since they involve only the constraints and are independent of the
geometric structure of the feasible set S (a broader definition, in terms of critical directions, is given in [8]). Equivalently, they correspond to conditions
which assure the positiveness of the cost multiplier λ0 in Theorem 2.8. Both
normality and regularity are constraint qualifications, as well as the characterization of normality mentioned above, known as the Mangasarian-Fromovitz
constraint qualification.
A combination of Theorems 2.5 and 2.7 yields the following basic result on
second order necessary conditions.
Theorem 2.9 Suppose (x0 , λ) ∈ E, F (x) = f (x)+hλ, g(x)i, and S1 = {x ∈
S | F (x) = f (x)}. If x0 solves P(S) locally and is a normal point of S1 , then
F ′′ (x0 ; h) ≥ 0 for all h ∈ RS1 (x0 ).
Let us turn now to a different notion of normality given in [18].
Definition 2.10 x0 ∈ S is s-normal relative to S if the linear functionals
(α ∈ I(x0 ) ∪ B) in h are linearly independent.
gα′ (x0 ; h)
By an application of the implicit function theorem, and making use of the
set of curvilinear tangent vectors of S at x0 given by
CS (x0 ) := {h ∈ Rn | ∃δ > 0, x: [0, δ) → S such that x(0) = x0 , ẋ(0) = h},
one can easily show that s-normality implies that RS (x0 ) ⊂ CS (x0 ). Since
CS (x0 ) ⊂ TS (x0 ) ⊂ RS (x0 ), this implies the following well-known result.
Theorem 2.11 If x0 is s-normal relative to S then x0 is a regular point
of S.
This statement is precisely Lemma 10.1, Chapter 1 of [18] and it has an
implication, in particular, on second order conditions. In Theorem 2.5, if x0
is a regular point of S1 , then F ′′ (x0 ; h) ≥ 0 for all h ∈ RS1 (x0 ). By Theorem
2.7, x0 is a regular point of S1 if it is a normal point of S1 , and thus Theorem
2.9 holds. In [18], on the other hand, the criterion for this condition to hold is
for x0 to be s-normal relative to S since then it would be s-normal relative to
445
Critical cones for regular controls
S1 , and hence a regular point of S1 . As one readily verifies, x0 ∈ S is s-normal
relative to S if it is normal relative to the set
S0 := {x ∈ Rn | gi (x) = 0 (i ∈ I(x0 ) ∪ B)}
and thus the criterion used in [18] for regularity relative to S1 is in general
stronger than that of normality relative to S1 used in [19]. Moreover, in most
textbooks (see a list of well-known references in [8]) not only the assumption of
normality relative to S1 is replaced by (the stronger assumption of) normality
relative to S0 , but the set of tangential constraints RS1 (x0 ) is replaced by (the,
in general, smaller set of tangential constraints)
RS0 (x0 ) = {h ∈ Rn | gi′ (x0 ; h) = 0 (i ∈ I(x0 ) ∪ B)}
This rather “well-worn result” (as Ben-Tal puts it [8]) is the following.
Theorem 2.12 Suppose (x0 , λ) ∈ E and F (x) = f (x) + hλ, g(x)i. If x0
solves P(S) locally and is a normal point of S0 , then F ′′ (x0 ; h) ≥ 0 for all
h ∈ RS0 (x0 ).
As pointed out in [8], “The source of this weaker result can be attributed
to the traditional way of treating the active inequality constraints as equality
constraints.”
Let us end this section with an example which illustrates some of the main
features on second order necessary conditions explained above. It should be
noted that, for this example, the solution x0 to the problem is a regular point
of S so that, by Theorem 2.4, ∃λ such that (x0 , λ) is an extremal. However, in
contrast with the conclusion of Theorem 2.9, we can exhibit h ∈ RS1 (x0 ) with
F ′′ (x0 ; h) < 0.
Example 2.13 Consider the problem of minimizing f (x, y) = y + x2 on
the set
S = {(x, y) | gα (x, y) ≤ 0 (α = 1, 2, 3)}
where g1 (x, y) = −y, g2 (x, y) = cos x − y, g3 (x, y) = cos 2x − y.
As one readily verifies, (0, 1) solves the problem. The set of active indices
at this point is I(0, 1) = {2, 3}, and it is a regular point of S since
TS (0, 1) = {(h, k) | k ≥ 0},
RS (0, 1) = {(h, k) | −k ≤ 0}.
By Theorem 2.4, there exists λ = (λ1 , λ2 , λ3 ) with λ1 = 0 and λ2 , λ3 ≥ 0 such
that, if
F (x, y) = f (x, y) + hλ, g(x, y)i
= y + x2 − (λ1 + λ2 + λ3 )y + λ2 cos x + λ3 cos 2x
446
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
then F ′ (0, 1) = 0. Since
F ′ (x, y) = (2x − λ2 sin x − 2λ3 sin 2x, 1 − λ1 − λ2 − λ3 )
we have, with λ1 = 0, that F ′ (0, 1) = (0, 1 − λ2 − λ3 ) and so λ2 + λ3 = 1. Now,
note that
2 − λ2 cos x − 4λ3 cos 2x 0
′′
F (x, y) =
0
0
and therefore F ′′ (0, 1; h, k) = (2 − λ2 − 4λ3 )h2 . Let λ2 = 0 and λ3 = 1. Then
RS1 (0, 1) = {(h, k) | −k ≤ 0, k = 0}
and so, for any h 6= 0, we have (h, 0) ∈ RS1 (0, 1) but
F ′′ (0, 1; h, 0) = −2h2 < 0.
3
Optimal control
In this section we shall deal with a fixed endpoint Lagrange optimal control
problem involving inequality and equality constraints in the control functions.
We are interested in deriving, for this optimal control problem, a theory parallel
to the one given in the previous section. We shall follow a similar line of thought
and, in particular, try to see if a result analogous to Theorem 2.9, that is,
second order conditions in terms of normality relative to the corresponding set
S1 , can be established.
Suppose we are given an interval T := [t0 , t1 ] in R, two points ξ0 , ξ1 in
Rn , and functions L and f mapping T × Rn × Rm to R and Rn respectively,
and ϕ = (ϕ1 , . . . , ϕq ) mapping Rm to Rq (q ≤ m). Denote by X the space
of piecewise C 1 functions mapping T to Rn , by Uk the space of piecewise
continuous functions mapping T to Rk (k ∈ N), set Z := X × Um ,
D := {(x, u) ∈ Z | ẋ(t) = f (t, x(t), u(t)) (t ∈ T ), x(t0 ) = ξ0 , x(t1 ) = ξ1 },
S := {(x, u) ∈ D | ϕα (u(t)) ≤ 0, ϕβ (u(t)) = 0 (α ∈ R, β ∈ Q, t ∈ T )}
where R = {1, . . . , r}, Q = {r+1, . . . , q}, and consider the functional I: Z → R
given by
Z
t1
I(x, u) :=
t0
L(t, x(t), u(t))dt ((x, u) ∈ Z).
The problem we shall deal with, which (again, but no confusion should arise)
we label P(S), is that of minimizing I over S.
Implicit in the statement of the problem is a relatively open subset O of
T × Rn for which the domain of L and f is O × Rm . Elements of Z will
be called processes, and a process (x, u) is admissible if it belongs to S and
Critical cones for regular controls
447
(t, x(t)) ∈ O (t ∈ T ). A process (x, u) solves P(S) (locally) if it is admissible
and (upon shrinking O if necessary) we have I(x, u) ≤ I(y, v) for all admissible
processes (y, v).
Given (x, u) ∈ Z we shall use the notation (x̃(t)) to represent (t, x(t), u(t)),
and ‘∗’ denotes transpose. With respect to the smoothness of the functions
delimiting the problem, we assume that L, f and ϕ are C 2 and the q × (m + r)dimensional matrix
∂ϕi
δiα ϕα
∂uk
(i = 1, . . . , q; α = 1, . . . , r; k = 1, . . . , m)
has rank q on U (here δαα = 1, δαβ = 0 (α 6= β)), where
U := {u ∈ Rm | ϕα (u) ≤ 0 (α ∈ R), ϕβ (u) = 0 (β ∈ Q)}.
This condition is equivalent to the condition that, at each point u in U , the
matrix
∂ϕi
(i = i1 , . . . , ip ; k = 1, . . . , m)
∂uk
has rank p, where i1 , . . . , ip are the indices i ∈ {1, . . . , q} such that ϕi (u) = 0
(see [15] for details).
For all (t, x, u, p, µ, λ) in T × Rn × Rm × Rn × Rq × R let
H(t, x, u, p, µ, λ) := hp, f (t, x, u)i − λL(t, x, u) − hµ, ϕ(u)i.
First order necessary conditions are well established (see, for example, [15, 18,
22]) and one version can be stated as follows.
Theorem 3.1 If (x0 , u0 ) solves P(S), ∃λ0 ≥ 0, p ∈ X, and µ ∈ Uq not
vanishing simultaneously on T , such that
a. µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T );
b. ṗ(t) = −Hx∗ (x̃0 (t), p(t), µ(t), λ0 ) and Hu (x̃0 (t), p(t), µ(t), λ0 ) = 0 on
every interval of continuity of u0 .
From this result we introduce the notion of normality as in Section 2, that
is, in such a way that the non-vanishing of the cost multiplier can be assured.
This is accomplished by having zero as the unique solution to the adjoint
equation whenever λ0 = 0.
Definition 3.2 (x0 , u0 ) ∈ S is a normal process of S if, given (p, µ) ∈
X × Uq satisfying
i. µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T );
ii. ṗ(t) = −fx∗ (x̃0 (t))p(t) [ = −Hx∗ (x̃0 (t), p(t), µ(t), 0) ] (t ∈ T );
iii. 0 = fu∗ (x̃0 (t))p(t) − ϕ′∗ (u0 (t))µ(t) [ = Hu∗ (x̃0 (t), p(t), µ(t), 0) ] (t ∈ T ),
then p ≡ 0. Note that, in this event, also µ ≡ 0.
448
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
We define the set of extremals as the set of all (x, u, p, µ) for which the
conditions of Theorem 3.1 hold with λ0 = 1.
Definition 3.3 Denote by E the set of all (x, u, p, µ) ∈ Z × X × Uq such
that
a. µα (t) ≥ 0 and µα (t)ϕα (u(t)) = 0 (α ∈ R, t ∈ T );
b. ṗ(t) = −fx∗ (x̃(t))p(t) + L∗x (x̃(t)) (t ∈ T );
c. fu∗ (x̃(t))p(t) = L∗u (x̃(t)) + ϕ′∗ (u(t))µ(t) (t ∈ T ).
Note 3.4 From the above definitions it follows that, in Theorem 3.1, if
(x0 , u0 ) is also a normal process of S, then ∃(p, µ) ∈ X × Uq such that
(x0 , u0 , p, µ) ∈ E.
Let us now introduce the second variation with respect to H, as well as the
set of curvilinear tangent processes. For any (x, u, p, µ) ∈ Z × X × Uq let
J((x, u, p, µ); (y, v)) :=
Z t1
t0
2Ω(t, y(t), v(t))dt ((y, v) ∈ Z)
where, for all (t, y, v) ∈ T × Rn × Rm ,
2Ω(t, y, v) := −[hy, Hxx (t)yi + 2hy, Hxu (t)vi + hv, Huu (t)vi]
and H(t) denotes H(x̃(t), p(t), µ(t), 1).
For any (x0 , u0 ) ∈ S, denote by CS (x0 , u0 ) the set of all processes (y, v) for
which there exist δ > 0 and (x(·, ǫ), u(·, ǫ)) ∈ S (0 ≤ ǫ < δ) such that
a. (x(t, 0), u(t, 0)) = (x0 (t), u0 (t)) (t ∈ T );
b. (xǫ (t, 0), uǫ (t, 0)) = (y(t), v(t)) (t ∈ T );
c. x(t, ǫ) is continuous and has continuous first and second derivatives with
respect to ǫ; u(t, ǫ) and its first and second derivatives with respect to ǫ are
piecewise continuous with respect to t.
Based on these definitions, we can establish the following fundamental result.
Theorem 3.5 Let (x0 , u0 ) ∈ S and suppose ∃(p, µ) ∈ X × Uq such that
(x0 , u0 , p, µ) ∈ E. Let
S1 [= S1 (µ)] := {(x, u) ∈ S | ϕα (u(t)) = 0 (α ∈ R, µα (t) > 0, t ∈ T )}.
If (x0 , u0 ) solves P(S) then
J((x0 , u0 , p, µ); (y, v)) ≥ 0 for all (y, v) ∈ CS1 (x0 , u0 ).
449
Critical cones for regular controls
Proof: Define
K(x, u) := hp(t1 ), ξ1 i − hp(t0 ), ξ0 i +
Z t1
t0
F (t, x(t), u(t))dt ((x, u) ∈ Z)
where, for all (t, x, u) ∈ T × Rn × Rm ,
F (t, x, u) := L(t, x, u) − hp(t), f (t, x, u)i + hµ(t), ϕ(u)i − hṗ(t), xi.
Observe that
F (t, x, u) = −H(t, x, u, p(t), µ(t), 1) − hṗ(t), xi
and, if (x, u) ∈ S, then
K(x, u) = I(x, u) +
Z t1
t0
hµ(t), ϕ(u(t))idt.
Note that
S1 = {(x, u) ∈ S | µα (t)ϕα (u(t)) = 0 (α ∈ R, t ∈ T )}
= {(x, u) ∈ S | K(x, u) = I(x, u)}.
Since (x0 , u0 ) ∈ S1 , (x0 , u0 ) minimizes K on S1 . Now, let (y, v) belong to
CS1 (x0 , u0 ) and let δ > 0 and (x(·, ǫ), u(·, ǫ)) ∈ S1 (0 ≤ ǫ < δ) be such that
(x(t, 0), u(t, 0)) = (x0 (t), u0 (t)) and (xǫ (t, 0), uǫ (t, 0)) = (y(t), v(t)).
Then g(ǫ) := K(x(·, ǫ), u(·, ǫ)) (0 ≤ ǫ < δ) satisfies
g(ǫ) = I(x(·, ǫ), u(·, ǫ)) ≥ I(x0 , u0 ) = K(x0 , u0 ) = g(0) (0 ≤ ǫ < δ).
Note that
Fx (x̃0 (t)) = −Hx (x̃0 (t), p(t), µ(t), 1) − ṗ∗ (t) = 0,
Fu (x̃0 (t)) = −Hu (x̃0 (t), p(t), µ(t), 1) = 0
and therefore g ′ (0) = 0. Consequently 0 ≤ g ′′ (0) = J((x0 , u0 , p, µ); (y, v)).
As in the finite dimensional case, we are interested in characterizing the
set CS (x0 , u0 ) in terms of a set of tangential constraints whose membership, in
contrast with CS (x0 , u0 ), may be easily verifiable.
Definition 3.6 For all (x0 , u0 ) ∈ S let
L(x0 , u0 ) := {(y, v) ∈ Z | ẏ(t) = A(t)y(t) + B(t)v(t) (t ∈ T ),
y(t0 ) = y(t1 ) = 0}
450
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
where A(t) = fx (x̃0 (t)), B(t) = fu (x̃0 (t)). Denote the set of active indices at
u ∈ Rm by
Ia (u) := {α ∈ R | ϕα (u) = 0}
and define the set of processes satisfying the tangential constraints at (x0 , u0 )
with respect to S by
RS (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | ϕ′α (u0 (t))v(t) ≤ 0 (α ∈ Ia (u0 (t)), t ∈ T ),
ϕ′β (u0 (t))v(t) = 0 (β ∈ Q, t ∈ T )}.
As we prove next, curvilinear tangent processes satisfy the tangential constraints.
Proposition 3.7 For all (x0 , u0 ) ∈ S, CS (x0 , u0 ) ⊂ RS (x0 , u0 ).
Proof: Let (y, v) ∈ CS (x0 , u0 ) and let δ > 0 and (x(·, ǫ), u(·, ǫ)) ∈ S (0 ≤
ǫ < δ) be such that
(x(t, 0), u(t, 0)) = (x0 (t), u0 (t)) and (xǫ (t, 0), uǫ (t, 0)) = (y(t), v(t)).
Then, for all 0 ≤ ǫ < δ,
ẋ(t, ǫ) = f (t, x(t, ǫ), u(t, ǫ)) (t ∈ T ),
x(t0 , ǫ) = ξ0 , x(t1 , ǫ) = ξ1
and so (y, v) solves L(x0 , u0 ). Also, for all (t, ǫ) ∈ T × [0, δ),
ϕα (u(t, ǫ)) ≤ 0 (α ∈ R),
ϕβ (u(t, ǫ)) = 0 (β ∈ Q).
Fix i ∈ R ∪ Q and t ∈ T , and set γ(ǫ) := ϕi (u(t, ǫ)) so that γ ′ (0) =
ϕ′i (u0 (t))v(t). If i ∈ Ia (u0 (t)) then γ ′ (0) ≤ 0. If i ∈ Q then γ ′ (0) = 0.
Definition 3.8 We say (x0 , u0 ) is a c-regular point of S if CS (x0 , u0 ) =
RS (x0 , u0 ).
Now, let us exhibit three sets of constraints and their corresponding sets
of tangential constraints. For any u ∈ Rm and µ ∈ Rq , let
τ0 (u) := {h ∈ Rm | ϕ′i (u)h = 0 (i ∈ Ia (u) ∪ Q)},
τ1 (u, µ) := {h ∈ Rm | ϕ′α (u)h ≤ 0 (α ∈ Ia (u), µα = 0),
ϕ′β (u)h = 0 (β ∈ R with µβ > 0, or β ∈ Q)},
τ2 (u) := {h ∈ Rm | ϕ′α (u)h ≤ 0 (α ∈ Ia (u)), ϕ′β (u)h = 0 (β ∈ Q)}.
451
Critical cones for regular controls
Note that
RS (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | v(t) ∈ τ2 (u0 (t)) (t ∈ T )}.
Also, we have
S1 (µ) = {(x, u) ∈ D | ϕα (u(t)) ≤ 0 (α ∈ R, µα (t) = 0, t ∈ T ),
ϕβ (u(t)) = 0 (β ∈ R with µβ (t) > 0, or β ∈ Q, t ∈ T )}
and therefore
RS1 (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | v(t) ∈ τ1 (u0 (t), µ(t)) (t ∈ T )}.
Similarly, if
S0 := {(x, u) ∈ D | ϕi (u(t)) = 0 (i ∈ Ia (u0 (t)) ∪ Q, t ∈ T )},
then
RS0 (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | v(t) ∈ τ0 (u0 (t)) (t ∈ T )}.
Let us now apply the definition of normality, given in Definition 3.2, to the
sets S0 and S1 .
Note 3.9 (x0 , u0 ) is a normal process of S0 if, given (p, µ) ∈ X × Uq satisfying
i. µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T );
ii. ṗ(t) = −fx∗ (x̃0 (t))p(t) and fu∗ (x̃0 (t))p(t) = ϕ′∗ (u0 (t))µ(t) (t ∈ T )
then p ≡ 0.
Note 3.10 (x0 , u0 ) is a normal process of S1 (µ) if, given (p, ν) ∈ X × Uq
satisfying
i. να (t) ≥ 0 and να (t)ϕα (u0 (t)) = 0 (α ∈ R, µα (t) = 0, t ∈ T );
ii. ṗ(t) = −fx∗ (x̃0 (t))p(t) and fu∗ (x̃0 (t))p(t) = ϕ′∗ (u0 (t))ν(t) (t ∈ T )
then p ≡ 0.
Normality relative to S0 , S1 and S can be characterized in terms of the
sets τ0 , τ1 and τ2 defined above. We refer to [28] for a detailed proof of these
characterizations.
Proposition 3.11 Let (x0 , u0 ) ∈ S and suppose µ ∈ Uq is such that µα (t) ≥
0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ). Then (x0 , u0 ) is a normal process of
S0 (respectively S1 , S) if and only if z ≡ 0 is the only solution of the system
ż(t) = −A∗ (t)z(t),
z ∗ (t)B(t)h ≤ 0
for all h ∈ τ0 (u0 (t)) (t ∈ T )
(respectively h ∈ τ1 (u0 (t), µ(t)), h ∈ τ2 (u0 (t))).
452
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
As one readily verifies (see also [28] for details), if (x0 , u0 ) is normal relative
to S0 (or strongly normal) then it is normal relative to S1 (with µ as above)
which in turn implies normality relative to S (or weak normality).
It is important to mention, at this stage, that second order conditions in
terms of RS0 (x0 , u0 ) assuming normality relative to S0 are well established in
the literature (see, for example, [15]). The result is similar to Theorem 2.12
and it can be stated as follows.
Theorem 3.12 Let (x0 , u0 ) ∈ S and suppose ∃(p, µ) ∈ X × Uq such that
(x0 , u0 , p, µ) ∈ E. If (x0 , u0 ) solves P(S) and is a normal process of S0 , then
J((x0 , u0 , p, µ); (y, v)) ≥ 0 for all (y, v) ∈ RS0 (x0 , u0 ).
As we show next, we are able to provide two examples which illustrate a
crucial aspect of the theory, namely, that if we modify the assumptions or the
space where the nonnegativity of the quadratic form holds, the conclusion of
Theorem 3.12 may no longer hold. Let us begin with slightly changing the
assumptions. In this first example, we exhibit a solution (x0 , u0 ) to a problem P(S) which is normal relative to S, together with a pair (p, µ) such that
(x0 , u0 , p, µ) ∈ E, but J((x0 , u0 , p, µ); (y, v)) < 0 for some (y, v) ∈ RS0 (x0 , u0 ).
Example 3.13 Let a > 0, k > 1 and consider the problem of minimizing
I(x, u) =
Z 1
0
{u2 (t) − u1 (t)}dt
subject to (x, u) ∈ Z,
ẋ(t) = 2u2 (t) − (k + 1)u1 (t) + au23 (t) (t ∈ [0, 1]),
x(0) = x(1) = 0, and
u2 (t) ≥ u1 (t),
ku1 (t) ≥ u2 (t) (t ∈ [0, 1]).
In this case T = [0, 1], n = 1, m = 3, r = q = 2, ξ0 = ξ1 = 0 and, for all
t ∈ T , x ∈ R, and u = (u1 , u2 , u3 ),
L(t, x, u) = u2 − u1 ,
f (t, x, u) = 2u2 − (k + 1)u1 + au23 ,
ϕ1 (u) = u1 − u2 ,
ϕ2 (u) = u2 − ku1 .
We have
H(t, x, u, p, µ) = p(2u2 − (k + 1)u1 + au23 ) + u1 − u2 − µ1 (u1 − u2 ) − µ2 (u2 − ku1 )
and so
Hu (t, x, u, p, µ) = (−p(k + 1) − µ1 + kµ2 + 1, 2p + µ1 − µ2 − 1, 2apu3 ),
453
Critical cones for regular controls
0 0 0
Huu (t, x, u, p, µ) =
0
0 0
0 0 2ap
so that, for all (x, u, p, µ) ∈ Z × X × U2 and (y, v) ∈ Z,
J((x, u, p, µ); (y, v)) = −
Z 1
0
2ap(t)v32 (t)dt.
Clearly (x0 , u0 ) ≡ (0, 0) is a solution to the problem. Note that (x0 , u0 , p, µ) is
an extremal if µα (t) ≥ 0 (α = 1, 2), ṗ(t) = 0 and
fu∗ (x̃0 (t))p(t) = L∗u (x̃0 (t)) + ϕ′∗ (u0 (t))µ(t) (t ∈ T ).
This last relation corresponds to
−(k + 1)
−1
µ1 (t) − kµ2 (t)
2
p(t) = 1 + −µ1 (t) + µ2 (t)
0
0
0
and thus, if µ = (µ1 , µ2 ) ≡ (0, 1) and p ≡ 1, (x0 , u0 , p, µ) ∈ E.
Now, since ϕ′1 (u) = (1, −1, 0) and ϕ′2 (u) = (−k, 1, 0), we have
τ0 (u0 (t)) = {h ∈ R3 | h1 − h2 = 0, h2 − kh1 = 0},
τ1 (u0 (t), µ(t)) = {h ∈ R3 | h1 − h2 ≤ 0, h2 − kh1 = 0},
τ2 (u0 (t)) = {h ∈ R3 | h1 − h2 ≤ 0, h2 − kh1 ≤ 0}.
Since fx (x̃0 (t)) = 0 and fu (x̃0 (t)) = (−(k + 1), 2, 0), the system
ż(t) = −A∗ (t)z(t) = 0,
z ∗ (t)B(t)h = z(t)(−(k + 1)h1 + 2h2 ) = 0
for all (h1 , h2 , h3 ) ∈ τ0 (u0 (t)) (t ∈ T ) has nontrivial solutions and so (x0 , u0 )
is not normal relative to S0 . On the other hand, z ≡ 0 is the only solution to
the system
ż(t) = 0, z(t)(−(k + 1)h1 + 2h2 ) ≤ 0
for all (h1 , h2 , h3 ) ∈ τ2 (u0 (t)) (t ∈ T ) since both (1, 1, 0) and (1, k, 0) belong
to τ2 (u0 (t)) implying that z(t) ≥ 0 and z(t) ≤ 0 (t ∈ T ) respectively, so that
(x0 , u0 ) is normal relative to S. Finally, the system
ż(t) = 0, z(t)(−(k + 1)h1 + 2h2 ) ≤ 0
for all (h1 , h2 , h3 ) ∈ τ1 (u0 (t), µ(t)) (t ∈ T ) corresponds to ż(t) = 0, z(t)(h2 −
h1 ) ≤ 0 with h2 − h1 ≥ 0, which has nontrivial solutions and so (x0 , u0 , µ) is
not a normal point of S1 .
Let v = (v1 , v2 , v3 ) ≡ (0, 0, 1) and y ≡ 0. Then (y, v) solves
ẏ(t) = −(k + 1)v1 (t) + 2v2 (t) (t ∈ T ),
y(0) = y(1) = 0, v(t) ∈ τ0 (u0 (t)), and J((x0 , u0 , p, µ); (y, v)) = −2a < 0.
454
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
Our second example deals with a change, in Theorem 3.12, of the critical
cone where the second order conditions hold. It shows that a solution to the
problem, which is a normal process of S0 , may yield a negative second variation
on RS (x0 , u0 ).
Example 3.14 Consider the problem of minimizing
I(x, u) =
Z 1
0
− exp(−u2 (t))dt
subject to (x, u) ∈ Z, ẋ(t) = u1 (t) + u22 (t) (t ∈ [0, 1]), x(0) = x(1) = 0,
u2 (t) ≥ 0 (t ∈ [0, 1]).
In this case T = [0, 1], n = 1, m = 2, r = q = 1, ξ0 = ξ1 = 0 and, for all
t ∈ T , x ∈ R, and u = (u1 , u2 ),
L(t, x, u) = −e−u2 ,
f (t, x, u) = u1 + u22 ,
ϕ1 (u) = −u2 .
We have H(t, x, u, p, µ) = p(u1 + u22 ) + µ1 u2 + e−u2 and so
Hu (t, x, u, p, µ) = (p, 2pu2 + µ1 − e
−u2
),
Huu (t, x, u, p, µ) =
0
0
0 2p + e−u2
so that, for all (x, u, p, µ) ∈ Z × X × U1 and (y, v) ∈ Z,
J((x, u, p, µ); (y, v)) = −
Z 1
0
(2p(t) + e−u2 (t) )v22 (t)dt.
Clearly (x0 , u0 ) ≡ (0, 0) solves the problem. Also, since fu (t, x, u) = (1, 2u2 ),
Lu (t, x, u) = (0, e−u2 ) and ϕ′ (u) = (0, −1), (x0 , u0 , p, µ) is an extremal if ṗ(t) =
0 and
1
0
0
p(t) =
+
.
0
1
−µ(t)
Thus, if (p, µ) ≡ (0, 1), (x0 , u0 , p, µ) ∈ E. Observe now that
τ0 (u0 (t)) = τ1 (u0 (t), µ(t)) = {h ∈ R2 | h2 = 0},
τ2 (u0 (t)) = {h ∈ R2 | −h2 ≤ 0}.
Since z ≡ 0 is the only solution to the system
ż(t) = 0,
z(t)h1 = 0 for all (h1 , h2 ) ∈ τ0 (u0 (t)) (t ∈ T ),
(x0 , u0 ) is normal relative to S0 . Now, if v ≡ (0, 1) and y ≡ 0 then v(t) ∈
τ2 (u0 (t)) and ẏ(t) = v1 (t) (t ∈ T ), y(0) = y(1) = 0, but
J((x0 , u0 , p, µ); (y, v)) = −1 < 0.
455
Critical cones for regular controls
In spite of these examples, Theorem 3.12 clearly looks too weak in comparison with the main result that holds for the finite dimensional case, namely,
Theorem 2.9. In the latter, one assumes normality relative to S1 and one can
assure that the second variation is nonnegative on the set of tangential constraints with respect to S1 . In the former, normality is assumed relative to S0
(in general a stronger assumption than normality on S1 ) and the condition on
the second variation holds on the tangential constraints with respect to S0 (in
general a set smaller than that with respect to S1 ). We pose the question if a
result, analogous to Theorem 2.9, holds for our optimal control problem, that
is, if the following result is valid or not.
Conjecture 3.15 Let (x0 , u0 ) ∈ S and suppose ∃(p, µ) ∈ X × Uq such that
(x0 , u0 , p, µ) ∈ E. Let
S1 := {(x, u) ∈ S | ϕα (u(t)) = 0 (α ∈ R, µα (t) > 0, t ∈ T )}.
If (x0 , u0 ) solves P(S) and is normal relative to S1 , then
J((x0 , u0 , p, µ); (y, v)) ≥ 0 for all (y, v) ∈ RS1 (x0 , u0 ).
Clearly, the role played by RS1 (x0 , u0 ) above is crucial. We shall end this
section by proving a new result, mentioned in the introduction, which entirely
characterizes this set of critical directions. In particular, we shall conclude
that, though the set S1 (µ) of constraints depends on the multiplier µ, this
dependence is no longer present in the set RS1 (x0 , u0 ) of tangential constraints
relative to S1 (µ).
In our characterization given below, we shall make use of the first variation
of I given by
I ′ (x0 , u0 ; y, v) =
Z t1
t0
{Lx (x̃0 (t))y(t) + Lu (x̃0 (t))v(t)}dt.
For all (x0 , u0 ) ∈ S, denote by M (x0 , u0 ) the set of all (p, µ) ∈ X × Uq such
that (x0 , u0 , p, µ) ∈ E, that is,
a. µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T );
b. ṗ(t) = −A∗ (t)p(t) + L∗x (x̃0 (t)) (t ∈ T );
c. B ∗ (t)p(t) = L∗u (x̃(t)) + ϕ′∗ (u(t))µ(t) (t ∈ T ),
where A(t) = fx (x̃0 (t)), B(t) = fu (x̃0 (t)). Recall that
RS (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | ϕ′α (u0 (t))v(t) ≤ 0 (α ∈ Ia (u0 (t)), t ∈ T ),
ϕ′β (u0 (t))v(t) = 0 (β ∈ Q, t ∈ T )}
where
L(x0 , u0 ) := {(y, v) ∈ Z | ẏ(t) = A(t)y(t) + B(t)v(t) (t ∈ T ),
y(t0 ) = y(t1 ) = 0}
456
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
and therefore, given (p, µ) ∈ M (x0 , u0 ), since
S1 [= S1 (µ)] := {(x, u) ∈ S | ϕα (u(t)) = 0 (α ∈ R, µα (t) > 0, t ∈ T )},
it follows that
RS1 (x0 , u0 ) = {(y, v) ∈ RS (x0 , u0 ) | ϕ′α (u0 (t))v(t) = 0
(α ∈ R, µα (t) > 0, t ∈ T )}.
Theorem 3.16 Suppose (x0 , u0 ) ∈ S and (p, µ) ∈ M (x0 , u0 ). Then
RS1 (µ) (x0 , u0 ) = {(y, v) ∈ RS (x0 , u0 ) | I ′ (x0 , u0 ; y, v) = 0}.
In particular, RS1 (µ) (x0 , u0 ) = RS1 (ν) (x0 , u0 ) for any (q, ν) ∈ M (x0 , u0 ).
Proof: We proceed first as in the proof of Theorem 3.5. Define
K(x, u) := hp(t1 ), ξ1 i − hp(t0 ), ξ0 i +
Z t1
t0
F (t, x(t), u(t))dt ((x, u) ∈ Z)
where, for all (t, x, u) ∈ T × Rn × Rm ,
F (t, x, u) := L(t, x, u) − hp(t), f (t, x, u)i + hµ(t), ϕ(u)i − hṗ(t), xi.
Since F (t, x, u) = −H(t, x, u, p(t), µ(t), 1) − hṗ(t), xi, we have
Fx (x̃0 (t)) = −Hx (x̃0 (t), p(t), µ(t), 1) − ṗ∗ (t) = 0,
Fu (x̃0 (t)) = −Hu (x̃0 (t), p(t), µ(t), 1) = 0
and, consequently, K ′ (x0 , u0 ; y, v) = 0 for all (y, v) ∈ Z. From this equality,
we obtain that
0 =
Z t1
t0
′
{[Lx (x̃0 (t)) − p∗ (t)A(t) − ṗ∗ (t)]y(t) +
[Lu (x̃0 (t)) − p∗ (t)B(t) + µ∗ (t)ϕ′ (u0 (t))]v(t)}dt
= I (x0 , u0 ; y, v) −
Z t1
t0
{p∗ (t)(A(t)y(t) + B(t)v(t)) +
ṗ∗ (t)y(t)}dt +
Z t1
t0
µ∗ (t)ϕ′ (u0 (t))v(t)dt.
If (y, v) ∈ L(x0 , u0 ), this implies that
I ′ (x0 , u0 ; y, v) +
Z t1
t0
µ∗ (t)ϕ′ (u0 (t))v(t)dt = 0.
457
Critical cones for regular controls
If also (y, v) ∈ RS (x0 , u0 ), we have
I ′ (x0 , u0 ; y, v) +
Z t1 X
t0 α∈R
µα (t)ϕ′α (u0 (t))v(t)dt = 0.
Since µα (t)ϕ′ (u0 (t))v(t) ≤ 0 (α ∈ R, t ∈ T ), we conclude that (y, v) belongs
to the set
{(y, v) ∈ RS (x0 , u0 ) | I ′ (x0 , u0 ; y, v) = 0}
if and only if ϕ′ (u0 (t))v(t) = 0 (α ∈ R, µα (t) > 0, t ∈ T ), that is, if and only
if (y, v) ∈ RS1 (µ) (x0 , u0 ).
4
Lagrangian independent of the state
The particular case where the Lagrangian does not depend on the state variable
has been studied in a recent paper [27]. It provides an excellent illustration
of how the theory of first and second order necessary conditions for the finite
dimensional case can be applied to optimal control problems. Moreover, it
implies a second order condition stronger than the one given in Conjecture 3.15
since the assumption of normality of a solution (x0 , u0 ) relative to S1 implies the
nonnegativity of the second variation in a set which generally strictly contains
RS1 (x0 , u0 ). In this section we derive some consequences of this result (see also
[13, 14], where the constraints may depend not only on the control functions
but also on the time variable).
Let us then consider the case where L(t, x, u) = L(t, u), i.e., L does not
depend on x. Our problem P(S) is that of minimizing the functional I(u) =
R t1
t0 L(t, u(t))dt on S, that is, subject to (x, u) ∈ Z and
ẋ(t)
= f (t, x(t), u(t)) (t ∈ T );
x(t0 ) = ξ0 , x(t1 ) = ξ1 ;
ϕα (u(t)) ≤ 0, ϕβ (u(t)) = 0 (α ∈ R, β ∈ Q, t ∈ T ).
Our assumptions include that L, f and ϕ are C 2 and the q×(m+r)-dimensional
matrix
∂ϕi
δiα ϕα
∂uk
(i = 1, . . . , q; α = 1, . . . , r; k = 1, . . . , m)
has rank q on the set of points u ∈ Rm satisfying ϕα (u) ≤ 0 (α ∈ R), ϕβ (u) = 0
(β ∈ Q).
Let C := {u ∈ Um | (t, u(t)) ∈ A (t ∈ T )}, where
A = {(t, u) ∈ T × Rm | ϕα (u) ≤ 0 (α ∈ R), ϕβ (u) = 0 (β ∈ Q)}.
458
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
Denote by (C) the problem of minimizing I(u) = tt01 L(t, u(t))dt over C. Note
that, if (x0 , u0 ) is admissible for P(S) and u0 solves (C), that is, I(u0 ) ≤ I(u)
for all u ∈ C, then (x0 , u0 ) solves P(S), that is, I(u0 ) ≤ I(u) for all u ∈ C such
that, if ẋ(t) = f (t, x(t), u(t)) (t ∈ T ) and x(t0 ) = ξ0 , then x(t1 ) = ξ1 . Clearly,
the converse may not hold. Two simple examples illustrate this fact.
R
Example 4.1 Consider the problem of minimizing I(u) =
to
R1
0
u(t)dt subject
ẋ(t) = u(t), x(0) = x(1) = 0, u(t) ≤ 0.
Then (x0 , u0 ) ≡ (0, 0) solves P(S), being the only admissible process, but
u0 ≡ 0 does not solve (C), that is, it does not minimize I over the set
C = {u ∈ U1 | u(t) ≤ 0 (t ∈ T )}.
Actually, in this example, u0 maximizes I on C. Note also that (x0 , u0 ) is not
normal with respect to S since
τ2 (u0 ) = {h ∈ R | h ≤ 0}
and so the system
ż(t) = 0,
z(t)h ≤ 0 for all h ∈ τ2 (u0 (t)) (t ∈ T )
has nontrivial solutions.
Example 4.2 Consider the problem of minimizing I(u) =
subject to
ẋ(t) = u2 (t), x(0) = x(2) = 0, u2 (t) ≤ 1.
R2
0
(t − 1)u(t)dt
Then (x0 , u0 ) ≡ (0, 0) solves P(S) but u0 ≡ 0 does not solve (C) since it does
not minimize I over the set
C = {u ∈ U1 | −1 ≤ u(t) ≤ 1 (t ∈ T )}.
In contrast with the previous example, here u0 does not minimize or maximize
I on C. As one readily verifies, the solution to (C) is given by the function
which equals 1 on [0, 1] and −1 on [1, 2] (for the maximum the signs are
reversed). As before, clearly (x0 , u0 ) is not normal with respect to S since
fu (t, x0 (t), u0 (t)) = 0.
Let us now state the result on second order conditions for problem P(S)
proved in [27].
Critical cones for regular controls
459
Theorem 4.3 Suppose (x0 , u0 ) solves P(S) and ∃(p, µ) ∈ X × Uq satisfying
a. µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T );
b. ṗ(t) = −A∗ (t)p(t) (t ∈ T );
c. p∗ (t)B(t) = Lu (t, u0 (t)) + µ∗ (t)ϕ′ (u0 (t)) (t ∈ T ).
Suppose also that u0 solves (C), and consider the following statements:
i. (x0 , u0 ) is normal relative to S1 (µ).
ii. p ≡ 0.
iii. J((x0 , u0 , p, µ); (y, v)) ≥ 0 for all (y, v) ∈ Z with v(t) ∈ τ1 (u0 (t), µ(t))
(t ∈ T ).
Then (i) ⇒ (ii) ⇒ (iii).
It should be noted that the assumptions in this theorem imply that the
second variation is nonnegative on the set of pairs (y, v) in Z satisfying v(t) ∈
τ1 (u0 (t), µ(t)) (t ∈ T ). This holds in contrast with the conclusion of Conjecture
3.15 where the nonnegativity of that function requires from (y, v) to also satisfy
the linear equation ẏ(t) = A(t)y(t) + B(t)v(t) (t ∈ T ) together with the
endpoint conditions y(t0 ) = y(t1 ) = 0.
Observe also that, as Example 4.1 illustrates, we may have a solution
(x0 , u0 ) to our problem P(S) for which u0 does not solve problem (C) and,
moreover, u0 maximizes I on C. Also, in that example, u0 ≡ 0 maximizes
(t, u) 7→ u on A and, in Example 4.2, u0 ≡ 0 minimizes (t, u) 7→ u2 on A.
Apart from this, recall that, in Theorem 4.3, we are assuming normality relative to S1 which in turn implies normality relative to S. These facts motivate
the following result, assuming that f (t, x, u) = f (t, u), i.e., f is independent
on x.
Proposition 4.4 Let (x0 , u0 ) be an admissible process and suppose that, for
some i ∈ {1, . . . , n}, f i (t, u0 (t)) ≤ f i (t, u) (or ≥ f i (t, u)) whenever (t, u) ∈ A.
Then (x0 , u0 ) is not normal relative to S.
Proof: Suppose, without loss of generality, that f i (t, u0 (t)) ≤ f i (t, u) whenever (t, u) ∈ A. Then there exists a unique ν ∈ Uq such that
fui (t, u0 (t)) + ν ∗ (t)ϕ′ (u0 (t)) = 0 (t ∈ T ).
Moreover, να (t) ≥ 0 and να (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ). By definition,
τ2 (u) := {h ∈ Rm | ϕ′i (u)h ≤ 0 (i ∈ Ia (u)), ϕ′j (u)h = 0 (j ∈ Q)}
and (x0 , u0 ) is a normal point of S if z ≡ 0 is the only solution to the system
ż(t) = 0,
z(t)fu (t, u0 (t))h ≤ 0 for all h ∈ τ2 (u0 (t)) (t ∈ T ).
460
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
Let zi (t) = −1 and zj (t) = 0 (j 6= i). Then
z(t)fu (t, u0 (t))h = −fui (t, u0 (t))h = ν ∗ (t)ϕ′ (u0 (t))h ≤ 0
for all h ∈ τ2 (u0 (t)) and therefore the above system has nontrivial solutions.
This proves the claim.
The following example illustrates an application of Theorem 4.3, namely,
the verification of normality by simply exhibiting an extremal (x0 , u0 , p, µ) for
which p 6≡ 0.
Example 4.5 Consider the problem of minimizing I(x, u) =
ject to (x, u) ∈ Z and
R1
0
u2 (t)dt sub-
= u21 (t) + u2 (t) − u3 (t) (t ∈ [0, 1]);
x(0) = x(1) = 0;
u2 (t) ≥ 0, u3 (t) ≥ 0 (t ∈ [0, 1]).
ẋ(t)
In this case T = [0, 1], n = 1, m = 3, r = q = 2, ξ0 = ξ1 = 0 and, for all
t ∈ T , x ∈ R, and u = (u1 , u2 , u3 ),
L(t, x, u) = u2 ,
f (t, x, u) = u21 + u2 − u3 ,
ϕ1 (u) = −u2 ,
ϕ2 (u) = −u3 .
We have H(t, x, u, p, µ) = p(u21 + u2 − u3 ) − u2 + µ1 u2 + µ2 u3 and so
Hu (t, x, u, p, µ) = (2pu1 , p − 1 + µ1 , −p + µ2 )
Clearly (x0 , u0 ) ≡ (0, 0) is a solution to both problems P(S) and (C). Let
µ = (µ1 , µ2 ) ≡ (0, 1) and p ≡ 1. Then, as one readily verifies, (x0 , u0 , p, µ) ∈ E,
that is, the first conditions of Theorem 4.3 hold. Since p 6≡ 0, we conclude that
the solution (x0 , u0 ) is not normal relative to S1 . Actually, for this example, the
pair (x0 , u0 ) is normal relative to S and, if v = (v1 , v2 , v3 ) ≡ (1, 0, 0) and y ≡ 0,
then (y, v) solves ẏ(t) = v2 (t) − v3 (t) (t ∈ T ), y(0) = y(1) = 0, v(t) ∈ τ0 (u0 (t)),
and J((x0 , u0 , p, µ); (y, v)) = −2 < 0.
Theorem 4.3 gives rise to several surmises and questions which we will be
solved completely by means of some illustrative examples. Some of these issues
were first posed, and left open, in [27].
A natural question is if, in Theorem 4.3, the assumption that u0 solves
(C) becomes redundant once normality relative to S1 is imposed. In other
words, if (x0 , u0 ) solves P(S), (x0 , u0 , p, µ) is an extremal, and (i) holds, then
necessarily u0 solves (C). Our first example does answer this question. We
provide a solution (x0 , u0 ) to P(S) together with an extremal which not only
satisfies (i) but (x0 , u0 ) is strongly normal, and u0 is not a solution to (C).
461
Critical cones for regular controls
Example 4.6 (x0 , u0 ) solves P(S) and is strongly normal, but u0 does not
solve (C).
R
Consider the problem P(S) of minimizing I(x, u) = 01 u(t)dt subject to
ẋ(t) = u(t)x(t) (t ∈ [0, 1]), x(0) = 1, x(1) = e, u(t) ≥ 0 (t ∈ [0, 1]).
For this problem T = [0, 1], ξ0 = 1, ξ1 = e,
L(t, u) = u,
f (t, x, u) = ux,
ϕ(u) = −u.
Also S = {(x, u) ∈ D | u(t) ≥ 0 (t ∈ T )} where
D = {(x, u) ∈ Z | ẋ(t) = u(t)x(t), x(0) = 1, x(1) = e}.
To begin with note Rthat, if (x, u) is admissible,
then x(t) = exp( 0t u(s)ds)
R1
1
and x(1) = e = exp( 0 u(t)dt) and so 0 u(t)dt = I(x, u) = 1. Therefore
(x0 (t), u0 (t)) := (et , 1) (t ∈ T ) is a solution to the problem P(S) and, clearly,
u0 ≡ 1 is not a solution to (C). Now, fx (x̃0 (t)) = 1, fu (x̃0 (t)) = et , and
ϕ(u0 (t)) = −1 < 0. Since, for this solution, there are no active constraints,
S0 = D. Clearly (x0 , u0 ) is a normal process of S0 since, given (p, µ) satisfying
R
µ(t)(−1) = 0, ṗ(t) = −p(t), et p(t) = −µ(t) = 0 (t ∈ [0, 1])
then p ≡ 0.
This example illustrates some more features of the theorem. In particular,
we can exhibit a pair (p, µ) such that (x0 , u0 , p, µ) is an extremal for which (i)
is satisfied, but (ii) and (iii) do not hold.
Example 4.7 (x0 , u0 ) solves P(S), (p, µ) ∈ M (x0 , u0 ) is such that, in Theorem 4.3, (i) is satisfied but (ii) and (iii) do not hold. Moreover,
J((x0 , u0 , p, µ); (y, v)) = 0 for all (y, v) ∈ L(x0 , u0 ).
Consider the previous example with (x0 , u0 ) defined as above. If p(t) = e−t
and µ ≡ 0 then (x0 , u0 , p, µ) ∈ E since ṗ(t) = −p(t) and et p(t) = 1 (t ∈ T ).
Also, as seen before, (i) holds but not (ii). Now, we have H = pux − u + µu
and so
Hu = px − 1 + µ,
Hx = pu,
Huu = Hxx = 0,
Hux = Hxu = p
all evaluated at (t, x, u, p, µ). Thus 2Ω(t, y, v) = −2p(t)yv and so
J((x0 , u0 , p, µ); (y, v)) = −2
Z 1
0
e−t y(t)v(t)dt.
462
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
Since there are no active constraints, τ0 (u0 (t)) = τ1 (u0 (t), µ(t)) = τ2 (u0 (t)) =
R. Thus, if y ≡ v ≡ 1 then clearly (iii) fails to hold.
Let us now show that J((x0 , u0 , p, µ); (y, v)) = 0 for all (y, v) ∈ L(x0 , u0 ).
Let (y, v) ∈ L(x0 , u0 ) and note that, by definition,
ẏ(t) = y(t) + et v(t) (t ∈ T ) and y(0) = y(1) = 0.
Let g(t) =
by parts
Rt
0
v(s)ds. Then y(t) = et g(t) and so g(0) = g(1) = 0. Integrating
Z 1
0
and therefore
g(t)v(t)dt = g
2
(t)|10
−
J((x0 , u0 , p, µ), (y, v)) = −2
Z 1
Z 1
0
0
g(t)v(t)dt
g(t)v(t)dt = 0.
Now, this particular solution gives rise to a different conjecture. It seems
to be crucial, for this example, the fact that there are no active constraints.
However, we can provide below a different solution for which the previous
conclusions also hold but the constraint is active on a subinterval of T .
Example 4.8 The conclusions of Examples 4.6 and 4.7 hold in the presence
of active constraints.
Consider again problem P(S) of Example 4.7. Let a = 1/2 and define the
process (x1 , u1 ) by
x1 (t) =
( 2t
e
e
if t ∈ [0, a]
if t ∈ [a, 1]
u1 (t) =
(
2 if t ∈ [0, a]
0 if t ∈ (a, 1].
Clearly (x1 , u1 ) is a solution to P(S) but not to (C). To check for normality,
suppose (p, µ) ∈ X × U1 satisfies
i. µ(t)ϕ(u1 (t)) = 0 (t ∈ T );
ii. ṗ(t) = −fx (x̃1 (t))p(t) and fu (x̃1 (t))p(t) = ϕ′ (u1 (t))µ(t) (t ∈ T ).
By (i), µ(t) = 0 on [0, a]. By (ii), since fx (x̃1 (t)) = u1 (t), fu (x̃1 (t)) = x1 (t),
and ϕ′ (u1 (t)) = −1, we have
ṗ(t) = −u1 (t)p(t),
x1 (t)p(t) = −µ(t) (t ∈ T )
which implies, in particular, that e2t p(t) = 0 on [0, a] and ṗ(t) = 0 on (a, 1].
Thus p(t) = 0 on [0, a] and p(t) is constant on (a, 1]. By continuity, p ≡ 0 and
so (x1 , u1 ) is strongly normal.
Now, for this solution, an extremal (x1 , u1 , p, µ) must satisfy
a. µ(t) ≥ 0 and µ(t)ϕ(u1 (t)) = 0 (t ∈ T );
b. ṗ(t) = −u1 (t)p(t) and p(t)x1 (t) = 1 − µ(t) (t ∈ T ).
463
Critical cones for regular controls
These conditions are clearly satisfied by (p, µ) with µ ≡ 0 and
p(t) =
( −2t
e
e
−1
if t ∈ [0, a]
if t ∈ (a, 1].
Observe now that, since ϕ(u1 (t)) = −u1 (t) (t ∈ T ), the constraint is inactive
on [0, a] but active on (a, 1]. Thus we have
τ0 (u1 (t)) =
(
R
if t ∈ [0, a]
{h | −h = 0} if t ∈ (a, 1]
τ2 (u1 (t)) =
(
R
if t ∈ [0, a]
{h | −h ≤ 0} if t ∈ (a, 1]
and τ1 (u1 (t)), µ(t)) = τ2 (u1 (t)) (t ∈ T ).
As before we have 2Ω(t, y, v) = −2p(t)yv and so, if y ≡ v ≡ 1, then
J((x1 , u1 , p, µ); (y, v)) < 0.
Now, (y, v) ∈ Z belongs to L(x1 , u1 ) if and only if y(0) = y(1) = 0 and
ẏ(t) =
Rt
and so, if g(t) =
0
(
2y(t) + e2t v(t) if t ∈ [0, a]
ev(t)
if t ∈ [a, 1]
v(s)ds, then
y(t) =
( 2t
e g(t)
eg(t)
if t ∈ [0, a]
if t ∈ (a, 1]
with g(0) = g(1) = 0. We reach the same conclusion as before since
J((x1 , u1 , p, µ), (y, v)) = −2
Z 1
0
g(t)v(t)dt = 0
for any (y, v) ∈ L(x1 , u1 ).
This last solution can be generalized by considering subintervals of length
an = 1/n and setting
xn (t) =
( nt
e
e
if t ∈ [0, an ]
if t ∈ [an , 1]
un (t) =
(
n if t ∈ [0, an ]
0 if t ∈ (an , 1].
Then (xn , un ) is a solution to the problem P(S) and essentially the same conclusions as before will follow.
We have thus shown, in particular, that we may have a strongly normal
solution (x0 , u0 ) to P(S) and u0 does not solve (C). In Example 4.6, however,
note that f depends on x and, again, this seems to be fundamental in the
analysis. In the following example we reach the same conclusion, but f does
not depend on x.
464
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
Example 4.9 The previous conclusions hold with the function f independent of x.
Consider the problem P(S) of minimizing
I(x, u) =
Z 1
0
{u1 (t) + u2 (t) + u3 (t)}dt
subject to (x, u) ∈ Z and
ẋ(t)
= u3 (t) (t ∈ [0, 1]);
x(0) = 0, x(1) = 1;
u1 (t) − 2u2 (t) ≤ 0, u2 (t) − u1 (t) ≤ 0 (t ∈ [0, 1]).
For this problem T = [0, 1], ξ0 = 0, ξ1 = 1,
L(t, u) = u1 + u2 + u3 ,
f (t, x, u) = u3 ,
ϕ1 (u) = u1 − 2u2 ,
ϕ2 (u) = u2 − u1 .
Clearly, (C) has no solution. Let x0 (t) := t and u0 (t) := (0, 0, 1) (t ∈ T ).
Then (x0 , u0 ) is a solution to problem P(S) since, for any (x, u) admissible,
u1 (t) + u2 (t) ≥ 0 (t ∈ T ) and therefore
I(x, u) ≥
Z 1
0
u3 (t)dt = 1 = I(x0 , u0 ).
It is also normal relative to S0 since
τ0 (u0 (t)) = {h ∈ R3 | h1 − 2h2 = 0, −h1 + h2 = 0} = {h | h1 = h2 = 0}
and so z ≡ 0 is the only solution to the system ż(t) = 0 and z(t)h3 = 0 for all
h3 ∈ R.
Let us now show that one can find an extremal for which (i) of Theorem
4.3 holds but not (ii). This clearly follows if p ≡ 1 and (µ1 , µ2 ) ≡ (2, 3), for
then (x0 , u0 , p, µ) ∈ E since ṗ(t) = 0 and
1 −2 0
−1 1 0
= (0, 0, 1) = p(t)fu (x̃0 (t)).
Lu (x̃0 (t)) + µ∗ (t)ϕ′ (u0 (t)) = (1, 1, 1) + (2, 3)
We end this paper with another important property derived from Theorem
4.3. Our final example provides a solution (x0 , u0 ) to a problem P(S) which is
normal relative to S and u0 solves (C), but (x0 , u0 ) is not normal relative to S1
(nor in consequence to S0 ) for any pair (p, µ) with (x0 , u0 , p, µ) an extremal.
Example 4.10 (x0 , u0 ) solves P(S), it is normal relative to S, and u0 solves
(C), but (x0 , u0 ) is not normal relative to S1 (µ) for any (p, µ) ∈ M (x0 , u0 ).
465
Critical cones for regular controls
Consider the problem P(S) of minimizing I(x, u) =
to
R2
0
(t − 1)u(t)dt subject
ẋ(t) = u(t) (t ∈ [0, 2]), x(0) = x(2) = 0, u2 (t) ≤ 1 (t ∈ [0, 2]).
Define the process (x0 , u0 ) as
x0 (t) =
t
if t ∈ [0, 1]
2 − t if t ∈ [1, 2]
u0 (t) =
(
1
if t ∈ [0, 1]
−1 if t ∈ (1, 2].
Clearly (x0 , u0 ) is a solution to problem P(S) and u0 solves (C). It is not
normal relative to S0 since, for example, p ≡ 2 and µ ≡ u0 show that the
system ṗ(t) = 0 and p(t) = 2u0 (t)µ(t) (t ∈ T ) has nonnull solutions. The
conclusion also follows straightforwardly by noting that τ0 (u0 (t)) = {0} and so
the system ż(t) = 0 and z(t)h ≤ 0 for all h ∈ τ0 (u0 (t)) has nonnull solutions.
Now, note that
τ2 (u0 (t)) =
(
{h | h ≤ 0} if t ∈ [0, 1]
{h | h ≥ 0} if t ∈ (1, 2]
and so (x0 , u0 ) is normal relative to S since z ≡ 0 is the only solution of the
system ż(t) = 0 and z(t)h ≤ 0 for all h ∈ τ2 (u0 (t)). Finally, if (x0 , u0 , p, µ) is
an extremal, then
µ(t) ≥ 0, ṗ(t) = 0 and p(t) = t − 1 + 2µ(t)u0 (t) (t ∈ T ).
This implies, in particular, that p is constant, p ≥ t − 1 on [0, 1] and p ≤ t − 1
on (1, 2]. Hence p ≡ 0 and
µ(t) =
(
(1 − t)/2 if t ∈ [0, 1]
(t − 1)/2 if t ∈ (1, 2].
τ1 (u0 (t), µ(t)) =
(
{0}
if t ∈ [0, 1) ∪ (1, 2]
{h | h ≤ 0} if t = 1
In this event we have
and therefore (x0 , u0 ) is not normal relative to S1 (µ).
References
[1] A.V. Arutyunov, Smooth abnormal problems in extremum theory and
analysis, Russian Mathematical Surveys, 67 (2012), 403-457.
https://doi.org/10.1070/rm2012v067n03abeh004793
466
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
[2] A.V. Arutyunov, D.Y. Karamzin, Necessary conditions for a weak minimum in an optimal control problem with mixed constraints, Differential
Equations, 41 (2005), 1532-1543.
https://doi.org/10.1007/s10625-005-0311-2
[3] A.V. Arutyunov, F.L. Pereira, Second-order necessary optimality conditions for problems without a priori normality assumptions, Mathematics
of Operations Research, 31 (2006), 1-12.
https://doi.org/10.1287/moor.1050.0172
[4] A.V. Arutyunov, Y.S. Vereshchagina, On necessary second-order conditions in optimal control problems, Differentsial’nye Uravneniya 38 (2002),
1443-1450. (Russian). Translation in Differential Equations 38 (2002),
1531-1540. https://doi.org/10.1023/a:1023624602611
[5] J.P. Aubin, H. Frankowska, Set-Valued Analysis, Birkhäuser, Boston,
1990.
[6] J.A. Becerril, J.F. Rosenblueth, Necessity for isoperimetric inequality
constraints, Discrete and Continuous Dynamical Systems - Series A, 37
(2017), 1129-1158. https://doi.org/10.3934/dcds.2017047
[7] J.A. Becerril, J.F. Rosenblueth, The importance of being normal, regular
and proper in the calculus of variations, Journal of Optimization Theory
& Applications, 172 (2017), 759-773.
https://doi.org/10.1007/s10957-017-1070-y
[8] A. Ben-Tal, Second-order and related extremality conditions in nonlinear
programming, Journal of Optimization Theory & Applications, 31 (1980),
143-165. https://doi.org/10.1007/bf00934107
[9] J.F. Bonnans, X. Dupuis, L. Pfeiffer, Second-order necessary conditions in
Pontryagin form for optimal control problems, SIAM Journal of Control
& Optimization, 52 (2014), 3887-3916.
https://doi.org/10.1137/130923452
[10] J.F. Bonnans, N.P. Osmolovskii, Quadratic growth conditions in optimal control problems, Contemporary Mathematics, 514 (2010), 85-98.
https://doi.org/10.1090/conm/514/10100
[11] J.F. Bonnans, A. Shapiro, Perturbation Analysis of Optimization Problems, Springer, New York, 2000.
https://doi.org/10.1007/978-1-4612-1394-9
[12] F. Clarke, Functional Analysis, Calculus of Variations and Optimal Control, Springer, London, 2013.
Critical cones for regular controls
467
[13] K.L. Cortez, J.F. Rosenblueth, A second order constraint qualification
for certain classes of optimal control problems, WSEAS Transactions on
Systems and Control, 11 (2016), 419-424.
[14] K.L. Cortez, J.F. Rosenblueth, Extended critical directions for timecontrol constrained problems, International Journal of Circuits, Systems
and Signal Processing, 11 (2017), 1-11.
[15] M.R. de Pinho, J.F. Rosenblueth, Mixed constraints in optimal control:
an implicit function theorem approach, IMA Journal of Mathematical
Control and Information, 24 (2007) 197-218.
https://doi.org/10.1093/imamci/dnl008
[16] E.G. Gilbert, D.S. Bernstein, Second order necessary conditions in optimal
control: accessory-problem results without normality conditions, Journal
of Optimization Theory & Applications, 41 (1983), 75-106.
https://doi.org/10.1007/bf00934437
[17] G. Giorgi, A. Guerraggio, J. Thierfelder, Mathematics of Optimization:
Smooth and Nonsmooth Case, Elsevier, Amsterdam, 2004.
[18] M.R. Hestenes, Calculus of Variations and Optimal Control Theory, John
Wiley, New York, 1966.
[19] M.R. Hestenes, Optimization Theory, The Finite Dimensional Case, John
Wiley, New York, 1975.
[20] D. Hoehener, Variational approach to second-order optimality conditions
for control problems with pure state constraints, SIAM Journal of Control
& Optimization, 50 (2012), 1139-1173.
https://doi.org/10.1137/110828320
[21] E. Levitin, A. Milyutin, N.P. Osomolovskii, Conditions of high order for
a local minimum for problems with constraints, Russian Math Surveys,
33 (1978), 97-168. https://doi.org/10.1070/rm1978v033n06abeh003885
[22] P.D. Loewen, H. Zheng, Generalized conjugate points for optimal control problems, Nonlinear Analysis, Theory, Methods & Applications, 22
(1994), 771-791. https://doi.org/10.1016/0362-546x(94)90226-7
[23] E.J. McShane, The Lagrange Multiplier Rule, The American Mathematical Monthly, 80 (1973), 922-925.
https://doi.org/10.1080/00029890.1973.11993409
468
Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth
[24] A.A. Milyutin, N.P. Osmolovskii, Calculus of Variations and Optimal
Control, Translations of Mathematical Monographs, Vol. 180, American
Mathematical Society, Providence, Rhode Island, 1998.
https://doi.org/10.1090/mmono/180
[25] N.P. Osmolovskii, Second order conditions for a weak local minimum in
an optimal control problem (necessity, sufficiency), Soviet Math. Dokl.,
16 (1975), 1480-1484.
[26] N.P. Osmolovskii, H. Maurer, Applications to Regular and Bang-Bang
Control, Second-Order Necessary and Sufficient Optimality Conditions
in Calculus of Variations and Optimal Control, Advances in Design and
Control, SIAM, Philadelphia, 2012.
https://doi.org/10.1137/1.9781611972368
[27] J.F. Rosenblueth, Modified Critical Directions for Inequality Control Constraints, WSEAS Transactions on Systems and Control, 10 (2015), 215227.
[28] J.F. Rosenblueth, G. Sánchez Licea, Cones of Critical Directions in Optimal Control, International Journal of Applied Mathematics and Informatics, 7 (2013), 55-67.
[29] G. Stefani, P.L. Zezza, Optimality conditions for a constrained control
problem, SIAM Journal on Control & Optimization, 34 (1996), 635-659.
https://doi.org/10.1137/s0363012994260945
[30] J. Warga, A second-order Lagrangian condition for restricted control problems, Journal of Optimization Theory & Applications, 24 (1978), 475-483.
https://doi.org/10.1007/bf00932890
Received: August 17, 2018; Published: September 19, 2018