Academia.eduAcademia.edu

Critical cones for regular controls with inequality constraints

2018, International Journal of Mathematical Analysis

Based on the notions of normality and regularity for constrained problems in optimization, a conjecture on second order necessary conditions related to different critical cones is posed for a wide range of optimal control problems involving equality and inequality constraints. Several properties of the sets and functions delimiting the problem are derived and, for a specific case where the surmise has been proved correct, some fundamental questions derived from this result are completely solved.

International Journal of Mathematical Analysis Vol. 12, 2018, no. 10, 439 - 468 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ijma.2018.8856 Critical Cones for Regular Controls with Inequality Constraints Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth IIMAS, Universidad Nacional Autónoma de México Apartado Postal 20-126, CDMX 01000, México Copyright c 2018 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth. This article is distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Abstract Based on the notions of normality and regularity for constrained problems in optimization, a conjecture on second order necessary conditions related to different critical cones is posed for a wide range of optimal control problems involving equality and inequality constraints. Several properties of the sets and functions delimiting the problem are derived and, for a specific case where the surmise has been proved correct, some fundamental questions derived from this result are completely solved. Mathematics Subject Classification: 49K15 Keywords: Nonlinear programming, optimal control, second order necessary conditions, normality, regularity 1 Introduction: three main questions When dealing with the theory of second order necessary conditions for constrained problems in optimization, one usually faces three fundamental questions which can be summarized in these words: who (that is, which function satisfies the conditions), when (under what assumptions), and where (on which set do they hold). For constrained problems in the finite dimensional case, they have been answered in a satisfactory way in terms of Lagrange multipliers but, 440 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth even for such a well-known theory, the functions, the assumptions, and the sets where the conditions hold, may vary from text to text. In this paper we shall be concerned with these same questions in the context of optimal control. Though second order conditions for such problems can be found in the literature, some fundamental questions still remain open, and our aim is to give answers to some of them. For the optimal control problem we shall deal with, the control functions are restricted to satisfy equality and inequality constraints for all points t in a fixed compact time interval. As pointed out in [12], these constraints make the problem much more complex than the mathematical programming problem, or even the isoperimetric problem (see [6, 7]), in part because one deals with infinitely many constraints, one for each t. In some cases like in [21, 25], the main results on second order conditions are not proved and, quoting [21], “the derivation of the conditions is very special and difficult.” There is an extensive literature explaining the importance of deriving second order conditions for optimal control problems, both from a theoretical and an applied point of view. Particular attention deserves the research developed in [1–4] where one of the main features is that the necessary conditions obtained make sense, and are derived, without a priori assumptions of normality, and second order conditions are expressed in terms of the maximum of a quadratic form over certain multipliers, for all the elements of the critical cone of the original set of constraints. Similarly, in [9–11, 24, 26], all essential references in the subject, second order necessary conditions in Pontryagin form are derived for problems with pure state and mixed (control-state) constraints, and varying the set of multipliers (those satisfying Pontryagin’s principle, or the larger set of Lagrange multipliers) where the maximum of the quadratic forms on the set of critical variations is taken. For a more complete study of the subject, see also [16, 18, 20, 22, 27–30] and references therein. Our approach, based on the notions of normal and regular controls, is rather different to the one found in those references. To begin with, one finds in those references results from an abstract optimization theory on Banach spaces applied to the control problem posed over L∞ -controls, a technique which does not work in our setting, where the fixed endpoint Lagrange control problem studied is posed over piecewise C 1 trajectories and piecewise continuous controls (the same setting can be found, for example, in [18, 22, 24]). Secondly, we shall be able to establish, for certain cases, the nonnegativity of a certain quadratic form under weaker assumptions and on larger sets than those appearing in those texts (see [22, 28], where a similar set of critical directions for our problem is proposed). Finally, the nature of our necessary conditions differs from the previous ones in that the nonnegativity of the second variation holds for the same multipliers without the need to take a maximum over different sets of multipliers. 441 Critical cones for regular controls This paper is organized as follows. To clearly situate the contributions of the article, we provide in Section 2 a summary of the main features, in terms of the notions of regularity and normality, that occur in the finite dimensional case. This is crucial in our exposition, not only because some of the results given are used later, but also due to the way it is presented and for comparison reasons with the infinite dimensional case. In Section 3 we pose the optimal control problem we shall deal with and follow a way of reasoning similar to that of the previous section, leading to a natural conjecture on second order conditions, as well as to a new fundamental result related to a characterization of a certain set of critical directions. A particular case, covering a wide range of problems and recently derived in [27], is studied in Section 4, where we solve some important surmises and questions related to the theory of second order necessary conditions. 2 The finite dimensional case In this section we shall deal with constrained minimum problems in finite dimensional spaces. We shall state, in a succinct and clear way, first and second order necessary conditions for optimality in terms of Lagrange multiplier rules. The approach we follow is based on the notions of regularity and normality. For a full account of these ideas we refer to [18, 19]. Suppose we are given a set S ⊂ Rn and a function f mapping Rn to R, and consider the problem, which we label P(S), of minimizing f on S. The notion of tangent cone given below is due to Hestenes [18] and, as shown in [17], it is equivalent to the one introduced by Bouligand (1932), also known as the contingent cone to S at x0 . Other authors such as Bazaraa, Goode, Nashed, Kurcyusz, Rockafellar, Saks, Rogak and many more, have given various equivalent definitions of such a cone (see [5, 17]). Definition 2.1 We shall say that a sequence {xq } ⊂ Rn converges to x0 in the direction h if h is a unit vector, xq 6= x0 , and lim |xq − x0 | = 0, q→∞ lim q→∞ xq − x0 = h. |xq − x0 | The tangent cone of S at x0 , denoted by TS (x0 ), is the (closed) cone determined by the unit vectors h for which there exists a sequence {xq } in S converging to x0 in the direction h. Equivalently (see [19]), TS (x0 ) is the set of all h ∈ Rn for which there exist a sequence {xq } in S and a sequence {tq } of positive numbers such that lim tq = 0, q→∞ lim q→∞ xq − x0 = h. tq 442 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth Clearly, if {xq } converges to x0 in the direction h and f has a differential at x0 , then f (xq ) − f (x0 ) lim = f ′ (x0 ; h). q→∞ |xq − x0 | If f has a second differential at x0 , then f (xq ) − f (x0 ) − f ′ (x0 ; xq − x0 ) 1 = f ′′ (x0 ; h). 2 q→∞ |xq − x0 | 2 lim From these facts, necessary conditions for P(S) follow straightforwardly. Theorem 2.2 Suppose x0 solves P(S) locally. If f has a differential at x0 , then f ′ (x0 ; h) ≥ 0 for all h ∈ TS (x0 ). If f has a second differential at x0 and f ′ (x0 ) = 0, then f ′′ (x0 ; h) ≥ 0 for all h ∈ TS (x0 ). Consider now problem P(S) with S defined by inequality and equality constraints. We are given functions f, gi : Rn → R (i ∈ A ∪ B) and S = {x ∈ Rn | gα (x) ≤ 0 (α ∈ A), gβ (x) = 0 (β ∈ B)} where A = {1, . . . , p}, B = {p + 1, . . . , m}. The cases p = 0 and p = m are to be given the obvious interpretations. Assume f and gi are continuous on a neighborhood of x0 ∈ S and possess second differentials at x0 . Definition 2.3 Define the set of active indices at x0 by I(x0 ) := {α ∈ A | gα (x0 ) = 0} and the set of vectors satisfying the tangential constraints at x0 by RS (x0 ) := {h ∈ Rn | gα′ (x0 ; h) ≤ 0 (α ∈ I(x0 )), gβ′ (x0 ; h) = 0 (β ∈ B)}. Note that TS (x0 ) ⊂ RS (x0 ). We shall say that x0 is a regular point of S if TS (x0 ) = RS (x0 ). The first order Lagrange multiplier rule is a consequence of Theorem 2.2 and some basic results on linear functionals. Let us define the set E of extremals as the set of all (x0 , λ) ∈ S × Rm such that i. λα ≥ 0 and λα gα (x0 ) = 0 (α ∈ A). ii. If F (x) := f (x) + hλ, g(x)i then F ′ (x0 ) = 0. Theorem 2.4 Suppose x0 solves P(S) locally. If x0 is a regular point of S, then ∃λ ∈ Rm such that (x0 , λ) ∈ E. The second order Lagrange multiplier rule follows by a simple application of Theorem 2.2. 443 Critical cones for regular controls Theorem 2.5 Suppose (x0 , λ) ∈ E, F (x) = f (x) + hλ, g(x)i, and S1 = {x ∈ S | F (x) = f (x)}. If x0 solves P(S) locally, then F ′′ (x0 ; h) ≥ 0 for all h ∈ TS1 (x0 ). Note that S1 = {x ∈ S | gα (x) = 0 (α ∈ Γ)} where Γ = {α ∈ A | λα > 0}, and RS1 (x0 ) = {h ∈ RS (x0 ) | gα′ (x0 ; h) = 0 (α ∈ Γ)} = {h ∈ RS (x0 ) | f ′ (x0 ; h) = 0}. In general, it may be difficult to test for regularity and one usually requires some criteria that implies that condition. A simple criterion is that of normality, defined in [19] as follows. Definition 2.6 We shall say that x0 ∈ S is normal relative to S if λ = 0 is the only solution to i. λα ≥ 0 and λα gα (x0 ) = 0 (α ∈ A). P ′ ii. m 1 λi gi (x0 ) = 0. To be precise, we should say in this definition that x0 ∈ S is normal relative to the functions defining the set S and not relative to the set S defined by those functions. In other words, if λ = 0 is the only solution to the above system, then x0 ∈ S is normal relative to the function g and the sets A and B of indices for inequality and equality constraints respectively. However, following the definition introduced in [18, 19], we shall use the above definition, and no confusion should arise if S is replaced with a different set defined also in terms of inequality and equality constraints. In [19], it is proved that x0 is normal relative to S if and only if the set ′ {gβ (x0 ) | β ∈ B} is linearly independent and, if p > 0, ∃h such that gα′ (x0 ; h) < 0 (α ∈ I(x0 )), gβ′ (x0 ; h) = 0 (β ∈ B). This characterization is used in [19] to prove the following, crucial result. Theorem 2.7 If x0 is a normal point of S then x0 is a regular point of S. The following extended rule, the Fritz John necessary optimality condition, yields in a natural way the above definition of normality (see [23] for a simple proof based on the theory of augmentability). Theorem 2.8 Suppose x0 solves P(S) locally. Then ∃λ0 ≥ 0 and λ ∈ Rm , not both zero, such that i. λα ≥ 0 and λα gα (x0 ) = 0 (α ∈ A). ii. If F0 (x) := λ0 f (x) + hλ, g(x)i then F0′ (x0 ) = 0. 444 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth Clearly, in this theorem, if x0 is also a normal point of S then λ0 > 0 and the multipliers can be chosen so that λ0 = 1. The real numbers λ1 , . . . , λm satisfying the first order Lagrange multiplier rule (or the Karush-Kuhn-Tucker conditions) are called the Kuhn-Tucker or Lagrange multipliers, and the function F is the standard Lagrangian. In general, if x0 affords a local minimum to f on S, those conditions may not hold at x0 , and some additional assumptions should be imposed to guarantee the existence of Lagrange multipliers. Assumptions of this nature are usually referred to as constraint qualifications (see [17]) since they involve only the constraints and are independent of the geometric structure of the feasible set S (a broader definition, in terms of critical directions, is given in [8]). Equivalently, they correspond to conditions which assure the positiveness of the cost multiplier λ0 in Theorem 2.8. Both normality and regularity are constraint qualifications, as well as the characterization of normality mentioned above, known as the Mangasarian-Fromovitz constraint qualification. A combination of Theorems 2.5 and 2.7 yields the following basic result on second order necessary conditions. Theorem 2.9 Suppose (x0 , λ) ∈ E, F (x) = f (x)+hλ, g(x)i, and S1 = {x ∈ S | F (x) = f (x)}. If x0 solves P(S) locally and is a normal point of S1 , then F ′′ (x0 ; h) ≥ 0 for all h ∈ RS1 (x0 ). Let us turn now to a different notion of normality given in [18]. Definition 2.10 x0 ∈ S is s-normal relative to S if the linear functionals (α ∈ I(x0 ) ∪ B) in h are linearly independent. gα′ (x0 ; h) By an application of the implicit function theorem, and making use of the set of curvilinear tangent vectors of S at x0 given by CS (x0 ) := {h ∈ Rn | ∃δ > 0, x: [0, δ) → S such that x(0) = x0 , ẋ(0) = h}, one can easily show that s-normality implies that RS (x0 ) ⊂ CS (x0 ). Since CS (x0 ) ⊂ TS (x0 ) ⊂ RS (x0 ), this implies the following well-known result. Theorem 2.11 If x0 is s-normal relative to S then x0 is a regular point of S. This statement is precisely Lemma 10.1, Chapter 1 of [18] and it has an implication, in particular, on second order conditions. In Theorem 2.5, if x0 is a regular point of S1 , then F ′′ (x0 ; h) ≥ 0 for all h ∈ RS1 (x0 ). By Theorem 2.7, x0 is a regular point of S1 if it is a normal point of S1 , and thus Theorem 2.9 holds. In [18], on the other hand, the criterion for this condition to hold is for x0 to be s-normal relative to S since then it would be s-normal relative to 445 Critical cones for regular controls S1 , and hence a regular point of S1 . As one readily verifies, x0 ∈ S is s-normal relative to S if it is normal relative to the set S0 := {x ∈ Rn | gi (x) = 0 (i ∈ I(x0 ) ∪ B)} and thus the criterion used in [18] for regularity relative to S1 is in general stronger than that of normality relative to S1 used in [19]. Moreover, in most textbooks (see a list of well-known references in [8]) not only the assumption of normality relative to S1 is replaced by (the stronger assumption of) normality relative to S0 , but the set of tangential constraints RS1 (x0 ) is replaced by (the, in general, smaller set of tangential constraints) RS0 (x0 ) = {h ∈ Rn | gi′ (x0 ; h) = 0 (i ∈ I(x0 ) ∪ B)} This rather “well-worn result” (as Ben-Tal puts it [8]) is the following. Theorem 2.12 Suppose (x0 , λ) ∈ E and F (x) = f (x) + hλ, g(x)i. If x0 solves P(S) locally and is a normal point of S0 , then F ′′ (x0 ; h) ≥ 0 for all h ∈ RS0 (x0 ). As pointed out in [8], “The source of this weaker result can be attributed to the traditional way of treating the active inequality constraints as equality constraints.” Let us end this section with an example which illustrates some of the main features on second order necessary conditions explained above. It should be noted that, for this example, the solution x0 to the problem is a regular point of S so that, by Theorem 2.4, ∃λ such that (x0 , λ) is an extremal. However, in contrast with the conclusion of Theorem 2.9, we can exhibit h ∈ RS1 (x0 ) with F ′′ (x0 ; h) < 0. Example 2.13 Consider the problem of minimizing f (x, y) = y + x2 on the set S = {(x, y) | gα (x, y) ≤ 0 (α = 1, 2, 3)} where g1 (x, y) = −y, g2 (x, y) = cos x − y, g3 (x, y) = cos 2x − y. As one readily verifies, (0, 1) solves the problem. The set of active indices at this point is I(0, 1) = {2, 3}, and it is a regular point of S since TS (0, 1) = {(h, k) | k ≥ 0}, RS (0, 1) = {(h, k) | −k ≤ 0}. By Theorem 2.4, there exists λ = (λ1 , λ2 , λ3 ) with λ1 = 0 and λ2 , λ3 ≥ 0 such that, if F (x, y) = f (x, y) + hλ, g(x, y)i = y + x2 − (λ1 + λ2 + λ3 )y + λ2 cos x + λ3 cos 2x 446 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth then F ′ (0, 1) = 0. Since F ′ (x, y) = (2x − λ2 sin x − 2λ3 sin 2x, 1 − λ1 − λ2 − λ3 ) we have, with λ1 = 0, that F ′ (0, 1) = (0, 1 − λ2 − λ3 ) and so λ2 + λ3 = 1. Now, note that   2 − λ2 cos x − 4λ3 cos 2x 0 ′′ F (x, y) = 0 0 and therefore F ′′ (0, 1; h, k) = (2 − λ2 − 4λ3 )h2 . Let λ2 = 0 and λ3 = 1. Then RS1 (0, 1) = {(h, k) | −k ≤ 0, k = 0} and so, for any h 6= 0, we have (h, 0) ∈ RS1 (0, 1) but F ′′ (0, 1; h, 0) = −2h2 < 0. 3 Optimal control In this section we shall deal with a fixed endpoint Lagrange optimal control problem involving inequality and equality constraints in the control functions. We are interested in deriving, for this optimal control problem, a theory parallel to the one given in the previous section. We shall follow a similar line of thought and, in particular, try to see if a result analogous to Theorem 2.9, that is, second order conditions in terms of normality relative to the corresponding set S1 , can be established. Suppose we are given an interval T := [t0 , t1 ] in R, two points ξ0 , ξ1 in Rn , and functions L and f mapping T × Rn × Rm to R and Rn respectively, and ϕ = (ϕ1 , . . . , ϕq ) mapping Rm to Rq (q ≤ m). Denote by X the space of piecewise C 1 functions mapping T to Rn , by Uk the space of piecewise continuous functions mapping T to Rk (k ∈ N), set Z := X × Um , D := {(x, u) ∈ Z | ẋ(t) = f (t, x(t), u(t)) (t ∈ T ), x(t0 ) = ξ0 , x(t1 ) = ξ1 }, S := {(x, u) ∈ D | ϕα (u(t)) ≤ 0, ϕβ (u(t)) = 0 (α ∈ R, β ∈ Q, t ∈ T )} where R = {1, . . . , r}, Q = {r+1, . . . , q}, and consider the functional I: Z → R given by Z t1 I(x, u) := t0 L(t, x(t), u(t))dt ((x, u) ∈ Z). The problem we shall deal with, which (again, but no confusion should arise) we label P(S), is that of minimizing I over S. Implicit in the statement of the problem is a relatively open subset O of T × Rn for which the domain of L and f is O × Rm . Elements of Z will be called processes, and a process (x, u) is admissible if it belongs to S and Critical cones for regular controls 447 (t, x(t)) ∈ O (t ∈ T ). A process (x, u) solves P(S) (locally) if it is admissible and (upon shrinking O if necessary) we have I(x, u) ≤ I(y, v) for all admissible processes (y, v). Given (x, u) ∈ Z we shall use the notation (x̃(t)) to represent (t, x(t), u(t)), and ‘∗’ denotes transpose. With respect to the smoothness of the functions delimiting the problem, we assume that L, f and ϕ are C 2 and the q × (m + r)dimensional matrix  ∂ϕi δiα ϕα ∂uk  (i = 1, . . . , q; α = 1, . . . , r; k = 1, . . . , m) has rank q on U (here δαα = 1, δαβ = 0 (α 6= β)), where U := {u ∈ Rm | ϕα (u) ≤ 0 (α ∈ R), ϕβ (u) = 0 (β ∈ Q)}. This condition is equivalent to the condition that, at each point u in U , the matrix   ∂ϕi (i = i1 , . . . , ip ; k = 1, . . . , m) ∂uk has rank p, where i1 , . . . , ip are the indices i ∈ {1, . . . , q} such that ϕi (u) = 0 (see [15] for details). For all (t, x, u, p, µ, λ) in T × Rn × Rm × Rn × Rq × R let H(t, x, u, p, µ, λ) := hp, f (t, x, u)i − λL(t, x, u) − hµ, ϕ(u)i. First order necessary conditions are well established (see, for example, [15, 18, 22]) and one version can be stated as follows. Theorem 3.1 If (x0 , u0 ) solves P(S), ∃λ0 ≥ 0, p ∈ X, and µ ∈ Uq not vanishing simultaneously on T , such that a. µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ); b. ṗ(t) = −Hx∗ (x̃0 (t), p(t), µ(t), λ0 ) and Hu (x̃0 (t), p(t), µ(t), λ0 ) = 0 on every interval of continuity of u0 . From this result we introduce the notion of normality as in Section 2, that is, in such a way that the non-vanishing of the cost multiplier can be assured. This is accomplished by having zero as the unique solution to the adjoint equation whenever λ0 = 0. Definition 3.2 (x0 , u0 ) ∈ S is a normal process of S if, given (p, µ) ∈ X × Uq satisfying i. µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ); ii. ṗ(t) = −fx∗ (x̃0 (t))p(t) [ = −Hx∗ (x̃0 (t), p(t), µ(t), 0) ] (t ∈ T ); iii. 0 = fu∗ (x̃0 (t))p(t) − ϕ′∗ (u0 (t))µ(t) [ = Hu∗ (x̃0 (t), p(t), µ(t), 0) ] (t ∈ T ), then p ≡ 0. Note that, in this event, also µ ≡ 0. 448 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth We define the set of extremals as the set of all (x, u, p, µ) for which the conditions of Theorem 3.1 hold with λ0 = 1. Definition 3.3 Denote by E the set of all (x, u, p, µ) ∈ Z × X × Uq such that a. µα (t) ≥ 0 and µα (t)ϕα (u(t)) = 0 (α ∈ R, t ∈ T ); b. ṗ(t) = −fx∗ (x̃(t))p(t) + L∗x (x̃(t)) (t ∈ T ); c. fu∗ (x̃(t))p(t) = L∗u (x̃(t)) + ϕ′∗ (u(t))µ(t) (t ∈ T ). Note 3.4 From the above definitions it follows that, in Theorem 3.1, if (x0 , u0 ) is also a normal process of S, then ∃(p, µ) ∈ X × Uq such that (x0 , u0 , p, µ) ∈ E. Let us now introduce the second variation with respect to H, as well as the set of curvilinear tangent processes. For any (x, u, p, µ) ∈ Z × X × Uq let J((x, u, p, µ); (y, v)) := Z t1 t0 2Ω(t, y(t), v(t))dt ((y, v) ∈ Z) where, for all (t, y, v) ∈ T × Rn × Rm , 2Ω(t, y, v) := −[hy, Hxx (t)yi + 2hy, Hxu (t)vi + hv, Huu (t)vi] and H(t) denotes H(x̃(t), p(t), µ(t), 1). For any (x0 , u0 ) ∈ S, denote by CS (x0 , u0 ) the set of all processes (y, v) for which there exist δ > 0 and (x(·, ǫ), u(·, ǫ)) ∈ S (0 ≤ ǫ < δ) such that a. (x(t, 0), u(t, 0)) = (x0 (t), u0 (t)) (t ∈ T ); b. (xǫ (t, 0), uǫ (t, 0)) = (y(t), v(t)) (t ∈ T ); c. x(t, ǫ) is continuous and has continuous first and second derivatives with respect to ǫ; u(t, ǫ) and its first and second derivatives with respect to ǫ are piecewise continuous with respect to t. Based on these definitions, we can establish the following fundamental result. Theorem 3.5 Let (x0 , u0 ) ∈ S and suppose ∃(p, µ) ∈ X × Uq such that (x0 , u0 , p, µ) ∈ E. Let S1 [= S1 (µ)] := {(x, u) ∈ S | ϕα (u(t)) = 0 (α ∈ R, µα (t) > 0, t ∈ T )}. If (x0 , u0 ) solves P(S) then J((x0 , u0 , p, µ); (y, v)) ≥ 0 for all (y, v) ∈ CS1 (x0 , u0 ). 449 Critical cones for regular controls Proof: Define K(x, u) := hp(t1 ), ξ1 i − hp(t0 ), ξ0 i + Z t1 t0 F (t, x(t), u(t))dt ((x, u) ∈ Z) where, for all (t, x, u) ∈ T × Rn × Rm , F (t, x, u) := L(t, x, u) − hp(t), f (t, x, u)i + hµ(t), ϕ(u)i − hṗ(t), xi. Observe that F (t, x, u) = −H(t, x, u, p(t), µ(t), 1) − hṗ(t), xi and, if (x, u) ∈ S, then K(x, u) = I(x, u) + Z t1 t0 hµ(t), ϕ(u(t))idt. Note that S1 = {(x, u) ∈ S | µα (t)ϕα (u(t)) = 0 (α ∈ R, t ∈ T )} = {(x, u) ∈ S | K(x, u) = I(x, u)}. Since (x0 , u0 ) ∈ S1 , (x0 , u0 ) minimizes K on S1 . Now, let (y, v) belong to CS1 (x0 , u0 ) and let δ > 0 and (x(·, ǫ), u(·, ǫ)) ∈ S1 (0 ≤ ǫ < δ) be such that (x(t, 0), u(t, 0)) = (x0 (t), u0 (t)) and (xǫ (t, 0), uǫ (t, 0)) = (y(t), v(t)). Then g(ǫ) := K(x(·, ǫ), u(·, ǫ)) (0 ≤ ǫ < δ) satisfies g(ǫ) = I(x(·, ǫ), u(·, ǫ)) ≥ I(x0 , u0 ) = K(x0 , u0 ) = g(0) (0 ≤ ǫ < δ). Note that Fx (x̃0 (t)) = −Hx (x̃0 (t), p(t), µ(t), 1) − ṗ∗ (t) = 0, Fu (x̃0 (t)) = −Hu (x̃0 (t), p(t), µ(t), 1) = 0 and therefore g ′ (0) = 0. Consequently 0 ≤ g ′′ (0) = J((x0 , u0 , p, µ); (y, v)). As in the finite dimensional case, we are interested in characterizing the set CS (x0 , u0 ) in terms of a set of tangential constraints whose membership, in contrast with CS (x0 , u0 ), may be easily verifiable. Definition 3.6 For all (x0 , u0 ) ∈ S let L(x0 , u0 ) := {(y, v) ∈ Z | ẏ(t) = A(t)y(t) + B(t)v(t) (t ∈ T ), y(t0 ) = y(t1 ) = 0} 450 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth where A(t) = fx (x̃0 (t)), B(t) = fu (x̃0 (t)). Denote the set of active indices at u ∈ Rm by Ia (u) := {α ∈ R | ϕα (u) = 0} and define the set of processes satisfying the tangential constraints at (x0 , u0 ) with respect to S by RS (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | ϕ′α (u0 (t))v(t) ≤ 0 (α ∈ Ia (u0 (t)), t ∈ T ), ϕ′β (u0 (t))v(t) = 0 (β ∈ Q, t ∈ T )}. As we prove next, curvilinear tangent processes satisfy the tangential constraints. Proposition 3.7 For all (x0 , u0 ) ∈ S, CS (x0 , u0 ) ⊂ RS (x0 , u0 ). Proof: Let (y, v) ∈ CS (x0 , u0 ) and let δ > 0 and (x(·, ǫ), u(·, ǫ)) ∈ S (0 ≤ ǫ < δ) be such that (x(t, 0), u(t, 0)) = (x0 (t), u0 (t)) and (xǫ (t, 0), uǫ (t, 0)) = (y(t), v(t)). Then, for all 0 ≤ ǫ < δ, ẋ(t, ǫ) = f (t, x(t, ǫ), u(t, ǫ)) (t ∈ T ), x(t0 , ǫ) = ξ0 , x(t1 , ǫ) = ξ1 and so (y, v) solves L(x0 , u0 ). Also, for all (t, ǫ) ∈ T × [0, δ), ϕα (u(t, ǫ)) ≤ 0 (α ∈ R), ϕβ (u(t, ǫ)) = 0 (β ∈ Q). Fix i ∈ R ∪ Q and t ∈ T , and set γ(ǫ) := ϕi (u(t, ǫ)) so that γ ′ (0) = ϕ′i (u0 (t))v(t). If i ∈ Ia (u0 (t)) then γ ′ (0) ≤ 0. If i ∈ Q then γ ′ (0) = 0. Definition 3.8 We say (x0 , u0 ) is a c-regular point of S if CS (x0 , u0 ) = RS (x0 , u0 ). Now, let us exhibit three sets of constraints and their corresponding sets of tangential constraints. For any u ∈ Rm and µ ∈ Rq , let τ0 (u) := {h ∈ Rm | ϕ′i (u)h = 0 (i ∈ Ia (u) ∪ Q)}, τ1 (u, µ) := {h ∈ Rm | ϕ′α (u)h ≤ 0 (α ∈ Ia (u), µα = 0), ϕ′β (u)h = 0 (β ∈ R with µβ > 0, or β ∈ Q)}, τ2 (u) := {h ∈ Rm | ϕ′α (u)h ≤ 0 (α ∈ Ia (u)), ϕ′β (u)h = 0 (β ∈ Q)}. 451 Critical cones for regular controls Note that RS (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | v(t) ∈ τ2 (u0 (t)) (t ∈ T )}. Also, we have S1 (µ) = {(x, u) ∈ D | ϕα (u(t)) ≤ 0 (α ∈ R, µα (t) = 0, t ∈ T ), ϕβ (u(t)) = 0 (β ∈ R with µβ (t) > 0, or β ∈ Q, t ∈ T )} and therefore RS1 (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | v(t) ∈ τ1 (u0 (t), µ(t)) (t ∈ T )}. Similarly, if S0 := {(x, u) ∈ D | ϕi (u(t)) = 0 (i ∈ Ia (u0 (t)) ∪ Q, t ∈ T )}, then RS0 (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | v(t) ∈ τ0 (u0 (t)) (t ∈ T )}. Let us now apply the definition of normality, given in Definition 3.2, to the sets S0 and S1 . Note 3.9 (x0 , u0 ) is a normal process of S0 if, given (p, µ) ∈ X × Uq satisfying i. µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ); ii. ṗ(t) = −fx∗ (x̃0 (t))p(t) and fu∗ (x̃0 (t))p(t) = ϕ′∗ (u0 (t))µ(t) (t ∈ T ) then p ≡ 0. Note 3.10 (x0 , u0 ) is a normal process of S1 (µ) if, given (p, ν) ∈ X × Uq satisfying i. να (t) ≥ 0 and να (t)ϕα (u0 (t)) = 0 (α ∈ R, µα (t) = 0, t ∈ T ); ii. ṗ(t) = −fx∗ (x̃0 (t))p(t) and fu∗ (x̃0 (t))p(t) = ϕ′∗ (u0 (t))ν(t) (t ∈ T ) then p ≡ 0. Normality relative to S0 , S1 and S can be characterized in terms of the sets τ0 , τ1 and τ2 defined above. We refer to [28] for a detailed proof of these characterizations. Proposition 3.11 Let (x0 , u0 ) ∈ S and suppose µ ∈ Uq is such that µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ). Then (x0 , u0 ) is a normal process of S0 (respectively S1 , S) if and only if z ≡ 0 is the only solution of the system ż(t) = −A∗ (t)z(t), z ∗ (t)B(t)h ≤ 0 for all h ∈ τ0 (u0 (t)) (t ∈ T ) (respectively h ∈ τ1 (u0 (t), µ(t)), h ∈ τ2 (u0 (t))). 452 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth As one readily verifies (see also [28] for details), if (x0 , u0 ) is normal relative to S0 (or strongly normal) then it is normal relative to S1 (with µ as above) which in turn implies normality relative to S (or weak normality). It is important to mention, at this stage, that second order conditions in terms of RS0 (x0 , u0 ) assuming normality relative to S0 are well established in the literature (see, for example, [15]). The result is similar to Theorem 2.12 and it can be stated as follows. Theorem 3.12 Let (x0 , u0 ) ∈ S and suppose ∃(p, µ) ∈ X × Uq such that (x0 , u0 , p, µ) ∈ E. If (x0 , u0 ) solves P(S) and is a normal process of S0 , then J((x0 , u0 , p, µ); (y, v)) ≥ 0 for all (y, v) ∈ RS0 (x0 , u0 ). As we show next, we are able to provide two examples which illustrate a crucial aspect of the theory, namely, that if we modify the assumptions or the space where the nonnegativity of the quadratic form holds, the conclusion of Theorem 3.12 may no longer hold. Let us begin with slightly changing the assumptions. In this first example, we exhibit a solution (x0 , u0 ) to a problem P(S) which is normal relative to S, together with a pair (p, µ) such that (x0 , u0 , p, µ) ∈ E, but J((x0 , u0 , p, µ); (y, v)) < 0 for some (y, v) ∈ RS0 (x0 , u0 ). Example 3.13 Let a > 0, k > 1 and consider the problem of minimizing I(x, u) = Z 1 0 {u2 (t) − u1 (t)}dt subject to (x, u) ∈ Z, ẋ(t) = 2u2 (t) − (k + 1)u1 (t) + au23 (t) (t ∈ [0, 1]), x(0) = x(1) = 0, and u2 (t) ≥ u1 (t), ku1 (t) ≥ u2 (t) (t ∈ [0, 1]). In this case T = [0, 1], n = 1, m = 3, r = q = 2, ξ0 = ξ1 = 0 and, for all t ∈ T , x ∈ R, and u = (u1 , u2 , u3 ), L(t, x, u) = u2 − u1 , f (t, x, u) = 2u2 − (k + 1)u1 + au23 , ϕ1 (u) = u1 − u2 , ϕ2 (u) = u2 − ku1 . We have H(t, x, u, p, µ) = p(2u2 − (k + 1)u1 + au23 ) + u1 − u2 − µ1 (u1 − u2 ) − µ2 (u2 − ku1 ) and so Hu (t, x, u, p, µ) = (−p(k + 1) − µ1 + kµ2 + 1, 2p + µ1 − µ2 − 1, 2apu3 ), 453 Critical cones for regular controls   0 0 0 Huu (t, x, u, p, µ) =  0  0 0  0 0 2ap so that, for all (x, u, p, µ) ∈ Z × X × U2 and (y, v) ∈ Z, J((x, u, p, µ); (y, v)) = − Z 1 0 2ap(t)v32 (t)dt. Clearly (x0 , u0 ) ≡ (0, 0) is a solution to the problem. Note that (x0 , u0 , p, µ) is an extremal if µα (t) ≥ 0 (α = 1, 2), ṗ(t) = 0 and fu∗ (x̃0 (t))p(t) = L∗u (x̃0 (t)) + ϕ′∗ (u0 (t))µ(t) (t ∈ T ). This last relation corresponds to       −(k + 1) −1 µ1 (t) − kµ2 (t)       2   p(t) =  1  +  −µ1 (t) + µ2 (t)  0 0 0 and thus, if µ = (µ1 , µ2 ) ≡ (0, 1) and p ≡ 1, (x0 , u0 , p, µ) ∈ E. Now, since ϕ′1 (u) = (1, −1, 0) and ϕ′2 (u) = (−k, 1, 0), we have τ0 (u0 (t)) = {h ∈ R3 | h1 − h2 = 0, h2 − kh1 = 0}, τ1 (u0 (t), µ(t)) = {h ∈ R3 | h1 − h2 ≤ 0, h2 − kh1 = 0}, τ2 (u0 (t)) = {h ∈ R3 | h1 − h2 ≤ 0, h2 − kh1 ≤ 0}. Since fx (x̃0 (t)) = 0 and fu (x̃0 (t)) = (−(k + 1), 2, 0), the system ż(t) = −A∗ (t)z(t) = 0, z ∗ (t)B(t)h = z(t)(−(k + 1)h1 + 2h2 ) = 0 for all (h1 , h2 , h3 ) ∈ τ0 (u0 (t)) (t ∈ T ) has nontrivial solutions and so (x0 , u0 ) is not normal relative to S0 . On the other hand, z ≡ 0 is the only solution to the system ż(t) = 0, z(t)(−(k + 1)h1 + 2h2 ) ≤ 0 for all (h1 , h2 , h3 ) ∈ τ2 (u0 (t)) (t ∈ T ) since both (1, 1, 0) and (1, k, 0) belong to τ2 (u0 (t)) implying that z(t) ≥ 0 and z(t) ≤ 0 (t ∈ T ) respectively, so that (x0 , u0 ) is normal relative to S. Finally, the system ż(t) = 0, z(t)(−(k + 1)h1 + 2h2 ) ≤ 0 for all (h1 , h2 , h3 ) ∈ τ1 (u0 (t), µ(t)) (t ∈ T ) corresponds to ż(t) = 0, z(t)(h2 − h1 ) ≤ 0 with h2 − h1 ≥ 0, which has nontrivial solutions and so (x0 , u0 , µ) is not a normal point of S1 . Let v = (v1 , v2 , v3 ) ≡ (0, 0, 1) and y ≡ 0. Then (y, v) solves ẏ(t) = −(k + 1)v1 (t) + 2v2 (t) (t ∈ T ), y(0) = y(1) = 0, v(t) ∈ τ0 (u0 (t)), and J((x0 , u0 , p, µ); (y, v)) = −2a < 0. 454 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth Our second example deals with a change, in Theorem 3.12, of the critical cone where the second order conditions hold. It shows that a solution to the problem, which is a normal process of S0 , may yield a negative second variation on RS (x0 , u0 ). Example 3.14 Consider the problem of minimizing I(x, u) = Z 1 0 − exp(−u2 (t))dt subject to (x, u) ∈ Z, ẋ(t) = u1 (t) + u22 (t) (t ∈ [0, 1]), x(0) = x(1) = 0, u2 (t) ≥ 0 (t ∈ [0, 1]). In this case T = [0, 1], n = 1, m = 2, r = q = 1, ξ0 = ξ1 = 0 and, for all t ∈ T , x ∈ R, and u = (u1 , u2 ), L(t, x, u) = −e−u2 , f (t, x, u) = u1 + u22 , ϕ1 (u) = −u2 . We have H(t, x, u, p, µ) = p(u1 + u22 ) + µ1 u2 + e−u2 and so Hu (t, x, u, p, µ) = (p, 2pu2 + µ1 − e −u2 ), Huu (t, x, u, p, µ) =  0 0 0 2p + e−u2  so that, for all (x, u, p, µ) ∈ Z × X × U1 and (y, v) ∈ Z, J((x, u, p, µ); (y, v)) = − Z 1 0 (2p(t) + e−u2 (t) )v22 (t)dt. Clearly (x0 , u0 ) ≡ (0, 0) solves the problem. Also, since fu (t, x, u) = (1, 2u2 ), Lu (t, x, u) = (0, e−u2 ) and ϕ′ (u) = (0, −1), (x0 , u0 , p, µ) is an extremal if ṗ(t) = 0 and       1 0 0 p(t) = + . 0 1 −µ(t) Thus, if (p, µ) ≡ (0, 1), (x0 , u0 , p, µ) ∈ E. Observe now that τ0 (u0 (t)) = τ1 (u0 (t), µ(t)) = {h ∈ R2 | h2 = 0}, τ2 (u0 (t)) = {h ∈ R2 | −h2 ≤ 0}. Since z ≡ 0 is the only solution to the system ż(t) = 0, z(t)h1 = 0 for all (h1 , h2 ) ∈ τ0 (u0 (t)) (t ∈ T ), (x0 , u0 ) is normal relative to S0 . Now, if v ≡ (0, 1) and y ≡ 0 then v(t) ∈ τ2 (u0 (t)) and ẏ(t) = v1 (t) (t ∈ T ), y(0) = y(1) = 0, but J((x0 , u0 , p, µ); (y, v)) = −1 < 0. 455 Critical cones for regular controls In spite of these examples, Theorem 3.12 clearly looks too weak in comparison with the main result that holds for the finite dimensional case, namely, Theorem 2.9. In the latter, one assumes normality relative to S1 and one can assure that the second variation is nonnegative on the set of tangential constraints with respect to S1 . In the former, normality is assumed relative to S0 (in general a stronger assumption than normality on S1 ) and the condition on the second variation holds on the tangential constraints with respect to S0 (in general a set smaller than that with respect to S1 ). We pose the question if a result, analogous to Theorem 2.9, holds for our optimal control problem, that is, if the following result is valid or not. Conjecture 3.15 Let (x0 , u0 ) ∈ S and suppose ∃(p, µ) ∈ X × Uq such that (x0 , u0 , p, µ) ∈ E. Let S1 := {(x, u) ∈ S | ϕα (u(t)) = 0 (α ∈ R, µα (t) > 0, t ∈ T )}. If (x0 , u0 ) solves P(S) and is normal relative to S1 , then J((x0 , u0 , p, µ); (y, v)) ≥ 0 for all (y, v) ∈ RS1 (x0 , u0 ). Clearly, the role played by RS1 (x0 , u0 ) above is crucial. We shall end this section by proving a new result, mentioned in the introduction, which entirely characterizes this set of critical directions. In particular, we shall conclude that, though the set S1 (µ) of constraints depends on the multiplier µ, this dependence is no longer present in the set RS1 (x0 , u0 ) of tangential constraints relative to S1 (µ). In our characterization given below, we shall make use of the first variation of I given by I ′ (x0 , u0 ; y, v) = Z t1 t0 {Lx (x̃0 (t))y(t) + Lu (x̃0 (t))v(t)}dt. For all (x0 , u0 ) ∈ S, denote by M (x0 , u0 ) the set of all (p, µ) ∈ X × Uq such that (x0 , u0 , p, µ) ∈ E, that is, a. µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ); b. ṗ(t) = −A∗ (t)p(t) + L∗x (x̃0 (t)) (t ∈ T ); c. B ∗ (t)p(t) = L∗u (x̃(t)) + ϕ′∗ (u(t))µ(t) (t ∈ T ), where A(t) = fx (x̃0 (t)), B(t) = fu (x̃0 (t)). Recall that RS (x0 , u0 ) := {(y, v) ∈ L(x0 , u0 ) | ϕ′α (u0 (t))v(t) ≤ 0 (α ∈ Ia (u0 (t)), t ∈ T ), ϕ′β (u0 (t))v(t) = 0 (β ∈ Q, t ∈ T )} where L(x0 , u0 ) := {(y, v) ∈ Z | ẏ(t) = A(t)y(t) + B(t)v(t) (t ∈ T ), y(t0 ) = y(t1 ) = 0} 456 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth and therefore, given (p, µ) ∈ M (x0 , u0 ), since S1 [= S1 (µ)] := {(x, u) ∈ S | ϕα (u(t)) = 0 (α ∈ R, µα (t) > 0, t ∈ T )}, it follows that RS1 (x0 , u0 ) = {(y, v) ∈ RS (x0 , u0 ) | ϕ′α (u0 (t))v(t) = 0 (α ∈ R, µα (t) > 0, t ∈ T )}. Theorem 3.16 Suppose (x0 , u0 ) ∈ S and (p, µ) ∈ M (x0 , u0 ). Then RS1 (µ) (x0 , u0 ) = {(y, v) ∈ RS (x0 , u0 ) | I ′ (x0 , u0 ; y, v) = 0}. In particular, RS1 (µ) (x0 , u0 ) = RS1 (ν) (x0 , u0 ) for any (q, ν) ∈ M (x0 , u0 ). Proof: We proceed first as in the proof of Theorem 3.5. Define K(x, u) := hp(t1 ), ξ1 i − hp(t0 ), ξ0 i + Z t1 t0 F (t, x(t), u(t))dt ((x, u) ∈ Z) where, for all (t, x, u) ∈ T × Rn × Rm , F (t, x, u) := L(t, x, u) − hp(t), f (t, x, u)i + hµ(t), ϕ(u)i − hṗ(t), xi. Since F (t, x, u) = −H(t, x, u, p(t), µ(t), 1) − hṗ(t), xi, we have Fx (x̃0 (t)) = −Hx (x̃0 (t), p(t), µ(t), 1) − ṗ∗ (t) = 0, Fu (x̃0 (t)) = −Hu (x̃0 (t), p(t), µ(t), 1) = 0 and, consequently, K ′ (x0 , u0 ; y, v) = 0 for all (y, v) ∈ Z. From this equality, we obtain that 0 = Z t1 t0 ′ {[Lx (x̃0 (t)) − p∗ (t)A(t) − ṗ∗ (t)]y(t) + [Lu (x̃0 (t)) − p∗ (t)B(t) + µ∗ (t)ϕ′ (u0 (t))]v(t)}dt = I (x0 , u0 ; y, v) − Z t1 t0 {p∗ (t)(A(t)y(t) + B(t)v(t)) + ṗ∗ (t)y(t)}dt + Z t1 t0 µ∗ (t)ϕ′ (u0 (t))v(t)dt. If (y, v) ∈ L(x0 , u0 ), this implies that I ′ (x0 , u0 ; y, v) + Z t1 t0 µ∗ (t)ϕ′ (u0 (t))v(t)dt = 0. 457 Critical cones for regular controls If also (y, v) ∈ RS (x0 , u0 ), we have I ′ (x0 , u0 ; y, v) + Z t1 X t0 α∈R µα (t)ϕ′α (u0 (t))v(t)dt = 0. Since µα (t)ϕ′ (u0 (t))v(t) ≤ 0 (α ∈ R, t ∈ T ), we conclude that (y, v) belongs to the set {(y, v) ∈ RS (x0 , u0 ) | I ′ (x0 , u0 ; y, v) = 0} if and only if ϕ′ (u0 (t))v(t) = 0 (α ∈ R, µα (t) > 0, t ∈ T ), that is, if and only if (y, v) ∈ RS1 (µ) (x0 , u0 ). 4 Lagrangian independent of the state The particular case where the Lagrangian does not depend on the state variable has been studied in a recent paper [27]. It provides an excellent illustration of how the theory of first and second order necessary conditions for the finite dimensional case can be applied to optimal control problems. Moreover, it implies a second order condition stronger than the one given in Conjecture 3.15 since the assumption of normality of a solution (x0 , u0 ) relative to S1 implies the nonnegativity of the second variation in a set which generally strictly contains RS1 (x0 , u0 ). In this section we derive some consequences of this result (see also [13, 14], where the constraints may depend not only on the control functions but also on the time variable). Let us then consider the case where L(t, x, u) = L(t, u), i.e., L does not depend on x. Our problem P(S) is that of minimizing the functional I(u) = R t1 t0 L(t, u(t))dt on S, that is, subject to (x, u) ∈ Z and   ẋ(t)   = f (t, x(t), u(t)) (t ∈ T ); x(t0 ) = ξ0 , x(t1 ) = ξ1 ;    ϕα (u(t)) ≤ 0, ϕβ (u(t)) = 0 (α ∈ R, β ∈ Q, t ∈ T ). Our assumptions include that L, f and ϕ are C 2 and the q×(m+r)-dimensional matrix  ∂ϕi δiα ϕα ∂uk  (i = 1, . . . , q; α = 1, . . . , r; k = 1, . . . , m) has rank q on the set of points u ∈ Rm satisfying ϕα (u) ≤ 0 (α ∈ R), ϕβ (u) = 0 (β ∈ Q). Let C := {u ∈ Um | (t, u(t)) ∈ A (t ∈ T )}, where A = {(t, u) ∈ T × Rm | ϕα (u) ≤ 0 (α ∈ R), ϕβ (u) = 0 (β ∈ Q)}. 458 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth Denote by (C) the problem of minimizing I(u) = tt01 L(t, u(t))dt over C. Note that, if (x0 , u0 ) is admissible for P(S) and u0 solves (C), that is, I(u0 ) ≤ I(u) for all u ∈ C, then (x0 , u0 ) solves P(S), that is, I(u0 ) ≤ I(u) for all u ∈ C such that, if ẋ(t) = f (t, x(t), u(t)) (t ∈ T ) and x(t0 ) = ξ0 , then x(t1 ) = ξ1 . Clearly, the converse may not hold. Two simple examples illustrate this fact. R Example 4.1 Consider the problem of minimizing I(u) = to R1 0 u(t)dt subject ẋ(t) = u(t), x(0) = x(1) = 0, u(t) ≤ 0. Then (x0 , u0 ) ≡ (0, 0) solves P(S), being the only admissible process, but u0 ≡ 0 does not solve (C), that is, it does not minimize I over the set C = {u ∈ U1 | u(t) ≤ 0 (t ∈ T )}. Actually, in this example, u0 maximizes I on C. Note also that (x0 , u0 ) is not normal with respect to S since τ2 (u0 ) = {h ∈ R | h ≤ 0} and so the system ż(t) = 0, z(t)h ≤ 0 for all h ∈ τ2 (u0 (t)) (t ∈ T ) has nontrivial solutions. Example 4.2 Consider the problem of minimizing I(u) = subject to ẋ(t) = u2 (t), x(0) = x(2) = 0, u2 (t) ≤ 1. R2 0 (t − 1)u(t)dt Then (x0 , u0 ) ≡ (0, 0) solves P(S) but u0 ≡ 0 does not solve (C) since it does not minimize I over the set C = {u ∈ U1 | −1 ≤ u(t) ≤ 1 (t ∈ T )}. In contrast with the previous example, here u0 does not minimize or maximize I on C. As one readily verifies, the solution to (C) is given by the function which equals 1 on [0, 1] and −1 on [1, 2] (for the maximum the signs are reversed). As before, clearly (x0 , u0 ) is not normal with respect to S since fu (t, x0 (t), u0 (t)) = 0. Let us now state the result on second order conditions for problem P(S) proved in [27]. Critical cones for regular controls 459 Theorem 4.3 Suppose (x0 , u0 ) solves P(S) and ∃(p, µ) ∈ X × Uq satisfying a. µα (t) ≥ 0 and µα (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ); b. ṗ(t) = −A∗ (t)p(t) (t ∈ T ); c. p∗ (t)B(t) = Lu (t, u0 (t)) + µ∗ (t)ϕ′ (u0 (t)) (t ∈ T ). Suppose also that u0 solves (C), and consider the following statements: i. (x0 , u0 ) is normal relative to S1 (µ). ii. p ≡ 0. iii. J((x0 , u0 , p, µ); (y, v)) ≥ 0 for all (y, v) ∈ Z with v(t) ∈ τ1 (u0 (t), µ(t)) (t ∈ T ). Then (i) ⇒ (ii) ⇒ (iii). It should be noted that the assumptions in this theorem imply that the second variation is nonnegative on the set of pairs (y, v) in Z satisfying v(t) ∈ τ1 (u0 (t), µ(t)) (t ∈ T ). This holds in contrast with the conclusion of Conjecture 3.15 where the nonnegativity of that function requires from (y, v) to also satisfy the linear equation ẏ(t) = A(t)y(t) + B(t)v(t) (t ∈ T ) together with the endpoint conditions y(t0 ) = y(t1 ) = 0. Observe also that, as Example 4.1 illustrates, we may have a solution (x0 , u0 ) to our problem P(S) for which u0 does not solve problem (C) and, moreover, u0 maximizes I on C. Also, in that example, u0 ≡ 0 maximizes (t, u) 7→ u on A and, in Example 4.2, u0 ≡ 0 minimizes (t, u) 7→ u2 on A. Apart from this, recall that, in Theorem 4.3, we are assuming normality relative to S1 which in turn implies normality relative to S. These facts motivate the following result, assuming that f (t, x, u) = f (t, u), i.e., f is independent on x. Proposition 4.4 Let (x0 , u0 ) be an admissible process and suppose that, for some i ∈ {1, . . . , n}, f i (t, u0 (t)) ≤ f i (t, u) (or ≥ f i (t, u)) whenever (t, u) ∈ A. Then (x0 , u0 ) is not normal relative to S. Proof: Suppose, without loss of generality, that f i (t, u0 (t)) ≤ f i (t, u) whenever (t, u) ∈ A. Then there exists a unique ν ∈ Uq such that fui (t, u0 (t)) + ν ∗ (t)ϕ′ (u0 (t)) = 0 (t ∈ T ). Moreover, να (t) ≥ 0 and να (t)ϕα (u0 (t)) = 0 (α ∈ R, t ∈ T ). By definition, τ2 (u) := {h ∈ Rm | ϕ′i (u)h ≤ 0 (i ∈ Ia (u)), ϕ′j (u)h = 0 (j ∈ Q)} and (x0 , u0 ) is a normal point of S if z ≡ 0 is the only solution to the system ż(t) = 0, z(t)fu (t, u0 (t))h ≤ 0 for all h ∈ τ2 (u0 (t)) (t ∈ T ). 460 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth Let zi (t) = −1 and zj (t) = 0 (j 6= i). Then z(t)fu (t, u0 (t))h = −fui (t, u0 (t))h = ν ∗ (t)ϕ′ (u0 (t))h ≤ 0 for all h ∈ τ2 (u0 (t)) and therefore the above system has nontrivial solutions. This proves the claim. The following example illustrates an application of Theorem 4.3, namely, the verification of normality by simply exhibiting an extremal (x0 , u0 , p, µ) for which p 6≡ 0. Example 4.5 Consider the problem of minimizing I(x, u) = ject to (x, u) ∈ Z and R1 0 u2 (t)dt sub- = u21 (t) + u2 (t) − u3 (t) (t ∈ [0, 1]); x(0) = x(1) = 0;    u2 (t) ≥ 0, u3 (t) ≥ 0 (t ∈ [0, 1]).   ẋ(t)   In this case T = [0, 1], n = 1, m = 3, r = q = 2, ξ0 = ξ1 = 0 and, for all t ∈ T , x ∈ R, and u = (u1 , u2 , u3 ), L(t, x, u) = u2 , f (t, x, u) = u21 + u2 − u3 , ϕ1 (u) = −u2 , ϕ2 (u) = −u3 . We have H(t, x, u, p, µ) = p(u21 + u2 − u3 ) − u2 + µ1 u2 + µ2 u3 and so Hu (t, x, u, p, µ) = (2pu1 , p − 1 + µ1 , −p + µ2 ) Clearly (x0 , u0 ) ≡ (0, 0) is a solution to both problems P(S) and (C). Let µ = (µ1 , µ2 ) ≡ (0, 1) and p ≡ 1. Then, as one readily verifies, (x0 , u0 , p, µ) ∈ E, that is, the first conditions of Theorem 4.3 hold. Since p 6≡ 0, we conclude that the solution (x0 , u0 ) is not normal relative to S1 . Actually, for this example, the pair (x0 , u0 ) is normal relative to S and, if v = (v1 , v2 , v3 ) ≡ (1, 0, 0) and y ≡ 0, then (y, v) solves ẏ(t) = v2 (t) − v3 (t) (t ∈ T ), y(0) = y(1) = 0, v(t) ∈ τ0 (u0 (t)), and J((x0 , u0 , p, µ); (y, v)) = −2 < 0. Theorem 4.3 gives rise to several surmises and questions which we will be solved completely by means of some illustrative examples. Some of these issues were first posed, and left open, in [27]. A natural question is if, in Theorem 4.3, the assumption that u0 solves (C) becomes redundant once normality relative to S1 is imposed. In other words, if (x0 , u0 ) solves P(S), (x0 , u0 , p, µ) is an extremal, and (i) holds, then necessarily u0 solves (C). Our first example does answer this question. We provide a solution (x0 , u0 ) to P(S) together with an extremal which not only satisfies (i) but (x0 , u0 ) is strongly normal, and u0 is not a solution to (C). 461 Critical cones for regular controls Example 4.6 (x0 , u0 ) solves P(S) and is strongly normal, but u0 does not solve (C). R Consider the problem P(S) of minimizing I(x, u) = 01 u(t)dt subject to ẋ(t) = u(t)x(t) (t ∈ [0, 1]), x(0) = 1, x(1) = e, u(t) ≥ 0 (t ∈ [0, 1]). For this problem T = [0, 1], ξ0 = 1, ξ1 = e, L(t, u) = u, f (t, x, u) = ux, ϕ(u) = −u. Also S = {(x, u) ∈ D | u(t) ≥ 0 (t ∈ T )} where D = {(x, u) ∈ Z | ẋ(t) = u(t)x(t), x(0) = 1, x(1) = e}. To begin with note Rthat, if (x, u) is admissible, then x(t) = exp( 0t u(s)ds) R1 1 and x(1) = e = exp( 0 u(t)dt) and so 0 u(t)dt = I(x, u) = 1. Therefore (x0 (t), u0 (t)) := (et , 1) (t ∈ T ) is a solution to the problem P(S) and, clearly, u0 ≡ 1 is not a solution to (C). Now, fx (x̃0 (t)) = 1, fu (x̃0 (t)) = et , and ϕ(u0 (t)) = −1 < 0. Since, for this solution, there are no active constraints, S0 = D. Clearly (x0 , u0 ) is a normal process of S0 since, given (p, µ) satisfying R µ(t)(−1) = 0, ṗ(t) = −p(t), et p(t) = −µ(t) = 0 (t ∈ [0, 1]) then p ≡ 0. This example illustrates some more features of the theorem. In particular, we can exhibit a pair (p, µ) such that (x0 , u0 , p, µ) is an extremal for which (i) is satisfied, but (ii) and (iii) do not hold. Example 4.7 (x0 , u0 ) solves P(S), (p, µ) ∈ M (x0 , u0 ) is such that, in Theorem 4.3, (i) is satisfied but (ii) and (iii) do not hold. Moreover, J((x0 , u0 , p, µ); (y, v)) = 0 for all (y, v) ∈ L(x0 , u0 ). Consider the previous example with (x0 , u0 ) defined as above. If p(t) = e−t and µ ≡ 0 then (x0 , u0 , p, µ) ∈ E since ṗ(t) = −p(t) and et p(t) = 1 (t ∈ T ). Also, as seen before, (i) holds but not (ii). Now, we have H = pux − u + µu and so Hu = px − 1 + µ, Hx = pu, Huu = Hxx = 0, Hux = Hxu = p all evaluated at (t, x, u, p, µ). Thus 2Ω(t, y, v) = −2p(t)yv and so J((x0 , u0 , p, µ); (y, v)) = −2 Z 1 0 e−t y(t)v(t)dt. 462 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth Since there are no active constraints, τ0 (u0 (t)) = τ1 (u0 (t), µ(t)) = τ2 (u0 (t)) = R. Thus, if y ≡ v ≡ 1 then clearly (iii) fails to hold. Let us now show that J((x0 , u0 , p, µ); (y, v)) = 0 for all (y, v) ∈ L(x0 , u0 ). Let (y, v) ∈ L(x0 , u0 ) and note that, by definition, ẏ(t) = y(t) + et v(t) (t ∈ T ) and y(0) = y(1) = 0. Let g(t) = by parts Rt 0 v(s)ds. Then y(t) = et g(t) and so g(0) = g(1) = 0. Integrating Z 1 0 and therefore g(t)v(t)dt = g 2 (t)|10 − J((x0 , u0 , p, µ), (y, v)) = −2 Z 1 Z 1 0 0 g(t)v(t)dt g(t)v(t)dt = 0. Now, this particular solution gives rise to a different conjecture. It seems to be crucial, for this example, the fact that there are no active constraints. However, we can provide below a different solution for which the previous conclusions also hold but the constraint is active on a subinterval of T . Example 4.8 The conclusions of Examples 4.6 and 4.7 hold in the presence of active constraints. Consider again problem P(S) of Example 4.7. Let a = 1/2 and define the process (x1 , u1 ) by x1 (t) = ( 2t e e if t ∈ [0, a] if t ∈ [a, 1] u1 (t) = ( 2 if t ∈ [0, a] 0 if t ∈ (a, 1]. Clearly (x1 , u1 ) is a solution to P(S) but not to (C). To check for normality, suppose (p, µ) ∈ X × U1 satisfies i. µ(t)ϕ(u1 (t)) = 0 (t ∈ T ); ii. ṗ(t) = −fx (x̃1 (t))p(t) and fu (x̃1 (t))p(t) = ϕ′ (u1 (t))µ(t) (t ∈ T ). By (i), µ(t) = 0 on [0, a]. By (ii), since fx (x̃1 (t)) = u1 (t), fu (x̃1 (t)) = x1 (t), and ϕ′ (u1 (t)) = −1, we have ṗ(t) = −u1 (t)p(t), x1 (t)p(t) = −µ(t) (t ∈ T ) which implies, in particular, that e2t p(t) = 0 on [0, a] and ṗ(t) = 0 on (a, 1]. Thus p(t) = 0 on [0, a] and p(t) is constant on (a, 1]. By continuity, p ≡ 0 and so (x1 , u1 ) is strongly normal. Now, for this solution, an extremal (x1 , u1 , p, µ) must satisfy a. µ(t) ≥ 0 and µ(t)ϕ(u1 (t)) = 0 (t ∈ T ); b. ṗ(t) = −u1 (t)p(t) and p(t)x1 (t) = 1 − µ(t) (t ∈ T ). 463 Critical cones for regular controls These conditions are clearly satisfied by (p, µ) with µ ≡ 0 and p(t) = ( −2t e e −1 if t ∈ [0, a] if t ∈ (a, 1]. Observe now that, since ϕ(u1 (t)) = −u1 (t) (t ∈ T ), the constraint is inactive on [0, a] but active on (a, 1]. Thus we have τ0 (u1 (t)) = ( R if t ∈ [0, a] {h | −h = 0} if t ∈ (a, 1] τ2 (u1 (t)) = ( R if t ∈ [0, a] {h | −h ≤ 0} if t ∈ (a, 1] and τ1 (u1 (t)), µ(t)) = τ2 (u1 (t)) (t ∈ T ). As before we have 2Ω(t, y, v) = −2p(t)yv and so, if y ≡ v ≡ 1, then J((x1 , u1 , p, µ); (y, v)) < 0. Now, (y, v) ∈ Z belongs to L(x1 , u1 ) if and only if y(0) = y(1) = 0 and ẏ(t) = Rt and so, if g(t) = 0 ( 2y(t) + e2t v(t) if t ∈ [0, a] ev(t) if t ∈ [a, 1] v(s)ds, then y(t) = ( 2t e g(t) eg(t) if t ∈ [0, a] if t ∈ (a, 1] with g(0) = g(1) = 0. We reach the same conclusion as before since J((x1 , u1 , p, µ), (y, v)) = −2 Z 1 0 g(t)v(t)dt = 0 for any (y, v) ∈ L(x1 , u1 ). This last solution can be generalized by considering subintervals of length an = 1/n and setting xn (t) = ( nt e e if t ∈ [0, an ] if t ∈ [an , 1] un (t) = ( n if t ∈ [0, an ] 0 if t ∈ (an , 1]. Then (xn , un ) is a solution to the problem P(S) and essentially the same conclusions as before will follow. We have thus shown, in particular, that we may have a strongly normal solution (x0 , u0 ) to P(S) and u0 does not solve (C). In Example 4.6, however, note that f depends on x and, again, this seems to be fundamental in the analysis. In the following example we reach the same conclusion, but f does not depend on x. 464 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth Example 4.9 The previous conclusions hold with the function f independent of x. Consider the problem P(S) of minimizing I(x, u) = Z 1 0 {u1 (t) + u2 (t) + u3 (t)}dt subject to (x, u) ∈ Z and    ẋ(t)  = u3 (t) (t ∈ [0, 1]); x(0) = 0, x(1) = 1;    u1 (t) − 2u2 (t) ≤ 0, u2 (t) − u1 (t) ≤ 0 (t ∈ [0, 1]). For this problem T = [0, 1], ξ0 = 0, ξ1 = 1, L(t, u) = u1 + u2 + u3 , f (t, x, u) = u3 , ϕ1 (u) = u1 − 2u2 , ϕ2 (u) = u2 − u1 . Clearly, (C) has no solution. Let x0 (t) := t and u0 (t) := (0, 0, 1) (t ∈ T ). Then (x0 , u0 ) is a solution to problem P(S) since, for any (x, u) admissible, u1 (t) + u2 (t) ≥ 0 (t ∈ T ) and therefore I(x, u) ≥ Z 1 0 u3 (t)dt = 1 = I(x0 , u0 ). It is also normal relative to S0 since τ0 (u0 (t)) = {h ∈ R3 | h1 − 2h2 = 0, −h1 + h2 = 0} = {h | h1 = h2 = 0} and so z ≡ 0 is the only solution to the system ż(t) = 0 and z(t)h3 = 0 for all h3 ∈ R. Let us now show that one can find an extremal for which (i) of Theorem 4.3 holds but not (ii). This clearly follows if p ≡ 1 and (µ1 , µ2 ) ≡ (2, 3), for then (x0 , u0 , p, µ) ∈ E since ṗ(t) = 0 and 1 −2 0 −1 1 0 = (0, 0, 1) = p(t)fu (x̃0 (t)). Lu (x̃0 (t)) + µ∗ (t)ϕ′ (u0 (t)) = (1, 1, 1) + (2, 3)   We end this paper with another important property derived from Theorem 4.3. Our final example provides a solution (x0 , u0 ) to a problem P(S) which is normal relative to S and u0 solves (C), but (x0 , u0 ) is not normal relative to S1 (nor in consequence to S0 ) for any pair (p, µ) with (x0 , u0 , p, µ) an extremal. Example 4.10 (x0 , u0 ) solves P(S), it is normal relative to S, and u0 solves (C), but (x0 , u0 ) is not normal relative to S1 (µ) for any (p, µ) ∈ M (x0 , u0 ). 465 Critical cones for regular controls Consider the problem P(S) of minimizing I(x, u) = to R2 0 (t − 1)u(t)dt subject ẋ(t) = u(t) (t ∈ [0, 2]), x(0) = x(2) = 0, u2 (t) ≤ 1 (t ∈ [0, 2]). Define the process (x0 , u0 ) as x0 (t) =  t if t ∈ [0, 1] 2 − t if t ∈ [1, 2] u0 (t) = ( 1 if t ∈ [0, 1] −1 if t ∈ (1, 2]. Clearly (x0 , u0 ) is a solution to problem P(S) and u0 solves (C). It is not normal relative to S0 since, for example, p ≡ 2 and µ ≡ u0 show that the system ṗ(t) = 0 and p(t) = 2u0 (t)µ(t) (t ∈ T ) has nonnull solutions. The conclusion also follows straightforwardly by noting that τ0 (u0 (t)) = {0} and so the system ż(t) = 0 and z(t)h ≤ 0 for all h ∈ τ0 (u0 (t)) has nonnull solutions. Now, note that τ2 (u0 (t)) = ( {h | h ≤ 0} if t ∈ [0, 1] {h | h ≥ 0} if t ∈ (1, 2] and so (x0 , u0 ) is normal relative to S since z ≡ 0 is the only solution of the system ż(t) = 0 and z(t)h ≤ 0 for all h ∈ τ2 (u0 (t)). Finally, if (x0 , u0 , p, µ) is an extremal, then µ(t) ≥ 0, ṗ(t) = 0 and p(t) = t − 1 + 2µ(t)u0 (t) (t ∈ T ). This implies, in particular, that p is constant, p ≥ t − 1 on [0, 1] and p ≤ t − 1 on (1, 2]. Hence p ≡ 0 and µ(t) = ( (1 − t)/2 if t ∈ [0, 1] (t − 1)/2 if t ∈ (1, 2]. τ1 (u0 (t), µ(t)) = ( {0} if t ∈ [0, 1) ∪ (1, 2] {h | h ≤ 0} if t = 1 In this event we have and therefore (x0 , u0 ) is not normal relative to S1 (µ). References [1] A.V. Arutyunov, Smooth abnormal problems in extremum theory and analysis, Russian Mathematical Surveys, 67 (2012), 403-457. https://doi.org/10.1070/rm2012v067n03abeh004793 466 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth [2] A.V. Arutyunov, D.Y. Karamzin, Necessary conditions for a weak minimum in an optimal control problem with mixed constraints, Differential Equations, 41 (2005), 1532-1543. https://doi.org/10.1007/s10625-005-0311-2 [3] A.V. Arutyunov, F.L. Pereira, Second-order necessary optimality conditions for problems without a priori normality assumptions, Mathematics of Operations Research, 31 (2006), 1-12. https://doi.org/10.1287/moor.1050.0172 [4] A.V. Arutyunov, Y.S. Vereshchagina, On necessary second-order conditions in optimal control problems, Differentsial’nye Uravneniya 38 (2002), 1443-1450. (Russian). Translation in Differential Equations 38 (2002), 1531-1540. https://doi.org/10.1023/a:1023624602611 [5] J.P. Aubin, H. Frankowska, Set-Valued Analysis, Birkhäuser, Boston, 1990. [6] J.A. Becerril, J.F. Rosenblueth, Necessity for isoperimetric inequality constraints, Discrete and Continuous Dynamical Systems - Series A, 37 (2017), 1129-1158. https://doi.org/10.3934/dcds.2017047 [7] J.A. Becerril, J.F. Rosenblueth, The importance of being normal, regular and proper in the calculus of variations, Journal of Optimization Theory & Applications, 172 (2017), 759-773. https://doi.org/10.1007/s10957-017-1070-y [8] A. Ben-Tal, Second-order and related extremality conditions in nonlinear programming, Journal of Optimization Theory & Applications, 31 (1980), 143-165. https://doi.org/10.1007/bf00934107 [9] J.F. Bonnans, X. Dupuis, L. Pfeiffer, Second-order necessary conditions in Pontryagin form for optimal control problems, SIAM Journal of Control & Optimization, 52 (2014), 3887-3916. https://doi.org/10.1137/130923452 [10] J.F. Bonnans, N.P. Osmolovskii, Quadratic growth conditions in optimal control problems, Contemporary Mathematics, 514 (2010), 85-98. https://doi.org/10.1090/conm/514/10100 [11] J.F. Bonnans, A. Shapiro, Perturbation Analysis of Optimization Problems, Springer, New York, 2000. https://doi.org/10.1007/978-1-4612-1394-9 [12] F. Clarke, Functional Analysis, Calculus of Variations and Optimal Control, Springer, London, 2013. Critical cones for regular controls 467 [13] K.L. Cortez, J.F. Rosenblueth, A second order constraint qualification for certain classes of optimal control problems, WSEAS Transactions on Systems and Control, 11 (2016), 419-424. [14] K.L. Cortez, J.F. Rosenblueth, Extended critical directions for timecontrol constrained problems, International Journal of Circuits, Systems and Signal Processing, 11 (2017), 1-11. [15] M.R. de Pinho, J.F. Rosenblueth, Mixed constraints in optimal control: an implicit function theorem approach, IMA Journal of Mathematical Control and Information, 24 (2007) 197-218. https://doi.org/10.1093/imamci/dnl008 [16] E.G. Gilbert, D.S. Bernstein, Second order necessary conditions in optimal control: accessory-problem results without normality conditions, Journal of Optimization Theory & Applications, 41 (1983), 75-106. https://doi.org/10.1007/bf00934437 [17] G. Giorgi, A. Guerraggio, J. Thierfelder, Mathematics of Optimization: Smooth and Nonsmooth Case, Elsevier, Amsterdam, 2004. [18] M.R. Hestenes, Calculus of Variations and Optimal Control Theory, John Wiley, New York, 1966. [19] M.R. Hestenes, Optimization Theory, The Finite Dimensional Case, John Wiley, New York, 1975. [20] D. Hoehener, Variational approach to second-order optimality conditions for control problems with pure state constraints, SIAM Journal of Control & Optimization, 50 (2012), 1139-1173. https://doi.org/10.1137/110828320 [21] E. Levitin, A. Milyutin, N.P. Osomolovskii, Conditions of high order for a local minimum for problems with constraints, Russian Math Surveys, 33 (1978), 97-168. https://doi.org/10.1070/rm1978v033n06abeh003885 [22] P.D. Loewen, H. Zheng, Generalized conjugate points for optimal control problems, Nonlinear Analysis, Theory, Methods & Applications, 22 (1994), 771-791. https://doi.org/10.1016/0362-546x(94)90226-7 [23] E.J. McShane, The Lagrange Multiplier Rule, The American Mathematical Monthly, 80 (1973), 922-925. https://doi.org/10.1080/00029890.1973.11993409 468 Jorge A. Becerril, Karla L. Cortez and Javier F. Rosenblueth [24] A.A. Milyutin, N.P. Osmolovskii, Calculus of Variations and Optimal Control, Translations of Mathematical Monographs, Vol. 180, American Mathematical Society, Providence, Rhode Island, 1998. https://doi.org/10.1090/mmono/180 [25] N.P. Osmolovskii, Second order conditions for a weak local minimum in an optimal control problem (necessity, sufficiency), Soviet Math. Dokl., 16 (1975), 1480-1484. [26] N.P. Osmolovskii, H. Maurer, Applications to Regular and Bang-Bang Control, Second-Order Necessary and Sufficient Optimality Conditions in Calculus of Variations and Optimal Control, Advances in Design and Control, SIAM, Philadelphia, 2012. https://doi.org/10.1137/1.9781611972368 [27] J.F. Rosenblueth, Modified Critical Directions for Inequality Control Constraints, WSEAS Transactions on Systems and Control, 10 (2015), 215227. [28] J.F. Rosenblueth, G. Sánchez Licea, Cones of Critical Directions in Optimal Control, International Journal of Applied Mathematics and Informatics, 7 (2013), 55-67. [29] G. Stefani, P.L. Zezza, Optimality conditions for a constrained control problem, SIAM Journal on Control & Optimization, 34 (1996), 635-659. https://doi.org/10.1137/s0363012994260945 [30] J. Warga, A second-order Lagrangian condition for restricted control problems, Journal of Optimization Theory & Applications, 24 (1978), 475-483. https://doi.org/10.1007/bf00932890 Received: August 17, 2018; Published: September 19, 2018