Applications to Regular and Bang-Bang Control DC24_Osmolovskii-MaurerFM.indd 1 10/8/2012 9:29:39 AM Advances in Design and Control SIAM’s Advances in Design and Control series consists of texts and monographs dealing with all areas of design and control and their applications. Topics of interest include shape optimization, multidisciplinary design, trajectory optimization, feedback, and optimal control. The series focuses on the mathematical and computational aspects of engineering design and control that are usable in a wide variety of scientific and engineering disciplines. Editor-in-Chief Ralph C. Smith, North Carolina State University Editorial Board Athanasios C. Antoulas, Rice University Siva Banda, Air Force Research Laboratory Belinda A. Batten, Oregon State University John Betts, The Boeing Company (retired) Stephen L. Campbell, North Carolina State University Michel C. Delfour, University of Montreal Max D. Gunzburger, Florida State University J. William Helton, University of California, San Diego Arthur J. Krener, University of California, Davis Kirsten Morris, University of Waterloo Series Volumes Osmolovskii, Nikolai P. and Maurer, Helmut, Applications to Regular and Bang-Bang Control: Second-Order Necessary and Sufficient Optimality Conditions in Calculus of Variations and Optimal Control Biegler, Lorenz T., Campbell, Stephen L., and Mehrmann, Volker, eds., Control and Optimization with Differential-Algebraic Constraints Delfour, M. C. and Zolésio, J.-P., Shapes and Geometries: Metrics, Analysis, Differential Calculus, and Optimization, Second Edition Hovakimyan, Naira and Cao, Chengyu, L1 Adaptive Control Theory: Guaranteed Robustness with Fast Adaptation Speyer, Jason L. and Jacobson, David H., Primer on Optimal Control Theory Betts, John T., Practical Methods for Optimal Control and Estimation Using Nonlinear Programming, Second Edition Shima, Tal and Rasmussen, Steven, eds., UAV Cooperative Decision and Control: Challenges and Practical Approaches Speyer, Jason L. and Chung, Walter H., Stochastic Processes, Estimation, and Control Krstic, Miroslav and Smyshlyaev, Andrey, Boundary Control of PDEs: A Course on Backstepping Designs Ito, Kazufumi and Kunisch, Karl, Lagrange Multiplier Approach to Variational Problems and Applications Xue, Dingyü, Chen, YangQuan, and Atherton, Derek P., Linear Feedback Control: Analysis and Design with MATLAB Hanson, Floyd B., Applied Stochastic Processes and Control for Jump-Diffusions: Modeling, Analysis, and Computation Michiels, Wim and Niculescu, Silviu-Iulian, Stability and Stabilization of Time-Delay Systems: An EigenvalueBased Approach ¸ Adaptive Control Tutorial Ioannou, Petros and Fidan, Barıs, Bhaya, Amit and Kaszkurewicz, Eugenius, Control Perspectives on Numerical Algorithms and Matrix Problems Robinett III, Rush D., Wilson, David G., Eisler, G. Richard, and Hurtado, John E., Applied Dynamic Programming for Optimization of Dynamical Systems Huang, J., Nonlinear Output Regulation: Theory and Applications Haslinger, J. and Mäkinen, R. A. E., Introduction to Shape Optimization: Theory, Approximation, and Computation Antoulas, Athanasios C., Approximation of Large-Scale Dynamical Systems Gunzburger, Max D., Perspectives in Flow Control and Optimization Delfour, M. C. and Zolésio, J.-P., Shapes and Geometries: Analysis, Differential Calculus, and Optimization Betts, John T., Practical Methods for Optimal Control Using Nonlinear Programming El Ghaoui, Laurent and Niculescu, Silviu-Iulian, eds., Advances in Linear Matrix Inequality Methods in Control Helton, J. William and James, Matthew R., Extending H1 Control to Nonlinear Systems: Control of Nonlinear Systems to Achieve Performance Objectives DC24_Osmolovskii-MaurerFM.indd 2 10/8/2012 9:29:39 AM Applications to Regular and Bang-Bang Control Second-Order Necessary and Sufficient Optimality Conditions in Calculus of Variations and Optimal Control Nikolai P. Osmolovskii Systems Research Institute Warszawa, Poland University of Technology and Humanities in Radom Radom, Poland University of Natural Sciences and Humanities in Siedlce Siedlce, Poland Moscow State University Moscow, Russia Helmut Maurer Institute of Computational and Applied Mathematics Westfälische Wilhelms-Universität Münster Münster, Germany Society for Industrial and Applied Mathematics Philadelphia DC24_Osmolovskii-MaurerFM.indd 3 10/8/2012 9:29:39 AM Copyright © 2012 by the Society for Industrial and Applied Mathematics 10 9 8 7 6 5 4 3 2 1 All rights reserved. Printed in the United States of America. No part of this book may be reproduced, stored, or transmitted in any manner without the written permission of the publisher. For information, write to the Society for Industrial and Applied Mathematics, 3600 Market Street, 6th Floor, Philadelphia, PA 19104-2688 USA. No warranties, express or implied, are made by the publisher, authors, and their employers that the programs contained in this volume are free of error. They should not be relied on as the sole basis to solve a problem whose incorrect solution could result in injury to person or property. If the programs are employed in such a manner, it is at the user’s own risk and the publisher, authors, and their employers disclaim all liability for such misuse. Trademarked names may be used in this book without the inclusion of a trademark symbol. These names are used in an editorial context only; no infringement of trademark is intended. GNUPLOT Copyright © 1986–1993, 1998, 2004 Thomas Williams, Colin Kelley. Figure 8.3 reprinted with permission from John Wiley & Sons, Ltd. Figure 8.5 reprinted with permission from Springer Science+Business Media. Figure 8.8 reprinted with permission from Elsevier. Library of Congress Cataloging-in-Publication Data Osmolovskii, N. P. (Nikolai Pavlovich), 1948Applications to regular and bang-bang control : second-order necessary and sufficient optimality conditions in calculus of variations and optimal control / Nikolai P. Osmolovskii, Helmut Maurer. p. cm. -- (Advances in design and control ; 24) Includes bibliographical references and index. ISBN 978-1-611972-35-1 1. Calculus of variations. 2. Control theory. 3. Mathematical optimization. 4. Switching theory. I. Maurer, Helmut. II. Title. QA315.O86 2012 515’.64--dc23 2012025629 is a registered trademark. DC24_Osmolovskii-MaurerFM.indd 4 10/8/2012 9:29:45 AM j For our wives, Alla and Gisela DC24_Osmolovskii-MaurerFM.indd 5 10/8/2012 9:29:45 AM Contents List of Figures xi Notation xiii Preface xvii Introduction 1 I Second-Order Optimality Conditions for Broken Extremals in the Calculus of Variations 7 1 Abstract Scheme for Obtaining Higher-Order Conditions in Smooth Extremal Problems with Constraints 9 1.1 Main Concepts and Main Theorem . . . . . . . . . . . . . . . . . . . 9 1.2 Proof of the Main Theorem . . . . . . . . . . . . . . . . . . . . . . . 15 1.3 Simple Applications of the Abstract Scheme . . . . . . . . . . . . . . 21 2 Quadratic Conditions in the General Problem of the Calculus of Variations 2.1 Statements of Quadratic Conditions for a Pontryagin Minimum . . . 2.2 Basic Constant and the Problem of Its Decoding . . . . . . . . . . . 2.3 Local Sequences, Higher Order γ , Representation of the Lagrange Function on Local Sequences with Accuracy up to o(γ ) . . . . . . . 2.4 Estimation of the Basic Constant from Above . . . . . . . . . . . . 2.5 Estimation of the Basic Constant from Below . . . . . . . . . . . . 2.6 Completing the Proof of Theorem 2.4 . . . . . . . . . . . . . . . . . 2.7 Sufficient Conditions for Bounded Strong and Strong Minima in the Problem on a Fixed Time Interval . . . . . . . . . . . . . . . . . . . 3 27 . 27 . 34 . . . . 39 54 75 102 . 115 Quadratic Conditions for Optimal Control Problems with Mixed Control-State Constraints 127 3.1 Quadratic Necessary Conditions in the Problem with Mixed ControlState Equality Constraints on a Fixed Time Interval . . . . . . . . . . 127 3.2 Quadratic Sufficient Conditions in the Problem with Mixed ControlState Equality Constraints on a Fixed Time Interval . . . . . . . . . . 138 vii viii Contents 3.3 3.4 4 Quadratic Conditions in the Problem with Mixed Control-State Equality Constraints on a Variable Time Interval . . . . . . . . . . . . . . . 150 Quadratic Conditions for Optimal Control Problems with Mixed Control-State Equality and Inequality Constraints . . . . . . . . . . . 164 Jacobi-Type Conditions and Riccati Equation for Broken Extremals 183 4.1 Jacobi-Type Conditions and Riccati Equation for Broken Extremals in the Simplest Problem of the Calculus of Variations . . . . . . . . . . . 183 4.2 Riccati Equation for Broken Extremal in the General Problem of the Calculus of Variations . . . . . . . . . . . . . . . . . . . . . . . . . . 214 II Second-Order Optimality Conditions in Optimal Bang-Bang Control Problems 5 6 7 Second-Order Optimality Conditions in Optimal Control Problems Linear in a Part of Controls 5.1 Quadratic Optimality Conditions in the Problem on a Fixed Time Interval . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2 Quadratic Optimality Conditions in the Problem on a Variable Time Interval . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.3 Riccati Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.4 Numerical Example: Optimal Control of Production and Maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Second-Order Optimality Conditions for Bang-Bang Control 6.1 Bang-Bang Control Problems on Nonfixed Time Intervals . . . . . . 6.2 Quadratic Necessary and Sufficient Optimality Conditions . . . . . . 6.3 Sufficient Conditions for Positive Definiteness of the Quadratic Form on the Critical Cone K . . . . . . . . . . . . . . . . . . . . . . . 6.4 Example: Minimal Fuel Consumption of a Car . . . . . . . . . . . . 6.5 Quadratic Optimality Conditions in Time-Optimal Bang-Bang Control Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.6 Sufficient Conditions for Positive Definiteness of the Quadratic Form on the Critical Subspace K for Time-Optimal Control Problems . 6.7 Numerical Examples of Time-Optimal Control Problems . . . . . . 6.8 Time-Optimal Control Problems for Linear Systems with Constant Entries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Bang-Bang Control Problem and Its Induced Optimization Problem 7.1 Main Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2 First-Order Derivatives of x(tf ; t0 , x0 , θ ) with Respect to t0 , tf , x0 , and θ. Lagrange Multipliers and Critical Cones . . . . . . . . . . . 7.3 Second-Order Derivatives of x(tf ; t0 , x0 , θ ) with Respect to t0 , tf , x0 , and θ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4 Explicit Representation of the Quadratic Form for the Induced Optimization Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . 221 223 . 223 . 237 . 245 . 248 255 . 255 . 259 . 266 . 272 . 274 . 281 . 286 . 293 299 . 299 . 305 . 310 . 319 Contents ix 7.5 Equivalence of the Quadratic Forms in the Basic and Induced Optimization Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . 333 8 Numerical Methods for Solving the Induced Optimization Problem and Applications 8.1 The Arc-Parametrization Method . . . . . . . . . . . . . . . . . . 8.2 Time-Optimal Control of the Rayleigh Equation Revisited . . . . . 8.3 Time-Optimal Control of a Two-Link Robot . . . . . . . . . . . . 8.4 Time-Optimal Control of a Single Mode Semiconductor Laser . . . 8.5 Optimal Control of a Batch-Reactor . . . . . . . . . . . . . . . . . 8.6 Optimal Production and Maintenance with L1 -Functional . . . . . 8.7 Van der Pol Oscillator with Bang-Singular Control . . . . . . . . . . . . . . . . . . . . . . . 339 339 344 346 353 357 361 365 Bibliography 367 Index 377 List of Figures 2.1 2.2 Neighborhoods of the control at a point t1 of discontinuity. . . . . . . . . 40 Definition of functions (t, v) on neighborhoods of discontinuity points. . 51 4.1 Tunnel-diode oscillator. I denotes inductivity, C capacity, R resistance, I electric current, and D diode. . . . . . . . . . . . . . . . . . . . . . . 203 Rayleigh problem with regular control. (a) State variables. (b) Control. (c) Adjoint variables. (d) Solutions of the Riccati equation (4.140). . . . 204 Top left: Extremals x (1) , x (2) (lower graph). Top right: Variational solutions y (1) and y (2) (lower graph) to (4.145). Bottom: Envelope of neighboring extremals illustrating the conjugate point tc = 0.674437. . . 206 4.2 4.3 5.1 5.2 6.1 6.2 6.3 6.4 Optimal production and maintenance, final time tf = 0.9. (a) State variables x1 , x2 . (b) Regular production control v and bang-bang maintenance control m. (c) Adjoint variables ψ1 , ψ2 . (d) Maintenance control m with switching function φm . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250 Optimal production and maintenance, final time tf = 1.1. (a) State variables x1 , x2 . (b) Regular production control v and bang-singular-bang maintenance control m. (c) Adjoint variables ψ1 , ψ2 . (d) Maintenance control m with switching function φm . . . . . . . . . . . . . . . . . . . . 252 Minimal fuel consumption of a car. (a) State variables x1 , x2 . (b) Bangbang control u. (c) Adjoint variable ψ2 . (d) Switching function φ. . . . Time-optimal solution of the van der Pol oscillator, fixed terminal state (6.120). (a) State variables x1 and x2 (dashed line). (b) Control u and switching function ψ2 (dashed line). (c) Phase portrait (x1 , x2 ). (d) Adjoint variables ψ1 and ψ2 (dashed line). . . . . . . . . . . . . . . . . . Time-optimal solution of the van der Pol oscillator, nonlinear boundary condition (6.129). (a) State variables x1 and x2 (dashed line). (b) Control u and switching function ψ2 (dashed line). (c) Phase portrait (x1 , x2 ). (d) Adjoint variables ψ1 and ψ2 (dashed line). . . . . . . . . . . . . . Time-optimal control of the Rayleigh equation. (a) State variables x1 and x2 (dashed line). (b) Control u and switching function φ (dashed line). (c) Phase portrait (x1 , x2 ). (d) Adjoint variables ψ1 and ψ2 (dashed line). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi . 274 . 287 . 289 . 292 xii List of Figures 8.1 8.2 8.3 8.4 8.5 8.6 8.7 8.8 8.9 8.10 8.11 8.12 8.13 Time-optimal control of the Rayleigh equation with boundary conditions (8.31). (a) Bang-bang control and scaled switching function (×4), (b) State variables x1 , and x2 . . . . . . . . . . . . . . . . . . . . . . . Time-optimal control of the Rayleigh equation with boundary condition (8.40). (a) Bang-bang control u and scaled switching function φ (dashed line). (b) State variables x1 , x2 . . . . . . . . . . . . . . . . . . . . . . . lower arm OP , and angles q1 Two-link robot [67]: upper arm OQ, and q2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Control of the two-link robot (8.44)–(8.47). (a) Control u1 and scaled switching function φ1 (dashed line). (b) Control u2 and scaled switching function φ2 (dashed line). (c) Angle q1 and velocity ω1 . (d) Angle q2 and velocity ω2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Control of the two-link robot (8.53)–(8.57). (a) Control u1 . (b) Control u2 . (c) Angle q1 and velocity ω1 . (d) Angle q2 and velocity ω2 [17]. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Control of the two-link robot (8.53)–(8.57): Second solution. (a) Control u1 . (b) Control u2 . (c)Angle q1 and velocity ω1 . (d)Angle q2 and velocity ω2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Time-optimal control of a semiconductor laser. (a) Normalized photon density S(t) × 10−5 . (b) Normalized photon density N (t) × 10−8 . (c) Electric current (control) I (t) with I (t) = I0 = 20.5 for t < 0 and I (t) = I∞ = 42.5 for t > tf . (d) Adjoint variables ψS (t), ψN (t). . . . . Normalized photon number S(t) for I (t) ≡ 42.5 mA and optimal I (t) [46]. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Schematics of a batch-reaction with two control variables. . . . . . . . Control of a batch reactor with functional (8.79). Top row: Control u = (FB , Q) and scaled switching functions. Middle row: Molar concentrations MA and MB . Bottom row: Molar concentrations (MC , MD ) and energy holdup H . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Control of a batch reactor with functional (8.89). Top row: Control u = (FB , Q) and scaled switching functions. Middle row: Molar concentrations MA , MB . Bottom row: Molar concentrations (MC , MD ) and energy holdup H . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Optimal production and maintenance with L1 -functional (8.95). (a) State variables x and y. (b) Control variables v and m. (c), (d) Control variables and switching functions. . . . . . . . . . . . . . . . . . . . . . . . . . Control of the van der Pol oscillator with regulator functional. (a) Bangsingular control u. (b) State variables x1 and x2 . . . . . . . . . . . . . . 345 . 346 . 347 . 349 . 351 . 352 . 356 . 356 . 357 . 360 . 362 . 364 . 366 Notation {x | P (x)} R ξ+ := max{ξ , 0} : : : x ∈ Rn : set of elements x with the property P . set of real numbers. positive ⎛ part ⎞ of ξ ∈ R. x1 ⎜ ⎟ x = ⎝ ... ⎠ , xi ∈ R. y ∈ Rn∗ x ∗ ∈ Rn∗ yx = ni=1 yi xi x, y = ni=1 xi yi d(a) · X [a, b] ⊂ X : : xn ⇔ y = (y1 , . . . , yn ), yi ∈ R. x ∗ = (x1 , . . . , xn ) for x ∈ Rn . : x ∈ Rn , y ∈ Rn∗ . : : : : A X∗ x ∗ , x : : : K∗ : g (x0 ) : J (x) → min : σ (δx) : : λ = (α0 , . . . , αk , y ∗ ) : x, y ∈ Rn . dimension of the vector a. norm in the normed space X. [a, b] := {x ∈ X | x = αa + (1 − α)b, where α ∈ [0, 1]}, closed interval in the linear space X with endpoints a, b. closure of the set A. dual space to the normed space X. value of the linear functional x ∗ ∈ X∗ at the point x ∈ X. K ∗ := {x ∗ ∈ X∗ | x ∗ , x ≥ 0 for all x ∈ K}, dual cone of K. Fréchet derivative of the mapping g : X → Y at x0 . J (x) → min , fi (x) ≤ 0, i = 1, . . . , k, g(x) = 0 (g : X → Y ); abstract optimization problem in the space X. σ (δx) = max{m(δx), g(x0 + δx) }, m(δx) = maxi=0,...,k fi (x0 + δx), f0 (x) := J (x) − J (x0 ), violation function in the abstract optimization problem at x0 . Lagrange multipliers in the abstract optimization problem. xiii xiv Notation L(λ, x) : 0 : : K : z(·) : x(t) u(t) = [t0 , tf ] w := (x, u) p = (x0 , xf ) : : : : : p := (t0 , x0 , tf , xf ) ψ(t) ẋ = f (t, x, u) H (t, x, u, ψ) : : : : Hx , Hu : Hux : w 0 (t) = (x 0 (t), u0 (t)) : δx(t) or x̄(t) δu(t) or ū(t) Hux x̄, ū : : : k L(λ, x) = α0 J (x) + i=1 αi fi (x) + y ∗ , g(x) , Lagrange function in the abstract optimization problem. 0 = λ | αi ≥ 0, (i = 0, . . . , m), αi fi (x̂) = 0 (i = 1, . . . , k), k ∗ = 1, L (λ, x̂) = 0 , x i=0 αi + y Lx = ∂L/∂x, set of the normed tuples of Lagrange multipliers at x̂. K = {x̄ ∈ X | fi (x0 ), x̄ ≤ 0, i ∈ I ∪ {0}; g (x0 )x̄ = 0}, cone of critical directions (critical cone) at the point x0 , I = {i ∈ {1, . . . , k} | fi (x0 ) = 0}, set of active indices at x0 . element of a function space; z(t) is the value of z(·) at t. state variable, x ∈ Rd(x) . control variable, u ∈ Rd(u) . time interval. state x and control u. ∈ R2d(x) , if is a fixed time interval, x0 := x(t0 ), xf := x(tf ). ∈ R2+2d(x) , if is a variable time interval. adjoint variable, ψ ∈ Rd(x)∗ . control system, where ẋ:= dx dt . = ψf (t, x, u), Pontryagin function or Hamiltonian partial derivatives of H with respect to x and ∂H ∂H n∗ u, e.g., Hx := ∂H ∂x = ∂x1 , . . . , ∂xn ∈ R , n = d(x). second partial derivative of H with respect to x and u, ⎞ ⎛ 2H ∂2H . . . ∂u∂ 1 ∂x ∂u1 ∂x1 n ⎟ ⎜ ... . . . ⎠, Hux = ⎝ . . . 2 2 ∂ H . . . ∂u∂mH ∂um ∂x1 ∂xn m = d(u), n = d(x). pair satisfying the constraints of an optimal control problem. variation of the state x 0 (t). variation of the control u0 (t). n ∂2H = m i=1 j =1 ∂ui ∂xj ūi x̄j , m = d(u), n = d(x). Notation xv C([t0 , tf ], Rn ) : PC([t0 , tf ], Rn ) : = {t1 , . . . , ts } : C 1 ([t0 , tf ], Rn ) : PC 1 ([t0 , tf ], Rn ) : L1 ([t0 , tf ], Rm ) : L2 ([t0 , tf ], Rm ) : L∞ ([t0 , tf ], Rm ) : W 1,1 ([t0 , tf ], Rn ) : W 1,2 ([t0 , tf ], Rn ) : W 1,∞ ([t0 , tf ], Rn ) : P W 1,1 ([t0 , tf ], Rd(x) ) : P W 1,2 ([t0 , tf ], Rd(x) ) : space of continuous vector functions x : [t0 , tf ] → Rn with norm x(·) ∞ = x(·) C = maxt∈[t0 ,tf ] |x(t)|. class of piecewise continuous functions u : [t0 , tf ] → Rn . set of discontinuity points of u0 (·) ∈ PC([t0 , tf ], Rn ), u0k− = u0 (tk −), u0k+ = u0 (tk +), [u0 ]k = u0k+ − u0k− . space of continuously differentiable functions x : [t0 , tf ] → Rn endowed with the norm x(·) C 1 = max{ x(·) ∞ , ẋ(·) ∞ }. class of continuous functions x : [t0 , tf ] → Rn with a piecewise continuous derivative. space of Lebesgue integrable functions u : [t0 , tf ] → Rm endowed with the norm t u(·) 1 = t0f |u(t)| dt. Hilbert space of square Lebesgue integrable functions u : [t0 , tf ] → Rm with the inner product t u(·), v(·) = t0f u(t), v(t) dt. space of bounded measurable functions u : [t0 , tf ] → Rm endowed with the norm u(·) ∞ = ess supt∈[t0 ,tf ] |u(t)|. space of absolutely continuous functions x : [t0 , tf ] → Rn endowed with the norm t x(·) 1,1 = |x(t0 )| + t0f |ẋ(t)| dt. Hilbert space of absolutely continuous functions x : [t0 , tf ] → Rn with square integrable derivative and inner product x(·), y(·) = t x(t0 ), y(t0 ) + t0f ẋ(t), ẏ(t) dt. space of Lipschitz continuous functions x : [t0 , tf ] → Rn endowed with the norm x(·) 1,∞ = |x(t0 )| + ẋ(·) ∞ . space of piecewise continuous functions x̄ : [t0 , tf ] → Rd(x) that are absolutely continuous on each of the intervals of the set (t0 , tf ) \ . Hilbert space of functions x̄(·) ∈ P W 1,1 ([t0 , tf ], Rd(x) ) such that the first derivative x̄˙ is Lebesgue square integrable; with [x̄]k = x̄ k+ − x̄ k− = x̄(tk +) − x̄(tk −) and x̄, ȳ = x̄(t0 ), ȳ(t0 ) + s innerk product tf ˙ k ˙ k=1 [x̄] , [ȳ] + t0 x̄(t), ȳ(t) dt. Preface The book is devoted to the theory and application of second-order necessary and sufficient optimality conditions in the Calculus of Variations and Optimal Control. The theory is developed for control problems with ordinary differential equations subject to boundary conditions of equality and inequality type and mixed control-state constraints of equality type. The book exhibits two distinctive features: (a) necessary and sufficient conditions are given in the form of no-gap conditions, and (b) the theory covers broken extremals, where the control has finitely many points of discontinuity. Sufficient conditions for regular controls that satisfy the strict Legendre condition can be checked either via the classical Jacobi condition or through the existence of solutions to an associated Riccati equation. Particular emphasis is given to the study of bang-bang control problems. Bang-bang controls induce an optimization problem with respect to the switching times of the control. It is shown that the classical second-order sufficient condition for the Induced Optimization Problem (IOP), together with the so-called strict bang-bang property, ensures second-order sufficient conditions (SSC) for the bang-bang control problem. Numerical examples in different areas of application illustrate the verification of SSC for both regular controls and bang-bang controls. SSC are crucial for exploring the sensitivity analysis of parametric optimal control problems. It is well known in the literature that for regular controls satisfying the strict Legendre condition, SSC allow us to prove the parametric solution differentiability of optimal solutions and to compute parametric sensitivity derivatives. This property has lead to efficient real-time control techniques. Recently, similar results have been obtained for bangbang controls via SSC for the IOP. Though the discussion of sensitivity analysis and the ensuing real-time control technques are an immediate consequence of the material presented in this book, a systematic treatment of these issues is beyond the scope of this book. The results of Sections 1.1–1.3 are due to Levitin, Milyutin, and Osmolovskii. The results of Section 6.8 were obtained by Milyutin and Osmolovskii. The results of Sections 2.1–3.4, 5.1, 5.2, 6.1, 6.2, and 6.5 were obtained by Osmolovskii; some important ideas used in these sections are due to Milyutin. The results of Sections 4.1 and 4.2 (except for Section 4.1.5) are due to Lempio and Osmolovskii. The results of Sections 5.3, 6.3, 6.6, and 7.1–7.5 were obtained by Maurer and Osmolovskii. All numerical examples in Sections 4.1, 5.4, 6.4, and Chapter 8 were collected and investigated by Maurer, who is grateful for the numerical assistance provided by Christof Büskens, Laurenz Göllmann, Jang-Ho Robert Kim, and Georg Vossen. Together we solved a lot more bang-bang and singular control problems than could be included in this book. H. Maurer is indebted to Yalçin Kaya for drawing his attention to the arc-parametrization method presented in Section 8.1.2. xvii xviii Preface Acknowledgments. We are thankful to our colleagues at Moscow State University and University of Münster for their support. A considerable part of the book was written by the first author during his work in Poland (System Research Institute of Polish Academy of Science in Warszawa, Politechnika Radomska in Radom, Siedlce University of Natural Sciences and Humanities), and during his stay in France (Ecole Polytechnique, INRIA Futurs, Palaiseau). We are grateful to J. Frédéric Bonnans from laboratory CMAP (INRIA) at Ecole Polytechnique for his support and for fruitful discussions. Many important ideas used in this book are due to A.A. Milyutin. N.P. Osmolovskii was supported by the grant RFBR 11-01-00795; H. Maurer was supported by the grant MA 691/18 of the Deutsche Forschungsgemeinschaft. Finally, we wish to thank Elizabeth Greenspan, Lisa Briggeman, and the SIAM compositors for their helpful, patient, and excellent work on our book. Introduction By quadratic conditions, we mean second-order extremality conditions formulated for a given extremal in the form of positive (semi)definiteness of the corresponding quadratic form. So, in the simplest problem of the calculus of variations tf F (t, x, ẋ) dt → min, x(t0 ) = a, x(tf ) = b, t0 considered on the space C 1 of continuously differentiable vector-valued functions x(·) on the given closed interval [t0 , tf ], the quadratic form is as follows: tf ˙ x̄˙ dt, x̄(t0 ) = x̄(tf ) = 0, Fxx x̄, x̄ + 2Fẋx x̄, x̄˙ + Fẋ ẋ x̄, ω= t0 where the second derivatives Fxx , Fẋx , and Fẋ ẋ are calculated along the extremal x(·) that is of interest to us. It is considered in the space W 1,2 of absolutely continuous functions ˙ By using ω for a given extremal, one formulates x̄ with square integrable derivative x̄. a necessary second-order condition for a weak minimum (the positive semidefiniteness of the form), as well as a sufficient second-order condition for a weak minimum (the positive definiteness of the form). As is well known, these quadratic conditions are equivalent (under the strengthened Legendre condition) to the corresponding Jacobi conditions. The simplest problem of the calculus of variations can also be considered in the space W 1,∞ of Lipschitz-continuous functions x(·), and then, in particular, there arises the problem of studying the extremality of broken extremals x(·), i.e., extremals such that the derivative ẋ(·) has finitely many points of discontinuity of the first kind. What are second-order conditions for broken extremals, and what is the corresponding quadratic form for them? A detailed study of this problem for the simplest problem was performed by the first author in the book [79], where quadratic necessary and sufficient conditions for the so-called “Pontryagin minimum” (corresponding to L1 -small variations of the control u = ẋ under the condition of their uniform L∞ -boundedness) were obtained, and also the relation between the obtained conditions and the conditions for the strong minimum (and also the so-called “-weak” and “bounded strong” minima) was established. For an extremal x(·) with one break at a point t∗ , the corresponding form becomes (cf. [39, 85, 86, 107]): tf ˙ x̄˙ dt, = a ξ̄ 2 + 2[Fx ]x̄av ξ̄ + Fxx x̄, x̄ + 2Fẋx x̄, x̄˙ + Fẋ ẋ x̄, t0 where ξ̄ is a numerical parameter, x̄(·) is a function that can have a nonzero jump [x̄] := x̄(t∗ +) − x̄(t∗ −) at the point t∗ and is absolutely continuous on the semiopen intervals 1 2 Introduction ˙ is square integrable, and, moreover, the following [t0 , t∗ ) and (t∗ , tf ], and the derivative x̄(·) conditions hold: [x̄] = [ẋ]ξ̄ , x̄(t0 ) = x̄(tf ) = 0. Here, [Fx ] and [ẋ] denote the jumps of the gradient Fx (t, x(t), ẋ(t)) and the derivative ẋ(t) of the extremal at the point t∗ , respectively (e.g., [ẋ] = ẋ(t∗ +) − ẋ(t∗ −)); a is the derivative in t of the function Fẋ (t, x(t), ẋ(t))[ẋ] − F (t, x(t), ẋ(t∗ +)) + F (t, x(t), ẋ(t∗ −)) at the same point (its existence is proved); and x̄av is the average value of the left-hand and the right-hand values of x̄ at t∗ , i.e., x̄av = 12 (x̄(t∗ −) + x̄(t∗ +)). The Weierstrass condition (the minimum principle) implies the inequality a ≥ 0, which complements the well-known Weierstrass–Erdmann conditions for broken extremals. In [96] and [97] it was shown (see also [90]) that the problem of the “sign” of the quadratic form can be studied by using methods analogous to the classical methods. The Jacobi conditions and criteria formulated by using the corresponding Riccati equation are extended to the case of a broken extremal. In all these conditions, a new aspect consists only of the fact that the solutions of the corresponding differential equations should have completely certain jumps at the point of break of the extremal. Moreover, it was shown in [97] that, as in the classical case, the quadratic form reduces to a sum of squares solving, by the corresponding Riccati equation satisfying (this is the difference from the classical case), a definite jump condition at the point t∗ , which also gives a criterion for positive definiteness of the form and, therefore, a sufficient extremality condition for a given extremal. In book [79], quadratic extremality conditions for discontinuous controls were also presented in the following problem on a fixed time interval [t0 , tf ]: J(x(·), u(·)) = J (x(t0 ), x(tf )) → min, F (x(t0 ), x(tf )) ≤ 0, K(x(t0 ), x(tf )) = 0, ẋ = f (t, x, u), g(t, x, u) = 0, (t, x, u) ∈ Q, where Q is an open set, x, u, F , K, and g are vector-valued functions, and J is a scalarvalued function. The functions J , F , K, f , and g belong to the class C 2 , and, moreover, the derivative gu has the maximum rank on the surface g = 0 (the nondegeneracy condition for the relation g = 0). We seek the minimum among pairs (x(·), u(·)) admissible by the constraints such that the function x(·) is absolutely continuous and u(·) is bounded and measurable. This statement corresponds to the general canonical optimal control problem in the Dubovitskii–Milyutin form, but, in contrast to the latter, it is considered on a fixed interval of time, and, which is of special importance, it does not contain pointwise (or local, in the Dubovitskii–Milyutin terminology) mixed inequality-type constraints ϕ(t, x, u) ≤ 0. Precisely, these constraints caused the major difficulties in the study of quadratic conditions [85, 88]. Also, due to the absence of local inequalities, we refer this problem to the calculus of variations (rather than to optimal control) and call it the general problem of the calculus of variations (with mixed equality-type constraints, on a fixed time interval). Its statement is close to the Mayer problem, but the existence of endpoint inequality-type constraints determines its specifics. Introduction 3 On the other hand, this problem, even being referred to as the calculus of variations, is sufficiently general and its statement is close to optimal control problems, especially owing to the local relation g(t, x, u) = 0. In [79], it was shown how, by using quadratic conditions for the general problem of calculus of variations, one can obtain quadratic (necessary and sufficient) conditions in optimal control problems in which the controls enter linearly and the constraint on the control is given in the form of a convex polyhedron under the assumption that the optimal control is piecewise-constant and (outside the switching points) belongs to vertices of the polyhedron (the so-called bang-bang control). To show this, in [79], we first used the property that the set V of vertices of a polyhedron U can be given by a nondegenerate relation g(u) = 0 on an open set Q consisting of disjoint open neighborhoods of vertices. This allows us to write quadratic necessary conditions for bang-bang controls. Further, in [79], it was shown that a sufficient minimality condition on V guarantees (when the control enters linearly) the minimum on its convexification U = co V . In this way, the quadratic sufficient conditions were obtained for bang-bang controls. However, in [79], there is a substantial gap stemming from the fact that, to avoid making the book too long, the authors decided to omit the proofs of quadratic conditions for the general problem of the calculus of variations and restricted themselves to their formulation and the presentation of proofs only for the simplest problem. Although the latter gives the idea of the proofs in the general case, there are no formal proofs of quadratic conditions for the general problem of the calculus of variations in [79]. Part I of the present book is devoted to removing this gap. Therefore, Part I can be considered as a necessary supplement to the book [79]. At the same time, the material contained in Part I is independent and is a complete theory of quadratic conditions for smooth problems of the calculus of variations and optimal control that are covered by the statement presented above. Part I is organized as follows. First, in Chapter 1, we present a fragment of the abstract theory of higher-order conditions of Levitin, Milyutin, and Osmolovskii [54, 55, 56], more precisely, a modification of this theory for smooth problems on the set of sequences determining one or another concept of minimum. In this theory, by a higher order, we mean a nonnegative functional determining a growth estimate of the objective function on admissible sequences of variations from . The main result of the abstract theory is that for a given class of problems, a given higher order γ , and a given set of sequences , one defines a constant Cγ (by using the Lagrange function) such that Cγ ≥ 0 is a necessary minimality condition (corresponding to ), and Cγ > 0 is a sufficient condition. The constant Cγ is said to be basic. In each concrete class of problems (for given and γ ), there arises the problem of “decoding” the basic constant. By decoding we mean the simplest method for calculating its sign. We illustrate the decoding of the basic constant by two simple examples, obtaining conditions of the order γ , where γ is a certain quadratic functional. In Chapter 2, on the basis of the results of Chapter 1, we create a quadratic theory of conditions for Pontryagin minimum in the general problem of the calculus of variations without local mixed constraint g(t, x, u) = 0. We perform the decoding of the basic constant for the set of the so-called “Pontryagin sequences” and a special higher order γ , which is characteristic for extremals with finitely many discontinuities of the first kind of control. We first estimate the basic constant from above, thus obtaining quadratic necessary conditions for Pontryagin minimum, and then we estimate it from below, thus obtaining sufficient conditions. After that (in Section 2.7), we establish the relation of the obtained sufficient conditions for Pontryagin minimum with conditions for strong minimum. 4 Introduction In Chapter 3, we extend the quadratic conditions obtained in Chapter 2 to the general problem with the local relation g(t, x, u) = 0 using a special method of projection contained in [79]. Moreover, we extend these conditions to the problem on a variable interval of time using a simple change of time variable. We also formulate, without proofs, quadratic conditions in an optimal control problem with local relations g(t, x, u) = 0 and ϕ(t, x, u) ≤ 0. The proofs are set forth in [94, 95]. In Chapter 4, following the results of papers [96] and [97], we derive the tests for positive semidefiniteness and that for positive definiteness of the quadratic form on the critical cone K (obtained in Chapter 2 for extremals with jumps of the control); these are necessary and sufficient conditions for local minimum, respectively. First, we derive such tests for the simplest problem of the calculus of variations and for extremal with only one corner point. In these tests we exploit the classical Jacobi and Riccati equations, but, as it was said, we use the discontinuous solutions to these equations satisfying specific jump conditions at the corner point of extremal. In the proofs we use the ideas in [25] and [26]. Namely, we consider a one parameter family of the auxiliary minimization problems and reduce the question of the “sign” of our quadratic form on the critical subspace to the condition of the existence of a nonzero point of minimum in the auxiliary problem for a certain value of the parameter. This condition was called in [25] the “passage of quadratic form through zero”. Then we obtain a dual test for minimum in the auxiliary problem. As a result we arrive at a generalization of the concept of conjugate point. Such a point is called a -conjugate, where is a singleton consisting of one corner point t∗ of extremal at hand. This generalization allows us to formulate both necessary and sufficient second-order optimality conditions for broken extremal. Next, we concentrate only on sufficient conditions for positive definiteness of the quadratic form in the auxiliary problem. Following [97], we show that if there exists a solution to the Riccati matrix equation satisfying a definite jump condition, then the quadratic form can be transformed into a perfect square, just as in the classical case. This gives the possibility of proving a sufficient condition for positive definiteness of the quadratic form in the auxiliary problem and thus to obtain one more sufficient condition for optimality of broken extremal. First we prove this result for extremal with one corner point in the simplest problem of the calculus of variations, and then for extremal with finitely many points of discontinuity of the control in the general problem of the calculus of variations. Part II is devoted to optimal control problems. In Chapter 5, we derive quadratic optimality conditions for optimal control problems with a vector control variable having two components: a continuous unconstrained control appearing nonlinearly in the control system and a bang-bang control appearing linearly and belonging to a convex polyhedron. Such type of control problem arises in many applications. The proofs of quadratic optimality conditions for the mixed continuous-bang case are very similar to the proofs given in [79] for the pure bang-bang case, but some modifications were inevitable. We demonstrate these modifications. In the proofs we use the optimality conditions obtained in Chapter 3 for extremal with jumps of the control in the general problem of calculus of variations. Further, we show that, also for the mixed case, there is a techniques for checking positive definiteness of the quadratic form on the critical cone via a discontinuous solution of an associated Riccati equation with corresponding jump conditions (for this solution) at the points of discontinuity of bang-bang control. This technique is applied to an economic control problem in optimal production and maintenance which was introduced by Cho, Abad, and Parlar [22]. We show that the numerical solution obtained in Maurer, Kim, and Introduction 5 Vossen [67] satisfies the second-order test derived in Chapter 5 while existing sufficiency results fail to hold. In Chapter 6, we investigate the pure bang-bang case. We obtain second-order necessary and sufficient optimality conditions for this case as a consequence of the conditions obtained in Chapter 5. In the pure bang-bang case, the conditions amount to testing the positive (semi)definiteness of a quadratic form on a finite-dimensional critical cone. Nevertheless, the assumptions are appropriate for numerical verification only in some special cases. Therefore, again we study various transformations of the quadratic form and the critical cone which will be tailored to different types of control problems in practice. In particular, by means of a solution to a linear matrix differential equation, the quadratic form can be converted to perfect squares. We demonstrate by practical examples that the obtained conditions can be verified numerically. We also study second-order optimality conditions for time optimal control problems with control appearing linearly. More specifically, we consider the special case of time optimal bang-bang controls with a given initial and terminal state. We aim at showing that an approach similar to the above-mentioned Riccati equation approach works as well for such problems. Again, the test requires us to find a solution of a linear matrix differential equation which satisfies certain jump conditions at the switching points. We discuss three numerical examples that illustrate the numerical procedure of verifying positive definiteness of the corresponding quadratic forms. Finally, following [79], we study secondorder optimality conditions in a simple, but important, class of time optimal control problems for linear systems with constant entries. Second-order optimality conditions in bang-bang control problems have been derived in the literature in two different forms. The first form was discussed above. The second form belongs to Agrachev, Stefani, and Zezza [1], who first reduce the bang-bang control problem to a finite-dimensional Induced Optimization Problem (IOP) and then show that well-known sufficient optimality conditions for the induced problem supplemented by the strict bang-bang property furnish sufficient conditions for the bang-bang control problem. In Chapter 7, we establish the equivalence of both forms of sufficient conditions. The proof of this equivalence make extensive use of explicit formulas for first- and secondorder derivatives of the trajectory with respect to variations of the optimization variable ζ comprising the switching times, the free initial and final time, and the free initial state. We formulate the IOP with optimization variable ζ which is associated with the bang-bang control problem. We give formulas for the first- and second-order derivatives of trajectories with respect to ζ which follow from elementary properties of ordinary differential equations (ODEs). The formulas are used to establish the explicit relations between the multipliers of Pontryagin’s minimum principle and the Lagrange multipliers, critical cones and quadratic forms of the original and IOPs. In our opinion, the resulting formulas seem to have been mostly unknown in the literature. These formulas provide the main technical tools to obtain explicit representations of the second-order derivatives of the Lagrangian. The remarkable fact to be noted here is that by using a suitable transformation, these derivatives are seen to involve only first-order variations of the trajectory with respect to ζ . This property facilitates considerably the numerical computation of the Hessian of the Lagrangian. Thus, we arrive at a representation of the quadratic form associated with the Hessian of the Lagrangian. Finally, Chapter 8 is devoted to numerical methods for solving the IOP and testing the second-order sufficient conditions in Theorem 7.10. After a brief survey on numerical 6 Introduction methods for solving optimal control problems, we present in Section 8.1.2 the arcparametrization method for computing bang-bang controls [44, 45, 66] and its extension to piecewise feedback controls [111, 112, 113]. Arc parametrization can be efficiently implemented using the code NUDOCCCS developed by Büskens [13, 14]. Several numerical examples illustrate the arc-parametrization method and the verification of second-order conditions. Chapter 1 Abstract Scheme for Obtaining Higher-Order Conditions in Smooth Extremal Problems with Constraints Here, we present the general theory of higher-order conditions [78] which will be used in what follows in obtaining quadratic optimality conditions in the canonical problem of the calculus of variations. In Section 1.1, we formulate the main result of the general theory in the smooth case. Section 1.2 is devoted to its proof. In Section 1.3, we present two simple applications of the general theory. 1.1 1.1.1 Main Concepts and Main Theorem Minimum on a Set of Sequences Let X and Y be Banach spaces. Let a set ⊂ X, functionals J : → R1 , fi : → R1 , i = 1, . . . , k, and an operator g : → Y be given. Consider the problem J (x) → min; fi (x) ≤ 0, i = 1, . . . , k; g(x) = 0; x ∈ . (1.1) Let a point x0 ∈ satisfy the constraints, and let us study its optimality. By {δxn }, and also by {x̄n }, we denote countable sequences in X. Denote by 0 the set of sequences {x̄n } converging in norm to zero in X. Let us introduce the set of sequences determining the type of the minimum at the point x0 , which will be used. Let be an arbitrary set of sequences {δxn } satisfying the following conditions: (a) is closed with respect to passing to a subsequence; (b) + 0 ⊂ ; i.e., the conditions {δxn } ∈ and {x̄n } ∈ 0 imply {δxn + x̄n } ∈ (in this case, we say that sustains a 0 -extension). Moreover, it is assumed that the following condition holds for , , and x0 : (c) for any sequence {δxn } ∈ , we have x0 + δxn ∈ for sufficiently large n (in this case, we say that the set is absorbing for at the point x0 ). We give the following definition for problem (1.1). Definition 1.1. We say that the minimum is attained at a point x0 on (or x0 is a point of -minimum) if there is no sequence {δxn } ∈ such that for all n, J (x0 + δxn ) − J (x0 ) < 0, fi (x0 + δxn ) ≤ 0 (i = 1, . . . , k), 9 g(x0 + δxn ) = 0. 10 Chapter 1. Abstract Scheme for Obtaining Higher-Order Conditions In a similar way, the strict minimum on at x0 is defined. We need only replace the strict inequality J (x0 + δxn ) − J (x0 ) < 0 in the previous definition with nonstrict inequality and require additionally that the sequence {δxn } contains nonzero terms. Obviously, the minimum on 0 is a local minimum. If 0 ⊂ , then the minimum on is not weaker than a local minimum. The inclusion 0 ⊂ holds iff contains a zero sequence. This condition holds in all applications of the general theory. In what follows, the point x0 is fixed, and, therefore, as a rule, it will be omitted in the definitions and notation. By δ, we denote the set of variations δx ∈ X such that x0 + δx ∈ . Note that 0 ∈ δ. We set f0 (x) = J (x) − J (x0 ) for x ∈ . Denote by S the system consisting of the functionals f0 , f1 , . . . , fk and the operator g. The concepts of minimum and strict minimum on are naturally extended to the system S. The concepts introduced below can also be related to problem (1.1), as well as to the system S. Sometimes, it is more convenient to speak about the system and not about the problem. For δx ∈ δ, we set m(δx) = max0≤i≤k fi (x0 + δx). Let g be the set of sequences {δxn } ∈ such that g(x0 + δxn ) = 0 for all sufficiently large n. Consider the condition m ≥ 0 | g . By definition, this condition means that for any sequence {δxn } ∈ g , there exists a number starting from which m(δxn ) ≥ 0. In what follows, such a notation will be used without additional explanations. The following proposition follows from the definitions directly. Proposition 1.2. If the minimum is attained on , then m ≥ 0 | g . Therefore, the condition m ≥ 0 | g is necessary for the minimum on . It will serve as a source of other, coarser necessary conditions. Now let us consider two obvious sufficient conditions. Let + be the set of all sequences from that do not vanish. Define + g analogously. The following proposition follows directly from the definitions. Proposition 1.3. The condition m > 0 | + g is equivalent to the strict minimum on . For δx ∈ δ, we set σ (δx) = max{m(δx), g(x0 + δx) }. We say that σ is the violation function. If, in problem (1.1), there is no equality-type constraint g(x) = 0, then we set σ (δx) = m+ (δx), where m+ = max{m, 0}. The following proposition is elementary. Proposition 1.4. Condition σ > 0 | + is equivalent to the strict minimum on . 1.1.2 Smooth Problem Let us formulate the assumptions in problem (1.1) that define it as a smooth problem. These assumptions are related not only to the functionals J , f1 , . . . , fk , the operator g, and the point x0 , but also to the set of sequences . First let us give several definitions. Let Z be a Banach space. A mapping h : → Z is said to be -continuous at x0 if the condition h(x0 + δxn ) − h(x0 ) → 0 (n → ∞) holds for any {δxn } ∈ . A -continuous mapping is said to be strictly -differentiable at x0 if there exists a linear operator H : X → Z such that for any sequences {δxn } ∈ and {x̄n } ∈ 0 , there exists a sequence {zn } in Z such that zn → 0, and for all sufficiently large n, we have the relation 1.1. Main Concepts and Main Theorem 11 h(x0 + δxn + x̄n ) = h(x0 + δxn ) + H x̄n + zn x̄n . The definition easily implies the uniqueness of the operator H . If contains the zero sequence (and hence 0 ⊂ ), then H is the Frechét derivative of the operator h at the point x0 , and the strict -differentiability implies the strict differentiability. In what follows, we set H = h (x0 ). The function of two variables, (δx, x̄) −→ h(x0 + δx) + h (x0 )x̄, that maps δ × X into Z, is called a fine linear approximation of the operator h at the point x0 on . These concepts are used for operators, as well as for functionals. It is assumed that all functionals J , f1 , . . . , fk and the operator g in problem (1.1) are -continuous at x0 . Introduce the set of active indices: I = {i ∈ {0, 1, . . . , k} fi (x0 ) = 0}, where f0 (x) = J (x) − J (x0 ). (1.2) Obviously, 0 ∈ I . It is assumed that the functionals fi , i ∈ I , and the operator g are strictly -differentiable at x0 . Also, it is assumed that either g (x0 )X = Y (in this case, we say that for g at x0 on , the Lyusternik condition holds) or the image g (x0 )X is closed in Y and has a direct complement which is a closed subspace in Y . Precisely, these assumptions define a smooth problem (smooth system) on at the point x0 . In this chapter we will consider only this type of problem. 1.1.3 Conditions of Order γ As usual, by the order of an extremality condition, one means the order of the highest derivative entering this condition. We give another definition of the order. A functional γ : δ → R1 is called an order on if it is nonnegative on δ, -continuous at zero, and γ (0) = 0. An order γ on is said to be higher if it is strictly -differentiable at zero, and hence γ (0) = 0. An order γ on is said to be strict if γ > 0 + . Let γ be strict higher order on . Define the following two conditions on : the γ -necessity and the γ -sufficiency. To this end, we set m . Cγ (m, g ) = inf lim inf g γ Let us explain that, in calculating this quantity, for each sequence {δxn } ∈ g that does not vanish, we first calculate the limit inferior (lim inf ) of the ratio m(δxn )/γ (δxn ) as n → ∞, and then we take the greatest lower bound of the limits inferior over the whole set of sequences from g that do not vanish. An analogous notation will be used for other functions and other sets of sequences. Proposition 1.2 implies the following assertion. Proposition 1.5. If we have the minimum on , then Cγ (m, g ) ≥ 0. The condition Cγ (m, g ) ≥ 0 is called the γ -necessity on . It is easy to see that the γ -necessity on is equivalent to the following condition: there are no ε > 0 and sequence {δxn } ∈ + such that fi (x0 + δxn ) ≤ −εγ (δxn ) (i = 0, . . . , k), g(x0 + δxn ) = 0. 12 Chapter 1. Abstract Scheme for Obtaining Higher-Order Conditions It is convenient to compare the concept of γ -necessity in this form with the concept of minimum on . Further, we set σ Cγ (σ , ) = inf lim inf . γ Propositions 1.3 and 1.4 imply the following assertion. Proposition 1.6. Each of the two conditions Cγ (m, g ) > 0 and Cγ (σ , ) > 0 is sufficient for the strict minimum on . The condition Cγ (σ , ) > 0 is said to be γ -sufficiency on . It is equivalent to the following condition: there exists C > 0 such that σ ≥ Cγ | + . Since, obviously, Cγ (m, g )+ = Cγ (m+ , g ) = Cγ (σ , g ) ≥ Cγ (σ , ), the inequality Cγ (σ , ) > 0 implies the inequality Cγ (m, g ) > 0. Therefore, the inequality Cγ (σ , ) > 0 is not a weaker sufficient condition for the strict minimum on than the inequality Cγ (m, g ) > 0. In what follows, we will show that if the Lyusternik condition holds, then these two sufficient conditions are equivalent. As the main sufficient condition, we will consider the inequality Cγ (σ , ) > 0, i.e., the γ -sufficiency on . Therefore, the γ -necessity and the γ -sufficiency on are obvious weakening and strengthening of the concept of minimum on , respectively. We aim at obtaining the criteria for the γ -conditions, i.e., the γ -necessity and the γ -sufficiency, that are formulated by using the Lagrange function. 1.1.4 Lagrange Function. Main Result By λ = (α, y ∗ ), we denote an arbitrary tuple of multipliers, where α = (α0 , α1 , . . . , αk ) ∈ Rk+1 , y ∗ ∈ Y ∗ . Denote by 0 the set of tuples λ such that αi ≥ 0 (i = 0, . . . , k), k αi + y ∗ = 1, i=0 αifi (x0 ) = 0 (i = 1, . . . , k), k (1.3) ∗ αifi (x0 ) + y g (x0 ) = 0, i=0 where y ∗ g (x0 ), x = y ∗ , g (x0 )x for all x ∈ X by definition (here, we prefer not to use Therefore, 0 is the set of normalized tuples the notation g (x0 )∗ for the adjoint operator). of Lagrange multipliers. The relation αi + y ∗ = 1 is the normalization condition here. Such a normalization is said to be standard. Introduce the Lagrange function L(λ, x) = k αifi (x) + y ∗ , g(x) , x ∈ , i=0 and the functions (λ, δx) = L(λ, x0 + δx) = 0 (δx) = max (λ, δx), λ∈0 k αifi (x0 + δx) + y ∗ , g(x0 + δx) , i=0 δx ∈ δ. Here and in what follows, we set max∅ (·) = −∞. 1.1. Main Concepts and Main Theorem 13 Denote by σ γ the set of sequences {δxn } ∈ satisfying the condition σ (δxn ) ≤ O(γ (δxn )). The latter means that there exists C > 0 depending on the sequence such that σ (δxn ) ≤ Cγ (δxn ) for all n. We set 0 . Cγ (0 , σ γ ) = inf lim inf σ γ γ The constant Cγ (0 , σ γ ) is said to be basic. It turns out that for an arbitrary higher strict order γ on , the constant Cγ (0 , σ γ ) allows us to formulate the following pair of adjacent conditions for : the inequality Cγ (0 , σ γ ) ≥ 0 is necessary for the minimum on , and the strict inequality Cγ (0 , σ γ ) > 0 is sufficient for the strict minimum on . Moreover, the following assertion holds. Theorem 1.7. (a) If g (x0 )X = Y , then the inequality Cγ (0 , σ γ ) ≥ 0 is equivalent to the inequality Cγ (m, g ) ≥ 0. If g (x0 )X = Y , then 0 ≥ 0, and , therefore, Cγ (0 , σ γ ) ≥ 0. (b) The inequality Cγ (0 , σ γ ) > 0 is always equivalent to the inequality Cγ (σ , ) > 0. In the case where g (x0 )X = Y , the following three inequalities are pairwise equivalent to each other: Cγ (0 , σ γ ) > 0, Cγ (σ , ) > 0, and Cγ (m, g ) > 0. Therefore, the γ -necessity on always implies the inequality Cγ (0 , σ γ ) ≥ 0, and the γ -sufficiency on is always equivalent to the inequality Cγ (0 , σ γ ) > 0. Theorem 1.7 is the main result of the abstract theory of higher-order conditions for smooth problems. Note thatit remains valid if, in the definition of the set 0 , we replace the standard normalization αi + y ∗ = 1 by any equivalent normalization. Let us make more precise what we mean by an equivalent normalization. 1.1.5 Equivalent Normalizations Let ν(λ) be a positively homogeneous function of the first degree. A normalization ν(λ) = 1 is said to be equivalent to the standard normalization if the condition that 0 is nonempty implies the inequalities 0 < inf ν(λ) ≤ sup ν(λ) < +∞. 0 0 The following assertion holds. Proposition 1.8. Let g (x0 )X = Y . Then the condition normalization. k i=0 αi = 1 defines an equivalent Proof. Let 0 be nonempty, and let λ ∈ 0 . Then I αifi (x0 ) + y ∗ g (x0 ) = 0, which ∗ implies y g (x0 ) ≤ ( αi ) maxI fi (x0 ) . Therefore, αi + y ∗ g (x0 ) ≤ αi 1 + max fi (x0 ) . αi ≤ I It remains to note that y ∗ g (x0 ) | and y ∗ g (x0 )X = Y . are two equivalent norms on Y ∗ , since 14 Chapter 1. Abstract Scheme for Obtaining Higher-Order Conditions Therefore, in thecase where the Lyusternik condition g (x0 )X = Y holds, we can use the normalization αi = 1 in the definition of 0 . This normalization is called the Lyusternik normalization. We need to compare the functions 0 for the standard and Lyusternik normalizations. Therefore, in the case of Lyusternik normalization, let us agree L to equip the set 0 and the function 0 with the subscript L; i.e., we write L 0 and 0 . The following assertion holds. Proposition 1.9. Let g (x0 )X = Y . Then there exists a number a, 0 < a ≤ 1, such that L 0 ≤ max{aL 0 , 0 }, 1 L ≤ max , 0 0 . 0 a (1.4) (1.5) Proof. We first prove inequality (1.4). If 0 is empty, then 0 = −∞, and hence inequality (1.4) holds. Suppose that 0 is not empty. By Proposition 1.8, there exists a, 0 < a ≤ 1, such that for any λ ∈ 0 , the inequality a ≤ αi holds. Moreover, the condition αi + y ∗ = 1 implies αi ≤ 1. Let λ = (α, y ∗ ) ∈ 0 . We set ν := αi . Then λ̂ := λ/ν ∈ L 0 , a ≤ ν ≤ 1. Therefore, for any δx ∈ δ, L (λ, δx) = (ν λ̂, δx) = ν(λ̂, δx) ≤ max{aL 0 (δx), 0 (δx)}. This implies estimate (1.4). L Now let us prove (1.5). If L 0 is empty, then 0 = −∞, and hence (1.5) holds. ∗ ∗ = (α̂, ŷ ∗ ) ∈ L Now let L 0 be nonempty, and let λ̂ 0 . We set μ := 1 + ŷ , λ = (α, y ) := 1 1 λ̂/μ. Then λ ∈ 0 . Moreover, a ≤ αi = μ α̂i = μ ≤ 1, which implies 1 ≤ μ ≤ 1/a. Therefore, for any δx ∈ δ, we have 1 (λ̂, δx) = μ(λ, δx) ≤ max 0 (δx), 0 (δx) . a This implies estimate (1.5). Proposition 1.9 immediately implies the following assertion. Proposition 1.10. Let g (x0 )X = Y . Then there exists a number a, 0 < a ≤ 1, such that L Cγ (0 , σ γ ) ≤ max{aCγ (L 0 , σ γ ), Cγ (0 , σ γ )}, (1.6) 1 Cγ (L 0 , σ γ ) ≤ max{Cγ (0 , σ γ ), Cγ (0 , σ γ )}. a (1.7) Therefore, if the Lyusternik condition holds, constants Cγ (0 , σ γ ) and Cγ (L 0, σ γ ) have the same signs, which in this case allows us to replace the first constant by the second in Theorem 1.7. 1.2. Proof of the Main Theorem 1.1.6 15 Sufficient Conditions As was already noted, the inequalities Cγ (0 , σ γ ) ≥ 0 and Cγ (0 , σ γ ) > 0 are a pair of adjacent conditions for a minimum on at the point x0 . The nonstrict inequality is a necessary condition, and the strict inequality is a sufficient condition. This is implied by Theorem 1.7. The necessary condition is not trivial as we will verify below in proving Theorem 1.7. As for the sufficient condition Cγ (0 , σ γ ) > 0, its sufficiency for the minimum on is simple (this is characteristic for sufficient conditions in general: their sources are simple as a rule). Let us prove the sufficiency of this condition. The following estimate easily follows from the definitions of the functions 0 and σ : 0 ≤ σ . Hence Cγ (0 , σ γ ) ≤ Cγ (σ , σ γ ). (1.8) Cγ (σ , σ γ ) = Cγ (σ , ). (1.9) Let us show that Indeed, the inclusion σ γ ⊂ implies the inequality Cγ (σ , σ γ ) ≥ Cγ (σ , ). Let us prove the converse inequality Cγ (σ , σ γ ) ≤ Cγ (σ , ). (1.10) If Cγ (σ , ) = ∞, then inequality (1.10) holds. Let Cγ (σ , ) < ∞, and let a number C be such that Cγ (σ , ) < C, i.e., σ inf lim inf < C. γ Then there exists a sequence {δxn } ∈ + such that σ (δxn )/γ (δxn ) < C for all n. This implies {δxn } ∈ σ γ and Cγ (σ , σ γ ) := inf lim inf σ γ σ ≤ C. γ We have shown that the inequality Cγ (σ , ) < C always implies inequality Cγ (σ , σ γ ) ≤ C. This implies inequality (1.10) and, therefore, relation (1.9). From (1.8) and (1.9) we obtain Cγ (0 , σ γ ) ≤ Cγ (σ , ). (1.11) Thus, the inequality Cγ (0 , σ γ ) > 0 implies the inequality Cγ (σ , ) > 0, i.e., the γ sufficiency on . Therefore, the inequality Cγ (0 , σ γ ) > 0 is sufficient for the strict minimum on . The latter assertion contained in Theorem 1.7 turns out to be very simple. However, a complete proof of Theorem 1.7 requires considerably greater effort. Before passing directly to its proof, we present all necessary auxiliary assertions. The next section is devoted to this. 1.2 Proof of the Main Theorem We will need the main lemma for the proof of Theorem 1.7. In turn, the proof of the main lemma is based on the following three important properties used in the extremum theory: the compatibility condition of a set of linear inequalities and equations, the Hoffman lemma, and the Lyusternik theorem. These three properties compose the basis of the abstract 16 Chapter 1. Abstract Scheme for Obtaining Higher-Order Conditions theory of higher order for smooth problems, and for the reader’s convenience they are formulated in this section. 1.2.1 Basis of the Abstract Theory As above, let X and Y be Banach spaces, and let X ∗ and Y ∗ be their duals. Let a tuple on l = {l1 , . . . , lm }, li ∈ X∗ , i = 1, . . . , m, and a linear surjective operator A : X −→ Y be given. With the tuple l and the operator A we associate the set = (l, A) consisting of the tuples of multipliers λ = (α, y ∗ ), α = (α1 , . . . , αm ) ∈ Rm∗ , y ∗ ∈ Y ∗ , satisfying the conditions αi ≥ 0 (i = 1, . . . , m), m m αi = 1, i=1 αi li + y ∗ A = 0. i=1 The compatibility criterion of a set of linear inequalities and equations has the following form. Lemma 1.11. Let ξ = (ξ1 , . . . , ξm )∗ ∈ Rm , y ∈ Y . The set of conditions li , x + ξi < 0 (i = 1, . . . , m), is compatible iff sup λ∈ m Ax + y = 0 ∗ αi ξi + y , y < 0. i=1 (By definition, sup∅ = −∞.) Along with this lemma, in studying problems with inequality-type constraints, an important role is played by the estimate of the distance to the set of solutions of a set of linear inequalities and equations [41], which is presented below. Lemma 1.12 (Hoffman). There exists a constant C = C(l, A) with the following property: if for certain ξ ∈ Rm and y ∈ Y the system li , x + ξi ≤ 0 (i = 1, . . . , m), Ax + y = 0 is compatible, then there exists its solution x satisfying the estimate x ≤ C max{ξ1 , . . . , ξm , y }. In the case where there is no equation Ax + y = 0, Lemma 1.12 holds with the estimate x ≤ C max{ξ1+ , . . . , ξm+ }. Finally, we present the Lyusternik-type theorem on the estimate of the distance to the level of the equality operator in the form which is convenient for us (see [78, Theorem 2]). Let a set of sequences satisfy the same conditions as in Section 1.1. The following theorem holds for an operator g : X → Y strictly -differentiable at a point x0 . Theorem 1.13. Let g(x0 ) = 0 and let g (x0 )X = Y . Then there exists C > 0 such that for any {δxn } ∈ , there exists {x̄n } ∈ 0 satisfying the following conditions for all sufficiently large n: g(x0 + δxn + x̄n ) = 0, x̄n ≤ C g(x0 + δxn ) . 1.2. Proof of the Main Theorem 1.2.2 17 Main Lemma We now turn to problem (1.1). Let all assumptions of Section 1.1 hold. Let g (x0 )X = Y . In the definition of the set 0 , we choose the normalization αi = 1. According to Proposition 1.8, it is equivalent to the standard normalization. The following assertion holds. Lemma 1.14 (Main Lemma). Let a sequence {δxn } and a sequence of numbers {ζn } be such that δxn ∈ δ for all n, ζn+ → 0, and L 0 (δxn ) + ζn < 0 for all n. Then there exists a sequence {x̄n } ∈ 0 such that the following conditions hold : (1) x̄n ≤ O(σ (δxn ) + ζn+ ); (2) fi (x0 + δxn + x̄n ) + ζn ≤ o( x̄n ), i ∈ I ; (3) g(x0 + δxn + x̄n ) = 0 for all sufficiently large n. Proof. For an arbitrary n, let us consider the following set of conditions on x̄: fi (x0 ), x̄ + fi (x0 + δxn ) + ζn < 0, i ∈ I ; g (x0 )x̄ + g(x0 + δxn ) = 0. (1.12) Let (αi )i∈I and y ∗ be a tuple from the set of system (1.12). We set αi = 0 for i ∈ / I. L ∗ Then λ = (α0 , . . . , αk , y ∗ ) ∈ L 0 . The converse is also true: if λ = (α0 , . . . , αk , y ) ∈ 0 , then αi = 0 for i ∈ / I and the tuple ((αi )i∈I , y ∗ ) belong to the set of system (1.12). Therefore, ∗ max αi (fi (x0 + δxn ) + ζn ) + y , g(x0 + δxn ) I = = max L 0 k ∗ αi (fi (x0 + δxn ) + ζn ) + y , g(x0 + δxn ) i=0 L 0 (δxn ) + ζn < 0. The latter relation is implied by the definition of the function L 0 and the normalization αi = 1. According to Lemma 1.11, system (1.12) is compatible. Then by Hoffman’s lemma (Lemma 1.12), there exist C > 0 and a sequence {x̄n } such that for all n, fi (x0 ), x̄n + fi (x0 + δxn ) + ζn ≤ 0, i ∈ I ; g (x0 )x̄n + g(x0 + δxn ) = 0; x̄n ≤ C max max{fi (x0 + δxn ) + ζn }, g(x0 + δxn ) . i∈I (1.13) (1.14) (1.15) It follows from (1.15) that for all sufficiently large n, x̄n ≤ C(σ (δxn ) + ζn+ ) → 0, (1.16) / I . Therefore, {x̄n } ∈ 0 . Since g is strictly since fi (x0 + δxn ) → fi (x0 ) < 0 for i ∈ differentiable at the point x0 , condition (1.14) implies g(x0 + δxn + x̄n ) = o( x̄n ). Then by the Lyusternik theorem (see Theorem 1.13), there exists {x̄n } ∈ 0 such that for all sufficiently large n, we have g(x0 + δxn + x̄n + x̄n ) = 0, x̄n = o( x̄n ). (1.17) (1.18) 18 Chapter 1. Abstract Scheme for Obtaining Higher-Order Conditions We set {x̄n } = {x̄n + x̄n }. Condition (1.17) implies g(x0 + δxn + x̄n ) = 0 for all sufficiently large n, and conditions (1.16) and (1.18) imply x̄n ≤ O(σ (δxn ) + ζn+ ). Further, we obtain from conditions (1.13) and the property of strong -differentiability of the functionals fi at the point x0 that fi (x0 + δxn + x̄n ) + ζn = fi (x0 + δxn ) + fi (x0 ), x̄n + fi (x0 ), x̄n + ζn + o( x̄n ) ≤ fi (x0 ), x̄n + o( x̄n ) = o1 ( x̄n ), i ∈ I. The latter relation holds because of (1.18). The lemma is proved. Now, we prove a number of assertions from which the main result (Theorem 1.7) will follow. Below we assume that the order γ is strict and higher on , and all assumptions of Section 1.1 hold for the set of sequences and problem (1.1) at the point x0 . 1.2.3 Case Where the Lyusternik Condition Holds We have the following theorem. Theorem 1.15. Let g (x0 )X = Y . Then Cγ (L 0 , σ γ ) = Cγ (m, g ). Proof. We first show that Cγ (L 0 , σ γ ) ≤ Cγ (m, g ). Indeed, L Cγ (L 0 , σ γ ) ≤ Cγ (0 , σ γ ∩ g ) ≤ Cγ (m, σ γ ∩ g ) = Cγ (m, g ). Here, the first inequality is obvious, and the second inequality follows from the obvious estimate L 0 ≤ m | g . The equality is proved in the same way as relation (1.9). Now let us prove inequality Cγ (m, g ) ≤ Cγ (L 0 , σ γ ), which will finish the proof , ) = +∞, the inequality holds. Let Cγ (L of the theorem. If Cγ (L σγ 0 0 , σ γ ) < +∞, and let C be such that Cγ (L 0 , σ γ ) := inf lim inf σ γ L 0 < −C. γ L Then there exists a sequence {δxn } ∈ + σ γ at which 0 (δxn ) + Cγ (δxn ) < 0, and, moreover, δxn ∈ δ for all n. We set ζn = Cγ (δxn ). According to the main lemma, there exists a sequence {x̄n } such that the following conditions hold: (α) (β) (γ ) x̄n ≤ O(σ (δxn ) + C + γ (δxn )); fi (x0 + δxn + x̄n ) + Cγ (δxn ) ≤ o( x̄n ), i ∈ I ; g(x0 + δxn + x̄n ) = 0 for all sufficiently large n. Since {δxn } ∈ σ γ , the first condition implies x̄n ≤ O(γ (δxn )). We set {δxn } = {δxn + x̄n }. Then condition (γ ) implies {δxn } ∈ g , and condition (β) implies fi (x0 + δxn ) + Cγ (δxn ) ≤ o(γ (δxn )), From this we obtain lim inf m(δxn ) ≤ −C. γ (δxn ) i ∈ I. 1.2. Proof of the Main Theorem 19 Since γ is a higher order, we have γ (δxn ) = γ (δxn + x̄n ) = γ (δxn ) + o( x̄n ) = γ (δxn ) + o(γ (δxn )). Therefore, lim inf m(δxn ) ≤ −C. γ (δxn ) Taking into account that {δxn } ∈ g , we obtain from this that Cγ (m, g ) ≤ −C. Therefore, we have proved that the inequality Cγ (L 0 , σ γ ) < −C always implies the inequality Cγ (m, g ) ≤ −C. Therefore, Cγ (m, g ) ≤ Cγ (L 0 , σ γ ). The theorem is completely proved. Theorem 1.15 and Proposition 1.10 imply the following theorem. Theorem 1.16. Let g (x0 )X = Y . Then the following three inequalities are pairwise equivalent: Cγ (m, g ) ≥ 0, Cγ (L 0 , σ γ ) ≥ 0, and Cγ (0 , σ γ ) ≥ 0. Now, consider the sequence of relations Cγ (0 , σ γ ) ≤ Cγ (σ , ) ≤ Cγ (σ , g ) = Cγ (m+ , g ) = Cγ (m, g )+ . (1.19) The first of these relations was proved in Section 1.1 (inequality (1.11)), and the other relations are obvious. The following assertion follows from (1.19), Theorem 1.15, and inequality (1.7). Corollary 1.17. Let g (x0 )X = Y . Then for 0 < a ≤ 1 the following inequalities hold : + Cγ (0 , σ γ ) ≤ Cγ (σ , ) ≤ Cγ (m, g )+ ≤ Cγ (L 0 , σ γ ) ≤ 1 Cγ (0 , σ γ )+ . a This implies the following theorem. Theorem 1.18. Let g (x0 )X = Y . Then the following four inequalities are pairwise equivalent: Cγ (σ , ) > 0, Cγ (m, g ) > 0, Cγ (0 , σ γ ) > 0, and Cγ (L 0 , σ γ ) > 0. Therefore, in the case where the Lyusternik condition holds, we have proved all the assertions of Theorem 1.7. 1.2.4 Case Where the Lyusternik Condition Is Violated To complete the proof of Theorem 1.7, we need to prove the following: if g (x0 )X = Y , then (a) 0 ≥ 0, and (b) the inequality Cγ (σ , ) > 0 is equivalent to the inequality Cγ (0 , σ γ ) > 0. We begin with the proof of (a). Proposition 1.19. If g (x0 )X = Y , then 0 (δx) ≥ 0 for all δx ∈ δ. 20 Chapter 1. Abstract Scheme for Obtaining Higher-Order Conditions Proof. Since the image Y1 := g (x0 )X is closed in Y , the condition Y1 = Y implies the existence of y ∗ ∈ Y , y ∗ = 1 such that y ∗ , y = 0 for all y ∈ Y1 , and hence λ = (0, y ∗ ) ∈ 0 and λ = (0, −y ∗ ) ∈ 0 . From this we obtain that for any δx ∈ δ, max (λ, δx) ≥ max{(λ , δx), (λ , δx)} = |y ∗ , g(x0 + δx) | ≥ 0. 0 The proposition is proved. Now let us prove assertion (b). Since the inequality Cγ (0 , σ γ ) ≤ Cγ (σ , ) always holds by (1.11), in order to prove (b), we need to prove the following lemma. Lemma 1.20. Let g (x0 )X = Y . Then there exists a constant b = b(g (x0 )) > 0 such that Cγ (σ , ) ≤ b Cγ (0 , σ γ )+ . (1.20) Proof. The proof uses a special method for passing from the system S = {f0 , . . . , fk , g} to a certain auxiliary system Ŝ. We set Y1 = g (x0 )X. According to the definition of the smooth problem, Y = Y1 ⊕ Y2 , where Y2 is a closed subspace in Y . Then Y ∗ = W1 ⊕ W2 , where W1 and W2 are such that any functional from W1 is annihilated on Y2 , and any functional from W2 is annihilated on Y1 . Without loss of generality, we assume that if y = y1 + y2 , y1 ∈ Y1 , and y2 ∈ Y2 , then y = max{ y1 , y2 }. Then for y ∗ = y1∗ + y2∗ , y1∗ ∈ W1 , y2∗ ∈ W2 , we have y ∗ = y1∗ + y2∗ . Let P1 : Y → Y1 and P2 : Y → Y2 be projections compatible with the decomposition of Y into a direct sum Y = Y1 ⊕ Y2 . Then P1 + P2 = I , P1 P2 = 0, and P2 P1 = 0. We set g1 = P1 g and g2 = P2 g. Then g = g1 + g2 , g1 (x0 )X = Y1 , and g2 (x0 )X = {0}. Introduce the functional fg (x) = g2 (x) . The condition g2 (x0 )X = {0} implies that fg is strictly -differentiable at the point x0 and fg (x0 ) = 0. Consider the system Ŝ consisting of the functionals f0 , . . . , fk , fg and the operator g1 . All subjects related to this system will be endowed with the sign ∧. Since g = g1 + g2 , we have g = max{ g1 , g2 }. Therefore, σ (δx) := max{f0 (x0 + δx), . . . , fk (x + δx), g(x0 + δx) } = max{f0 (x0 + δx), . . . , fk (x0 + δx), fg (x0 + δx), g1 (x0 + δx) } =: σ̂ (δx). This implies Cγ (σ , ) = Cγ (σ̂ , ). (1.21) Further, since the Lyusternik condition g1 (x0 )X = Y1 holds for the system Ŝ, by Corollary 1.17, there exists â > 0 such that Cγ (σ̂ , ) ≤ 1 ˆ 0 , σ̂ γ )+ . Cγ ( â (1.22) ˆ 0 is empty, then ˆ 0 = −∞, and hence the inequality ˆ 0 ≤ 0 . If Now let us show that ˆ holds. Let 0 be nonempty, and let λ̂ = (α0 , . . . , αk , αg , y1∗ ) be an arbitrary element of the ˆ 0 . Then set αi fi (x0 ) = 0 (i = 1, . . . k), αg ≥ 0; y1∗ ∈ W1 , αi ≥ 0 (i = 0, . . . , k), k k αi + αg + y1∗ = 1, αi fi (x0 ) + αg fg (x0 ) + y1∗ g1 (x0 ) = 0. i=0 i=0 1.3. Simple Applications of the Abstract Scheme 21 Moreover, fg (x0 ) = 0. Let δx ∈ δ be an arbitrary element. Choose y2∗ ∈ W2 so that the following conditions hold: y2∗ = αg , y2∗ , g2 (x0 + δx) = αg g2 (x0 + δx) . We set y ∗ = y1∗ + y2∗ , and λ = (α0 , . . . , αk , y ∗ ). As is easily seen, then we have λ ∈ 0 and ˆ λ̂, δx) = (λ, δx). Therefore, for arbitrary δx and λ̂ ∈ ˆ 0 , there exists λ ∈ 0 such that ( ˆ the indicated relation holds. This implies 0 (δx) ≤ 0 (δx). Also, taking into account that σ̂ = σ , we obtain ˆ 0 , σ̂ γ ) ≤ Cγ (0 , σ γ ). Cγ ( (1.23) It follows from (1.21)–(1.23) that Cγ (σ , ) ≤ 1 Cγ (0 , σ γ )+ . â (1.24) It remains to set b = 1/â. The lemma is proved. Therefore, we have shown that in the case of violation of the Lyusternik condition, the inequalities Cγ (σ , ) > 0 and Cγ (0 , σ γ ) > 0 are equivalent. Thus, we have completed the proof of Theorem 1.7. 1.3 Simple Applications of the Abstract Scheme In this section, following [55], we shall obtain quadratic conditions in a smooth problem in Rn and in the problem of Bliss with endpoint inequalities. 1.3.1 A Smooth Problem in Rn Let X = Rn , Y = Rm , and = X. Consider the problem J (x) → min; fi (x) ≤ 0 (i = 1, . . . , k), g(x) = 0. (1.25) We assume that the functions J : Rn → R1 , fi : Rn → R1 , i = 1, . . . , k, and the operator g : Rn → Rm are twice differentiable at each point. Let a point x0 satisfy the constraints, and let us study its optimality. We define f0 and I as in relation (1.2). Let = 0 := {{δxn } | δxn → 0 (n → ∞)}. Obviously, (1.25) is a smooth problem on 0 at the point x0 , and the minimum on 0 is a local minimum. For an order we take a unique (up to a nonsingular transformation) quadratic positive definite functional γ (δx) = δx, δx in Rn . Obviously, γ is a strict higher order (see the definition in Section 1.1.3). Let us define the violation function σ as in Section 1.1.1. Denote by σ γ the set of sequences {δxn } ∈ satisfying the condition σ (δxn ) ≤ O(γ (δxn )). Define the following necessary condition for a local minimum at a point x0 . Condition ℵγ : g (x0 )X = Y or γ -necessity holds. By results in Section 1.1.4, the inequality Cγ (0 , σ γ ) ≥ 0 is a necessary condition for a local minimum and is equivalent to Condition ℵγ , and Cγ (0 , σ γ ) > 0 is a sufficient 22 Chapter 1. Abstract Scheme for Obtaining Higher-Order Conditions condition for a local minimum and is equivalent to γ -sufficiency. Below we transform the expression for 0 (δxn ) Cγ (0 , σ γ ) := inf lim inf {δxn }∈σ γ n→∞ γ (δxn ) into an equivalent simpler form. We recall that (λ, δx) := k αi fi (x0 + δx) + y ∗ , g(x0 + δx) = L(λ, x0 + δx), i=0 0 (δx) := max (λ, δx), λ∈0 where the set 0 is defined by (1.3). We set 1 Lxx (λ, x0 )x̄, x̄ , 0 (x̄) := max λ (x̄), λ∈0 2 K := {x̄ ∈ X | fi (x0 ), x̄ ≤ 0, i ∈ I ; g (x0 )x̄ = 0}, σ (x̄) := fi (x0 ), x̄ + + g (x0 )x̄ , λ (x̄) := Cγ (0 , K) := i∈I inf {0 (x̄) | x̄ ∈ K, γ (x̄) = 1}. If K = {0}, then as usual we set Cγ (0 , K) = +∞. Obviously, K = {x̄ ∈ X | σ (x̄) = 0}. We call K the critical cone. Theorem 1.21. The following equality holds: Cγ (0 , σ γ ) = Cγ (0 , K). Proof. Evidently, 0 (δx) = 0 (δx) + o(γ (δx)) as δx → 0, σ (δx) = σ (δx) + O(γ (δx)) as δx → 0. We set σ γ (1.26) (1.27) := {δxn } ∈ 0 | σ (δx) ≤ O(γ (δx)) . It follows from (1.27) that σ γ = σ γ ; hence, by taking into account (1.26), we immediately obtain Cγ (0 , σ γ ) = Cγ (0 , σ γ ), where Cγ (0 , σ γ ) := inf {δxn }∈σ lim inf γ n→∞ 0 (δxn ) . γ (δxn ) Then we set σ := {δxn } ∈ 0 | σ (δx) = 0 , Cγ (0 , σ ) := inf {δxn }∈σ lim inf n→∞ 0 (δxn ) . γ (δxn ) Since, obviously, σ ⊂ σ γ , we have Cγ (0 , σ ) ≥ Cγ (0 , σ γ ). We show that, in fact, equality holds. Suppose that Cγ (0 , σ γ ) = +∞. Take any ε > 0. Let {δxn } ∈ σ γ be a nonvanishing sequence such that lim n→∞ 0 (δxn ) ≤ Cγ (0 , σ γ ) + ε. γ (δxn ) 1.3. Simple Applications of the Abstract Scheme 23 By applying the Hoffman lemma (see Lemma 1.12) for each δxn (n = 1, 2, . . . ) to the system f (x0 ), δxn + x̄ ≤ 0, i ∈ I , g (x0 )(δxn + x̄) = 0, regarded as a system in the unknown x̄, and by bearing in mind that σ (δxn ) ≤ O(γ (δxn )), we obtain the following assertion: we can find an {x̄n } such that σ (δxn + x̄n ) = 0 and x̄n ≤ O(γ (δxn )). Consequently, lim n→∞ 0 (δxn + x̄n ) 0 (δxn ) = lim ≤ Cγ (0 , σ γ ) + ε, n→∞ γ (δxn ) γ (δxn + x̄n ) and {δxn + x̄n } ∈ σ . This implies that Cγ (0 , σ ) ≤ Cγ (0 , σ γ ). Consequently, the equality Cγ (0 , σ ) = Cγ (0 , σ γ ) holds, from which we also obtain Cγ (0 , σ γ ) = Cγ (0 , σ ). But since 0 and γ are positively homogeneous of degree 2, by applying the definition of the cone K, in an obvious way we obtain that, in turn, Cγ (0 , σ ) = Cγ (0 , K). Thus, Cγ (0 , σ γ ) = Cγ (0 , K), and the theorem is proved. Corollary 1.22. The condition Cγ (0 , K) ≥ 0 is equivalent to Condition ℵγ and so is necessary for a local minimum. The condition Cγ (0 , K) > 0 is equivalent to γ -sufficiency and so is sufficient for a strict local minimum. In particular, K = {0} is sufficient for a strict local minimum. It is obvious, that Cγ (0 , K) ≥ 0 is equivalent to the condition 0 ≥ 0 on K, and since K is finite-dimensional, Cγ (0 , K) > 0 is equivalent to 0 > 0 on K \ {0}. We also remark that these conditions are stated by means of the maximum of the quadratic forms, and they cannot be reduced in an equivalent way to a condition on one of these forms. Here is a relevant example. Example 1.23 (Milyutin). Let X = R2 , and let ϕ and ρ be polar coordinates in R2 . Let the four quadratic forms Qi , i = 1, 2, 3, 4, be defined by their traces qi on a circle of unit radius: q1 (ϕ) = sin 2ϕ − ε, q2 (ϕ) = − sin 2ϕ − ε, q3 (ϕ) = cos 2ϕ − ε, q4 (ϕ) = − cos 2ϕ − ε. We choose the constant ε > 0 so that max1≤i≤4 qi (ϕ) > 0, 0 ≤ ϕ < 2π . We consider the system S formed from the functionals fi (x) = Qi (x), i = 1, 2, 3, 4, in a neighborhood of x0 = 0. For γ (x) = x, x this system has γ -sufficiency. In fact, since fi (0) = 0, i = 1, 2, 3, 4, we have K = R2 , 0 = {α ∈ R4 | αi ≥ 0, αi = 1} and 0 (x) = max1≤i≤4 Qi (x), and so 0 (x) ≥ εγ (x) for all x ∈ X. But no form 4i=1 αi Qi , where αi ≥ 0 and αi = 1, is nonnegative, since its trace q on a circle of unit radius has the form q(ϕ) = A sin 2ϕ + B cos 2ϕ − ε. 1.3.2 The Problem of Bliss with Endpoint Inequalities We consider the following problem. It is required to minimize the function of the initial and final states (1.28) J (x(t0 ), x(tf )) → min under the constraints Fi (x(t0 ), x(tf )) ≤ 0 (i = 1, . . . , k), K(x(t0 ), x(tf )) = 0, ẋ = f (t, x, u), (1.29) 24 Chapter 1. Abstract Scheme for Obtaining Higher-Order Conditions where x ∈ Rn , u ∈ Rr , J ∈ R, Fi ∈ R, K ∈ Rs , f ∈ Rn , and the interval [t0 , tf ] is fixed. Strictly speaking, the problem of Bliss [6] includes the local equation g(t, x, u) = 0, where g ∈ Rq , and it is traditional to require that the matrix gu has maximal rank at points (t, x, u) such that g(t, x, u) = 0. This statement will be considered later. Here, for simplicity, we consider the problem without the local equation. We set W = W 1,1 ([t0 , tf ], Rn ) × L∞ ([t0 , tf ], Rr ), where W 1,1 ([t0 , tf ], Rn ) is the space of n-dimensional absolutely continuous functions, and L∞ ([t0 , tf ], Rr ) is the space of r-dimensional bounded measurable functions. We consider problem (1.28)–(1.29) in W . We denote the pair (x, u) by w, and we define the norm in W by w = x 1,1 + u ∞ = |x(t0 )| + tf ẋ(t) dt + ess sup |u(t)|. t∈[t0 ,tf ] t0 Clearly, a local minimum in this space is weak. This is the minimum on the set of sequences 0 := {{δwn } | δwn → 0 (n → ∞)}. Again we set = 0 . We denote the argument of the functions J , Fi , and K by p = (x0 , xf ), where x0 ∈ Rn and xf ∈ Rn . All relations containing measurable sets and functions are understood with accuracy up to a set of measure zero. We assume the following. Assumption 1.24. The functions J , Fi , and K are twice continuously differentiable with respect to p; the function f is twice differentiable with respect to w; the function f and its second derivative fww are uniformly bounded and equicontinuous with respect to w on any bounded set of values (t, w) and are measurable in t for any fixed w. Evidently, f satisfies these conditions if f and fww are continuous jointly in both variables. Let w 0 (·) ∈ W be a trajectory satisfying all constraints that is being investigated for an optimal situation. We set p 0 = (x 0 (t0 ), x 0 (tf )), F0 (p) = J (p) − J (p 0 ), I = {i ∈ {0, 1, . . . , k} | Fi (p0 ) = 0}. Obviously, problem (1.28)–(1.29) is smooth on 0 at the point w0 . The set 0 consists of aggregates λ = (α, β, ψ) for which the local form of Pontryagin’s minimum principle holds: α ∈ R(k+1)∗ , α ≥ 0, β ∈ Rs∗ , ψ(·) ∈ W 1,1 ([t0 , tf ], Rn∗ ), αi Fi (p0 ) = 0 (i = 1, . . . , k), k αi + i=0 ψ̇ = −Hx (t, w0 , ψ), ψ(t0 ) = −lx0 (p0 , α, β), Hu (t, w 0 , ψ) = 0, s (1.30) |βj | = 1, (1.31) j =1 ψ(tf ) = lxf (p0 , α, β), (1.32) (1.33) where H = ψf , l = αF + βK. (1.34) The notation Rn∗ stands for the space of n-dimensional row vectors. We emphasize the dependence of the Pontryagin function H and the endpoint Lagrange function l on λ that is defined by (1.34) by writing H = H λ (t, w) and l = l λ (p). Under our assumptions 0 is a finite-dimensional compact set, each point of which is uniquely determined by its projection (α, β). 1.3. Simple Applications of the Abstract Scheme 25 For any δw ∈ W and λ ∈ 0 the Lagrange function has the form t (λ, δw) = l λ (p0 + δp) + t0f H λ (t, w0 + δw) − H λ (t, w0 ) dt − tf t0 ψδ ẋ dt, (1.35) where δp = (δx(t0 ), δx(tf )). For arbitrary w̄ ∈ W and λ ∈ 0 we set 1 λ 0 1 ωλ (w̄) := lpp (p )p̄, p̄ + 2 2 tf t0 λ Hww (t, w0 )w̄, w̄ dt, ω0 (w̄) = max ωλ (w̄), (1.36) λ∈0 where p̄ = (x̄(t0 ), x̄(tf )). Set γ (δw) = δx(t0 ), δx(t0 ) + tf tf δu(t), δu(t) dt. (1.37) Obviously, γ is a strict higher order. We define the cone of critical variations K = w̄ ∈ W | Fi (p0 )p̄ ≤ 0, i ∈ I , K (p0 )p̄ = 0, x̄˙ = fw (t, w0 )w̄ (1.38) t0 δ ẋ(t), δ ẋ(t) dt + t0 and the constant Cγ (ω0 , K) = inf {ω0 (w̄) | w̄ ∈ K, γ (w̄) = 1} . (1.39) (Let us note that the sign of the constant Cγ (ω0 , K) will not change if we replace, in its t definition, the functional γ with the functional γ̄ (w̄) = x̄(t0 ), x̄(t0 ) + t0f ū(t), ū(t) dt.) We define Condition ℵγ as in Section 1.3.1. Theorem 1.25. The condition Cγ (ω0 , K) ≥ 0 is equivalent to the condition ℵγ and so is necessary for a local minimum; the condition Cγ (ω0 , K) > 0 is equivalent to γ -sufficiency and so is sufficient for a strict local minimum. This result among others, is given, in [84]. But it is not difficult to derive it directly by following the scheme indicated in our discussion of the finite-dimensional case. Because the problem is smooth, Theorem 1.25 follows from Cγ (0 , σ γ ) = Cγ (ω0 , K), which is established in the same way as in Section 1.3.1. Notes on SSC for abstract optimization problems. Maurer and Zowe [75] considered optimization problems in Banach spaces with fully infinite-dimensional equality and inequality constraints defined by cone constraints and derived SSC for quadratic functionals γ . Maurer [62] showed that the SSC in [75] can be applied to optimal control problems by taking into account the so-called “two-norm discrepancy.” Chapter 2 Quadratic Conditions in the General Problem of the Calculus of Variations In this chapter, on the basis of the results of Chapter 1, we create the quadratic theory of conditions for a Pontryagin minimum in the general problem of the calculus of variations without local mixed constraint g(t, x, u) = 0. Following [92], we perform the decoding of the basic constant for the set of the so-called “Pontryagin sequences” and a special higher order γ , which is characteristic for extremals with finitely many discontinuities of the first kind of control. In Section 2.1, we formulate both necessary and sufficient quadratic conditions for a Pontryagin minimum, which will be obtained as a result of the decoding. In Sections 2.2 and 2.3, we perform some preparations for the decoding. In Section 2.4, we estimate the basic constant from above, thus obtaining quadratic necessary conditions for Pontryagin minimum, and in Section 2.5, we estimate it from below, thus obtaining sufficient conditions. In Section 2.7, we establish the relation of the obtained sufficient conditions for Pontryagin minimum with conditions for strong minimum. 2.1 2.1.1 Statements of Quadratic Conditions for a Pontryagin Minimum Statement of the Problem and Assumptions In this chapter, we consider the following general problem of the calculus of variations on a fixed time interval := [t0 , tf ]: J (x(t0 ), x(tf )) → min, F (x(t0 ), x(tf )) ≤ 0, K(x(t0 ), x(tf )) = 0, ẋ = f (t, x, u), (x(t0 ), x(tf )) ∈ P , (t, x, u) ∈ Q, (2.1) (2.2) (2.3) (2.4) where P ⊂ R2d(x) and Q ⊂ R1+d(x)+d(u) are open sets. By d(a) we denote the dimension of vector a. Problem (2.1)–(2.4) also will be called the canonical problem. For the sake of brevity, we set x(t0 ) = x0 , x(tf ) = xf , (x0 , xf ) = p, 27 (x, u) = w. 28 Chapter 2. Quadratic Conditions in the Calculus of Variations We seek the minimum among pairs of functions w = (x, u) such that x(t) is an absolutely continuous function on = [t0 , tf ] and u(t) is a bounded measurable function on . Recall that by W 1,1 (, Rd(x) ) we denote the space of absolutely continuous functions x : → t Rd(x) , endowed with the norm x 1,1 := |x(t0 )| + t0f |ẋ(t)| dt, and L∞ (, Rd(u) ) denotes the space of bounded measurable functions u : → Rd(u) , endowed with the norm u ∞ := ess supt∈[t0 ,tf ] |u(t)|. We set W = W 1,1 (, Rd(x) ) × L∞ (, Rd(u) ). We define the norm in the space W as the sum of the norms in the spaces W 1,1 (, Rd(x) ) and L∞ (, Rd(u) ): w = x 1,1 + u ∞ . The space W with this norm is a Banach space. Therefore, we seek the minimum in the space W . A pair w = (x, u) is said to be admissible, if w ∈ W and constraints (2.2)–(2.4) are satisfied by w. We assume that the functions J (p), F (p), and K(p) are defined and twice continuously differentiable on the open set P , and the function f (t, w) is defined and twice continuously differentiable on the open set Q. These are the assumptions on the functions of the problem. Before formulating the assumptions on the point w 0 ∈ W being studied, we give the following definition. Definition 2.1. We say that t∗ ∈ (t0 , tf ) is an L-point (or Lipschitz point) of a function ϕ : [t0 , tf ] → Rn if at t∗ , there exist the left and right limit values lim ϕ(t) = ϕ(t∗ −), t→t∗ t<t∗ lim ϕ(t) = ϕ(t∗ +) t→t∗ t>t∗ and there exist L > 0 and ε > 0 such that |ϕ(t) − ϕ(t∗ −)| ≤ L|t − t∗ | |ϕ(t) − ϕ(t∗ +)| ≤ L|t − t∗ | ∀ t ∈ (t∗ − ε, t∗ ) ∩ [t0 , tf ], ∀ t ∈ (t∗ , t∗ + ε) ∩ [t0 , tf ]. A point of discontinuity of the first kind that is an L-point will be called a point of L-discontinuity. Let w0 = (x 0 , u0 ) be a pair satisfying the constraints of the problem whose optimality is studied. We assume that the control u0 (· ) is piecewise continuous. Denote by = {t1 , . . . , ts } the set of points of discontinuity for the control u0 , where t0 < t1 < · · · < ts < tf . We assume that is nonempty (in the case where is empty, all the results remain valid and are obviously simplified). We assume that each tk ∈ is a point of L-discontinuity. By u0k− = u0 (tk −) and u0k+ = u0 (tk +) we denote the left and right limit values of the function u0 (t) at the point tk ∈ , respectively. For a piecewise continuous function u0 (t), the condition (t, x 0 , u0 ) ∈ Q means that (t, x 0 (t), u0 (t)) ∈ Q for all t ∈ [t0 , tf ]\. We also assume that (tk , x 0 (tk ), u0k− ) ∈ Q and (tk , x 0 (tk ), u0k+ ) ∈ Q for all tk ∈ . As above, all relations and conditions involving measurable functions are assumed to be valid with accuracy up to a set of zero measure even if this is not specified. 2.1.2 Minimum on the Set of Sequences Let S be an arbitrary set of sequences {δwn } in the space W invariant with respect to the operation of passing to a subsequence. According to Definition 1.1, w0 is a minimum point 2.1. Statements of Quadratic Conditions for a Pontryagin Minimum 29 on S if there is no sequence {δwn } ∈ S such that the following conditions hold for all its members: J (p 0 + δpn ) < J (p0 ), F (p0 + δpn ) ≤ 0, K(p 0 + δpn ) = 0, ẋ 0 + δ ẋn = f (t, w0 + δwn ), (p0 + δpn ) ∈ P , (t, w0 + δwn ) ∈ Q, where p 0 = (x 0 (t0 ), x 0 (tf )), δwn = (δxn , δun ), and δpn = (δxn (t0 ), δxn (tf )). In a similar way, we define the strict minimum on S: it is necessary to only replace the strict inequality J (p0 + δpn ) < J (p 0 ) by nonstrict in the previous definition and additionally assume that the sequence {δwn } contains only nonzero members. We can define any local (in the sense of a certain topology) minimum as a minimum on the corresponding set of sequences. For example, a weak minimum is a minimum on the set of sequences {δwn } in W such that δxn C + δun ∞ → 0, where x C = maxt∈[t0 ,tf ] |x(t)| is the norm in the space of continuous functions. Let 0 be the set of sequences {δwn } in W such that δwn = δxn 1,1 + δun ∞ → 0. (We note that 0 corresponds to the set of sequences 0 introduced in Section 1.1.) It is easy to see that a minimum on 0 is also a weak minimum. Therefore, we can define the same type of minimum using various sets of sequences. We often use this property when choosing a set of sequences for the type of minimum considered which is most convenient for studying. In particular, this refers to the definition of Pontryagin minimum studied in this chapter. 2.1.3 Pontryagin Minimum t Let u 1 := t0f |u(t)| dt be the norm in the space L1 (, Rd(u) ) of functions u : [t0 , tf ] → Rd(u) Lebesque integrable with first degree. Denote by the set of sequences {δwn } in W satisfying the following two conditions: (a) δxn 1,1 + δun 1 → 0; (b) there exists a compact set C ⊂ Q (for each sequence) such that, starting from a certain number, the following condition holds: (t, w 0 (t) + δwn (t)) ∈ C a.e. on [t0 , tf ]. A minimum on is called a Pontryagin minimum. For convenience, let us formulate an equivalent definition of the Pontryagin minimum. The pair w 0 = (x 0 , u0 ) is a point of Pontryagin minimum iff for each compact set C ⊂ Q there exists ε > 0 such that J (p) ≥ J (p 0 ) (where p 0 = (x 0 (t0 ), x 0 (tf )), p = (x(t0 ), x(tf ))) for all admissible pairs w = (x, u) such that (a) maxt∈ |x(t) − x 0 (t)| < ε, (b) |u(t) − u0 (t)| dt < ε, (c) (t, x(t), u(t)) ∈ C a.e. on [t0 , tf ]. We can show that it is impossible to define a Pontryagin minimum as a local minimum with respect to a certain topology. Therefore the concept of minimum on a set of sequences is more general than the concept of local minimum. Since ⊃ 0 , a Pontryagin minimum implies a weak minimum. 2.1.4 Pontryagin Minimum Principle Define two sets 0 and M0 of tuples of Lagrange multipliers. They are related to the firstorder necessary conditions for the weak and Pontryagin minimum, respectively. We set 30 Chapter 2. Quadratic Conditions in the Calculus of Variations l = α0 J + αF + βK, H = ψf , where α0 is a number and α, β, and ψ are row vectors of the same dimension as F , K, and f , respectively (note that x, u, w, F , K, and f are column vectors). Denote by (Rn )∗ the space of row vectors of the dimension n. The functions l and H depend on the following variables: l = l(p, α0 , α, β), H = H (t, w, ψ). Denote by λ an arbitrary tuple (α0 , α, β, ψ(· )) such that α0 ∈ R1 , α ∈ (Rd(F ) )∗ , β ∈ (Rd(K) )∗ , ψ(· ) ∈ W 1,∞ (, (Rd(x) )∗ ), where W 1,∞ (, (Rd(x) )∗ ) is the space of Lipschitz continuous functions mapping [t0 , tf ] into (Rd(x) )∗ . For arbitrary λ, w, and p, we set l λ (p) = l(p, α0 , α, β), H λ (t, w) = H (t, w, ψ(t)). We introduce an analogous notation for partial derivatives (except for the derivative with ∂l respect to t) lxλ0 = ∂x (p, α0 , α, β), Hxλ (t, w) = ∂H ∂x (t, w, ψ(t)), etc. Denote by 0 the set of 0 tuples λ such that α0 ≥ 0, α ≥ 0, αF (p ) = 0, 0 α0 + d(F ) i=1 ψ̇ = −Hxλ (t, w 0 ), ψ(t0 ) = −lxλ0 (p0 ), αi + d(K) |βj | = 1, (2.5) j =1 ψ(tf ) = lxλf (p 0 ), Huλ (t, w0 ) = 0. (2.6) Here, αi are components of the row vector α and βj are components of the row vector β. If a point w0 yields a weak minimum, then 0 is nonempty. This was shown in [79, Part 1]. We set U(t, x) = {u ∈ Rd(u) | (t, x, u) ∈ Q}. Denote by M0 the set of tuples λ ∈ 0 such that for all t ∈ [t0 , tf ]\, the condition u ∈ U(t, x 0 (t)) implies the inequality H (t, x 0 (t), u, ψ(t)) ≥ H (t, x 0 (t), u0 (t), ψ(t)). (2.7) If w 0 is a point of Pontryagin minimum, then M0 is nonempty; i.e., the Pontryagin minimum principle holds. This also was shown in [79, Part 1]. The sets 0 and M0 are finite-dimensional compact sets, and, moreover, the projection λ → (α0 , α, β) is injective on the largest set 0 and, therefore, on M0 . Denote by co 0 the convex hull of the set 0 , and let M0co be the set of all λ ∈ co 0 such that for all t ∈ [t0 , tf ]\, the condition u ∈ U(t, x 0 (t)) implies inequality (2.7). We now formulate a quadratic necessary condition for the Pontryagin minimum. For this purpose, along with the set M0 , we need to define a critical cone K and a quadratic form on it. 2.1.5 Critical Cone Denote by P W 1,2 (, Rd(x) ) the space of piecewise continuous functions x̄(t) : [t0 , tf ] → Rd(x) absolutely continuous on each of the intervals of the set (t0 , tf )\ such that their first derivative is square Lebesgue integrable. We note that all points of discontinuity of functions in P W 1,2 (, Rd(x) ) are contained in . Below, for tk ∈ and x̄ ∈ P W 1,2 (, Rd(x) ), we set x̄ k− = x̄(tk −), x̄ k+ = x̄(tk +), and [x̄]k = x̄ k+ − x̄ k− . Let Z2 () be the space of triples z̄ = (ξ̄ , x̄, ū) such that ξ̄ = (ξ̄1 , . . . , ξ̄s ) ∈ Rs , x̄ ∈ P W 1,2 (, Rd(x) ), ū ∈ L2 (, Rd(u) ), 2.1. Statements of Quadratic Conditions for a Pontryagin Minimum 31 where L2 (, Rd(u) ) is the space of Lebesgue square integrable functions ū(t) : [t0 , tf ] → Rd(u) . Let IF (w 0 ) = {i ∈ {1, . . . , d(F )} | Fi (p0 ) = 0} be the set of active subscripts of the constraints Fi (p) ≤ 0 at the point w0 . Denote by K the set of z̄ = (ξ̄ , x̄, ū) ∈ Z2 () such that Jp (p 0 )p̄ ≤ 0, Fip (p0 )p̄ ≤ 0 ∀ i ∈ IF (w 0 ), x̄˙ = fw (t, w0 )w̄, [x̄]k = [ẋ 0 ]k ξ̄k ∀ tk ∈ , Kp (p0 )p̄ = 0, (2.8) (2.9) where p̄ = (x̄(t0 ), x̄(tf )), w̄ = (x̄, ū), and [ẋ 0 ]k is the jump of the function ẋ 0 (t) at the point tk , i.e., [ẋ 0 ]k = ẋ 0k+ − ẋ 0k− = ẋ 0 (tk +) − ẋ 0 (tk −). Clearly, K is a convex polyhedral cone. It will be called the critical cone. The following question is of interest: which inequalities in the definition of K can be replaced by equalities not changing K? An answer to this question follows from the next proposition. For λ = (α0 , α, β, ψ) ∈ 0 , denote by [H λ ]k the jump of the function H (t, x 0 (t), u0 (t), ψ(t)) at the point tk ∈ , i.e., [H λ ]k = H λk+ − H λk− , where H λk+ = H (tk , x 0 (tk ), u0k+ , ψ(tk )), H λk− = H (tk , x 0 (tk ), u0k− , ψ(tk )). λ k Let 0 be the subset of λ ∈ 0 such that [H ] = 0 for all tk ∈ . Then 0 is a finite dimensional compact set, and, moreover, M0 ⊂ 0 ⊂ 0 . Proposition 2.2. The following conditions hold for any λ ∈ 0 and any z̄ ∈ K: α0 Jp (p0 )p̄ = 0, αi Fip (p0 )p̄ = 0 ∀ i ∈ IF (w 0 ). (2.10) Also, the following question is of interest: in which case can one of the inequalities in the definition of K be omitted not changing K? For example, in which case can we omit the inequality Jp (p0 )p̄ ≤ 0? Proposition 2.3. If there exist λ ∈ 0 such that α0 > 0, then the conditions Fip (p0 )p̄ ≤ 0, αi Fi (p0 )p̄ = 0 ∀ i ∈ IF (w 0 ), x̄˙ = fw (t, w0 )w̄, [x̄]k = [ẋ 0 ]k ξ¯k ∀ tk ∈ Kp (p 0 )p̄ = 0, (2.11) (2.12) imply Jp (p0 )p̄ = 0, i.e., conditions (2.11) and (2.12) determine K as before. An analogous assertion holds for any other inequality Fip (p 0 )p̄ ≤ 0, i ∈ IF (w 0 ), in the definition of K. 32 2.1.6 Chapter 2. Quadratic Conditions in the Calculus of Variations Quadratic Form For λ ∈ 0 , tk ∈ , we set (k H λ )(t) := H (t, x 0 (t), u0k+ , ψ(t)) − H (t, x 0 (t), u0k− , ψ(t)). The following assertion holds: for any λ ∈ 0 , tk ∈ , the function (k H λ )(t) has the derivative at the point tk . We set1 D k (H λ ) = − d (k H λ )(t)|t=tk . dt The quantity D k (H λ ) can be calculated by the formula D k (H λ ) = −Hxλk+ Hψk− + Hxλk− Hψk+ − [Htλ ]k , where Hxλk+ = ψ(tk )fx (tk , x 0 (tk ), u0k+ ), Hxλk− = ψ(tk )fx (tk , x 0 (tk ), u0k− ), Hψk+ = f (tk , x 0 (tk ), u0k+ ) = ẋ 0k+ , Hψk− = f (tk , x 0 (tk ), u0k− ) = ẋ 0k− , ẋ 0k+ = ẋ(tk +), ẋ 0k− = ẋ(tk −), and [Htλ ]k is the jump of the function Htλ = ψ(t)ft (t, x 0 (t), u0 (t)) at the point tk , i.e., [Htλ ]k = Htλk+ − Htλk− = ψ(tk )ft (tk , x 0 (tk ), u0k+ ) − ψ(tk )ft (tk , x 0 (tk ), u0k− ). We note that D k (H λ ) depends on λ linearly, and D k (H λ ) ≥ 0 for any λ ∈ M0 and any tk ∈ . Let [Hxλ ]k be the jump of the function Hxλ = ψ(t)fx (t, x 0 (t), u0 (t)) = −ψ̇(t) at the point tk ∈ , i.e., [Hxλ ]k = Hxλk+ − Hxλk− . For λ ∈ M0 , z̄ ∈ Z2 (), we set (see also Henrion [39]) 1 k λ 2 k D (H )ξ̄k + 2[Hxλ ]k x̄av ξ̄k 2 k=1 tf 1 λ λ lpp + (p 0 )p̄, p̄ + Hww (t, w0 )w̄, w̄ dt , 2 t0 s λ (z̄) = (2.13) k = 1 (x̄ k− + x̄ k+ ), p̄ = (x̄(t ), x̄(t )). Obviously, λ (z̄) is a quadratic form in z̄ where x̄av 0 f 2 and a linear form in λ. 2.1.7 Necessary Quadratic Condition for a Pontryagin Minimum In the case where the set M0 is nonempty, we set 0 (z̄) = max λ (z̄). λ∈M0 (2.14) If M0 is empty, we set 0 (· ) = −∞. Theorem 2.4. If w 0 is a Pontryagin minimum point, then M0 is nonempty and the function 0 (· ) is nonnegative on the cone K. 1 In the book [79], the value D k (H λ ) was defined by the formula D k (H λ ) = d ( H λ )(t ), since there k k dt the Pontryagin maximum principle was used. 2.1. Statements of Quadratic Conditions for a Pontryagin Minimum 33 We say that Condition A holds for the point w 0 if the set M0 is nonempty and the function 0 is nonnegative on the cone K. According to Theorem 2.4, the condition A is necessary for a Pontryagin minimum. 2.1.8 Sufficient Quadratic Condition for a Pontryagin Minimum A natural strengthening of Condition A is sufficient for the strict Pontryagin minimum. (We will show in Section 2.7 that it is also sufficient for the so-called “bounded strong minimum”.) We now formulate this strengthening. For this purpose, we define the set M0+ of elements λ ∈ M0 satisfying the so-called “strict minimum principle” and its subset Leg+ (M0+ ) of “strictly Legendre elements.” We begin with the definition of the concept of strictly Legendre element. An element λ = (α0 , α, β, ψ) ∈ 0 is said to be strictly Legendre if the following conditions hold: (a) [H λ ]k = 0, D k (H λ ) > 0 for all tk ∈ ; (b) for any t ∈ [t0 , tf ]\, the inequality Huu (t, x 0 (t), u0 (t), ψ(t))ū, ū > 0 holds for all ū ∈ Rd(u) , ū = 0; (c) for each point tk ∈ , the inequalities Huu (tk , x 0 (tk ), u0k− , ψ(tk ))ū, ū > 0 and Huu (tk , x 0 (tk ), u0k+ , ψ(tk ))ū, ū > 0 hold for all ū ∈ Rd(u) , ū = 0. We note that each element λ ∈ M0 is Legendre in nonstrict sense, i.e., [H λ ]k = 0, ≥ 0 for all tk ∈ , and in conditions (b) and (c) nonstrict inequalities hold for the quadratic form Huu ū, ū . In other words, Leg(M0 ) = M0 , where Leg(M) is a subset of Legendre (in nonstrict sense) elements of the set M ⊂ 0 . Further, denote by M0+ the set of λ ∈ M0 such that (a) H (t, x 0 (t), u, ψ(t)) > H (t, x 0 (t), u0 (t), ψ(t)) if t ∈ [t0 , tf ]\, u ∈ U(t, x 0 (t)), u = u0 (t); (b) H (tk , x 0 (tk ), u, ψ(tk )) > H λk− = H λk+ if tk ∈ , u ∈ U(tk , x 0 (tk )), u∈ / {u0k− , u0k+ }. D k (H λ ) Denote by Leg+ (M0+ ) the subset of all strictly Legendre elements λ ∈ M0+ . We set tf γ̄ (z̄) = ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) + ū(t), ū(t) dt. (2.15) t0 We say that Condition B holds for the point M ⊂ Leg+ (M0+ ) and a constant C > 0 such that w0 if there exist a nonempty compact set max λ (z̄) ≥ C γ̄ (z̄) ∀ z̄ ∈ K. λ∈M Theorem 2.5. If Condition B holds, then w0 is a strict Pontryagin minimum point. In conclusion, we note that the space P W 1,2 (, Rd(x) ) with the inner product tf s ˙¯ ¯ 0) + ¯ k + ˙ x̄(t) x̄, x̄¯ = x̄(t0 ), x̄(t [x̄]k , [x̄] x̄(t), dt, t0 k=1 and also the space Z2 () with the inner product z̄, z̄¯ = ξ̄ , ξ̄¯ + x̄, x̄¯ + tf t0 ¯ ū(t), ū(t) dt, 34 Chapter 2. Quadratic Conditions in the Calculus of Variations are Hilbert spaces. Moreover, the functional γ̄ (z̄) is equivalent to the norm z̄ Z2 () = √ z̄, z̄ on the subspace {z̄ ∈ Z2 () | x̄˙ = fw w̄, [x̄]k = [ẋ 0 ]k ξ̄k , k ∈ I ∗ }, where I ∗ = {1, . . . , s}. 2.2 Basic Constant and the Problem of Its Decoding 2.2.1 Verification of Assumptions in the Abstract Scheme The derivation of quadratic conditions for the Pontryagin minimum in the problem (2.1)– (2.4) is based on the abstract scheme presented in Chapter 1. We begin with the verification of the property that the problem (2.1)–(2.4), the set of sequences , and the point w0 satisfy all the assumptions of the abstract scheme. As the Banach space X, we consider the space W = W 1,1 (, Rd(x) ) × L∞ (, Rd(u) ) of pairs of functions w = (x, u). As the set entering the statement of the abstract problem (1.1), we consider the set W of pairs w = (x, u) ∈ W such that (x(t0 ), x(tf )) ∈ P and there exists a compact set C ⊂ Q such that (t, x(t), u(t)) ∈ C a.e. on [t0 , tf ]. Let w 0 = (x 0 , u0 ) ∈ W be a point satisfying the constraints of the canonical problem (2.1)–(2.4) and the assumptions of Section 2.1. As the set of sequences defining the type of minimum at this point, we take the set of Pontryagin sequences in the space W . Obviously, is invariant with respect to the operation of passing to a subsequence. Also, it is elementary verified that is invariant with respect to the 0 -extension, where 0 = {{w̄n } | w̄n = x̄n 1,1 + ūn ∞ → 0}. As was already mentioned, 0 corresponds to the set of sequences 0 introduced in Section 1.1. (Here, the notation 0 is slightly more convenient.) Finally, it is easy to see that the set W is absorbing for at the point w 0 . Therefore, all the assumptions on w0 , W , and from Section 1.1 hold. Furthermore, consider the functional w(· ) = (x(· ), u(· )) −→ J (x(t0 ), x(tf )). (2.16) Since the function J (p) is defined on P , we can consider this functional as a functional given on W . We denote it by Jˆ. Therefore, the functional Jˆ : W → R1 is given by formula (2.16). Analogously, we define the following functionals on W : F̂i : w(· ) = (x(· ), u(· )) ∈ W −→ Fi (x(t0 ), x(tf )), i = 1, . . . , d(F ), (2.17) and the operator ĝ : w(· ) = (x(· ), u(· )) ∈ W −→ (ẋ − f (t, w), K(p)) ∈ Y , (2.18) L1 (, Rd(x) ) × Rd(K) . We omit the verification of the property that all the where Y = functionals Jˆ(w), F̂i (w), i = 1, . . . , d(F ) and the operator ĝ(w) : W −→ Y are -continuous and strictly -differentiable at the point w 0 , and, moreover, their Fréchet derivatives at this point are linear functionals Jˆ (w 0 ) : w̄ = (x̄, ū) ∈ W −→ Jp (p 0 )p̄, F̂i (w0 ) : w̄ = (x̄, ū) ∈ W −→ Fip (p0 )p̄, i = 1, . . . , d(F ), and a linear operator ĝ (w 0 ) : w̄ = (x̄, ū) ∈ W −→ (x̄˙ − fw (t, w0 )w̄, Kp (p 0 )p̄) ∈ Y , (2.19) 2.2. Basic Constant and the Problem of Its Decoding 35 respectively; moreover, here, p̄ = (x̄(t0 ), x̄(tf )). The verification of this assertion is also very elementary. We prove only that if ĝ (w 0 )W = Y , then the image ĝ (w 0 )W is closed in Y and has a direct complement that is a closed subspace in Y . Note that the operator w̄ = (x̄, ū) ∈ W −→ x̄˙ − fw (t, w0 )w̄ ∈ L1 (, Rd(x) ) (2.20) is surjective. Indeed, for an arbitrary function v̄ ∈ L1 (, Rd(x) ), there exists a function x̄ satisfying x̄˙ − fx (t, w0 )x̄ = v̄ and x̄(t0 ) = 0. Then under the mapping given by this operator, the pair w̄ = (x̄, ū) yields v̄ in the image. Further, we note that the operator w̄ = (x̄, ū) ∈ W −→ Kp (p0 )p̄ ∈ Rd(K) (2.21) is finite-dimensional. The surjectivity of operator (2.20) and the finite-dimensional property of operator (2.21) imply the closedness of the image of operator (2.19). The following more general assertion holds. Lemma 2.6. Let X, Y , and Z be Banach spaces, let A : X → Y be a linear operator with closed range, and let B : X → Z be a linear operator such that the image B(Ker A) of the kernel of the operator A under the mapping of the operator B is closed in Z. Then the operator T : X → Y × Z defined by the relation Tx = (Ax, Bx) for all x ∈ X has a closed range. Proof. Let a sequence {xn } in X be such that Txn = (Axn , Bxn ) → (y, z) ∈ Y × Z. Since the range of AX is closed in Y , by the Banach theorem on the inverse operator there exists a convergent subsequence xn → x in X such that Axn = Axn → y and, therefore, Ax = y. Then A(xn − xn ) = 0 for all n and B(xn − xn ) → z − Bx. By the closedness of the range B(Ker A), there exists x ∈ X such that Ax = 0 and Bx = z − Bx. Then B(x + x ) = z and A(x + x ) = y as required. The lemma is proved. Since the image of any subspace under the mapping defined by a finite-dimensional operator is a finite-dimensional and, therefore, closed subspace, Lemma 2.6 implies the following assertion. Corollary 2.7. Let X, Y , and Z be Banach spaces, let A : X → Y be a surjective linear operator, and let B : X → Z be a finite-dimensional linear operator. Then the operator T : X → Y × Z defined by the relation Tx = (Ax, Bx) for all x ∈ X has a closed range. This implies that operator (2.19) has a closed range. Since the range of operator (2.19) is of finite codimension, this range has a direct complement that is finite-dimensional, and, therefore, it is a closed subspace. Therefore, all the conditions defining a smooth problem on at the point w 0 hold for the problem (2.1)–(2.4). Hence we can apply the main result of the abstract theory of higher-order conditions, Theorem 1.7. 2.2.2 Set 0 In what follows, for the sake of brevity we make the following agreement: if a point at which the derivative of a given function is taken is not indicated, then for the functions J , F , K, 36 Chapter 2. Quadratic Conditions in the Calculus of Variations and l λ , p0 is such a point, and for the functions f and H λ , it is the point (t, w0 (t)). Using the definition given by relations (1.3), we denote by 0 the set of tuples λ = (α0 , α, β, ψ) such that α0 ∈ R, α ∈ (Rd(F ) )∗ , β ∈ (Rd(K) )∗ , ψ(· ) ∈ L∞ (, (Rd(x) )∗ ) and the following conditions hold: α0 ≥ 0, α0 + d(F ) αi ≥ 0 (i = 1, . . . , d(F )), αi + |βj | + ψ = 1, (2.22) (2.23) j =1 i=1 α0 Jp p̄ + d(K) αi Fi (p0 ) = 0 (i = 1, . . . , d(F )), d(F ) αi Fip p̄ + βKp p̄ − tf ψ(x̄˙ − fw w̄) dt = 0 (2.24) t0 i=1 for all w̄ = (x̄, ū) ∈ W . Here, p̄ = (x̄(t0 ), x̄(tf )). Let us show that the normalization (2.23) is equivalent to the normalization α0 + d(F ) i=1 αi + d(K) |βj | = 1 (2.25) j =1 (the definition of equivalent normalizations was given in Section 1.1.5). The upper estimate is obvious: α0 + αi + |βj | ≤ 1 for any λ ∈ 0 . It is required to establish the lower estimate: there exists ε > 0 such that α0 + αi + |βj | ≥ ε (2.26) for any λ ∈ 0 . Suppose that this is not true. Then there exists a sequence λn = (α0n , αn , βn , ψn ) ∈ 0 such that α0n → 0, αn → 0, and βn → 0. By (2.24), this implies that the functional w̄ ∈ t W −→ t0f ψn (x̄˙ − fw w̄) dt is norm-convergent to zero. Since the operator w̄ ∈ W −→ (x̄˙ − fw w̄) ∈ L1 (, Rd(x) ) is surjective, this implies ψn ∞ → 0. Therefore, α0n + |αn | + |βn | + ψn → 0, which contradicts the condition α0n + αin + |βj n | + ψn = 1, which follows from λn ∈ 0 . Therefore, estimate (2.26) also holds with some ε > 0. The equivalence of normalizations (2.23) and (2.25) is proved. In what follows, we will use normalization (2.25) preserving the old notation for the new 0 . Let us show that the set 0 with normalization (2.25) coincides with the set 0 defined by (1.3) in Section 1.1.4. For λ = (α0 , α, β, ψ), let conditions (2.22), (2.24), and (2.25) hold. We rewrite condition (2.24) in the form tf lp p̄ − ψ(x̄˙ − fw w̄) dt = 0 ∀ w̄ ∈ W , (2.27) t0 where l = α0 J + αF + βK. We set x̄ = 0 in (2.27). Then t0f ψfu ū dt = 0 for all ū ∈ L∞ . Therefore, ψfu = 0 or Hu = 0, where H = ψf . Then we obtain from (2.27) that tf tf lp p̄ − ψ x̄˙ dt + ψfx x̄ dt = 0 (2.28) t t0 t0 2.2. Basic Constant and the Problem of Its Decoding 37 for all x̄ ∈ W 1,1 . Let us show that this implies the conditions ψ ∈ W 1,∞ , −ψ̇ = Hx , ψ(t0 ) = −lx0 , ψ(tf ) = lxf . (2.29) Let ψ satisfy the conditions ψ ∈ W 1,∞ , −ψ̇ = ψ fx , and ψ (t0 ) = −lx0 . Integrating by parts, we obtain tf tf tf tf ˙ ψ x̄ dt = ψ x̄ |t0 − ψ fx x̄ dt ψ̇ x̄ dt = lx0 x̄0 + ψ (tf )x̄f + t0 t0 t0 for any x̄ ∈ W 1,1 . (Hereafter x̄0 := x̄(t0 ), x̄f := x̄(tf )). Therefore, tf tf ˙ ψ x̄ dt − ψ fx x̄ dt − lx0 x̄0 − ψ (tf )x̄f = 0. t0 (2.30) t0 Adding (2.30) to (2.28), we obtain (lxf − ψ (tf ))x¯f + tf (ψ − ψ)x̄˙ dt + t0 tf (ψ − ψ )fx x̄ dt = 0 t0 for all x̄ ∈ W 1,1 . Let c̄ = lxf − ψ (tf ) and ψ̄ = ψ − ψ. Then tf c̄x̄f + ψ̄(x̄˙ − fx x̄) dt = 0. (2.31) t0 This is true for any x̄ ∈ W 1,1 . Let ā ∈ Rn , v̄ ∈ L1 . Let us find x̄ ∈ W 1,1 such that x̄˙ − fx x̄ = v̄ t and x̄(tf ) = ā. Then c̄ā + t0f ψ̄ v̄ dt = 0. Since this relation holds for an arbitrary ā ∈ Rn , v̄ ∈ L1 , we obtain that c̄ = 0 and ψ̄ = 0, i.e., ψ = ψ and ψ (tf ) = lxf . Therefore, conditions (2.29) hold. Conversely, if conditions (2.29) hold for ψ, then applying the integrationt t t by-parts formula t0f ψ x̄˙ dt = ψ x̄ |tf0 − t0f ψ̇ x̄ dt, we obtain condition (2.28). From (2.28) and the condition Hu = 0, condition (2.27), and therefore, condition (2.24) follow. Therefore, we have shown that the set 0 defined in this section coincides with the set 0 of Section 1.1.4. We note that the set 0 is a finite-dimensional compact set, and the projection λ = (α0 , α, β, ψ) −→ (α0 , α, β) is injective on 0 . Indeed, according to (α0 , α, β), the vector lxλ0 is uniquely defined, and from the conditions −ψ̇ = ψfx and ψ(t0 ) = lx0 the function ψ is uniquely found. The same is also true for the set co 0 . 2.2.3 Lagrange Function We set F0 (p) = J (p) − J (p0 ). Introduce the Lagrange function of the problem (2.1)–(2.4). Let δW be the set of variations δw ∈ W such that (w 0 + δw) ∈ W , i.e., (p 0 + δp) ∈ P , where δp = (δx(t0 ), δx(tf )), and there exists a compact set C ⊂ Q such that (t, w0 (t) + δw(t)) ∈ C a.e. on [t0 , tf ]. For λ = (α0 , α, β, ψ) ∈ co 0 and δw = (δx, δu) ∈ δW , we set (λ, δw) = α0F0 (p0 + δp) + αF (p0 + δp) + βK(p 0 + δp) − tf t0 ψ(ẋ 0 + δ ẋ − f (t, w0 + δw)) dt. 38 Chapter 2. Quadratic Conditions in the Calculus of Variations We set δFi = Fi (p0 + δp) − Fi (p 0 ), i = 1, . . . , d(F ), δK = K(p 0 + δp) − K(p 0 ) = K(p 0 + δp), δf = f (t, w 0 + δw) − f (t, w0 ). Then (λ, δw) = = d(F ) tf αi δFi + βδK − ψ(δ ẋ − δf ) dt t0 i=0 tf tf ψδ ẋ dt + δH λ dt, δl λ − t0 t0 where lλ = d(F ) αi Fi + βK, H λ = ψf , δl λ = l λ (p 0 + δp) − l λ (p 0 ), i=0 δH λ = H (t, w 0 + δw, ψ) − H (t, w0 , ψ) = ψδf . Note that in contrast to the classical calculus of variations, where δJ stands for the first variation of the functional, we denote by δJ , δf , etc. the full increments corresponding to the variation δw. 2.2.4 Basic Constant For δw ∈ δW , we set 0 (δw) = max (λ, δw). 0 R1 Let γ = γ (δw) : δW −→ be an arbitrary strict higher order on whose definition is given in Section 1.1.3. We set σ γ = {{δwn } ∈ | σ (δwn ) ≤ O(γ (δwn ))}, where σ (δw) = max{F0 (p 0 + δp), . . . , Fd(F ) (p0 + δp), |δK|, δ ẋ − δf 1 } is the violation function of the problem (4.1)–(4.4). In what follows, we shall use the shorter notation lim for the limit inferior instead of lim inf . We set 0 Cγ (0 , σ γ ) = inf lim . σ γ γ Theorem 1.7 implies the following theorem. Theorem 2.8. The condition Cγ (0 , σ γ ) ≥ 0 is necessary for a Pontryagin minimum at the point w0 , and the condition Cγ (0 , σ γ ) > 0 is sufficient for a strict Pontryagin minimum at this point. Further, there arises the problem of the choice of a higher order γ and decoding the constant Cγ (0 , σ γ ) corresponding to the chosen order. The constant Cγ (0 , σ γ ) is said to be basic, and by decoding of the basic constant, we mean the simplest method for calculating its sign. In what follows, we will deal with the choice of γ and decoding of the basic constant. As a result, we will obtain theorems on quadratic conditions for the Pontryagin minimum formulated in Section 2.1. 2.3. Local Sequences, Representation of the Lagrange Function 2.3 2.3.1 39 Local Sequences, Higher Order γ , Representation of the Lagrange Function on Local Sequences with Accuracy up to o(γ ) Local Sequences and Their Structure As before, let w0 ∈ W be a point satisfying the constraints of the canonical problem and the assumptions of Section 2.1. For convenience, we assume that u0 is left continuous at each point of discontinuity tk ∈ . Denote by u0 the closure of the graph of u0 (t). Let loc be the set of sequences {δwn } in the space W satisfying the following two conditions: (a) δxn 1,1 → 0; (b) for any neighborhood V of the compact set u0 there exists n0 ∈ N with (t, u0 (t) + δun (t)) ∈ V a.e. on [t0 , tf ] ∀ n ≥ n0 . (2.32) Sequences from loc are said to be local. Obviously, 0 ⊂ loc ⊂ . Although the set of local sequences loc is only a part of the set of Pontryagin sequences, all main considerations in obtaining quadratic conditions for the Pontryagin minimum are related namely to the set loc . Let us consider the structure of local sequences. Denote by loc u the set of sequences {δun } in L∞ (, Rd(u) ) such that for any neighborhood V of the compact set u0 there exists a number starting from which condition (2.32) holds. We briefly write the 0 0 condition defining the sequences from loc u in the form (t, u + δun ) → u . Therefore, loc 0 0 = {{δwn } | δxn 1,1 → 0, (t, u + δun ) → u }. In what follows, in order not to abuse the notation, we will omit the number n in sequences. Let Qtu be the projection of Q under the mapping (t, x, u) → (t, u). Then Qtu is an open set in Rd(u)+1 containing the compact set u0 . Denote by u0 (tk−1 , tk ) the closure in Rd(u)+1 of the intersection of the compact set u0 with the layer {(t, u) | u ∈ Rd(u) , t ∈ (tk−1 , tk )}, where k = 1, . . . , s + 1, and ts+1 = tf . In other words, u0 (tk−1 , tk ) is the closure of the graph of the restriction of the function u0 (· ) to the interval (tk−1 , tk ). Obviously, u0 is the union k of disjoint compact sets u0 (tk−1 , tk ). For brevity, we set u0 (tk−1 , tk ) = u0 . k Let Vk ⊂ Qut be fixed disjoint bounded neighborhoods of the compact sets u0 , k = 1, . . . , s + 1. We set s+1 Vk . (2.33) V= k=1 Without loss of generality, we assume that V, together Then V is a neighborhood of with its closure, is contained in Qut . Recall that for brevity, we set I ∗ = {1, . . . , s}. By the superscript “star” we denote the functions and sets related to the set of points of discontinuity of the function u0 . Define the following subsets of the neighborhood V; cf. the illustration in Figure 2.1 for k = 1: u0 . ∗ Vk− = {(t, u) ∈ Vk+1 | t < tk }; ∗ ∗ ∪ Vk+ , Vk∗ = Vk− k ∈ I ∗; ∗ Vk+ = {(t, u) ∈ Vk | t > tk }; ∗ V = Vk∗ ; V 0 = V\V ∗ . k∈I ∗ 40 Chapter 2. Quadratic Conditions in the Calculus of Variations 6 u ' u1+ ∗ V1− u(t) V2 & $ u1− u(t) V1 - V∗ 1+ % t1 t0 tf t - Figure 2.1. Neighborhoods of the control at a point t1 of discontinuity. While the superscripts k−, k+, and k were used for designation of left and right limit values and ordinary values of functions at the point tk ∈ , the same subscripts will be used for enumeration of sets and functions related to the point tk ∈ . The notation vraimaxt∈M u(t) will be often used to denote the essential supremum (earlier denoted also by ess sup) of a function u(·) on a set M. 0 ∗ 0 Further, let {δu} ∈ loc u , i.e., (t, u + δu) → u , and let k ∈ I . For a sequence {δu}, ∗ ∗ ∗ is 0 introduce the sequence of sets Mk− = {t | (t, u (t) + δu(t)) ∈ Vk− } and assume that χk− ∗ ∗ ∗ the characteristic function of the set Mk− . We set {δuk− } = {δuχk− }. Then vraimax |u0 (t) + δu∗k− (t) − u0k+ | → 0, ∗ where t∈Mk− u0k+ = u0 (tk +). ∗ → u0k+ . Analogously, for the sequence {δu}, In short, we write this fact as (u0 + δu∗k− ) |Mk− we define ∗ ∗ = {t | (t, u0 (t) + δu(t)) ∈ Vk+ } and Mk+ ∗ {δu∗k+ } = {δuχk+ }, ∗ is the characteristic function of the set M ∗ . Then (u0 +δu∗ ) | ∗ → u0k− , i.e., where χk+ k+ k+ Mk+ |u0 (t) + δu∗k+ (t) − u0k− | → 0, vraimax ∗ where t∈Mk+ u0k− = u0 (tk −). The sequence {δu∗k− } belongs to the set of sequences ∗uk− defined as follows: ∗uk− consists of sequences {δu} in L∞ (, Rd(u) ) such that (a) vraimaxt∈M t ≤ tk , where M = {t | δu(t) = 0}, i.e., the support M of each member δu of the sequence {δu} is located to the left from tk (here and in what follows, by the support of a measurable function, we mean the set of points at which it is different from zero); for brevity, we write this fact in the form M ≤ tk ; (b) vraimaxt∈M |t − tk | → 0, i.e., the support M tends to tk ; for brevity, we write this fact as M → tk ; (c) vraimaxt∈M |u0 (t) + δu(t) − u0k+ | → 0, i.e., the values of the function u0 + δu on the support M tend to u0k+ ; in short, we write this fact in the form (u0 + δu) |M → u0k+ . 2.3. Local Sequences, Representation of the Lagrange Function 41 Analogously, we define the set ∗uk+ consisting of sequences {δu} such that M ≥ tk , M → tk , (u0 + δu) |M → u0k− , where M = {t | δu(t) = 0}. Clearly, {δu∗k+ } ∈ ∗uk+ . We set ∗uk− + ∗uk+ = ∗uk , ∗uk = ∗u . k∈I ∗ By definition, the sum of sets of sequences consists of all sums of sequences belonging to these sets. ∗ ∗ ∗ ∗ As before, let {δu} ∈ loc u . Let the sets Mk− , Mk+ and the functions χk− , χk+ correspond to the members of this sequence. We set ∗ ∪ M∗ , M∗ = Mk∗ = Mk− Mk∗ , k+ k∈I ∗ ∗ ∗ ∗ ∗ χk = χk− + χk+ , χ = χk∗ , ∗ k∈I δu∗k = δuχk∗ = δu∗k− + δu∗k+ , δu∗ = δuχ ∗ = δu∗k . k∈I ∗ Then the sequence {δu∗ } belongs to the set ∗u . Further, for members δu of the sequence {δu}, we set δu0 = δu − δu∗ . Then starting from a certain number, we have for the sequence {δu0 } that (t, u0 (t) + δu0 (t)) ∈ V 0 a.e. on [t0 , tf ]. Moreover, δu0 χ ∗ = 0. Here, we assume that members of the sequences {δu0 } and {χ ∗ } having the same number are multiplied. This remark refers to all relations which contain members of distinct sequences. In what follows, we do not make such stipulations. Clearly, δu0 ∞ → 0. We denote the set of sequence in L∞ (, Rd(u) ) having this property by 0u . Therefore, we have shown that an arbitrary sequence {δu} ∈ loc u admits the representation {δu} = {δu0 } + {δu∗ }, {δu0 } ∈ 0u , {δu∗ } ∈ ∗u , {δu0 χ ∗ } = {0}, where χ ∗ is the characteristic function of the set M ∗ = {t | δu∗ (t) = 0}. Such a representation is said to be canonical. It is easy to see that the canonical representation is unique. In 0 ∗ particular, the existence of the canonical representation implies loc u ⊂ u +u . Obviously, loc 0 ∗ the converse inclusion holds. Therefore, the relation u = u + u holds. We now introduce the set ∗ consisting of sequences {δw ∗ } = {(0, δu∗ )} in the space W such that {δu∗ } ∈ ∗u (the component δx of these sequence vanishes identically). Proposition 2.9. We have the relation loc = 0 + ∗ . Moreover, we can represent any sequence {δw} = {(δx, δu)} ∈ loc in the form {δw} = {δw0 } + {δw∗ }, where {δw0 } = {(δx, δu0 )} ∈ 0 , {δw∗ } = {(0, δu∗ )} ∈ ∗ , and , moreover, for all members of sequences with the same numbers, the condition δu0 χ ∗ = 0 holds; here, χ ∗ is the characteristic function of the set {t | δu∗ (t) = 0}. Recall that by definition, 0 = {{δw} | δx 1,1 + δu ∞ → 0}. Proposition 2.9 fol0 ∗ lows from the relation loc u = u + u and the existence of the canonical representation for sequences in loc u . 42 Chapter 2. Quadratic Conditions in the Calculus of Variations The representation of a sequence from loc in the form of a sum of sequences from and ∗ with the condition {δu0 χ ∗ } = {0}, which was indicated in Proposition 2.9, will also be called canonical. Obviously, the canonical representation is unique. It will play an important role in what follows. We introduce one more notation related to an arbitrary sequence {δw} ∈ loc . For such a sequence, we set 0 ∗ = (u0 + δu∗ − u0k+ )χ ∗ , δvk− = (u0 + δu − u0k+ )χk− k− k− ∗ = (u0 + δu∗ − u0k− )χ ∗ , 0 0k− δvk+ = (u + δu − u )χk+ k+ k+ {δvk } = {δvk− } + {δvk+ }, {δv} = {δvk }. k∈I ∗ ∗ , M ∗ , M ∗ , and M ∗ , Then the supports of δvk− , δvk+ , δvk , and δv are the sets Mk− k+ k ∗ ∗ respectively. Moreover, it is obvious that δvk− ∞ → 0, δvk+ ∞ → 0, δvk ∞ → 0, and ∗ }, {δv ∗ }, {δv }, and {δv} belong to 0 . δv ∞ → 0, i.e., the sequences {δvk− k u k+ 2.3.2 Representation of the Function f (t, x, u) on Local Sequences Let {δw} ∈ loc . Recall that 0 (δw) := max (λ, δw) = max (λ, δw), λ∈0 tf t0 λ∈co 0 tf t0 where (λ, δw) = δl λ − ψδ ẋ dt + δH λ dt, δH λ = ψδf , and co 0 is the convex hull of 0 . Consider δf on the sequences {δw}. Represent {δw} in the canonical form: {δw} = {δw0 } + {δw ∗ }, {δw∗ } ∈ ∗ , {δw 0 } ∈ 0 , {δu0 χ ∗ } = {0}. Then δf = f (t, w 0 + δw) − f (t, w0 ) = f (t, w 0 + δw) − f (t, w0 + δw ∗ ) + f (t, w 0 + δw ∗ ) − f (t, w 0 ) = f (t, w 0 + δw ∗ + δw 0 ) − f (t, w0 + δw ∗ ) + δ ∗ f 1 = fw (t, w0 + δw ∗ )δw0 + fww (t, w0 + δw ∗ )δw0 , δw 0 + r + δ ∗ f , 2 (2.34) where δ ∗ f = f (t, w0 + δw∗ ) − f (t, w0 ), and the residue term r with the components ri , i = 1, . . . , d(f ), is defined by the mean value theorem as follows: 1 ri = (fiww (t, w 0 + δw ∗ + ζi δw 0 ) − fiww (t, w 0 + δw∗ ))δw 0 , δw0 , 2 ζi = ζi (t), 0 ≤ ζi (t) ≤ 1, i = 1, . . . , d(f ). Therefore, r 1 =o δx 2 C + tf |δu | dt . 0 2 (2.35) t0 Further, fw (t, w0 + δw ∗ )δw0 = fw (t, w 0 )δw0 + (δ ∗ fw )δw0 , where δ ∗ fw = fw (t, w0 + δw ∗ ) − fw (t, w0 ). Since (δ ∗ fw )δw0 = (δ ∗ fx )δx + (δ ∗ fu )δu0 = (δ ∗ fx )δx (because δu0 χ ∗ = 0), we have (2.36) fw (t, w0 + δw ∗ )δw0 = fw δw 0 + (δ ∗ fx )δx. 2.3. Local Sequences, Representation of the Lagrange Function 43 Here and in what follows, we set fw = fw (t, w0 ) for brevity. Analogously, fww (t, w 0 + δw ∗ )δw 0 , δw0 = fww (t, w0 )δw0 , δw0 + (δ ∗ fww )δw0 , δw0 = fww (t, w0 )δw0 , δw0 + (δ ∗ fxx )δx, δx . Setting fww (t, w0 ) = fww for brevity, we obtain fww (t, w0 + δw ∗ )δw0 , δw 0 = fww δw 0 , δw0 + (δ ∗ fxx )δx, δx , where δ ∗ fxx = fxx (t, w0 + δw∗ ) − fxx (t, w0 ). Moreover, δ ∗ fxx This implies (δ ∗ fxx )δx, δx 1 = o( δx 2C ). 1 (2.37) ≤ const· meas M∗ → 0. (2.38) Substituting (2.36) and (2.37) in (2.34) and taking into account estimates (2.35) and (2.38), we obtain the following assertion. Proposition 2.10. Let {δw} ∈ loc . Then the following formula holds for the canonical representation {δw} = {δw 0 } + {δw∗ }: 1 δf = fw δw 0 + fww δw 0 , δw0 + δ ∗ f + (δ ∗ fx )δx + r̃, 2 where r̃ 1 =o δx 2 C + tf |δu | dt , 0 2 δ ∗ f = f (t, w 0 + δw ∗ ) − f (t, w0 ), t0 δ ∗ fx = fx (t, w0 + δw ∗ ) − fx (t, w0 ), fw = fw (t, w 0 ), fww = fww (t, w0 ). Therefore, for any λ ∈ co 0 , we have tf tf 1 tf λ λ λ δH dt = Hx δx dt + Hww δw0 , δw0 dt 2 t0 t0 t0 tf tf + δ ∗ H λ dt + (δ ∗ Hxλ δx) dt + ρ λ , t0 where supλ∈co 0 |ρ λ | = o( δx 2 C + tf t0 (2.39) t0 |δu0 |2 dt), δ ∗ H λ = ψδ ∗ f , and δ ∗ Hxλ = ψδ ∗ fx . Here, we have used the relation Huλ = 0 for all λ ∈ co 0 . As above, all derivatives whose argument is not indicated are taken for w = w0 (t). 2.3.3 Representation of the Integral Sequences tf t0 (δ ∗ Hxλ )δx dt on Local Proposition 2.11. Let two sequences {δx} and {δu∗ } such that δx C → 0 and {δu} ∈ ∗u be given. Let λ ∈ co 0 . Then tf tf ∗ ∗ (δ ∗ Hxλ )δx dt = [Hxλ ]k δx(χk− − χk+ ) dt + ρ ∗λ , (2.40) t0 k∈I ∗ t0 44 Chapter 2. Quadratic Conditions in the Calculus of Variations where supλ∈co 0 |ρ ∗λ | = o( δx 2 + λ k I ∗ Mk∗ |δtk | dt), δtk = tk −t, [Hx ] C ψ(tk )(fxk+ − fxk− ) = ψ(tk )(fx (tk , x 0 (tk ), u0k+ ) − fx (tk , x 0 (tk ), u0k− )). Proof. Since χ ∗ = tf t0 δ ∗ Hxλ δx dt = χk∗ , we have δ ∗ Hxλ = tf t0 I∗ δk∗ Hxλ δx dt, = Hxλk+ −Hxλk− = δk∗ Hxλ , and, therefore, δk∗ Hxλ = ψδk∗ f = ψ(f (t, w 0 + δwk∗ ) − f (t, w0 )). Further, for ψ = ψ(t) we have ψ(t) = ψ(tk ) + k ψ, where k ψ = ψ(tk + δtk ) − ψ(tk ), and, moreover, supco 0 |k ψ| ≤ const |δtk |, since −ψ̇ = ψfx , and supco 0 ψ ∞ < +∞. ∗ + χ ∗ , we have δ ∗ f = δ ∗ f + δ ∗ f , where the inConsider δk∗ fx . Since χk∗ = χk− k+ k x k− x k+ x ∗ ∗ crements δk− fx and δk+ fx correspond to the variations δu∗k− and δu∗k+ , respectively. For ∗ f , we have δk− x ∗ ∗ ∗ ∗ fx = [fx ]k χk− + (δk− fx − [fx ]k )χk− , δk− where [fx ]k = fxk+ − fxk− , fxk− = fx (tk , x 0 (tk ), u0k− ), fxk+ = fx (tk , x 0 (tk ), u0k+ ). ∗ f −[f ]k )χ ∗ = (f (t, x 0 , u0 +δu∗ )−f k+ )χ ∗ −(f (t, x 0 , u0 )− Further, let ηk− := (δk− x x x x x k− k− k− ∗ ∗ k− 0 0k+ 0 ∗ ∗ fx )χk− . Then ηk− ∞ → 0, since u + δuk− |Mk− → u and u |Mk− → u0k− . There∗ ∗ ∗ k fore, δk− fx = [fx ] χk− + ηk− , ηk− χk− = ηk− , ηk− ∞ → 0. Therefore, ∗ Hλ δk− x λ where supco 0 ηk− = = λ χ ∗ = ηλ . This implies → 0, ηk− k− k− ∞ tf t0 λ = where ρk− tf t0 ∗ δk− Hxλ δx dt = tf ∗ λ [Hxλ ]k δxχk− dt + ρk− , t0 (2.41) λ δx dt. Let us estimate ρ λ . We have ηk− k− λ λ | ≤ sup ηk− |ρk− co 0 ∗ f = (ψ(t ) + ψ)([f ]k + η )χ ∗ ψδk− x k k x k− k− ∗ + ηλ = [H λ ]k χ ∗ + ηλ , ψ(tk )[fx ]k χk− x k− k− k− ∞ δx ∗ C meas Mk− 1 sup ηλ 2 co 0 k− ≤ ∗ )2 ≤ 2 We use the obvious estimate (meas Mk− ∗ |δtk | dt. Mk− λ |=o sup |ρk− 2 C δx co 0 Analogously, we prove that tf ∗ δk+ Hxλ δx dt = − t0 tf t0 ∞ + ∗ Mk− δx 2 C ∗ 2 + (meas Mk− ) . Then |δtk | dt . ∗ λ [Hxλ ]k δxχk+ dt + ρk+ , (2.42) (2.43) 2.3. Local Sequences, Representation of the Lagrange Function where sup co 0 λ |ρk+ |=o δx 2 C + 45 ∗ Mk+ |δtk | dt . (2.44) ∗ H λ + δ ∗ H λ , we obtain from (2.41)–(2.43) that Taking into account that δk∗ Hxλ = δk− x k+ x tf tf ∗ ∗ δk∗ Hxλ δx dt = [Hxλ ]k δx(χk− − χk+ ) dt + ρkλ , t0 where supco 0 |ρkλ | = o t0 δx 2 C + Mk∗ |δtk | dt . This and the relation δ ∗ Hxλ = δk∗ Hxλ im- ply (2.40). The proposition is proved. 2.3.4 Representation of the Integral tf t0 δ ∗ H λ dt on Local Sequences t We consider the term t0f δ ∗ H λ dt on an arbitrary sequence {δu∗ } ∈ ∗u . Since χ ∗ = χk∗ , ∗ H λ + δ ∗ H λ , which we have δ ∗ H λ = δk∗ H λ , where δk∗ H λ = ψδk∗ f . In turn, δk∗ H λ = δk− k+ ∗ ∗ ∗ ∗ Hλ = corresponds to the representation δuk = δuk− + δuk+ . Consider the increment δk− ∗ ∗ ∗ ∗ f = 0 0 0 0 ψδk− f . By definition, δk− f = f (t, x , u + δuk− ) − f (t, x , u ), and, moreover, δk− ∗ ∗ δk− f χk− . Recall that we have introduced the function ∗ ∗ = (u0 + δu∗k− − u0k+ )χk− . δvk− ∗ , we have t = t + δt , u0 + δu∗ = u0k+ + δv , x 0 (t) = x 0 (t + δt ) = x 0 (t ) + On Mk− k k k− k k k k− 0 k x , where k x 0 = x 0 (tk + δtk ) − x 0 (tk ). But k x 0 = ẋ 0k− δtk + o(δtk ), where ẋ 0k− = ∗ , ẋ 0 (tk −). Therefore, |k x 0 | = O(δtk ). This implies that on Mk− f (t, x 0 , u0 + δu∗k− ) = f (tk + δtk , x 0 (tk ) + k x 0 , u0k+ + δvk− ) = f k+ + ftk+ δtk + fxk+ k x 0 + fuk+ δvk− 1 + (f )k+ (δtk , k x 0 , δvk− ), (δtk , k x 0 , δvk− ) 2 + o |δtk |2 + |k x 0 |2 + |δvk− |2 , where f k+ , ftk+ , fxk+ , fuk+ , and (f )k+ are values of the function f and its derivatives at the point (tk , x 0 (tk ), u0k+ ). Taking into account that k x 0 = ẋ 0k− δtk + o(δtk ), we obtain ∗ , from this that on Mk− f (t, x 0 , u0 + δu∗k− ) = f k+ + ftk+ δtk + fxk+ ẋ 0k− δtk + fuk+ δvk− 1 k+ + fuu δvk− , δvk− + o(|δtk | + |δvk− |2 ). 2 (2.45) ∗ . For t < t , we set u0 = u0 (t) − u0k− . Note that Further, consider f (t, x o , u0 ) on Mk− k k 0 |k u | = O(δtk ) by the assumption that tk is an L-point of the function u0 . Moreover, k x 0 = ẋ 0k− δtk + o(δtk ). Therefore, f (t, x 0 , u0 ) = f (tk + δtk , x 0 (tk ) + k x 0 , u0k− + k u0 ) = f k− + ftk− δtk + fxk− ẋ 0k− δtk + fuk− k u0 + o(δtk ), (2.46) 46 Chapter 2. Quadratic Conditions in the Calculus of Variations where f k− , ftk− , fxk− , and fuk− are the values of the function f and its derivatives at the point (tk , x 0 (tk ), u0k− ). Subtracting (2.46) from (2.45), we obtain the following on Mk− : δ ∗ k− f = [f ]k + [ft ]k δtk + [fx ]k ẋ 0k− δtk + fuk+ δvk− − fuk− k u0 1 k+ + fuu δvk− , δvk− + o(|δtk | + |δvk− |2 ). 2 (2.47) ∗ f and δv ∗ ∗ Taking into account that δk− k− are concentrated on Mk− and δtk = −|δtk | on Mk− , we obtain from this that ∗ − [f ]k + [f ]k ẋ 0k− |δt |χ ∗ + f k+ δv δ ∗ k− f = [f ]k χk− t x k k− k− u (2.48) 1 2 ∗ ∗ + f k+ δv , δv −fuk− k u0 χk− k− k− + o(|δtk | + |δvk− | )χk− . uu 2 ∗ f , we have the formula Analogously, for δk+ ∗ − [f ]k + [f ]k ẋ 0k+ |δt |χ ∗ + f k− δv δ ∗ k+ f = −[f ]k χk+ t x k k+ k+ u 1 2 ∗ ∗ + f k− δv , δv −fuk+ k u0 χk+ k+ k+ + o(|δtk | + |δvk+ | )χk+ , 2 uu (2.49) where k u0 = u0 (t) − u0k+ for t > tk , and ẋ 0k+ = ẋ 0 (tk +). We mention a consequence of formulas (2.48) and (2.49), which will be needed in what follows. Adding (2.48) and (2.49) and taking into account that |k u0 | = O(δtk ), we obtain ∗ ∗ δk∗ f = [f ]k (χk− − χk+ ) + fuk+ δvk− + fuk− δvk+ + O(|δtk | + |δvk |2 )χk∗ . (2.50) This formula holds for an arbitrary sequence {δu∗k } ∈ ∗uk . ∗ Hλ = Let us return to formula (2.48) and use it to obtain the expression for δk− ∗ f . On M ∗ , we have ψδk− k− ψ ψ(t) = ψ(tk + δtk ) = ψ(tk ) + ψ̇ k− δtk + ηk− |δtk | ψ = ψ(tk ) − ψ̇ k− |δtk | + ηk− |δtk |, (2.51) ψ where ψ̇ k− = ψ̇(tk −) and supco 0 |ηk− | → 0 as |δtk | → 0, since ψ̇ = −Hxλ is left continuous at the point tk . We obtain from (2.48) and (2.51) that ∗ ∗ ∗ δk− H λ = [H λ ]k χk− − [Htλ ]k + [Hxλ ]k ẋ 0k− + ψ̇ k− [Hψ ]k |δtk |χk− 1 λk+ λ + Huu δvk− , δvk− + η̃k− (|δtk | + |δvk− |2 ), 2 (2.52) λ χ ∗ = η̃λ and sup λ where η̃k− λ∈co 0 η̃k− ∞ → 0. Here, we have taken into account that k− k− λk+ λk− Hu = Hu = 0 for all λ ∈ co 0 and [f ]k = [Hψ ]k . ∗ Hλ = We now turn to formula (2.49) and use it to obtain the expression for δk+ ∗ ∗ ψδk+ f . On Mk+ , we have ψ(t) ψ = ψ(tk + δtk ) = ψ(tk ) + ψ̇ k+ δtk + ηk+ δtk ψ = ψ(tk ) + ψ̇ k+ |δtk | + ηk+ |δtk |, (2.53) 2.3. Local Sequences, Representation of the Lagrange Function 47 ψ where ψ̇ k+ = ψ̇(tk +), and supco 0 |ηk+ | → 0. Analogously to formula (2.52), we obtain from (2.49) and (2.53) that ∗ ∗ ∗ δk+ H λ = −[H λ ]k χk+ − [Htλ ]k + [Hxλ ]k ẋ 0k+ + ψ̇ k+ [Hψ ]k |δtk |χk+ 1 λk− λ + Huu δvk+ , δvk+ + η̃k+ (|δtk | + |δvk+ |2 ), 2 (2.54) λ χ ∗ = η̃λ and sup λ where η̃k+ λ∈co 0 η̃k+ ∞ → 0. k+ k+ A remarkable fact is that the coefficients of |δtk | in formulas (2.52) and (2.54) coincide and are the derivative at the point t = tk of the function {k H λ }(t) introduced in Section 2.1.6. Let us show this. Let k ∈ I ∗ and λ ∈ co 0 . Recall that by definition, (k H λ )(t) = H (t, x 0 (t), u0k+ , ψ(t)) − H (t, x 0 (t), u0k− , ψ(t)) = ψ(t)(k f )(t), where (k f )(t) = f (t, x 0 (t), u0k+ )−f (t, x 0 (t), u0k− ). In what follows, we will omit the superscript λ of H . The conditions ψ̇(t) = −ψ(t)fx (t, x 0 (t), u0 (t)), ẋ 0 (t) = f (t, x 0 (t), u0 (t)), and the property of the function u0 (t) implies the existence of the left and right derivatives of the functions ψ(t) and x 0 (t) at the point tk , and, moreover, the left derivative is the left limit of the derivatives, and the right derivative is the right limit of the derivatives. On each of the intervals of the set [t0 , tf ]\, the derivatives ψ̇, ẋ 0 are continuous. Analogous assertions hold for the function (k H )(t). Its left derivative at the point tk can be calculated by the formula d d H (t, x 0 (t), u0k+ , ψ(t)) − H (t, x 0 (t), u0k− , ψ(t)) (k H )(tk ) = dt − dt − t=tk = [Ht ]k + [Hx ]k ẋ 0k− + ψ̇ k− [Hψ ]k . (2.55) ∗ H . The right derivative This derivative is the coefficient of |δtk | in expression (2.52) for δk− k of the function {H } (t) at the point tk can be calculated by the formula d d 0 0k+ 0 0k− H (t, x (t), u , ψ(t)) − H (t, x (t), u , ψ(t)) (k H )(tk ) = dt + dt + t=tk = [Ht ]k + [Hx ]k ẋ 0k+ + ψ̇ k+ [Hψ ]k . (2.56) ∗ H . We show that This derivative is the coefficient of |δtk | in expression (2.54) for δk+ d d (k H )(tk ) = (k H )(tk ); dt − dt + (2.57) i.e., the function (k H )(tk ) is differentiable at the point tk . Indeed, d (k H )(tk ) = [Ht ]k + [Hx ]k ẋ 0k− + ψ̇ k− [Hψ ]k dt − = [Ht ]k + (Hxk+ − Hxk− )Hψk− − Hxk− (Hψk+ − Hψk− ) = [Ht ]k + Hxk+ Hψk− − Hxk− Hψk+ . (2.58) 48 Chapter 2. Quadratic Conditions in the Calculus of Variations But the right derivative has the same form: d (k H )(tk ) dt + = [Ht ]k + [Hx ]k ẋ 0k+ + ψ̇ k+ [Hψ ]k = [Ht ]k + (Hxk+ − Hxk− )Hψk+ − Hxk+ (Hψk+ − Hψk− ) = [Ht ]k + Hxk+ Hψk− − Hxk− Hψk+ . We denote the derivative of the function −(k H λ )(t) at the point tk by D k (H λ ). Therefore, we have proved the following assertion. Lemma 2.12. The function (k H )(t) is differentiable at each point tk ∈ . Its derivative at this point can be calculated by the formulas d (k H )(tk ) dt = [Htλ ]k + Hxλk+ Hψλk− − Hxλk− Hψλk+ = [Htλ ]k + [Hxλ ]k ẋ k− + ψ̇ k− [Hψλ ]k = [Htλ ]k + [Hxλ ]k ẋ k+ + ψ̇ k+ [Hψλ ]k . In particular, these formulas imply d d (k H )(tk −) = (k H )(tk +), dt dt and, therefore, (k H )(t) is continuously differentiable at each point tk ∈ . By definition − δ∗H λ d (k H )(tk ) = D k (H ). dt ∗ H λ + δ ∗ H λ and We now turn to formulas (2.52) and (2.54). Since δk∗ H λ = δk− k+ ∗ λ = k δk H , the following assertions follows from these formulas and Lemma 2.12. Proposition 2.13. Let {δw∗ } = {(0, δu∗ )} ∈ ∗ and λ = (α0 , α, β, ψ) ∈ co 0 . Then, 1 λk+ ∗ ∗ − χk+ ) + D k (H λ )|δtk |χk∗ + Huu δvk− , δvk− δ∗H λ = [H λ ]k (χk− 2 k∈I ∗ 1 λk− + Huu δvk+ , δvk+ + η̃kλ (|δtk | + |δvk |2 ) , 2 where supλ∈co 0 η̃kλ tf t0 δ ∗ H λ dt = ∞ → 0, η̃kλ χk∗ = η̃kλ , k ∈ I ∗ . Therefore, ∗ ∗ [H λ ]k (meas Mk− − meas Mk+ ) + D k (H λ ) k∈I ∗ 1 tf λk+ λk− Huu δvk− , δvk− + Huu δvk+ , δvk+ 2 t0 εkλ (|δtk | + |δvk |2 ) dt, + + k∈I ∗ Mk∗ where supλ∈co 0 |εkλ | → 0, k ∈ I ∗ . Mk∗ | δtk | dt dt 2.3. Local Sequences, Representation of the Lagrange Function 2.3.5 tf Representation of the Integral t0 49 δH λ dt on Local Sequences Propositions 2.10, 2.11, and 2.13 imply the following assertion. Proposition 2.14. Let {δw} ∈ loc be represented in the canonical form: {δw} = {δw 0 } + {δw∗ }, where {δw 0 } = {(δx, δu0 )}, {δw∗ } = {(0, δu∗ )}, {δu0 χ ∗ } = {0}. Then the following formula holds for any λ ∈ co 0 : tf tf 1 tf λ λ λ δH dt = Hx δx dt + Hww δw 0 , δw0 dt 2 t0 t0 t0 λ k ∗ ∗ k λ [H ] (meas Mk− − meas Mk+ ) + D (H ) + |δtk | dt Mk∗ k∈I ∗ tf 1 λk+ λk− Huu δvk− , δvk− + Huu δvk+ , δvk+ dt + 2 t0 tf ∗ ∗ λ, +[Hxλ ]k δx(χk− − χk+ ) dt + ρH t0 λ = ελ ( where ρH H 2.3.6 k∈I ∗ Mk∗ (|δtk |+|δvk | 2 ) dt + tf t0 |δu0 |2 dt + δx 2 ), sup λ λ∈co 0 |εH | → 0. C Expansion of the Lagrange Function on Local Sequences tf t0 Now recall that (λ, δw) = δl λ − the decomposition ψδ ẋ dt + tf t0 δH λ dt for {δw} ∈ loc . For δl λ , we have 1 λ δl λ = l λ (p0 + δp) = lpλ δp + lpp δp, δp + εlλ |δp|2 , 2 t where supλ∈co 0 |εlλ | → 0. We integrate the term t0f ψδ ẋ dt by parts and use the transversality conditions and also the adjoint equation: tf tf ψδ ẋ dt = −ψ(t0 )δx(t0 ) + ψ(tf )δx(tf ) − ψ̇δx dt t0 t0 = lxλ0 δx(t0 ) + lxλf δx(tf ) + From this we obtain δl λ − tf t0 tf t0 Hxλ δx dt = lpλ δp + 1 λ ψδ ẋ dt = lpp δp, δp − 2 tf t0 tf t0 Hxλ δx dt. Hxλ δx dt + εlλ |δp|2 . (2.59) For the sequence {δw} ∈ loc represented in the canonical form, we set γ (δw) = δx 2 C + tf t0 |δu | dt + 2 0 2 s ∗ k=1 Mk tf |δtk | dt + t0 Formula (2.59) and Proposition 2.14 imply the following assertion. |δv|2 dt. (2.60) 50 Chapter 2. Quadratic Conditions in the Calculus of Variations Proposition 2.15. For any sequence {δw} ∈ loc represented in the canonical form, {δw} = {δw 0 } + {δw∗ }, {δw0 } = {(δx, δu0 )} ∈ 0 , {δw ∗ } = {(0, δu∗ )} ∈ ∗ , {δu0 χ ∗ } = {0}, and for any λ ∈ co 0 , we have the formula λ γ (δw), (λ, δw) = 1λ (δw) + ε where 1 λ 1 tf λ (δw) = lpp δp, δp + Hww δw 0 , δw0 dt 2 2 t0 s ∗ ∗ [H λ ]k (meas Mk− − meas Mk+ ) + 1λ k=1 +D (H ) + 1 2 k λ tf t0 |δtk | dt + [Hxλ ]k ∗ Mk tf t0 ∗ ∗ δx(χk− − χk+ ) dt λk+ λk− Huu δvk− , δvk− + Huu δvk+ , δvk+ dt (2.61) λ | → 0. and supλ∈co 0 |ε ∗ − In expression (2.61) for 1λ (δw), all terms, except for sk=1 [H λ ]k (meas Mk− ∗ ), are estimated through γ on any sequence {δw} ∈ loc starting from a certain meas Mk+ number. For example, tf 1 ∗ ∗ δx(χk− − χk+ ) dt ≤ δx C meas Mk∗ ≤ ( δx 2C + (meas Mk∗ )2 ) ≤ γ (δw), 2 t0 since (meas Mk∗ )2 ≤ 4 1 ∗ 2 Mk∗ |δtk | dt. (This estimate follows from the estimates 2 (meas Mk− ) 1 ∗ 2 ∗ ∗ ∗ |δtk | dt, and the equality meas Mk = meas Mk− + 2 (meas Mk+ ) ≤ Mk+ ≤ M∗ |δtk | dt, k− ∗ .) meas Mk+ Recall that by 0 we denote the set consisting of those λ ∈ 0 for which the conditions [H λ ]k = 0 for all k ∈ I ∗ hold. Proposition 2.15 implies the following assertion. Proposition 2.16. Let the set 0 be nonempty. Then there exists a constant C > 0 such that the following estimate holds at any sequence {δw} ∈ loc represented in the canonical form, starting from a certain number: max |(λ, δw)| ≤ C γ (δw). λ∈co 0 We will need this estimate later, in Section 2.5. We have made an important step in the way of distinguishing the quadratic form. Also, we have defined the functional γ on local sequences. We must extend this functional to Pontryagin sequences. Precisely, this functional will define the higher order which we will use to obtain quadratic conditions in the problem considered. Note that (λ, δw 0 ) := 1 λ 1 tf λ 0 0 2 lpp δp, δp + 2 t0 Hww δw , δw dt is the second variation of the Lagrange functional. 2.3. Local Sequences, Representation of the Lagrange Function 51 u 6 (t, v) = |v − u(t)|2 (t, v) > 0 ' ? (t, v) = |v − u1+ |2 + 2|t − t1 | - u1+ u(t) V2 & $ u(t) 6 u1− - (t, v) > 0 6 V1 % (t, v) = |v − u1− |2 + 2|t − t1 | (t, v) = |v − u(t)|2 t1 t0 tf t - Figure 2.2. Definition of functions (t, v) on neighborhoods of discontinuity points. 2.3.7 Higher Order γ We first define the concept of admissible function (t, u); cf. the illustration in Figure 2.2 for k = 1. Definition 2.17. A function (t, u) : Qtu → R is said to be admissible (or an order function) if it is continuous on Qtu and there exist disjoint neighborhoods Vk ⊂ Qtu of the compact sets u0 (tk−1 , tk ) such that the following five conditions hold: (1) (t, u) = |u − u0 (t)|2 if (t, u) ∈ Vk , t ∈ (tk−1 , tk ), k = 1, . . . , s + 1; (2) (t, u) = 2|t − tk | + |u − u0k− |2 if (t, u) ∈ Vk , t > tk , k = 1, . . . , s; (3) (t, u) = 2|t − tk | + |u − u0k+ |2 if (t, u) ∈ Vk+1 , t < tk , k = 1, . . . , s; (4) (t, u) > 0 on Qtu \V, where V = s+1 k=1 Vk ; (5) for any compact set F ⊂ Qtu \V, there exists a constant L > 0 such that |(t, u ) − (t, u )| ≤ L|u − u | if (t, u ) and (t, u ) belong to F . Let us show that there exists at least one admissible function . Fix arbitrary disjoint neighborhoods Vk ⊂ Qtu of the compact sets u0 (tk−1 , tk ) and define on V = Vk by conditions (1)–(3). We set Vε = {(t, u) ∈ V | (t, u) < ε}. For a sufficiently small ε = ε0 > 0, the set Vε0 is a neighborhood of u0 contained in V together with its closure. For the above ε0 , we set (t, u) if (t, u) ∈ Vε0 , 0 (t, u) = ε0 if (t, u) ∈ Qtu \Vε0 . Then the function 0 is admissible. An admissible function is not uniquely defined, but any two of them coincide in a sufficiently small neighborhood of the compact set u0 . 52 Chapter 2. Quadratic Conditions in the Calculus of Variations Let (t, u) be a certain admissible function. We set tf γ (δw) = δx 2C + (t, u0 + δu) dt. (2.62) t0 This functional is defined for pairs δw = (δx, δu) ∈ W such that (t, u0 + δu) ∈ Qtu a.e. on [t0 , tf ]. Such pairs are said to be admissible with respect to Qtu (also, in this case, the variation δu is said to be admissible with respect to Qtu ). It is easy to see that for any local sequence {δw} ∈ loc , the values of γ (δw) can be calculated by formula (2.60) starting from a certain number, and, therefore in the definition of γ and in formula (2.60), we have used the same notation. Let us verify that γ is a strict higher order on , where is the set of Pontryagin sequences. Obviously, γ ≥ 0, γ (0) = 0, and for any variation δw admissible on Qtu , the condition γ (δw) = 0 implies δw = 0. Let us show that the functional γ is -continuous at zero. It is required to show that γ (δw) → 0 for any Pontryagin sequence {δw}. Since the condition δx 1,1 → 0 holds for {δw} ∈ and δx C ≤ δx 1,1 , it suffices to show that for {δw} ∈ , the condition tf 0 1+d(u) . 0 0 t0 (t, u + δu) dt → 0 holds. Let Uε (u ) be an ε-neighborhood of the set u in R Assume that ε > 0 is chosen so that Uε (u0 ) ⊂ Qtu . Represent δu in the form δu = δuε + δuε , where δu(t) if (t, u0 (t) + δu(t)) ∈ Uε (u0 ), δuε (t) = 0 otherwise. We set M ε = {t | δuε (t) = 0}, Mε = [t0 , tf ]\Mε . Since |δuε | ≥ ε on M ε , we have tf tf ε ε |δu | dt ≤ |δu| dt → 0. ε meas M ≤ t0 t0 meas Mε Therefore, for any fixed ε, we have → 0. But then we can choose a subsequence ε → +0 such that meas M ε → 0. (Recall that M ε is defined by a member of the sequence {δu} and the corresponding member of the sequence {ε}; when defining Mε , we take the members of the sequences {δu} and {ε} with the same numbers.) Fix such a sequence t {ε}. Since δuε ∞ ≤ O(1), we have (t, u0 + δuε ) ∞ ≤ O(1). Therefore, t0f (t, u0 + δuε ) dt ≤ (t, u0 +δuε ) ∞ meas Mε → 0. Moreover, the condition ε → +0 implies {δuε } ∈ tf 0 0 loc u , and therefore, (t, u + δuε ) ∞ → 0, which implies t0 (t, u + δuε ) dt → 0. Then tf tf tf (t, u0 + δu) dt = (t, u0 + δuε ) dt + (t, u0 + δuε ) dt → 0; t0 t0 t0 this is what was required to be proved. Therefore, we have shown that the functional γ is -continuous at zero. Therefore, γ is an order. Moreover, γ is a strict order. Let us verify that γ is a higher order. Let {δw} ∈ , and let {w̄} ∈ 0 . We need to show that γ (δw + w̄) = γ (δw) + o( w̄ ), where w̄ = x̄ 1,1 + ū ∞ . Since δx + x̄ 2C = δx 2C + o( x̄ C ) and x̄ C ≤ x̄ 1,1 , it suffices to show that tf tf (t, u0 + δu + ū) dt = (t, u0 + δu) dt + o( ū ∞ ). t0 t0 As above, represent {δu} in the form {δu} = {δuε } + {δuε }, where ε → 0, meas M ε → 0. 2.3. Local Sequences, Representation of the Lagrange Function Then tf (t, u0 + δu + ū) dt = t0 = tf (t, u0 + δuε + ū) dt + Mε Mε (t, u0 + δu) dt + t0 53 (t, u0 + δuε + ū) dt ((t, u0 + δuε + ū) Mε − (t, u0 + δuε )) dt ((t, u0 + δuε + ū) − (t, u0 + δuε )) dt. + Mε Here, we have used the relations (t, u0 + δu) = (t, u0 + δu)(χε + χ ε ) = (t, u0 + δuε ) + (t, u0 + δuε ), where χε and χ ε are the characteristic functions of the sets Mε and M ε , respectively. By property (5) of Definition 2.17, we have (t, u0 + δuε + ū) − (t, u0 + δuε ) dt ≤ const(meas Mε ) ū ∞ = o( ū ∞ ). Mε Therefore, it suffices to show that ((t, u0 + δuε + ū) − (t, u0 + δuε )) dt = o( ū ∞) or, which is the same, tf ((t, u0 + δuε + ūε ) − (t, u0 + δuε )) dt = o( ū ∞ ), Mε t0 where ūε = ūχε . As was already noted, {δuε } ∈ loc u . Moreover, ūε 0u . Therefore, it suffices to prove the following assertion. ∞ (2.63) → 0, i.e., {ūε } ∈ 0 Proposition 2.18. The following estimate holds for any {δu} ∈ loc u and {ū} ∈ u : tf tf (t, u0 + δu + ū) dt = (t, u0 + δu) dt + o( ū ∞ ). t0 t0 Proof. Represent {δu} in the canonical form {δu} = {δu0 } + {δu∗ }, {δu0 } ∈ 0u , {δu∗ } ∈ ∗u , |δu0 |· |δu∗ | = 0. The latter property holds for all members of the sequences {δu0 } and {δu∗ } with the same numbers. According to the definition of the function (t, u), we have tf tf 0 0 2 (t, u + δu) dt = |δu | dt + (2|δtk | + |δvk∗ |2 ) dt. t0 Let M = [t0 , tf ]\M∗ . Then tf (t, u0 + δu + ū) dt t0 = t0 = |δu + ū| dt + 0 M tf t0 The proposition is proved. k 2 Mk∗ k Mk∗ (t, u + δu) dt + o( ū 0 (2|δtk | + |δvk∗ + ū|2 ) dt ∞ ). 54 Chapter 2. Quadratic Conditions in the Calculus of Variations According to Proposition 2.18, formula (2.63) holds. Therefore, γ is a higher order on . Obviously, γ is a strict order on . We will perform decoding of the constant Cγ := Cγ (0 , σ γ ) precisely with this order. According to Theorem 2.8, the inequality Cγ ≥ 0 is necessary for the Pontryagin minimum at the point w0 , and the strict inequality Cγ > 0 is sufficient for the strict Pontryagin minimum at this point. As was already mentioned, the decoding of the basic constant Cγ consists of the following two stages: estimating Cγ from above and estimating Cγ from below. 2.4 2.4.1 Estimation of the Basic Constant from Above Passing to Local Sequences and Needles Recall that 0 , σ γ = {{δw} ∈ | σ ≤ O(γ )}, γ σ = max{F0 (p0 + δp), . . . , Fd(F ) (p0 + δp), |δK|, δ ẋ − δf 1 }. Cγ = Cγ (0 , σ γ ) = inf lim σ γ We will estimate Cγ from above. Since Cγ ≥ 0 is a necessary condition for the Pontryagin minimum, the nonnegativity of any upper estimate for Cγ is also a necessary condition for the Pontryagin minimum. Therefore, this stage of decoding can be considered as obtaining a necessary condition for the Pontryagin minimum. Let t ∈ (t0 , tf )\, ε > 0, and let [t − ε, t + ε] be entirely contained in one of the intervals of the set (t0 , tf )\. Let a point u ∈ Rd(u) be such that (t , x 0 (t ), u ) ∈ Q, u = u0 (t ). Define the needle-shaped variation u − u0 (t), t ∈ [t − ε, t + ε], δu = δu (t; t , ε, u ) = 0 otherwise. Consider a sequence of needle-shaped variations {δu } := {δu (· , t , n1 , u )}, enumerated by the parameter ε so that ε = εn = 1/n. Clearly, {δw } := {(0, δu )} is a Pontryagin sequence. Obviously, γ := γ (δw ) = [t −ε,t +ε] (t, u ) dt is of order ε. We denote by the set of sequences {δw } = {(0, δu )} such that {δu } is a sequence of needle-shaped variations. We loc set loc σ γ = ∩ σ γ . Therefore, loc | σ ≤ O(γ )}. loc σ γ = {{δw} ∈ We have the following assertion. loc Lemma 2.19. Let {δw loc } ∈ loc σ γ and {δw } ∈ be such that γ ≤ O(γ ), where γ = loc loc γ (δw ) and γ = γ (δw ). Then lim max 0 (λ, δw loc ) + tf t0 δ H λ dt γ loc + γ where δ H λ = H (t, x 0 , u0 + δu , ψ) − H (t, x 0 , u0 , ψ). ≥ Cγ , 2.4. Estimation of the Basic Constant from Above 55 To prove Lemma 2.19, we need the following proposition. Proposition 2.20. Let ϕ(t, w) : Q → Rd(ϕ) be a continuous function. Let {δwloc } ∈ loc , {δw } ∈ , {δw} = {δw loc + δw }. Then δϕ = δ loc ϕ + δ ϕ + rϕ , where rϕ 1 = o(γ ), rϕ ∞ → 0, δϕ = ϕ(t, w 0 + δw) − ϕ(t, w0 ), δ loc ϕ = ϕ(t, w0 + δwloc ) − ϕ(t, w 0 ), δ ϕ = ϕ(t, w0 + δw ) − ϕ(t, w0 ). Proof. The following relations hold: = ϕ(t, w0 + δwloc + δw ) − ϕ(t, w 0 + δw loc ) + δ loc ϕ = δ̄ ϕ + δ loc ϕ = δ̄ ϕχ + δ loc ϕ, δϕ where δ̄ ϕ = ϕ(t, w 0 + δwloc + δw ) − ϕ(t, w0 + δw loc ) and χ is the characteristic function of the set M = {t | δu = 0}. Further, let {δwloc } = {δw 0 } + {δw∗ } be the canonical representation, where {δw 0 } ∈ 0 , {δw∗ } ∈ ∗ , and |δu0 |· |δu∗ | = 0. It follows from the definitions of sequences {δw } and {δw∗ } that |δu |· |δu∗ | = 0 starting from a certain number. Therefore, (δ̄ ϕ)χ = ϕ(t, w0 + δw 0 + δw ) − ϕ(t, w0 + δw 0 ) χ = ϕ(t, w0 + δw 0 + δw ) − ϕ(t, w0 + δw ) − δ 0 ϕ + δ ϕ χ = rϕ + δ ϕ, where = (δ̄ 0 ϕ − δ 0 ϕ)χ , = ϕ(t, w 0 + δw + δw 0 ) − ϕ(t, w 0 + δw ), = ϕ(t, w 0 + δw 0 ) − ϕ(t, w0 ). rϕ δ̄ 0 ϕ δ0ϕ Therefore, δϕ = δ loc ϕ + δ ϕ + rϕ . Since δ̄ 0 ϕ ∞ → 0, δ 0 ϕ ∞ → 0, meas M = O(γ ), we have rϕ ∞ ≤ δ̄ 0 ϕ ∞ + δ 0 ϕ ∞ → 0, rϕ 1 ≤ rϕ ∞ meas M = o(γ ). The proposition is proved. Proposition 2.20 implies the following assertion. Proposition 2.21. Let {δwloc } ∈ loc , {δw } ∈ , {δw} = {δwloc + δw }. Then for any t λ ∈ 0 , we have (λ, δw) = (λ, δw loc ) + t0f δ H λ dt + ρ λ , where sup0 |ρ λ | = o(γ ). Moreover, γ = γ loc + γ + o(γ ), where γ = γ (δw), γ loc = γ (δwloc ), and γ = γ (δw ). Finally, δ ẋ − δf 1 = δ ẋ − δ loc f 1 + O(γ ). Proof. By Proposition 2.20, we have tf ψδ ẋ dt + (λ, δw) = δl λ − + t0 tf tf ψδ f dt + tf ψδf dt = δ loc l λ − t0 t0 tf t0 ψrf dt = (λ, δw t0 loc )+ ψδ ẋ dt + = γ loc + γ + o(γ ). ψδ loc f dt t0 tf δ H λ dt + ρ λ , t0 where sup0 |ρ λ | ≤ sup0 ψ ∞ rf 1 = o(γ ). Further, tf tf δ dt = δx 2C + δ loc dt + γ (δw) = δx 2C + t0 tf t0 tf t0 δ dt + o(γ ) 56 Chapter 2. Quadratic Conditions in the Calculus of Variations Finally, δ ẋ − δf 1 = δ ẋ − δ loc f − δ f − rf 1 = δ ẋ − δ loc f 1 + O(γ ), since δ f const meas M = O(γ ) and rf 1 = o(γ ). The proposition is proved. 1 ≤ Proof of Lemma 2.19. Let {δw} = {δw loc } + {δw }, where {δw loc } ∈ loc σ γ , {δw } ∈ loc loc and γ ≤ O(γ ). Then, according to Proposition 2.21, γ = γ + γ + o(γ ). However, γ ≤ O(γ loc ). Therefore, γ ≤ O(γ loc ). On the other hand, since γ = γ loc + (1 + o(1))γ , γ ≥ 0, we have γ loc ≤ O(γ ). Therefore, γ and γ loc are of the same order of smallness. Obviously, {δw} ∈ . Let us show that {δw} ∈ σ γ . Indeed, by Proposition 2.21, δ ẋ − δf 1 = δ ẋ − δ loc f 1 + O(γ ). Therefore, = max{Fi (p 0 + δp), |δK|, δ ẋ − δf 1 } ≤ σ (δw loc ) + O(γ ) ≤ O1 (γ loc ) ≤ O2 (γ ). σ (δw) Thus, {δw} ∈ σ γ . Further, according to Proposition 2.21, tf δ H λ dt + ρ λ = (λ, δw), (λ, δwloc ) + t0 where sup0 |ρ λ | = o(γ ) = o1 (γ ). Therefore, lim max 0 (λ, δw loc ) + tf t0 δ H λ dt γ loc + γ = lim max 0 (λ, δw) 0 0 = lim ≥ inf lim = Cγ . σ γ γ γ γ The inequality holds, since {δw} ∈ σ γ . The lemma is proved. We can now use the results of Section 2.3. Lemma 2.19 and Proposition 2.15 imply the following assertion. Lemma 2.22. Let {δw} ∈ loc σ γ and {δw } ∈ be such that γ ≤ O(γ ), where γ = γ (δw ) and γ = γ (δw). Then lim max 0 1λ (δw) + tf t0 δ H λ dt γ +γ ≥ Cγ , where δ H λ = H (t, x 0 , u0 + δu , ψ) − H (t, x 0 , u0 , ψ) and the function 1λ (δw) is defined by formula (2.61). 2.4.2 Replacement of the Functions in the Definition of the Set loc σγ by their Decompositions up to First-Order Terms We represent the sequences {δw} ∈ loc in the canonical form: {δw} = {δw0 } + {δw ∗ }, where {δw0 } = {(δx, δu0 )} ∈ 0 , {δw ∗ } = {0, δu∗ } ∈ ∗ , and, moreover |δu0 |· |δu∗ | = 0. We set I = {i ∈ {0, 1, . . . , d(F )} : Fi (p0 ) = 0} = IF (w 0 ) ∪ {0}. Then loc σ γ consists of seloc quences {δw} ∈ such that δFi ≤ O(γ ), i ∈ I ; |δK| ≤ O(γ ); δ ẋ − δf 1 ≤ O(γ ). (2.64) (2.65) (2.66) 2.4. Estimation of the Basic Constant from Above 57 Since δFi = Fip δp + O(|δp|2 ), i ∈ I ; δK = Kp δp + O(|δp|2 ); |δp|2 ≤ 2 δx 2 C ≤ 2γ , conditions (2.64) and (2.65) are equivalent to the conditions Fip δp ≤ O(γ ), i ∈ I ; |Kp δp| ≤ O(γ ), (2.67) (2.68) respectively. Further, consider condition (2.66). By Proposition 2.10, 1 δf = fw δw0 + fww δw0 , δw0 + δ ∗ f + δ ∗ fx δx + r̃, 2 tf 2 0 where r̃ 1 = o δx C + t0 |δu |2 dt = o1 (γ ). Here, fww δw 0 , δw0 (δ ∗ fx )δx 1 1 ≤ O( δx ≤ const δx 2 C + tf |δu0 |2 dt) ≤ O1 (γ ), t0 C meas M ∗ ≤ 1 const( δx 2 2 C + (meas M∗ )2 ). But, as was already mentioned, (meas Mk∗ )2 ≤ 4 M∗ |δtk | dt ≤ 2γ . Therefore, (δ ∗ fx )δx k O(γ ). Further, δ ∗ f = k δk∗ f . According to formula (2.50), we have 1≤ ∗ ∗ δk∗ f = [f ]k (χk− − χk+ ) + fuk+ δvk− + fuk− δvk+ + O(|δtk | + |δvk |2 )χk∗ . Therefore, condition (2.66) is equivalent to the following condition: δ ẋ − fw δw 0 − ∗ ∗ ([f ]k (χk− − χk+ ) − fuk+ δvk− − fuk− δvk+ ) ≤ O(γ ). (2.69) 1 k loc such that conditions (2.67)– We have shown that loc σ γ consists of sequences {δw} ∈ (2.69) hold for their canonical representations. 2.4.3 Narrowing the Set of Sequences loc σγ In what follows, in the formulation of Lemma 2.22, we narrow the set loc σ γ up to its subset defined by the following conditions: (a) δv = 0; (b) for any λ ∈ 0 , ∗ ∗ [H λ ]k (meas Mk− − meas Mk+ ) ≤ 0. (2.70) k These conditions should hold for each member δw of the sequence {δw}. We denote by 1 loc loc σ γ the set of sequences {δw} ∈ σ γ satisfying these conditions. 1 For any sequence {δw} ∈ loc σ γ , we obviously have 1λ (δw) ≤ 2λ (δw) ∀ λ ∈ 0 , (2.71) 58 Chapter 2. Quadratic Conditions in the Calculus of Variations where 1λ is as defined in (2.61) and 1 (λ, δw 0 ) 2 tf ∗ ∗ D k (H λ ) + |δtk | dt + [Hxλ ]k δx(χk− − χk+ ) dt , ∗ M t 0 k tf k λ λ = lpp δp, δp + Hww δw0 , δw0 dt. 2λ (δw) = (λ, δw0 ) t0 Moreover, for any sequence {δw} ∈ loc1 σγ , γ1 (δw) = δx 2 C + tf we have γ (δw) = γ1 (δw), where |δu0 |2 dt + 2 |δtk | dt. t0 Mk∗ k Finally, condition (2.69) passes to the following condition on these sequences: ∗ ∗ δ ẋ − fw δw0 − [f ]k (χk− − χk+ ) ≤ O(γ1 ). 1 k ∗ 0 We note that in the definitions of 2 , γ1 , and loc1 σ γ , only δw and Mk participate and, moreover, the variation δw is uniquely reconstructed by Mk∗ and δw0 by using the conditions δu0 χ ∗ = 0, δv = 0. We denote the pairs (δw, M ∗ ) by b. Introduce the set of sequences of pairs {b} = {(δw, M ∗ )} such that Mk∗ , Mk∗ → tk (k ∈ I ∗ ), δuχ ∗ = 0, {δw} = {(δx, δu)} ∈ 0 , M∗ = k∈I ∗ Fip δp ≤ O(γ1 ) (i ∈ I ), |Kp δp| ≤ O(γ1 ), ∗ ∗ [f ]k (χk− − χk+ ) ≤ O(γ1 ), δ ẋ − fw δw − 1 k λ k ∗ ∗ [H ] (meas Mk− − meas Mk+ ) ≤ 0 ∀ λ ∈ 0 . k As above, we denote this set of sequences by loc1 σ γ . In what follows, we denote by {δw} the sequences from 0 . Lemma 2.22 and inequality (2.71) imply the following assertion. Lemma 2.23. Let {b} = {(δw, M ∗ )} ∈ loc1 σ γ and {δw } ∈ be such that γ ≤ O(γ1 ), where tf 2 2 |δu| dt + 2|δtk | dt, γ = γ (δw ). γ1 = γ1 (b) := δx C + t0 k Then, lim max 2λ + tf t0 δ H λ dt γ1 + γ 0 Mk∗ ≥ Cγ , where 2λ = 1 2λ (b) := (λ, δw) 2 k λ λ k D (H ) + |δtk | dt + [Hx ] k δ Hλ := Mk∗ H (t, x 0 , u0 + δu , ψ) − H (t, x 0 , u0 , ψ). tf t0 ∗ ∗ δx(χk− − χk+ ) dt , 2.4. Estimation of the Basic Constant from Above 2.4.4 59 Replacement of δx2C by |δx(t0 )|2 in the Definition of Functional γ1 We set γ2 = |δx(t0 )|2 + tf |δu|2 dt + t0 k Mk∗ 2|δtk | dt. Since |δx(t0 )| ≤ δx C , we have γ2 ≤ γ1 . Let us show that the following estimate also holds on the sequence from loc1 σ γ : γ1 ≤ const γ2 , where const > 0 is independent of the sequence. For this purpose, it suffices to show that δx 2C ≤ const γ2 on a sequence from loc1 σ γ . Let us prove the following assertion. Proposition 2.24. There exists a const > 0 such that for any sequence {b} = {δw, M ∗ } satisfying the conditions M∗ = ∪k Mk∗ , Mk∗ → tk (k ∈ I ∗ ), √ ∗ ∗ δ ẋ − fw δw − [f ]k (χk− − χk+ ) = o( γ1 ) {δw} ∈ 0 , (2.72) 1 k starting from a certain number the following inequality holds: δx 2 C ≤ const γ2 . Proof. Let {b} satisfy conditions (2.72). Then ∗ ∗ [f ]k (χk− − χk+ ) + r, δ ẋ = fx δx + fu δu + k √ where r 1 = o( γ1 ). As is known, this implies the estimate δx 1,1 ≤ |δx(t0 )| + const fu δu + ∗ ∗ [f ]k (χk− − χk+ )+r Since √ . 1 k ∗ − χ∗ χk− ≤ meas Mk∗ , k+ 1 ∗ )2 + (meas M ∗ )2 ≤ 2 |δtk | dt ≤ γ2 , k ∈ I ∗ , (meas Mk− k+ δu 1 ≤ tf − t0 δu 2 , Mk∗ we have δx 2C ≤ δx sition is proved. 2 1,1 ≤ const γ2 + o(γ1 ). This implies what was required. The propo- Therefore, γ1 and γ2 are equivalent on the set of sequences satisfying conditions (2.72), i.e., on each such sequence, they estimate one another from above and from below: γ2 ≤ γ1 ≤ γ2 ( > 1) (2.73) (the constant is independent of the sequence). In particular, inequalities (2.73) hold on loc1 sequences from loc1 σ γ . First, this implies that the set σ γ does not change if we replace γ1 by γ2 in its definition. Further, inequality (2.73) implies the inequalities γ2 + γ ≤ γ1 + γ ≤ (γ2 + γ ), whence γ1 + γ ≤ . (2.74) 1≤ γ2 + γ 60 Chapter 2. Quadratic Conditions in the Calculus of Variations Let {b} ∈ loc1 σ γ , {δw } ∈ , and let γ ≤ O(γ2 ). Then, we obtain from (2.74) and Lemma 2.23 that lim max 0 2λ + tf t0 δ H λ dt γ2 + γ t 2λ + t0f δ H λ dt γ1 + γ = lim max γ2 + γ 0 γ1 + γ Cγ , Cγ ≥ 0, ≥ min{Cγ , Cγ } = Cγ , Cγ < 0. We have proved the following assertion. Lemma 2.25. The following inequality holds for any {b} ∈ loc1 σ γ and {δw } ∈ such that γ ≤ O(γ2 ): t 2λ (b) + t0f δ H λ dt lim max ≥ min{Cγ , Cγ }. 0 γ2 (b) + γ 2.4.5 Passing to Sequences with Discontinuous State Variables Denote by P W 1,1 (, Rd(x) ) the space of functions x̄ : [t0 , tf ] → Rd(x) piecewise continuous on [t0 , tf ] and absolutely continuous on each of the intervals of the set (t0 , tf ) \ (points of discontinuity of such functions are possible only at points of the set ). The differential constraint in the set loc1 σ γ is represented by the condition δ ẋ − fw δw − ∗ ∗ [f ]k (χk− − χk+ ) ≤ O(γ2 ). (2.75) 1 k ∗ − χ ∗ ) on δx in this condition? We What is the influence of the terms k [f ]k (χk− k+ 1,1 show below that the variations δx ∈ W (, Rd(x) ) can be replaced by variations ∗ − meas M ∗ , k ∈ I ∗ , x̄ ∈ P W 1,1 (, Rd(x) ) such that [x̄]k = [f ]k ξk , where ξk = meas Mk− k+ and, moreover, (2.75) passes to the condition x̄˙ − fx x̄ − fu δu 1 ≤ O(γ2 ). We will prove a slightly more general assertion, which will be used later in estimating Cγ from below. Therefore, we assume that there is a sequence {b} = {(δw, M ∗ )} such that {δw} ∈ 0 , ∗ M = ∪Mk∗ , Mk∗ → tk , k ∈ I ∗ . Moreover, let the following condition (which is weaker than (2.75)) hold: √ ∗ ∗ δ ẋ = fx δx + fu δu + [f ]k (χk− − χk+ ) + r, r 1 = o( γ2 ). (2.76) k For each member b = (δx, δu, M∗ ) of the sequence {b}, let us define the functions δxk∗ and x̄k∗ by the following conditions: ∗ − χ ∗ ), δx ∗ (t ) = 0, x̄˙ ∗ = 0, δ ẋk∗ = [f ]k (χk− k+ k 0 k k ∗ ∗ − meas M ∗ . k [x̄k ] = [f ] ξk , ξk = meas Mk− k+ x̄k∗ (t0 ) = 0, (2.77) 2.4. Estimation of the Basic Constant from Above 61 Therefore, x̄k∗ is the jump function: x̄k∗ (t) = 0 if t < tk and x̄k∗ (t) = [f ]k ξk if t > tk , and, moreover, the value of the jump is equal to [f ]k ξk . We set ∗ ∗ ∗ ∗ δx k = x̄k∗ − δxk∗ , k ∈ I ∗ , δx = δx k , x̄ = δx + δx = δx + (x̄k∗ − δxk∗ ). k k Note that x̄ ∈ P W 1,1 (, Rd(x) ). Since the functions x̄k∗ and δxk∗ coincide outside Mk∗ , we ∗ ∗ ∗ ∗ ∗ ∗ have δx k χk∗ = δx k for all k. Hence δx χ ∗ = δx . Let us estimate δx ∞ and δx 1 . We have ∗ δx k ∞ ≤ x̄k∗ ∞ + δxk∗ ∞ ≤ [f ]k · ξk + [f ]k meas Mk∗ . Moreover, ξk ≤ meas M ∗ ≤ 4 k ∗ √ ∗ δx χ ∗ Mk∗ δtk dt 1 2 ≤ 2γ2 . ∗ ∗ ∗ = δx , we have δx 1 ≤ δx ∞ meas M∗ ≤ Hence δx ∞ ≤ const γ2 . Since const γ2 . What equation does x̄ satisfy? We obtain from (2.76) and (2.77) that ∗ x̄˙ = fx δx + fu δu + r = fx x̄ + fu δu − fx δx + r. ∗ Since δx 1 ≤ O(γ2 ), we have x̄˙ − fx x̄ − fu δu 1 ≤ O(γ2 ) + r 1 . Note that the replacement of δx by x̄ does not influence the value of γ2 , since x̄(t0 ) = δx(t0 ) and M ∗ and δu are preserved. Now let us show that tf ∗ ∗ k δx(χk− − χk+ ) dt = x̄av ξk + o(γ2 ), (2.78) t0 ∗ − meas M ∗ where ξk = meas Mk− k+ k = 1 (x̄ k− + x̄ k+ ) = 1 (x̄(t −) + x̄(t +)). Recall and x̄av k k 2 2 0 + δx ∗ + x , where that δx satisfies equation (2.76). Represent δx in the form δx = δx r ẋr = r, xr (t0 ) = 0, and δx ∗ = δxk∗ . Then δ ẋ 0 = fx δx + fu δu and δx 0 (t0 ) = δx(t0 ). This and the conditions δx C → 0, δu ∞ → 0 imply δ ẋ 0 ∞ → 0. t ∗ − χ ∗ ) dt. We set δx 0 := δx 0 + x . Since Now let us consider t0f δx(χk− r r k+ δx = (δx − δx ∗ ) + δx ∗ = δx 0 + xr + δx ∗ = δxr0 + δx ∗ , we have tf t0 ∗ ∗ δx(χk− − χk+ ) dt = tf t0 ∗ ∗ δxr0 (χk− − χk+ ) dt + tf t0 ∗ ∗ δx ∗ (χk− − χk+ ) dt. (2.79) Let us estimate each summand separately. We have tf tf 0 ∗ ∗ ∗ ∗ δxr (χk− − χk+ ) dt = δxr0 (tk )(χk− − χk+ ) dt t0 t0 + + tf t0 tf t0 = ∗ ∗ (δx 0 (t) − δx 0 (tk ))(χk− − χk+ ) dt ∗ ∗ (xr (t) − xr (tk ))(χk− − χk+ ) dt δxr0 (tk )ξk + o(γ2 ). (2.80) 62 Chapter 2. Quadratic Conditions in the Calculus of Variations Here, we have used the following estimates: √ √ (a) xr ∞ ≤ xr 1,1 = ẋr 1 = r 1 = o( γ2 ), meas Mk∗ ≤ 2γ2 , whence tf t0 ∗ ∗ (xr (t) − xr (tk ))(χk− − χk+ ) dt = o(γ2 ); (b) δx 0 (t) − δx 0 (tk ) ≤ δx˙ 0 tf t0 since δ ẋ 0 ∞ (2.81) δtk , and hence ∗ ∗ (δx 0 (t) − δx 0 (tk ))(χk− − χk+ ) dt ≤ δ ẋ 0 ∞ Mk∗ δtk dt = o(γ2 ), (2.82) → 0. Relation (2.80) follows from (2.81) and (2.82). Further, the conditions x̄˙ = fx δx + fu δu + r, δ ẋr0 = fx δx + fu δu + r, 0 δxr (t0 ) = δx(t0 ), x̄(t0 ) = δx(t0 ) ∞ imply δxr0 = x̄ − k x̄k∗ outside , and hence δxr0 (tk ) = x̄ k− − j <k [x̄]j . We obtain from this and (2.80) that ∗ ∗ δxr0 (χk− − χk+ ) dt = x̄ k− − [x̄]j ξk + o(γ2 ). tf t0 (2.83) j <k ∗ − χ ∗ , y ∗ (t ) = 0. Then Further, let yk∗ (t) be defined by the conditions ẏk∗ = χk− k+ k 0 tf t0 1 1 ∗ ∗ 2 yk∗ ẏk∗ dt = (meas Mk− − meas Mk+ ) = ξk2 . 2 2 Obviously, δxk∗ = [f ]k yk∗ . Hence tf t0 ∗ ∗ δxk∗ (χk− − χk+ ) dt = tf t0 1 1 1 [f ]k yk∗ ẏk∗ dt = [f ]k ξk2 = [x̄k∗ ]k ξk = [x̄]k ξk . 2 2 2 We obtain from this that tf ∗ ∗ δx ∗ (χk− − χk+ ) dt t0 tf ∗ ∗ = δxj∗ (χk− − χk+ ) dt + t0 j <k = 1 [x̄]j ξk + [x̄]k ξk , 2 tf t0 ∗ ∗ δxk∗ (χk− − χk+ ) dt j <k since we have the following for j < k: δxj∗ χk∗ = x̄j∗ χk∗ = [x̄j∗ ]j χk∗ = [x̄]j χk∗ . (2.84) 2.4. Estimation of the Basic Constant from Above 63 We obtain from (2.79), (2.83), and (2.84) that tf 1 ∗ ∗ k δx(χk− − χk+ ) dt = x̄ k− + [x̄]k ξk + o(γ2 ) = x̄av ξk + o(γ2 ), 2 t0 as required. Finally, let us show that tf t0 λ Hww w̄, w̄ dt = tf t0 λ Hww δw, δw dt + ρ λ , (2.85) ∗ ∗ where w̄ = (x̄, ū), ū = δu, and sup0 ρ λ = o(γ2 ). Indeed, x̄ = δx + δx , where δx ∗ 0, δx 1 = O(γ2 ). Hence tf t0 λ Hww w̄, w̄ dt tf t0 tf = = = λ λ λ Hxx δx, δx + 2Hux δx, δu + Huu δu, δu dt tf t0 tf t0 → λ λ λ Hxx x̄, x̄ + 2Hux x̄, ū + Huu ū, ū dt t0 + ∞ λ 2Hxx δx, δx ∗ ∗ ∗ λ λ + 2Hux δx , δu + Hxx δx , δx ∗ dt supρ λ = o(γ2 ), λ Hww δw, δw dt + ρ λ , 0 since t f λ ∗ λ Hxx δx, δx dt ≤ sup Hxx 0 t0 t f λ ∗ ≤ sup H λ H δx , δu dt ux ux 0 t0 t f λ ∗ ∗ λ Hxx δx , δx dt ≤ sup Hxx 0 t0 ∗ ∞ δx C δx ∞ δu ∞ δx ∞ δx ∗ ∞ 1 = o(γ2 ), 1 = o(γ2 ), ∗ δx ∗ 1 = o(γ2 ). Therefore, formula (2.85) holds. We set ∗ − meas M ∗ , b̄ = (w̄, M ∗ ); ξk = meas Mk− k+ tf λ λ w̄, w̄ dt, (λ, w̄) = lpp p̄, p̄ + t0 Hww where p̄ = (x̄(t0 ), x̄(tf )); 1 δtk dt + [H λ ]k x̄ k ξk ; 3λ (b̄) = (λ, w̄) + D k (H λ ) x av ∗ 2 M k k tf |ū|2 dt + 2 |δtk | dt. γ2 (b̄) = |x̄(t0 )|2 + t0 k Mk∗ (2.86) 64 Chapter 2. Quadratic Conditions in the Calculus of Variations Since p̄ = δp for the entire sequence {b}, we obtain from (2.78) and (2.85) that 2λ (b) = 3λ (b̄) + ρ λ , where sup0 |ρ λ | = o(γ2 ). We have proved the following assertion. Lemma 2.26. Let a sequence {b} = {(δw, M∗ )} be such that {δw} = {(δx, δu)} ∈ 0 , M∗ = ∪Mk∗ , Mk∗ → tk , k ∈ I ∗ , and , moreover, let δ ẋ = fx δx + fu δu + ∗ ∗ [f ]k (χk− − χk+ ) + r, k √ where r 1 = o( γ2 ). Let a sequence {b̄} = {(w̄, M∗ )} be such that w̄ = (x̄, ū), x̄ = ∗ ∗ δx + δx , and ū = δu, where δx = x̄ ∗ − δx ∗ = x̄k∗ − δxk∗ , and x̄k∗ and δxk∗ are defined by formulas (2.77). Let δp = (δx(t0 ), δx(tf )) and p̄ = (x̄(t0 ), x̄(tf )). Then ∗ {δp} = {p̄}, γ2 (b) = γ2 (b̄), x̄˙ = fx x̄ + fu ū − fx δx + r, ∗ ∗ − meas M ∗ ; [x̄]k = [f ]k ξk ∀ k, ξk = meas Mk− δx k+ 2λ (b) = 3λ (b̄) + ρ λ , sup0 |ρ λ | = o(γ2 ), 1 ≤ O(γ2 ); where 3λ (b̄) is defined by formula (2.86). We will need this lemma in estimating Cγ from below. We now use the corollary of Lemma 2.26, which is formulated below. Corollary 2.27. Let a sequence {b̄} = {(w̄, M ∗ )} be such that w̄ = (x̄,ū), x̄ ∈ P W 1,1 (, Rd(x) ), ū ∈ L∞ (, Rd(u) ), x̄ M∗ = Mk∗ , Mk∗ → tk (k ∈ I ∗ ), x̄˙ − fw w̄ 1 ≤ O(γ2 ). ∞+ ū ∞ → 0, ∗ Let a sequence {b} = {(δw, M ∗ )} be such that δw = (δx, δu), δx = x̄ − δx , and δu = ū, ∗ where δx = x̄ ∗ − δx ∗ = x̄k∗ − δxk∗ , and x̄k∗ and δxk∗ are defined by formulas (2.77). Then {δw} ∈ 0 , {δp} = {p̄}, γ2 (b) = γ2 (b̄), ∗ ∗ δ ẋ − fw δw − [f ]k (χk− − χk+ ) ≤ O(γ2 ), 1 k 2λ (b) = 3λ (b̄) + ρ λ , sup0 |ρ λ | = o(γ2 ). Proof. Indeed, by the condition of Corollary 2.27, it follows that x̄˙ = fx x̄ + fu ū + r̃, ∗ ∗ r̃ 1 ≤ O(γ2 ). Then x̄˙ = fx x̄ + fu ū − fx δx + r, where r = r̃ + fx δx , and, moreover, ∗ r 1 ≤ r̃ 1 + fx δx 1 ≤ O(γ2 ). We obtain from this that δ ẋ = fx δx + fu δu + ∗ ∗ [f ]k (χk− − χk+ ) + r. k ∗ −χ ∗ ) Consequently, δ ẋ −fw δw − k [f ]k (χk− k+ lary 2.27 follow from Lemma 2.26 directly. 1 ≤ O(γ2 ). The other assertions of Corol- 2.4. Estimation of the Basic Constant from Above 65 Denote by S 2 the set of sequences {b̄} = {(w̄, M ∗ )} such that w̄ = (x̄, ū), ū ∈ L∞ (, Rd(u) ), x̄ ∈ P W 1,1 (, Rd(x) ), ū ∞ → 0, x̄ ∞ → 0, ∗ ∗ M = Mk , Mk∗ → tk (k ∈ I ∗ ), ūχ ∗ = 0, k Fip p̄ ≤ O(γ2 ) (i ∈ I ), |Kp p̄| ≤ O(γ2 ), x̄˙ − fw w̄ λ k ∗ ∗ [H ] (meas Mk− − meas Mk+ ) ≤ 0 ∀ λ ∈ 0 . 1 ≤ O(γ2 ), k Then Lemma 2.25 and also Corollary 2.27 imply the following assertion. Lemma 2.28. The following inequality holds for any sequences {b̄} ∈ S 2 and {δw } ∈ such that γ ≤ O(γ2 (b̄)): lim max 3λ (b̄) + tf t0 δ H λ dt γ2 (b̄) + γ 0 ≥ min{Cγ , Cγ }. Passing to the Quadratic Form 2.4.6 We remove the condition ūχ ∗ = 0 from the definition of S 2 . We denote by S 3 the resulting new set of sequences. Let us show that Lemma 2.28 remains valid under the replacement of S 2 by S 3 . Indeed, assume that there is a sequence {b̄} = {(w̄, M∗ )} ∈ S 3 , where w̄ = (x̄, ū). We set ū¯ = ū(1−χ ∗ ) = ū− ū∗ , where ū∗ = ūχ ∗ . Then ū∗ ∞ → 0 and ū∗ 1 = M∗ |ū| dt ≤ √ meas M∗ ū 2 ≤ O(γ2 ). Consequently, tf |ū∗ |2 dt ≤ ū∗ ∞ ū∗ 1 = o(γ2 ), t0 tf t0 tf t0 λ Huu ū, ū dt = λ Hux x̄, ū dt = tf t0 tf t0 λ ¯ ¯ Huu ū, ū dt + λ Hux x̄, ū¯ dt + tf t0 tf t0 λ ∗ ∗ Huu ū , ū dt = λ Hux x̄, ū∗ dt = tf t0 tf t0 λ ¯ ¯ Huu ū, ū dt + ρ1λ , λ Hux x̄, ū¯ dt + ρ2λ , ¯ = {(x̄, ū, ¯ M∗ )}. Then where sup0 |ρiλ | = o(γ2 ), i = 1, 2. We set {b̄} ¯ = γ (b̄) + o(γ ), γ2 (b̄) 2 2 ¯ = 3λ (b̄) + ρ λ , 3λ (b̄) sup |ρ λ | = o(γ2 ). (2.87) 0 ¯ ∈ S 2 . Indeed, Moreover, it is easy to see that {b̄} x̄˙ − fx x̄ − fu ū¯ 1 = ≤ x̄˙ − fx x̄ − fu ū − fu ū∗ 1 x̄˙ − fw w̄ 1 + fu ū∗ 1 ≤ O(γ2 ). ¯ ∈ S2 ¯ belonging to S 2 obviously hold. Conditions (2.87) and {b̄} The other conditions of {b̄} imply that Lemma 2.28 remains valid under the replacement of S 2 by S 3 . 66 Chapter 2. Quadratic Conditions in the Calculus of Variations Further, we narrow the set of sequences S 3 up to the set S 4 by adding the following conditions to the definition of the set S 3 : (i) Each set Mk∗ is a segment adjusting to tk , i.e., Mk∗ = [tk − ε, tk ] or Mk∗ = [tk , tk + ε], where ε → +0. In this case (see formula (2.13)), ∗ − meas M ∗ |δtk | dt = ξk2 , where ξk = meas Mk− 2 k+ Mk∗ tf γ2 (b̄) = ξk2 + |x̄(t0 )|2 + |ū|2 dt = γ̄ ; t0 3λ (b̄) = 1 1 k (λ, w̄) + D k (H λ )ξk2 + [Hxλ ]k x̄av ξk =: λ . 2 2 k (ii) Also, the following relations hold: Fip p̄ ≤ 0, i ∈ I , Kp p̄ = 0, x̄˙ = fw w̄. ∗ −meas M ∗ We note that for sequences from S 4 , each of the quantities ξk = meas Mk− k+ − + − ∗ uniquely defines the set Mk := [tk − ξk , tk + ξk ]. Here, ξk = max{0, −ξk }, and ξk+ = max{0, ξk }. Moreover, we note that and γ̄ depend on ξ̄ , x̄, and ū. Therefore, S 4 can be identified with the set of sequences {z̄} = {(ξ̄ , w̄)} such that ξ̄ ∈ Rs , w̄ = (x̄, ū), x̄ ∈ P W 1,1 (, Rd(x) ), ū ∈ L∞ (, Rd(u) ), ū ∞ → 0, |ξ̄ | → 0, x̄ ∞ → 0, Fip p̄ ≤ 0 (i ∈ I ), Kp p̄ = 0, x̄˙ = fw w̄, [x̄]k = [f ]k ξ̄k (k ∈ I ∗ ), [H λ ]k ξ̄k ≤ 0 ∀ λ ∈ 0 . (2.88) k Therefore, the following assertion holds. Lemma 2.29. The following inequality holds for any sequences {z̄} ∈ S 4 and {δw } ∈ such that γ̄ (z̄) ≤ O(γ ): λ (z̄) + tf t0 δ H λ dt ≥ min{Cγ , Cγ }. γ̄ (z̄) + γ Now let us show that the condition k [H λ ]k ξ̄k ≤ 0 for all λ ∈ 0 holds automatically for the elements of the critical cone K, and therefore, it is extra in the definition of the set of sequences S 4 , i.e., it can be removed. We thus will prove that S 4 consists of the sequences of elements of the critical cone K that satisfy the condition |ξ̄ | + x̄ ∞ + ū ∞ → 0. lim max 0 2.4.7 Properties of Elements of the Critical Cone Proposition 2.30. Let λ = (α0 , α, β, ψ) ∈ 0 , (ξ̄ , x̄, ū) ∈ K. Then the function ψ x̄ is constant on each of the intervals of the set (t0 , tf )\ and hence is piecewise constant on [t0 , tf ]. 2.4. Estimation of the Basic Constant from Above 67 Proof. The conditions x̄˙ = fw w̄, ψ̇ = −ψfx , and ψfu = 0 imply d 0 = ψ(x̄˙ − fx x̄ − fu ū) = ψ x̄˙ − ψfx x̄ = ψ x̄˙ + ψ̇ x̄ = (ψ x̄). dt Moreover, ψ ∈ W 1,∞ (, Rd(x) ) and x̄ ∈ P W 1,2 (, Rd(x) ). Therefore, ψ x̄ = const on any interval of the set (t0 , tf )\. The proposition is proved. Proposition 2.31. Let λ = (α0 , α, β, ψ) ∈ 0 , and let (ξ̄ , x̄, ū) ∈ K. Then [H λ ]k ξ̄k = lpλ p̄ ≤ 0. k Proof. By Proposition 2.30, dtd (ψ x̄) = 0 a.e. on [t0 , tf ]. Hence tf t d (ψ x̄) dt = ψ x̄ tf − [ψ x̄]k 0 = 0 dt t0 k = ψ(tf )x̄(tf ) − ψ(t0 )x̄(t0 ) − ψ(tk )[x̄]k k = lxλf x̄(tf ) + lxλ0 x̄(t0 ) − ψ(tk )[f ]k ξ̄k = lpλ p̄ − [H λ ]k ξ̄k . We obtain from this that αi ≥ 0 ∀ i ∈ I , k λ k k [H ] ξ̄k = lpλ p̄. αi = 0 ∀ i ∈ / I, k Moreover, the conditions Fip p̄ ≤ 0 ∀ i ∈ I, Kp p̄ = 0 imply lpλ p̄ ≤ 0. The proposition is proved. Proposition 2.31 implies the following assertion. Proposition 2.32. Let λ = (α0 , α, β, ψ) ∈ 0 be such that [H λ ]k = 0 for all k ∈ I ∗ . Let z̄ = (ξ̄ , x̄, ū) ∈ K. Then α0 (Jp p̄) = 0, αi (Fip p̄) = 0, i = 1, . . . , d(F ). λ k ∗ Proof. By Proposition 2.31, the conditions z̄ ∈ K, λ ∈ 0 , [H ] = 0 for all k ∈ I imply λ lp p̄ = 0, where lp p̄ = α0 (Jp p̄) + αi (Fip p̄) + βj (Kjp p̄). This and the conditions α0 ≥ 0, Jp p̄ ≤ 0, αi ≥ 0, Fip p̄ ≤ 0 ∀ i ∈ IF (w 0 ), / I := IF (w 0 ) ∪ {0}, Kp p̄ = 0 αi = 0 ∀ i ∈ imply what was required. The proposition is proved. In fact, Proposition 2.32 is equivalent to Proposition 2.2. Therefore, using Proposition 2.31, we have proved Proposition 2.2. Proposition 2.3 is proved analogously (we leave this proof to the reader). Further, we use Proposition 2.31. We set Z() = Rs × P W 1,1 (, Rd(x) ) × L∞ (, Rd(u) ). (2.89) Consider sequences of the form {ε z̄}, where ε → +0 and z̄ ∈ K ∩ Z() is a fixed element. According to Proposition 2.31, such a sequence belongs to S 4 . Therefore, Lemma 2.29 implies the following assertion. 68 Chapter 2. Quadratic Conditions in the Calculus of Variations Lemma 2.33. Let z̄ ∈ K ∩ Z(), ε → +0, {δw } ∈ , and γ ≤ O(ε 2 ). Then lim max ε2 λ (z̄) + tf t0 δ H λ dt ε2 γ̄ (z̄) + γ 0 ≥ min{Cγ , Cγ }. Below, we set K ∩ Z() = KZ . 2.4.8 Cone C in the Space of Affine Functions on the Compact Set co 0 To obtain the next upper estimate of the basic constant Cγ , in the estimate of Lemma 2.33, we need to replace the set 0 by the set M co (C) of tuples satisfying the minimum principle of “strictness C” (the meaning of this will be explained later) and, simultaneously, to remove the sequence of needle-shaped variations. We will attain this using the method of Milyutin, which has become a standard method [27, 86, 92, 95] for this stage of decoding higher-order conditions in problems of the calculus of variations and optimal control. The method of Milyutin is called the “cone technique.” The meaning of such a term will become clear after its description. We now give several definitions. Denote by L0 = Lin(co 0 ) the linear span of the set co 0 . Since 0 is a finitedimensional compact set, L0 is also of finite dimension. We denote by l(λ) an arbitrary linear function l : L0 → R1 , and by a(λ) an arbitrary affine function a : L0 → R1 , i.e., a function of the form a(λ) = l(λ) + c, where c is a number. Since L0 is of finite dimension, the affine functions a(λ) : L0 → R1 compose a finite-dimensional space, and each of these functions is uniquely defined by its values on co 0 . In what follows, we consider affine functions a(λ) on the compact set co 0 (of course they can be defined directly on co 0 , not passing to the linear span). With each sequence {δw } ∈ , we associate the following sequence of linear functions on co 0 : tf λ t0 δ H dt l(λ) = , γ t where δ H λ = H λ (t, w 0 + δw ) − H λ (t, w0 ), γ = t0f dt, and = (t, u0 + δu ). It follows from the definition of the sequence {δw } ∈ that γ > 0 on it, and, therefore, this definition is correct. Let A0 be the set of all limit points of the sequences {l(λ)} obtained by the above method from all sequences {δw } ∈ . Clearly, A0 is a closed subset in the finitedimensional space of linear functions l(λ) : co 0 → R1 . We note that each convergent sequence {l(λ)} converges to its limit uniformly on co 0 . Further, to each number C, we associate the set of affine functions AC obtained from A0 by means of the shift by (−C): AC = {a(λ) = l(λ) − C l(· ) ∈ A0 }. Denote by C := con AC the cone spanned by the set AC . Fix an arbitrary element z̄ ∈ KZ , z̄ = 0 (recall that KZ := K ∩ Z()). Proposition 2.34. Let a number C be such that lim max 0 ε 2 λ (z̄) + tf t0 δ H λ dt ε2 γ̄ (z̄) + γ ≥C (2.90) 2.4. Estimation of the Basic Constant from Above 69 for any pair of sequences {ε}, {δw } such that ε → +0, {δw } ∈ , γ ≤ O(ε 2 ). Then inf max{λ (z̄) + a(λ)} ≥ C γ̄ (z̄). a∈C co 0 (2.91) Proof. Let a(· ) ∈ C , i.e., a(λ) = ρ(l(λ) − C), where ρ > 0, l(· ) ∈ A0 . The latter means that there exists {δw } ∈ such that tf t0 δ H λ dt γ → l(λ) ∀ λ ∈ co 0 . (2.92) To the sequence {δw }, we write the sequence {ε} = {ε(δw )} of positive numbers such that tf γ ε 2 = , where γ = (t, u0 + δu ) dt. (2.93) ρ t0 Then γ = O(ε2 ), and, therefore, inequality (2.90) holds. This inequality and conditions (2.92) and (2.93) easily imply max 0 λ (z̄) ρ + l(λ) γ̄ (z̄) ρ +1 ≥ C. Clearly, the maximum over 0 in this inequality can be replaced by the maximum over co 0 , since λ and l(λ) are linear in λ. Therefore, multiplying the inequality by γ̄ (z̄) + ρ, we obtain max(λ (z̄) + ρl(λ)) ≥ C(γ̄ (z̄)) + ρ). co 0 Since ρ(l(λ) − C) = a(λ), this implies max(λ (z̄) + a(λ)) ≥ C γ̄ (z̄). co 0 (2.94) It remains to recall that a(· ) is an arbitrary element of C , and hence (2.94) implies (2.91). The proposition is proved. Lemma 2.33 and Proposition 2.34 imply the following assertion. Lemma 2.35. Let Cγ > −∞. Then for any C ≤ min{Cγ , Cγ } and for any z̄ ∈ KZ , we have the inequality inf max {λ (z̄) + a(λ)} ≥ C γ̄ (z̄). (2.95) a∈C λ∈co 0 In what follows, we will need the convexity property of the cone C . It is implied by the following assertion. Proposition 2.36. The set A0 is convex. Proof. Let l1 (· ) ∈ A0 , l2 (· ) ∈ A0 , p > 0, q > 0, p + q = 1. It is required to show that l (· ) = pl1 (· ) + ql2 (· ) ∈ A0 . Let {δwi } ∈ , i = 1, 2, be two sequences of needle-shaped 70 Chapter 2. Quadratic Conditions in the Calculus of Variations variations such that tf λ t0 δi H dt γi → li (λ), i = 1, 2, t where δi H λ and γi = t0f i dt correspond to the sequences {δwi }, i = 1, 2. Using the sequences {δwi }, i = 1, 2, we construct a sequence {δw } ∈ such that tf t0 δ H λ dt → l(λ). γ (2.96) Thus, the convexity of A0 will be proved. For brevity, we set Then ξi (λ) → li (λ), i = 1, 2. γi tf λ t0 δi H dt = ξi (λ), i = 1, 2. Moreover, we set pγ2 =α , qγ1 + pγ2 qγ1 =β . qγ1 + pγ2 Then α > 0, β > 0, α + β = 1, and α γ1 = p, α γ1 + β γ2 β γ2 = q. α γ1 + β γ2 Consequently, α ξ1 (λ) + β ξ2 (λ) ξ1 (λ) ξ2 (λ) =p +q → pl1 (λ) + ql2 (λ) = l(λ). α γ1 + β γ2 γ1 γ2 Therefore, α tf λ t0 δ1 H dt + (1 − α t α t0f 1 dt + (1 − α ) ) tf λ t0 δ2 H dt tf t0 2 dt → l(λ). (2.97) Assume now that there is a sequence of functions {α(t)} in L∞ , each member of which satisfies the following conditions: (i) α(t) assumes two values, 0 or 1, only; t t t t (ii) α t0f δ1 H λ dt = t0f α(t)δ1 H λ dt, α t0f δ2 H λ dt = t0f α(t)δ2 H λ dt for all λ ∈ co 0 ; t t t t (iii) α t0f 1 dt = t0f α(t)1 dt and α t0f 2 dt = t0f α(t)2 dt. We note that conditions (ii) hold for all elements of λ ∈ co 0 whenever they hold for finitely many linearly independent elements of co 0 that compose a basis in L0 = Lin(co 0 ). Therefore, in conditions (ii) and (iii) we in essence speak about the preservation of finitely many integrals. We set δw = α(t)δw1 + (1 − α(t))δw2 . Then, obviously, {δw } ∈ , and, moreover, tf tf tf λ λ α δ1 H dt + (1 − α ) δ2 H dt = α(t)δ1 H λ + (1 − α(t))δ2 H λ dt t0 t0 t0 tf = t0 δ H λ dt ∀ λ ∈ co 0 , 2.4. Estimation of the Basic Constant from Above 71 where δ H λ corresponds to the sequence {δw }. Analogously, tf tf tf 1 dt + (1 − α ) 2 dt = dt, α t0 t0 t0 where = (δw , t). This and (2.97) imply (2.96). Therefore, the convexity of A0 will be proved if we ensure the existence of a sequence {α(t)} satisfying conditions (ii) and (iii). The existence of such a sequence is implied by the Blackwell lemma, which is well known in optimal control theory and is contiguous to the Lyapunov theorem on the convexity of the range of a vector-valued measure. However, we note that we need not satisfy conditions (ii) and (iii) exactly: it suffice to do this with an arbitrary accuracy; i.e., given an arbitrary sequence ε → +0 in advance, we need to ensure the fulfillment of conditions (ii) and (iii) with accuracy up to ε for each serial number of the sequence. Also, in this case, condition (2.96) certainly holds. With such a weakening of conditions (ii) and (iii), we can refer to Theorem 16.1 in [79, Part 2]. The proposition is proved. 2.4.9 Narrowing of the Set co 0 up to the Set M co (C ) In what follows, we will deal with the transformation of the left-hand side of inequality (2.95). For this purpose, we need the following abstract assertion. Lemma 2.37. Let X be a Banach space, F : X → R1 be a sublinear (i.e., convex and positively homogeneous) functional, K ⊂ X be a nonempty convex cone, and x0 ∈ X be a fixed point. Then the following formula holds: inf F (x0 + x) = x∈K sup x ∗ ∈∂F ∩K ∗ x ∗ , x0 , where ∂F = {x ∗ ∈ X∗ | x ∗ , x ≤ F (x) for all x ∈ X} is the set of support functionals of F and K ∗ = {x ∗ ∈ X∗ | x ∗ , x ≥ 0 for all x ∈ K} is the dual cone of K. We use this lemma in order to transform the expression inf max {λ (z̄) + a(λ)} a∈C λ∈co 0 in the left-hand side of inequality (2.91). Denote by A the set of all affine functions a(λ) : co 0 → R1 . As was already noted, A is a finite-dimensional space. In this space, we consider the sublinear functional F : a(· ) ∈ A → max a(λ). λ∈co 0 Since co 0 is a convex compact set, we can identify it with the set of support functionals of F ; more precisely, there is a one-to-one correspondence between each support functional a∗ ∈ ∂F and the element λ ∈ co 0 such that a∗ , a = a(λ) for all a ∈ A. Moreover, according to this formula, a certain support functional a∗ ∈ ∂F corresponds to every element λ ∈ co 0 . Further, let C be a number such that the cone C defined above is nonempty. Then what was said above implies that ∂F ∩ ∗C can be identified with the set def M(C ; co 0 ) = λ ∈ co 0 a(λ) ≥ 0 ∀ a ∈ C . 72 Chapter 2. Quadratic Conditions in the Calculus of Variations By Lemma 2.37, we obtain from this that inf max {λ (z̄) + a(λ)} = a∈C λ∈co 0 max λ∈M(C ;co 0 ) λ (z̄). (2.98) It suffices to make more precise what the set M(C ; co 0 ) means. For an arbitrary C, denote by M co (C) the set of λ ∈ co 0 such that (2.99) H (t, x 0 (t), u, ψ(t)) − H (t, x 0 (t), u0 (t), ψ(t)) ≥ C(t, u) 0 d(u) (t, x, u) ∈ Q}. Namely in this if t ∈ [t0 , tf ] \ , u ∈ U(t, x (t)), where U(t, x) = {u ∈ R case, we say that the minimum principle “of strictness C” holds for λ. For a positive C and λ ∈ 0 it is a strengthening of the usual minimum principle. Also, we note that the set M co (C) for C = 0 coincides with the set M0co defined in the same way as the set M0 from Section 2.1.4, with the only difference being that in the definition of the latter, it is necessary to replace the set 0 by its convex hull co 0 . Proposition 2.38. For any real C, we have M(C ; co 0 ) ⊂ M co (C ). (2.100) Proof. Let C ∈ R. Let λ̂ ∈ M(C ; co 0 ), i.e., λ̂ ∈ co 0 and a(λ̂) ≥ 0 for all a ∈ C . Hence l(λ̂) ≥ C for all l ∈ A0 . Using this inequality, we show that λ̂ ∈ M co (C). Fix an arbitrary point t and a vector u such that t ∈ (t0 , tf ) \ , u ∈ Rd(u) , (t , x 0 (t ), u ) ∈ Q, Let ε > 0. Define the needle-shaped variation u − u0 (t), δu (t) = 0 u = u0 (t ). (2.101) t ∈ [t − ε, t + ε], otherwise. For ε → +0, we have the sequence {δw } ∈ , where δw = (0, δu ). For each λ ∈ co 0 , there exists the limit tf λ δ H λ t0 δ H dt = lim tf ε→+0 dt t0 t=t for this sequence. According to the definition of the set A0 , this limit is l(λ), where l(· ) ∈ A0 . Since l(λ̂) ≥ C, we have δ H λ̂ ≥ C. t=t In other words, for λ = λ̂, inequality (2.99) holds for arbitrary u and t satisfying conditions (2.101). This implies λ̂ ∈ M co (C). The proposition is proved. (In fact, we have the relation M(C ; co 0 ) = M co (C), but we need only the inclusion for decoding the constant Cγ .) We obtain from relation (2.98) and inclusion (2.100) that inf max {λ (z̄) + a(λ)} ≤ a∈C λ∈co 0 max λ∈M co (C) This and Lemma 2.35 imply the following assertion. λ (z̄). 2.4. Estimation of the Basic Constant from Above 73 Lemma 2.39. Let min{Cγ , Cγ } ≥ C > −∞. Then max λ∈M co (C) λ (z̄) ≥ C γ̄ (z̄) ∀ z̄ ∈ KZ . (2.102) The distinguishing of the cone C by using the set of sequences of needle-shaped variations and the “narrowing” of the set co 0 up to the set M co (C) by using formula (2.98) referred to the duality theory represents Milyutin’s method which is called the “cone technique” bearing in mind the cone C of affine functions on the convex compact set co 0 . Also, we note that the use of this method necessarily leads to the convexification of the compact set 0 in all the corresponding formulas. 2.4.10 Closure of the Cone KZ in the Space Z2 ( ) It remains to prove that Lemma 2.39 remains valid if we replace KZ by K in it. This is implied by the following assertion. Proposition 2.40. The closure of the cone KZ in the space Z2 () coincides with the cone K. The proof of this proposition uses the Hoffman lemma on the estimation of the distance to the solution set of a system of linear inequalities, i.e., Lemma 1.12. More precisely, we use the following consequence of the Hoffman lemma. Lemma 2.41. Let X and Y be two Banach spaces, let li : X → R1 , i = 1, . . . , k, be linear functionals on X, and let A : X → Y be a linear operator with closed range. Then there exists a constant N = N(l1 , . . . , lk , A) > 0 such that for any point x0 ∈ X, there exists x̄ ∈ X such that li , x0 + x̄ ≤ 0, i = 1, . . . , k; A(x0 + x̄) = 0, k + li , x0 + Ax0 . x̄ ≤ N (2.103) (2.104) i=1 Indeed, system (2.103) is compatible, since it admits the solution x̄ = −x0 , and then by Lemma 1.12, there exists a solution satisfying estimate (2.104) (clearly, the surjectivity condition for the operator A : X → Y in Lemma 1.12 can be replaced by the closedness of the range of this operator considering the image AX as Y ). Proof of Proposition 2.40. Since KZ ⊂ K and K is closed in Z2 (), it suffices to only show that K ⊂ [KZ ]2 , where [· ]2 is the closure in Z2 (). Let z̄ = (ξ̄ , x̄, ū) ∈ K. We show that there exists a sequence in KZ that converges to z̄ in Z2 (). Take a sequence N → ∞. For each member of this sequence, we set ū(t), |ū(t)| ≥ N , N ū (t) = 0, |ū(t)| < N . t The sequence {ūN } satisfies the condition t0f ūN , ūN dt → 0. Let x̄ N be defined from the conditions x̄˙ N = fx x̄ N + fu ūN , x̄ N (t0 ) = 0, x̄ N ∈ W 1,2 (, Rd(x) ), where W 1,2 (, Rd(x) ) 74 Chapter 2. Quadratic Conditions in the Calculus of Variations is the space of absolutely continuous functions x(t) : [t0 , tf ] → Rd(x) having the Lebesgue square integrable derivatives; it is endowed with the norm tf 1/2 x 1,2 = x(t0 ), x(t0 ) + ẋ, ẋ dt . t0 Then x̄ N 1,2 → 0, and hence x̄ N C → 0. We set z̄N = (0, x̄ N , ūN ) and z̄N = z̄ − z̄N . Then z − z̄N Z2 () = z̄N Z2 () → 0. The conditions z̄ ∈ K and x̄ N C → 0 imply (Fip p̄N )+ + |Kp p̄N | → 0, (2.105) i∈I where p̄N corresponds to the sequence z̄N . Moreover, {z̄N } belongs to the subspace T2 ⊂ Z2 () defined by the conditions x̄˙ = fx x̄ + fu ū, [x̄]k = [f ]k ξ̄k , k = 1, . . . , s. Applying Lemma 2.41 on T2 , we obtain that for a sequence {z̄N } in T2 satisfying condition (2.105), there exists a sequence {z̄N } in KZ such that z̄N − z̄N Z2 () → 0. But the following condition also holds for {z̄N }: z̄ − z̄N Z2 () → 0. Therefore, z̄ ∈ [KZ ]2 . Since z̄ is an arbitrary element from K, we have K ⊂ [KZ ]2 , and then K = [KZ ]2 . The proposition is proved. Lemma 2.39 and Proposition 2.40 imply the following assertion. Lemma 2.42. Let min{Cγ , Cγ } ≥ C > −∞. Then max λ∈M co (C) λ (z̄) ≥ C γ̄ (z̄) ∀ z̄ ∈ K. (2.106) We now recall that Cγ ≥ 0 is a necessary condition for the Pontryagin minimum. Assume that it holds. Then, setting C = 0 in (2.106), we obtain the following result. Theorem 2.43. Let w 0 = (x 0 , u0 ) be a Pontryagin minimum point. Then max λ (z̄) ≥ 0 ∀ z̄ ∈ K. λ∈M0co Therefore, we have obtained the quadratic necessary condition for the Pontryagin minimum, which is slightly weaker than Condition A of Theorem 2.4. It is called Condition Aco . Using Condition Aco , we will show below that Condition A is also necessary for the Pontryagin minimum. We thus will complete the proof of Theorem 2.4. Section 2.6 is devoted to this purpose. But first we complete (in Section 2.5) the decoding of the constant Cγ . Denote by CK the least upper bound of C such that M co (C) is nonempty and condition (2.106) holds. Then Lemma 2.42 implies CK ≥ min{Cγ , Cγ }, (2.107) i.e., the constant CK estimates the constant Cγ from above with accuracy up to a constant multiplier. We will prove that the constant CK estimates the constant Cγ from below with 2.5. Estimation of the Basic Constant from Below 75 accuracy up to a constant multiplier. This will allow us to obtain a sufficient condition for the Pontryagin minimum. Remark 2.44. Lemma 2.42 also implies the following assertion: if Cγ ≥ 0, then M0co is nonempty and maxco λ (z̄) ≥ 0 ∀ z̄ ∈ K. λ∈M0 This assertion serves as an important supplement to inequality (2.107). 2.5 2.5.1 Estimation of the Basic Constant from Below Method for Obtaining Sufficient Condition for a Pontryagin Minimum Since Cγ > 0 is a sufficient condition for the Pontryagin minimum, the positivity requirement for any quantity that estimates Cγ from below is also a sufficient condition. Therefore, the second stage of decoding, the estimation of Cγ from below, can also be considered as a method for obtaining a sufficient condition for the Pontryagin minimum. As was already noted, the sufficiency of the condition Cγ > 0 for the Pontryagin minimum is a sufficiently elementary fact not requiring the constructions of Chapter 1 for its proof. Therefore a source for obtaining a sufficient condition for the Pontryagin minimum is very simple, which is characteristic for sufficient conditions in general. However, in contrast to other sources and methods for obtaining sufficient conditions, which often are of arbitrary character, in this case, we are familiar with connection of the sufficient condition being used with the necessary condition; it consists of the passage from the strict inequality Cγ > 0 to the nonstrict inequality. This fact is already not so obvious, it is nontrivial, and it is guaranteed by the abstract scheme. 2.5.2 Extension of the Set σ γ For convenience, we recall the following main definitions: 0 , 0 (δw) = max (λ, δw) = max (λ, δw), λ∈0 λ∈co 0 γ tf (λ, δw) = δl λ − (ψδ ẋ − δH λ ) dt, σ γ = {{δw} ∈ | σ ≤ O(γ )}, Cγ = inf lim σ γ t0 σ = max{Fi (p0 + δp) (i ∈ I ), |δK|, δ ẋ − δf 1 }. Let M ⊂ co 0 be an arbitrary nonempty compact set. Then we have the following for any variation δw ∈ δW : def 0 (δw) = max (λ, δw) ≥ max (λ, δw) = M (δw). co 0 M Consequently, Cγ := inf lim σ γ 0 M ≥ inf lim . σ γ γ γ (2.108) 76 Chapter 2. Quadratic Conditions in the Calculus of Variations We set √ o(√γ ) = {{δw} ∈ | σ = o( γ )}. Since σ γ ⊂ o(√γ ) , we have inf lim σ γ M M ≥ inf√ lim . o( γ ) γ γ (2.109) Inequalities (2.108) and (2.109) imply Cγ ≥ inf√ lim o( γ) M . γ (2.110) Let C ∈ R1 be such that M co (C) is nonempty. We set C = max (λ, δw), co M (C) Cγ C ; o(√γ ) = inf√ lim o( γ) C . γ Then (2.110) implies the following assertion. Lemma 2.45. The following inequality holds for an arbitrary C such that M co (C) is nonempty: Cγ ≥ Cγ (C ; o(√γ ) ). (2.111) In what follows, we will fix an arbitrary C such that the set M co (C) is nonempty. 2.5.3 Passing to Local Sequences We set In other words, √ loc √ loc | σ = o( γ )}. o( γ ) = {{δw} ∈ (2.112) loc √ √ loc o( γ ) = o( γ ) ∩ . Our further goal consists of passing from the constant Cγ (C , o(√γ ) ) defined by the √ set of sequences o(√γ ) to a constant defined by the set of sequences loc o( γ ) . For such a passage, we need the estimate |C | ≤ O(γ ) | loc , (2.113) i.e., |C (δw)| ≤ O(γ (δw)) for any sequence {δw} ∈ loc . To prove estimate (2.113), we need a certain property of the set M co (C) analogous to one of the Weierstrass–Erdmann conditions of the classical calculus of variations. Let us formulate this analogue. Proposition 2.46. The following conditions hold for any λ ∈ M co (C): [H λ ]k = 0 ∀ tk ∈ . 2.5. Estimation of the Basic Constant from Below 77 Proof. Fix arbitrary λ ∈ M co (C) and tk ∈ . We set tε = tk − ε, ε > 0, and δε H λ = H (tε , x 0 (tε ), u0 (tε ) + [u0 ]k , ψ(tε )) − H (tε , x 0 (tε ), u0 (tε ), ψ(tε )). Then for a small ε > 0, the condition λ ∈ M co (C) implies δε H λ ≥ Cδε , where δε = (tε , u0 (tε ) + [u0 ]k ) − (tε , u0 (tε )) = (tε , u0 (tε ) + [u0 ]k ). Taking into account that δε H λ → [H λ ]k and δε → 0 as ε → +0, we obtain [H λ ]k ≥ 0. Constructing an analogous sequence t ε = tk + ε to the right from the point tk , we obtain −[H λ ]k ≥ 0. Therefore, [H λ ]k = 0. The proposition is proved. λ k In Section 2.1.5, we have defined 0 as the set of tuples λ ∈ 0 such that [H ] = 0 co for all tk ∈ . Proposition 2.46 means that M (C) ⊂ co 0 . This and Proposition 2.16 imply estimate (2.113). By the way, we note that under the replacement of M co (C) by co 0 , estimate (2.113) does not hold in general. Recall that by u0 we have denoted the closure of the graph of the function u0 (t) assuming that u0 (t) is left continuous. By Qtu we have denoted the projection of the set Q under the mapping (t, x, u) → (t, u). Denote by {V } an arbitrary sequence of neighborhoods of the compact set u0 contained in Qtu such that V → u0 . The latter means that for any neighborhood V ⊂ Qtu , of the compact set u0 , there exists a number starting from which V ⊂ V . Let {δw} ∈ be an arbitrary sequence. For members δw = (δx, δu) and V of the sequences {δw} and {V }, respectively, which have the same numbers, we set δu(t) if (t, u0 (t) + δu(t)) ∈ V , δuV (t) = 0 otherwise, δuV = δu(t) − δuV (t), δw loc = (δx, δuV ), δw V = (0, δuV ). Then {δw} = {δwloc } + {δw V }, {δw loc } ∈ loc , and {δw V } ∈ . Proposition 2.47. The following relation holds for the above representation of the sequence {δw}: δf = δ loc f + δ Vf + rfV , where δf δ Vf = f (t, w 0 + δw) − f (t, w 0 ), δ locf = f (t, w 0 + δw V ) − f (t, w0 ), rfV = f (t, w 0 + δwloc ) − f (t, w0 ) = (δ̄x f − δx f )χ V ; the function χ V is the characteristic function of the set M V = {t | δuV = 0}; and δx f = f (t, x 0 + δx, u0 ) − f (t, x 0 , u0 ), δ̄x f = f (t, x 0 + δx, u0 + δu) − f (t, x 0 , u0 + δu). Moreover, rfV ∞ ≤ δ̄x f ∞+ δx f ∞ def = εf → 0 and rfV 1 ≤ εf meas M V . 78 Chapter 2. Quadratic Conditions in the Calculus of Variations Proof. We have δf = δf (1 − χ V ) + δf χ V = δ loc f (1 − χ V ) + (f (t, x 0 + δx, u0 + δu) − f (t, x 0 , u0 + δu) + f (t, x 0 , u0 + δu) − f (t, x 0 , u0 ))χ V = δ loc f − δ loc f χ V + δ̄x f χ V + δ Vf = δ loc f + δ Vf + (δ̄x f − δx f )χ V = δ loc f + δ Vf + rfV . The estimates for rfV are obvious. The proposition is proved. Also, it is obvious that the representation γ = γ loc + γ V corresponds to the representation {δw} = {δw loc } + {δwV }, where γ = γ (δw), γ loc = γ (δw loc ), and γ V = γ (δwV ). This is implied by the relation (t, u0 + δu) = (t, u0 + δuV ) + (t, u0 + δuV ). Now let us define a special sequence {V } = {V (ε)} of neighborhoods of the compact 0 set u that converges to this compact set. Namely, for each ε > 0, we set V = V (ε) = Oε (t, u), (u,t) ∈ u0 where Oε (t, u) = {(t , u ) | |t − t | < ε 2 , |u − u | < ε }. Fix a sequence {δw} ∈ and define a sequence {ε} for it such that ε → +0 sufficiently slowly. The following proposition explains what “sufficiently slowly” means. Proposition 2.48. Let ε → +0 so that εf → 0 and ε2 √ γ → 0, 2 ε where εf is the same as in Proposition 2.47. Then rfV meas MV = o( γ V ). 1 ≤ εf meas MV = o(γ V ) and Proof. Let there exist the sequences {δw}, {ε}, and {V } = {V (ε)} defined as above. Since V → u0 , we have that, starting from a certain number, V ⊂ V, where V ⊂ Qtu is the neighborhood of the compact set u0 from the definition of the function (t, u) given in Section 2.3. Condition V ⊂ V implies (δuV )V = δuV (the meaning of this notation is the same as above). Hence δuV = δuVV + δuV , where (t, u0 + δuVV ) ∈ V \ V on MVV = / V on M V = {t | δuV = 0}. The definitions of V, , and {t | δuVV = 0} and (t, u0 + δuV ) ∈ V V . This implies γ V ≥ ε 2 meas M V , where γ V := 0 V = V (ε) imply (t, u + δuV ) ≥ ε 2 on MV V V V γV . Further, it follows from the ε2 t V V V V definitions of and δu that meas M ≤ const γ , where γ := t0f (t, u0 + δuV ) dt, and const depends on the entire sequence {δuV } but not on its specific member. Since γ V ≤ γ V tf t0 V ≤ (t, u0 + δuVV ) dt. Since γVV ≤ γ V , we have meas MV (because MV ⊂ MV ), we have meas M V ≤ const γ V . Therefore, V γ 1 V meas M V = meas MV + meas MV ≤ 2 + const γ V = + const γ V . ε ε2 (2.114) 2.5. Estimation of the Basic Constant from Below 79 Taking into account that εf /ε 2 → 0, we obtain from (2.114) that εf meas MV = o(γ V ). √ Moreover, since γ /ε 2 → 0 and γ V ≤ γ , (2.114) also implies ! ! ! γV V V V =o + const γ γ γV . meas M ≤ ε2 The proposition is proved. In what follows, for brevity we set δ = (λ, δw), δ loc λ = (λ, δw loc ), λ δC = C (δw), δ loc C = C (δwloc ). Assume that there is an arbitrary sequence {δw} ∈ . We define a sequence {ε} for it such that the conditions of Proposition 2.48 hold. Also, we define the corresponding sequences {V } = {V (ε)}, {δwloc }, and {δw V }. Then tf tf λ λ δ := δl − ψδ ẋ dt + δH λ dt t t 0tf 0tf = δl λ − ψδ ẋ dt + ψ(δ loc f + δ Vf + rfV ) dt t0 t0 tf tf tf tf λ loc λ V λ ψδ ẋ dt + δ H dt + δ H dt + ψrfV dt = δl − t0 t0 t0 t0 tf loc λ = δ + δ V H λ dt + ρ V λ , t0 where ρV λ = tf t0 ψrfV dt. Since supco 0 ψ ∞ < +∞ and rfV 1 = o(γ V ), we have sup |ρ V λ | = o(γ V ). (2.115) co 0 Finally, we obtain δλ = δ loc λ + tf δ V H λ dt + ρ V λ . t0 Moreover, the conditions δf = δ loc f + δ Vf + rfV , δ Vf and rfV 1 = o(γ V ) imply δ ẋ − δ loc f 1 ≤ δ ẋ − δf ≤ const meas MV = o 1 1 +o ! γV . (2.116) γV , (2.117) Therefore, the following assertion holds. Proposition 2.49. Let {δw} ∈ be an arbitrary sequence, and let a sequence {ε} satisfy the conditions of Proposition 2.48. Then conditions (2.115)–(2.117) hold for the corresponding sequences {V (ε)}, {δwloc }, and {δwV }. We now are ready to pass to local sequences. We set √ Cγ C ; loc o( γ ) = inf lim √ loc o( γ ) We have the following assertion. C . γ 80 Chapter 2. Quadratic Conditions in the Calculus of Variations Lemma 2.50. The following inequality holds: √ Cγ C ; o(√γ ) ≥ min C, Cγ C ; loc . o( γ ) (2.118) Proof. Let {δw} ∈ o(√γ ) be a sequence with γ > 0. Choose the sequence {ε} and the corresponding sequence {V } as in Proposition 2.48. With the sequences {δw} and {V }, we associate the splitting {δw} = {δwloc } + {δwV }. Let us turn to formula (2.116). The definition of the set M co (C) implies δ V H λ ≥ C(t, u0 + δuV ) ∀ λ ∈ M co (C) for all sufficiently large numbers. Then (2.115) and (2.116) imply δC ≥ δ loc C + Cγ V + o(γ V ), (2.119) where γ V = γ (δw V ). We set γ = γ (δw) and γ loc = γ (δwloc ). Then γ = γ loc + γ V . The following two cases are possible: loc Case (a): lim γγ = 0. loc Case (b): lim γγ > 0. Let us consider each of them. loc Case (a): Let lim γγ = 0. Choose subsequences such that γ loc /γ → 0 for them, and, therefore, γV → 1. (2.120) γ loc = o(γ ) and γ We preserve the above notation for the subsequences. It follows from (2.119), (2.120), and estimate (2.113) that δC ≥ Cγ + o(γ ). Hence lim δC ≥ C, γ where the lower limit is taken over the chosen subsequence. loc ). Case (b): Now let limγ loc /γ > 0, and, therefore, γ ≤ O(γ loc ) and γ V ≤ O(γ √ Since δ ẋ − δf 1 = o( γ ), it follows from (2.117) that δ ẋ − δ loc f 1 = o( γ loc ). Moreover, ! ! √ √ δ loc Fi = δFi ≤ o( γ ) = o1 γ loc , i ∈ I ; |δ loc K| = |δK| = o( γ ) = o γ loc . √ Hence {δw loc } ∈ loc o( γ ) . Further, we obtain from (2.119) that δC δ loc C + Cγ V γ loc δ loc C γ V lim + ≥ lim = lim C γ γ γ γ γ loc δ loc C δ loc C ≥ lim min , C = min lim ,C . γ loc γ loc 2.5. Estimation of the Basic Constant from Below 81 The second equation follows from the condition γ loc γ V + = 1. γ γ Furthermore, lim δ loc C √ ≥ Cγ C ; loc o( γ ) , γ loc √ since {δwloc } ∈ loc o( γ ) . Hence lim δC √ ≥ min Cγ C ; loc , C . o( γ ) γ Therefore, we have shown that for every sequence {δw} ∈ o(√γ ) on which γ > 0, there exists a subsequence such that lim δC /γ is not less than the right-hand side of inequality (2.118). This implies inequality (2.118). The lemma is proved. The method used in this section is very characteristic for decoding higher-order conditions in optimal control in order to obtain sufficient conditions. Now this method is related to the representation of the Pontryagin sequence as the sum {δw} = {δw loc } + {δw V } and the use of the maximum principle of strictness C for {δwV }: δ V H λ ≥ C(t, u0 + δuV ), λ ∈ M co (C). In what follows, analogous consideration will be given to various consequences of the minimum principle of strictness C, the Legendre conditions, and conditions related to them. Lemmas 2.45 and 2.50 imply the following assertion. Lemma 2.51. Let the set M co (C) be nonempty. Then √ Cγ ≥ min Cγ C ; loc , C . o( γ ) 2.5.4 Simplifications in the Definition of Cγ C √ ; loc o( γ ) loc satisfying the conditions √ By definition, loc o( γ ) consists of sequences {δw} ∈ √ δFi ≤ o( γ ) ∀ i ∈ I , √ δ ẋ − δf 1 = o( γ ). √ |δK| = o( γ ), (2.121) (2.122) Obviously, conditions (2.121) are equivalent to the conditions √ √ Fip δp ≤ o( γ ) ∀ i ∈ I , |Kp δp| = o( γ ). Let us consider condition (2.122). Assume that the sequence {δw} is represented in the canonical form (see Proposition 2.9) {δw} = {δw 0 } + {δw ∗ }, {δw 0 } = {(δx, δu0 )} ∈ 0 , {δw∗ } = {(0, δu∗ )} ∈ ∗ , |δu0 | · |δu∗ | = 0. 82 Chapter 2. Quadratic Conditions in the Calculus of Variations By Proposition 2.10, 1 δf = fw δw 0 + fww δw0 , δw0 + δ ∗ f + δ ∗ fx δx + r̃, 2 t where r̃||1 = o(γ 0 ) and γ 0 := δx 2C + t0f |δu0 |2 dt. According to formula (2.50), ∗ ∗ δk∗ f = [f ]k (χk− − χk+ ) + fuk+ δvk− + fuk− δvk+ + O(|δtk | + |δvk |2 ). δ∗f = As was shown in Section 2.4 in proving the equivalence of conditions (2.66) and (2.69), we have the estimates tf 0 0 0 fww δw , δw 1 ≤ o(γ ), δ ∗ fx δx dt ≤ o(γ ). t0 Moreover, ≤ |fuk+ | meas Mk− δvk− k− 1 ≤ |fu | meas Mk+ δvk+ fuk+ δvk− where γ ∗ := = o( γ ∗ ), ∗ 2 = o( γ ), 1 fuk− δvk+ 2 tf tf 2 1/2 is k ( Mk∗ |δt| dt + t0 |δvk | dt) and v 2 = ( t0 v(t), v(t) dt) L2 (, Rd(u) ) of Lebesgue square integrable functions v(t) : [t0 , tf ] of the space Therefore, condition (2.122) is equivalent to the condition √ ∗ ∗ δ ẋ − fw δw 0 − [f ]k (χk− − χk+ ) = o( γ ). the norm → Rd(u) . 1 Finally, we obtain √ loc o( γ ) = √ {δw} ∈ loc | Fip δp ≤ o( γ ) ∀ i ∈ I ; ∗ ∗ [f ]k (χk− − χk+ ) δ ẋ − fw δw0 − 1 √ |Kp δp| = o( γ ); √ = o( γ ) . In this relation, we use the canonical representation of the sequence {δw} = {δw 0 } + {δw ∗ }, where {δw0 } ∈ 0 , {δw∗ } ∈ ∗ , and |δu0 | · |δu∗ | = 0. √ By Proposition 2.15, in calculating Cγ (C ; loc o( γ ) ), we can use the function 1λ (δw) 1C (δw) := max co M (C) (see definition (2.61) for 1λ ) instead of the function C = maxM co (C) (λ, δw), and since by Proposition 2.46the conditions [H λ ]k = 0 for all tk ∈ hold for any λ ∈ M co (C), ∗ − meas M ∗ ) in the definition of 1λ , thus we can omit the terms k [H λ ]k (meas Mk− k+ passing to the function ˜ 1λ (δw) := 1 (λ, δw 0 ) 2 s k λ D (H ) + Mk k=1 1 + 2 |δtk | dt + [Hxλ ]k ∗ tf t0 tf t0 ∗ ∗ δx(χk− − χk+ ) dt λk+ λk− Huu δvk− , δvk− + Huu δvk+ , δvk+ dt , 2.5. Estimation of the Basic Constant from Below where λ (λ, δw) = lpp δp, δp + tf t0 83 λ Hww δw, δw dt. √ Therefore, the constant Cγ C ; loc o( γ ) does not change, if we replace the function C ˜ 1 , where by the function C ˜ 1λ (δw). ˜ 1C (δw) = max co M (C) Finally, we recall that the functional γ has the following form on local sequences represented in the canonical form: tf tf 2 |δu0 |2 dt + |δtk | dt + |δvk |2 dt γ (δw) = δx 2C + t0 Mk∗ k t0 or, in short, γ = γ 0 + γ ∗ , where γ 0 = δx 2 C + tf |δu0 |2 dt, γ∗ = t0 k 2 Mk∗ |δtk | dt + tf |δvk |2 dt . t0 We have proved that loc√ ˜1 √ C C ; loc o( γ ) = Cγ C ; o( γ ) , where ˜ 1C ; loc√ Cγ o( γ ) = inf lim √ loc o( γ ) (2.123) ˜1 C . γ ˜ 1λ We call attention to the fact that in the definition of this constant, the functions 0 ∗ and γ are defined indeed on the set of triples (δw , M , δv) such that ∗ ∪ M∗ , δw0 = (δx, δu0 ) ∈ W , M∗ = ∪Mk∗ , Mk∗ = Mk− k+ ∗ ∗ ∗ Mk− ⊂ (tk − ε, tk ), Mk+ ⊂ (tk , tk + ε), k ∈ I , ε > 0; δv = δvk , δvk = δvk− + δvk+ , k ∈ I ∗ ; ∗ , {t | δvk− = 0} ⊂ Mk− ∗ , {t | δvk+ = 0} ⊂ Mk+ k ∈ I ∗. Denote by a these triples, and denote by {a} an arbitrary sequences of triples a such that {δw 0 } ∈ 0 ; Mk∗ → tk , k ∈ I ∗; δv ∞ → 0. To each sequence {δw} from loc represented in canonical form, we naturally associate the sequence of triples {a} = {a(δw)} = {(δw 0 , M ∗ , δv)}. However, note that for an arbitrary sequence {a}, we do not require the condition δu0 χ ∗ = 0, which holds for the sequences {a} corresponding to the sequences {δw} ∈ loc in their canonical representations (as before, χ ∗ is the characteristic function of the set M∗ ). Therefore, on the set of sequences {a}, we 84 Chapter 2. Quadratic Conditions in the Calculus of Variations ˜ 1λ = ˜ 1λ (a), λ ∈ M co (C) and γ = γ (a) = γ 0 + γ ∗ . We set have defined the functions √ √ {a} = {(δw0 , M ∗ , δv)} | Fip δp ≤ o( γ ) ∀ i ∈ I , |Kp δp| = o( γ ), S1 = √ ∗ ∗ [f ]k (χk− − χk+ ) = o( γ ) , δ ẋ − fw δw0 − 1 ∗ where χk− ∗ and χk+ Also, we set ∗ and M ∗ , respectively. are the characteristic functions of the sets Mk− k+ ˜ 1C (a) = max ˜ 1λ (a), co M (C) ˜ 1C ; S1 ) = inf lim Cγ ( S1 ˜1 C . γ The following inequality holds: ˜ 1C ; loc√ ˜1 Cγ o( γ ) ≥ Cγ C ; S1 . (2.124) √ Indeed, the sequence {a} = {a(δw)} ∈ S1 corresponds to every sequence [δw} ∈ loc o( γ ) , 1 ˜ and, moreover, the values of and γ are preserved under this correspondence. There is C no converse correspondence, since we omit the condition δu0 χ ∗ = 0 in the definition of S1 . Lemma 2.51 and formulas (2.123) and (2.124) imply the following assertion. ˜ 1 ; S1 ), C}. Lemma 2.52. Let the set M co (C) be nonempty. Then Cγ ≥ min{Cγ ( C ˜ 1 ; S1 ) from below. In what follows, we estimate the constant Cγ ( C 2.5.5 Use of Legendre Conditions Now our goal is to pass to sequences with δv = 0. This will be done by using the Legendre conditions for points tk ∈ . Let us formulate the following general assertion on the Legendre conditions. Proposition 2.53. Let λ ∈ M co (C). Then we have that (a) the following condition holds for any t ∈ [t0 , tf ] \ : 1 Huu (t, x 0 (t), u0 (t), ψ(t))ū, ū ≥ Cū, ū 2 ∀ ū ∈ Rd(u) ; (b) the following conditions hold for any tk ∈ : 1 λk+ ū, ū ≥ Cū, ū H 2 uu 1 λk− H ū, ū ≥ Cū, ū 2 uu ∀ ū ∈ Rd(u) ; ∀ ū ∈ Rd(u) . Proof. Let λ ∈ M co (C), t ∈ [t0 , tf ]. Choose ε > 0 so small that the conditions ũ ∈ Rd(u) and |ũ| < ε imply (t, x 0 (t), u0 (t) + ũ) ∈ Q and (t, u0 (t) + ũ) = |ũ|2 , and hence H (t, x 0 (t), u0 (t) + ũ, ψ(t)) − H (t, x 0 (t), u0 (t), ψ(t)) ≥ C|ũ|2 . 2.5. Estimation of the Basic Constant from Below 85 In other words, the function ϕ(ũ) := H (t, x 0 (t), u0 (t) + ũ, ψ(t)) − C|ũ|2 defined on a neighborhood of the origin of the space Rd(u) has a local minimum at zero. This implies ϕ (0) = 0 and ϕ (0)ū, ū ≥ 0 for all ū ∈ Rd(u) . The first condition is equivalent to Hu (t, x 0 (t), u0 (t), ψ(t)) = 0, and the second is equivalent to Huu (t, x 0 (t), u0 (t), ψ(t))ū, ū − 2Cū, ū ≥ 0 ∀ ū ∈ Rd(u) . This implies assertion (a) of the proposition. Assertion (b) is obtained from assertion (a) by passing to the limit as t → tk + 0 and t → tk − 0, k ∈ I ∗ . The proposition is proved. We now use assertion (b) only. Denote by b the pair (δw0 , M ∗ ) and by {b} the sequence of pairs such that {δw0 } ∈ 0 , Mk∗ → tk for all k, M∗ = ∪Mk∗ . On each such sequence, we define the functions tf 1 0 k λ λ k ∗ ∗ D (H ) (λ, δw ) + 2λ (b) = |δtk | dt + [Hx ] δx(χk− − χk+ ) dt , 2 Mk∗ t0 2C (b) γ1 (b) k = max 2λ (b), M co (C) tf = δx 2C + |δu0 |2 dt + t0 k Mk∗ 2|δtk | dt. We set S2 = √ √ {b} = {(δw0 , M ∗ )} | Fip δp ≤ o( γ1 ), i ∈ I ; |Kp δp| = o( γ1 ); √ ∗ − χ∗ ) δ ẋ − fw δw 0 − [f ]k (χk− k+ 1 = o( γ1 ) . ˜ 1 , γ , and We obtain the definitions of 2 , γ1 , and S2 from the corresponding definitions of S1 setting δv = 0 everywhere. We set Cγ1 (2C ; S2 ) = inf lim S2 2C . γ1 Lemma 2.54. The following inequality holds: ˜ 1C ; S1 ) ≥ min Cγ1 (2C ; S2 ), C . Cγ ( (2.125) Proof. Let {a} ∈ S1 be an arbitrary sequence such that γ > 0 for all its members. For this sequence, we set tf 0 ∗ |δv|2 dt. {b} = {(δw , M )}, γ̂ (δv) = t0 Then γ (a) = γ1 (b)+ γ̂ (δv), or, for short, γ = γ1 + γ̂ . Let λ ∈ M co (C). Proposition 2.53(b), ˜ 1λ (δw0 ) ≥ 2λ (δw0 )+C γ̂ (δv). Consequently, ˜ 1 (δw0 ) ≥ 2 (δw 0 )+C γ̂ (δv), implies C C ˜ 1 ≥ 2 + C γ̂ . We consider the following two possible cases for the seor, briefly, C C quence {a}. 86 Chapter 2. Quadratic Conditions in the Calculus of Variations Case (a): Let lim(γ1 /γ ) = 0. Extract a subsequence such that γ1 = o(γ ) on it. Let this condition hold for the sequence {a} itself. Then we obtain from the inequality ˜ 1 ≥ 2 + C γ̂ and the obvious estimate |2 | ≤ O(γ1 ) that C C C lim ˜1 2 + C γ̂ o(γ ) + C γ̂ γ̂ C ≥ lim C = lim = C lim = C, γ γ γ γ since γ = γ1 + γ̂ = o(γ ) + γ̂ . Case (b): Assume now that lim(γ1 /γ ) > 0, and hence γ ≤ const γ1 on the subse˜ 1 ≥ 2 + C γ̂ implies quence. The inequality C C ˜1 2C + C γ̂ 2C γ̂ γ1 2C C + C ≥ min ,C , ≥ = γ γ γ γ1 γ γ1 since γ1 + γ̂ = γ , γ1 ≥ 0, and γ̂ ≥ 0. Consequently, ˜1 2C C lim ≥ min lim ,C . γ γ1 But the conditions {a} ∈ S and γ ≤ const γ1 immediately imply {b} ∈ S2 . We obtain from this that ˜1 lim C ≥ Cγ1 (2C , S2 ). γ1 Consequently, lim ˜1 C ≥ min Cγ1 (2C ; S2 ), C . γ Therefore, we have shown that from any sequence {a} ∈ S1 at which γ > 0, it is possible to ˜ 1 /γ on it is not less than the rightextract a subsequence such that the lower limit lim C hand side of inequality (2.125). This obviously implies inequality (2.125). The lemma is proved. Lemmas 2.52 and 2.54 imply the following assertion. Lemma 2.55. Let the set M co (C) be nonempty. Then Cγ ≥ min Cγ1 (2C ; S2 ), C . In what follows, we will estimate the constant Cγ1 (2C ; S2 ) from below. 2.5.6 Replacement of δx2C by |δx(t0 )|2 in the Definition of the Functional γ1 We set γ2 = γ2 (b) = |δx(t0 )|2 + tf t0 |δu0 |2 dt + k Mk∗ 2|δtk | dt. 2.5. Estimation of the Basic Constant from Below 87 Therefore, γ2 differs from γ1 by the fact that the term δx 2C is replaced by |δx(t0 )|2 . By definition, the sequences {b} = {(δw, M∗ )} from S2 satisfy the condition √ ∗ ∗ δ ẋ − fw δw − [f ]k (χk− − χk+ ) = o( γ1 ). 1 Therefore, according to Proposition 2.24, there exists a constant 0 < q ≤ 1, q = 1/ (see (2.73)), such that for any sequence {b} ∈ S2 , there exists a number starting from which we have qγ1 (b) ≤ γ2 (b), or, briefly, qγ1 ≤ γ2 . Moreover, since |δx(t0 )| ≤ δx C , we have γ2 ≤ γ1 . Hence γ2 ≤ 1. q≤ γ1 This implies that the following relations hold for any sequence {b} ∈ S2 : lim 2C 2 γ2 = lim C · ≥ min Cγ2 (2C ; S2 ), qCγ2 (2C ; S2 ) . γ1 γ2 γ1 Consequently, Cγ1 (2C ; S2 ) ≥ min Cγ2 (2C ; S2 ), qCγ2 (2C ; S2 ) . (2.126) Here, Cγ2 (2C ; S2 ) = inf lim S2 2C . γ2 Inequality (2.126) and Lemma 2.55 imply the following assertion. Lemma 2.56. Let the set M co (C) be nonempty. Then Cγ ≥ min Cγ2 (2C ; S2 ), qCγ2 (2C ; S2 ), C , 0 < q ≤ 1. In what follows, we will estimate the constant Cγ2 (2C ; S2 ) from below. 2.5.7 Passing to Sequences with Discontinuous State Variables As in Section 2.4, denote by {b̄} = {(w̄, M ∗ )} a sequence such that M ∗ = ∪Mk∗ , Mk∗ → tk (k ∈ I ∗ ), w̄ = (x̄, ū), x̄ ∈ P W 1,1 (, Rd(x) ), ū ∈ L∞ (, Rd(u) ), Also, we set 1 D k (H λ ) 3λ (b̄) = (λ, w̄) + 2 k w̄ → 0. Mk∗ k |δtk | dt + [Hxλ ]k x̄av ξk , ∗ − meas M ∗ , 3 (b̄) = max co 3λ where ξk = meas Mk− M (C) (b̄), and k+ C tf γ2 (b̄) = |x̄(t0 )|2 + |ū|2 dt + 2 |δtk | dt. t0 ∞ k Mk∗ 88 Chapter 2. Quadratic Conditions in the Calculus of Variations Let S3 be the set of sequences {b̄} such that √ √ Fip p̄ ≤ o( γ2 ) (i ∈ I ), |Kp p̄| = o( γ2 ), √ x̄˙ − fw w̄ 1 = o( γ2 ), [x̄]k = [f ]k ξk ∀ tk ∈ , ∗ − meas M ∗ . According to Lemma 2.26, for any sequence {b} ∈ S where ξk = meas Mk− 2 k+ for which γ2 > 0, there exists a sequence {b̄} such that √ δp = p̄, x̄˙ − fw w̄ 1 = o( γ2 ), [x̄]k = [f ]k ξk ∀ tk ∈ ; γ2 (b) = γ2 (b̄), 2C (b) = 3C (b̄) + o(γ2 ). Consequently, {b̄} ∈ S3 and lim 2C (b) 3 (b̄) 3 def = lim C ≥ inf lim C = Cγ2 (3C ; S3 ). S3 γ2 (b) γ2 γ2 (b̄) Since {b} ∈ S2 is an arbitrary sequence on which γ2 > 0, we have Cγ2 (2C ; S2 ) ≥ Cγ2 (3C ; S3 ). (2.127) This inequality and Lemma 2.56 imply the following assertion. Lemma 2.57. Let the set M co (C) be nonempty. Then Cγ ≥ min Cγ2 (3C ; S3 ), qCγ2 (3C ; S3 ), C . In what follows, we will estimate the constant Cγ2 (3C ; S3 ). 2.5.8 Additional Condition of Legendre, Weierstrass–Erdmann Type Related to Varying Discontinuity Points of the Control Now our goal is to pass from the sequences {(w̄, M ∗ )} ∈ S3 to the sequences {(ξ̄ , w̄)}, where ∗ − meas M ∗ , k = 1, . . . , s. ξ̄ = (ξ̄1 , . . . , ξ̄s ), ξ̄k = meas Mk− k+ Proposition 2.58. The following inequality holds for any λ ∈ M co (C): D k (H λ ) ≥ 2C, k ∈ I ∗. Proof. Fix λ ∈ M co (C), k ∈ I ∗ . Take a small ε > 0 and construct a variation δu(t) in the left neighborhood (tk − ε, tk ) of the point tk such that t ∈ (tk − ε, tk ), u0k+ − u0 (t), δu(t) = 0, t∈ / (tk − ε, tk ). For a sufficiently small ε > 0, we have (t, x 0 (t), u0 (t) + δu(t)) ∈ Q. Consequently, δH λ ≥ C(t, u0 (t) + δu(t)), 2.5. Estimation of the Basic Constant from Below 89 where δH λ = H λ (t, x 0 , u0 + δu) − H λ (t, x 0 , u0 ). It follows from the definition of (t, u) that the following relation holds for a sufficiently small ε > 0: ∗ , (t, u0 (t) + δu(t)) = 2|δtk |χk− ∗ is the characteristic function of the set M ∗ = {t | δu(t) = 0} = (t − ε, t ). where χk− k k k− Moreover, according Proposition 2.13, with the conditions [H λ ]k = 0, δvk− = (u0 + δu − ∗ = 0 taken into account, we have the following for δu: δH λ = (D k (H λ )|δt | + u0k+ )χk− k ∗ . Consequently, D k (H λ )|δt | + o(|δt |) ≥ 2C|δt | on (t − ε, t ). This implies o(|δtk |))χk− k k k k k D k (H λ ) ≥ 2C. The lemma is proved. Denote by S4 the set of sequences {z̄} = {(ξ̄ , x̄, ū)} of elements of the space Z() (defined by (2.89)) such that |ξ̄ | + ||w̄ ∞ → 0 and, moreover, √ √ Fip p̄ ≤ o( γ̄ ) (i ∈ I ), |Kp p̄| = o( γ̄ ), √ ∀ tk ∈ . x̄˙ − fw w̄ 1 = o( γ̄ ), [x̄]k = [f ]k ξ̄k Recall that γ̄ = γ̄ (z̄) := |x̄(t0 )|2 + λ (z̄) := tf t0 |ū|2 dt + |ξ̄ |2 . Also, recall that (see formula (2.13)) 1 k λ 2 1 k (D (H )ξ̄k + 2[Hxλ ]k x̄av ξ̄k ) + (λ, w̄), 2 2 k λ p̄, p̄ + where (λ, w̄) = lpp tf λ t0 Hww w̄, w̄ dt. We set C (z̄) = max λ (z̄) co M (C) and Cγ̄ (C ; S4 ) = inf lim S4 C . γ̄ Using Proposition 2.58, we prove the following estimate. Lemma 2.59. The following inequality holds: Cγ2 (3C ; S3 ) ≥ min{Cγ̄ (C ; S4 ), C}. (2.128) Proof. Let {b̄} = {(w̄, M ∗ )} ∈ S3 be a sequence on which γ2 > 0. The following inequalities hold for each member of this sequence: 2 |δtk | dt ≥ ξ̄k2 , k ∈ I ∗ , Mk∗ ∗ − meas M ∗ . We set where ξ̄k = meas Mk− k+ μk = 2 Mk∗ |δtk | dt − ξ̄k2 , k ∈ I ∗, μ= μk . 90 Chapter 2. Quadratic Conditions in the Calculus of Variations Then μk ≥ 0 for all k, and, therefore, μ ≥ 0. To the sequence {b̄} = {(w̄, M ∗ )}, we naturally associate the sequence {z̄} = {ξ̄ , w̄)} with the same components w̄ and with ∗ − meas M ∗ , k ∈ I ∗ . Then γ (b̄) = γ̄ (z̄) + μ, or, for short, γ = γ̄ + μ. ξ̄k = meas Mk− 2 2 k+ Moreover, 1 1 k λ 2 λ k k 3λ (λ, w̄) + D (H )(ξ̄k + μk ) + [Hx ] x̄av ξ̄k (b̄) = 2 2 k 1 k λ D (H )μk . = λ (z̄) + 2 k According to Proposition 2.58, D k (H λ ) ≥ 2C for all k for any λ ∈ M co (C). Consequently, 3C (b̄) ≥ C (z̄) + Cμ or, briefly, 3C ≥ C + Cμ. Therefore, we have the following for the sequence {b̄}: 3C C + Cμ ≥ , γ2 γ2 (2.129) where γ2 = γ̄ + μ. Consider the following two cases. Case (a). Let lim γ̄ /γ2 = 0. Choose a subsequence such that γ̄ = o(γ2 ) on it. Let this condition hold on the sequence {b̄} itself. Since γ2 = γ̄ + μ, we have μ/γ2 → 1. Let us show that the following estimate holds in this case: |C | = o(γ2 ). (2.130) Indeed, the definition of the functional λ and the boundedness of the set M co (C) imply the existence of a constant C > 0 such that |C | ≤ C ( x̄ Further, since {b̄} ∈ S3 , we have x̄˙ = fx x̄ + fu ū + r̄; Consequently, x̄ ∞ r̄ 1 1 ū √ = o( γ2 ); ≤ O(|x̄(t0 )| + |ξ̄ | + ū |x̄(t0 )| + |ξ̄ | + ū 2 ∞+ 1+ r̄ ≤ O( γ̄ ), 2 2 2 + |ξ̄ | ). [x̄]k = [f ]k ξ̄k , 1 ). (2.131) k ∈ I ∗. Since r̄ 1 √ = o( γ2 ), we have x̄ 2∞ ≤ O(γ̄ ) + o(γ2 ). This and estimate (2.131) imply |C | ≤ O(γ̄ ) + o(γ2 ). But γ̄ = o(γ2 ). Consequently, estimate (2.130) holds. Taking into account estimate (2.130), we obtain from inequality (2.129) that lim 3C C + Cμ Cμ ≥ lim = lim = C. γ2 γ2 γ2 Case (b). Let lim γ̄ /γ2 > 0. Then γ2 = O(γ̄ ). This and the condition {b̄} ∈ S3 easily imply that {z̄} ∈ S4 . We obtain from inequality (2.129) that 3C μ γ̄ C C + C ≥ min ,C , ≥ · γ2 γ2 γ̄ γ2 γ̄ 2.5. Estimation of the Basic Constant from Below 91 since γ̄ ≥ 0, μ ≥ 0, and γ2 = γ̄ + μ. Consequently, 3 C ; C ≥ min Cγ̄ (C , S4 ), C . lim C ≥ min lim γ2 γ̄ The latter inequality holds, since {z̄} ∈ S4 . Therefore, we have proved that it is possible to extract a subsequence from any sequence {b̄} ∈ S3 on which γ2 > 0 such that the following inequality holds on it: lim 3C ≥ min{Cγ̄ (C ; S4 ), C}. γ2 This implies inequality (2.128). The lemma is proved. Lemmas 2.57 and 2.59 imply the following assertion. Lemma 2.60. Let the set M co (C) be nonempty. Then Cγ ≥ min{Cγ̄ (C , S4 ), qCγ̄ (C , S4 ), C, qC}. In what follows, we will estimate the constant Cγ̄ (C ; S4 ) from below. 2.5.9 Passing to Equality in the Differential Relation We restrict the set of sequences S4 to the set of sequences in which each member satisfies the conditions x̄˙ = fw w̄, [x̄]k = [f ]k ξ̄k , k ∈ I ∗ . We denote by S5 this new set of sequences. Let us show that under such a restriction, the constant Cγ̄ does not increase. In other words, let us prove the following lemma. Lemma 2.61. The following relation holds: Cγ̄ (C ; S4 ) = Cγ̄ (C ; S5 ). Proof. Let {z̄} ∈ S4 be such that γ̄ > 0 on it. Then √ √ Fip p̄ ≤ o( γ̄ ) (i ∈ I ), |Kp p̄| = o( γ̄ ), √ r̄ = o( γ̄ ); [x̄]k = [f ]k ξ̄k x̄˙ = fx x̄ + fu ū + r̄, (k ∈ I ∗ ). (2.132) Let a sequence {δx} satisfy the conditions δ ẋ = fx δx + r̄ and δx(t0 ) = 0. Then δx C ≤ δx 1,1 ≤ const · r̄ 1 = o( γ̄ ). We set δu = 0, δw = (δx, 0), w̄ = w̄ − δw, ξ̄ = ξ̄ , z̄ = (ξ̄ , w̄ ). For the sequence {z̄ }, we have γ̄ (z̄ ) = γ̄ (z̄), or, for short, γ̄ = γ̄ . Moreover, Fip p̄ ≤ o( γ̄ ) (i ∈ I ), |Kp p̄ | = o( γ̄ ), w̄ ∞ + |ξ̄ | → 0. x̄˙ = fw w̄ , [x̄ ]k = [f ]k ξ (k ∈ I ∗ ), k Hence {z̄ } ∈ S5 . Moreover, λ (z̄) = 1 λ (z̄ ) + lpp p̄, δp + lpp δp, δp 2 1 tf λ λ + Hxx δx, δx + 2Hxx δx, x̄ dt + [Hx ]k δx(tk )ξ̄k . 2 t0 k (2.133) 92 Chapter 2. Quadratic Conditions in the Calculus of Variations √ Conditions (2.132) easily imply x̄ 2∞ ≤ O(γ̄ ). Moreover, δx C = o( γ̄ ). This and conditions (2.133) imply C (z̄) = C (z̄ ) + o(γ̄ ). Consequently, lim C (z̄) C (z̄ ) = lim ≥ Cγ̄ (C ; S5 ). γ̄ (z̄) γ̄ (z̄ ) The inequality holds, since {z̄ } ∈ S5 . Since {z̄} is an arbitrary sequence from S4 on which γ̄ > 0, we obtain from this that Cγ̄ (C ; S4 ) ≥ Cγ̄ (C ; S5 ). The inclusion S5 ⊂ S4 implies the converse inequality. Therefore, we have an equality here. The lemma is proved. Lemmas 2.60 and 2.61 imply the following assertion. Lemma 2.62. Let the set M co (C) be nonempty. Then Cγ ≥ min{Cγ̄ (C ; S5 ), qCγ̄ (C ; S5 ), C, qC}. We now estimate the constant Cγ̄ (C ; S5 ) from below. 2.5.10 Passing to the Critical Cone Introduce the constant Cγ̄ (C ; S5 ) = inf C (z̄) γ̄ (z̄) z̄ ∈ K \ {0} . Lemma 2.63. The following inequality holds: Cγ̄ (C ; S5 ) ≥ Cγ̄ (C ; K). Proof. Let {z̄} be an arbitrary nonvanishing sequence from S5 . For this sequence, we have Fip p̄ ≤ o( γ̄ ) (i ∈ I ), |Kp p̄| = o( γ̄ ), w̄ ∞ + |ξ̄ | → 0, (2.134) ˙x̄ = fw w̄, [x̄]k = [f ]k ξ̄k (k ∈ I ∗ ). (2.135) Moreover, γ̄ (z̄) > 0 on the whole sequence, since it contains nonzero members. Let T be the subspace in Z() defined by conditions (2.135). According to Lemma 2.41 (which ¯ = follows from the Hoffman lemma), for the sequence {z̄}, there exists a sequence {z̄} ¯ ¯ ū)} ¯ in the subspace T such that {(ξ̄ , x̄, ¯ ≤0 Fip (p̄ + p̄) (i ∈ I ), ¯ = 0, Kp (p̄ + p̄) √ ¯ As in the proof of and, moreover, x̄¯ ∞ + ū¯ ∞ + |ξ̄¯ | = o( γ̄ ). We set {z̄ } = {z̄ + z̄}. Lemma 2.61, for the sequence {z̄ }, we have C (z̄) = C (z̄ ) + o(γ̄ ). Moreover, γ̄ (z̄) = γ̄ (z̄ ) + o(γ̄ ). (2.136) √ (2.137) ¯ ¯ ¯ These conditions √ follow from the estimate x̄ ∞ + ū ∞ + |ξ̄ | = o( γ̄ ) and the estimate x̄ ∞ ≤ O( γ̄ ), which, in turn, follows from (2.135). It follows from (2.137) and the 2.5. Estimation of the Basic Constant from Below 93 condition γ̄ (z̄) > 0 that γ̄ (z̄ ) > 0 starting from a certain number. According to (2.136) and (2.137), we have C (z̄) C (z̄ ) (2.138) = lim ≥ Cγ̄ (C ; K). lim γ̄ (z̄) γ̄ (z̄ ) The inequality holds, since z̄ ∈ K \ {0}. Since this inequality holds for an arbitrary nonvanishing sequence {z̄} ∈ S5 , this implies Cγ̄ (C ; S5 ) ≥ Cγ̄ (C ; K). Lemmas 2.62 and 2.63 imply the following assertion. Lemma 2.64. The following inequality holds for any real C such that the set M co (C) is nonempty: (2.139) Cγ ≥ min{Cγ̄ (C ; K), qCγ̄ (C ; K), C, qC}. Let C be such that M co (C) is nonempty and C (z̄) ≥ C γ̄ (z̄) ∀ z̄ ∈ K. (2.140) Then it follows from the definition of the constant Cγ̄ (C ; K) that Cγ̄ (C ; K) ≥ C. This and (2.139) imply Cγ ≥ min{C, qC}. This inequality holds for all C such that the set M co (C) is nonempty and condition (2.140) holds. Therefore, it also holds for the least upper bound of these C. At the end of Section 2.4, we have denoted this upper bound by CK . Therefore, we have proved the following theorem. Theorem 2.65. The following inequality holds: Cγ ≥ min{CK , qCK }, 0 < q ≤ 1. (2.141) Earlier, at the end of Section 2.4, we obtained the following estimate (see (7.43)): 1 CK ≥ min{Cγ , Cγ }, = ≥ 1, q which can be written in the equivalent form: 1 max CK , CK ≥ Cγ , 2.5.11 ≥ 1. (2.142) Decoding Result Combining inequalities (2.141) and (2.142), we obtain the following decoding result. Theorem 2.66. The following inequalities hold : min{CK , qCK } ≤ Cγ ≤ max{CK , qCK }. (2.143) These inequalities are equivalent to min{Cγ , Cγ } ≤ CK ≤ max{Cγ , Cγ } (2.144) 94 Chapter 2. Quadratic Conditions in the Calculus of Variations Recall that ≥ 1, 0 < q ≤ 1, and = 1/q. Bearing in mind inequalities (2.143) and (2.144), we write const C γ = CK . (2.145) This is the main result of decoding. We now obtain the following important consequences of this result. Case (a). Let CK ≥ 0. Then Cγ ≥ 0 by (2.143); according to Remark 2.44, this implies the condition M0co = ∅; max λ (z̄) ≥ 0 ∀ z̄ ∈ K. (2.146) co M0 Conversely, if condition (2.146) holds, then CK ≥ 0. Therefore, condition (2.146) is equivalent to the inequality CK ≥ 0; by Theorem 2.66, the latter is equivalent to the inequality Cγ ≥ 0. Case (b). Let CK > 0. Let the constant C be such that CK > C > 0. Then, by the definition of CK , we have M co (C) = ∅; max λ (z̄) ≥ C γ̄ (z̄) M co (C) ∀ z̄ ∈ K. (2.147) Conversely, if for a certain C > 0, condition (2.147) holds, then CK ≥ C, and hence CK > 0. Therefore, the existence of C > 0 such that (2.147) holds is equivalent to the inequality CK > 0. By Theorem 2.66, the latter is equivalent to the inequality Cγ > 0. The following theorem summarizes what was said above. Theorem 2.67. (a) The inequality Cγ ≥ 0 is equivalent to condition (2.146). (b) The inequality Cγ > 0 is equivalent to the existence of C > 0 such that condition (2.147) holds. 2.5.12 Sufficient Conditions for the Pontryagin Minimum We give the following definition. Definition 2.68. Let (t, u) be an admissible function. We say that Condition Bco () holds at the point w 0 if there exists C > 0 such that condition (2.147) holds. As we already know, the condition Cγ > 0 is sufficient for the strict Pontryagin minimum at the point w0 , and, according to Theorem 2.67, it is equivalent to Condition Bco (). Therefore, Condition Bco () is also sufficient for the strict Pontryagin minimum. Let us consider Condition Bco (). First of all, we show that it is equivalent to Condition B() in whose definition we have the set M(C) instead of the set M co (C). By definition, the set M(C) consists of tuples λ = (α0 , α, β, ψ) ∈ 0 such that ψ(t)(f (t, x 0 (t), u) − f (t, x 0 (t), u0 (t))) ≥ C(t, u) if t ∈ [t0 , tf ]\ and (t, x 0 (t), u) ∈ Q. Therefore, the set M(C) differs from the set M co (C) by that co 0 is replaced by 0 in the definition of the latter. Condition B() means that the set M(C) is nonempty for a certain C > 0 and max λ (z̄) ≥ C γ̄ (z̄) M(C) ∀ z̄ ∈ K. (2.148) Since 0 ⊂ co 0 , and hence M(C) ⊂ M co (C), Condition B() implies Condition Bco (). It is required to prove the converse statement: Condition Bco () implies Condition B(). For this purpose, we prove the following lemma. 2.5. Estimation of the Basic Constant from Below 95 Lemma 2.69. For any C > 0 such that M co (C) is nonempty, there exists 0 < ε < 1 such that C co , M (C) ⊂ [ε, 1] ◦ M ε where [ε, 1] ◦ M is the set of tuples λ = ρ λ̃ such that ρ ∈ [ε, 1] and λ̃ ∈ M. Proof. For an arbitrary λ = (α0 , α, β, ψ) ∈ co 0 , we set ν(λ) = α0 + |α| + |β|. Since the function ν(λ) is convex and equals 1 on 0 , we have ν(λ) ≤ 1 for all λ ∈ co 0 . Further, let C > 0 be such that M co (C) is nonempty. Then as is easily seen, the compact set M co (C) does not contain zero. This implies ε= min λ∈M co (C) ν(λ) > 0, since the conditions λ ∈ co 0 and ν(λ) = 0 imply λ = 0. Therefore, the inequalities ε ≤ ν(λ) ≤ 1 hold for an arbitrary λ ∈ M co (C). We set λ̃ = λ/ν(λ). Then ν(λ̃) = 1, and hence λ̃ ∈ 0 . Moreover, the condition ψδu f ≥ C implies C C ψ̃δu f ≥ ≥ , ν(λ) ε where ψ̃ is a component of λ̃, δu f = f (t, x 0 , u) − f (t, x 0 , u0 ), and = (t, u). Hence, λ̃ ∈ M( Cε ). Therefore, for an arbitrary λ ∈ M co (C), we have found the representation λ = ν(λ)λ̃, where λ̃ ∈ M( Cε ), ε ≤ ν(λ) ≤ 1. This implies C M co (C) ⊂ [ε, 1] ◦ M . ε The lemma is proved. Lemma 2.69 implies the following theorem. Theorem 2.70. Condition Bco () is equivalent to Condition B(). Proof. Let Condition Bco () hold, i.e., there exists C > 0 such that M co (C) is nonempty and max λ (z̄) ≥ C γ̄ (z̄) ∀ z̄ ∈ K. co M (C) Then, by Lemma 2.69, there exist ε > 0 such that λ max C (z̄) ≥ C γ̄ (z̄) [ε,1]◦M ε ∀ z̄ ∈ K. By the linearity of λ in λ and the positivity of C, we obtain from this that λ max C (z̄) ≥ C γ̄ (z̄) M ε ∀ z̄ ∈ K. Therefore, Condition B() holds. We have shown that Condition Bco () implies Condition B(). As mentioned above, the converse is also true. Therefore, Conditions Bco () and B() are equivalent. The theorem is proved. 96 Chapter 2. Quadratic Conditions in the Calculus of Variations Theorems 2.67(b) and 2.70 imply the following theorem. Theorem 2.71. The inequality Cγ > 0 is equivalent to Condition B(). There is a certain inconvenience in Condition B(), in that it is difficult to verify that λ ∈ M(C). Therefore, it is desirable to pass to the sufficient condition, which can be more easily verified. Recall that in Section 2.1.8 we introduced the set M0+ consisting of those λ ∈ M0 for which the strict minimum principle holds outside the discontinuity points of the control. In the same section, we have introduced the set Leg+ (M0+ ) consisting of all strictly Legendrian elements λ ∈ M0+ . The definition of M(C) ⊂ M co (C), its compactness, and Propositions 2.53 and 2.58 imply the following assertion. Lemma 2.72. For any admissible function (t, u) and any C > 0, the set M(C) is a compact set contained in the set Leg+ (M0+ ). Also, the following assertion holds. Lemma 2.73. For any nonempty compact set M ⊂ Leg+ (M0+ ), there exist an admissible function (t, u) and a constant C > 0 such that M ⊂ M(C). Before proving Lemma 2.73, we prove a slightly simpler property. Let U be an arbitrary neighborhood of the compact set u0 containing in Qtu . With the subscript U, we denote all objects referring to the canonical problem complemented by the constraint (t, u) ∈ U. For example, we write M0U , M U (C), etc. Denote by Leg+ (0 ) the subset of all strictly Legendre elements λ ∈ 0 . Lemma 2.74. Let M ⊂ Leg+ (0 ) be a nonempty compact set, and let (t, u) be an admissible function. Then there exist a neighborhood U of the compact set u0 and a constant C > 0 such that M ⊂ M U (C). To prove Lemma 2.74, we need several auxiliary assertions. Proposition 2.75. Assume that there is a nonempty compact set M ⊂ 0 such that the following conditions hold for each of its elements λ: (a) for any t ∈ [t0 , tf ] \ , 1 Huu (t, x 0 (t), u0 (t), ψ(t))ū, ū > 0 ∀ ū ∈ Rd(u) \ {0}; 2 (b) for any tk ∈ , 1 λk+ H ū, ū > 0 ∀ ū ∈ Rd(u) \ {0} 2 uu (2.149) 1 λk− ū, ū > 0 ∀ ū ∈ Rd(u) \ {0}. H 2 uu (2.150) and 2.5. Estimation of the Basic Constant from Below 97 Then there exist C > 0 and ε > 0 such that for any λ ∈ M, the conditions t ∈ [t0 , tf ] \ and |u − u0 (t)| < ε (2.151) H (t, x 0 (t), u, ψ(t)) − H (t, x 0 (t), u0 (t), ψ(t)) ≥ C|u − u0 (t)|2 . (2.152) imply Proof. Assume the contrary. Let the compact set M ⊂ 0 be such that conditions (a) and (b) of the proposition hold for each of its element, but there are no C > 0 and ε > 0 such that conditions (2.151) imply inequality (2.152). Then there exist sequences {Cn }, {tn }, {λn }, and {ūn } such that Cn → +0, tn ∈ [t0 , tf ] \ , λn ∈ M, ūn ∈ Rd(u) , |ūn | → 0, H λn (tn , xn0 , u0n + ūn ) − H λn (tn , xn0 , u0n ) < Cn |ūn |2 , (2.153) where xn0 = x 0 (tn ) and u0n = u0 (tn ). Without loss of generality, we assume that tn → tˆ ∈ [t0 , tf ], λn → λ̂ ∈ M, ūn → ū. |un | Then ūn = εn (ū + ũn ), where εn = |ūn | → 0, |ūn | → 0. In this case, we obtain from (2.153) that 1 λn (2.154) H (tn , xn0 , u0n )ūn , ūn + o(εn2 ) < Cn εn2 . 2 uu Here, we have taken into account that Huλn (tn , xn0 , u0n ) = 0. We first assume that tˆ ∈ / . Dividing (2.154) by εn2 and passing to the limit, we obtain 1 λ̂ H (tˆ, x 0 (tˆ), u0 (tˆ))ū, ū ≤ 0. 2 uu But this contradicts condition (a), since ū = 0. Analogously, in the case where tˆ ∈ , we arrive at a contradiction to one of the conditions in (b). The proposition is proved. In what follows, we need to use the assumption that each tk ∈ is an L-point of the control u0 . Proposition 2.76. Let M ⊂ 0 be a nonempty compact set such that the following conditions hold for a fixed point tk ∈ and any λ ∈ M: [H λ ]k = 0, D k (H λ ) > 0, 1 λk− ū, ū > 0 ∀ ū ∈ Rd(u) \ {0}. H 2 uu (2.155) (2.156) Then there exist C > 0 and ε > 0 such that for any λ ∈ M, the conditions tk < t < tk + ε and |u − u0k− | < ε (2.157) imply the inequality 1 H (t, x 0 (t), u, ψ(t)) − H (t, x 0 (t), u0 (t), ψ(t)) ≥ C |t − tk | + |u − u0k− |2 . 2 (2.158) 98 Chapter 2. Quadratic Conditions in the Calculus of Variations Proof. Let a compact set M ⊂ 0 satisfy the condition of the proposition, and let there be no C > 0 and ε > 0 such that for any λ ∈ M, conditions (2.157) imply inequality (2.158). Then there exist a sequence {C} and a sequence of triples {(t, u, λ)} such that C → +0, t → tk + 0, u → u0k+ , H λ (t, x 0 (t), u) − H λ (t, x 0 (t), u0 (t)) < C 1 |t − tk | + |u − u0k− |2 2 (2.159) (we omit the serial numbers of members). We set t − tk = δt > 0, u − u0k− = δv, u0 (t) − u0k+ = δu0 , x 0 (tk ) = x 0k , x 0 − x 0k = δx 0 , etc. Then we get H λ (t, x 0 , u) = H (t, x 0 , u, ψ) = H (tk + δt, x 0k + δx 0 , u0k− + δv, ψ k + δψ) = H λk− + Htλk− + Hxλk+ ẋ 0k+ + ψ̇ k+ Hψλk− δt + Huλk− δv 1 λk− + Huu δv, δv + o(|δt| + |δv|2 ), 2 and taking into account that |δu0 | ≤ L|δt| (L > 0) by assumption, we obtain H λ (t, x 0 , u0 ) = H (t, x 0 , u0 , ψ) = H (tk + δt, x 0k + δx 0 , u0k+ + δu0 , ψ k + δψ) = H λk+ + Htλk+ + Hxλk+ ẋ 0k+ + ψ̇ k+ Hψλk+ δt + Huλk+ δu0 + o(δt). Subtracting the latter relation from the previous one, taking into account that Huλk+ = Huλk− = 0 and [H λ ]k = 0, and also taking into account inequality (2.159), we obtain 1 λk− − [Htλ ]k + [Hxλ ]k ẋ 0k+ + ψ̇ k+ [Hψλ ]k δt + Huu δv, δv + o(|δt| + |δv|2 ) 2 1 2 < C |δt| + |δv| . 2 But [Htλ ]k + [Hxλ ]k ẋ 0k+ + ψ̇ k+ [Hψλ ]k = −D k (H λ ) and C → +0. Hence 1 λk− δv, δv < o1 (|δt| + |δv|2 ). D k (H λ )δt + Huu 2 (2.160) Without loss of generality, we assume that λ → λ̂ ∈ M. We consider the following two possible cases for the sequences {δt} and {δv}. Case (a). Assume that there exists a subsequence such that the following relation holds on it: |δv|2 = o(|δt|). Then it follows from (2.160) that D k (H λ )δt < o(δt). Since δt → +0 and λ → λ̂, we obtain from this that D k (H λ̂ ) ≤ 0, which contradicts the condition λ̂ ∈ M. Case (b). Assume now that |δv|2 > 0, lim |δt| i.e., |δt| ≤ O(|δv|2 ). In this case, from (2.160) and the conditions δt > 0 and D k (H λ ) > 0, we obtain 1 λk− H δv, δv < o(|δv|2 ). (2.161) 2 uu 2.5. Estimation of the Basic Constant from Below 99 Without loss of generality, we assume that δv → v̄, |δv| |v̄| = 1. Dividing inequality (2.161) by |δv|2 and passing to the limit, we obtain 1 λ̂k− v̄, v̄ ≤ 0. H 2 uu Since |v̄| = 1, this also contradicts the condition λ̂ ∈ M. Therefore, our assumption that there exist the sequences {C} and {(t, u, λ)} with the above property is not true. We thus have proved the proposition. The following assertion is proved analogously. Proposition 2.77. Let M ⊂ 0 be a nonempty compact set such that the following conditions hold for a fixed point tk ∈ and any λ ∈ M: [H λ ]k = 0, D k (H λ ) > 0, and 1 λk+ ū, ū > 0 ∀ ū ∈ Rd(u) \ {0}. H 2 uu Then there exist C > 0 and ε > 0 such that for any λ ∈ M, the conditions tk − ε < t < tk , and |u − u0k+ | < ε imply 1 0 0 0 0k+ 2 H (t, x (t), u, ψ(t)) − H (t, x (t), u (t), ψ(t)) ≥ C |t − tk | + |u − u | . 2 Propositions 2.75, 2.76, and 2.77 directly imply Lemma 2.74. Proof of Lemma 2.73. Assume that there exists a nonempty compact set M ⊂ Leg+ (M0+ ). Let 1 (t, u) be a certain admissible function (as was shown in Section 2.3.7, there exists at least one such function). According to Lemma 2.73, there exist a neighborhood U ⊂ Qtu of the compact set u0 and a constant C > 0 such that M ⊂ M U (C1 ), i.e., M ⊂ 0 and the conditions t ∈ [t0 , tf ] \ , (t, u) ∈ U, and (t, x 0 (t), u) ∈ Q imply H λ (t, x 0 (t), u) − H λ (t, x 0 (t), u0 (t)) ≥ C1 (t, u). We set 1 min {H λ (t, x 0 (t), u) − H λ (t, x 0 (t), u0 (t))}, C λ∈M (t, u) = min{h(t, u), 1 (t, u)}. h(t, u) = (2.162) It is easy to see that the function (t, u) (defined by (2.162)) is admissible, and, moreover, M ⊂ M(C). The lemma is proved. We now recall the definition given in Section 2.1.8. We say that Condition B holds for the point w0 if there exist a nonempty compact set M ⊂ Leg+ (M0+ ) and a constant C > 0 100 Chapter 2. Quadratic Conditions in the Calculus of Variations such that max λ (z̄) ≥ C γ̄ (z̄) M ∀ z̄ ∈ K. (2.163) The following assertion holds. Theorem 2.78. Condition B is equivalent to the existence of an admissible function (t, u) such that Condition B() holds. Proof. Let Condition B hold; i.e., there exist a nonempty compact set M ⊂ Leg+ (M0+ ) and a constant C > 0 such that condition (2.163) holds. Then, according to Lemma 2.73, there exist an admissible function (t, u) and a constant C1 > 0 such that M ⊂ M(C1 ). We set C2 = min{C, C1 }. Then M(C1 ) ⊂ M(C2 ). Consequently, M ⊂ M(C2 ) and max λ (z̄) ≥ max λ (z̄) ≥ C γ̄ (z̄) ≥ C2 γ̄ (z̄) M(C2 ) M ∀ z̄ ∈ K (the second inequality holds by (2.163)). Therefore, Condition B() holds. Conversely, let there exist an admissible function such that Condition B() holds. Then there exists C > 0 such that M(C) is nonempty and condition (2.148) holds. By Lemma 2.72, M(C) ⊂ Leg+ (M0+ ), and M(C) is a compact set. Therefore, Condition B also holds. The theorem is proved. According to Theorem 2.71, the inequality Cγ > 0 is equivalent to Condition B(). This is true for every admissible function . This and Theorem 2.78 imply the following theorem. Theorem 2.79. Condition B is equivalent to the existence of an admissible function such that the condition Cγ > 0 holds for the order γ corresponding to it. In Section 2.2, we have verified all the assumptions of the abstract scheme for the canonical problem, the point w 0 , the sets P and Q (the absorbing set = W corresponds to them in the space W ), and the set of Pontryagin sequences in the space W . In Section 2.3, the corresponding assumptions of the abstract scheme were also verified for a higher order γ on . Therefore, Theorem 1.7 is applicable. According to this theorem, the condition Cγ > 0 is not only sufficient for the Pontryagin minimum at the point w 0 but is equivalent to the γ -sufficiency on . The latter will be also called the Pontryagin γ -sufficiency. For convenience, we define this concept here. Let be an admissible function, and let γ be the higher order corresponding to it. Definition 2.80. We say that the point w 0 yields the Pontryagin γ -sufficiency if there exists ε > 0 such that for any sequence {δw} ∈ , there exists a number, starting from which the condition σ (δw) ≥ εγ (δw) holds. The equivalent condition for the Pontryagin γ -sufficiency consists of the following: There is no sequence {δw} ∈ such that σ = o(γ ) on it. The violation function σ was already defined in Section 2.2.4. In what follows, it is convenient to use the following expression for σ : tf σ = (δJ )+ + Fi+ (p0 + δp) + |δK| + |δ ẋ − δf | dt, (2.164) t0 2.5. Estimation of the Basic Constant from Below 101 where δJ = J (p0 +δp)−J (p 0 ), δK = K(p0 +δp)−K(p 0 ), δf = f (t, w0 +δw)−f (t, w 0 ), and a + = max{a, 0}. This expression differs from the expression from Section 2.2 by only a constant multiplier (more precisely, they estimate each other from above and from below with constant multipliers), and, therefore, it also can be used in all the formulations. Therefore, Theorem 1.7(b) implies the following theorem. Theorem 2.81. The condition Cγ > 0 is equivalent to the Pontryagin γ -sufficiency at the point w 0 . Theorems 2.79 and 2.81 imply the following theorem. Theorem 2.82. Condition B is equivalent to the existence of an admissible function such that the Pontryagin γ -sufficiency holds at the point w0 for the order γ corresponding to it. Since the Pontryagin γ -sufficiency implies the strict Pontryagin minimum, Condition B is also sufficient for the latter. Therefore, Theorem 2.5 is proved. However, Theorem 2.82 is a considerably stronger result than Theorem 2.5. It allows us to proceed more efficiently in analyzing sufficient conditions. We will see this in what follows. 2.5.13 An Important Estimate We devote this section to a certain estimate, which will be needed in Section 2.7 for obtaining the sufficient conditions for the strong minimum. Let an admissible function and a constant C be such that the set M(C) is nonempty. Let M be a nonempty compact set in M(C). According to (2.111), we have Cγ ≥ Cγ (M ; o(√γ ) ), where Cγ (M ; o(√γ ) ) = inf√ lim o( γ) (2.165) M γ and M (δw) = maxλ∈M (λ, δw). We can further estimate Cγ (M ; o(√γ ) ) from below exactly in the same way as was done for the constant Cγ (C ; o(√γ ) ) when M = M co (C). All the arguments are repeated literally (see relations (2.118), (2.121), (2.124)–(2.128) and Lemmas 2.61, 2.63). As a result, we arrive at the following estimate: Cγ (M ; o(√γ ) ) ≥ min{Cγ̄ (M ; K), qCγ̄ (M ; K), C, qC)}, where 0 < q ≤ 1, M (z̄) = maxλ∈M λ (z̄), and M (z̄) Cγ̄ (M ; K) = inf γ̄ (z̄) (2.166) z̄ ∈ K \ {0} . Now let M ⊂ Leg+ (M0+ ) be a nonempty compact set, and let there exist a constant C > 0 such that (2.167) max λ (z̄) ≥ C γ̄ (z̄) ∀ z̄ ∈ K M 102 Chapter 2. Quadratic Conditions in the Calculus of Variations (i.e., Condition B holds). Then by Lemma 2.73, there exist an admissible function and a constant C1 such that M ⊂ M(C1 ). Condition (2.167) implies Cγ̄ (M ; K) ≥ C. Then (2.166) implies Cγ (M ; o(√γ ) ) ≥ q min{C, C1 }. We set CM = 12 q min{C, C1 }. Then Cγ (M ; o(√γ ) ) > CM . Therefore, M ≥ CM · γ | o(√γ ) , (2.168) i.e., for any sequence {δw} ∈ o(√γ ) , there exists a number starting from which we have M ≥ CM γ . We have obtained the following result. Lemma 2.83. Let M ⊂ Leg+ (M0+ ) be a nonempty compact set, and let there exist C > 0 such that condition (2.167) holds. Then there exists a constant CM > 0 such that condition (2.168) holds. 2.6 2.6.1 Completing the Proof of Theorem 2.4 Replacement of M0co by M0 in the Necessary Conditions The purpose of this section is to complete the proof of Theorem 2.4, which we began in Section 2.4. Here we will not use the results of Section 2.5. Instead, we shall need some constructions from [79, Part 1, Chapter 2, Section 7]. Let us note that the proofs in this section are rather technical and could be omitted in a first reading of the book. We now turn to the following quadratic necessary Condition Aco for the Pontryagin minimum obtained in Section 2.4 (see Theorem 2.43): max λ (z̄) ≥ 0 co M0 ∀ z̄ ∈ K. As was already noted, it is slightly weaker than the necessary condition of Theorem 2.4, since in Condition A, we have the set M0 , which is more narrow than the set M0co . However, we will show that the obtained necessary condition remains valid under the replacement of M0co by M0 , i.e., the necessary Condition A holds. We thus will complete the proof of Theorem 2.4. The passage to the auxiliary problem in [79, Part 1, Chapter 2, Section 7] and the trajectory of this problem corresponding to the index ζ chosen in a special way allows us to do this. For this trajectory, we write the necessary Condition Aco in the auxiliary problem with the subsequent transform of this condition into the initial problem. Such a method was already used in [79, Section 7, Part 1] in proving the maximum principle. We use the notation, the concepts, and the results of [79, Section 7, Part 1], briefly mentioning the main constructions. We stress that in contrast to [79, Section 7, Part 1], all the constructions here refer to the problem on a fixed closed interval of time [t0 , tf ]. We write the condition that the endpoints of the closed interval of time [t0 , tf ] are fixed as follows: t0 = t00 , tf = tf0 . Therefore, let us consider the problem (2.1)–(2.4) in the form which corresponds to the general problem considered in [79, Section 7, Part 1], J (x0 , xf ) → min, (x0 , xf ) ∈ P , F (x0 , xf ) ≤ 0, t0 − t00 dx = f (t, x, u), dt = 0, K(x0 , xf ) = 0, tf − tf0 = 0; (t, x, u) ∈ Q, (2.169) (2.170) 2.6. Completing the Proof of Theorem 2.4 103 where x0 = x(t0 ), xf = x(tf ), and let ŵ0 = (x̂ 0 (t), û0 (t) | t ∈ [t0 , tf ]) (2.171) be a Pontryagin minimum point in this problem. (Here the components x 0 and u0 of the pair w0 are denoted by x̂ 0 and û0 , respectively, as in [79, Section 7, Part 1].) Then the minimum principle holds, and hence the set M0 is nonempty. 2.6.2 Two Cases We have the following two possibilities: (a) There exists λ ∈ M0 such that −λ ∈ M0 . (b) There is no λ ∈ M0 such that −λ ∈ M0 , i.e., M0 ∩ (−M0 ) = ∅. In case (a), the necessary Condition A holds trivially, since for any z̄, at least one of the quadratic forms λ (z̄) and (−λ) (z̄) is nonnegative. Therefore, we consider case (b). As in [79, Section 7, Part 1], for a given number N , we denote by ζ = (t i , uik ) a vector 2 N in R × RN d(u) with components t i ∈ (t00 , tf0 ), i = 1, . . . , N , and uik ∈ Rd(u) , i, k = 1, . . . , N , such that t i < t i+1 , i = 1, . . . , N − 1; (t i , x̂ 0 (t i ), uik ) ∈ Q, i, k = 1, . . . , N . Here and below in this section, the fixed time interval is denoted by [t00 , tf0 ], while all t i , i = 1, . . . , N are internal points of this interval. Denote by D() the set of all ζ satisfying the condition t i ∈ / , i = 1, . . . , N . Further, recall the definition of the set ζ in [79, Section 7, Part 1]. For the problem (2.169), (2.170), it consists of tuples μ = (α0 , α, β) such that α0 ≥ 0, α ≥ 0, αF (x̂00 , x̂f0 ) = 0, α0 + αi + |β| = 1, (2.172) and, moreover, there exist absolutely continuous functions ψ̂x (t) and ψ̂t (t) such that ψ̂x (t00 ) = −lx0 , ψ̂x (tf0 ) = lxf , d ψ̂x = ψ̂x (t)fx (t, x̂ 0 (t), û0 (t)), dt d ψ̂t − = ψ̂x (t)ft (t, x̂ 0 (t), û0 (t)), dt t0 f ψ̂x (t)f (t, x̂ 0 (t), û0 (t)) + ψ̂t (t) dt = 0, (2.173) − t00 ψ̂x (t i )f (t i , x̂ 0 (t i ), uik ) + ψ̂t (t i ) ≥ 0, i, k = 1, . . . , N . (2.174) (2.175) (2.176) Here, l = l(x0 , xf , α0 , α, β) = α0 J (x0 , xf ) + αF (x0 , xf ) + βK(x0 , xf ) (2.177) and the gradients lx0 and lxf in the transversality conditions (2.173) are taken at the point (x̂00 , x̂10 , α0 , α, β) (note that the components ψ of the tuple λ are denoted by ψx here). Let " = ζ , ζ 104 Chapter 2. Quadratic Conditions in the Calculus of Variations where the intersection is taken over all subscripts ζ . At the end of Section 7 in [79, Part 1], we have shown that elements μ ∈ satisfy the following minimum principle: ψ̂x (t)f (t, x̂ 0 (t), u) + ψ̂t (t) ≥ 0 if t ∈ [t00 , tf0 ], u ∈ Rd(u) , ψ̂x (t)f (t, x̂ 0 (t), û0 (t)) + ψ̂t (t) = 0 (t, x 0 (t), u) ∈ Q; a.e. on [t00 , tf0 ]. (2.178) (2.179) By continuity, the latter condition extends to all points of the set [t00 , tf0 ] \ . This implies ⊂ N0 , where N0 is the projection of the set M0 under the injective mapping λ = (α0 , α, β, ψx ) → μ = (α0 , α, β). We now consider the set " () = ζ . ζ ∈D() Clearly, elements μ ∈ () satisfy condition (2.178) at all points t ∈ [t00 , tf0 ] \ ; however, by continuity, this condition extends to all points of the interval [t00 , tf0 ]. Consequently, () = , and, therefore, " ζ ⊂ N0 ζ ∈D() (in fact, we have an equality here). In case (b) the following assertion holds. Proposition 2.84. There exists a subscript ζ such that ζ ∩ (−ζ ) = ∅, and, moreover, the following condition holds for all instants of time t i , i = 1, . . . , N , entering the definition of the subscript ζ : t i ∈ / , i = 1, . . . , N . Proof. Assume that the proposition does not hold. Then each of the sets Fζ := ζ ∩ (−ζ ), ζ ∈ D(), is not empty. These sets compose a centered system of nonempty compact sets, and hence their intersection " F := Fζ ζ ∈D() is nonempty. Moreover, F ⊂ () ⊂ N0 , and the condition Fζ = −Fζ for all ζ ∈ D() implies F = −F . Let μ = (α0 , α, β) ∈ F . Then μ ∈ N0 and (−μ) ∈ N0 ; therefore, we have λ ∈ M0 and (−λ) ∈ M0 for the corresponding element λ. But this contradicts case (b) considered. The proposition is proved. 2.6.3 Problem ZN and Trajectory κN Fix the subscript ζ from Proposition 2.84. For a given N , consider Problem ZN on a fixed closed interval of time [τ0 , τf ], where τ0 = 0, τf = tf − t0 + N 2 . Problem ZN has the form J (x(τ0 ), x(τf )) → inf ; F (x(τ0 ), x(τf )) ≤ 0, K(x(τ0 ), x(τf )) = 0, (x(τ0 ), x(τf )) ∈ P , t(τ0 ) − t00 = 0, t(τf ) − tf0 = 0, −z(τ0 ) ≤ 0, (2.180) 2.6. Completing the Proof of Theorem 2.4 dx = (ϕ(η)z)f (t, x, u); dτ (t, x, u) ∈ Q, η ∈ Q1 . dt = ϕ(η)z; dτ 105 dz = 0; dτ (2.181) Here, z and η are of dimension N 2 + 1 and have the components zθ , zik i, k = 1, . . . , N , and ηθ , ηik , i, k = 1, . . . , N , respectively.2 The open set Q1 is the union of disjoint neighborhoods Qθ and Qik of the 2 points eθ and eik , i, k = 1, . . . , N , respectively, which are the standard basis of RN +1 (eθ has θ ik the unit component eθ , and other components of e are zero, while e has the unit component eik , and other components of eik are zero), and φ(η) : Q1 → Rd(η) is a function mapping each of the mentioned neighborhoods into the element of the basis whose neighborhood it is. Note that the functions u(τ ) and η(τ ) are controls, while the functions z(τ ), x(τ ), and t(τ ) are state variables in Problem ZN . Recall the definition of the point κ ζ = (z0 (τ ), x 0 (τ ), t 0 (τ ), u0 (τ ), η0 (τ )) (in Problem ZN on the closed interval [τ0 , τf ]) corresponding to the subscript ζ and the trajectory (x̂ 0 (t), û0 (t)), t ∈ [t00 , tf0 ]. We “insert” N closed intervals of unit length adjusting to each other into each point t i of the closed interval [t00 , tf0 ]; thus, we enlarge the length of the closed interval by N 2 . Place the left endpoint of the new closed interval at zero. We obtain the closed interval [τ0 , τf ] (τ0 = 0) with N 2 closed intervals (denoted by ik , i, k = 1, . . . , N ) placed on it; moreover, i1 , . . . , iN are closed intervals adjusted to each other, located in the same order, and corresponding to the point t i , i = 1, . . . , N . We set E = (τ0 , τf ) \ n ij . i,j =1 Let χE and χij be the characteristic functions of the sets E and ij , respectively. We set zθ0 = 1, 0 zij = 0, i, j = 1, . . . , N , i.e., z0 = eθ . Further, we set η0 (τ ) = eθ χE (τ ) + eij χij (τ ). i,j Since ϕ(η0 ) = η0 , we have ϕ(η0 )z0 = η0 z0 = η0 eθ = χE . Define t 0 (τ ) by the conditions dt 0 = ϕ(η0 )z0 , dτ t 0 (τ0 ) = t00 . Then t 0 (τf ) = tf0 , since meas E = tf0 − t00 . We set x 0 (τ ) = x̂ 0 (t 0 (τ )), u0 (τ ) = û0 (t 0 (τ )). As was shown in [79, Part 1, Section 7, Proposition 7.1], the point κ ζ defined in such a way 2 We preserve notation accepted in [79, Section 7, Part 1], where z and η were used to denote “zeroθ θ components” of vectors z and η, respectively. 106 Chapter 2. Quadratic Conditions in the Calculus of Variations that it is a Pontryagin minimum point in Problem ZN on the fixed closed interval [τ0 , τf ]. In what follows, all the functions and sets related to N or ζ are endowed with the indices N or ζ , respectively. Since κ ζ yields the Pontryagin minimum in Problem ZN , the necessary Condition co ζ holds for it in this problem. We show that the necessary Condition Aζ also holds for A ζ ζ the chosen index ζ . For this purpose, in Problem ZN , we consider the sets ζ , 0 , co 0 , ζ co ζ M0 , and M0 for the trajectory κ ζ and find the relations between them. The definition of the set ζ was given in [79, Section 7, Part 1], and the other sets were defined in [79, Part 2]. Condition Aζ 2.6.4 The function l N has the form lN = α0 J + αF + βK − αz z0 + βt0 (t0 − t00 ) + βtf (tf − tf0 ) = l − αz z0 + βt0 (t0 − t00 ) + βtf (tf − tf0 ). The Pontryagin function H N has the form H N = ψx (ϕ(η)z)f (t, x, u) + ψt (ϕ(η)z) + ψz · 0 = (ϕ(η)z)(H + ψt ) + ψz · 0, ζ where H = ψx f (t, x, u). The set 0 consists of tuples λN = (α0 , α, β, αz , βt0 , βtf , ψx (τ ), ψt (τ ), ψz (τ )) (2.182) such that α0 ≥ 0, α ≥ 0, αz ≥ 0, αF (x00 , xf0 ) = 0, α0 + αi + αzi + |β| + |βt0 | + |βtf | = 1, ψx (τ0 ) = −lx0 , αz z0 (τ0 ) = 0, ψx (τf ) = lxf , ψt (τ0 ) = −βt0 , ψt (τf ) = βtf , ψz (τ0 ) = αz , ψz (τf ) = 0, dψx 0 − = ϕ(η (τ ))z0 ψx (τ )fx (t 0 (τ ), x 0 (τ ), u0 (τ )), dτ dψt 0 = ϕ(η (τ ))z0 ψx (τ )ft (t 0 (τ ), x 0 (τ ), u0 (τ )), − dτ dψz = ϕ(η0 (τ )) ψx (τ )f (t 0 (τ ), x 0 (τ ), u0 (τ )) + ψt (τ ) , − dτ 0 ϕ(η (τ ))z0 ψx (τ )fu (t 0 (τ ), x 0 (τ ), u0 (τ )) = 0. (2.183) (2.184) (2.185) (2.186) (2.187) (2.188) (2.189) (2.190) (2.191) The gradients lx0 and lxf are taken at the point (x00 , xf0 , α0 , α, β). In [79, Section 7, Part 1], we have shown that there is the following equivalent ζ normalization for the set 0 : α0 + αi + |β| = 1 (2.192) 2.6. Completing the Proof of Theorem 2.4 107 (the conditions α0 = 0, α = 0, and β = 0, and also conditions (2.183), (2.185)–(2.191) ζ imply αz = 0, βt0 = 0, and βtf = 0). Therefore, in the definition of 0 , we can replace normalization (2.184) by the equivalent normalization (2.192). In this case, the quadratic Condition Aco ζ remains valid. Assume that we have made this replacement. The new set ζ is denoted by 0 as before. In [79, Section 7, Part 1], it was also shown that the element ζ (α0 , α, β) ∈ ζ corresponds to an element λN ∈ 0 and has the same components α0 , α, ζ and β, i.e., the projection λN → (α0 , α, β) maps 0 into ζ . ζ Proposition 2.85. The convex hull co 0 does not contain zero. ζ Proof. Assume that this is not true. Then there exist an element λN ∈ 0 and a number ζ ρ > 0 such that −ρλN ∈ 0 . This implies that all nonnegative components α0 , α, and αz of ζ ζ the element λN (see (2.182)) vanish. But then the condition λN ∈ 0 implies −λN ∈ 0 , i.e., we may set ρ = 1. Let μN = (α0 , α, β) = (0, 0, β) be the projection of the element λN . Then μN and N −μ belong to ζ . But the existence of such a μN contradicts the choice of the index ζ . Therefore, the assumption that 0 ∈ co ζ is wrong. The proposition is proved. Proposition 2.85 implies the following assertion. ζ ζ Corollary 2.86. For any λN ∈ co 0 , there exists ρ > 0 such that ρλN ∈ 0 . ζ ζ Proof. Let λN ∈ co 0 . Then by Proposition 2.85, λN = 0. Obviously, co 0 is contained ζ ζ def ζ in the cone con 0 spanned by 0 . The conditions λN ∈ con 0 , λN = 0 imply ν(λN ) = α0 + |α| + |β| > 0 (since ν = 1 is a normalization condition). We set λ̃N = λN . ν(λN ) ζ ζ Then λ̃N ∈ con 0 and ν(λ̃N ) = 1. Therefore, λ̃N ∈ 0 . It remains to set ρ = 1/ν(λN ). The proposition is proved. co ζ Corollary 2.86 and the definitions of the sets M0 assertion. ζ and M0 imply the following co ζ ζ Corollary 2.87. Let λN ∈ M0 . Then there exists ρ > 0 such that ρλN ∈ M0 . The condition Aco ζ for the point κ ζ in Problem ZN has the form max ζ (λN ; z̄N ) ≥ 0 co ζ ∀ z̄N ∈ K ζ . M0 Here, K ζ is the critical cone and ζ is the quadratic form of Problem ZN at the point κ ζ . Let us show that this implies Condition Aζ : max ζ (λN ; z̄N ) ≥ 0 ζ M0 ∀ z̄N ∈ K ζ . 108 Chapter 2. Quadratic Conditions in the Calculus of Variations co ζ Indeed, let z̄N ∈ K ζ . Condition Aco ζ implies the existence of λN ∈ M0 such that ζ (λN ; z̄N ) ≥ 0. (2.193) ζ According to Corollary 2.87, there exists ρ > 0 such that λ̃N = ρλN ∈ M0 . Multiplying (2.193) by ρ > 0, we obtain ζ (λ̃N ; z̄N ) ≥ 0. Hence, maxM ζ ζ (·, z̄N ) ≥ 0. Since z̄N is an 0 arbitrary element in K ζ , this implies Condition Aζ . Thus, we have proved the following lemma. Lemma 2.88. Let ŵ0 be a Pontryagin minimum point in the problem (2.169), (2.170), and let M0 ∩ (−M0 ) = ∅. Then there exists a superscript ζ ∈ D() such that Condition Aζ holds. In what follows, we fix a superscript ζ ∈ D() such that Condition Aζ holds. Now our goal is to reveal which information about the trajectory ŵ 0 can be extracted from Condition Aζ of superscript ζ . We show that Condition Aζ implies Condition A at the point ŵ0 in the initial problem. For this purpose, consider in more detail the definitions of the ζ set M0 , cone K ζ , and quadratic form ζ at the point κ ζ in Problem ZN . 2.6.5 ζ Relation Between the Sets M0 and M0 ζ ζ ζ Consider the conditions defining the set M0 . By definition, M0 is the set of λN ∈ 0 such that the following inequality holds for all τ in the closed interval [τ0 , τf ], except for a finite set of discontinuity points of the controls u0 (τ ) and η0 (τ ): (ϕ(η)z0 ) ψx (τ )f (t 0 (τ ), x 0 (τ ), u) + ψt (τ ) ≥ (ϕ(η0 )z0 ) ψx (τ )f (t 0 (τ ), x 0 (τ ), u0 (τ )) + ψt (τ ) (2.194) for all u ∈ Rd(u) such that (t 0 (τ ), x 0 (τ ), u) ∈ Q and all η ∈ Q1 . Let us analyze condition (2.194). Choose a function η = η(τ ) ∈ Q1 so that the following condition holds: ϕ(η(τ ))z0 = 0. (2.195) Such a choice is possible, since the condition z0 = eθ and the definition of the function ϕ imply 1, η ∈ Qθ , 0 ϕ(η)z = ϕθ (η) = 0, η ∈ / Qθ , and, therefore, we may set η(τ ) = η∗ , where η∗ is an arbitrary point in Q1 \ Qθ , for example, η∗ = e11 . Therefore, condition (2.195) holds for η(τ ) ≡ e11 . It follows from (2.194) and (2.195) that the right-hand side of inequality (2.194) is nonpositive. But the integral of it over the interval [τ0 , τf ] vanishes (this was shown in [79, Section 7, Part 1]; moreover, this follows from the adjoint equation (2.190), conditions (2.183), and the transversality conditions (2.187) considered for the component ψzθ only). But if the integral of a nonpositive 2.6. Completing the Proof of Theorem 2.4 109 function over a closed interval vanishes, then this function equals zero almost everywhere on this closed interval. Hence (ϕ(η0 (τ ))z0 ) ψx (τ )f (t 0 (τ ), x 0 (τ ), u0 (τ )) + ψt (τ ) = 0 (2.196) a.e. on [τ0 , τf ]. Further, setting η = η0 (τ ) in (2.194) and taking into account (2.196), we obtain (ϕ(η0 (τ ))z0 ) ψx (τ )f (t 0 (τ ), x 0 (τ ), u) + ψt (τ ) ≥ 0 if (t 0 (τ ), x 0 (τ ), u) ∈ Q. (2.197) This condition also holds for almost all τ ∈ [τ0 , τf ]. We may rewrite conditions (2.196) and (2.197) for the independent variable t. Recall that in [79, Section 7, Part 1], we have denoted by Ê the image of the set E under the mapping t 0 (τ ). Also, we note that t 0 (τ ) defines a one-to-one and bi-absolutely continuous correspondence between E and Ê , and, moreover, [t00 , tf0 ] \ Ê is a finite set of points {t i }N i=1 , and hence Ê is of full measure in [t00 , tf0 ]. We have denoted by τ 0 (t) the inverse function mapping Ê onto E . The function τ 0 (t) monotonically increases on Ê . Let us extend it to the whole closed interval [t00 , tf0 ] so that the extended function is left continuous. As before, this function is denoted by τ 0 (t). We set We note that ψ̂x (t) = ψx (τ 0 (t)), ψ̂t (t) = ψt (τ 0 (t)). (2.198) x̂ 0 (t) = x 0 (τ 0 (t)), û0 (t) = u0 (τ 0 (t)). (2.199) The first equation holds on [t00 , tf0 ], and the second holds at every continuity point of the function û0 (t), i.e., on the set [t00 , tf0 ] \ . Also, recall that ϕ(η0 (τ ))z0 = χE (τ ), and hence ϕ(η0 (τ 0 (t)))z0 = χE (τ 0 (t)) = 1 (2.200) a.e. on [t00 , tf0 ]. Setting τ = τ 0 (t) in conditions (2.196) and (2.197) and taking into account (2.198)–(2.200), for almost all t ∈ [t00 , tf0 ], we obtain ψ̂x (t)f (t, x̂ 0 (t), û0 (t)) + ψ̂t (t) = 0, (2.201) ψ̂x (t)f (t, x̂ (t), u) + ψ̂t (t) ≥ 0 (2.202) 0 if (t, x̂ 0 (t), u) ∈ Q, u ∈ Rd(u) . Condition (2.201), which holds a.e. on [t00 , tf0 ], also holds at every continuity point of the function û0 (t), i.e., on [t00 , tf0 ] \ ; condition (2.202) holds for all t ∈ [t00 , tf0 ], since all functions entering this condition are continuous. In [79, Section 7, Part 1], we have proved that equations (2.188) and (2.189) imply the equations d ψ̂x = ψ̂x (t)fx (t, x̂ 0 (t), û0 (t)), dt d ψ̂t = ψ̂t (t)ft (t, x̂ 0 (t), û0 (t)). − dt − (2.203) (2.204) 110 Chapter 2. Quadratic Conditions in the Calculus of Variations In proving this, we use the change τ = τ 0 (t) and the condition dt 0 = ϕ(η0 (τ ))z0 . dτ Finally, the transversality conditions (2.185) imply the following transversality conditions: ψ̂x (t00 ) = −lx0 (x̂00 , x̂10 ), ψ̂x (tf0 ) = lxf (x̂00 , x̂10 ), (2.205) since t 0 (τ0 ) = t00 and t 0 (τf ) = tf0 . Conditions (2.201)–(2.205) and conditions (2.183) and (2.192), which hold for a tuple (α0 , α, β, ψx (t), ψt (t)), imply that its projection (α0 , α, β, ψx (t)) belongs to the set M0 of the problem (2.169), (2.170) at the point ŵ 0 (t). Therefore, for the superscript ζ indicated in Lemma 2.88 and corresponding function τ 0 (t) (defined above), we have proved the following assertion. Lemma 2.89. Let a tuple λN = (α0 , α, β, αz , βt0 , βtf , ψx (τ ), ψt (τ ), ψz (τ )), belong to the ζ set M0 of Problem ZN at the point κ ζ . We set ψ̂x (t) = ψx (τ 0 (t)). Then the tuple λ = (α0 , α, β, ψ̂x (t)) belongs to the set M0 of the problem (2.169), (2.170) at the point ŵ0 . 2.6.6 Critical Cone K ζ and Its Relation to the Critical Cone K The discontinuity points τk = τ 0 (tk ), k = 1, . . . , s, of the function u0 (τ ) = û0 (t 0 (τ )) correspond to the discontinuity points tk ∈ , k = 1, . . . , s, of the function û0 (t). We set ˜ ζ = {τ̃i }s̃ be the ζ = {τk }sk=1 . The condition ζ ∈ D() implies ζ ⊂ E . Further, let i=1 0 set of discontinuity points of the control η (τ ). The definition of the function η0 (τ ) implies ˜ ζ does not intersect the open set E . Therefore, the sets ζ and ˜ ζ are disjoint. We that denote their union by (ζ ). By definition, the critical cone K ζ for the trajectory κ ζ in Problem ZN consists of the tuples z̄N = (ξ̄ , ξ̃ , t¯(τ ), x̄(τ ), z̄(τ ), ū(τ ), η̄(τ )) (2.206) such that x̄ ∈ P(ζ ) W 1,2 ([τ0 , τf ], Rd(x) ), t¯ ∈ P(ζ ) W 1,2 ([τ0 , τf ], R1 ), z̄ ∈ P(ζ ) W 1,2 ([τ0 , τf ], Rd(z) ), ū ∈ L2 ([τ0 , τf ], Rd(u) ), η̄ ∈ L2 ([τ0 , τf ], Rd(η) ), ξ̄ ∈ Rs , ξ̃ ∈ Rs̃ , Jp p̄ ≤ 0, Fip p̄ ≤ 0 (i ∈ I ), Kp p̄ = 0, (2.207) (2.208) where p̄ = (x̄(τ0 ), x̄(τf )) and the gradients Jp , Fip , and Kp are taken at the point (x̂00 , x̂10 ) = p̂ 0 , (2.209) t¯(τ0 ) = 0, t¯(τf ) = 0, d x̄ dτ = (ϕ(η0 (τ ))z0 ) ft t¯(τ ) + fx x̄(τ ) + fu ū(τ ) + ((ϕη (η0 (τ ))η̄(τ ))z̄0 )f (t 0 (τ ), x 0 (τ ), u0 (τ )) + (ϕ(η0 (τ ))z̄(τ ))f (t 0 (τ ), x 0 (τ ), u0 (τ )), (2.210) 2.6. Completing the Proof of Theorem 2.4 111 where the gradients fx , fu , and ft are taken at the trajectory (t 0 (τ ), x 0 (τ ), u0 (τ )), d t¯ = (ϕη (η0 (τ ))η̄(τ ))z0 + ϕ(η0 (τ ))z̄, dτ d z̄ = 0, dτ (2.211) [t¯](τk ) = [ϕ(η0 )z0 ](τk )ξ̄k , [x̄](τk ) = [(ϕ(η0 )z0 )f (t 0 , x 0 , u0 )](τk )ξ̄k , z̄(τk ) = 0, k = 1, . . . , s, [x̄](τ̃i ) = [(ϕ(η0 )z0 )f (t 0 , x 0 , u0 )](τ̃i )ξ̃i , [z̄](τ̃i ) = 0, i = 1, . . . , s̃. (2.212) [t¯](τ̃i ) = [ϕ(η0 )z0 ](τ̃i )ξ̃i , (2.213) Here, [ · ](τk ) is the jump at the point τk , and [·](τ̃i ) is the jump at the point τ̃i . We set t¯(τ ) = 0, z̄(τ ) = 0, η̄(τ ) = 0, ξ̃ = 0. (2.214) ζ These conditions define the subcone K0 of the cone K ζ such that the following conditions hold: x̄(τ ) ∈ Pζ W 1,2 ([τ0 , τf ], Rd(x) ), ū(τ ) ∈ L2 ([τ0 , τf ], Rd(u) ), Jp p̄ ≤ 0, Fip p̄ ≤ 0 (i ∈ I ), Kp p̄ = 0, d x̄ = (ϕ(η0 (τ ))z0 ) fx (t 0 (τ ), x 0 (τ ), u0 (τ ))x̄(τ ) dτ +fu (t 0 (τ ), x 0 (τ ), u0 (τ ))ū(τ ) ; # $ [x̄](τk ) = (ϕ(η0 )z0 )f (t 0 , x 0 , u0 ) (τk )ξ̄k , k = 1, . . . , s. ξ̄ ∈ Rs , (2.215) (2.216) (2.217) (2.218) The following assertion holds. Lemma 2.90. Let ˆ ū(t) ˆ z̄ˆ = ξ̄ˆ , x̄(t), (2.219) be an arbitrary element of the critical cone K of the problem (2.169), (2.170) at the point ŵ 0 . We set ˆ 0 (τ )), ū(τ ) = ū(t ˆ 0 (τ )), ξ̄ = ξ̄ˆ , x̄(τ ) = x̄(t (2.220) t¯(τ ) = 0, z̄(τ ) = 0, η̄(τ ) = 0, ξ̃ = 0. Then z̄N = (ξ̄ , ξ̃ , t¯(τ ), x̄(τ ), z̄(τ ), ū(τ ), η̄(τ )) ζ K0 Kζ , (2.221) Kζ is an element of the cone ⊂ where is the critical cone of Problem ZN at the ζ ζ point κ , and K0 is defined by conditions (2.214)–(2.218). Proof. Let z̄ˆ be an arbitrary element of the critical cone K of the problem (2.169), (2.170) at the point ŵ0 having the form (2.219). Then by the definition of the cone K, we have ξ̄ˆ ∈ Rs , ˆ ∈ P W 1,2 ([t00 , tf0 ], Rd(x) ), x̄(t) Jp (p̂0 )p̄ˆ ≤ 0, Fip (p̂0 )p̄ˆ ≤ 0 (i ∈ I ), ˆ ∈ L2 ([t00 , tf0 ], Rd(u) ), ū(t) Kp (p̂0 )p̄ˆ = 0, ˆ d x̄(t) ˆ + fu (t, x̂ 0 (t), û0 (t))ū(t), ˆ = fx (t, x̂ 0 (t), û0 (t))x̄(t) dt ˆ k ) = [f (·, x 0 , u0 )](tk )ξ̄ˆk , k = 1, . . . , s. [x̄](t (2.222) (2.223) (2.224) (2.225) 112 Chapter 2. Quadratic Conditions in the Calculus of Variations Let conditions (2.220) hold. We show that all conditions (2.214)–(2.218) defining the cone ζ K0 hold for the element z̄N (having form (2.221)). Conditions (2.214) follow from (2.220). Conditions (2.215) follow from (2.222). ˆ Indeed, the function x̄(t) is piecewise absolutely continuous, and the function t 0 (τ ) is ˆ 0 (τ )) is a piecewise absolutely continuous function Lipschitz continuous. Hence x̄(τ ) = x̄(t whose set of discontinuity points is contained in ζ . Further, ˆ 0 (τ )) d x̄ˆ dt 0 (τ ) d x̄ˆ d x̄(τ ) d x̄(t · . (2.226) = = = χE (τ ) dτ dτ dt 0 dτ dt 0 t=t (τ ) Since χE2 = χE = τf τ0 d x̄(τ ) dτ dt 0 dτ , t=t (τ ) we have 2 dτ = τf τ0 2 d x̄ˆ 0 (t (τ )) dt dt 0 (τ ) dτ = dτ tf0 t00 ˆ d x̄(t) dt 2 dt < +∞. Hence the derivative d x̄/dτ is square Lebesgue integrable. Therefore, x̄(·) ∈ Pζ W 1,2 ([τ0 , τf ], Rd(x) ). Further, consider the integral τf ū(τ )2 dτ = ū(τ )2 dτ + E τ0 ū(τ )2 dτ . [τ0 ,τf ]\E ˆ 0 (τ )) = ū(τ ), assumes finitely many values The function t 0 (τ ), and hence the function ū(t on [τ0 , τf ] \ E ; hence the second integral in the sum is finite. For the first integral, we have 0 dt 0 (τ ) ˆ 0 (τ ))2 dt (τ ) dτ ū(τ )2 dτ = ū(τ )2 χE dτ = ū(τ )2 dτ = ū(t dτ dτ E E E E ˆ 2 dt = ˆ 2 dt < +∞, ū(t) ū(t) = Ê [t00 ,tf0 ] since ūˆ is Lebesgue square integrable. Hence, τ0f ū(τ )2 dτ < +∞, i.e., ū(·) ∈ L2 ([τ0 , tf ], Rd(u) ). Further, condition (2.223) implies condition (2.216), since t 0 (τ0 ) = t00 , ˆ 0 ), x̄(τf ) = x̄(t ˆ 0 ). Consider the variational equation t 0 (τf ) = tf0 , and, therefore, x̄(τ0 ) = x̄(t f 0 (2.224). Making the change t = t 0 (τ ) in it, multiplying by χE (τ ), and taking into account (2.226), we obtain τ d x̄ = χE (τ )(fx (t 0 (τ ), x 0 (τ ), u0 (τ ))x̄(τ ) + fu (t 0 (τ ), x 0 (τ ), u0 (τ ))ū(τ )). dτ But χE (τ ) = ϕ(η0 (τ ))z0 . Therefore, the variational equation (2.217) holds for x̄ and ū. Finally, we show that the jump conditions (2.218) hold. Note that tk = t 0 (τk ), τk ∈ E , k = 1, . . . , s. Consequently, each τk is a continuity point of the function ϕ(η0 (τk ))z0 = 1, η0 (τ ) k = 1, . . . , s. (2.227) and (2.228) 2.6. Completing the Proof of Theorem 2.4 113 It follows from (2.227) and (2.228) that [(ϕ(η0 )z0 )f (t 0 , x 0 , u0 )](τk ) = [f (·, x̂ 0 , û0 )](tk ), k = 1, . . . , s. (2.229) Analogously, ˆ k ), [x̄](τk ) = [x̄](t k = 1, . . . , s. (2.230) Conditions (2.218) follow from (2.225), (2.229), and (2.230) and the relation ξ̄ = ξ̄ˆ . Thereζ fore, all the conditions defining the cone K0 hold for the tuple z̄N . 2.6.7 Quadratic Form ζ and its Relation to the Quadratic Form ζ ζ Let λN ∈ M0 and z̄N ∈ K0 ; hence let condition (2.214) hold for z̄N . The value of the quadratic form ζ (corresponding to the tuple of Lagrange multipliers λN , at the point κ ζ in Problem ZN ) at the element z̄N is denoted by ζ (λN ; z̄N ). Taking into account conditions (2.214), by definition, we obtain 2ζ (λN , z̄N ) = s k (D k (H N )ξ̄k2 + 2[HxN ]k x̄av ξ̄k ) k=1 τf N +lpp p̄, p̄ + Hww w̄(τ ), w̄(τ ) dτ . (2.231) τ0 Here, p̄ = (x̄(τ0 ), x̄(τf )), lpp = lpp (x̂00 , x̂f0 ; α0 , α, β), (2.232) N = (ϕ(η0 (τ ))z0 )Hww (t 0 (τ ), x 0 (τ ), u0 (τ ), ψx (τ )) Hww = χE (τ )Hww (t 0 (τ ), x 0 (τ ), u0 (τ ), ψx (τ )). (2.233) Further, [HxN ]k = [HxN ](τk ), k = 1, . . . , s, where HxN = χE (τ )Hx (t 0 (τ ), x 0 (τ ), u0 (τ ), ψx (τ )). Let ψ̂x (t) = ψx (τ 0 (t)), Hx = Hx (t, x̂ 0 (t), û0 (t), ψ̂x (t)), [Hx ]k = [Hx ](tk ), k = 1, . . . , s. Taking into account that χE (τk ) = 1, ψx (τk ) = ψ̂x (tk ), x 0 (τk ) = x̂ 0 (tk ), u0 (τk −) = û0 (tk −), t 0 (τk ) = tk , k = 1, . . . , s, u0 (τk +) = û0 (tk +), we obtain [HxN ](τk ) = [Hx ](tk ) = [Hx ]k , k = 1, . . . , s. Thus, [HxN ]k = [Hx ]k , k = 1, . . . , s. Finally, by definition, d N D (H ) = − (k H ) dτ k N τ =τk , k = 1, . . . , s. (2.234) 114 Chapter 2. Quadratic Conditions in the Calculus of Variations Since τk is a continuity point of the function η0 (τ ) and ϕ(η0 (τk ))z0 = 1, we have (k H N )(τ ) 0 )ψ (τ ) f (t 0 (τ ), x 0 (τ ), u0 (τ +)) − f (t 0 (τ ), x 0 (τ ), u0 (τ −)) = (ϕ(η0 (τk ))z x k k = ψ̂x (t 0 (τ )) f (t 0 (τ ), x̂ 0 (t 0 (τ )), û0 (tk +)) − f (t 0 (τ )), x̂ 0 (t 0 (τ )), û0 (tk −)) = (k H )(t 0 (τ )). Consequently, d d dt 0 N (k H ) =− · D (H ) = − (k H ) τ =τk dτ dt t=tk dτ τ =τk k N = D k (H )χE (τk ) = D k (H ). (2.235) ˆ = (ξ̄ˆ , x̄(t), ˆ ū(t)) ˆ Let z̄ˆ = (ξ̄ˆ , w̄) be an arbitrary element of the critical cone K. Let = (ξ̄ , ξ̃ , t¯, x̄, z̄, ū, η̄) be the tuple defined according to z̄ˆ by using formulas (2.220). Then ζ by Lemma 2.90, z̄N ∈ K0 . According to (2.233), z̄N tf0 t00 ˆ ˆ Hww (t, x̂ 0 (t), û0 (t), ψ̂(t))w̄(t), w̄(t) dt τf = τ0 τf ˆ 0 (τ )), w̄(t ˆ 0 (τ )) Hww (t 0 (τ ), x̂ 0 (t 0 (τ )), û0 (t 0 (τ )), ψ̂x (t 0 (τ )))w̄(t = = τ 0τf τ0 dt 0 (τ ) dτ dτ Hww (t 0 (τ ), x 0 (τ ), u0 (τ ), ψx (τ ))w̄(τ ), w̄(τ ) χE (τ ) dτ N Hww w̄(τ ), w̄(τ ) dτ . (2.236) ζ Let λ = (α0 , α, β, ψ̂x (t)) be the element corresponding to the tuple λN ∈ M0 , where ψ̂x (t) = ψx (τ 0 (t)). Then λ ∈ M0 according to Lemma 2.89. Recall that by definition, the ˆ for the problem (2.169), (2.170) at the point ŵ0 corresponding to the quadratic form λ (z̄) tuple λ of Lagrange multipliers and calculated at the element z̄ˆ has the form ˆ = 2λ (z̄) s k ˆ ˆ p̄ˆ + ξ̄k ) + lpp p̄, (D k (H )ξ̄ˆk2 + 2[Hx ]k x̄ˆav k=1 tf0 t00 ˆ w̄ˆ dt. Hww w̄, (2.237) ˆ 0 )). Note that ˆ 0 ), x̄(t Here, p̄ˆ = (x̄(t f 0 ˆ p̄ˆ = lpp p̄, p̄ , lpp p̄, (2.238) ˆ 0 ) = x̄(t ˆ 0 (τ0 )) = x̄(τ0 ) and x̄(t ˆ 0 ) = x̄(t ˆ 0 (τf )) = x̄(τf ). Formulas (2.231) and since x̄(t f 0 (2.234)–(2.238), imply ˆ = ζ (λN ; z̄N ). λ (z̄) Therefore, we have proved the following assertion. (2.239) 2.7. Sufficient Conditions for the Strong Minimum 115 Lemma 2.91. Let z̄ˆ be an arbitrary element (2.219) of the critical cone K, and let z̄N ζ ζ be element (2.221) of the cone K0 ⊂ K ζ obtained by formulas (2.220). Let λN ∈ M0 be an arbitrary tuple (2.182), and let λ = (α0 , α, β, ψ̂x (t)) ∈ M0 be the tuple with the same components α0 , α, β and with ψ̂x (t) = ψx (τ 0 (t)) corresponding to it by Lemma 2.89. Then relation (2.239) holds for the quadratic forms. 2.6.8 Proof of Theorem 2.4 Thus, let ŵ0 be a Pontryagin minimum point in the problem (2.169), (2.170). Then the set M0 is nonempty. If, moreover, M0 ∩ (−M0 ) = ∅ (case (a) in Section 2.6.2), then as already mentioned, Condition A holds trivially. Otherwise (case (b) in Section 2.6.2), by ζ Lemma 2.88, there exists a superscript ζ such that Condition Aζ holds, i.e., the set M0 is nonempty and max ζ (λN , z̄N ) ≥ 0 ζ M0 ∀ z̄N ∈ K ζ . (2.240) Let us show that Condition A holds: the set M0 is nonempty and max λ (z̄) ≥ 0 M0 ∀ z̄ ∈ K. (2.241) ζ Take an arbitrary element z̄ˆ ∈ K. According to Lemma 2.90, the element z̄N ∈ K0 ⊂ K ζ ζ corresponds to it by formulas (2.220). By (2.240), for z̄N , there exists λN ∈ M0 such that ζ (λN , z̄N ) ≥ 0. (2.242) ζ The element λ ∈ M0 corresponds to λN ∈ M0 by Lemma 2.89, and, moreover, by Lemma 2.91, we have relation (2.239). It follows from (2.239) and (2.242) that ˆ ≥ 0, λ (z̄) λ ∈ M0 . Since z̄ˆ is an arbitrary element of K, this implies condition (2.241). Therefore, Condition A also holds in case (b). The theorem is completely proved. 2.7 2.7.1 Sufficient Conditions for Bounded Strong and Strong Minima in the Problem on a Fixed Time Interval Strong Minimum In [79, Part 1], we considered the strong minimum conditions related to the solutions of the Hamilton-Jacobi equation. We can say that they were obtained as a result of development of the traditional approach to sufficient strong minimum conditions accepted in the calculus of variations. However, it is remarkable that there exists another, nontraditional approach to the strong minimum sufficient conditions using the strengthening of the quadratic sufficient conditions for a Pontryagin minimum. Roughly speaking, the strengthening consists of assuming certain conditions on the behavior of the function H at infinity. 116 Chapter 2. Quadratic Conditions in the Calculus of Variations This fact, which had been previously absent in the classical calculus of variations, was first discovered by Milyutin when studying problems of the calculus variations and optimal control. We use this fact in this section. We first define the concept of strong minimum, which will be considered here. It is slightly different from the usual concept from the viewpoint of strengthening. The usual concept used in the calculus of variations corresponds to the concept of minimum on the set of sequences {δw} in the space W such that δx C → 0. It is not fully correct to extend it to the canonical problem without any changes. Indeed, in the classical calculus of variations, t it is customary to minimize an integral functional of the form J = t0f F (t, x, u) dt, where u = ẋ. In passing to the canonical problem, we write the integral functional as the terminal functional: J = y(tf ) − y(t0 ), but there arises a new state variable y such that ẏ = F (t, x, u). Clearly, the requirement δy C → 0 must be absent in the canonical problem if we do not want to distort the original concept of strong minimum in rewriting the problem. How can this be taken into account if we have the canonical form in advance, and it is not known from which problem it originates? It is easy to note that the new state variables y arising in rewriting the integral functionals are characterized by the property that they affinely enter the terminal functionals of the canonical form and are completely absent in the control system of the canonical form. These variables are said to be unessential and the other variables are said to be essential. In defining the strong minimum, we take into account only the essential variables. Let us give the precise definition. As before, we consider the canonical problem (2.1)–(2.4) on a fixed closed interval of time [t0 , tf ]. Definition 2.92. A state variable xi (the component xi of a vector x) is said to be unessential if the function f is independent of it and the functions J , F , and K affinely depend on xi0 := xi (t0 ), xif = xi (tf ). The state variables xi without these properties are said to be essential. (One can also use the terms “main” (or “basic”) and “complementary” (or “auxiliary”) variables.) Respectively, we speak about the essential components of the vector x. Denote by x the vector composed of the essential components of the vector x. Similarly, denote by δx the vector-valued function composed of the essential components of the variation δx. Denote by S the set of sequences {δw} in the space W such that |δx(t0 )|+ δx C → 0. Let us give the following definition for problem (2.1)–(2.4). Definition 2.93. We say that w0 is a strong minimum point (with respect to the essential state variables) if it is a minimum point on S . In what follows, the strong minimum with respect to the essential variables will be called the strong minimum, for brevity. By the strict strong minimum we mean the strict minimum on S . Since ⊂ S , the strong minimum implies the Pontryagin minimum. 2.7.2 Bounded Strong Minimum, Sufficient Conditions We now define the concept of bounded strong minimum, which occupies an intermediate place between the strong and Pontryagin minima. 2.7. Sufficient Conditions for the Strong Minimum 117 Definition 2.94. We say that w0 is a bounded strong minimum point if it is a minimum point on the set of sequences {δw} in W satisfying the following conditions: (a) |δx(t0 )| + δx C → 0. (b) For each sequence there exists a compact set C ⊂ Q such that the following condition holds starting from a certain number: (t, x 0 (t), u0 (t) + δu(t)) ∈ C a.e. on [t0 , tf ]. S Denote by the set of sequences {δw} in W satisfying conditions (a) and (b) and also the following additional conditions: (c) starting from a certain number, (p0 + δp) ∈ P , (t, w 0 + δw) ∈ Q. (d) σ (δw) → 0, where σ is the violation function (2.164). Conditions (c) and (d) hold on every sequence “violating the minimum.” Therefore, we S may treat the bounded strong minimum as a minimum on . We will proceed in this way in what follows, since we will need conditions (c) and (d). By the strict bounded strong S S minimum, we mean the strict minimum on . Since ⊂ ⊂ S , the strong minimum implies the bounded strong minimum, and the latter implies the Pontryagin minimum. A remarkable property is that the sufficient conditions obtained in Section 2.5 guarantee not only the Pontryagin minimum but also the bounded strong minimum. This follows from the theorem, which now will be proved. In what follows, w 0 is an admissible point satisfying the standard assumptions of Section 2.1. Theorem 2.95. For a point w 0 , let there exist an admissible function (t, u) and a constant S C > 0 such that the set M(C) is nonempty. Then = , and hence the Pontryagin minimum is equivalent to the bounded strong minimum. To prove this, we need several auxiliary assertions. S t Proposition 2.96. Let λ ∈ 0 , {δw} ∈ . Then ( t0f δH λ dt)+ → 0, where a+ = max{a, 0} and δH λ = H (t, x 0 + δx, u0 + δu, ψ) − H (t, x 0 , u0 , ψ). Proof. Let λ ∈ 0 , and let δw be an admissible variation with respect to Q, i.e., (t, w 0 + δw) ∈ Q. Then tf λ δl − ψ(δ ẋ − δf ) dt ≤ const σ (δw). (2.243) t0 On the other hand, we have shown earlier that the conditions −ψ̇ = Hxλ , ψ(t0 ) = −lxλ0 , and ψ(tf ) = lxλf imply tf t0 t ψδ ẋ dt = ψδx tf − 0 tf ψ̇δx dt = lp δp + t0 tf t0 Hxλ δx dt. Taking into account that ψδf = δH λ , we obtain from inequality (2.243) that δl λ − lp δp − tf t0 Hxλ δx dt + tf t0 δH λ dt ≤ const σ (δw). (2.244) 118 Chapter 2. Quadratic Conditions in the Calculus of Variations S Let {δw} ∈ . The condition δx C → 0 implies tf t0 Hxλ δx dt → 0 and (δl λ − lp δp) → 0. Moreover, σ (δw) → 0. Therefore, condition (2.244) implies ( proposition is proved. tf t0 δH λ dt)+ → 0. The Using Proposition 2.96, we prove the following assertion. Proposition 2.97. Let there exist an admissible function (t, u) and a constant C > 0 such that the set M(C) is nonempty. Then the following condition holds for any sequence S t {δw} ∈ : t0f (t, u0 + δu) dt → 0. Proof. Let C > 0, and let λ ∈ M(C). According to Proposition 2.96, ( Represent δH λ as δH λ = δ̄x H λ + δu H λ , where tf t0 δH λ dt)+ → 0. δ̄x H λ = H λ (t, x 0 + δx, u0 + δu) − H λ (t, x 0 , u0 + δu), δu H λ = H λ (t, x 0 , u0 + δu) − H λ (t, x 0 , u0 ). The conditions δx C → 0, (t, x 0 , u0 + δu) ∈ C, where C ⊂ Q is a compact set, imply t δ̄x H λ ∞ → 0. Hence ( t0f δu H λ dt)+ → 0. Further, the condition λ ∈ M(C) implies tf tf λ 0 t0 δu H dt ≥ C t0 (t, u + δu) dt ≥ 0. This tf imply t0 (t, u0 + δu) dt → 0. The proposition is δu H λ ≥ C(t, u0 + δu) ≥ 0. Consequently, and the condition ( proved. tf λ t0 δu H dt)+ →0 In what follows, (t, u) is a function admissible for the point w 0 . Proposition 2.98. Let C ⊂ Q be a compact set, and let the variation δu ∈ L∞ (, Rd(u) ) be such that (t, x 0 , u0 + δu) ∈ C a.e. on [t0 , tf ]. Then we have the estimate δu 1 ≤ t const( t0f (t, u0 + δu) dt)1/2 , where const depends only on C. Proof. Let V be the neighborhood of the compact set u0 from the definition of the function (t, u) in Section 2.3.7, and let V 0 and V ∗ be subsets of the neighborhood V defined in Section 2.3.1. Represent δu as δu = δuv + δuv , where δu if (t, u0 + δu) ∈ V, δuv = δuv = δu − δuv . 0 otherwise, Further, let the representation δuv = δu0 + δu∗ correspond to the partition V = V 0 ∪ V ∗ : δuv if (t, u0 + δuv ) ∈ V 0 , δuv if (t, u0 + δuv ) ∈ V ∗ , 0 ∗ δu = δu = 0 otherwise, 0 otherwise. Then δu 1 = δuv 1+ δu0 1+ δu∗ 1 . (2.245) Let us estimate each of the summands separately. / V for δuv = 0, by the definition (1) Since (t, x 0 , u0 + δuv ) ∈ C and (t, u0 + δuv ) ∈ of the function (t, u), there exists ε = ε(C) > 0 such that (t, u0 + δuv ) ≥ ε if δuv = 0. 2.7. Sufficient Conditions for the Strong Minimum 119 Moreover, there exists a constant N = N (C) > 0 such that δuv ∞ ≤ δu Consequently, N tf (t, u0 + δuv ) dt. δuv 1 ≤ δuv ∞ · meas{t δuv = 0} ≤ ε t0 ≤ N. − t0 ) ≤ N (tf − t0 ), we obtain from this tf v 2 v v δu 1 ≤ δu ∞ (tf − t0 ) δu 1 ≤ const (t, u0 + δuv ) dt, (2.246) Also, taking into account that δuv that 1 ≤ δuv ∞ ∞ (tf t0 where const > 0 depends on C only. t (2) Since (t, u0 + δu0 ) ∈ V 0 , we have δu0 22 = t0f (t, u0 + δu0 ) dt. Consequently, 1/2 tf (t, u0 + δu0 ) dt . (2.247) δu0 1 ≤ tf − t0 δu0 2 = tf − t0 t0 (3) Obviously, δu∗ 1 = |δu∗ | dt ≤ δu∗ ∞ · meas M ∗ ≤ N · meas M ∗ , where M ∗ = ∗ ∗ ∪ M∗ , {t δu = 0}. Further, as in Section 2.3.1, represent M∗ = ∪ Mk∗ , Mk∗ = Mk− k+ k = 1, . . . , s. Then tf ∗ 2 (meas Mk− ) ≤2 |δtk | dt ≤ (t, u0 + δu∗ ) dt; ∗ Mk− t0 tf ∗ 2 |δtk | dt ≤ (t, u0 + δu∗ ) dt; (meas Mk+ ) ≤ 2 ∗ Mk+ t0 ∗ ∗ meas Mk− + meas Mk+ . meas M∗ = tf t0 k k Consequently, δu∗ 1 ≤ const tf (t, u0 + δu∗ ) dt 1/2 , t0 where const depends only on C. It follows from (2.245)–(2.248) that 1/2 tf 1/2 tf (t, u0 + δuv ) dt + (t, u0 + δu0 ) dt δu 1 ≤ const t0 + tf (t, u0 + δu∗ ) dt t0 1/2 . t0 Now, to obtain the required estimate, it remains to use the inequality ! a + b + c ≤ 3(a 2 + b2 + c2 ), which holds for any numbers a, b, and c, and also the relation tf tf tf 0 v 0 0 (t, u + δu ) dt + (t, u + δu ) dt + (t, u0 + δu∗ ) dt t0 = t0 tf (t, u0 + δu) dt. t0 The proposition is proved. t0 (2.248) 120 Chapter 2. Quadratic Conditions in the Calculus of Variations Proof of Theorem 2.95. Assume that the conditions of the theorem hold. Let us prove ¯ S ⊂ . Let {δw} ∈ ¯ S . Then it follows from Propositions 2.97 and 2.98 the inclusion that δu 1 → 0. Further, the condition σ (δw) → 0 implies δ ẋ − δf 1 → 0. But δf = δ̄x f + δu f , where δ̄x f = f (t, x 0 + δx, u0 + δu) − f (t, x 0 , u0 + δu), δu f = f (t, x 0 , u0 + δu) − f (t, x 0 , u0 ). Since δx C → 0 and there exists a compact set C ⊂ Q such that (t, x 0 , u0 + δu) ∈ C starting from a certain number, we have δ̄x f ∞ → 0. The conditions δu 1 → 0 and (t, x 0 , u0 + δu) ∈ C imply δu f 1 → 0. Consequently, δ ẋ 1 ≤ δ ẋ − δf 1+ δf 1 ≤ δ ẋ − δf 1+ δ̄x f ∞ (tf − t0 ) + δu f 1 → 0. The conditions δ ẋ 1 → 0 and |δx(t0 )| → 0 imply δx 1,1 → 0. Therefore, {δw} ∈ . The ¯ S ⊂ is proved. The converse inclusion always holds. Therefore, ¯ S = . inclusion The theorem is proved. Lemma 2.73 and Theorem 2.95 imply the following theorem. ¯ S , and hence the PonTheorem 2.99. Let the set Leg+ (M0+ ) be nonempty. Then = tryagin minimum is equivalent to the bounded strong minimum. Proof. Assume that Leg+ (M0+ ) is nonempty. Choose an arbitrary compact set M ⊂ Leg+ (M0+ ), e.g., a singleton. According to Lemma 2.73, there exist an admissible function (t, u) and a constant C > 0 such that M ⊂ M(C). Therefore, M(C) is nonempty. Then ¯ S by Theorem 2.95. The theorem is proved. = For a point w 0 and the higher order γ corresponding to an admissible function , we give the following definition. Definition 2.100. We say that the point w 0 is a point of bounded strong γ -sufficiency if ¯ S such that σ = o(γ ) on it. there is no sequence {δw} ∈ The condition that the set Leg+ (M0+ ) is nonempty is a counterpart of Condition B. Therefore, Theorems 2.82 and 2.99 imply the following theorem. Theorem 2.101. Condition B is equivalent to the existence of an admissible function such that the bounded strong γ -sufficiency holds at the point w0 for the higher order γ corresponding to it. The bounded strong γ -sufficiency implies the strict bounded strong minimum. Therefore, Theorem 2.101 implies the following theorem. Theorem 2.102. Condition B is sufficient for the strict bounded strong minimum at the point w0 . At this point, we complete the consideration of conditions for the bounded strong minimum. Before passing to sufficient conditions for the strong minimum, we prove some estimate for the function , which will be needed in what follows. 2.7. Sufficient Conditions for the Strong Minimum 2.7.3 Estimate for the Function 121 on Pontryagin Sequences Recall that in Section 2.1, we introduced the set 0 consisting of those λ ∈ 0 for which [H λ ]k = 0 for all tk ∈ , and in Section 2.3, we showed that there exists a constant C > 0 such that the following estimate holds for any sequence {δw} ∈ loc starting from a certain number: max |(λ, δw)| ≤ C γ (δw) co 0 (see Proposition 2.16). Let us show that the same estimate also holds on any Pontryagin sequence but with a constant depending on the order γ and the sequence. Lemma 2.103. Let the set 0 be nonempty. Let (t, u) be an admissible function, and let γ be the higher order corresponding to it. Then for any sequence {δw} ∈ , there exists a constant C > 0 such that max |(λ, δw)| ≤ Cγ (δw). co 0 Briefly, this property will be written as maxco || ≤ O(γ ) . 0 Proof. Proposition 2.16 implies the following assertion: there exist constants C > 0 and ε > 0 and a neighborhood V of the compact set u0 such that the conditions δw ∈ W , δx C ≤ ε, and (t, u0 + δu) ∈ V imply the estimate maxco |(λ, δw)| ≤ Cγ (δw). Let us 0 use this estimate. Let {δw} be an arbitrary sequence from . For each member δw = (δx, δu) of the sequence {δw}, represent δu as δu = δuV + δuV , where δu if (t, u0 + δu) ∈ V , δuV = δu − δuV . δuV = 0 otherwise, We set δwV = (δx, δuV ) and δwV = (0, δuV ). Owing to a possible decrease of V , we can assume that both sequences {δwV } and {δwV } are admissible with respect to Q; i.e., the conditions (t, x 0 + δx, u0 + δuV ) ∈ Q and (t, x 0 , u0 + δuV ) ∈ Q hold starting from a certain number (such a possibility follows from the definition of Pontryagin sequence). We assume that this condition holds for all numbers and δx C ≤ ε holds for all numbers. We set MV = {t | δuV = 0}. The definitions of admissible function (t, u) and Pontryagin sequence imply the existence of constants 0 < a < b such that a ≤ (t, u0 + δuV ) ≤ b | M V for all members of the sequence. This implies a· meas MV ≤ γ V ≤ b· meas M V , where t γ V = t0f (t, u0 + δuV ) dt = γ (δw V ). Therefore, γ V and meas MV are of the same order of smallness. Moreover, the definitions of γ , δwV , and δwV imply γ (δw) = γ (δwV ) + γ (δwV ), or, briefly, γ = γV + γ V . In what follows, we will need the formula δf = δV f + δ̄ Vf , (2.249) where δf = f (t, w 0 + δw) − f (t, w 0 ), δV f = f (t, w 0 + δwV ) − f (t, w 0 ), and δ̄ Vf = f (t, x 0 + δx, u0 + δuV ) − f (t, x 0 + δx, u0 ). The fulfillment of this formula is proved by the following calculation: δf = f (t, x 0 + δx, u0 + δu) − f (t, x 0 + δx, u0 + δuV ) + δV f = f (t, x 0 + δx, u0 + δu) − f (t, x 0 + δx, u0 + δuV ) χ V + δV f = δ̄ Vf χ V + δV f = δ̄ Vf + δV f , 122 Chapter 2. Quadratic Conditions in the Calculus of Variations where χ V is the characteristic function of the set M V . Formula (2.249) implies the following representation for (λ, δw) on the sequence {δw}: tf tf λ ψδ ẋ dt + ψδf dt (λ, δw) := δl − t0 tf = δl λ − ψδ ẋ dt + t0 = (λ, δwV ) + t0 tf t0 tf tf ψδV f dt + ψ δ̄ Vf dt t0 δ̄ V H λ dt, t0 where δ̄ V H λ = ψ δ̄ Vf and λ ∈ co 0 . This implies the estimate max |(λ, δw)| ≤ max |(λ, δwV )| + max ψ co 0 co 0 co 0 ∞ tf |δ̄ Vf | dt. t0 According to the choice of V and ε, the first term of the sum on the right-hand side of the inequality is estimated through γ (δwV ). The second term of this sum is estimated through meas MV and hence through γ V . Since γV ≤ γ and γ V ≤ γ , the total sum is estimated through γ with a certain positive constant as a multiplier. The lemma is proved. 2.7.4 Sufficient Conditions for the Strong Minimum In this section, we assume that the set Q has the form Q = Qt × Qx × Qu , where Qt ⊂ R, Qx ⊂ Rd(x) , and Qu ⊂ Rd(u) are open sets. Set Qtu = Qt × Qu . We now give those additional requirements which, together with Condition B, turn out to be sufficient for the strong minimum whose definition was given in Section 2.7.1. For (t, x, u) ∈ Q, we set δ¯u f = f (t, x, u) − f (t, x, u0 (t)). Further, for (t, x, u) ∈ Q and λ = (α0 , α, β, ψ) ∈ 0 , we set δ̄u H λ = ψ δ̄u f . The following theorem holds. Theorem 2.104. Let the following conditions hold for the point w0 : (1) There exists a nonempty compact set M ⊂ Leg+ (M0+ ) such that (a) for a certain C > 0, maxλ∈M λ (z̄) ≥ C γ̄ (z̄) for all z̄ ∈ K, i.e., Condition B holds; (b) for any ε > 0, there exist δ > 0 and a compact set C ⊂ Qtu such that for all t ∈ [t0 , tf ] \ , the conditions (t, w) ∈ Q, |x − x 0 (t)| ≤ δ, (t, u) ∈ / C imply minλ∈M δ̄u H λ ≥ −ε|δ̄u f |; (2) there exist δ0 > 0, ε0 > 0, a compact set C0 ⊂ Qtu , and an element λ0 ∈ M0 such that for all t ∈ [t0 , tf ] \ the conditions (t, w) ∈ Q, |x − x 0 (t)| < δ0 , (t, u) ∈ / C0 imply δ̄u H λ0 ≥ ε0 |δ̄u f | > 0. Then w0 is a strict strong minimum point. Remark 2.105. For the point w 0 , let there exist δ0 , ε0 , C0 , and λ0 satisfying condition (2) of Theorem 2.104. Moreover, let λ0 ∈ Leg+ (M0+ ), and for a certain C > 0, let λ0 (z̄) ≥ C γ̄ (z̄) for all z̄ ∈ K. Then, as is easily seen, all the conditions of Theorem 2.104 hold, and, therefore, w0 is a strict strong minimum point. 2.7. Sufficient Conditions for the Strong Minimum 2.7.5 123 Proof of Theorem 2.104 Assume that for a subset M ⊂ Leg+ (M0+ ), C > 0, δ0 > 0, ε0 > 0, C0 ⊂ Qtu , λ0 ∈ M0 , all conditions of the theorem hold, but there is no strict strong minimum at the point w 0 . Let us show that this leads to a contradiction. Since w 0 is not a strict strong minimum point, there exists a sequence {δw} such that |δx(t0 )| + δx C → 0 (i.e., {δw} ∈ S ), and the following conditions hold for all members of this sequence: (p0 +δp) ∈ P , (t, w0 +δw) ∈ Q, σ (δw) = 0, δw = 0. The condition σ (δw) = 0 implies δ ẋ − δf = 0, δK = 0, δJ ≤ 0, F (p 0 + δp) ≤ 0. Hence, for any λ ∈ 0 , we have the following on the sequence {δw}: (λ, δw) = δl λ ≤ σ (δw) = 0. (2.250) Assume that there is an arbitrary compact set C satisfying the condition C0 ⊂ C ⊂ Qtu . (2.251) For each member δw = (δx, δu) of the sequence {δw}, we set δu if (t, u0 + δu) ∈ / C, C δu = 0 otherwise, δwC = (0, δuC ), Then {δw} = {δwC } + {δwC }. δuC = δu − δuC , The relation δH λ δwC = (δx, δuC ). λ = δC H + δ̄ C H λ , λ ∈ 0 , where δC H λ = H λ (t, w0 + δwC ) − H λ (t, w 0 ), δ̄ C H λ = H λ (t, w0 + δw) − H λ (t, w0 + δwC ) = H λ (t, x 0 + δx, u0 + δu) − H λ (t, x 0 + δx, u0 + δuC ) = H λ (t, x 0 + δx, u0 + δuC ) − H λ (t, x 0 + δx, u0 ), corresponds to this representation of the sequence {δw}. This and condition (2.250) imply that for any λ ∈ 0 , we have the following inequality on the sequence {δw}: tf λ (2.252) δ̄ C H λ dt ≤ 0. (δwC ) + t0 We set γC = γ (δwC ), δ̄ Cf = f (t, x 0 + δx, u0 + δuC ) − f (t, x 0 + δx, u0 ), ϕ C = t0f |δ̄ Cf | dt. Since C ⊃ C0 and δx C → 0, condition (2) of the theorem implies ϕ C > 0 for all nonzero members of the sequence {δwC } with sufficiently large numbers. t Proposition 2.106. The following conditions hold : (a) ϕ C → 0, (b) {δwC } ∈ , and hence γC → 0. Proof. By (2.252), we have the following for the sequence {δw} and the element λ = λ0 : tf tf tf δ̄ C H dt ≤ 0 δl − ψδ ẋ dt + δC H dt + (2.253) t0 t0 t0 124 Chapter 2. Quadratic Conditions in the Calculus of Variations (we omit λ = λ0 in this proof). Represent δC H in the form δC H = H (t, w 0 + δwC ) − H (t, w0 ) = δ̂Cx H + δCu H , where δ̂Cx H = H (t, x 0 + δx, u0 + δuC ) − H (t, x 0 , u0 + δuC ), δCu H = H (t, x 0 , u0 + δuC ) − H (t, x 0 , u0 ). Then we obtain from (2.253) that tf tf δl − ψδ ẋ dt + δ̂Cx H dt + t0 The conditions δx t0 C tf t0 δCu H dt + tf δ̄ C H dt ≤ 0. (2.254) t0 → 0 and (t, u0 + δuC ) ∈ C imply tf δ̂Cx H dt → 0. (2.255) t0 Further, the condition δx C → 0 also imply tf t ψδ ẋ dt = δl − ψδx tf + δl − 0 t0 = δl − lp δp − tf tf ψ̇δx dt t0 Hx δx dt → 0. (2.256) t0 Conditions (2.254)–(2.256) imply tf δCu H dt + t0 tf C δ̄ H dt + t0 → 0. (2.257) Since λ = λ0 and C ⊃ C0 , according to assumption (2) of the theorem, the following inequalities hold for all members of the sequence {δw C } with sufficiently large numbers: tf (2.258) δ̄ C H dt ≥ ε0 ϕ C > 0. t0 Conditions (2.257) and (2.258) imply tf δCu H dt + ε0 ϕ C → 0. t0 (2.259) + Since λ = λ0 ∈ M0 and (t, x 0 , u0 + δuC ) ∈ Q, we have δCu H ≥ 0. Therefore, both terms in (2.259) are nonnegative for all sufficiently large numbers of the sequence {δw}. But then t (2.259) implies t0f δCu H dt → 0, ϕ C → 0. ¯ S . For this purpose, we prove that σ (δwC ) → 0 (the We now show that {δwC } ∈ ¯ S obviously hold). Since other conditions of {δwC } belonging to the set of sequences σ (δw) = 0, we need show only that δ ẋ − δC f 1 → 0. We have the following for the sequence {δw}: (2.260) δ ẋ − δf = 0, δf = δC f + δ̄ Cf . 2.7. Sufficient Conditions for the Strong Minimum 125 The condition ϕ C → 0 means that δ̄ Cf 1 → 0. This and (2.260) imply δ ẋ − δC f 1 → 0. ¯ S. Therefore, {δwC } ∈ We now recall that, by condition (1) of the theorem, the set Leg+ (M0+ ) is nonempty. ¯ S = . Therefore, {δwC } ∈ . The proposition By Theorem 2.99, it follows from this that is proved. We continue the proof of the theorem. Consider the following two possible cases for the sequence {δw}. Case (a). Assume that there exist a compact set C satisfying conditions (2.251) and a subsequence of the sequence {δw} such that the following conditions hold on this subsequence: ϕ C > 0, γC = o(ϕ C ). (2.261) Assume that these conditions hold for the sequence {δw} itself. Inequality (2.252) and the conditions λ0 ∈ M0 ⊂ 0 imply that the following inequality holds on the sequence {δw}: tf t0 δ̄ C H λ0 dt ≤ −λ0 (δwC ). (2.262) As was already mentioned in the proof of Proposition 2.106, the following inequalities hold for all members of the sequence {δw} having sufficiently large numbers: ε0 ϕ C ≤ tf δ̄ C H λ0 dt. (2.263) t0 On the other hand, according to Lemma 2.103, the conditions {δwC } ∈ , λ0 ∈ M0 ⊂ 0 imply the estimate (2.264) −λ0 (δwC ) ≤ O(γC ). We obtain from (2.261)–(2.264) that 0 < ϕ C ≤ O(γC ) = o(ϕ C ). This is a contradiction. Case (b). Consider the second possibility. Assume that for any compact set C satisfying conditions (2.251), there exists a constant N > 0 such that the following estimate holds on the sequence {δw}: (2.265) ϕ C ≤ N γC . We show that this also leads to a contradiction. We will thus prove the theorem. First of all, we note that the constant N in (2.265) can be chosen common for all compact sets C satisfying conditions (2.251). Indeed, let N0 correspond to a compact set C0 , i.e., the estimate ϕ C0 ≤ N0 γC0 holds on the sequence {δw}. Let C be an arbitrary compact set such that (2.251) hold. Then we have the following on the sequence {δw} for all sufficiently large numbers: ϕ C ≤ ϕ C0 ≤ N0 γC0 ≤ N0 γC . Therefore, N = N0 is also appropriate for C. Also, we note that for any C satisfying (2.251), there exists a (serial) number of the sequence starting from which γC > 0. Indeed, otherwise, there exist a compact set C, satisfying (2.251) and a subsequence of the sequence {δw} such that γC = 0 on the subsequence, and then ϕ C = 0 by (2.265). By assumption (2) of the theorem, this implies that all members of the subsequence vanish. The latter is impossible, since the sequence {δw} contains nonzero members by assumption. 126 Chapter 2. Quadratic Conditions in the Calculus of Variations Now let the compact set C satisfy conditions (2.251). Inequality (2.252) and the inclusion M ⊂ 0 imply that the following inequality holds on the sequence {δw}: tf max (λ, δwC ) ≤ − min (2.266) δ̄ C H λ dt. M Obviously, M tf min M δ̄ C H λ dt ≥ t0 tf t0 t0 min δ̄ C H λ dt. M (2.267) Condition 1(b) of the theorem implies that, for any ε > 0, there exists a compact set C satisfying (2.251) such that tf min δ̄ C H λ dt ≥ −ε· ϕ C (2.268) t0 M for all sufficiently large numbers of the sequence. We obtain from (2.265)–(2.268) that for any ε > 0, there exists a compact set C satisfying (2.251) such that following estimate holds starting from a certain number: max (λ, δwC ) ≤ εN γC , M (2.269) where N = N0 is independent of C. We now estimate the left-hand side of inequality (2.269) from below. For this purpose, we show that for any compact set C satisfying (2.251), the sequence {δwC } belongs to o(√γ ) . Let C be an arbitrary compact set satisfying (2.251). According to Proposi tion 2.106, {δwC } ∈ . Since σ (δw) = 0, we have (δJ )+ + Fi+ (p 0 + δp) + |δK| = 0 on the whole sequence. Moreover, the conditions δ ẋ = δf , δf = δC f + δ̄ Cf , δ¯C f 1 = ϕ C ≤ O(γC ) imply δ ẋ − δC f 1 ≤ O(γC ). Therefore, {δwC } ∈ σ γ ⊂ o(√γ ) . But then, by Lemma 2.83, condition 1(a) of the theorem implies that there exists a constant CM > 0 such that, starting from a certain number, max (λ, δwC ) ≥ CM γC > 0. M (2.270) Moreover, the constant CM is independent of the sequence from o(√γ ) and hence is independent of C. Comparing estimates (2.269) and (2.270), we obtain the following result: For any ε > 0, there exists a compact set C satisfying (2.251) such that, starting from a certain number, 0 < CM γC ≤ εN γC . Choosing 0 < ε < CM /N , we obtain a contradiction. The theorem is proved. Chapter 3 Quadratic Conditions for Optimal Control Problems with Mixed Control-State Constraints In Sections 3.1 and 3.2 of this chapter, following [92], we extend the quadratic conditions obtained in Chapter 2 to the general problem with the local relation g(t, x, u) = 0 using a special method of projection contained in [79]. In Section 3.3, we extend these conditions to the problem on a variable interval of time using a simple change of time variable. In Section 3.4, we formulate (without proofs) quadratic conditions in an optimal control problem with the local relations g(t, x, u) = 0 and ϕ(t, x, u) ≤ 0. 3.1 Quadratic Necessary Conditions in the Problem with Mixed Control-State Equality Constraints on a Fixed Time Interval 3.1.1 Statement of the Problem with a Local Equality and Passage to an Auxiliary Problem without Local Constraints We consider the following problem on a fixed interval [t0 , tf ] with a local equality-type constraint: F (x(t0 ), x(tf )) ≤ 0, J (x(t0 ), x(tf )) → min, K(x(t0 ), x(tf )) = 0, (x(t0 ), x(tf )) ∈ P , ẋ = f (t, x, u), g(t, x, u) = 0, (t, x, u) ∈ Q. (3.1) (3.2) It is assumed that the functions J , F , and K are twice continuously differentiable on the open set P ⊂ R2d(x) , and f and g are twice continuously differentiable on the open set Q ⊂ R1+d(x)+d(u) . Moreover, the following full-rank condition is assumed for the local equality: rank gu (t, x, u) = d(g) (3.3) for all (t, x, u) ∈ Q such that g(t, x, u) = 0. As in Section 2.1.2, we define a (strict) minimum on a set of sequences S : w 0 is a (strict) minimum point on S in problem (3.1), (3.2) if there is no sequence {δw} ∈ S such 127 128 Chapter 3. Quadratic Conditions for Optimal Control Problems that the following conditions hold for all its members: J (p 0 + δp) < J (p0 ) (J (p 0 + δp) ≤ J (p 0 ), δw = 0), F (p0 + δp) ≤ 0, K(p 0 + δp) = 0, ẋ 0 + δ ẋ = f (t, w 0 + δw), g(t, w 0 + δw) = 0, (p 0 + δp) ∈ P , (t, w 0 + δw) ∈ Q, where p0 = (x 0 (t0 ), x 0 (tf )), δw = (δx, δu), and δp = (δx(t0 ), δx(tf )). A (strict) minimum on (see Section 2.1.3) is said to be a (strict) Pontryagin minimum. Our goal is to obtain quadratic conditions in problem (3.1), (3.2) using the quadratic conditions obtained in problem (2.1)–(2.4). In this case, we will use the same method for passing to a problem without local constraints, which was already used in [79, Section 17, Part 1] for obtaining first-order conditions. Recall that in [79, Section 17, Part 1], we have introduced the set G = {(t, x, u) ∈ Q | g(t, x, u) = 0}. We have shown that there exist a neighborhood Q1 ⊂ Q of the set G and a continuously differentiable function U (t, x, u) : Q1 → Rd(x) such that (i) (t, x, U (t, x, u)) ∈ G ∀ (t, x, u) ∈ Q1 , (ii) U (t, x, u) = u ∀ (t, x, u) ∈ G. (3.4) Owing to these properties, U is called a projection. Since g is a twice continuously differentiable function on Q, we can choose the function U, together with the neighborhood Q1 , so that U is a twice continuously differentiable function on Q1 . This can be easily verified by analyzing the scheme for proving the existence of a projection presented in [79, Section 17, Part 1]. We fix certain Q1 and U with the above properties. Instead of system (3.2) with the local equality-type constraint g(t, x, u) = 0, consider the following system without local constraint: ẋ = f (t, x, U (t, x, u)), (t, x, u) ∈ Q1 . (3.5) We find a connection between the necessary conditions in the problem (3.1), (3.2) and those in the problem (3.1), (3.5). Preparatorily, we prove the following assertion. Proposition 3.1. Let (x 0 , u0 ) = w 0 be a Pontryagin minimum point in the problem (3.1), (3.2). Then w 0 is a Pontryagin minimum point in the problem (3.1), (3.5). Proof. The property that w0 is a Pontryagin minimum point in the problem (3.1), (3.2) implies that w0 is an admissible point in the problem (3.1), (3.2), and then by the second property (3.4) of the projection, w0 is also admissible in the problem (3.1), (3.5). Suppose that w0 is not a Pontryagin minimum point in the problem (3.1), (3.5). Then there exist C ∈ Q1 and a sequence {δw} = {(δx, δu)} such that δx 1,1 → 0, δu 1 → 0, (3.6) and the following conditions hold for all members of the sequence: (t, w 0 + δw) ∈ C, δJ < 0, F (p0 + δp) ≤ 0, K(p 0 + δp) = 0, x˙0 + δ ẋ = f (t, x 0 + δx, U (t, x 0 + δx, u0 + δu)), (3.7) (3.8) (3.9) 3.1. Quadratic Necessary Conditions on a Fixed Time Interval 129 where p0 = (x 0 (t0 ), x 0 (tf )) and δp = (δx(t0 ), δx(tf )). Therefore, {δw} is a Pontryagin sequence “violating the Pontryagin minimum” at the point w 0 in the problem (3.1), (3.5). We set {δw1 } = {(δx, δu1 )}, where δu1 = U (t, x 0 + δx, u0 + δu) − U (t, x 0 , u0 ) = U (t, x 0 + δx, u0 + δu) − u0 . We show that {δw1 } is a Pontryagin sequence “violating the Pontryagin minimum” at w0 in the problem (3.1), (3.2). First of all, we show that {δw 1 } is a Pontryagin sequence for system (3.2) at w 0 . For this purpose, we represent δu1 in the form δu1 = U (t, x 0 + δx, u0 + δu) − U (t, x 0 , u0 + δu) + U (t, x 0 , u0 + δu) − U (t, x 0 , u0 ). (3.10) This representation is correct, since the conditions δx C →0 and (t, x 0 + δx, u0 + δu) ∈ C ⊂ Q1 (3.11) imply (t, x 0 , u0 + δu) ∈ Q1 for all sufficiently large serial numbers (here, we use the compactness of C and the openness of Q1 ). Representation (3.10) implies δu1 1 ≤ U (t, x 0 + δx, u0 + δu) − U (t, x 0 , u0 + δu) + U (t, x 0 , u0 + δu) − U (t, x 0 , u0 ) 1 . ∞ (tf − t0 ) This, condition (3.11), and also the condition δu 1 → 0 imply δu1 1 → 0. Further, denote by C1 the image of the compact set C under the mapping (t, x, u) → (t, x, U (t, x, u)). Then C ⊂ G, and hence C1 ⊂ Q1 . Moreover, C1 is a compact set and (t, x 0 + δx, u0 + δu1 ) = (t, x 0 + δx, U (t, x 0 + δx, u0 + δu)) ∈ C1 . This implies g(t, x 0 + δx, u0 + δu1 ) = 0. Finally, we note that the sequence {w 0 + δw1 } satisfies the differential equation of system (3.2): ẋ 0 + δ ẋ = f (t, x 0 + δx, U (t, x 0 + δx, u0 + δu)) = f (t, x 0 + δx, u0 + δu1 ) and the local equality constraint g(t, x 0 + δx, u0 + δu1 ) = g(t, x 0 + δx, U (t, x 0 + δx, u0 + δu)) = 0. Therefore, we have shown that the mapping (δx, δu) → δx, U (t, x 0 + δx, u0 + δu) − u0 transforms the Pontryagin sequence {δw} of the system (3.1), (3.5) at the point w 0 into the Pontryagin sequence {δw1 } of the system (3.1), (3.2) at the same point. Since the sequence {δw} “violates” the Pontryagin minimum in the problem (3.1), (3.5) at the point w 0 , conditions (3.8) hold. But these conditions can be also referred to as the sequence {δw1 }, since the members δx of these two sequence coincide. Therefore, {δw 1 } “violates” the Pontryagin minimum at the point w0 in the problem (3.1), (3.2). Therefore, the absence 130 Chapter 3. Quadratic Conditions for Optimal Control Problems of the Pontryagin minimum at the point w 0 in the problem (3.1), (3.5) implies the same case in the problem (3.1), (3.2). This implies what was required. Let w 0 = (x 0 , u0 ) be a fixed point of the Pontryagin minimum in the problem (3.1), (3.2). Then w0 is a Pontryagin minimum point in the problem (3.1), (3.5). We now write the quadratic necessary conditions for the Pontryagin minimum at the point w0 in the problem (3.1), (3.5) (which contains no local constraints) so that the projection U can be excluded from these conditions. Then we will obtain the quadratic necessary conditions for the Pontryagin minimum in the problem (3.1), (3.2). 3.1.2 Set M0 For the problem (3.1), (3.2), we set l = α0 J + αK, H = ψf , and H̄ = H + νg, where ν ∈ (Rd(g) )∗ . Therefore, H̄ = H̄ (t, x, u, ψ, ν). We also set H (t, x, ψ) = min {u|(t,x,u)∈G} ψf (t, x, u). For problem (3.1), (3.2) and the point w 0 , we introduce the set M0 consisting of tuples λ = (α0 , α, β, ψ(t), ν(t)) such that α0 ≥ 0, α ≥ 0, αF (p 0 ) = 0, ψ(t0 ) = −lx0 , ψ(tf ) = lxf , α0 + |α| + |β| = 1, − ψ̇ = H̄x , H̄u = 0, H (t, x 0 , u0 , ψ) = H (t, x 0 , ψ). (3.12) (3.13) (3.14) (3.15) Here, ψ(t) is an absolutely continuous function and ν(t) is a bounded measurable function. All the derivatives are taken for p = p 0 and w = w 0 (t). The results of [79, Section 17, Part 1] imply that the set M0 of problem (3.1), (3.5) at the point w0 can be represented in this form. More precisely, the linear projection (α0 , α, β, ψ, ν) → (α0 , α, β, ψ) yields a one-to-one correspondence between the elements of the set M0 of the problem (3.1), (3.2) at the point w 0 and the elements of the set M0 of the problem (3.1), (3.5) at the same point. To differentiate these two sets from one another, we denote the latter set by M0U . We will equip all objects referring to the problem (3.1), (3.5) with the superscript U . In what follows, we will assume that all assumptions of Section 2.1 hold for the point w 0 , i.e., u0 (t) is a piecewise continuous function whose set of discontinuity points is = {t1 , . . . , ts } ⊂ (t0 , tf ), and each point of the set is an L-point. The condition −fu∗ (t, x 0 (t), u0 (t))ψ ∗ (t) = gu∗ (t, x 0 (t), u0 (t))ν ∗ (t), (3.16) which is equivalent to the condition H̄u = 0, and also the full-rank condition (3.3) imply that ν(t) has the same properties as u0 (t): the function ν(t) is piecewise continuous and each of its point of discontinuity is an L-point which belongs to . To verify this, it suffices to premultiply the above relation by the matrix gu (t, x 0 (t), u0 (t)), −gu (t, x 0 (t), u0 (t))fu∗ (t, x 0 (t), u0 (t))ψ ∗ (t) = gu (t, x 0 (t), u0 (t))gu∗ (t, x 0 , (t), u0 (t))ν ∗ (t), and use the properties of the functions g, f , x 0 , and u0 ; in particular, the property | det gu (t, x 0 , u0 )gu∗ (t, x 0 , u0 )| ≥ const > 0, 3.1. Quadratic Necessary Conditions on a Fixed Time Interval 131 which is implied by the full-rank condition (3.3). Therefore, the basic properties of the function ν(t) are proved exactly in the same way as in [79, Section 17, Part 1] for the bounded measurable control u0 (t). 3.1.3 Critical Cone We now consider the conditions defining critical cone at the point w 0 in the problem (3.1), (3.5). The variational equation has the form x̄˙ = (fx + fu Ux )x̄ + fu Uu ū. (3.17) All the derivatives are taken for w = w0 (t). Setting ũ = Ux x̄ + Uu ū, we obtain x̄˙ = fx x̄ + fu ũ. (3.18) This is the usual variational equation, but for the pair w̃ = (x̄, ũ). Let us show that the pair (x̄, ũ) also satisfies the condition gx x̄ + gu ũ = 0. (3.19) By the first condition in (3.4), we have g(t, x, U (t, x, u)) = 0 ∀ (t, x, u) ∈ Q1 . (3.20) Differentiating this relation in x, u, and t, as in [79, Section 17, Part 1], we obtain gx + gu Ux = 0, gu Uu = 0, gt + gu Ut = 0. (3.21) (3.22) (3.23) These relations hold on Q1 , but it suffices to consider them only on the trajectory (t, x 0 (t), u0 (t)), and, moreover, we now need only the first two conditions. By (3.21) and (3.22), we have gx x̄ + gu ũ = gx x̄ + gu (Ux x̄ + Uu ū) = (gx + gu Ux )x̄ + gu Uu ū = 0. Therefore, we have proved the following proposition. Proposition 3.2. Let a pair of functions (x̄, ū) satisfy the variational equation (3.17) of system (3.5). We set ũ = Ux x̄ + Uu ū. Then conditions (3.18) and (3.19) hold for (x̄, ũ). In this proposition, x̄ ∈ P W 1,2 (, Rd(x) ), ū ∈ L2 (, Rd(u) ), and ũ ∈ L2 (, Rd(u) ), where = [t0 , tf ]. Also, we are interested in the possibility of the converse passage from conditions (3.18) and (3.19) to condition (3.17). For this purpose, we prove the following proposition. Proposition 3.3. Let a pair of functions (x̄, ũ) be such that gx x̄ + gu ũ = 0. Then setting ū = ũ − Ux x̄, we obtain Uu ū = ū, and hence ũ = Ux x̄ + Uu ū. Here, as above, x̄ ∈ P W 1,2 (, Rd(x) ), ū ∈ L2 (, Rd(u) ), and ũ ∈ L2 (, Rd(u) ); all the derivatives are taken for x = x 0 (t) and u = u0 (t). 132 Chapter 3. Quadratic Conditions for Optimal Control Problems Proof. First of all, we note that properties (3.4) of the function U (t, x, u) imply the following assertion: at each point (t, x, u) ∈ G, the finite-dimensional linear operator ū ∈ Rd(u) → Uu (t, x, u)ū ∈ Rd(u) (3.24) is the linear projection of the space Rd(u) on the subspace Lg (t, x, u) = {ū ∈ Rd(u) | gu (t, x, u)ū = 0}. (3.25) Indeed, let (t, x, u) ∈ G, and then condition (3.22) implies that the image of operator (3.24) is contained in subspace (3.25). The condition Uu (t, x, u)ū = ū ∀ ū ∈ Lg (t, x, u) (3.26) is easily proved by using the Lyusternik theorem [28]. Indeed, let ū ∈ Lg (t, x, u), i.e., gu (t, x, u)ū = 0. Let ε → +0. Then g(t, x, u + εū) = g(t, x, u) + gu (t, x, u)εū + rg (ε) = rg (ε), where |rg (ε)| = o(ε). By the Lyusternik theorem [28], there exists a “sequence” {ug (ε)} such that ug (ε) ∈ Rd(u) , |ug (ε)| = o(ε), and, moreover, g(t, x, u + ε ū + ug (ε)) = 0, i.e., (t, x, u + εū + ug (ε)) ∈ G. Hence U (t, x, u + εū + ug (ε)) = u + ε ū + ug (ε). But U (t, x, u + ε ū + ug (ε)) = U (t, x, u) + Uu (t, x, u)(ε ū + ug (ε)) + rU (ε), where |rU (ε)| = o(ε). We obtain from the latter two conditions and the condition U (t, x, u) = u that ε ū + ug (ε) = Uu (t, x, u)(ε ū + ug (ε)) + rU (ε). Dividing this relation by ε and passing to the limit as ε → +0, we obtain ū = Uu (t, x, u)ū. Condition (3.26) is proved. Condition (3.26) holds at each point (t, x, u) ∈ G, but we use it only at the trajectory (t, x 0 (t), u0 (t)) | t ∈ [t0 , tf ]. If a pair of functions x̄(t), ũ(t) satisfies the condition gx x̄ + gu ũ = 0, then by (3.21), we have −gu Ux x̄ + gu ũ = 0, i.e., −Ux x̄ + ũ ∈ Lg (t, x 0 , u0 ). Then the condition Uu ū = ū also holds for ū = −Ux x̄ + ũ, and hence ũ = Ux x̄ + Uu ū. The proposition is proved. Proposition 3.3 implies the following assertion. Proposition 3.4. Let a pair of functions (x̄, ũ) be such that conditions (3.18) and (3.19) hold : x̄˙ = fx x̄ +fu ũ and gx x̄ +gu ũ = 0. We set ū = −Ux x̄ + ũ. Then Uu ū = ū, and the variational equation (3.17) of system (3.5) holds for the pair of functions (x̄, ū) at the point w 0 . Proof. Indeed, x̄˙ = fx x̄ + fu ũ = fx x̄ + fu (Ux x̄ + ū) = fx x̄ + fu (Ux x̄ + Uu ū) = (fx + fu Ux )x̄ + fu Uu ū as required. The proposition is proved. We now give the following definition. Definition 3.5. The critical cone K of problem (3.1), (3.2) at the point w 0 is the set of triples z̄ = (ξ̄ , x̄, ū) satisfying the following conditions: 3.1. Quadratic Necessary Conditions on a Fixed Time Interval ξ̄ ∈ Rs , x̄ ∈ P W 1,2 (, Rd(x) ), ū ∈ L2 (, Rd(u) ), Jp p̄ ≤ 0, Fip p̄ ≤ 0 ∀ i ∈ I , Kp p̄ = 0, x̄˙ = fx x̄ + fu ū, [x̄]k = [f ]k ξ̄k ∀ tk ∈ , gx x̄ + gu ū = 0. 133 (3.27) (3.28) (3.29) (3.30) (3.31) Let us compare this definition with the definition of the critical cone K U of the problem (3.1), (3.5) at the point w0 . According to Section 2.1, the latter is defined by the same conditions (3.27) and (3.28), the variational equation (3.17), and the jump condition (3.30). Thus, all the conditions referring to the components ξ̄ and x̄ in the definitions of two critical cones coincide. This and Proposition 3.2 imply the following assertion. Lemma 3.6. Let z̄ = (ξ̄ , x̄, ū) be an arbitrary element of the critical cone K U of the problem (3.1), (3.5) at the point w0 . We set ũ = Ux x̄ + Uu ū. Then z̃ = (ξ̄ , x̄, ũ) is an element of the critical cone K of the problem (3.1), (3.2) at the point w0 . Respectively, Proposition 3.4 implies the following lemma. Lemma 3.7. Let z̃ = (ξ̄ , x̄, ũ) be an arbitrary element of the critical cone K of the problem (3.1), (3.2) at the point w0 . We set ū = ũ − Ux x̄. Then z̄ = (ξ̄ , x̄, ū) is an element of the critical cone K U of the problem (3.1), (3.5) at the point w0 , and, moreover, Uu ū = ū, which implies ũ = Ux x̄ + Uu ū. This is the connection between the critical cones at the point w 0 in the problems (3.1), (3.2) and (3.1), (3.5). We will need Lemma 3.6 later in deducing quadratic sufficient conditions in the problem with a local equality; now, in deducing necessary conditions, we use Lemma 3.7. Preparatorily, we find the connection between the corresponding quadratic forms. 3.1.4 Quadratic Form We write the quadratic form U (z̄) for the point w 0 in the problem (3.1), (3.5) in accordance with its definition in Section 2.1. Let λU = (α0 , α, β, ψ) be an element of the set M0U of the problem (3.1), (3.5) at the point w 0 , and let λ = (α0 , α, β, ψ, ν) be the corresponding element of the set M0 of the problem (3.1), (3.2) at the same point. As above, we set H (t, x, u, ψ) = ψf (t, x, u); H̄ (t, x, u, ψ) = ψf (t, x, u) + νg(t, x, u). Also, we introduce the notation f U (t, x, u) = f (t, x, U (t, x, u)), H U (t, x, u, ψ) = ψf U (t, x, u) = H (t, x, U (t, x, u), ψ). (3.32) (3.33) We omit the superscripts λ and λU in the notation. For each t ∈ [t0 , tf ], let us calculate the quadratic form U U U U Hww w̄, w̄ = Hxx x̄, x̄ + 2Hxu ū, x̄ + Huu ū, ū , (3.34) 134 Chapter 3. Quadratic Conditions for Optimal Control Problems where x̄ ∈ Rd(x) , ū ∈ Rd(u) , and all the second derivatives are taken at the point (t, x, u) = (t, x 0 (t), u0 (t)). U x̄, x̄ . It follows from (3.33) that Let us calculate Hxx HxU = Hx + Hu Ux . (3.35) Differentiating this equation in x and twice multiplying it by x̄, we obtain U x̄, x̄ Hxx = Hxx x̄, x̄ + 2Hxu (Ux x̄), x̄ + Huu (Ux x̄), (Ux x̄) + (Hu Uxx )x̄, x̄ , (3.36) d(u) where (Hu Uxx )x̄, x̄ = ( i=1 Hui Uixx )x̄, x̄ by definition. Further, let us calculate U ū, x̄ . Differentiating (3.35) in u and multiplying it by ū and x̄, we obtain Hxu U ū, x̄ = Hxu (Uu ū), x̄ + Huu (Uu ū), (Ux x̄) + (Hu Uxu )ū, x̄ , Hxu (3.37) d(u) U ū, ū . It follows where (Hu Uxu )ū, x̄ = ( i=1 Hui Uixu )ū, x̄ . Finally, let us calculate Huu U from (3.33) that Hu = Hu Uu . Differentiating this equation in u and twice multiplying it by ū, we obtain U Huu ū, ū = Huu (Uu ū), (Uu ū) + (Hu Uuu )ū, ū , (3.38) d(u) where (Hu Uuu )ū, ū = ( i=1 Hui Uiuu )ū, ū . Formulas (3.34), (3.36)–(3.38) imply U w̄, w̄ Hww = Hxx x̄, x̄ + 2Hxu (Ux x̄), x̄ + Huu (Ux x̄), (Ux x̄) + (Hu Uxx )x̄, x̄ + 2Hxu (Uu ū), x̄ + 2Huu (Uu ū), (Ux x̄) + 2(Hu Uxu )ū, x̄ + Huu (Uu ū), (Uu ū) + (Hu Uuu )ū, ū = Hxx x̄, x̄ + 2Hxu ũ, x̄ + Huu ũ, ũ + (Hu Uxx x̄, x̄ + 2(Hu Uxu )ū, x̄ + (Hu Uuu )ū, ū , (3.39) where ũ = Ux x̄ + Uu ū. Further, differentiating in x the relations gix + giu Ux = 0, i = 1, . . . , d(g), (3.40) which hold on Q1 , and twice multiplying the result by x̄, we obtain gixx x̄, x̄ + 2gixu (Ux x̄), x̄ + giuu (Ux x̄), (Ux x̄) + giu Uxx x̄, x̄ = 0, i = 1, . . . , d(g). Multiplying each of these relations by the ith component νi of the vectorvalued function ν(t), summing with respect to i, and using the relation Hu + νgu = 0, we obtain (3.41) νgxx x̄, x̄ + 2νgxu (Ux x̄), x̄ + νguu (Ux x̄), (Ux x̄) − Hu Uxx x̄, x̄ = 0, where νgxx = νi gixx , νgxu = νi gixu , and νguu = νi giuu . Differentiating the same relations (3.40) in u and multiplying by ū and x̄, we obtain gixu (Uu ū), x̄ + giuu (Uu ū), (Ux x̄) + (giu Uxu )ū, x̄ = 0. 3.1. Quadratic Necessary Conditions on a Fixed Time Interval 135 Multiplying each of these relations by 2νi , summing with respect to i, and using the property that Hu + νgu = 0, we obtain (3.42) 2νgxu (Uu ū), x̄ + 2νguu (Uu ū), (Ux x̄) − 2(Hu Uxu )ū, x̄ = 0, where νgxu = νi gixu and νguu = νi giuu . Finally, differentiating in u the relations giu Uu = 0, i = 1, . . . , d(g), which hold on Q1 , and twice multiplying the result by ū, we obtain giuu (Uu ū), (Uu ū) + (giu Uuu )ū, ū = 0. Multiplying each of these equations by νi and using the property that Hu + νgu = 0, we obtain νguu (Uu ū), (Uu ū) − Hu Uuu ū, ū = 0. (3.43) U ū, ū = Huu (Uu ū), (Uu ū) + νguu (Uu ū), (Uu ū) . Huu (3.44) This and (3.38) imply Using the notation H̄ = H + νg, we present this relation in the form U ū, ū = H̄uu (Uu ū), (Uu ū) . Huu (3.45) We will use this relation later in Section 3.2. Summing relations (3.41)–(3.43), we obtain νgxx x̄, x̄ + 2νgxu ũ, x̄ + 2νguu ũ, ũ − Hu Uxx x̄, x̄ − 2Hu Uxu ū, x̄ − Hu Uuu ū, ū = 0, (3.46) U w̄, w̄ = H where ũ = Ux x̄ + Uu ū. It follows from (3.39) and (3.46) that Hww ww w̃, w̃ + νgww w̃, w̃ or U w̄, w̄ = H̄ww w̃, w̃ , (3.47) Hww where w̃ = (x̄, ũ) = (x̄, Ux x̄ + Uu ū). We now consider the terms referring to the points of discontinuity of the control u0 . We set (3.48) D k (H̄ ) = −H̄xk+ H̄ψk− + H̄xk− H̄ψk+ − [H̄t ]k , k = 1, . . . , s, where H̄xk+ = H̄x (tk , x 0 (tk ), u0k+ , ψ(tk ), ν k+ ), H̄xk− = H̄x (tk , x 0 (tk ), u0k− , ψ(tk ), ν k− ), H̄ψk+ = f (tk , x 0 (tk ), u0k+ ) = Hψk+ , H̄ψk− = f (tk , x 0 (tk ), u0k− ) = Hψk− , [H̄t ]k = H̄tk+ − H̄tk− = ψ(tk )[ft ]k + [νgt ]k = ψ(tk )(ft (tk , x 0 (tk ), u0k+ ) − ft (tk , x 0 (tk ), u0k− )) + ν k+ gt (tk , x 0 (tk ), u0k+ ) − ν k− gt (tk , x 0 (tk ), u0k− ), k+ ν = ν(tk +), ν k− = ν(tk −). Therefore, the definition of D k (H̄ ) is analogous to that of D k (H ). We can define D k (H̄ ) using another method, namely, as the derivative of the “jump of H̄ ” at the point tk . Introduce the function (k H̄ )(t) = (k H )(t) + (k (νg))(t) = ψ(t) f (t, x 0 (t), u0k+ ) − f (t, x 0 (t), u0k− ) + ν k+ g(t, x 0 (t), u0k+ ) − ν k− g(t, x 0 (t), u0k− ) . (3.49) 136 Chapter 3. Quadratic Conditions for Optimal Control Problems Similarly to what was done for (k H )(t) in Section 2.3 (see Lemma 2.12), we can show that the function (k H̄ )(t) is continuously differentiable at the point tk ∈ , and its derivative at this point coincides with −D k (H̄ ). Therefore, we can obtain the value of D k (H̄ ) calculating the left or right limit of the derivatives of the function (k H̄ )(t) defined by formula (3.49): d D k (H̄ ) = − (k H̄ )(tk ). dt We now show that D k (H̄ ) = D k (H U ), k = 1, . . . , s. (3.50) Indeed, by definition, −D k (H U ) = HxU k+ HψU k− − HxU k− HψU k+ + [HtU ]k . (3.51) HxU = Hx + Hu Ux = Hx − νgu Ux = Hx + νgx = H̄x . (3.52) Furthermore, Here, we have used the formulas Hu + νgu = 0, gx + gu Ux = 0, and H̄ = H + νg. Also, it is obvious that for (t, w) = (t, w0 (t)), HψU = f U = f = Hψ = H̄ψ . (3.53) HtU = Ht + Hu Ut = Ht − νgu Ut = Ht + νgt = H̄t , (3.54) Finally, since gt + gu Ut = 0. The formulas (3.48), (3.51)–(3.54) imply relation (3.50). Further, note that relations (3.52) imply [HxU ]k = [H̄x ]k , k = 1, . . . , s, (3.55) where [H̄x ]k = H̄x (tk , x 0 (tk ), u0k+ , ψ(tk ), ν k+ ) − H̄x (tk , x 0 (tk ), u0k− , ψ(tk ), ν k− ) (3.56) is the jump of the function H̄x (t, x 0 (t), u0 (t), ψ(t), ν(t)) at the point tk ∈ . For the problem (3.1), (3.2) and the point w0 , we define the following quadratic form in z̄ = (ξ̄ , x̄, ū) for each λ = (α0 , α, β, ψ, ν) ∈ M0 : 2λ (z̄) = s k=1 k λ D k (H̄ λ )ξ̄k2 + 2[H̄xλ ]k x̄av p̄, p̄ + ξ̄k + lpp tf t0 λ H̄ww w̄, w̄ dt. (3.57) Therefore, the quadratic form in the problem with local equality-type constraints is defined in the same way as in the problem without local constraints; the only difference is that instead of the function H = ψf in the definition of the new quadratic form, we must use the function H̄ = H + νg. 3.1. Quadratic Necessary Conditions on a Fixed Time Interval 137 According to Section 2.1, the quadratic form takes the following form for the problem (3.1), (3.5) and the point w 0 : U 2U λ (z̄) % & U U k ξ̄k D k H U λ ξ̄k2 + 2 HxU λ x̄av k=1 tf U λU p̄, p̄ + H U λ w̄, w̄ dt. + lpp s = (3.58) t0 Here, as above, we have used the superscript U in the notation U of the quadratic form in order to stress that this quadratic form corresponds to the problem (3.1), (3.5) being considered for a given projection U (t, x, u). We have denoted by λU the tuple (α0 , α, β, ψ), which uniquely defines the tuple λ = (α0 , α, β, ψ, ν) by the condition ψfu + νgu = 0. In what follows, these tuples correspond to one another and belong to the sets M0U and M0 of the problems (3.1), (3.5) and (3.1), (3.2), respectively. Formulas (3.47), (3.50), (3.55), (3.57), and (3.58) imply the following assertion. Lemma 3.8. Let z̄ = (ξ̄ , x̄, ū) be an arbitrary element of the space Z2 (), and let z̃ = (ξ̄ , x̄, ũ) = (ξ̄ , w̃), where ũ = Ux x̄ + Uu ū. Let λU be an arbitrary element of M0U , and let λ U be the corresponding element of M0 . Then λ (z̃) = U λ (z̄). 3.1.5 Necessary Quadratic Conditions The following theorem holds. Theorem 3.9. If w 0 is a Pontryagin minimum point in the problem (3.1), (3.2), then the following Condition A holds: the set M0 is nonempty and max λ (z̄) ≥ 0 ∀ z̄ ∈ K, λ∈M0 (3.59) where K is the critical cone at the point w 0 defined by conditions (3.27)–(3.31), λ (z̄) is the quadratic form at the same point defined by (3.57), and M0 is the set of tuples of Lagrange multipliers satisfying the minimum principle defined by (3.12)–(3.15). Proof. Let w0 be a Pontryagin minimum point in the problem (3.1), (3.2). Then according to Proposition 3.1, w0 is a Pontryagin minimum point in the problem (3.1), (3.5). Hence, by Theorem 2.4, the following necessary Condition AU holds at the point w 0 in the problem (3.1), (3.5): the set M0U is nonempty and U max U λ (z̄) ≥ 0 M0U ∀ z̄ ∈ K U . (3.60) Let us show that the necessary Condition A holds at the point w 0 in the problem (3.1), (3.2). Let z̃ = (ξ̄ , x̄, ũ) be an arbitrary element of the critical cone K at the point w0 in the problem (3.1), (3.2). According to Lemma 3.7, there exists a function ū(t) such that ũ = Ux x̄ + Uu ū, and, moreover, z̄ = (ξ̄ , x̄, ū) is an element of the critical cone K U at the point w0 in the problem (3.1), (3.5) (we can set ū = ũ − Ux x̄). Since the necessary 138 Chapter 3. Quadratic Conditions for Optimal Control Problems U Condition AU holds, there exists an element λU ∈ M0U such that U λ (z̄) ≥ 0. According to Section 3.1.2, an element λ ∈ M0 corresponds to the element λU ∈ M0U . Moreover, by U Lemma 3.8, λ (z̃) = U λ (z̄). Therefore, λ (z̃) ≥ 0. Since z̃ is an arbitrary element of K, Condition A holds. The theorem is proved. Therefore, we have obtained the final form of the quadratic necessary condition for the Pontryagin minimum in the problem with local equality-type constraints, Condition A, in which there is no projection U . This condition is a natural generalization of the quadratic necessary Condition A in the problem without local constraints. 3.2 Quadratic Sufficient Conditions in the Problem with Mixed Control-State Equality Constraints on a Fixed Time Interval 3.2.1 Auxiliary Problem V Since there is the projection U (t, x, u), the problem (3.1), (3.5) has the property that the strict minimum is not attained at any point. For this reason, the problem (3.1), (3.5) cannot be directly used for obtaining quadratic sufficient conditions that guarantee the strict minimum. To overcome this difficulty, we consider a new auxiliary problem adding the additional constraint tf (u − U (t, x, u))2 dt ≤ 0, t0 where (u − U )2 = u − U , u − U . Representing this constraint as an endpoint constraint by introducing a new state variable y, we arrive at the following problem on a fixed interval [t0 , tf ]: J (x0 , xf ) → min, (3.61) F (x0 , xf ) ≤ 0, K(x0 , xf ) = 0, (x0 , xf ) ∈ P , y0 = 0, yf ≤ 0, 1 ẏ = (u − U (t, x, u))2 , 2 ẋ = f (t, x, U (t, x, u)), (t, x, u) ∈ Q1 , (3.62) (3.63) (3.64) where x0 = x(t0 ), xf = x(tf ), y0 = y(t0 ), and yf = y(tf ). Problem (3.61)–(3.64) is called the auxiliary problem V . By the superscript V we denote all objects referring to this problem. Therefore, the auxiliary problem V differs from the auxiliary problem (3.1), (3.5) or problem U by the existence of the additional constraints (3.62) and (3.63). If (y, x, u) is an admissible triple in problem (3.61)–(3.64), then y = 0, and the pair w = (x, u) satisfies the constraints of problem (3.1), (3.2). Indeed, (3.62) and (3.63) imply y(t) = 0, U (t, x(t), u(t)) = u(t), (3.65) 3.2. Quadratic Sufficient Conditions on a Fixed Time Interval and then 139 ẋ(t) − f (t, x(t), u(t)) = ẋ(t) − f (t, x(t), U (t, x(t), u(t))) = 0, g(t, x(t), u(t)) = g(t, x(t), U (t, x(t), u(t))) = 0, since (t, x(t), u(t)) ∈ Q1 . The converse is also true: if w = (x, u) is an admissible pair in problem (3.1), (3.2) and y = 0, then the triple (y, x, u) is admissible in Problem V , since conditions (3.65) hold for it. 3.2.2 Bounded Strong γ -Sufficiency Fix an admissible point w 0 = (x 0 , u0 ) ∈ W in problem (3.1), (3.2) satisfying the assumption of Section 2.1. Let y 0 = 0. Then (y 0 , w 0 ) = (0, w 0 ) is an admissible point in Problem V . Let (t, u) be the admissible function defined in Section 2.3 (see Definition 2.17). The higher order at the point (0, w 0 ) in Problem V is defined by the relation tf (t, u0 + δu) dt = δy 2C + γ (δw). (3.66) γ V (δy, δw) = δy 2C + δx 2C + t0 The violation function is defined by the relation σ V (δy, δw) = (δJ )+ + |F (p 0 + δp)+ | + |δK| + δ ẋ − δf U 1 1 + δ ẏ − (u0 + δu − U (t, x 0 + δx, u0 + δu))2 1 + (δyf )+ + |δy0 |, (3.67) 2 where δf U = f U (t, w 0 + δw) − f U (t, w0 ) = f (t, x 0 + δx, U (t, x 0 + δx, u0 + δu)) − f (t, x 0 , u0 ). (3.68) We will use the concept of a bounded strong minimum and also that of a bounded strong γ V -sufficiency at the point (0, w 0 ) in the auxiliary problem V . Let us introduce analogous concepts for problem (3.1), (3.2) at the point w 0 . Definition 3.10. We say that a point w0 = (x 0 , u0 ) is a point of strict bounded strong minimum in problem (3.1), (3.2) if there is no sequence {δw} in the space W that does not contain zero members, there is no compact set C such that |δx(t0 )| → 0, δx C → 0, and the following conditions hold for all members of the sequence {δw}: (t, w 0 + δw) ∈ C, δJ = J (p 0 + δp) − J (p0 ) ≤ 0, F (p0 + δp) ≤ 0, K(p 0 + δp) = 0, ẋ 0 + δ ẋ = f (t, w0 + δw), g(t, w0 + δw) = 0. As in Section 2.7 (see Definition 2.92), we denote by δx the tuple of essential components of the variation δx. (The definition of unessential components in problem (3.1), (3.2) is the same as in problem (2.1)–(2.4), but now neither functions f nor g depends on these components.) Definition 3.11. We say that w 0 = (x 0 , u0 ) is a point of bounded strong γ -sufficiency in problem (3.1), (3.2) if there is no sequence {δw} in the space W without zero members and 140 Chapter 3. Quadratic Conditions for Optimal Control Problems there is no compact set C ⊂ Q such that |δx(t0 )| → 0, δx C → 0, σ (δw) = o(γ (δw)), (3.69) and the following conditions hold for all members of the sequence {δw}: g(t, w0 + δw) = 0 Here, and (t, w 0 + δw) ∈ C. σ (δw) = (δJ )+ + |F (p 0 + δp)+ | + |δK| + δ ẋ − δf (3.70) 1. (3.71) Therefore, the violation function σ (δw) in problem (3.1), (3.2) contains no term related to the local constraint g(t, x, u) = 0; however, in the definition of the bounded strong γ -sufficiency of this problem, it is required that the sequence {w0 + δw} satisfies this constraint. The local constraint does not have the same rights as the other constraints. Obviously, the following assertion holds. Proposition 3.12. The bounded strong γ -sufficiency at the point w0 in problem (3.1), (3.2) implies the strict bounded strong minimum. Our goal is to obtain a sufficient condition for the bounded strong γ -sufficiency in problem (3.1), (3.2) using the sufficient condition for the bounded strong γ V -sufficiency in the auxiliary problem V without local constraints. For this purpose, we prove the following assertion. Proposition 3.13. Let a point (0, w 0 ) be a point of bounded strong γ V -sufficiency in problem V. Then w0 is a point of bounded strong γ -sufficiency in problem (3.1), (3.2). Proof. Suppose that w0 is not a point of the bounded strong γ -sufficiency in problem (3.1), (3.2). Then there exist a sequence {δw} containing nonzero members and a compact set C ⊂ Q such that conditions (3.69) and (3.70) hold. We show that in this case, (0, w0 ) is not a point of bounded strong γ V -sufficiency in Problem V . Consider the sequence {(0, δw)} with δy = 0. Condition (3.70) implies U (t, x 0 + δx, u0 + δu) = u0 + δu, (3.72) δf = f (t, x 0 + δx, u0 + δu) − f (t, x 0 , u0 ) = f (t, x 0 + δx, U (t, x 0 + δx, u0 + δu)) − f (t, x 0 , u0 ) = δf U . (3.73) and then It follows from (3.66)–(3.73) that σ V (0, δw) = σ (δw) = o(γ (δw)) = o(γ V (0, δw)). Therefore, (0, w 0 ) is not a point of the bounded strong γ V -sufficiency in Problem V . The proposition is proved. Next, we formulate the main result of this section: the quadratic sufficient conditions for the bounded strong γ -sufficiency at the point w 0 in problem (3.1), (3.2). Then we 3.2. Quadratic Sufficient Conditions on a Fixed Time Interval 141 show that these conditions guarantee the bounded strong γ V -sufficiency in Problem V , and hence, by Proposition 3.13, this implies the bounded strong γ -sufficiency at the point w 0 in problem (3.1), (3.2). This is our program. We now formulate the main result. 3.2.3 Quadratic Sufficient Conditions For an admissible point w0 = (x 0 , u0 ) in problem (3.1), (3.2), we give the following definition. Definition 3.14. An element λ = (α0 , α, β, ψ, ν) ∈ M0 is said to be strictly Legendre if the following conditions hold: (1) D k (H̄ λ ) > 0 for all tk ∈ ; (2) for any t ∈ [t0 , tf ] \ , the form H̄uu (t, w0 (t), ψ(t), ν(t))ū, ū (3.74) quadratic in ū is positive definite on the subspace of vectors ū ∈ Rd(u) such that gu (t, w0 (t))ū = 0; (3.75) (3) the following condition C k− holds for each point tk ∈ : the form H̄uu (tk , x 0 (tk ), u0k− , ψ(tk ), ν k− )ū, ū (3.76) quadratic in ū is positive definite on the subspace of vectors ū ∈ Rd(u) such that gu (tk , x 0 (tk ), u0k− )ū = 0 (3.77) (4) the following condition C k+ holds for each point tk ∈ : the quadratic form H̄uu (tk , x 0 (tk ), u0k+ , ψ(tk ), ν k+ )ū, ū (3.78) is positive definite on the subset of vectors ū ∈ Rd(u) such that gu (tk , x 0 (tk ), u0k+ )ū = 0. (3.79) Further, denote by M0+ the set of λ ∈ M0 such that the following conditions hold: H (t, x 0 (t), u, ψ(t)) > H (t, x 0 (t), u0 (t), ψ(t)), if t ∈ [t0 , tf ]\, u ∈ U(t, x 0 (t)), u = u0 (t), (3.80) (3.81) where U(t, x) = {u ∈ Rd(u) | (t, x, u) ∈ Q, g(t, x, u) = 0}; H (tk , x 0 (tk ), u, ψ(tk )) > H (tk , x 0 (tk ), u0k− , ψ(tk )) = H (tk , x 0 (tk ), u0k+ , ψ(tk )) (3.82) / {u0k− , u0k+ }. (3.83) if tk ∈ , u ∈ U(tk , x 0 (tk )), u ∈ Denote by Leg+ (M0+ ) the set of all strictly Legendrian elements λ ∈ M0+ . 142 Chapter 3. Quadratic Conditions for Optimal Control Problems Definition 3.15. We say that Condition B holds at the point w 0 in problem (3.1), (3.2) if the set Leg+ (M0+ ) is nonempty and there exist a compact set M ⊂ Leg+ (M0+ ) and a constant C > 0 such that max λ ≥ C γ̄ (z̄) ∀ z̄ ∈ K, (3.84) λ∈M the critical cone K for problem (3.1), (3.2), and the point where the quadratic form w0 were defined by relations (3.57) and (3.27)–(3.31), respectively, and tf ū(t), ū(t) dt (3.85) γ̄ (z̄) = ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) + λ (z̄), t0 (as in (2.15)). We have the following theorem. Theorem 3.16. If Condition B holds for the point w 0 in the problem (3.1), (3.2), then we have the bounded strong γ -sufficiency at this point. We now prove this theorem. As was said, by Proposition 3.13, it suffices to show that Condition B guarantees the bounded strong γ V -sufficiency at the point (0, w0 ) in the problem without local constraints. For this purpose, we write the quadratic sufficient condition of Section 2.1 for the auxiliary Problem V at the point (0, w0 ). 3.2.4 Proofs of Sufficient Quadratic Conditions We first write the set M0V of normalized Lagrange multipliers satisfying the maximum principle for the point (0, w9 ) in Problem V . The Pontryagin function in Problem V has the form 1 H V (t, y, x, u, ψy , ψx ) = ψx f (t, x, U (t, x, u)) + ψy (u − U (t, x, u))2 2 1 U = H (t, x, u, ψx ) + ψy (u − U (t, x, u))2 , 2 (3.86) and the endpoint Lagrange function is defined by the relation l V (y0 , x0 , yf , xf , α0 , αy , α, βy , β) = α0 J (x0 , xf ) + αF (x0 , xf ) + βK(x0 , xf ) + αy yf + βy y0 = l(x0 , xf , α0 , α, β) + αy yf + βy y0 . (3.87) The set M0V consists of tuples (α0 , αy , α, βy , β, ψy , ψx ) such that α0 ≥ 0, α ≥ 0, αy ≥ 0, (3.88) αF (p ) = 0, αy yf0 = 0, α0 + |α| + αy + |βy | + |β| = 1, 0 ψ̇y = 0, ψy (t0 ) = −βy , ψ̇x = −HxV , ψx (t0 ) = −lx0 , ψx (tf ) = lxf , (3.89) (3.90) (3.91) ψy (tf ) = αy , (3.92) (3.93) (3.94) 3.2. Quadratic Sufficient Conditions on a Fixed Time Interval 143 H V (t, y 0 (t), x 0 (t), u, ψy (t), ψx (t)) ≥ H V (t, y 0 (t), x 0 (t), u0 (t), ψy (t), ψx (t)) (3.95) if t ∈ [t0 , tf ] \ and (t, x 0 (t), u) ∈ Q1 . Let us analyze these conditions. Since yf0 = y 0 (tf ) = 0, condition (3.90) holds automatically, and we can exclude it from consideration. Further, since HxV = HxU − ψy (u0 − U (t, w0 ))Ux (t, w0 ) = HxU , condition (3.93) is equivalent to the condition ψ̇x = −HxU . (3.96) ψy = const = −βy = αy . (3.97) It follows from (3.92) that Therefore, the normalization condition (3.91) is equivalent to the condition α0 + |α| + |β| + αy = 1. (3.98) We now turn to the minimum condition (3.95). It follows from (3.86) and (3.97) that it is equivalent to the condition 1 H U (t, x 0 (t), u, ψx (t)) + αy (u − U (t, x 0 (t), u))2 2 ≥ H U (t, x 0 (t), u0 (t), ψx (t)) (3.99) whenever t ∈ [t0 , tf ] \ and (t, x 0 (t), u) ∈ Q1 . Therefore, we can identify the set M0V with the set of tuples λV = (α0 , α, β, ψx , αy ) such that conditions (3.88), (3.89), (3.98), (3.96), (3.94), and (3.99) hold. Let there exist an element λ = (α0 , α, β, ψ, ν) of the set M0 of problem (3.1), (3.2) at the point w 0 . Then its projection λU = (α0 , α, β, ψ) is an element of the set M0U of problem (3.1), (3.5) (of problem U ) at the same point. Let 0 ≤ αy ≤ 1. We set λV = ((1 − αy )λU , αy ). (3.100) Let us show that λV ∈ M0V . Indeed, (1 − αy ) α0 + (1 − αy )|α| + (1 − αy )|β| + αy = (1 − αy )(α0 + |α| + |β|) + αy = 1; i.e., the normalization condition (3.98) holds for λV . Also conditions (3.88), (3.89), (3.96), and (3.94) hold. Let us verify the minimum condition (3.99). Since λU ∈ M0U, the conditions t ∈ [t0 , tf ] \ , (t, x 0 (t), u) ∈ Q1 (3.101) imply H U (t, x 0 (t), u, ψx (t)) ≥ H U (t, x 0 (t), u0 (t), ψx (t)). (3.102) 144 Chapter 3. Quadratic Conditions for Optimal Control Problems Moreover, the condition αy ≥ 0 implies 1 αy (u − U (t, x 0 (t), u))2 ≥ 0. 2 (3.103) Adding inequalities (3.102) and (3.103), we obtain the minimum condition (3.99). Therefore, we have proved the following assertion. Proposition 3.17. The conditions λ = (α0 , α, β, ψ, ν) ∈ M0 , 0 ≤ αy ≤ 1, λU = (α0 , α, β, ψ), and λV = ((1 − αy )λU , αy ) imply λV ∈ M0V . Further, let λ = (α0 , α, β, ψ, ν) ∈ M0+ , i.e., λ ∈ M0 , and the following conditions of the strict minimum principle hold: (3.81) implies (3.80) and (3.83) implies (3.82). We set λU = (α0 , α, β, ψ). It follows from the condition λ ∈ M0 that λU ∈ M0U . Let 0 < αy < 1 and λV = ((1 − αy )λU , αy ). (3.104) Then according to Proposition 3.17, λV ∈ M0V . Let us show that λV ∈ M0V + , i.e., the strict minimum principle at the point (0, w0 ) in Problem V holds for λV . First, let t ∈ [t0 , tf ] \ , and let u ∈ Rd(u) , (t, x 0 (t), u) ∈ Q1 , u = u0 (t). (3.105) If g(t, x 0 (t), u) = 0 in this case, then U (t, x 0 (t), u) = u (3.106) and the strict inequality (3.80) holds. It follows from (3.106) and (3.80) that H U (t, x 0 (t), u, ψx (t)) > H U (t, x 0 (t), u0 (t), ψx (t)). (3.107) Taking into account that αy > 0 and (u − U )2 ≥ 0, we obtain 1 H U (t, x 0 (t), u, ψx (t)) + αy (u − U (t, x 0 (t), u))2 2 > H U (t, x 0 (t), u0 (t), ψx (t)), (3.108) i.e., the strict minimum condition holds for the function H V defined by relation (3.86). If, along with conditions (3.105), the condition g(t, x 0 (t), u) = 0 holds, then this implies U (t, x 0 (t), u) = u, (3.109) 1 αy (u − U (t, x 0 (t), u))2 > 0, 2 (3.110) and then since αy > 0. Since λU ∈ M0U, (3.105) implies the nonstrict inequality H U (t, x 0 (t), u, ψx (t)) ≥ H U (t, x 0 (t), u0 (t), ψx (t)). (3.111) 3.2. Quadratic Sufficient Conditions on a Fixed Time Interval 145 Again, inequalities (3.110) and (3.111) imply the strict inequality (3.108). Therefore, for t∈ / , we have the strict minimum in u at the point u0 (t) for H V (t, 0, x 0 (t), u, ψy , ψx (t)). The case t = tk ∈ is considered analogously. Therefore, we have proved the following assertion. Proposition 3.18. The conditions λ = (α0 , α, β, ψ, ν) ∈ M0+ , 0 < αy < 1, λU = (α0 , α, β, ψ), and λV = ((1 − αy )λU , αy ) imply λV ∈ M0V + . Let Leg+ (M0 ) be the set of all strictly Legendre elements λ ∈ M0 in the problem (3.1), (3.2) at the point w0 . The definition of these elements was given in Section 3.2.3. Also, denote by Leg+ (M0V ) the set of all strictly Legendre elements λV ∈ M0V of Problem V at the point (0, w0 ). The definition of these elements was given in Section 2.1. Let us prove the following assertion. Proposition 3.19. The conditions λ = (α0 , α, β, ψ, ν) ∈ Leg+ (M0 ), 0 < αy < 1, λU = (α0 , α, β, ψ), and λV = ((1 − αy )λU , αy ) imply λV ∈ Leg+ (M0V ). Proof. The definitions of the element λV and the function H V imply 1 H V = (1 − αy )H U + αy (u − U )2 , 2 (3.112) where H V corresponds to the element λV and H U corresponds to the element λU . It follows from (3.112) that the relation HxV = (1 − αy )HxU holds on the trajectory (t, w 0 (t)). But, according to (3.52), HxU = H̄x , where H̄x = H̄xλ corresponds to the element λ. Therefore, HxV = (1 − αy )H̄x . (3.113) Further, since H V is independent of y, we have HyV = 0. (3.114) Also, (3.112) implies that HtV = (1 − αy )HtU on the trajectory (t, w0 (t)). But, according to (3.54), HtU = H̄t . Hence HtV = (1 − αy )H̄t . (3.115) Finally, the definitions of the functions H V and H̄ imply that the following relations hold on the trajectory (t, w 0 (t)): HψVx = f U (t, w0 ) = f (t, w0 ) = H̄ψx , (3.116) 1 HψVy = (u0 − U (t, w0 ))2 = 0. 2 (3.117) 146 Chapter 3. Quadratic Conditions for Optimal Control Problems We obtain from the definitions of D k (H V ) and D k (H̄ ) and also from conditions (3.113)– (3.117) that −D k (H V ) = HxV k+ HψVxk− + HyV k+ HψVyk− − (HxV k− HψVxk+ + HyV k− HψVyk+ ) + [HtV ]k = HxV k+ HψVxk− − HxV k− HψVxk+ + [HtV ]k = (1 − αy )H̄xk+ H̄ψk− − (1 − αy )H̄xk− H̄ψk+ + (1 − αy )[H̄t ]k x x = −(1 − αy )D k (H̄ ) ∀ tk ∈ . (3.118) Since λ ∈ Leg+ (M0 ) by condition, D k (H̄ ) > 0 for all tk ∈ . This and (3.118) together with the inequality 1 − αy > 0 imply D k (H V ) > 0 ∀ tk ∈ . (3.119) Let us verify the conditions for the strict Legendre property of the element λV . For V ū, ū , where ū ∈ Rd(u) , for this element this purpose, we calculate the quadratic form Huu 0 on the trajectory (t, w (t)). Differentiating relation (3.112) in u and multiplying it by ū, we obtain HuV ū = (1 − αy )HuU ū + αy (u − U ), (ū − Uu ū) . The repeated differentiation in u and the multiplication by ū yield V U ū, ū = (1 − αy )Huu ū, ū + αy (ū − Uu ū)2 − αy (u − U )Uuu ū, ū . Huu Substituting (t, w) = (t, w 0 (t)) and, moreover, taking into account that u0 = U (t, w0 ), V ū, ū = (1 − α )H U ū, ū + α (ū − U ū)2 . Finally, according to (3.45), we obtain Huu y y u uu U λ corresponds to the element λ. Hence Huu ū, ū = H̄uu (Uu ū), (Uu ū) , where H̄uu = H̄uu V Huu ū, ū = (1 − αy )H̄uu (Uu ū), (Uu ū) + αy (ū − Uu ū)2 . (3.120) The values of the derivatives are taken for (t, w) = (t, w 0 (t)). It is easy to verify that for each t ∈ [t0 , tf ], form (3.120) quadratic in ū is positivedefinite on Rd(u) . Indeed, suppose first that t ∈ [t0 , tf ] \ . Recall that the mapping ū ∈ Rd(u) → Uu (t, w 0 (t))ū ∈ Rd(u) is the projection on the subspace {ū ∈ Rd(u) | gu (t, w0 (t))ū = 0}. (3.121) This and the condition λ ∈ Leg+ (M0 ) imply that the quadratic form H̄uu (Uu ū), (Uu ū) is positive semidefinite on Rd(u) and positive definite on subspace (3.121). Furthermore, the quadratic form (ū − Uu ū)2 is positive semidefinite on Rd(u) and positive outside subspace (3.121). This and the conditions 0 < αy < 1 imply the positivity of the quadratic V ū, ū outside the origin of the space Rd(u) and, therefore, the positive definiteness form Huu d(u) on R . The case t ∈ [t0 , tf ] \ has been considered. The case t = tk ∈ is considered similarly. Therefore, all the conditions needed for element λV to belong to the set Leg+ (M0V ) hold. The proposition is proved. 3.2. Quadratic Sufficient Conditions on a Fixed Time Interval 147 Propositions 3.18 and 3.19 imply the following assertion. Lemma 3.20. The conditions λ = (α0 , α, β, ψ, ν) ∈ Leg+ (M0+ ), 0 < αy < 1, λU = (α0 , α, β, ψ), λV = ((1 − αy )λU , αy ) imply λV ∈ Leg+ (M0V + ). Fix a number αy such that 0 < αy < 1, e.g., αy = 1/2, and consider the linear operator λ = (α0 , α, β, ψ, ν) → λV = ((1 − αy )λU , αy ), (3.122) = (α0 , α, β, ψ) is the projection of the element λ. Lemma 3.20 implies the where following assertion. λU Lemma 3.21. Operator (3.122) transforms an arbitrary nonempty compact set M ⊂ Leg+ (M0+ ) into a nonempty compact set M V ⊂ Leg+ (M0V + ). Now let us consider the critical cone K V of Problem V at the point (0, w0 ). According to the definition given in Section 2.1, it consists of elements z̄V = (ξ̄ , ȳ, w̄) = (ξ̄ , ȳ, x̄, ū) such that the following conditions hold: z̄ = (ξ̄ , x̄, ū) = (ξ̄ , w̄) ∈ Z2 (), ȳ ∈ P W 1,2 (, R), Jp p̄ ≤ 0, Fip p̄ ≤ 0 ∀ i ∈ I , Kp p̄ = 0, ȳ0 = 0, ȳf ≤ 0, ȳ˙ = 0, [ȳ]k = 0 ∀ tk ∈ , x̄˙ = fx x̄ + fu (Uw w̄), [x̄]k = [f ]k ξ̄k ∀ tk ∈ . These conditions imply ȳ = 0, z̄ = (ξ̄ , x̄, ū) ∈ K U , where K U is the critical cone of problem U , i.e., the problem (3.1), (3.5), at the point w 0 . Then, according to Lemma 3.6, z̃ = (ξ̄ , x̄, Uw w̄) is an element of the critical cone K of the problem (3.1), (3.2) at the point w 0 . Therefore, we have proved the following assertion. Lemma 3.22. Let z̄V = (ξ̄ , ȳ, x̄, ū) = (ξ̄ , ȳ, w̄) ∈ K V . Then ȳ = 0, z̃ = (ξ̄ , x̄, Uw w̄) ∈ K. Therefore, the linear operator (ξ̄ , ȳ, x̄, ū) → (ξ̄ , x̄, Ux x̄ + Uu ū) (3.123) transforms the critical cone K V into the critical cone K. We now consider the quadratic forms. Let λ ∈ M0 be an arbitrary element, and let λV be its image under mapping (3.122). According to Proposition 3.17, λV ∈ M0V . For V Problem V and the point (0, w0 ), let us consider the quadratic form V λ (corresponding to the element λV ), which was defined in Section 2.1, and let us study its relation with the quadratic form λ (which corresponds to the element λ) at the point w0 of the problem (3.1), (3.2). We have already shown that D k (H V ) = (1 − αy )D k (H̄ ), tk ∈ (3.124) (we omit the superscripts λV and λ of H V and H̄ , respectively). It follows from (3.113) that [HxV ]k = (1 − αy )[H̄x ]k , tk ∈ , (3.125) 148 Chapter 3. Quadratic Conditions for Optimal Control Problems and (3.114) implies [HyV ]k = 0, tk ∈ . (3.126) We obtain from (3.112) that V w̄, w̄ Hww U = (1 − αy )Hww w̄, w̄ 1 + αy 2 ' ( ∂2 2 (u − U ) w̄, w̄ , ∂w 2 (3.127) where w̄ = (x̄, ū), x̄ ∈ P W 1,2 (, Rd(x) ), ū ∈ L2 (, Rd(u) ). According to (3.47), U w̄, w̄ = H̄ww w̃, w̃ , Hww (3.128) where w̃ = (x̄, ũ) = (x̄, Uw w̄). Let us calculate the second summand in formula (3.127). We have 1 ∂ (u − U )2 w̄ = (u − U )(ū − Uw w̄). 2 ∂w Therefore, for w = w 0 (t), we have 1 2 ' ( ) * ∂2 2 2 0 0 w̄, w̄ = ( ū − U (u − U ) w̄) − u − U (w , t) U w̄, w̄ w ww ∂w2 = (ū − Uw w̄)2 = (ū − ũ)2 . (3.129) We obtain from (3.127)–(3.129) that V w̄, w̄ = (1 − αy )H̄ww w̃, w̃ + αy (ū − ũ)2 , Hww (3.130) where ũ = Uw w̄ and w̃ = (x̄, ũ). Since H V is independent of y, we have V =0 Hyy and V Hyw = 0. (3.131) Let zV = (ξ̄ , ȳ, x̄, ū) be an arbitrary tuple such that z̄ = (ξ̄ , x̄, ū) = (ξ̄ , w̄) ∈ Z2 (), ȳ ∈ P W 1,2 . (3.132) V The definitions of the quadratic forms V λ and λ and also relations (3.124)–(3.126), (3.130), and (3.131) imply tf V (ū − ũ)2 dt, (3.133) V λ (z̄V ) = (1 − αy )λ (z̃) + αy t0 where ũ = Ux x̄ + Uu ū and z̃ = (ξ̄ , x̄, ũ). Therefore, we have proved the following assertion. Lemma 3.23. Let an element λ ∈ M0 and a tuple z̄V = (ξ̄ , ȳ, x̄, ū) satisfying conditions (3.132) be given. Let λV be the image of λ under mapping (3.122). Then formula (3.133) V holds for the quadratic form V λ calculated for Problem V, the point (0, w 0 ), and the element λV and also holds for the quadratic form λ calculated for the problem (3.1), (3.2), the point w0 , and the element λ. 3.2. Quadratic Sufficient Conditions on a Fixed Time Interval 149 We now assume that the following sufficient Condition B holds at the point w0 in problem (3.1), (3.2): there exist a nonempty compact set M ⊂ Leg+ (M0+ ) and a constant C > 0 such that max λ (z̄) ≥ C γ̄ (z̄) ∀ z̄ ∈ K. (3.134) M Let us show that in this case, the sufficient condition of Section 2.1 (denoted by BV ) holds at the point (0, w 0 ) in Problem V . Let z̄V = (ξ̄ , ȳ, x̄, ū) be an arbitrary element of the critical cone K V of Problem V at the point (0, w0 ). Then ȳ = 0, and, by Lemma 3.22, z̃ = (ξ̄ , x̄, ũ) ∈ K, where ũ = Uw w̄. Condition (3.134) implies the existence of λ ∈ M such that λ (z̃) ≥ C γ̄ (z̃). (3.135) λV Let be the image of λ under the mapping defined by operator (3.122). Then, by Lemma 3.23, formula (3.133) holds. It follows from (3.133) and (3.135) that tf V λV V (z̄ ) ≥ (1 − αy )C γ̄ (z̃) + αy (ū − ũ)2 dt. (3.136) t0 Therefore, V max V λ (z̄V ) ≥ (1 − αy )C γ̄ (z̃) + αy λV ∈M V tf (ū − ũ)2 dt, (3.137) t0 where M V is the image of the compact set M under the mapping defined by operator (3.122). By Lemma 3.21, (3.138) M V ⊂ Leg+ (M0V + ). Conditions (3.137) and (3.138) imply that the sufficient Condition BV holds at the point (0, w 0 ) in Problem V . To verify this, it suffices to show that the right-hand side of inequality (3.137) on the cone K V is estimated from below by the functional tf V V 2 2 2 ū2 dt = γ̄ (z̄) + ȳ02 γ̄ (z̄ ) := ξ̄ + ȳ0 + x̄0 + t0 with a small coefficient ε > 0. Since ȳ = 0 for all z̄ ∈ K V , it suffices to prove that there exists ε > 0 such that (1 − αy )C γ̄ (z̃) + αy (ū − ũ)2 dt ≥ εγ̄ (z̄) or tf tf tf (1 − αy )C ξ̄ 2 + x̄02 + ũ2 dt + αy (ū − ũ)2 dt ≥ ε ξ̄ 2 + x̄02 + ū2 dt . (3.139) t0 t0 t0 We set ū − ũ = û. Then ū2 = û2 + 2ûũ + ũ2 ≤ 2û2 + 2ũ2 . This obviously implies the estimate required. Therefore, we have proved the following assertion. Lemma 3.24. Let the sufficient Condition B hold at the point w 0 in problem (3.1), (3.2). Then the sufficient Condition BV of Section 2.1 holds at the point (0, w 0 ) in Problem V . By Theorem 2.101, Condition BV implies the existence of the bounded strong at the point (0, w0 ) in Problem V ; by Proposition 3.13, this implies the existence of the bounded strong γ -sufficiency at the point w0 in the problem (3.1), (3.2). Also, taking into account Lemma 3.24, we obtain that Condition B is sufficient for the bounded strong γ -sufficiency at the point w 0 in problem (3.1), (3.2). Therefore, we have proved Theorem 3.16, to which this section is devoted. γ V -sufficiency 150 3.3 3.3.1 Chapter 3. Quadratic Conditions for Optimal Control Problems Quadratic Conditions in the Problem with Mixed Control-State Equality Constraints on a Variable Time Interval Statement of the Problem Here, quadratic optimality conditions, both necessary and sufficient, are presented (as in [93]) in the following canonical Dubovitskii–Milyutin problem on a variable time interval. Let T denote a trajectory (x(t), u(t) | t ∈ [t0 , tf ]), where the state variable x(·) is a Lipschitzcontinuous function, and the control variable u(·) is a bounded measurable function on a time interval = [t0 , tf ]. The interval is not fixed. For each trajectory T we denote by p = (t0 , x(t0 ), tf , x(tf )) the vector of the endpoints of time-state variable (t, x). It is required to find T minimizing the functional J(T ) := J (p) → min (3.140) F (p) ≤ 0, K(p) = 0, ẋ(t) = f (t, x(t), u(t)), g(t, x(t), u(t)) = 0, p ∈ P , (t, x(t), u(t)) ∈ Q, (3.141) (3.142) (3.143) (3.144) subject to the constraints where P and Q are open sets, and x, u, F , K, f , and g are vector functions. We assume that the functions J , F , and K are defined and twice continuously differentiable on P , and that the functions f and g are defined and twice continuously differentiable on Q. It is also assumed that the gradients with respect to the control giu (t, x, u), i = 1, . . . , d(g) are linearly independent at each point (t, x, u) ∈ Q such that g(t, x, u) = 0. Here d(g) is a dimension of the vector g. 3.3.2 Necessary Conditions for a Pontryagin Minimum Let T be a fixed admissible trajectory such that the control u(·) is a piecewise Lipschitzcontinuous function on the interval with the set of discontinuity points = {t1 , . . . , ts }, t0 < t1 < · · · < ts < tf . Let us formulate a first-order necessary condition for optimality of the trajectory T . We introduce the Pontryagin function H (t, x, u, ψ) = ψf (t, x, u) (3.145) and the augmented Pontryagin function H̄ (t, x, u, ψ, ν) = H (t, x, u, ψ) + νg(t, x, u), (3.146) where ψ and ν are row vectors of the dimensions d(x) and d(g), respectively. Let us define the endpoint Lagrange function l(p, α0 , α, β) = α0 J (p) + αF (p) + βK(p), (3.147) 3.3. Quadratic Conditions on a Variable Time Interval 151 where p = (t0 , x0 , tf , xf ), x0 = x(t0 ), xf = x(tf ), α0 ∈ R, α ∈ (Rd(F ) )∗ , β ∈ (Rd(K) )∗ . Also we introduce a tuple of Lagrange multipliers λ = (α0 , α, β, ψ(·), ψ0 (·), ν(·)) (3.148) such that ψ(·) : → (Rd(x) )∗ and ψ0 (·) : → R1 are piecewise smooth functions, continuously differentiable on each interval of the set \ , and ν(·) : → (Rd(g) )∗ is a piecewise continuous function, Lipschitz continuous on each interval of the set \ . Denote by M0 the set of the normed tuples λ satisfying the conditions of the minimum principle for the trajectory T : α0 ≥ 0, α ≥ 0, αF (p) = 0, α0 + αi + |βj | = 1, ψ̇ = −H̄x , ψ̇0 = −H̄t , H̄u = 0, t ∈ \ , ψ(t0 ) = −lx0 , ψ(tf ) = lxf , ψ0 (t0 ) = −lt0 , ψ0 (tf ) = ltf , min u∈U(t,x(t)) H (t, x(t), u, ψ(t)) = H (t, x(t), u(t), ψ(t)), H (t, x(t), u(t), ψ(t)) + ψ0 (t) = 0, (3.149) t ∈ \ , t ∈ \ , where U(t, x) = {u ∈ Rd(u) | g(t, x, u) = 0, (t, x, u) ∈ Q}. The derivatives lx0 and lxf are at (p, α0 , α, β), where p = (t0 , x(t0 ), tf , x(tf )), and the derivatives H̄x , H̄u , and H̄t are at (t, x(t), u(t), ψ(t), ν(t)), where t ∈ \ . (Condition H̄u = 0 follows from the others conditions in this definition, and therefore could be excluded; yet we need to use it later.) Let us give the definition of Pontryagin minimum in problem (3.140)–(3.144) on a variable interval [t0 , tf ]. Definition 3.25. The trajectory T affords a Pontryagin minimum if there is no sequence of admissible trajectories T n = (x n (t), un (t) | t ∈ [t0n , tfn ]), n = 1, 2, . . . such that (a) J(T n ) < J(T ) for all n; (b) t0n → t0 , tfn → tf (n → ∞); (c) maxn ∩ |x n (t) − x(t)| → 0 (n → ∞), where n = [t0n , tfn ]; (d) n ∩ |un (t) − u(t)| dt → 0 (n → ∞); (e) there exists a compact set C ⊂ Q such that (t, x n (t), un (t)) ∈ C a.e. on n for all n. For convenience, let us give an equivalent definition of the Pontryagin minimum. Definition 3.26. The trajectory T affords a Pontryagin minimum if for each compact set C ⊂ Q there exists ε > 0 such that J(T˜ ) ≥ J(T ) for all admissible trajectories T˜ = (x̃(t), ũ(t) | t ∈ [t˜0 , t˜f ]) satisfying the following conditions: (a) |t˜0 − t0 | < ε, |t˜f − tf | < ε; ˜ = [t˜0 , t˜f ]; |x̃(t) − x(t)| < ε, where (b) max∩ ˜ (c) ∩ | ũ(t) − u(t)| dt < ε; ˜ ˜ (d) (t, x̃(t), ũ(t)) ∈ C a.e. on . The condition M0 = ∅ is equivalent to the Pontryagin minimum principle. It is a firstorder necessary condition of Pontryagin minimum for the trajectory T . Thus, the following theorem holds. 152 Chapter 3. Quadratic Conditions for Optimal Control Problems Theorem 3.27. If the trajectory T affords a Pontryagin minimum, then the set M0 is nonempty. Assume that M0 is nonempty. Using the definition of the set M0 and the full rank condition of the matrix gu on the surface g = 0, one can easily prove the following statement. Proposition 3.28. The set M0 is a finite-dimensional compact set, and the mapping λ → (α0 , α, β) is injective on M0 . As in Section 3.1, for each λ ∈ M0 , tk ∈ , we set D k (H̄ ) = −H̄xk+ H̄ψk− + H̄xk− H̄ψk+ − [H̄t ]k , (3.150) where H̄xk− = H̄x (tk , x(tk ), u(tk −), ψ(tk ), ν(tk −)), H̄xk+ = H̄x (tk , x(tk ), u(tk +), ψ(tk ), ν(tk +)), [H̄t ]k = H̄tk+ − H̄tk− , etc. Theorem 3.29. For each λ ∈ M0 the following conditions hold: D k (H̄ ) ≥ 0, k = 1, . . . , s. (3.151) Thus, conditions (3.151) follows from the minimum principle conditions (3.149). The following is an alternative method for calculating D k (H̄ ): For λ ∈ M0 , tk ∈ , consider the function (k H̄ )(t) = H̄ (tk , x(t), u(tk +), ψ(t), ν(tk +)) − H̄ (tk , x(t), u(tk −), ψ(t), ν(tk −)). Proposition 3.30. For each λ ∈ M0 the following equalities hold : d d (k H̄ )t=t − = (k H̄ )t=t + = −D k (H̄ ), k k dt dt k = 1, . . . , s. (3.152) Hence, for λ ∈ M0 the function (k H̄ )(t) has a derivative at the point tk ∈ equal to −D k (H̄ ), k = 1, . . . , s. Let us formulate a quadratic necessary condition of a Pontryagin minimum for the trajectory T . First, for this trajectory, we introduce a Hilbert space Z2 () and the critical cone K ⊂ Z2 (). We denote by P W 1,2 (, Rd(x) ) the Hilbert space of piecewise continuous functions x̄(·) : → Rd(x) , absolutely continuous on each interval of the set \ and such that their first derivative is square integrable. For each x̄ ∈ P W 1,2 (, Rd(x) ), tk ∈ , we set x̄ k− = x̄(tk −), x̄ k+ = x̄(tk +), [x̄]k = x̄ k+ − x̄ k− . Further, we denote z̄ = (t¯0 , t¯f , ξ̄ , x̄, ū), where t¯0 ∈ R1 , t¯f ∈ R1 , ξ̄ ∈ Rs , x̄ ∈ P W 1,2 (, Rd(x) ), ū ∈ L2 (, Rd(u) ). 3.3. Quadratic Conditions on a Variable Time Interval 153 Thus, z̄ ∈ Z2 () := R2 × Rs × P W 1,2 (, Rd(x) ) × L2 (, Rd(u) ). Moreover, for given z̄ we set w̄ = (x̄, ū), x̄0 = x̄(t0 ), x̄f = x̄(tf ), x̄¯0 = x̄(t0 ) + t¯0 ẋ(t0 ), x̄¯f = x̄(tf ) + t¯f ẋ(tf ), (3.153) p̄¯ = (t¯0 , x̄¯0 , t¯f , x̄¯f ). (3.154) By IF (p) = {i ∈ {1, . . . , d(F )} | Fi (p) = 0}, we denote the set of active indices of the constraints Fi (p) ≤ 0. Let K be the set of all z̄ ∈ Z2 () satisfying the following conditions: J (p)p̄¯ ≤ 0, Fi (p)p̄¯ ≤ 0 ∀ i ∈ IF (p), ˙ = fw (t, w(t))w̄(t) for a.a. t ∈ [t0 , tf ], x̄(t) K (p)p̄¯ = 0, [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, gw (t, w(t))w̄(t) = 0 for a.a. t ∈ [t0 , tf ], (3.155) where p = (t0 , x(t0 ), tf , x(tf )), w = (x, u). It is obvious that K is a convex cone in the Hilbert space Z2 (), and we call it the critical cone. If the interval is fixed, then we set p := (x0 , xf ) = (x(t0 ), x(tf )), and in the definition of K we have t¯0 = t¯f = 0, x̄¯0 = x̄0 , x̄¯f = x̄f , and p̄¯ = p̄ := (x̄0 , x̄f ). Let us introduce a quadratic form on Z2 (). For λ ∈ M0 and z̄ ∈ K, we set ¯ p̄¯ − 2ψ̇(tf )x̄(tf )t¯f − ψ̇(tf )ẋ(tf ) + ψ̇0 (tf ) t¯f2 ωe (λ, z̄) = lpp p̄, + 2ψ̇(t0 )x̄(t0 )t¯0 + ψ̇(t0 )ẋ(t0 ) + ψ̇0 (t0 ) t¯02 , (3.156) where lpp = lpp (p, α0 , α, β), p = (t0 , x(t0 ), tf , x(tf )). We also set ω(λ, z̄) = ωe (λ, z̄) + tf H̄ww w̄(t), w̄(t) dt, (3.157) t0 where H̄ww = H̄ww (t, x(t), u(t), ψ(t), ν(t)). Finally, we set 2(λ, z̄) = ω(λ, z̄) + s k ξ̄k , D k (H̄ )ξ̄k2 − 2[ψ̇]k x̄av (3.158) k=1 k = 1 (x̄ k− + x̄ k+ ), [ψ̇]k = ψ̇ k+ − ψ̇ k− . where x̄av 2 Now, we formulate the main necessary quadratic condition of Pontryagin minimum in the problem on a variable time interval. Theorem 3.31. If the trajectory T yields a Pontryagin minimum, then the following Condition A holds: the set M0 is nonempty and max (λ, z̄) ≥ 0 λ∈M0 ∀ z̄ ∈ K. 154 Chapter 3. Quadratic Conditions for Optimal Control Problems 3.3.3 Sufficient Conditions for a Bounded Strong Minimum Next, we give the definition of a bounded strong minimum in problem (3.140)–(3.144) on a variable interval [t0 , tf ]. To this end, let us give the definition of essential component of vector x in this problem: the ith component xi of vector x is called unessential if the functions f and g do not depend on this component and the functions J , F , and K are affine in xi0 = xi (t0 ), xif = x(tf ); otherwise the component xi is called essential. We denote by x a vector composed of all essential components of vector x. Definition 3.32. We say that the trajectory T affords a bounded strong minimum if there is no sequence of admissible trajectories T n = (x n (t), un (t) | t ∈ [t0n , tfn ]), n = 1, 2, . . . such that (a) J(T n ) < J(T ); (b) t0n → t0 , tfn → tf , x n (t0n ) → x(t0 ) (n → ∞); (c) maxn ∩ |x n (t) − x(t)| → 0 (n → ∞), where n = [t0n , tfn ]; (d) there exists a compact set C ⊂ Q such that (t, x n (t), un (t)) ∈ C a.e. on n for all n. An equivalent definition has the following form. Definition 3.33. The trajectory T affords a bounded strong minimum if for each compact set C ⊂ Q there exists ε > 0 such that J(T˜ ) ≥ J(T ) for all admissible trajectories T˜ = (x̃(t), ũ(t) | t ∈ [t˜0 , t˜f ]) satisfying the following conditions: (a) |t˜0 − t0 | < ε, |t˜f − tf | < ε, |x̃(t˜0 ) − x(t0 )| < ε; ˜ = [t˜0 , t˜f ]; |x̃(t) − x(t)| < ε, where (b) max∩ ˜ ˜ (c) (t, x̃(t), ũ(t)) ∈ C a.e. on . The strict bounded strong minimum is defined in a similar way, with the nonstrict inequality J(T˜ ) ≥ J(T ) replaced by the strict one and the trajectory T˜ required to be different from T . Finally, we define a (strict) strong minimum in the same way but omit condition (c) in the last definition. The following statement is quite obvious. Proposition 3.34. If there exists a compact set C ⊂ Q such that {(t, x, u) ∈ Q | g(t, x, u) = 0} ⊂ C, then a (strict) strong minimum is equivalent to a (strict) bounded strong minimum. Let us formulate a sufficient optimality Condition B, which is a natural strengthening of the necessary Condition A. The condition B is sufficient not only for a Pontryagin minimum, but also for a strict bounded strong minimum. To formulate the condition B, we introduce, for λ ∈ M0 , the following conditions of the strict minimum principle: (MP+ \ ) H (t, x(t), u, ψ(t)) > H (t, x(t), u(t), ψ(t)) for all t ∈ \ , u = u(t), u ∈ U(t, x(t)), and (MP+ ) H (tk , x(tk ), u, ψ(tk )) > H k for all tk ∈ , u ∈ U(tk , x(tk )), u = u(tk −), u = u(tk +), where H k := H k− = H k+ , H k− = H (tk , x(tk ), u(tk −), ψ(tk )), H k+ = H (tk , x(tk ), u(tk +), ψ(tk )). We denote by M0+ the set of + all λ ∈ M0 satisfying conditions (MP+ \ ) and (MP ). 3.3. Quadratic Conditions on a Variable Time Interval 155 For λ ∈ M0 we also introduce the strengthened Legendre–Clebsch conditions: (SLC\ ): For each t ∈ \ , the quadratic form H̄uu (t, x(t), u(t), ψ(t), ν(t))ū, ū is positive definite on the subspace of vectors ū ∈ Rd(u) such that gu (t, x(t), u(t))ū = 0. (SLCk− ): For each tk ∈ , the quadratic form H̄uu (tk , x(tk ), u(tk −), ψ(tk ), ν(tk −))ū, ū is positive definite on the subspace of vectors ū ∈ Rd(u) such that gu (tk , x(tk ), u(tk −))ū = 0. (SLCk+ ): this condition is symmetric to condition (SLCk− ) by replacing (tk −) everywhere by (tk +). Note that for each λ ∈ M0 the nonstrengthened Legendre–Clebsch conditions hold; i.e., the same quadratic forms are nonnegative on the corresponding subspaces. We denote by Leg+ (M0+ ) the set of all λ ∈ M0+ satisfying the strengthened Legendre– Clebsch conditions (SLC\ ), (SLCk− ), (SLCk+ ), k = 1, . . . , s, and also the conditions D k (H̄ ) > 0 ∀ k = 1, . . . , s. (3.159) Let us introduce the functional γ̄ (z̄) = t¯02 + t¯f2 + ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) + tf ū(t), ū(t) dt, (3.160) k = 1, . . . , s, (3.161) t0 which is equivalent to the norm squared on the subspace x̄˙ = fw (t, x(t), u(t))w̄; [x̄]k = [ẋ]k ξ̄k , of Hilbert space Z2 (). Recall that the critical cone K is contained in the subspace (3.161). Theorem 3.35. For the trajectory T , assume that the following Condition B holds: The set Leg+ (M0+ ) is nonempty and there exist a nonempty compact set M ⊂ Leg+ (M0+ ) and a number C > 0 such that max (λ, z̄) ≥ C γ̄ (z̄) λ∈M for all z̄ ∈ K. Then the trajectory T affords a strict bounded strong minimum. (3.162) 156 Chapter 3. Quadratic Conditions for Optimal Control Problems 3.3.4 Proofs The proofs are based on the quadratic optimality conditions, obtained in this chapter for problems on a fixed interval of time. We will give the proofs but omit some details. In order to extend the proofs to the case of a variable interval [t0 , tf ] we use a simple change of the time variable. Namely, we associate the fixed admissible trajectory T = (x(t), u(t) | t ∈ [t0 , tf ]) in the problem on a variable time interval (3.140)–(3.144) with a trajectory T τ = (v(τ ), t(τ ), x(τ ), u(τ ) | τ ∈ [τ0 , τf ]), considered on a fixed interval [τ0 , τf ], where τ0 = t0 , τf = tf , t(τ ) ≡ τ , v(τ ) ≡ 1. This is an admissible trajectory in the following problem on a fixed interval [τ0 , τf ]: Minimize the cost function J(T τ ) := J (t(τ0 ), x(τ0 ), t(τf ), x(τf )) → min (3.163) subject to the constraints F (t(τ0 ), x(τ0 ), t(τf ), x(τf )) ≤ 0, K(t(τ0 ), x(τ0 ), t(τf ), x(τf )) = 0, dx(τ ) dt(τ ) dv(τ ) = v(τ )f (t(τ ), x(τ ), u(τ )), = v(τ ), = 0, dτ dτ dτ g(t(τ ), x(τ ), u(τ )) = 0, (t(τ0 ), x(τ0 ), t(τf ), x(τf )) ∈ P , (t(τ ), x(τ ), u(τ )) ∈ Q. (3.164) (3.165) (3.166) (3.167) In this problem, x(τ ), t(τ ), and v(τ ) are state variables, and u(τ ) is a control variable. For brevity, we refer to problem (3.140)–(3.144) as problem P (on a variable interval = [t0 , tf ]), and to problem (3.163)–(3.167) as problem P τ (on a fixed interval [τ0 , τf ]). We denote by Aτ the necessary quadratic Condition A for problem P τ on a fixed interval [τ0 , τf ]. Similarly, we denote by Bτ the sufficient quadratic Condition B for problem P τ on a fixed interval [τ0 , τf ]. Recall that the control u(·) is a piecewise Lipschitz continuous function on the interval = [t0 , tf ] with the set of discontinuity points = {t1 , . . . , ts }, where t0 < t1 < · · · < ts < tf . Hence, for each λ ∈ M0 , the function ν(t) is also piecewise Lipschitz continuous on the interval , and, moreover, all discontinuity points of ν belong to . This easily follows from the equation H̄u = 0 and the full-rank condition for matrix gu . Consequently, u̇ and ν̇ are bounded measurable functions on . The proof of Theorem 3.31 is composed of the following chain of implications: (i) A Pontryagin minimum is attained on the trajectory T in problem P =⇒ (ii) A Pontryagin minimum is attained on the trajectory T τ in problem P τ =⇒ (iii) Condition Aτ holds for the trajectory T τ in problem P τ =⇒ (iv) Condition A holds for the trajectory T in problem P . The first implication is readily verified, the second follows from Theorem 3.9. The verification of the third implication (iii) ⇒ (iv) is not short and rather technical: we have to compare the sets of Lagrange multipliers, the critical cones, and the quadratic forms in both problems. This will be done below. In order to prove the sufficient conditions in problem P , given by Theorem 3.35, we have to check the following chain of implications: (v) Condition B holds for the trajectory T in problem P =⇒ (vi) Condition Bτ holds for the trajectory T τ in problem P τ =⇒ 3.3. Quadratic Conditions on a Variable Time Interval 157 (vii) A bounded strong minimum is attained on the trajectory T τ in problem P τ =⇒ (viii) A bounded strong minimum is attained on the trajectory T in problem P . The verification of the first implication (v) ⇒ (vi) is similar to the verification of the third implication (iii) ⇒ (iv) in the proof of the necessary conditions, the second implication (vi) ⇒ (vii) follows from Theorem 3.16, and the third (vii) ⇒ (viii) is readily verified. Thus, it remains to compare the sets of Lagrange multipliers, the critical cones, and the quadratic forms in problems P and P τ for the trajectories T and T τ , respectively. Comparison of the sets of Lagrange multipliers. Let us formulate the Pontryagin minimum principle in problem P τ for the trajectory T τ . The endpoint Lagrange function l, the Pontryagin function H , and the augmented Pontryagin function H̄ (all of them are equipped with the superscript τ ) have the form l τ = α0 J + αF + βK = l, H τ = ψvf + ψ0 v + ψv · 0 = v(ψf + ψ0 ), H̄ τ = H τ + νg. The set M0τ in problem P τ for the trajectory T τ consists of all tuples of Lagrange multipliers λτ = (α0 , α, β, ψ, ψ0 , ψv , ν) such that the following conditions hold: α0 + |α| + |β| = 1, dψ dψ0 dψv − = vψfx + νgx , − = vψft + νgt , − = ψf + ψ0 dτ dτ dτ ψ(τ0 ) = −lx0 , ψ(τf ) = lxf , ψ0 (τ0 ) = −lt0 , ψ0 (τf ) = ltf , ψv (τ0 ) = ψv (τf ) = 0, vψfu + νgu = 0, v(τ ) ψ(τ )f (t(τ ), x(τ ), u) + ψ0 (τ ) ≥ v(τ ) ψ(τ )f (t(τ ), x(τ ), u(τ )) + ψ0 (τ ) . (3.168) The last inequality holds for all u ∈ Rd(u) such that g(t(τ ), x(τ ), u) = 0, (t(τ ), x(τ ), u) ∈ Q. Recall that v(τ ) ≡ 1, t(τ ) ≡ τ , τ0 = t0 , and τf = tf . In (3.168), the function f and its derivatives fx , fu , ft , gx gu , and gt are taken at (t(τ ), x(τ ), u(τ )), τ ∈ [τ0 , τf ] \ , while the derivatives lt0 , lx0 , ltf , and lxf are calculated at (t(τ0 ), x(τ0 ), t(τf ), x(τf )) = (t0 , x(t0 ), tf , x(tf )). τ Conditions −dψv /dτ = ψf + ψ0 and ψv (τ0 ) = ψv (τf ) = 0 imply that τ0f (ψf + ψ0 ) dτ = 0. As is well known, conditions (3.168) of the minimum principle also imply that ψf + ψ0 = const, whence ψf + ψ0 = 0 and ψv = 0. Taking this fact into account and comparing the definitions of the sets M0τ (3.168) and M0 (3.149), we see that the projector α0 , α, β, ψ, ψ0 , ψv , ν → α0 , α, β, ψ, ψ0 , ν (3.169) realizes a one-to-one correspondence between these two sets. (Moreover, in the definition of the set M0τ one could replace the relations −dψv /dτ = ψf + ψ0 and ψv (τ0 ) = ψv (τf ) = 0 with ψf + ψ0 = 0, and thus identify M0τ with M0 .) We say that an element λτ ∈ M0τ corresponds to an element λ ∈ M0 if λ is the projection of λτ under the mapping (3.169). Comparison of the critical cones. For brevity, we set ! = (v, t, x, u) = (v, t, w). Let us define the critical cone K τ in problem P τ for the trajectory T τ . It consists of all 158 Chapter 3. Quadratic Conditions for Optimal Control Problems tuples (ξ̄ , v̄, t¯, x̄, ū) = (ξ̄ , !) ¯ satisfying the relations Jt0 t¯(τ0 ) + Jx0 x̄(τ0 ) + Jtf t¯(τf ) + Jxf x̄(τf ) ≤ 0, Fit0 t¯(τ0 ) + Fix0 x̄(τ0 ) + Fitf t¯(τf ) + Fixf x̄(τf ) ≤ 0, i ∈ IF (p), Kt0 t¯(τ0 ) + Kx0 x̄(τ0 ) + Ktf t¯(τf ) + Kxf x̄(τf ) = 0, d x̄ = v̄f + v ft t¯ + fx x̄ + fu ū , [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, dτ d v̄ d t¯ = v̄, [t¯]k = 0, k = 1, . . . , s, = 0, [v̄]k = 0, k = 1, . . . , s, dτ dτ gt t¯ + gx x̄ + gu ū = 0, (3.170) (3.171) (3.172) (3.173) (3.174) (3.175) where the derivatives Jt0 , Jx0 , Jtf Jxf , etc. are calculated at (t(τ0 ), x(τ0 ), t(τf ), x(τf )) = (t0 , x(t0 ), tf , x(tf )), while f , ft , fx , fu gt , gx , and gu are taken at (t(τ ), x(τ ), u(τ )), τ ∈ [τ0 , τf ] \ . Let (ξ̄ , v̄, t¯, x̄, ū) be an element of the critical cone K τ . We can use the following change of variables: x̃ = x̄ − t¯ẋ, ũ = ū − t¯u̇, (3.176) or, briefly, w̃ = w̄ − t¯ẇ. (3.177) Since v = 1, ẋ = f , and t = τ , equation (3.173) is equivalent to d x̄ = v̄ ẋ + ft t¯ + fw w̄. dt (3.178) Using the relation x̄ = x̃ + t¯ẋ in (3.178) along with t˙¯ = v̄, we get x̃˙ + t¯ẍ = t¯ft + fw w̄. (3.179) By differentiating the equation ẋ(t) = f (t, w(t)), we obtain ẍ = ft + fw ẇ. (3.180) x̃˙ = fw w̃. (3.181) Using this relation in (3.179), we get The relations [x̄]k = [ẋ]k ξ̄k , x̄ = x̃ + t¯ẋ imply [x̃]k = [ẋ]k ξ̃k , where ξ̃k = ξ̄k − t¯k , t¯k = t¯(tk ), (3.182) k = 1, . . . , s. (3.183) Further, relation (3.175) may be written as gt t¯ + gw w̄ = 0. Differentiating the relation g(t, w(t)) = 0, we obtain (3.184) gt + gw ẇ = 0. 3.3. Quadratic Conditions on a Variable Time Interval 159 These relations along with (3.177) imply that gw w̃ = 0. (3.185) Finally, note that since x̄ = x̃ + t¯ẋ and τ0 = t0 , τf = tf , we have p̄ = t¯0 , x̄(t0 ), t¯f , x̄(tf ) = t¯0 , x̃(t0 ) + t¯0 ẋ(t0 ), t¯f , x̃(tf ) + t¯f ẋ(tf ) , (3.186) where t¯0 = t¯(t0 ) and t¯f = t¯(tf ). The vector in the right-hand side of the last equality has the same form as the vector p̄¯ in definition (3.154). Consequently, all relations in definition (3.155) of the critical cone K in problem P are satisfied for the element z̃ = (t¯0 , t¯f , ξ̃ , w̃). We have proved that the obtained element z̃ belongs to the critical cone K in problem P . Conversely, if (t¯0 , t¯f , ξ̃ , w̃) is an element of the critical cone in problem P , then by setting v̄ = t¯f − t¯0 , tf − t0 t¯ = v̄(τ − τ0 ) + t¯0 , w̄ = w̃ + t¯ẇ, ξ̄k = ξ̃k + t¯(τk ), k = 1, . . . , s, we obtain an element (ξ̄ , v̄, t¯, w̄) of the critical cone (3.170)–(3.175) in problem P τ . Thus, we have proved the following lemma. Lemma 3.36. If (ξ̄ , v̄, t¯, w̄) is an element of the critical cone (3.170)–(3.175) in problem P τ for the trajectory T τ and t¯0 = t¯(t0 ), t¯f = t¯(tf ), w̃ = w̄ − t¯ẇ, ξ̃k = ξ̄k − t¯(tk ), k = 1, . . . , s, (3.187) then (t¯0 , t¯f , ξ̃ , w̃) is an element of the critical cone (3.155) in problem P for the trajectory T . Moreover, relations (3.187) define a one-to-one correspondence between elements of the critical cones in problems P τ and P . We say that an element (ξ̄ , v̄, t¯, w̄) of the critical cone in problem P τ corresponds to an element (t¯0 , t¯f , ξ̃ , w̃) of the critical cone in problem P if relations (3.187) hold. Comparison of the quadratic forms. Assume that the element λτ ∈ M0τ corresponds to the element λ ∈ M0 . Let us show that the quadratic form τ (λτ , ·), calculated on the element (ξ̄ , v̄, t¯, w̄) of the critical cone in problem P τ for the trajectory T τ , can be transformed into the quadratic form (λ, ·) calculated on the corresponding element (t¯0 , t¯f , ξ̃ , w̃) of the critical cone in problem P for the trajectory T . (i) The relations H̄ τ = v(H + ψ0 ) + νg, H̄ = H + νg, v = 1 imply τ !, ¯ !¯ = H̄ww w̄, w̄ + 2H̄tw w̄t¯ + H̄tt t¯2 + 2v̄(Hw w̄ + Ht t¯), H̄!! (3.188) where ! = (v, t, w), !¯ = (v̄, t¯, w̄). Since w̄ = w̃ + t¯ẇ, we have H̄ww w̄, w̄ = H̄ww w̃, w̃ + 2H̄ww ẇ, w̄ t¯ − H̄ww ẇ, ẇ t¯2 . Moreover, using the relations Hw = H̄w − νgw , Ht = H̄t − νgt , −ψ̇ = H̄x , −ψ̇0 = H̄t , H̄u = 0, gw w̄ + gt t¯ = 0, (3.189) 160 Chapter 3. Quadratic Conditions for Optimal Control Problems we obtain Hw w̄ + Ht t¯ = H̄w w̄ + H̄t t¯ − ν(gw w̄ + gt t¯) = H̄w w̄ + H̄t t¯ = H̄x x̄ + H̄t t¯ = −ψ̇ x̄ − ψ̇0 t¯. (3.190) Relations (3.188)–(3.190) imply τ !, H̄!! ¯ !¯ = H̄ww w̃, w̃ + 2H̄ww ẇ, w̄ t¯ + 2H̄tw w̄t¯ − H̄ww ẇ, ẇ t¯2 + H̄tt t¯2 − 2v̄ ψ̇ x̄ + ψ̇0 t¯ . (3.191) (ii) Let us transform the terms 2H̄ww ẇ, w̄ t¯ + 2H̄tw w̄t¯ in (3.191). By differentiating −ψ̇ = H̄x with respect to t, we obtain −ψ̈ = H̄tx + (ẇ)∗ H̄wx + ψ̇ H̄ψx + ν̇ H̄νx . Here we have H̄ψx = fx and H̄νx = gx . Therefore −ψ̈ = H̄tx + (ẇ)∗ H̄wx + ψ̇fx + ν̇gx . (3.192) Similarly, by differentiating H̄u = 0 with respect to t, we obtain 0 = H̄tu + (ẇ)∗ H̄wu + ψ̇fu + ν̇gu . (3.193) Multiplying (3.192) by x̄ and (3.193) by ū and summing the results, we get −ψ̈ x̄ = H̄tw w̄ + H̄ww ẇ, w̄ + ψ̇fw w̄ + ν̇gw w̄. (3.194) Since (ξ̄ , v̄, t¯, w̄) is an element of the critical cone in problem P τ , from (3.173) and (3.175) we get fw w̄ = x̄˙ − v̄ ẋ − ft t¯, gw w̄ = −gt t¯. Therefore, equation (3.194) can be represented in the form d H̄tw w̄ + H̄ww ẇ, w̄ = v̄(ψ̇ ẋ) − (ψ̇ x̄) + ψ̇ft + ν̇gt t¯, (3.195) dt which implies 2H̄ww ẇ, w̄ t¯ + 2H̄tw w̄ t¯ = 2t¯v̄(ψ̇ ẋ) − 2t¯ d (ψ̇ x̄) + 2 ψ̇ft + ν̇gt t¯2 . dt (3.196) (iii) Let us transform the term −H̄ww ẇ, ẇ t¯2 in (3.191). Multiplying (3.192) by ẋ and (3.193) by u̇ and summing the results, we obtain −ψ̈ ẋ = H̄tw ẇ + H̄ww ẇ, ẇ + ψ̇fw ẇ + ν̇gw ẇ. (3.197) From (3.180) and (3.184), we get fw ẇ = ẍ − ft , gw ẇ = −gt , respectively. Then (3.197) implies d H̄tw ẇ + H̄ww ẇ, ẇ = − (ψ̇ ẋ) + ψ̇ft + ν̇gt . (3.198) dt Multiplying this relation by −t¯2 , we get −H̄ww ẇ, ẇ t¯2 = H̄tw ẇt¯2 + t¯2 d (ψ̇ ẋ) − ψ̇ft + ν̇gt t¯2 . dt (3.199) 3.3. Quadratic Conditions on a Variable Time Interval 161 (iv) Finally, let us transform the term H̄tt t¯2 in (3.191). Differentiating −ψ̇0 = H̄t with respect to t and using the relations H̄ψt = ft and H̄νt = gt , we get Consequently, −ψ̈0 = H̄tt + H̄tw ẇ + ψ̇ft + ν̇gt . (3.200) H̄tt t¯2 = −ψ̈0 t¯2 − H̄tw ẇt¯2 − ψ̇ft + ν̇gt t¯2 . (3.201) (v) Summing (3.199) and (3.201), we obtain d −H̄ww ẇ, ẇ t¯2 + H̄tt t¯2 = −ψ̈0 t¯2 − 2 ψ̇ft + ν̇gt t¯2 + t¯2 (ψ̇ ẋ). dt (3.202) Using relations (3.196) and (3.202) in (3.191), we get τ H̄!! !, ¯ !¯ = H̄ww w̃, w̃ + 2t¯v̄(ψ̇ ẋ) − 2t¯ − ψ̈0 t¯2 + t¯2 But ψ̈0 t¯2 + 2v̄ t¯ψ̇0 = d ψ̇0 t¯2 , dt 2t¯v̄(ψ̇ ẋ) + t¯2 d (ψ̇ x̄) dt d (ψ̇ ẋ) − 2v̄ ψ̇ x̄ + ψ̇0 t¯ . dt t¯ (3.203) d d (ψ̇ x̄) + v̄(ψ̇ x̄) = t¯ψ̇ x̄ , dt dt d d (ψ̇ ẋ) = (ψ̇ ẋ t¯2 ). dt dt Therefore, τ H̄!! !, ¯ !¯ = H̄ww w̃, w̃ + d (ψ̇ ẋ)t¯2 − ψ̇0 t¯2 − 2ψ̇ x̄ t¯ . dt (3.204) Finally, using the change of the variable x̄ = x̃ + t¯ẋ in the right-hand side of this relation, we obtain d τ !, ¯ !¯ = H̄ww w̃, w̃ − (3.205) H̄!! (ψ̇0 + ψ̇ ẋ)t¯2 + 2ψ̇ x̃ t¯ . dt We have proved the following lemma. Lemma 3.37. Let (ξ̄ , v̄, t¯, w̄) = (ξ̄ , !) ¯ be an element of the critical cone K τ in problem P τ τ for the trajectory T . Set w̃ = w̄ − t¯ẇ. Then formula (3.205) holds. (vi) Recall that λτ is an arbitrary element of the set M0τ (consequently ψv = 0) and λ is the corresponding element of the set M0 , i.e., λ is the projection of λτ under the mapping (3.169). The quadratic form τ (λτ , ·) in problem P τ for the trajectory T τ has the following representation: ¯ = τ (λτ ; ξ̄ , !) s k k ξ̄k − 2[ψ̇0 ]k t¯av ξ̄k D k (H̄ τ )ξ̄k2 − 2[ψ̇]k x̄av k=1 τf τ H̄!! !, ¯ !¯ dτ . + lpp p̄, p̄ + τ0 (3.206) 162 Chapter 3. Quadratic Conditions for Optimal Control Problems Comparing the definitions of D k (H̄ τ ) and D k (H̄ ) (see (3.152)) and taking into account that H̄ τ = v(ψf + ψ0 ) + νg and v = 1, we get D k (H̄ τ ) = D k (H̄ ). (3.207) ¯ = (ξ̄ , v̄, t¯, x̄, ū) be an element of the critical cone K τ in the problem P τ Let z̄τ = (ξ̄ , !) for the trajectory T τ , and let z̃ = (t¯0 , t¯f , ξ̃ , x̃, ũ) be the corresponding element of the critical cone K in the problem P for the trajectory T ; i.e., relations (3.187) hold. Since [t¯]k = 0, k = 1, . . . , s, we have k = t¯k , k = 1, . . . , s (3.208) t¯av where t¯k = t¯(tk ), k = 1, . . . , s. Also recall that τ0 = t0 , τf = tf , t(τ ) = τ , dt = dτ . Since the functions ψ̇0 , ψ̇, ẋ, and x̃ may have discontinuities only at the points of the set , the following formula holds: tf t0 tf d (ψ̇0 + ψ̇ ẋ)t¯2 + 2ψ̇ x̃ t¯ dt = (ψ̇0 + ψ̇ ẋ)t¯2 + 2ψ̇ x̃ t¯ t0 dt s [ψ̇0 + ψ̇ ẋ]k t¯(tk )2 + 2[ψ̇ x̃]k t¯(tk ) . − (3.209) k=1 Relations (3.205)–(3.209) imply the following representation of the quadratic form τ on the element (ξ̄ , !) ¯ of the critical cone K τ : τ (λτ ; ξ̄ , !) ¯ = s k D k (H̄ )ξ̄k2 − 2[ψ̇]k x̄av ξ̄k − 2[ψ̇0 ]k t¯(tk )ξ̄k k=1 + [ψ̇0 + ψ̇ ẋ]k t¯(tk )2 + 2[ψ̇ x̃]k t¯(tk ) + lpp p̄, p̄ t t − (ψ̇0 + ψ̇ ẋ)t¯2 + 2ψ̇ x̃ t¯ tf + t0f H̄ww w̃, w̃ dτ . (3.210) 0 Let us transform the terms related to the discontinuity points tk of the control u(·), k = 1, . . . , s. For any λ ∈ M0 , the following lemma holds. Lemma 3.38. Let z̄ = (ξ̄ , !) ¯ = (ξ̄ , v̄, t¯, w̄) be an element of the critical cone K τ in the τ problem P for the trajectory T τ . Let the pair (ξ̃ , x̃) be defined by the relations ξ̃k = ξ̄k − t¯(tk ), k = 1, . . . , s, x̃ = x̄ − t¯ẋ. (3.211) Then for any k = 1, . . . , s the following formula holds: k ξ̄ − 2[ψ̇ ]k t¯(t )ξ̄ + [ψ̇ + ψ̇ ẋ]k t¯(t )2 + 2[ψ̇ x̃]k t¯(t ) D k (H̄ )ξ̄k2 − 2[ψ̇]k x̄av k 0 k k 0 k k k ξ̃ . = D k (H̄ )ξ̃k2 − 2[ψ̇]k x̃av k (3.212) Proof. In this proof, we omit the subscript and superscript k. We also write t¯ instead of t¯(tk ). Set a = D(H̄ ). Using the relations ξ̄ = ξ̃ + t¯, x̄av = x̃av + t¯ẋav , (3.213) 3.3. Quadratic Conditions on a Variable Time Interval 163 we obtain a ξ̄ 2 − 2[ψ̇]x̄av ξ̄ − 2[ψ̇0 ]t¯ξ̄ + [ψ̇0 + ψ̇ ẋ]t¯2 + 2[ψ̇ x̃]t¯ = a ξ̃ 2 + 2a ξ̃ t¯ + a t¯2 − 2[ψ̇]x̃av ξ̄ − 2[ψ̇]ẋav t¯ξ̄ − 2[ψ̇0 ]t¯ξ̄ + [ψ̇0 + ψ̇ ẋ]t¯2 + 2[ψ̇ x̃]t¯ = a ξ̃ 2 − 2[ψ̇]x̃av ξ̃ + r, (3.214) where r = 2a ξ̃ t¯ + a t¯2 − 2[ψ̇]x̃av t¯ − 2[ψ̇]ẋav t¯ξ̄ − 2[ψ̇0 ]t¯ξ̄ + [ψ̇0 + ψ̇ ẋ]t¯2 + 2[ψ̇ x̃]t¯. (3.215) It suffices to show that r = 0. Using the relations (3.213) in formula (3.215), we get r 2a(ξ̄ − t¯)t¯ + a t¯2 − 2[ψ̇](x̄av − t¯ẋav )t¯ − 2[ψ̇]ẋav t¯ξ̄ − 2[ψ̇0 ]t¯ξ̄ + [ψ̇0 + ψ̇ ẋ]t¯2 + 2[ψ̇(x̄ − t¯ẋ)]t¯ = t¯2 − a + 2[ψ̇]ẋav + [ψ̇0 ] − [ψ̇ ẋ] + 2t¯ξ̄ a − [ψ̇]ẋav − [ψ̇0 ] + 2t¯ − [ψ̇]x̄av + [ψ̇ x̄] . = The coefficient of t¯2 in the right-hand side of the last equality vanishes: −a + 2[ψ̇]ẋav + [ψ̇0 ] − [ψ̇ ẋ] = − ψ̇ + ẋ − − ψ̇ − ẋ + + [ψ̇0 ] + (ψ̇ + − ψ̇ − )(ẋ + + ẋ − ) + [ψ̇0 ] − ψ̇ + ẋ + + ψ̇ − ẋ − = 0. The coefficient of 2t¯ξ̄ is equal to 1 = ψ̇ + ẋ − − ψ̇ − ẋ + + [ψ̇0 ] − (ψ̇ + − ψ̇ − )(ẋ − + ẋ + ) − [ψ̇0 ] 2 1 1 + − = ψ̇ ẋ − ψ̇ − ẋ + − [ψ̇ ẋ]. 2 2 The coefficient of 2t¯ is equal to 1 −[ψ̇]x̄av + [ψ̇ x̄] = − (ψ̇ + − ψ̇ − )(x̄ − + x̄ + ) + (ψ̇ + x̄ + − ψ̇ − x̄ − ) 2 1 1 + ψ̇ [x̄] + ψ̇ − [x̄] = ψ̇av [ẋ]ξ̄ , = 2 2 since [x̄] = [ẋ]ξ̄ . Consequently, 1 1 + − r = 2t¯ξ̄ ψ̇ ẋ − ψ̇ − ẋ + − [ψ̇ ẋ] + ψ̇av [ẋ] 2 2 + − − + + = t¯ξ̄ (ψ̇ ẋ − ψ̇ ẋ ) − (ψ̇ ẋ + − ψ̇ − ẋ − ) + (ψ̇ − + ψ̇ + )(ẋ + − ẋ − ) = 0. a − [ψ̇]ẋav − [ψ̇0 ] In view of (3.214) the equality r = 0 proves the lemma. Relation (3.210) along with equality (3.212) gives the following transformation of ¯ of the critical cone K τ , quadratic form τ (see (3.206)) on the element z̄τ = (ξ̄ , !) ¯ τ (λτ ; ξ̄ , !) = s k D k (H̄ )ξ̃k2 − 2[ψ̇]k x̃av ξ̃k tf + lpp p̄, p̄ − (ψ̇0 + ψ̇ ẋ)t¯2 + 2ψ̇ x̃ t¯ + k=1 t0 tf t0 H̄ww w̃, w̃ dτ . (3.216) 164 Chapter 3. Quadratic Conditions for Optimal Control Problems Taking into account (3.186) and definitions (3.156)–(3.158) of quadratic forms ωe , ω, and , we see that the right-hand side of (3.216) is the quadratic form (λ, z̃) (see (3.158)) in problem P for the trajectory T , where z̃ = (t¯0 , t¯f , ξ̃ , w̃) is the corresponding element of the critical cone K. Thus we have proved the following theorem. Theorem 3.39. Let z̄τ = (ξ̄ , v̄, t¯, w̄) be an element of the critical cone K τ in problem P τ for the trajectory T τ . Let z̃ = (t¯0 , t¯f , ξ̃ , w̃) be the corresponding element of the critical cone K in problem P for the trajectory T , i.e., relations (3.187) hold. Then for any λτ ∈ M0τ and the corresponding projection λ ∈ M0 (under the mapping (3.169)) the following equality holds: τ (λτ , z̄τ ) = (λ, z̃). This theorem proves the implications (iii) ⇒ (iv) and (v) ⇒ (vi) (see the beginning of this section), and thus completes the proofs of Theorems 3.31 and 3.35. 3.4 Quadratic Conditions for Optimal Control Problems with Mixed Control-State Equality and Inequality Constraints In this section, we give a statement of the general optimal control problem with mixed control-state equality and inequality constraints on fixed and variable time intervals, recall different concepts of minimum, and formulate optimality conditions. 3.4.1 General Optimal Control Problem on a Fixed Time Interval Statement of the problem. We consider the following optimal control problem on a fixed interval of time [t0 , tf ]: (3.217) Minimize J(x, u) = J x(t0 ), x(tf ) subject to the constraints F x(t0 ), x(tf ) ≤ 0, K x(t0 ), x(tf ) = 0, x(t0 ), x(tf ) ∈ P , ẋ = f (t, x, u), g(t, x, u) = 0, ϕ(t, x, u) ≤ 0, (t, x, u) ∈ Q, (3.218) (3.219) where P and Q are open sets and x, u, F , K, f , g, and ϕ are vector functions, d(g) ≤ d(u). We use the notation x(t0 ) = x0 , x(tf ) = xf , (x0 , xf ) = p, and (x, u) = w. We seek the minimum among the pairs of functions w(·) = (x(·), u(·)) such that x(·) ∈ W 1,1 ([t0 , tf ], Rd(x) ), u(·) ∈ L∞ ([t0 , tf ], Rd(u) ). Therefore, we seek for the minimum in the space W := W 1,1 ([t0 , tf ], Rd(x) ) × L∞ ([t0 , tf ], Rd(u) ). A pair w = (x, u) ∈ W is said to be admissible in problem (3.217)–(3.219) if constraints (3.218)–(3.219) are satisfied by w. Assumption 3.40. (a) The functions J (p), F (p), and K(p) are defined and twice continuously differentiable on the open set P ⊂ R2d(x) , and the functions f (t, w), g(t, w), and ϕ(t, w) are defined and twice continuously differentiable on the open set Q ⊂ Rd(x)+d(u)+1 . (b) The gradients with respect to the control giu (t, w), i = 1, . . . , d(g), ϕj u (t, w), j ∈ Iϕ (t, w) 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 165 are linearly independent at all points (t, w) ∈ Q such that g(t, w) = 0 and ϕ(t, w) ≤ 0. Here gi and ϕj are the components of the vector functions g and ϕ, respectively, and Iϕ (t, w) = {j ∈ {1, . . . , d(ϕ)} | ϕj (t, w) = 0} (3.220) is the set of indices of active inequality constraints ϕj (w, t) ≤ 0 at (t, w) ∈ Q. We refer to (b) as the linear independence assumption for the gradients of the active mixed constraints with respect to the control. Let a pair w0 (·) = (x 0 (·), u0 (·)) ∈ W satisfying constraints (3.218)–(3.219) of the problem be the point tested for optimality. Assumption 3.41. The control u0 (·) is a piecewise continuous function such that all its discontinuity points are L-points (see Definition 2.1). Let = {t1 , . . . , ts }, t0 < t1 < · · · < ts < tf be the set of all discontinuity points of u0 (·). It is also assumed that (tk , x 0 (tk ), u0k− ) ∈ Q, (tk , x 0 (tk ), u0k+ ) ∈ Q, k = 1, . . . , s, where u0k− = u0 (tk − 0), u0k+ = u0 (tk + 0). In what follows, we assume for definiteness that the set of discontinuity points of u0 is nonempty. Whenever this set is empty, all statements admit obvious simplifications. Minimum on a set of sequences. Weak and Pontryagin minimum. Let S be an arbitrary set of sequences {δwn } in the space W closed with respect to the operation of taking subsequences. For problem (3.217)–(3.219), let us define a concept of minimum on S at the admissible point w0 = (x 0 , u0 ). Set p0 = (x 0 (t0 ), x 0 (tf )). Definition 3.42. We say that w0 is a (strict) minimum on S if there exists no sequence {δwn } ∈ S such that the following conditions hold for all its members: J (p0 + δpn ) < J (p0 ) (J (p0 + δpn ) ≤ J (p 0 ), δwn = 0), F (p 0 + δpn ) ≤ 0, K(p 0 + δpn ) = 0, ẋ 0 + δ ẋn = f (t, w 0 + δwn ), g(t, w 0 + δwn ) = 0, ϕ(t, w0 + δwn ) ≤ 0, (p 0 + δpn ) ∈ P , (t, w0 + δwn ) ∈ Q, (3.221) (3.222) (3.223) (3.224) where δpn = (δxn (t0 ), δxn (tf )), and δwn = (δxn , δun ). Any sequence from S which satisfies conditions (3.221)–(3.224) is said to violate the (strict) minimality on S. Let S 0 be the set of sequences {δwn } in W such that δwn = δxn A weak minimum is a minimum on S 0 . 1,1 + δun ∞ → 0. Definition 3.43. We say that w0 is a point of Pontryagin minimum if this point is a minimum on the set of sequences {δwn } in W satisfying the following two conditions: t (a) δxn 1,1 + δun 1 → 0, where δun 1 = t0f |δun | dt; (b) there exists a compact set C ⊂ Q (which depends on the choice of the sequence) such that for all sufficiently large n, we have (t, w 0 (t) + δwn (t)) ∈ C a.e. on [t0 , tf ]. Any sequence satisfying conditions (a) and (b) will be referred to as a Pontryagin sequence on Q. Obviously, every Pontryagin minimum is a weak minimum. 166 Chapter 3. Quadratic Conditions for Optimal Control Problems Minimum principle. Let us state the well-known first-order necessary conditions for both a weak and for a Pontryagin minimum. These conditions are often referred to as the local and integral minimum principle, respectively. The local minimum principle, which we give first, is conveniently identified with the nonemptiness of the set 0 defined below. Let (3.225) l = α0 J + αF + βK, H = ψf , H̄ = H + νg + μϕ, where α0 is a scalar, and α, β, ψ, ν, and μ are row vectors of the same dimensions as F , K, f , g, and ϕ, respectively. The dependence of the functions l, H , and H̄ on the variables is as follows: l = l(p, α0 , α, β), H = H (t, w, ψ), H̄ = H̄ (t, w, ψ, ν, μ). The function l is said to be the endpoint Lagrange function, H is the Pontryagin function (or the Hamiltonian), and H̄ is the augmented Pontryagin function (or the augmented Hamiltonian). Denote by Rn∗ the space of n-dimensional row-vectors. Set λ = (α0 , α, β, ψ(·), ν(·), μ(·)), (3.226) where α0 ∈ R1 , α ∈ Rd(F )∗ , β ∈ Rd(K)∗ , ψ(·) ∈ W 1,1 ([t0 , tf ], Rd(x)∗ ), ν(·) ∈ L∞ ([t0 , tf ], Rd(g)∗ ), μ(·) ∈ L∞ ([t0 , tf ], Rd(ϕ)∗ ). Denote by 0 the set of all tuples λ satisfying the conditions α0 ≥ 0, α0 + d(F ) i=1 α ≥ 0, αi + d(K) αF (p 0 ) = 0, |βj | = 1, (3.227) (3.228) j =1 μ(t) ≥ 0, μ(t)ϕ(t, w 0 (t)) = 0, ψ̇ = −H̄x (t, w 0 (t), ψ(t), ν(t), μ(t)), ψ(t0 ) = −lx0 (p 0 , α0 , α, β), ψ(tf ) = lxf (p0 , α0 , α, β), (3.229) (3.230) (3.231) H̄u (t, w0 (t), ψ(t), ν(t), μ(t)) = 0, (3.232) where αi and βj are components of the row vectors α and β, respectively, and H̄x , H̄u , lx0 , and lxf are gradients with respect to the corresponding variables. It is well known that if w0 is a weak minimum, then 0 is nonempty (see, e.g., [30]). The latter condition is just the local minimum principle. Note that 0 can consist of more than one element. The following result pertain to this possibility. Proposition 3.44. The set 0 is a finite-dimensional compact set, and the projection λ = (α0 , α, β, ψ, ν, μ) → (α0 , α, β) is injective on 0 . This property of 0 follows from the linear independence assumption for the gradients of the active mixed constraints with respect to the control. This assumption also guarantees the following property. Proposition 3.45. Let λ ∈ 0 be an arbitrary tuple. Then its components ν(t) and μ(t) are continuous at each point of continuity of the control u0 (t). Consequently, ν(t) and μ(t) are piecewise continuous functions such that all their discontinuity points belong to the set . The adjoint variable ψ(t) is a piecewise smooth function such that all its break points belong to the set . 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 167 In a similar way, the integral minimum principle, which is a first-order necessary condition for a Pontryagin minimum at w0 , can be stated as the nonemptiness of the set M0 defined below. Let U(t, x) = u ∈ Rd(u) | (t, x, u) ∈ Q, g(t, x, u) = 0, ϕ(t, x, u) ≤ 0 . (3.233) Denote by M0 the set of all tuples λ ∈ 0 such that for all t ∈ [t0 , tf ] \ , the inclusion u ∈ U(t, x 0 (t)) implies the inequality H (t, x 0 (t), u, ψ(t)) ≥ H (t, x 0 (t), u0 (t), ψ(t)). (3.234) It is known [30] that if w0 is a Pontryagin minimum, then M0 is nonempty. The latter condition is just the integral (or Pontryagin) minimum principle. Inequality (3.234), satisfied for all t ∈ [t0 , tf ] \ , is called the minimum condition of Pontryagin’s function H with respect to u. (In the case of a measurable control u0 (t), inequality (3.234) is fulfilled for a.a. t ∈ [t0 , tf ].) Note that, just like 0 , the set M0 can contain more than one element. Since this set is closed and M0 ⊂ 0 , it follows from Proposition 3.44 that M0 is also a finite-dimensional compact set. Let us note one more important property of the set M0 (see [30]). Proposition 3.46. Let λ ∈ M0 be an arbitrary tuple. Then there exists an absolutely continuous function ψt (t) from [t0 , tf ] into R1 such that ψ̇t = −H̄t (t, w0 (t), ψ(t), ν(t), μ(t)), H (t, w0 (t), ψ(t)) + ψt (t) = 0. (3.235) (3.236) Consequently, ψt (t) is a piecewise smooth function whose break points belong to . Particularly, this implies the following assertion. Let λ ∈ M0 be an arbitrary tuple. Then the function H (t, w 0 (t), ψ(t)) satisfies the following condition: [H λ ]k = 0 ∀ tk ∈ , (3.237) where [H λ ]k is a jump of the function H (t, x 0 (t), u0 (t), ψ(t)) at the point tk ∈ , defined by the relations [H λ ]k = H λk+ − H λk− , H λk− = H (tk , x 0 (tk ), u0k− , ψ(tk )), and H λk+ = H (tk , x 0 (tk ), u0k+ , ψ(tk )). Here, by definition, u0k− = u0 (tk − 0), and u0k+ = u0 (tk + 0). Let 0 be the set of all tuples λ ∈ 0 satisfying condition (3.237). From Propositions 3.45 and 3.46 we have the following. Proposition 3.47. The set 0 is a finite-dimensional compact set such that M0 ⊂ 0 ⊂ 0 . Note that from minimum condition (3.234), the inequality H (tk , x 0 (tk ), u, ψ(tk )) ≥ H λk ∀ u ∈ U(tk , x 0 (tk )) (3.238) follows by continuity, where by definition H λk := H λk− = H λk+ . Condition (3.238) holds for any tk ∈ and any λ ∈ M0 . Now we will state two properties of elements of the set M0 which follow from the minimum principle. The first is a necessary optimality condition related to each discontinuity point of the control u0 . The second is a generalization of the Legendre condition. 168 Chapter 3. Quadratic Conditions for Optimal Control Problems The value D k (H̄ λ ). Let λ ∈ 0 and tk ∈ . According to Proposition 3.45 the quantities μk− = μ(tk − 0), μk+ = μ(tk + 0), ν k− = ν(tk − 0), ν k+ = ν(tk + 0) are well defined. Set H̄xλk− := H̄x (tk , x 0 (tk ), u0k− , ψ(tk ), ν k− , μk− ), (3.239) H̄xλk+ := H̄x (tk , x 0 (tk ), u0k+ , ψ(tk ), ν k+ , μk+ ). Similarly, set H̄ψλk− = f k− := f (tk , x 0 (tk ), u0k− ), H̄ψλk+ = f k+ := f (tk , x 0 (tk ), u0k+ ), (3.240) H̄tλk− := H̄t (tk , x 0 (tk ), u0k− , ψ(tk ), ν k− , μk− ), H̄tλk+ := H̄t (tk , x 0 (tk ), u0k+ , ψ(tk ), ν k+ , μk+ ). (3.241) D k (H̄ λ ) := −H̄xλk+ H̄ψλk− + H̄xλk− H̄ψλk+ − [H̄tλ ]k , (3.242) Finally, set where [H̄tλ ]k = H̄tλk+ − H̄tλk− is the jump of H̄t (t, x 0 (t), u0 (t), ψ(t), ν(t), μ(t)) at tk . Note that D k (H̄ λ ) is linear in λ. Theorem 3.48. Let λ ∈ M0 . Then D k (H̄ λ ) ≥ 0 for all tk ∈ . Since conditions D k (H̄ λ ) ≥ 0 for all tk ∈ follow from the minimum principle, they are necessary conditions for the Pontryagin minimum at the point w 0 . As in previous problems, there is another way, convenient for practical use, to calculate the quantities D k (H̄ λ ). Given any λ ∈ 0 and tk ∈ , we set (k H̄ λ )(t) = H̄ (t, x 0 (t), u0k+ , ψ(t), ν k+ , μk+ ) − H̄ (t, x 0 (t), u0k− , ψ(t), ν k− , μk− ). (3.243) The function (k H̄ λ )(t) is continuously differentiable at each point of the set [t0 , tf ] \ , since this property hold for x 0 (t) and ψ(t). The latter follows from the equation ẋ 0 (t) = f (t, x 0 (t), u0 (t)), (3.244) adjoint equation (3.230), and Assumption 3.41. Proposition 3.49. For any λ ∈ 0 and any tk ∈ , the following equalities hold: D k (H̄ λ ) = − d d (k H̄ λ )(tk − 0) = − (k H̄ λ )(tk + 0). dt dt (3.245) Finally, note that equation (3.244) can be written in the form ẋ 0 (t) = H̄ψ (t, x 0 (t), u0 (t), ψ(t), ν(t), μ(t)). (3.246) Relations (3.244), (3.230), (3.235), (3.236), and formula (3.242) can be used for obtaining one more representation of the value D k (H̄ λ ). 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 169 Proposition 3.50. For any λ ∈ 0 and tk ∈ , the following equality holds: D k (H̄ λ ) = ψ̇ k+ ẋ 0k− − ψ̇ k− ẋ 0k+ + [ψ̇t ]k , (3.247) where the function ψt (t) is defined by ψt (t) = −H (t, x 0 (t), u0 (t), ψ(t)), the value [ψ̇t ]k = ψ̇tk+ − ψ̇tk− is the jump of the derivative ψ̇t (t) at the point tk , and the vectors ẋ 0k− , ψ̇ k− , ψ̇tk− and ẋ 0k+ , ψ̇ k+ , ψ̇tk+ are the left and the right limit values of the derivatives ẋ 0 (t), ψ̇(t) and ψ̇t (t) at tk , respectively. Legendre–Clebsch condition. For any λ = (α0 , α, β, ψ, ν, μ) ∈ 0 , let us define the following three conditions: (LC) For any t ∈ [t0 , tf ] \ , the quadratic form H̄uu (t, x 0 (t), u0 (t), ψ(t), ν(t), μ(t))ū, ū (3.248) of the variable ū is positive semidefinite on the cone formed by the vectors ū ∈ Rd(u) such that gu (t, x 0 (t), u0 (t))ū = 0, (3.249) ϕj u (t, x 0 (t), u0 (t))ū ≤ 0 ∀ j ∈ Iϕ (t, x 0 (t), u0 (t)), μj (t)ϕj u (t, x 0 (t), u0 (t))ū = 0 ∀ j ∈ Iϕ (t, x 0 (t), u0 (t)), where H̄uu is the matrix of second derivatives with respect to u of the function H̄ , and Iϕ (t, x, u) is the set of indices of active inequality constraints ϕj (t, x, u) ≤ 0 at (t, x, u), defined by (3.220). (LC− ) For any tk ∈ , the quadratic form H̄uu (tk , x 0 (tk ), u0k− , ψ(tk ), ν k− , μk− )ū, ū (3.250) of the variable ū is positive semidefinite on the cone formed by the vectors ū ∈ Rd(u) such that gu (tk , x 0 (tk ), u0k− )ū = 0, ϕj u (tk , x 0 (tk ), u0k− )ū ≤ 0 ∀ j ∈ Iϕ (tk , x 0 (tk ), u0k− ), (3.251) k− 0 0k− 0 0k− μj ϕj u (tk , x (tk ), u )ū = 0 ∀ j ∈ Iϕ (tk , x (tk ), u ) (LC+ ) For any tk ∈ , the quadratic form H̄uu (tk , x 0 (tk ), u0k+ , ψ(tk ), ν k+ , μk+ )ū, ū (3.252) of the variable ū is positive semidefinite on the cone formed by the vectors ū ∈ Rd(u) such that gu (tk , x 0 (tk ), u0k+ )ū = 0, ϕj u (tk , x 0 (tk ), u0k+ )ū ≤ 0 ∀ j ∈ Iϕ (tk , x 0 (tk ), u0k+ ), (3.253) 0 (t ), u0k+ )ū = 0 ∀ j ∈ I (t , x 0 (t ), u0k+ ). μk+ ϕ (t , x ju k k ϕ k k j We say that element λ ∈ 0 satisfies the Legendre–Clebsch condition if conditions (LC), + − (LC− ) and (LC ) hold. Clearly, these conditions are not independent: conditions (LC ) + and (LC ) follow from condition (LC) by continuity. 170 Chapter 3. Quadratic Conditions for Optimal Control Problems Theorem 3.51. For any λ ∈ M0 , the Legendre–Clebsch condition holds. Thus, the Legendre–Clebsch condition is also a consequence of the minimum principle. Legendrian elements. An element λ ∈ 0 is said to be Legendrian if, for this element, the Legendre–Clebsch condition is satisfied and also the following conditions hold: [H λ ]k = 0, D k (H̄ λ ) ≥ 0 ∀ tk ∈ . (3.254) Let M be an arbitrary subset of the compact set 0 . Denote by Leg(M) the subset of all Legendrian elements λ ∈ M. It follows from Theorems 3.48 and 3.51 and Proposition 3.46 that Leg(M0 ) = M0 . (3.255) Now we introduce the critical cone K and the quadratic form λ (·) which will be used for the statement of the quadratic optimality condition. Critical cone. As above, we denote by P W 1,2 ([t0 , tf ], Rd(x) ) the space of piecewise continuous functions x̄(·) : [t0 , tf ] → Rd(x) that are absolutely continuous on each interval of the set [t0 , tf ] \ and have a square integrable first derivative. Given tk ∈ and x̄(·) ∈ P W 1,2 ([t0 , tf ], Rd(x) ), we use the notation x̄ k− = x̄(tk − 0), x̄ k+ = x̄(tk + 0), [x̄]k = x̄ k+ − x̄ k− . Denote by Z2 () the space of triples z̄ = (ξ̄ , x̄, ū) such that ξ̄ = (ξ̄1 , . . . , ξ̄s ) ∈ Rs , x̄ ∈ P W 1,2 ([t0 , tf ], Rd(x) ), ū ∈ L2 ([t0 , tf ], Rd(u) ). Thus, Z2 () = Rs × P W 1,2 ([t0 , tf ], Rd(x) ) × L2 ([t0 , tf ], Rd(u) ). Denote by IF (p0 ) = {i ∈ {1, . . . , d(F )} | Fi (p 0 ) = 0} the set of indices of active inequality constraints Fi (p) ≤ 0 at the point p 0 , where Fi are the components of the vector function F . Let K denote the set of z̄ = (ξ̄ , x̄, ū) ∈ Z2 () such that J (p 0 )p̄ ≤ 0, Fi (p0 )p̄ ≤ 0, i ∈ IF (p0 ), K (p 0 )p̄ = 0, ˙ = fw (t, w 0 (t))w̄(t), [x̄]k = [ẋ 0 ]k ξ̄k , tk ∈ , x̄(t) gw (t, w0 (t))w̄(t) = 0, ϕj w (t, w 0 (t))w̄(t) ≤ 0 a.e. on M0 (ϕj0 ), j = 1, . . . , d(ϕ), (3.256) (3.257) (3.258) (3.259) where M0 (ϕj0 ) = {t ∈ [t0 , tf ] | ϕj (t, w0 (t)) = 0}, p̄ = (x̄(t0 ), x̄(tf )), w̄ = (x̄, ū), and [ẋ 0 ]k is the jump of the function ẋ 0 (t) at the point tk ∈ , i.e., [ẋ 0 ]k = ẋ 0k+ − ẋ 0k− = ẋ 0 (tk + 0) − ẋ 0 (tk − 0), tk ∈ . Obviously, K is a closed convex cone in the space Z2 (). We call it the critical cone of problem (3.217)–(3.219) at the point w 0 . The following question is of interest: Which inequalities in the definition of K can be replaced by equalities without affecting K? This question is answered below. 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 171 Proposition 3.52. For any λ = (α0 , α, β, ψ, ν, μ) ∈ 0 and z̄ = (ξ̄ , x̄, ū) ∈ K, we have α0 J (p 0 )p̄ = 0, αi Fi (p 0 )p̄ = 0, μj (t)ϕj w (t, w (t))w̄(t) = 0, 0 i ∈ IF (p0 ), j = 1, . . . , d(ϕ), (3.260) (3.261) where αi and μj are the components of the vectors α and μ, respectively. Note that conditions (3.260) and (3.261) can be written in brief as α0 J (p 0 )p̄ = 0, αF (p0 )p̄ = 0, and μ(t)ϕw (t, w0 (t))w̄(t) = 0. Proposition 3.52 gives an answer to the question posed above. According to this proposition, for any λ ∈ 0 , conditions (3.256)– (3.261) also define K. It follows that if, for some λ = (α0 , α, β, ψ, ν, μ) ∈ 0 , the condition α0 > 0 holds, then, in the definition of K, the inequality J (p 0 )p̄ ≤ 0 can be replaced by the equality J (p0 )p̄ = 0. If, for some λ ∈ 0 and i0 ∈ {1, . . . , d(F )}, the condition αi0 > 0 holds, then the inequality Fi0 (p0 )p̄ ≤ 0 can be replaced by the equality Fi0 (p0 )p̄ = 0. 0 Finally, for any j ∈ {1, . . . , d(ϕ)} and λ ∈ 0 , the inequality ϕj w (t, w (t))w̄(t) ≤ 0 can be 0 replaced by the equality ϕj w (t, w (t))w̄(t) = 0 a.e. on the set {t ∈ [t0 , tf ] | μj (t) > 0} ⊂ M0 (ϕj0 ). Every such change gives an equivalent system of conditions still defining K. The following question is also of interest: Under what conditions can one of the endpoint inequalities in the definition of K be omitted without affecting K? In particular, when can the inequality J (p 0 )p̄ ≤ 0 be omitted? Proposition 3.53. Suppose that there exists λ ∈ 0 such that α0 > 0. Then the relations Fi (p0 )p̄ ≤ 0, i ∈ IF (p0 ), αi Fi (p0 )p̄ = 0, i ∈ IF (p0 ), K (p0 )p̄ = 0, (3.262) combined with (3.257)–(3.259) and (3.261) imply that J (p0 )p̄ = 0; i.e., K can be defined by conditions (3.257)–(3.259), (3.261), and (3.262)) as well. Therefore, if for some λ ∈ 0 all inequalities (3.256) corresponding to positive αi are replaced by the equalities and each inequality ϕj w (t, w0 (t))w̄(t) ≤ 0 is replaced by the equality on the set {t ∈ [t0 , tf ] | μj (t) > 0} ⊂ M0 (ϕj ), then, after all such changes, the equality Jp (p 0 )p̄ = 0 corresponding to positive α0 can be excluded, and the obtained new system of conditions still defines K. Quadratic form. Let us introduce the following notation. Given any λ = (α0 , α, β, ψ, ν, μ) ∈ 0 , we set [H̄xλ ]k = H̄xλk+ − H̄xλk− , k = 1, . . . , s, (3.263) where H̄xλk− and H̄xλk+ are defined by (3.239). Thus, [H̄xλ ]k denotes a jump of the function H̄x (t, x 0 (t), u0 (t), ψ(t), ν(t), μ(t)) at t = tk ∈ . It follows from the adjoint equation (3.230) that [H̄xλ ]k = −[ψ̇]k , k = 1, . . . , s, (3.264) 172 Chapter 3. Quadratic Conditions for Optimal Control Problems where the row vector [ψ̇]k = ψ̇ k+ − ψ̇ k− = ψ̇(tk + 0) − ψ̇(tk − 0) is the jump of the derivative ψ̇(t) at tk ∈ . Furthermore, for brevity we set λ (p0 ) = lpp ∂ 2l 0 ∂ 2 H̄ λ 0 (p , α , α, β), H̄ (w ) = (t, x 0 (t), u0 (t), ψ(t), ν(t), μ(t)). 0 ww ∂p 2 ∂w2 (3.265) Finally, for x̄ ∈ P W 1,2 ([t0 , tf ], Rd(x) ), we set 1 k = (x̄ k− + x̄ k+ ), x̄av 2 k = 1, . . . , s. (3.266) k is an average value of the function x̄ at t ∈ . Here x̄av k We are now ready to introduce the quadratic form, which takes into account the discontinuities of the control u0 . For any λ ∈ 0 and z̄ = (ξ̄ , x̄, ū) ∈ Z2 (), we set 1 k λ 2 k D (H̄ )ξ̄k + 2[H̄xλ ]k x̄av ξ̄k 2 k=1 1 λ 0 1 tf λ + lpp (p )p̄, p̄ + H̄ww (t, w0 )w̄, w̄ dt, 2 2 t0 s λ (z̄) = (3.267) where w̄ = (x̄, ū), p̄ = (x̄(t0 ), x̄(tf )). Recall that the value D k (H̄ λ ) is defined by (3.242), and it is nonpositive for any λ ∈ M0 . Obviously, λ is quadratic in z̄ and linear in λ. Set 1 k λ 2 k D (H̄ )ξ̄k + 2[H̄xλ ]k x̄av ξ̄k . 2 s λ ω (ξ̄ , x̄) = (3.268) k=1 This quadratic form is related to the discontinuities of control u0 , and we call it the internal form. According to (3.247) and (3.264) it can be written as follows: 1 2 s λ (ξ̄ , x̄) = ω k ξ̄k . ψ̇ k+ ẋ 0k− − ψ̇ k− ẋ 0k+ + [ψ̇t ]k ξ̄k2 − 2[ψ̇]k x̄av (3.269) k=1 Furthermore, we set 1 λ 0 1 (p )p̄, p̄ + ωλ (w̄) = lpp 2 2 tf t0 λ H̄ww (t, w 0 )w̄, w̄ dt. (3.270) This quadratic form is the second variation of the Lagrangian of problem (3.217)–(3.219) at the point w 0 . We call it the external form. Thus, the quadratic form λ (z̄) is a sum of the internal and external forms: λ (ξ̄ , x̄) + ωλ (w̄). λ (z̄) = ω (3.271) 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 173 Quadratic necessary condition for Pontryagin minimum. Now, we formulate the main necessary quadratic condition for Pontryagin minimum in problem (3.217)–(3.219) at the point w0 . Theorem 3.54. If w0 is a Pontryagin minimum, then the following Condition A holds: The set M0 is nonempty and max λ (z̄) ≥ 0 ∀ z̄ ∈ K. λ∈M0 The proof of this theorem was given in [86] and published in [95]. Next we formulate the basic sufficient condition in problem (3.217)–(3.219). We call it briefly Condition B(). It is sufficient not only for a Pontryagin minimum, but also for a bounded strong minimum defined below. We give a preliminary definition of a strong minimum which is slightly different from the commonly used definition. Strong minimum. A state variable xi (the ith component of x) is said to be unessential if the functions f , g, and ϕ do not depend on this variable and the functions J , F , and K are affine in pi = (xi (t0 ), xi (tf )). A state variable which does not possess this property is said to be essential. The vector comprised of the essential components xi of x is denoted by x. Similarly, δx denotes the vector consisting of the essential components of a variation δx. Denote by S the set of all sequences {δwn } in W such that |δxn (t0 )| + δx n C → 0. A minimum on S is called strong. Bounded strong minimum. A sequence {δwn } in W is said to be bounded strong on Q if {δwn } ∈ S and there exists a compact set C ⊂ Q such that for all sufficiently large n one has (t, x 0 (t), u0 (t) + δun (t)) ∈ C a.e. on [t0 , tf ]. By a (strict) bounded strong minimum we mean a (strict) minimum on the set of all bounded strong on Q sequences. Every strong minimum is bounded strong. Hence, it is a Pontryagin minimum. We know that the bounded strong minimum is equivalent to the strong minimum if there exists a compact set C ⊂ Q such that the conditions t ∈ [t0 , tf ] and u ∈ U(t, x 0 (t)) imply (t, x 0 (t), u) ∈ C, where U(t, x) is the set defined by (3.233). Let us state a quadratic sufficient condition for a point w0 = (x 0 , u0 ) to be a bounded strong minimum. Again we assume that u0 is a piecewise continuous control and = {t1 , . . . , ts } is the set of its discontinuity points, every element of being an L-point. Set M(C ). Let be an order function (see Definition 2.17). For any C > 0, we denote by M(C) the set of all λ ∈ M0 such that the following condition holds: H (t, x 0 (t), u, ψ(t)) − H (t, x 0 (t), u0 (t), ψ(t)) ≥ C(t, u), ∀ t ∈ [t0 , tf ] \ , u ∈ U(t, x 0 (t)). (3.272) Condition (3.272) strengthens the minimum condition (3.234), and we call (3.272) the minimum condition of strictness C (or C-growth condition for H ). For any C > 0, M(C) is a closed subset in M0 and, therefore, a finite dimensional compact set. 174 Chapter 3. Quadratic Conditions for Optimal Control Problems Basic sufficient condition. Let γ̄ (z̄) = ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) + tf ū(t), ū(t) dt. t0 On the subspace (3.257)–(3.258) of the space Z 2 (), the value γ̄ (z̄) is a norm, which is equivalent to the norm of the space Z 2 (). Let be an order function. Definition 3.55. We say that a point w0 satisfies condition B() if there exists C > 0 such that the set M(C) is nonempty and max λ (z̄) ≥ C γ̄ (z̄) λ∈M(C) ∀ z̄ ∈ K. Theorem 3.56. If there exists an order function (t, u) such that Condition B() holds, then w0 is a strict bounded strong minimum. Condition B() obviously holds if for some C > 0 the set M(C) is nonempty, and if the cone K consists only of zero. Therefore, Theorem 3.56 implies the following. Corollary 3.57. If for some C > 0 the set M(C) is nonempty, and if K = {0}, then w 0 is a strict bounded strong minimum. Corollary 3.57 states the first-order sufficient condition of a bounded strong minimum. γ -sufficiency. Quadratic Condition B() implies not only a bounded strong minimum, but also a certain strengthening of this concept which is called the γ -sufficiency on the set of bounded strong sequences. Below, we introduce the (higher) order γ and formulate two concepts: γ -sufficiency for Pontryagin minimum and γ -sufficiency for bounded strong minimum. Regarding the point w0 = (x 0 , u0 ), tested for optimality, we again use Assumption 3.41. Let (t, u) be an order function. Set γ (δw) = δx 2 C + tf (t, u0 + δu) dt. t0 The functional γ is defined on the set of all variations δw = (δx, δu) ∈ W such that (t, x 0 (t), u0 (t) + δu(t)) ∈ Q a.e. on [t0 , tf ]. The functional γ is the order associated with the order function (t, u). Following the general theory [55], we also call γ the higher order. Thus, with the point w0 we associate the family of order functions and the family of corresponding orders γ . Let us denote the latter family by Ord(w0 ). Let us introduce the violation function of problem (3.217)–(3.219) at the point w0 : σ (δw) = max σJFK (δp), σf (δw), σgϕ (δw) , (3.273) 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 175 where σJFK (δp) = max J (p 0 + δp) − J (p0 ), max Fi (p0 + δp), |K(p 0 + δp)| , i=1,...,d(F ) tf 0 0 σf (δw) = |ẋ + δ ẋ − f (t, w + δw)| dt, t0 σgϕ (δw) = ess sup[t0 ,tf ] max{|g(t, w0 + δw)|, max i=1,...,d(ϕ) ϕi (t, w0 + δw)}, δw = (δx, δu) ∈ W , δp = (δx(t0 ), δx(tf )), p0 + δp ∈ P , (t, w0 (t) + δw(t)) ∈ Q. Obviously, σ (δw) ≥ 0 and σ (0) = 0. Let S be a set of sequences {δwn } in W , closed with respect to the operation of taking subsequences. Evidently, w 0 is a point of a strict minimum on S iff, for any sequence {δwn } ∈ S containing nonzero terms, one has σ (δwn ) > 0 for all sufficiently large n. A strengthened version of the last condition is suggested by the following definition. Definition 3.58. We say that w0 is a point of γ -sufficiency on S if there exists an ε > 0 such that, for any sequence {δwn } ∈ S, we have σ (δwn ) ≥ εγ (δwn ) for all sufficiently large n. Let us now introduce a set of sequences related to a bounded strong minimum. Denote by Sbs the set of all sequences {δwn } which are bounded strong on Q and satisfy the following conditions: (a) (p0 + δpn ) ∈ P for all sufficiently large n, (b) σ (δwn ) → 0 as n → ∞. Conditions (a) and (b) hold for every sequence {δwn } that violates minimality, so a bounded strong minimum can be treated as a minimum on Sbs (the subscript bs means “bounded strong”). Theorem 3.59. Condition B() is equivalent to γ -sufficiency on Sbs . The proof of this theorem was given in [86] and published in [94, 95]. Theorem 3.59 particularly shows a nontrivial character of minimum guaranteed by condition B(). Let us explain this in more detail. A sequence {δwn } is said to be admissible if the sequence {w0 + δwn } satisfies all constraints of the canonical problem. We say that w0 is a point of γ -minimum on Sbs (or the γ -growth condition for the cost function holds on Sbs ) if there exists ε > 0 such that, for any admissible sequence {δwn } ∈ Sbs , we have J (p0 + δpn ) − J (p0 ) ≥ εγ (δwn ) for all sufficiently large n. Clearly, γ -sufficiency on Sbs implies γ minimum on Sbs . In fact, it is the sufficient Condition B() that ensures γ -minimum on Sbs . A nontrivial character of γ -minimum on Sbs is caused by a nontrivial definition of the order function which specifies the higher-order γ . Now, let us discuss an important question concerning characterization of Condition λ ∈ M(C). Local quadratic growth condition of the Hamiltonian. λ ∈ 0 . We set Fix an arbitrary tuple δH [t, v] := H (t, x 0 (t), u0 (t) + v, ψ(t)) − H (t, x 0 (t), u0 (t), ψ(t)). (3.274) 176 Chapter 3. Quadratic Conditions for Optimal Control Problems Definition 3.60. We say that, at the point w 0 , the Hamiltonian satisfies a local quadratic growth condition if there exist ε > 0 and α > 0 such that for all t ∈ [t0 , tf ] \ the following inequality holds: δH [t, v] ≥ α|v|2 if v ∈ Rd(u) , g(t, x 0 (t), u0 (t) + v) = 0, (3.275) 0 0 ϕ(t, x (t), u (t) + v) ≤ 0, |v| < ε. Recall the definition of H̄ in (3.225). Let us denote by H̄u (t) := H̄u (t, x 0 (t), u0 (t), ψ(t), ν(t), μ(t)), H̄uu (t) := H̄uu (t, x 0 (t), u0 (t), ψ(t), ν(t), μ(t)) the first and second derivative with respect to u of the augmented Hamiltonian, and adopt a similar notation for the Hamiltonian function H . Similarly, we denote gu (t) := gu (t, x 0 (t), u0 (t)), ϕi (t) := ϕi (t, x 0 (t), u0 (t)), ϕiu (t) := ϕiu (t, x 0 (t), u0 (t)), i = 1, . . . , d(ϕ). We shall formulate a generalization of the strengthened Legendre condition using the quadratic form H̄uu (t)v, v complemented by some special nonnegative term ρ(t, v) which will be homogeneous (not quadratic) of the second degree with respect to v. Let us define this additional term. For any number a, we set a + = max{a, 0} and a − = max{−a, 0}, so that a + ≥ 0, − a ≥ 0, and a = a + − a − . Denote by χi (t) := χ{ϕi (τ )<0} (t) (3.276) the characteristic function of the set {τ | ϕi (τ ) < 0}, i = 1, . . . , d(ϕ). If d(ϕ) > 1, then, for any t ∈ [t0 , tf ] and any v ∈ Rd(u) , we set ρ(t, v) = d(ϕ) j =1 max 1≤i≤d(ϕ) − + μj (t) . χi (t) ϕj u (t)v ϕiu (t)v |ϕi (t)| (3.277) Here, by definition, μj (t) χi (t) = 0 |ϕi (t)| if ϕi (t) = 0, i, j = 1, . . . , d(ϕ). Particularly, for d(ϕ) = 2 the function ρ has the form ρ(t, v) = − + μ1 (t) χ2 (t) ϕ1u (t)v ϕ2u (t)v |ϕ2 (t)| − + μ2 (t) χ1 (t) ϕ2u (t)v ϕ1u (t)v . + |ϕ1 (t)| (3.278) In the case d(ϕ) = 1, we set ρ(t, v) ≡ 0. For any > 0 and any t ∈ [t0 , tf ] \ , denote by Ct () the set of all vectors v ∈ Rd(u) satisfying gu (t)v = 0, ϕj u (t)v ≤ 0 if ϕj (t) = 0, (3.279) ϕj u (t)v = 0 if μj (t) > , j = 1, . . . , d(ϕ). 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 177 Definition 3.61. We say that the Hamiltonian satisfies the generalized strengthened Legendre condition if ∃ α > 0, > 0 such that ∀ t ∈ [tf , tf ] \ : (3.280) 1 2 ∀ v ∈ C (). t 2 H̄uu (t)v, v + ρ(t, v) ≥ α|v| Theorem 3.62. A local quadratic growth condition for the Hamiltonian is equivalent to the generalized strengthened Legendre condition. This theorem was proved in [8] for the control constrained problem (without mixed constraints). We note that Ct () is in general a larger set than the local cone Ct of critical directions for the Hamiltonian, i.e., the directions v ∈ Rd(u) , such that gu (t)v = 0, ϕj u (t)v ≤ 0 if ϕj (t) = 0, ϕj u (t)v = 0 if μj (t) > 0, j = 1, . . . , d(ϕ). (3.281) A simple sufficient condition for local quadratic growth of the Hamiltonian. Consider the following second-order condition for the Hamiltonian: ∃ α > 0, > 0 such that ∀ t ∈ [0, T ]: (3.282) 1 2 ∀ v ∈ C (). t 2 H̄uu (t)v, v ≥ α|v| Let us note that this inequality is stronger than (3.280), since the function ρ(t, v) is nonnegative. Theorem 3.63. Condition (3.282) implies a local quadratic growth of the Hamiltonian. Characterization of condition λ ∈ M(C ). An element λ ∈ 0 is said to be strictly Legendrian if, for this element, the generalized strengthened Legendre condition (3.280) is satisfied and also the following conditions hold: [H λ ]k = 0, D k (H̄ λ ) > 0 ∀ tk ∈ . (3.283) Denote by M0+ the set of λ ∈ M0 such that the following conditions hold: (a) H (t, x 0 (t), u, ψ(t)) > H (t, x 0 (t), u0 (t), ψ(t)) if t ∈ [t0 , tf ]\, u ∈ U(t, x 0 (t)), u = u0 (t), where U(t, x) = {u ∈ Rd(u) | (t, x, u) ∈ Q, g(t, x, u) = 0, ϕ(t, x, u) ≤ 0}; (b) H (tk , x 0 (tk ), u, ψ(tk )) > H k if tk ∈ , u ∈ U(tk , x 0 (tk )), u ∈ / {u0k− , u0k+ }, where k 0 0k− H := H (tk , x (tk ), u , ψ(tk )) = H (tk , x 0 (tk ), u0k+ , ψ(tk )). Denote by Leg+ (M0+ ) the set of all strictly Legendrian elements λ ∈ M0+ . Theorem 3.64. An element λ ∈ Leg+ (M0+ ) iff there exists C > 0 such that λ ∈ M(C). The proof will be published elsewhere. 178 3.4.2 Chapter 3. Quadratic Conditions for Optimal Control Problems General Optimal Control Problem on a Variable Time Interval Statement of the problem. Here, quadratic optimality conditions, both necessary and sufficient, are presented in the following optimal control problem on a variable time interval. Let T denote a trajectory (x(t), u(t) | t ∈ [t0 , tf ]), where the state variable x(·) is a Lipschitz continuous function, and the control variable u(·) is a bounded measurable function on a time interval = [t0 , tf ]. The interval is not fixed. For each trajectory T , we denote by p = (t0 , x(t0 ), tf , x(tf )) the vector of the endpoints of time-state variable (t, x). It is required to find T minimizing the functional J(T ) := J (t0 , x(t0 ), tf , x(tf )) → min (3.284) F (t0 , x(t0 ), tf , x(tf )) ≤ 0, K(t0 , x(t0 ), tf , x(tf )) = 0, ẋ(t) = f (t, x(t), u(t)), g(t, x(t), u(t)) = 0, ϕ(t, x(t), u(t)) ≤ 0, p ∈ P , (t, x(t), u(t)) ∈ Q, (3.285) (3.286) (3.287) (3.288) subject to the constraints where P and Q are open sets, and x, u, F , K, f , g, and ϕ are vector functions. We assume that the functions J , F , and K are defined and twice continuously differentiable on P , and the functions f , g, and ϕ are defined and twice continuously differentiable on Q; moreover, g and ϕ satisfy the linear independence assumption (see Assumption 3.40). Necessary conditions for a Pontryagin minimum. Let T be a fixed admissible trajectory such that the control u(·) is a piecewise Lipschitz continuous function on the interval with the set of discontinuity points = {t1 , . . . , ts }, t0 < t1 < · · · < ts < tf . Let us formulate a first-order necessary condition for optimality of the trajectory T . We introduce the Pontryagin function H (t, x, u, ψ) = ψf (t, x, u) (3.289) and the augmented Pontryagin function H̄ (t, x, u, ψ, ν, μ) = H (t, x, u, ψ) + νg(t, x, u) + μϕ(t, x, u), (3.290) where ψ, ν, and μ are row vectors of the dimensions d(x), d(g), and d(ϕ), respectively. Let us define the endpoint Lagrange function l(p, α0 , α, β) = α0 J (p) + αF (p) + βK(p), (3.291) where p = (t0 , x0 , tf , xf ), x0 = x(t0 ), xf = x(tf ), α0 ∈ R, α ∈ (Rd(F ) )∗ , β ∈ (Rd(K) )∗ . Also we introduce a tuple of Lagrange multipliers λ = (α0 , α, β, ψ(·), ψ0 (·), ν(·), μ(·)) (3.292) such that ψ(·) : → (Rd(x) )∗ and ψ0 (·) : → R1 are piecewise smooth functions, continuously differentiable on each interval of the set \ , and ν(·) : → (Rd(g) )∗ and 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 179 μ(·) : → (Rd(ϕ) )∗ are piecewise continuous functions, Lipschitz continuous on each interval of the set \ . Denote by M0 the set of the normed tuples λ satisfying the conditions of the minimum principle for the trajectory T : α0 ≥ 0, α ≥ 0, αF (p) = 0, α0 + d(F ) αi + i=1 d(K) |βj | = 1, j =1 ψ̇ = −H̄x , ψ̇0 = −H̄t , H̄u = 0, t ∈ \ , ψ(t0 ) = −lx0 , ψ(tf ) = lxf , ψ0 (t0 ) = −lt0 , ψ0 (tf ) = ltf , min H (t, x(t), u, ψ(t)) = H (t, x(t), u(t), ψ(t)), t ∈ \ , u∈U(t,x(t)) H (t, x(t), u(t), ψ(t)) + ψ0 (t) = 0, (3.293) t ∈ \ , where U(t, x) = {u ∈ Rd(u) | g(t, x, u) = 0, ϕ(t, x, u) ≤ 0, (t, x, u) ∈ Q}. The derivatives lx0 and lxf are at (p, α0 , α, β), where p = (t0 , x(t0 ), tf , x(tf )), and the derivatives H̄x , H̄u , and H̄t are at (t, x(t), u(t), ψ(t), ν(t), μ(t)), where t ∈ \ . (Condition H̄u = 0 follows from the others conditions in this definition, and therefore could be excluded; yet we need to use it later.) We define the Pontryagin minimum in problem (3.284)–(3.288) on a variable interval [t0 , tf ] as in Section 3.3.2 (see Definition 3.25). The condition M0 = ∅ is equivalent to the Pontryagin’s minimum principle. It is a first-order necessary condition of Pontryagin minimum for the trajectory T . Thus, the following theorem holds (see, e.g., [76]). Theorem 3.65. If the trajectory T affords a Pontryagin minimum, then the set M0 is nonempty. Assume that M0 is nonempty. Using the definition of the set M0 and the linear independence assumption for g and ϕ one can easily prove the following statement. Proposition 3.66. The set M0 is a finite-dimensional compact set, and the mapping λ → (α0 , α, β) is injective on M0 . As in Section 3.3, for each λ ∈ M0 , tk ∈ , we set D k (H̄ ) = −H̄xk+ H̄ψk− + H̄xk− H̄ψk+ − [H̄t ]k , (3.294) where H̄xk− = H̄x (tk , x(tk ), u(tk −), ψ(tk ), ν(tk −), μ(tk −)), H̄xk+ = H̄x (tk , x(tk ), u(tk +), ψ(tk ), ν(tk +), μ(tk +)), [H̄t ]k = H̄tk+ − H̄tk− , etc. Theorem 3.67. For each λ ∈ M0 , the following conditions hold: D k (H̄ ) ≥ 0, k = 1, . . . , s. (3.295) Thus, conditions (3.295) follow from the minimum principle conditions (3.293). Let us formulate a quadratic necessary condition of a Pontryagin minimum for the trajectory T . First, for this trajectory, we introduce a Hilbert space Z2 () and the critical cone K ⊂ Z2 (). We denote by P W 1,2 (, Rd(x) ) the Hilbert space of piecewise continuous functions x̄(·) : → Rd(x) , absolutely continuous on each interval of the set \ 180 Chapter 3. Quadratic Conditions for Optimal Control Problems and such that their first derivative is square integrable. For each x̄ ∈ P W 1,2 (, Rd(x) ), tk ∈ , we set x̄ k− = x̄(tk −), x̄ k+ = x̄(tk +), [x̄]k = x̄ k+ − x̄ k− . Further, we let z̄ = (t¯0 , t¯f , ξ̄ , x̄, ū), where t¯0 ∈ R1 , Thus, t¯f ∈ R1 , ξ̄ ∈ Rs , x̄ ∈ P W 1,2 (, Rd(x) ), ū ∈ L2 (, Rd(u) ). z̄ ∈ Z2 () := R2 × Rs × P W 1,2 (, Rd(x) ) × L2 (, Rd(u) ). Moreover, for given z̄, we set w̄ = (x̄, ū), x̄0 = x̄(t0 ), x̄f = x̄(tf ), x̄¯0 = x̄(t0 ) + t¯0 ẋ(t0 ), x̄¯f = x̄(tf ) + t¯f ẋ(tf ), (3.296) p̄¯ = (x̄¯0 , t¯0 , x̄¯f , t¯f ). (3.297) By IF (p) = {i ∈ {1, . . . , d(F )} | Fi (p) = 0} we denote the set of active indices of the constraints Fi (p) ≤ 0. Let K be the set of all z̄ ∈ Z2 () satisfying the following conditions: J (p)p̄¯ ≤ 0, Fi (p)p̄¯ ≤ 0 ∀ i ∈ IF (p), K (p)p̄¯ = 0, ˙ = fw (t, w(t))w̄(t) for a.a. t ∈ [t0 , tf ], x̄(t) [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, gw (t, w(t))w̄(t) = 0 for a.a. t ∈ [t0 , tf ], ϕj w (t, w0 (t))w̄(t) ≤ 0 a.e. on M0 (ϕj0 ), (3.298) j = 1, . . . , d(ϕ), where M0 (ϕj0 ) = {t ∈ [t0 , tf ] | ϕj (t, w 0 (t)) = 0}, p = (t0 , x(t0 ), tf , x(tf )), w = (x, u). It is obvious that K is a convex cone in the Hilbert space Z2 (), and we call it the critical cone. If the interval is fixed, then we set p := (x0 , xf ) = (x(t0 ), x(tf )), and in the definition of K we have t¯0 = t¯f = 0, x̄¯0 = x̄0 , x̄¯f = x̄f , and p̄¯ = p̄ := (x̄0 , x̄f ). Define quadratic forms ωe , ω, and by formulas (3.156), (3.157), and (3.158), respectively. Now, we formulate the main necessary quadratic condition of a Pontryagin minimum in the problem on a variable time interval. Theorem 3.68. If the trajectory T yields a Pontryagin minimum, then the following Condition A holds: The set M0 is nonempty and max (λ, z̄) ≥ 0 ∀ z̄ ∈ K. λ∈M0 Sufficient conditions for a bounded strong minimum. The ith component xi of vector x is called unessential if the functions f , g, and ϕ do not depend on this component and the functions J , F , and K are affine in xi0 = xi (t0 ), xif = xi (tf ); otherwise the component xi is called essential. We denote by x a vector composed of all essential components of vector x and we define the (strict) bounded strong minimum as in Definition 3.32. Let be an order function (see Definition 2.17). We formulate a sufficient optimality Condition B(), which is a natural strengthening of the necessary Condition A. Let us introduce the functional tf γ̄ (z̄) = t¯02 + t¯f2 + ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) + ū(t), ū(t) dt, (3.299) t0 3.4. Quadratic Conditions for Mixed Control-State Constrained Problems 181 which is equivalent to the norm squared on the subspace x̄˙ = fw (t, x(t), u(t))w̄; [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, (3.300) of Hilbert space Z2 (). Recall that the critical cone K is contained in the subspace (3.300). For any C > 0, we denote by M(C) the set of all λ ∈ M0 such that condition (3.272) holds. Theorem 3.69. For the trajectory T , assume that the following Condition B() holds: There exists C > 0 such that the set M(C) is nonempty and max (λ, z̄) ≥ C γ̄ (z̄) λ∈M(C) (3.301) for all z̄ ∈ K. Then the trajectory T affords a strict bounded strong minimum. Again we can use the characterization of the condition λ ∈ M(C) formulated in the previous section. Chapter 4 Jacobi-Type Conditions and Riccati Equation for Broken Extremals Here we derive tests for the positive semidefiniteness, respectively, positive definiteness of the quadratic form on the critical cone K (introduced in Chapter 2 for extremals with jumps of the control). In Section 4.1, we derive such tests for the simplest problem of the calculus of variations and for an extremal with only one corner point. We come to a generalization of the concept of conjugate point which allows us to formulate both necessary and sufficient second-order optimality conditions for broken extremals. Three numerical examples illustrate this generalization. Further, we concentrate on sufficient conditions for positive definiteness of the quadratic form in the auxiliary problem. We show that if there exists a solution to the Riccati matrix equation satisfying a certain jump condition, then the quadratic form can be transformed into a perfect square. This gives a possibility of proving a sufficient condition for positive definiteness of the quadratic form in the auxiliary problem and thus to obtain one more sufficient condition for optimality of broken extremals. At the end of Section 4.1, we obtain such condition for the simplest problem of the calculus of variations, and then, in Section 4.2, we prove it for the general problem (without constraint g(t, x, u) = 0). 4.1 Jacobi-Type Conditions and Riccati Equation for Broken Extremals in the Simplest Problem of the Calculus of Variations 4.1.1 An Auxiliary Problem for Broken Extremal Let a closed interval [t0 , tf ], two points b0 , bf ∈ Rm , an open set Q ⊂ R2m+1 , and a function F : Q → R of class C 2 be fixed. The simplest problem of the calculus of variations can be formulated as follows: tf (SP) Minimize J (x(·), u(·)) := F (t, x(t), u(t)) dt (4.1) t0 under the constraints ẋ(t) = u(t), x(t0 ) = b0 , (t, x(t), u(t)) ∈ Q. 183 x(tf ) = bf , (4.2) (4.3) 184 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Here, x(·) : [t0 , tf ] → Rm is absolutely continuous, and u(·) : [t0 , tf ] → Rm is bounded and measurable. Set w(·) = (x(·), u(·)). Then w(·) is an element of the space W := W 1,1 ([t0 , tf ], Rm ) × L∞ ([t0 , tf ], Rm ). We say that w(·) = (x(·), u(·)) is an admissible pair if w(·) ∈ W and the constraints (4.2), (4.3) hold for it. Let an admissible pair w(·) = (x(·), u(·)) be given. Assume that u(·) is a piecewise continuous function with a unique point of discontinuity t∗ ∈ (t0 , tf ). Denote by the singleton {t∗ }. For t∗ , we set u− = u(t∗ −), u+ = u(t∗ +), and [u] = u+ − u− . Thus, [u] is a jump of the function u(·) at the point t∗ . In correspondence with (4.3), assume that (t, x(t), u(t)) ∈ Q for all t ∈ [t0 , tf ] \ , and (t∗ , x(t∗ ), u− ) ∈ Q, (t∗ , x(t∗ ), u+ ) ∈ Q. Also assume that there exist a constant C > 0 and a small number ε > 0 such that |u(t) − u− | |u(t) − u+ | ≤ ≤ C|t − t∗ | C|t − t∗ | ∀ t ∈ (t∗ − ε, t∗ ) ∩ [t0 , tf ], ∀ t ∈ (t∗ , t∗ + ε) ∩ [t0 , tf ]. The pair w(·) = (x(·), u(·)) is called an extremal if ψ(t) := −Fu (t, x(t), u(t)) (4.4) is a Lipschitz-continuous function and a condition equivalent to the Euler equation is satisfied: −ψ̇(t) = Fx (t, x(t), u(t)) ∀ t ∈ [t0 , tf ]\. (4.5) Here ψ ∈ (Rm )∗ is a row vector while x, u ∈ R m are column vectors. For the Pontryagin function H (t, x, u, ψ) = ψu + F (t, x, u), (4.6) we set H − = H (t∗ , x(t∗ ), u− , ψ(t∗ )), H + = H (t∗ , x(t∗ ), u+ , ψ(t∗ )), and [H ] = H + − H − . The equalities [H ] = 0, [ψ] = 0 (4.7) are the Weierstrass–Erdmann conditions. They are known as necessary conditions for the strong minimum. However, they are also necessary for the Pontryagin minimum introduced in Chapter 2. For convenience, let us recall the definitions of Pontryagin and bounded strong minima in the simplest problem. We say that the pair of functions w = (x, u) is a point of Pontryagin minimum in problem (4.1)–(4.3) if for each compact set C ⊂ Q there exists ε > 0 such that J (w̃) ≥ J (w) for all admissible pairs w̃ = (x̃, ũ) such that (a) max[t0 ,tf ] |x̃(t) − x(t)| < ε, t (b) t0f |ũ(t) − u(t)| dt < ε, (c) (t, x̃(t), ũ(t)) ∈ C. 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 185 We say that the pair w = (x, u) is a point of bounded strong minimum in problem (4.1)–(4.3), if for each compact set C ⊂ Q there exists ε > 0 such that J (w̃) ≥ J (w) for all admissible pairs w̃ = (x̃, ũ), such that (a) max[t0 ,tf ] |x̃(t) − x(t)| < ε, (b) (t, x̃(t), ũ(t)) ∈ C. Clearly, the following implications hold: strong minimum =⇒ bounded strong minimum =⇒ Pontryagin minimum =⇒ weak minimum. As already mentioned, the Weierstrass–Erdmann conditions are necessary for the Pontryagin minimum. As shown in Chapter 2, they can be supplemented by additional condition of the same type. We set a = D(H ) := ψ̇ + ẋ − − ψ̇ − ẋ + − [Ft ], (4.8) where ψ̇ − = ψ̇(t∗ −), ψ̇ + = ψ̇(t∗ +), ẋ − = ẋ(t∗ −), ẋ + = ẋ(t∗ +), and [Ft ] = Ft (t∗ , x(t∗ ), u+ )− Ft (t∗ , x(t∗ ), u− ). Then a ≥ 0 is a necessary condition for the Pontryagin minimum. As we know, the value D(H ) can be computed in a different way. Consider the function (H )(t) := H (t, x(t), u+ , ψ(t)) − H (t, x(t), u− , ψ(t)) (4.9) = ψ(t)[ẋ] + F (t, x(t), u+ ) − F (t, x(t), u− ) . Using (4.5), we obtain d (H )|t∗ +0 = ψ̇ + [ẋ] − [ψ̇]ẋ + + [Ft ], dt d (H )|t∗ −0 = ψ̇ − [ẋ] − [ψ̇]ẋ − + [Ft ]. dt (4.10) (4.11) Since ψ̇ + [ẋ] − [ψ̇]ẋ + = ψ̇ − ẋ + − ψ̇ + ẋ − = ψ̇ − [ẋ] − [ψ̇]ẋ − , we obtain from this that d d (H )|t∗ −0 = (H )|t∗ +0 = −D(H ). dt dt (4.12) Note that the inequality a ≥ 0 and the Weierstrass–Erdmann conditions are implied by the conditions of the minimum principle, which is equivalent to the Weierstrass condition in this problem. Here, the minimum principle has the form: H (t, x(t), u, ψ(t)) ≥ H (t, x(t), u(t), ψ(t)) if t ∈ [t0 , tf ] \ , u ∈ Rm , (t, x(t), u) ∈ Q. Let us also formulate the strict minimum principle: (a) H (t, x(t), u, ψ(t)) > H (t, x(t), u(t), ψ(t)) for all t ∈ [t0 , tf ] \ , u ∈ Rm , (t, x(t), u) ∈ Q, u = u(t), (b) H (t∗ , x(t∗ ), u, ψ(t∗ )) > H (t∗ , x(t∗ ), u− , ψ(t∗ )) = H (t∗ , x(t∗ ), u+ , ψ(t∗ )) for all u ∈ Rm , (t, x(t), u) ∈ Q, u = u− , u = u+ . Now we define a quadratic form that corresponds to an extremal w(·) with a cor ner point. As in Section 2.1.5, denote by P W 1,2 [t0 , tf ], Rm the space of all piecewise continuous functions x̄(·) : [t0 , tf ] −→ Rm that are absolutely continuous on each of the intervals in [t0 , tf ]\ whose derivatives are square Lebesgue integrable. For t∗ , we set 186 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals x̄ − = x̄(t∗ −), x̄ + = x̄(t∗ +), [x̄] = x̄ + − x̄ − . Recall that the space P W 1,2 ([t0 , tf ], Rm ) t ˙ ˙ ȳ(t) dt is a Hilbert space. with the inner product (x̄, ȳ) = x̄(0), ȳ(0) + [x̄], [ȳ] + t0f x̄(t), We set Z2 () = R1 × P W 1,2 [t0 , tf ], Rm × L2 [t0 , tf ], Rm and denote by z̄ = (ξ̄ , x̄, ū) an element of the space Z2 (), where ξ̄ ∈ R1 , x̄(·) ∈ P W 1,2 [t0 , tf ], Rm , ū(·) ∈ L2 [t0 , tf ], Rm . In the Hilbert space Z2 (), we define a subspace and a quadratic form by setting ˙ = ū(t), K = z̄ ∈ Z2 () | x̄(t) and x̄(t0 ) = x̄(tf ) = 0, 1 1 (z̄) = a ξ̄ 2 − 2[ψ̇]x̄av ξ̄ + 2 2 tf [x̄] = [ẋ]ξ̄ Fww w̄(t), w̄(t) dt, (4.13) (4.14) t0 respectively, where 1 − x̄ + x̄ + , Fww = Fww (t, x(t), u(t)), 2 Fxx Fxu Fww = , w̄(·) = (x̄(·), ū(·)). Fux Fuu x̄av = w = (x, u), The condition of positive semidefiniteness of on K implies the following Legendre condition: (L) for any v ∈ Rm , ∀ t ∈ [t0 , tf ]\, Fuu (t, x(t), u(t))v, v ≥ 0 − + Fuu v, v ≥ 0, Fuu v, v ≥ 0. (4.15) (4.16) − = F (t , x(t ), u− ), F + = which is a necessary condition for a weak minimum. Here Fuu uu ∗ ∗ uu + Fuu (t∗ , x(t∗ ), u ). The condition of positive definiteness of on K implies the strengthened Legendre condition: (SL) for any v ∈ Rm \{0}, Fuu (t, x(t), u(t))v, v > 0 ∀ t ∈ [t0 , tf ]\, − + v, v > 0, Fuu v, v > 0. Fuu (4.17) (4.18) The auxiliary minimization problem for an extremal w(·) with a corner point is formulated as follows: (AP) minimize (z̄) under the constraint z̄ ∈ K. This setting is stipulated by the following two theorems. Theorem 4.1. If w(·) is a Pontryagin minimum, then the following conditions hold: (a) the Euler equation, (b) the minimum principle (the Weierstrass condition), 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 187 (c) the Legendre condition (L), (d) a ≥ 0, (e) (z̄) ≥ 0 for all z̄ ∈ K. Theorem 4.2. If the following conditions hold, then w(·) is a point of a strict bounded strong minimum: (a) the Euler equation, (b) the strict minimum principle (the strict Weierstrass condition), (c) the strengthened Legendre condition (SL), (d) a > 0, t (e) there exists ε > 0 such that (z̄) ≥ ε(ξ̄ 2 + t0f ū(t), ū(t) dt) for all z̄ ∈ K. Theorems 4.1 and 4.2 follow from Theorems 2.4 and 2.102, respectively. Note that the functional tf γ̄ (z̄) = ξ̄ 2 + ū(t), ū(t) dt (4.19) t0 on the subspace K is equivalent to the squared norm: γ̄ ∼ (z̄, z̄). Now our goal is to derive the tests for positive semidefiniteness and positive definiteness of the quadratic form on the subspace K in the case of an extremal with a single corner point. We give such tests in the form of the Jacobi-type conditions and in terms of solutions to the Riccati equations. 4.1.2 Jacobi-Type Conditions and the Riccati Equation In this section, we assume that for an extremal w(·) = (x(·), u(·)) with a single corner point t∗ ∈ (t0 , tf ), the condition (SL) holds, and a > 0, where a is defined by (4.8). It follows from condition (SL) that at each point t ∈ [t0 , tf ] \ , the matrix Fuu = Fuu (t, w(t)) has the −1 . We set inverse matrix Fuu A C −1 F , −1 , = −Fuu B = −Fuu ux −1 ∗ −1 . = −Fxu Fuu Fux + Fxx , A = −Fxu Fuu (4.20) All derivatives are computed along the trajectory (t, w(t)). Note that B ∗ = B, C ∗ = C, |det B| ≥ const > 0 on [t0 , tf ], (4.21) and A, B, and C are matrices with piecewise continuous entries on [t0 , tf ] that are continuous on each of the intervals of the set (t0 , tf ) \ . Further, we formulate Jacobi-type conditions for an extremal w(·) with a single corner point t∗ . Denote by X(t) and "(t) two square matrices of order m, where t ∈ [t0 , tf ]. For X(t) and "(t), we consider the set of differential equations Ẋ ˙ −" with the initial conditions = = X(t0 ) = O, AX + B", CX + A∗ " "(t0 ) = −I , where O and I are the zero matrix and the identity matrix, respectively. (4.22) (4.23) 188 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Recall that a continuous (and hence a piecewise smooth) solution X(t), "(t) to the Cauchy problem (4.22), (4.23) allows one to formulate the classical concept of the conjugate point. Namely, a point τ ∈ (t0 , tf ] is called conjugate (to the point t0 ) if det X(τ ) = 0. The absence of a conjugate point in (t0 , tf ) is equivalent to the positive semidefiniteness of the quadratic form ω= tf Fww w̄, w̄ dt t0 on the subspace K0 consisting of pairs w̄ = (x̄, ū) such that x̄˙ = ū, x̄(t0 ) = x̄(tf ) = 0, x̄ ∈ W 1,2 ([t0 , tf ], Rm ), ū ∈ L2 ([t0 , tf ], Rm ). The latter condition is necessary for the weak minimum. The absence of a conjugate point in (t0 , tf ] is equivalent to the condition of positive definiteness of ω on K0 , which is a sufficient condition for the strict weak minimum. This is the classical Jacobi condition. We note that and K pass into ω and K0 , respectively, if we set ξ̄ = 0 in the definition of the first pair. In our tests of positive semidefiniteness and positive definiteness of the quadratic form on the subspace K, we use a discontinuous solution X, " to the Cauchy problem (4.22), (4.23) with certain jump conditions at the point t∗ . Namely, let a pair X(t), "(t) be a continuous solution to the problem (4.22), (4.23) on the half-interval [t0 , t∗ ). We set X− = X(t∗ −), and " − = "(t∗ −). The jumps [X] and ["] of the matrixvalued functions X(t) and "(t) at the point t∗ are uniquely defined by using the relations a[X] = [ẋ] −[ẋ]∗ " − + [ψ̇]X− , (4.24) ∗ ∗ − − a["] = [ψ̇] −[ẋ] " + [ψ̇]X , (4.25) where [ẋ] and [ψ̇]∗ are column matrices, while [ẋ]∗ and [ψ̇] are row matrices. Let us define the right limits X + and " + of the functions X and " at t∗ in the following way: X+ = X − + [X], " + = " − + ["]. Then we continue the process of solution of system (4.22) on (t∗ , T ] by using the initial conditions for X and " at t∗ given by the conditions X(t∗ +) = X + , "(t∗ +) = " + . Thus, on [t0 , tf ], we obtain a piecewise continuous solution X(t), "(t) to system (4.22) with the initial conditions (4.23) at t0 and the jump conditions (4.24) and (4.25) at t∗ . On each of the intervals of the set [t0 , tf ] \ , the matrix-valued functions X(t) and "(t) are smooth. Briefly, this pair of functions will be called a solution to the problem (4.22)–(4.25) on [t0 , tf ]. Theorem 4.3. The form is positive semidefinite on K iff the solution X(t), "(t) to the problem (4.22)–(4.25) on [t0 , tf ] satisfies the conditions det X(t) = 0 ∀ t ∈ (t0 , tf ) \ , det X − = 0, det X + = 0, a − [ẋ]∗ Q− − [ψ̇] [ẋ] > 0, where Q(t) = "(t)X−1 (t), Q− = Q(t∗ −), and X −1 (t) is the inverse matrix to X(t). (4.26) (4.27) (4.28) 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 189 Theorem 4.4. The form is positive definite on K iff the solution X(t), "(t) to the problem (4.22)–(4.25) on [t0 , tf ] satisfies conditions (4.26)–(4.28), together with the additional condition det X(tf ) = 0. (4.29) The conditions for positive semidefiniteness, and those for positive definiteness of on K, which are given by these two theorems can easily be reformulated in terms of a solution to the corresponding matrix Riccati equation. Indeed, if X(t), "(t) is a solution to system (4.22) on a certain interval ⊂ [t0 , tf ] with det X(t) = 0 on , then, as is well known, the matrix-valued function Q(t) = "(t)X −1 (t) satisfies the Riccati equation Q̇ + QA + A∗ Q + QBQ + C = 0 (4.30) on . Let us prove this assertion. Differentiating the equality " = QX and using (4.22), we obtain ˙ = Q̇X + QẊ = Q̇X + Q(AX + B"). −CX − A∗ " = " Consequently, CX + A∗ " + Q̇X + QAX + QB" = 0. Multiplying this equation by X−1 from the right, we obtain (4.30). Using (4.20), we can also represent (4.30) as −1 (Q + Fux ) + Fxx = 0. Q̇ − (Q + Fxu )Fuu (4.31) The solution Q = "X −1 has a singularity at the zero point, since X(t0 ) = O. The question is: How do we correctly assign the initial condition for Q? We can do this in the following way. In a small half-neighborhood [t0 , t0 + ε), ε > 0, we find a solution to the Riccati equation for R = Q−1 = X" −1 with the initial condition R(t0 ) = O, (4.32) which is implied by (4.23). This Riccati equation for R can easily be obtained. Namely, differentiating the equality X = R" (4.33) and using (4.22), we obtain, for small ε > 0, Ṙ = AR + RA∗ + RCR + B, t ∈ [t0 , t0 + ε]. (4.34) Using (4.20), we can transform this Riccati equation into the form −1 (Fux R + I ) − RF xx R = 0, Ṙ + (RF xu + I )Fuu t ∈ [t0 , t0 + ε]. (4.35) Thus, we solve the Riccati equation (4.34) or (4.35) with initial condition (4.32) in a certain half-neighborhood [t0 , t0 +ε) of t0 . Recall that the matrices B and C are symmetric on [t0 , tf ]. Consequently, R is also symmetric, and therefore, Q(t0 + ε) = R −1 (t0 + ε) is symmetric. (4.36) 190 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Let Q(t) be the continuous solution of the Riccati equation (4.30) with initial condition (4.36) on the interval (t0 , t∗ ). The existence of such a solution is a necessary condition for positive semidefiniteness of on K. Since B(t) and C(t) are symmetric on [t0 , tf ] and Q(ε) is also symmetric, we have that Q(t) is symmetric on (t0 , t∗ ). Consequently, Q− = Q(t∗ −) is symmetric. Further, we define a jump condition for Q at t∗ that corresponds to the jump conditions (4.24) and (4.25). This condition has the form a − (q− )[ẋ] [Q] = (q− )∗ (q− ), (4.37) where q− = [ẋ]∗ Q− − [ψ̇]. [ẋ]∗ (4.38) )∗ Note that q− and are row vectors while (q− and [ẋ] are column vectors. The jump [Q] of the matrix Q at the point t∗ is uniquely defined by using (4.37) since, according to Theorem 4.3, the condition a − (q− )[ẋ] > 0 is necessary for positive semidefiniteness of on K. Note that (q− )∗ (q− ) is a symmetric positively semidefinite matrix. Hence, [Q] is a symmetric negatively semidefinite matrix. The right limit Q+ = Q(t∗ +) is defined by the relation Q+ = Q− + [Q]. (4.39) It follows that Q+ is symmetric. Using Q+ as the initial condition for Q at the point t∗ , we continue the solution of the Riccati equation (4.30) for t > t∗ . The matrix Q(t) is also symmetric for t > t∗ . Assume that this symmetric solution Q(t) is extended to a certain interval (t0 , τ ) or half-interval (t0 , τ ], where t∗ < τ ≤ tf . It will be called a solution to problem (4.30), (4.36), (4.37) on (t0 , τ ) or on (t0 , τ ], respectively. Theorem 4.5. The form is positive semidefinite on K iff there exists a solution Q to the problem (4.30), (4.36), (4.37) on (t0 , tf ) that satisfies a − (q− )[ẋ] > 0, (4.40) where q− is defined by (4.38). Theorem 4.6. The form is positive definite on K iff there exists a solution Q to the problem (4.30), (4.36), (4.37) on (t0 , tf ] that satisfies inequality (4.40). We note that the inequality (4.40) is equivalent to condition (4.28). Moreover, set b− = a − (q− )[ẋ]. Then the inequality (4.40) and the jump condition (4.37) obtain the form b− > 0, (b− )[Q] = (q− )∗ (q− ), respectively. 4.1.3 Passage of the Quadratic Form through Zero Now our goal consists of obtaining the tests for positive semidefiniteness and positive definiteness of on K, which were stated in Section 4.1.2. In the present section we shall 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 191 use some ideas from [25, 26]. Everywhere below, we assume that condition (SL) holds. It follows from condition (SL) that is a Legendre form on K (cf., e.g., [42]) in the abstract sense; i.e., it is a weakly lower semicontinuos functional on K, and the conditions z̄n ∈ K ∀ n, z̄ ∈ K, ξ̄ n → ξ̄ , ūn → ū weakly in L2 , (z̄n ) → (z̄) imply ūn − ū 2 (4.41) (4.42) (4.43) → 0, (4.44) tf t0 v(t), v(t) and hence → z̄ strongly in Z2 (). Here v 2 = ( is the norm in L2 . Further, consider a monotonically increasing one-parameter family of subspaces K(τ ) in K, each of which is defined by the relation z̄n dt)1/2 K(τ ) = z̄ = (ξ̄ , x̄, ū) ∈ K | ū(t) = 0 in [τ , tf ] , (4.45) where τ ∈ [t0 , tf ]. It is clear that K(t0 ) = {0}, K(tf ) = K. (4.46) We now study the problem of how the property of positive definiteness of on K(τ ) depends on the change of the parameter τ . This will allow us to obtain Jacobi-type conditions for on K. The form is positive on K(τ ) if (z̄) > 0 for all z̄ ∈ K(τ )\{0}. As is well known, for Legendrian forms, the positivity of on K(τ ) is equivalent to its positive definiteness on K(τ ), i.e., to the following property: There exists ε > 0 such that (z̄) ≥ ε γ̄ (z̄) for all z̄ ∈ K(τ ), where γ̄ is defined by relation (4.19). Obviously, for τ = t0 the form is positive definite on K(t0 ) = {0}. Below, we will prove that, due to condition (SL), is also positive definite on K(τ ) for all sufficiently small τ − t0 > 0. Let us increase the value of τ . For τ = tf , we have three possibilities: Case (1). is positive definite on K. Case (2). is positive semidefinite on K, but is not positive definite on K. Case (3). is not positive semidefinite on K. In Cases (2) and (3) we define τ0 := sup τ ∈ [t0 , tf ] | is positive definite on K(τ ) . (4.47) We will show that (·) ≥ 0 on K(τ0 ), but is not positive on K(τ0 ). Consequently, there exists z̄ ∈ K(τ0 )\{0} such that (z̄) = 0. This fact was called in [25] “the passage of quadratic form through zero.” This property of the form follows from condition (SL) and plays a crucial role in our study of the problem of the definiteness of on K. (Note that another possibility is that is still positive on K(τ0 ). In this case, “does not pass through zero.” Such examples are presented in [25] for a quadratic form that does not satisfy the strengthened Legendre condition.) Now, our goal consists of proving the following theorem. Theorem 4.7. If is not positive definite on K, then there exists τ0 ∈ (t0 , tf ] such that (a) (·) is positive definite on K(τ ) for all τ ∈ (t0 , τ0 ) and (b) (·) ≥ 0 on K(τ0 ) and there exists z̄ ∈ K(τ0 )\{0} such that (z̄) = 0. 192 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Using this theorem, we can define τ0 as the minimum value among all τ ∈ (t0 , tf ] such that the quadratic form has a nontrivial zero z̄ on the subspace K(τ ). To prove this theorem, we need two auxiliary assertions. Proposition 4.8. There exists τ ∈ (t0 , tf ] such that is positive definite on K(τ ). Proof. Let #L > 0 be such that for each v ∈ Rm , Fuu (t, x(t), u(t))v, v ≥ #L |v|2 ∀ t ∈ [t0 , tf ]\. (4.48) Choose a certain τ ∈ (t0 , t∗ ) and let z̄ = (ξ̄ , x̄, ū) ∈ K(τ )\{0} be an arbitrary element. Then ū = 0 and ūχ[t0 ,τ ] = ū, x̄˙ = ū, x̄(t0 ) = 0, x̄χ[t0 ,t∗ ] = x̄, x̄(τ ) = x̄ − , x̄ − + [ẋ]ξ̄ = 0, where χM is the characteristic function of a set M. Consequently, √ √ |x̄ − | τ − t0 x̄ ∞ ≤ ū 1 ≤ τ − t0 ū 2 , |ξ̄ | = ≤ ū 2 , |[ ẋ]| |[ ẋ]| √ |2x̄av | = |x̄ − | ≤ τ − t0 ū 2 . Therefore, t f Fxx x̄, x̄ dt ≤ Fxx ∞ x̄ 2∞ (t∗ − t0 ) ≤ Fxx ∞ (t∗ − t0 )(τ − t0 ) ū 22 , t0 t f ≤ 2 Fxu ∞ x̄ ∞ ū 1 ≤ 2 Fxu ∞ (τ − t0 ) ū 2 , 2F ū, x̄ dt xu 2 t0 |a| |[Fx ]| (τ − t0 ) ū 22 , |2[Fx ]x̄av ξ̄ | ≤ |aξ 2 | ≤ (τ − t0 ) ū 22 . |[ẋ]| |[ẋ]|2 Consequently, (z̄) ≥ tf Fuu ū, ū dt − M(τ − t0 ) ū t0 2 2 ≥ #L ū where M = Fxx ∞ (t∗ − t0 ) + 2 Fxu ∞+ 2 2 − M(τ − t0 ) ū 22 , (4.49) |a| |[Fx ]| . + |[ẋ]| |[ẋ]|2 Let τ be such that t0 < τ < t0 + #L /M. Then (4.49) implies that is positive definite on K(τ ). Proposition 4.9. Assume that is not positive definite on K. Then is positive semidefinite on K(τ0 ), where τ0 is defined as in (4.47). We omit the simple proof of this proposition. Now we are ready to prove the theorem. Proof of Theorem 4.7. We have already proved that is positive definite on K(τ ) for all τ > t0 sufficiently close to t0 (Proposition 4.8). Consequently, τ0 > t0 . Further, we consider 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 193 only the nontrivial case where τ0 < tf . We know that (·) ≥ 0 on K(τ0 ) (Proposition 4.9). We have to show that is not positive on K(τ0 ), i.e., there exists z̄ ∈ K(τ0 )\{0} such that (z̄) = 0 (the passage through zero). Now we follow [26]. For any τ > τ0 (τ ≤ tf ), is not positive on K(τ ) . Therefore, for each τn = τ0 + there exists 1 < tf n (4.50) z̄n ∈ K(τn ) (4.51) (z̄n ) ≤ 0, γ̄ (z̄n ) = 1. (4.52) (4.53) such that The sequence {z̄n } is bounded in K. Therefore, without loss of generality, we assume that z̄n −→ z̄0 weakly. (4.54) Since each subspace K(τn ) ⊂ K is weakly closed, we have z̄0 ∈ K(τn ) and, therefore, z̄0 ∈ " ∀n K(τn ) = K(τ0 ). (4.55) (4.56) n By Proposition 4.9, it follows from (4.56) that (z̄0 ) ≥ 0. (4.57) On the other hand, is weakly lower semicontinuous on K. Thus, (4.54) implies lim inf (z̄n ) ≥ (z̄0 ). (4.58) We obtain from (4.52), (4.57), and (4.58) that and then (z̄n ) −→ (z̄0 ) = 0, (4.59) z̄n −→ z̄0 (4.60) strongly since is a Legendre form. Conditions (4.53) and (4.60) imply γ̄ (z̄0 ) = 1. Consequently, z̄0 = 0, (z̄0 ) = 0, i.e., is not positive on K(τ0 ). For τ ∈ [t0 , tf ], let us consider the problem (APτ ) minimize (z) under the constraint z ∈ K(τ ). (4.61) 194 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Assume that a nonzero element z̄ = (ξ̄ , x̄, ū) = (ξ̄ , w̄) ∈ K(τ ) (4.62) yields the minimum in this problem. Then the following first-order necessary optimality condition holds: (z̄), z̃ = 0 ∀ z̃ ∈ K(τ ), (4.63) where z̃ = (ξ̃ , x̃, ũ) = (ξ̃ , w̃), (z̄) is the Fréchet derivative of the functional at the point z̄, and (·, ·) is the inner product in Z2 (); in more detail, ( (z̄), z̃) = a ξ̄ ξ̃ − [ψ̇]x̄av ξ̃ − [ψ̇]x̃av ξ̄ + tf Fww w̄, w̃ dt. (4.64) t0 Thus, Theorem 4.7 implies the following. Corollary 4.10. Assume that is not positive definite on K. Then, for τ = τ0 ∈ (t0 , tf ] (given by (4.47)), there exists a nonzero element z̄ that satisfies (4.62) and (4.63). On the other hand, we obviously have 1 ( (z̄), z̄) = (z̄). 2 (4.65) This implies the following. Proposition 4.11. If, for certain τ ∈ [t0 , tf ], there exists a nonzero element z̄ satisfying (4.62) and (4.63), then (z̄) = 0, and hence, is not positive definite on K(τ ). Corollary 4.10 and Proposition 4.11 imply the following. Theorem 4.12. Assume that is not positive definite on K. Then τ0 , given by (4.47), is minimal among all τ ∈ (t0 , tf ] such that there exists a nonzero element z̄ satisfying conditions (4.62) and (4.63). 4.1.4 -Conjugate Point In this section, we obtain a dual test for condition (4.63), and then we use it to obtain an analogue of a conjugate point for a broken extremal. The most important role is played by the following lemma. Lemma 4.13. Let τ ∈ (t0 , tf ]. A triple z̄ = (ξ̄ , x̄, ū) satisfies conditions (4.62) and (4.63) iff there exists a function ψ̄(·) : [t0 , tf ] −→ (Rm )∗ such that for the tuple ξ̄ , x̄, ū, ψ̄ , 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 195 the following conditions hold: ξ̄ ∈ R1 , x̄ ∈ P W 1,2 , x̄(t0 ) = x̄(tf ) = 0, x̄˙ = ū, ūχ[t0 ,τ ] = ū, ∗ ū ∈ L2 , ψ̄ ∈ P W 1,2 , (4.66) (4.67) (4.68) ∗ −ψ̄ = x̄ Fxu + ū Fuu a.e. on [t0 , τ ], −ψ̄˙ = x̄ ∗ Fxx + ū∗ Fux a.e. on [t0 , tf ], (4.70) [x̄] = [ẋ]ξ̄ , [ψ̄] = [ψ̇]ξ̄ , a ξ̄ = −ψ̄av [ẋ] + [ψ̇]x̄av . (4.71) (4.72) (4.73) (4.69) Proof. Let z̄ = (ξ̄ , x̄, ū) satisfy conditions (4.62) and (4.63). Consider condition (4.63). Define the following subspace: L̃2 ([τ , tf ], Rm ) := {ṽ ∈ L2 ([t0 , tf ], Rm ) | ṽ = 0 a.e. on [t0 , τ ]}. The operator z̃ −→ (x̃˙ − ũ, ũχ[τ ,tf ] ) maps the space Z2 () onto the space L2 ([t0 , tf ], Rm ) × L̃2 ([τ , tf ], Rm ). The operator z̃ −→ ([x̃] − [ẋ]ξ̃ , x̃(t0 ), x̃(tf )) is finite dimensional. Consequently, the image of the operator z̃ → [x̃] − [ẋ]ξ̃ , x̃˙ − ũ, ũχ[τ ,tf ] , x̃(t0 ), x̃(tf ) , which maps from Z2 () into Rn × L2 ([t0 , tf ], Rm ) × L̃2 ([τ , tf ], Rm ) × Rm × Rm is closed. The kernel of this operator is equal to K(τ ). Consequently, an arbitrary linear functional z$ that vanishes on the kernel of this operator admits the following representation: tf tf $ ˙ ν̄ ũ dt + c̄0 x̃(t0 ) + c̄f x̃(tf ), (4.74) ψ̄(x̃ − ũ) dt + (z , z̃) = ζ̄ ([x̃] − [ẋ]ξ̃ ) − t0 t0 where ζ̄ ∈ (Rn )∗ , ψ̄ ∈ L2 ([t0 , tf ], (Rm )∗ ), ν̄ ∈ L̃2 ([τ , tf ], (Rm )∗ ), c̄0 , c̄f ∈ (Rm )∗ . (4.75) Consequently, the condition (4.63) is equivalent to the existence of ζ̄ , ψ̄, ν̄, c̄0 , and c̄f that satisfy (4.75) and are such that for z$ defined by formula (4.74), we have ( (z̄), z̃) + (z$ , z̃) = 0 ∀ z̃ ∈ Z2 (). The exact representation of the latter condition has the form a ξ̄ ξ̃ − [ψ̇]x̄av ξ̃ − [ψ̇]x̃av ξ̄ tf (Fxx x̄, x̃ + Fxu ū, x̃ + Fux x̄, ũ + Fuu ū, ũ ) dt + t0 tf tf ν̄ ũ dt + c̄0 x̃(t0 ) + c̄f x̃(tf ) ψ̄(x̃˙ − ũ) dt + + ζ̄ ([x̃] − [ẋ]ξ̃ ) − =0 t0 ∀ z̃ = (ξ̃ , x̃, ũ) ∈ Z2 (). t0 (4.76) 196 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Let us examine this condition. (a) We set ξ̃ = 0 and x̃ = 0 in (4.76). Then tf tf ψ̄ ũ dt + (Fux x̄, ũ + Fuu ū, ũ ) dt + t0 t0 Consequently, tf ν̄ ũ dt = 0 ∀ ũ ∈ L2 . (4.77) t0 x̄ ∗ Fxu + ū∗ Fuu + ψ̄ = −ν̄. The latter equation is equivalent to condition (4.69). (b) We set ξ̃ = 0 and ũ = 0 in (4.76). Then tf (Fxx x̄, x̃ ) + Fxu ū, x̃ ) dt + ζ̄ [x̃] − −[ψ̇]x̃av ξ̄ + t0 (4.78) tf ψ̄ x̃˙ dt t0 + c̄0 x̃(t0 ) + c̄f x̃(tf ) = 0 ∀ x̃ ∈ P W 1,2 . Using (4.79), it is easy to show that ψ̄ ∈ P W 1,2 ([t0 , tf ], (Rm )∗ ). Consequently, tf tf t ψ̄ x̃˙ dt = ψ̄ x̃ tf − [ψ̄ x̃] − ψ̄˙ x̃ dt. 0 t0 (4.79) (4.80) t0 Using (4.80) and the definitions x̃av = 12 (x̃ − + x̃ + ), [x̃] = x̃ + − x̃ − in (4.79), we obtain tf 1 − + Fxx x̄ + Fxu ū, x̃ dt − [ψ̇](x̃ + x̃ )ξ̄ + 2 t0 + ζ̄ (x̃ + − x̃ − ) − ψ̄(tf )x̃(tf ) + ψ̄(t0 )x̃(t0 ) + ψ̄ + x̃ + − ψ̄ − x̃ − tf + ∀ x̃ ∈ P W 1,2 . ψ̄˙ x̃ dt + c̄0 x̃(t0 ) + c̄f x̃(tf ) = 0 (4.81) t0 Equation (4.81) implies that the coefficients of x̃ − , x̃ + , x̃(t0 ), x̃(tf ) and the coefficient of x̃ in the integral vanish: 1 − [ψ̇]ξ̄ − ψ̄ − − ζ̄ = 0, 2 1 − [ψ̇]ξ̄ + ψ̄ + + ζ̄ = 0, 2 ψ̄(t0 ) + c̄0 = 0, − ψ̄(tf ) + c̄f = 0, x̄ ∗ F + ū∗ F + ψ̄˙ = 0. xx ux (4.82) (4.83) (4.84) (4.85) (4.86) Adding (4.82) and (4.83), we obtain −[ψ̇]ξ̄ + [ψ̄] = 0. (4.87) Thus, (4.70) and (4.72) hold. Subtracting (4.82) from (4.83) and dividing the result by two, we obtain (4.88) ζ̄ = −ψ̄av . 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 197 (c) We set x̃ = 0 and ũ = 0 in (4.76). Then a ξ̄ ξ̃ − [ψ̇]x̄av ξ̃ − ζ̄ [ẋ]ξ̃ = 0 ∀ ξ̃ ∈ R1 . (4.89) Consequently, the coefficient of ξ̃ vanishes: a ξ̄ − [ψ̇]x̄av − ζ̄ [ẋ] = 0. (4.90) Using (4.88) in (4.90), we obtain (4.73). Conditions (4.67), (4.68), and (4.71) and the first three conditions in (4.66) are implied by (4.62). Thus, all conditions (4.66)–(4.73) hold. Conversely, if a tuple (ξ̄ , x̄, ū, ψ̄) satisfies conditions (4.66)–(4.73), then one easily verifies that z̄ = (ξ̄ , x̄, ū) satisfies (4.62) and (4.63). Let z̄ = (ξ̄ , x̄, ū) ∈ K. Obviously, the condition z̄ = 0 is equivalent to x̄(·) = 0. Thus, we obtain the following theorem from Theorem 4.12 and Lemma 4.13 under the above condition (SL). Theorem 4.14. Assume that is not positive definite on K. Then τ0 (given by equation (4.47)) is a minimal among all τ ∈ (t0 , tf ] such that there exists a tuple (ξ̄ , x̄, ū, ψ̄) that satisfies (4.66)–(4.73) and the condition x̄(·) = 0. In what follows, we will assume that a > 0 and, as above, condition (SL) holds. Assume that τ0 < tf . We know that ≥ 0 on K(τ0 ). Theoretically, it is possible that ≥ 0 on K(τ1 ) for a certain τ1 > τ0 . In this case, the closed interval [τ0 , τ1 ] is called a table [25]. Tables occur in optimal control problems [25], but, for a smooth extremal, they never arise in the calculus of variations. We now show that for the simplest problem of the calculus of variations, the closed interval [τ0 , tf ] cannot serve as a table in the case of a broken extremal. To this end, we complete Lemma 4.13 by the following two propositions. Proposition 4.15. If the functions x̄, ψ̄ ∈ P W 1,2 and ū ∈ L2 satisfy the system x̄˙ = ū, −ψ̄˙ = x̄ ∗ Fxx + ū∗ Fux , −ψ̄ = x̄ ∗ Fxu + ū∗ Fuu (4.91) on a certain closed interval ⊂ [t0 , tf ], then the functions x̄ and ψ̄ satisfy the system x̄˙ = Ax̄ + B ψ̄ ∗ , −ψ̄˙ = x̄ ∗ C + ψ̄A (4.92) on the same closed interval , where A, B, and C are the same as in (4.20). Proof. The third equation in (4.91) implies −1 ∗ −1 Fux x̄, ψ̄ − Fuu ū = −Fuu −1 −1 ū∗ = −ψ̄Fuu − x̄ ∗ Fxu Fuu . Substituting these expressions for ū and ū∗ into the first and second equations in (4.91), respectively, we obtain (4.92). Proposition 4.16. Conditions (4.71)–(4.73) imply a ξ̄ = −ψ̄ + [ẋ] + [ψ̇]x̄ + . (4.93) 198 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Proof. Using the expressions ψ̄av = ψ̄ + − 0.5[ψ̄] and x̄av = x̄ + − 0.5[x̄], together with (4.71) and (4.72), in (4.73), we obtain (4.93). Theorem 4.17. If ≥ 0 on K, then is positive definite on K(τ ) for all τ < tf . Proof. Assume the contrary, i.e., ≥ 0 on K and τ0 < tf . Choose an element z̄ = (ξ̄ , x̄, ū) ∈ K(τ0 ) \ {0} such that (z̄) = 0. Then z̄ yields the minimum in problem (APτ ) for τ = tf , and hence, (4.63) holds for K(τ ) = K. By Lemma 4.13, there exists a function ψ̄, defined on [t0 , tf ], such that the conditions (4.66)–(4.73) hold for τ = tf . Then, according to Propositions 4.15 and 4.16, x̄ and ψ̄ also satisfy system (4.92) on [t0 , tf ], and relation (4.93) holds for them. The conditions z̄ ∈ K(τ0 ), τ0 < tf , together with (4.69) and the condition τ = tf imply x̄(tf ) = 0, ψ̄(tf ) = 0 and hence x̄(t) = 0, ψ̄(t) = 0 on [t∗ , tf ] since x̄, ψ̄ satisfy (4.92) on the same closed interval. Therefore, x̄ + = 0, ψ̄ + = 0 and then (4.93) and the condition a > 0 imply ξ̄ = 0. By (4.71) and (4.72), we have [x̄] = 0, [ψ̄] = 0. Consequently, x̄ − = 0, ψ̄ − = 0, and then x̄(t) = 0, ψ̄(t) = 0 on [t0 , t∗ ], since x̄ and ψ̄ satisfy (4.92) on the same closed interval. Thus, x̄(t) = 0 on [t0 , tf ], and hence ū(t) = 0 on [t0 , tf ]. We arrive at a contradiction with condition z̄ = 0; this proves the theorem. Further, we examine conditions (4.66)–(4.73). Using Proposition 4.15, we exclude ū from these conditions. We obtain the following system: ξ̄ ∈ R1 , x̄, ψ̄ ∈ P W 1,2 , x̄(t0 ) = x̄(tf ) = 0, x̄˙ = Ax̄ + B ψ̄ ∗ , −ψ̄˙ = x̄ ∗ C + ψ̄A on [t0 , τ ], x̄˙ = 0, −ψ̄˙ = x̄ ∗ F on [τ , t ], (4.97) [x̄] = [ẋ]ξ̄ , [ψ̄] = [ψ̇]ξ̄ , a ξ̄ = −ψ̄ − [ẋ] + [ψ̇]x̄ − . (4.98) (4.99) xx f (4.94) (4.95) (4.96) The latter condition is obtained from (4.71)–(4.73) similarly to condition (4.93). If ψ̄(t0 ) = 0, then (4.94)–(4.99) imply x̄(·) = 0. Consequently, we can assign the nontriviality condition in the form (4.100) ψ̄(t0 ) = 0. Definition 4.18. A point τ ∈ (t0 , tf ] is called -conjugate (to t0 ) if there exists a triple (ξ̄ , x̄, ψ̄) that satisfies conditions (4.94)–(4.100). Obviously, a point τ ∈ (t0 , tf ] is -conjugate to t0 iff, for a given τ , there exists a quadruple (ξ̄ , x̄, ū, ψ̄) that satisfies conditions (4.66)–(4.73) and the condition x̄(·) = 0. Consequently, Theorems 4.14 and 4.17 imply the following. Theorem 4.19. The form is positive semidefinite on K iff there is no point that is -conjugate to t0 on the interval (t0 , tf ). The form is positive definite on K iff there is no point that is -conjugate to t0 on the half-interval (t0 , tf ]. Now let us examine the condition for positive definiteness of on K(t∗ ). This condition implies the positive definiteness of ω on K0 (t∗ ) which is defined as a subspace of 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 199 pairs w̄ = (x̄, ū) ∈ K0 (see Section 4.1.2) such that ū(t) = 0 on [t∗ , tf ]. Let X(t), "(t) be a matrix-valued solution of the Cauchy problem (4.22)–(4.23) on [t0 , t∗ ]. As is well known, the positive definiteness of ω on K0 (t∗ ) is equivalent to the condition det X(t) = 0 ∀ t ∈ (t0 , t∗ ]. (4.101) Assume that this condition holds. We set Q(t) = "(t)X−1 (t), t ∈ (t0 , t∗ ]. Consider conditions (4.94)–(4.100) for τ ≤ t∗ . These conditions imply that x̄(τ ) = x̄ − = x̄(t) ∀ t ∈ (τ , t∗ ), ψ̄ − = ψ̄(τ ) − x̄(τ )∗ (τ ), where (τ ) = to the system t∗ τ Fxx (t, w(t)) dt. x̄ − + [ẋ]ξ̄ = x̄ + = 0, Hence, for τ ≤ t∗ the system (4.94)–(4.100) is equivalent ξ̄ ∈ R1 , x̄, ψ̄ ∈ P W 1,2 , x̄(t0 ) = 0, ψ̄(t0 ) = 0, x̄˙ = Ax̄ + B ψ̄ ∗ , −ψ̄˙ = x̄ ∗ C + ψ̄A on [t0 , τ ], x̄(τ ) + [ẋ]ξ̄ = 0, a ξ̄ = − ψ̄(τ ) − x̄(τ )∗ (τ ) [ẋ] + [ψ̇]x̄(τ ). (4.102) (4.103) (4.104) (4.105) (4.106) Let (ξ̄ , x̄, ψ̄) be a solution of this system on the closed interval [t0 , τ ], where τ ∈ (t0 , t∗ ]. Then there exists c̄ ∈ Rm such that x̄(t) = X(t)c̄, ψ̄(t) = c̄∗ " ∗ (t). (4.107) Consequently, ψ̄(t)∗ = "(t)c̄ = "(t)X−1 (t)(X(t)c̄) = Q(t)x̄(t). Using this relation, together with (4.105), in (4.106), we obtain a ξ̄ − [ẋ]∗ (Q(τ ) − (τ ))[ẋ]ξ̄ + [ψ̇][ẋ]ξ̄ = 0. (4.108) If ξ̄ = 0, then from (4.105) and (4.107), we obtain x̄(·) = 0, ψ̄(·) = 0; this contradicts (4.103). Therefore ξ̄ = 0, and then (4.108) implies a − [ẋ]∗ (Q(τ ) − (τ ))[ẋ] + [ψ̇][ẋ] = 0. (4.109) We have obtained this relation from the system (4.102)–(4.106). Conversely, if (4.109) holds, then, setting ξ̄ = −1 and c̄ = X −1 (τ )[ẋ] and defining x̄, ψ̄ by formulas (4.107), we obtain a solution of the system (4.102)–(4.106). We have proved the following lemma. Lemma 4.20. Assume that condition (4.101) holds. Then, τ ∈ (t0 , t∗ ] is a point that is -conjugate to t0 iff condition (4.109) holds. We set μ(t) = a − [ẋ]∗ (Q(t) − (t))[ẋ] + [ψ̇][ẋ]. (4.110) Then, by Lemma 4.20, the absence of a -conjugate point in (t0 , t∗ ] is equivalent to the condition (4.111) μ(t) = 0 ∀ t ∈ (t0 , t∗ ]. 200 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals We show further that the function μ(t) does not increase on (t0 , t∗ ] and μ(t0 + 0) = +∞. Consequently, condition (4.111) is equivalent to μ(t∗ ) > 0, which is another form of condition (4.28) or condition (4.40). Proposition 4.21. Assume that a symmetric matrix Q(t) satisfies the Riccati equation (4.31) on (t0 , τ ), where τ ≤ t∗ . Then μ̇(t) ≤ 0 ∀ t ∈ (t0 , τ ). (4.112) Moreover, if Q satisfies the initial condition (4.36), then μ(t0 + 0) = +∞. (4.113) μ̇ = −(Q̇ + Fxx )[ẋ], [ẋ] . (4.114) Proof. For t ∈ (t0 , τ ), we have Using (4.31) in (4.114), we obtain −1 (Q + Fux )[ẋ], [ẋ] μ̇ = −(Q + Fxu )Fuu −1 (Q + Fux )[ẋ], (Q + Fux )[ẋ] ≤ 0 = −Fuu on (t0 , τ ), (4.115) −1 is positive definite. Assume additionally that Q satisfies (4.36). Then Q = "X −1 , since Fuu where X and " satisfy (4.22) and (4.23). It follows from (4.22) and (4.23) that X(t0 ) = O, −1 Ẋ(t0 ) = −B(t0 ) = Fuu (t0 , w(t0 )), "(t0 ) = −I . (4.116) Consequently, −1 (t0 , w(t0 )) + o(t), X(t) = (t − t0 )Fuu "(t) = −I + o(1) as t → t0 + 0. (4.117) Thus, Q(t) = − 1 (Fuu (t0 , w(t0 )) + o(1)) t − t0 as t → t0 + 0; (4.118) this implies −Q(t)[ẋ], [ẋ] = 1 (Fuu (t0 , w(t0 ))[ẋ], [ẋ] + o(1)) → +∞ t − t0 as t → t0 + 0. (4.119) Now (4.110) and (4.119) imply (4.113). Lemma 4.20 and Proposition 4.21 imply the following. Theorem 4.22. Assume that condition (4.101) holds. Then the absence of a point τ that is -conjugate to t0 on (t0 , t∗ ] is equivalent to condition (4.28). Consequently, the positive definiteness of on K(t∗ ) is equivalent to the validity of conditions (4.101) and (4.28). We examine further the system (4.94)–(4.100) for τ > t∗ . Using (4.99) and the condition a > 0, we can exclude ξ̄ from this system. Moreover, for x̄, ψ̄, we can formulate 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 201 all conditions on [t0 , τ ] only. As a result, we arrive at the following equivalent system on [0, τ ]: x̄, ψ̄ ∈ P W 1,2 [t0 , τ ], x̄(0) = x̄(τ ) = 0, ψ̄(0) = 0, x̄˙ = Ax̄ + B ψ̄ ∗ , −ψ̄˙ = x̄ ∗ C + ψ̄A on [t0 , τ ], a[x̄] = [ẋ] − ψ̄ − [ẋ] + [ψ̇]x̄ − , a[ψ̄] = [ψ̇] − ψ̄ − [ẋ] + [ψ̇]x̄ − . (4.120) (4.121) (4.122) (4.123) (4.124) We have proved the following lemma. Lemma 4.23. A point τ ∈ (t∗ , tf ] is -conjugate to t0 iff there exists a pair of functions x̄, ψ̄ that satisfies conditions (4.120)–(4.124). Now we can prove the following theorem. Theorem 4.24. A point τ ∈ (t∗ , tf ] is a -conjugate to t0 iff det X(τ ) = 0, (4.125) where X(t), "(t) is a solution to the problem (4.22)–(4.25). Proof. Assume that condition (4.125) holds, in which X, " is a solution to the problem (4.22)–(4.25). Then there exists c̄ ∈ Rm such that X(τ )c̄ = 0, We set x̄ = X c̄, c̄ = 0. ψ̄ = c̄∗ " ∗ . (4.126) (4.127) Then x̄ and ψ̄ satisfy (4.120)–(4.124). Conversely, let x̄ and ψ̄ satisfy (4.120)–(4.124). We set c̄ = −ψ̄(t0 )∗ . Then conditions (4.126) and (4.127) hold. Conditions (4.126) imply condition (4.125). Note that Theorems 4.19, 4.22, and 4.24 imply Theorems 4.3 and 4.4. To complete the proof of the results of Section 4.1.2, we have to consider a jump condition for a solution Q to the Riccati equation (4.30) with initial condition (4.36). In what follows, we assume that is nonnegative on K and that the pair X, " is a solution to the problem (4.22)–(4.25). Then conditions (4.26) and (4.27) hold. We set Q = "X−1 . Then " = QX. Using the relations q− = [ẋ]∗ Q− − [ψ̇] and " − = Q− X− in the jump conditions (4.24) and (4.25) for X and ", we obtain Proposition 4.25. Condition holds. a[X] = −[ẋ](q− )X − , a["] = −[ψ̇]∗ (q− )X − . (4.128) (4.129) det aI − [ẋ](q− ) = 0 (4.130) 202 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Proof. Relation (4.128) implies aX + = aI − [ẋ](q− ) X− . (4.131) Now (4.130) follows from this relation considered together with (4.27) and the inequality a > 0. Proposition 4.26. The following equality holds: −1 −1 (q− )∗ (q− ) aI − [ẋ](q− ) = (q− )∗ (q− ) a − (q− )[ẋ] , (4.132) where a − (q− )[ẋ] > 0. Proof. Equality (4.132) is equivalent to (q− )∗ (q− ) aI − [ẋ](q− ) = (q− )∗ (q− ) a − (q− )[ẋ] . Let us show that this equality holds. Indeed, we have (q− )∗ (q− ) aI − [ẋ](q− ) = a(q− )∗ (q− ) − (q− )∗ (q− ) [ẋ](q− ) = a(q− )∗ (q− ) − (q− )∗ (q− )[ẋ] (q− ) = a(q− )∗ (q− ) − (q− )[ẋ] (q− )∗ (q− ) = (q− )∗ (q− ) a − (q− )[ẋ] since (q− )[ẋ] is a number. Proposition 4.27. The jump condition (4.37) holds. Proof. The relation " = QX implies ["] = Q+ X+ − Q− X− . Multiplying it by a and using relations (4.129) and (4.131), we obtain −[ψ̇]∗ (q− )X − = Q+ aI − [ẋ](q− ) X − − aQ− X− . Since det X− = 0 and Q+ = Q− + [Q], we get −[ψ̇]∗ (q− ) = a[Q] − Q− [ẋ](q− ) − [Q][ẋ](q− ). This relation and the formula (q− )∗ = Q− [ẋ] − [ψ̇]∗ imply the equality [Q] aI − [ẋ](q− ) = (q− )∗ (q− ). By virtue of (4.130) this equality is equivalent to −1 [Q] = (q− )∗ (q− ) aI − [ẋ](q− ) . According to (4.132) this relation can be rewritten as −1 [Q] = (q− )∗ (q− ) a − (q− )[ẋ] . This implies jump condition (4.37). 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 4.1.5 203 Numerical Examples Rayleigh problem with regulator functional. The following optimal control problem (Rayleigh problem) was thoroughly investigated in [65]. Consider the electric circuit (tunnel-diode oscillator) shown in Figure 4.1. The state variable x1 (t) is taken as the electric current I at time t, and the control variable u(t) is induced by the voltage V0 at the generator. After a suitable transformation of the voltage V0 (t), we arrive at the following specific Rayleigh equation with a scalar control u(t) , ẍ(t) = −x(t) + ẋ(t) ( 1.4 − 0.14 ẋ(t)2 ) + 4 u(t). A numerical analysis reveals that the Rayleigh equation with zero control u(t) ≡ 0 has a limit cycle in the (x, ẋ)-plane. The goal of the control process is to avoid the strong oscillations on the limit cycle by steering the system toward a small neighborhood of the origin (x, ẋ) = (0, 0) using a suitable control function u(t). With the state variables x1 = x and x2 = ẋ we arrive at the following control problem with fixed final time tf > 0: Minimize the functional tf F (x, u) = ( u(t)2 + x1 (t)2 ) dt (4.133) 0 subject to ẋ1 (t) = x2 (t), ẋ2 (t) = −x1 (t) + x2 (t) ( 1.4 − 0.14 x2 (t)2 ) + 4 u(t), (4.134) x1 (0) = x2 (0) = −5, (4.135) x1 (tf ) = x2 (tf ) = 0. As in [65], we choose the final time tf = 4.5 . The Pontryagin function (Hamiltonian), which corresponds to the minimum principle, becomes H (x1 , x2 , ψ1 , ψ2 , u) = u2 + x12 + ψ1 x2 + ψ2 (−x1 + x2 (1.4 − 0.14x22 ) + 4 u), (4.136) where ψ1 and ψ2 are the adjoint variables associated with x1 and x2 . The optimal control that minimizes the Pontryagin function is determined by Hu = 2 u + 4 ψ2 = 0, L i.e., u(t) = −2 ψ2 (t). (4.137) I # V0 "! ? IC ID ? IR ? C AA D # ? R V 6 t "! Figure 4.1. Tunnel-diode oscillator. I denotes inductivity, C capacity, R resistance, I electric current, and D diode. 204 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals The strict Legendre–Clebsch condition holds in view of Huu (t) ≡ 2 > 0 . The adjoint equation ψ̇ = −Hx yields ψ̇1 = ψ2 − 2x1 , ψ̇2 = 0.42 ψ2 x22 − 1.4 ψ2 − ψ1 . (4.138) Since the final state is specified, there are no boundary conditions for the adjoint variable. The boundary value problem (4.133), (4.134) and (4.138) with control u = −2ψ2 was solved using the multiple shooting code BNDSCO developed by Oberle and Grimm [82]. The optimal state, control, and adjoint variables are shown in Figure 4.2. We get the following initial and final values for the adjoint variables: ψ1 (0) = −9.00247067, ψ1 (4.5) = −0.04456054, ψ2 (0) = −2.67303084, ψ2 (4.5) = −0.00010636. (4.139) Nearly identical numerical results can be obtained by solving the discretized control problem with a high number of gridpoints. The optimality of this extremal solution may be checked by producing a finite solution of the Riccati equation (4.30). Since the Rayleigh problem has dimension n = 2 , we (a) state variables x1 and x2 (b) 7 6 5 4 3 2 1 0 -1 -2 5 4 3 2 1 0 -1 -2 -3 -4 -5 -6 0 0.5 (c) 1 1.5 2 2.5 adjoint variables 3 1 and 3.5 4 4.5 control u 0 0.5 (d) 2 2 2.5 0 2 -2 1.5 -4 1 -6 0.5 -8 0 1 1.5 2 2.5 3 3.5 4 4.5 solutions Q11, Q12, Q22 of Riccati equation -0.5 -10 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 Figure 4.2. Rayleigh problem with regular control. (a) State variables. (b) Control. (c) Adjoint variables. (d) Solutions of the Riccati equation (4.140). 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 205 consider a symmetric 2 × 2 matrix Q(t) of the form ⎞ ⎛ Q11 (t) Q12 (t) ⎠. Q(t) = ⎝ Q12 (t) Q22 (t) The Riccati equation (4.30) then reads explicitly as Q̇11 = 2Q12 + 8Q212 − 2, Q̇12 = −Q11 + ( 0.42 x22 − 1.4 ) Q12 + Q22 + 8 Q12 Q22 , Q̇22 = −2 Q12 + 2 ( 0.42 x22 − 1.4 ) Q22 + 8 Q222 + 0.84 ψ2 x2 . (4.140) Since the final state is fixed, no boundary condition is prescribed for Q(t1 ) . Choosing for convenience the boundary condition Q(t1 ) = 0 , we obtain the bounded solution of the Riccati equation shown in Figure 4.2. The initial values are computed as Q11 (0) = 2.006324, Q12 (0) = 0.4705135, Q22 (0) = −0.3516654. Thus the unconstrained solution shown in Figure 4.2 provides a local minimum. To conclude the discussion of the Rayleigh problem, we consider the control problem with free final state x(tf ). The solution is quite similar to that shown in Figure 4.2, with the only difference being that the boundary inequality Q(tf ) > 0 should hold. However, it suffices to find a solution Q(t) satisfying the boundary condition Q(tf ) = 0 which was imposed earlier. Due to the continuous dependence of solutions on initial or terminal conditions, equation (4.140) then has a solution with Q(tf ) = # · I2 > 0 for # > 0 small. We obtain x1 (4.5) = −0.0957105, x2 (4.5) = −0.204377, ψ1 (0) = −9.00126, ψ2 (0) = −2.67259, (4.141) 2.00607, Q22 (0) = 0.470491, Q11 (0) = Q22 (0) = −0.351606. Variational problem with a conjugate point. The following variational problem was studied in Maurer and Pesch [71]: 1 1 ( x(t)3 + ẋ(t)2 ) dt subject to x(0) = 4, x(1) = 1. (4.142) Minimize F (x, u) = 2 0 Defining the control variable by u = ẋ as usual, the Pontryagin function (Hamiltonian), corresponding to the minimum principle, becomes 1 H (x, u, ψ) = (x 3 + u2 ) + ψ u. 2 The strict Legendre–Clebsch condition holds in view of Huu = 1 > 0. The minimizing control satisfies Hu = 0 which gives u = −ψ. Using the adjoint equation ψ̇ = −Hx = −3x 2 /2, we get the boundary value problem (Euler–Lagrange equation) 3 ẍ = x 2 , 2 x(0) = 4, x(1) = 1. (4.143) 206 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals The unknown initial value ẋ(0) can be determined by shooting methods; cf. Stoer and Bulirsch [106]. The boundary value problem (4.143) has the explicit solution x (1) (t) = 4/(1 + t)2 and a second solution x (2) (t) with initial values ẋ (1) (0) = −8, ẋ (2) (0) = −35.858549. (4.144) Both solutions x (k) (t), k = 1, 2, may be tested for optimality by the classical Jacobi condition. The variational system (4.22), (4.23), along the two extremals, yields for k = 1, 2 ẍ (k) (t) ÿ (k) (t) = = x (k) (0) = 4, y (k) (0) = 0, 3 2 x (k) (t), 3 x (k) (t) y (k) (t), ẋ (k) (0) as in (4.144), ẏ (k) (0) = 1. (4.145) For k = 1, 2 the extremals x (k) (k = 1, 2) and variational solutions y (k) (k = 1, 2) are displayed in Figure 4.3. The extremal x (1) (t) = 4/(1 + t)2 is optimal in view of y (1) (t) > 0 ∀ 0 < t ≤ 1, whereas the second extremal x (2) is not optimal, since it exhibits the conjugate point tc = 0.674437 with y (2) (tc ) = 0. The conjugate point tc has the property that the envelope of variations y (1) and y (2) x(1) and x(2) 4 2.5 2 2 0 1.5 -2 1 -4 -6 0.5 -8 0 -10 -0.5 -12 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 Envelope of x(2) and conjugate point tc = 0.674437 4 2 0 -2 -4 -6 -8 -10 -12 0 0.2 0.4 0.6 0.8 1 Figure 4.3. Top left: Extremals x (1) , x (2) (lower graph). Top right: Variational solutions y (1) and y (2) (lower graph) to (4.145). Bottom: Envelope of neighboring extremals illustrating the conjugate point tc = 0.674437. 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 207 extremals corresponding to neighboring initial slopes ẋ(0) touches the extremal x (2) (t) at tc ; cf. Figure 4.3. Example with a Broken Extremal Consider the following problem: Minimize J (x(·), u(·)) := 1 2 tf (f (u(t)) − x 2 (t)) dt (4.146) 0 under the constraints ẋ(t) = u(t), x(0) = x(tf ) = 0, (4.147) where f (u) = min{(u + 1)2 , (u − 1)2 }, tf > 0 is fixed, and u = 0. This example was examined in [79]. Here, we will study it using the results of this section. In [79] it was shown that for each closed interval [0, tf ] we have an infinite number of extremals. For a given closed interval [0, tf ], each extremal x(t) is a periodic piecewise smooth function with a finite number of switching points of the control tk = (2k − 1)ϕ, k = 1, . . . , s, where 0 < ϕ < π2 , tf = 2sϕ < sπ . In particular, for s = 1 we have tf = 2ϕ < π . Set = {t1 , . . . , ts }. For each s ≥ 1 the function x(t) is defined by the conditions ẍ = −x for t ∈ [0, tf ] \ , x(0) = x(tf ) = 0, [ẋ]k = (−1)k 2, k = 1, . . . , s (we identify the extremals differing only by sign). Consequently, x(t) = ρ sin t for t ∈ [0, ϕ) and x(t) = ρ sin(t − 2ϕ) for t ∈ (ϕ, 2ϕ), etc., where ρ = (cos ϕ)−1 . Further, in [79] it was shown that for tf > π each extremal does not satisfy the necessary second-order condition. The same is true for the case tf ≤ π and s > 1. Consequently, the only possibility for the quadratic form to be nonnegative on the critical subspace K is the case s = 1. In this case we have tf = 2ϕ < π and 1 tf 2 = (tan ϕ)ξ̄ 2 + (ū − x̄ 2 ) dt. 2 0 The critical subspace K is defined by the conditions x̄˙ = ū, x̄(0) = x̄(tf ) = 0, [x̄] = −2ξ̄ . The Riccati equation (4.34) has the form (for a small ε > 0) −Ṙ = 1 + R 2 , t ∈ [0, ε], R(0) = 0. The solution of this Cauchy problem is R(t) = − tan t. Consequently, Q(t) = R −1 (t) = − cot t is the solution of the Cauchy problem −Q̇ + Q2 + 1 = 0, Q(ε) = R −1 (ε) on the interval (0, ϕ). It follows that Q− = − cot ϕ. Using the equalities [ẋ] = [u] = −2, [ψ̇] = [x] = 0, and a = D(H ) = 2 tan ϕ, we obtain q− = [ẋ]Q− − [ψ̇] = −2Q− = 2 cot ϕ, b− = a − q− [ẋ] = 2 tan ϕ + 4 cot ϕ > 0, 208 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals and then [Q] = (b− )−1 (q− )2 = (b− )−1 4 cot 2 ϕ, Q+ = Q− + [Q] = − cot ϕ + 4 cot 2 ϕ = −2(b− )−1 < 0. b− Now, for t ∈ (ϕ, tf ], we have to solve the Cauchy problem −Q̇ + Q2 + 1 = 0, Q(ϕ+) = −2(b− )−1 . Since Q(ϕ+) < 0, the solution has the form Q(t) = − cot(t + τ ) with certain τ > 0. It is clear that this solution is defined at least on the closed interval [ϕ, ϕ + 0.5π ] which contains the closed interval [ϕ, tf ]. By Theorem 4.6 the quadratic form is positive definite on the subspace K. It means that the extremal with single switching point satisfies the sufficient second-order optimality condition. By Theorem 4.2, such an extremal is a point of a strict bounded strong minimum in the problem. 4.1.6 Transformation of the Quadratic Form to Perfect Squares for Broken Extremals in the Simplest Problem of the Calculus of Variations Assume again that w(·) = (x(·), u(·)) is an extremal in the simplest problem of the calculus of variations (4.1)–(4.3), with a single corner point t∗ ∈ (t0 , tf ). Assume that condition (SL) is satisfied, and a > 0, where a is defined by (4.8). Set = {t∗ }. Let Q(t) be a symmetric matrix function on [t0 , tf ] with piecewise continuous entries which are continuously differentiable on each of the two intervals of the set [t0 , tf ]\ (hence, a jump of Q is possible only at the point t∗ ). We continue to use the following notation Q− = Q(t∗ − 0), Q+ = Q(t∗ + 0), [Q] = Q+ − Q− , q− = [ẋ]∗ Q− − [ψ̇], b− = a − q− [ẋ], −1 (t, w(t))F (t, w(t)), A(t) = −Fuu ux −1 (t, w(t)), B(t) = −Fuu −1 (t, w(t))F (t, w(t)) + F (t, w(t)). C(t) = −Fxu (t, w(t))Fuu ux xx (4.148) (4.149) Note that B(t) is negative definite. Assume that Q satisfies (a) the Riccati equation Q̇(t) + Q(t)A(t) + A(t)∗ Q(t) + Q(t)B(t)Q(t) + C(t) = 0 (4.150) for all t ∈ [t0 , tf ]\, (b) the inequality b− > 0, (4.151) [Q] = (b− )−1 (q− )∗ (q− ), (4.152) (c) the jump condition 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 209 where q− is a row vector and (q− )∗ is a column vector. (Recall that (q− )∗ (q− ) is a symmetric positive semidefinite matrix.) In this case, we will say that the matrix Q(t) is the symmetric solution of the problem (4.150)–(4.152) on [t0 , tf ]. For the quadratic form on the subspace K, defined by relations (4.14) and (4.13), respectively, the following theorem holds. Theorem 4.28. If there exists a symmetric solution Q(t) of the problem (4.150)–(4.152) on [t0 , tf ], then the quadratic form has the following transformation into a perfect square on K: tf −1 − 2 B(t)v̄(t), v̄(t) dt, (4.153) 2(z̄) = (b− ) (a ξ̄ + q− x̄ ) + t0 where v̄(t) = Q(t) + Fux (t, w(t)) x̄(t) + Fuu (t, w(t))ū(t). (4.154) The proof of this theorem consists of the following four propositions. Proposition 4.29. Let Q(t) be a symmetric matrix on [t0 , tf ] with piecewise continuous entries, which are absolutely continuous on each interval of the set [t0 , tf ]\. Then 2(z̄) = b− ξ̄ 2 + 2q− x̄ + ξ̄ + [Q]x̄ + , x̄ + tf (4.155) (Fxx + Q̇)x̄, x̄ + 2(Fxu + Q)ū, x̄ + Fuu ū, ū dt + t0 for all z̄ = (ξ̄ , x̄, ū) ∈ K. Proof. For z̄ ∈ K, we obviously have tf d Qx̄, x̄ dt = Qx̄, x̄ dt t0 tf − [Qx̄, x̄ ]. (4.156) t0 Using the conditions x̄(t0 ) = x̄(tf ) = 0 and x̄˙ = ū in (4.156), we obtain tf 0 = Q+ x̄ + , x̄ + − Q− x̄ − , x̄ − + (Q̇x̄, x̄ + 2Qx̄, ū ) dt. (4.157) t0 Adding this zero form to the form − + 2(z̄) = a ξ̄ − [ψ̇](x̄ + x̄ )ξ̄ + 2 tf (Fxx x̄, x̄ + 2Fxu ū, x̄ + Fuu ū, ū ) dt, t0 we obtain 2(z̄) = a ξ̄ 2 − [ψ̇](x̄ − + x̄ + )ξ̄ − Q− x̄ − , x̄ − + Q+ x̄ + , x̄ + tf + (Fxx + Q̇)x̄, x̄ + 2(Fxu + Q)ū, x̄ + Fuu ū, ū dt. t0 (4.158) 210 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Consider the form ω∗ (ξ̄ , x̄) := a ξ̄ 2 − [ψ̇](x̄ − + x̄ + )ξ̄ − Q− x̄ − , x̄ − + Q+ x̄ + , x̄ + (4.159) connected to the jump point t∗ ∈ of the control u(·), and let us represent it as a function of x̄ + , ξ̄ , ω∗ (ξ̄ , x̄) = a ξ̄ 2 − [ψ̇](2x̄ + − [x̄])ξ̄ − Q− (x̄ + − [x̄]), x̄ + − [x̄] + Q+ x̄ + , x̄ + = a ξ̄ 2 − 2[ψ̇]x̄ + ξ̄ + [ψ̇][x̄]ξ̄ − Q− x̄ + , x̄ + + 2Q− [x̄], x̄ + − Q− [x̄], [x̄] + Q+ x̄ + , x̄ + . Using the relations [x̄] = [ẋ]ξ̄ and Q+ − Q− = [Q], we obtain ω∗ (ξ̄ , x̄) = (a − Q[ẋ] − [ψ̇]∗ , [ẋ] )ξ̄ 2 + 2Q− [ẋ] − [ψ̇]∗ , x̄ + ξ̄ + [Q]x̄ + , x̄ + . Now using definitions (4.148) of q− and b− , we obtain ω∗ (ξ̄ , x̄) = b− ξ̄ 2 + 2q− x̄ + ξ̄ + [Q]x̄ + , x̄ + . (4.160) Formulas (4.158), (4.159), and (4.160) imply formula (4.155). Proposition 4.30. If Q satisfies the jump condition b− [Q] = (q− )∗ (q− ), then b− ξ̄ 2 + 2q− x̄ + ξ̄ + [Q]x̄ + , x̄ + = (b− )−1 (a ξ̄ + q− x̄ − )2 . (4.161) Proof. Using the jump conditions for [Q] and [x̄] and the definition (4.148) of b− , we obtain b− ξ̄ 2 + 2q− x̄ + ξ̄ + [Q]x̄ + , x̄ + = (b− )−1 (b− )2 ξ̄ 2 + 2q− x̄ + b− ξ̄ + (q− x̄ + )2 2 = (b− )−1 b− ξ̄ + q− x̄ + 2 = (b− )−1 (a − q− [ẋ])ξ̄ + q− x̄ + 2 = (b− )−1 a ξ̄ − q− [x̄] + q− x̄ + 2 = (b− )−1 a ξ̄ + q− x̄ − . Proposition 4.31. The Riccati equation (4.150) is equivalent to the equation −1 Q̇ − (Q + Fxu )Fuu (Q + Fux ) + Fxx = 0 on [t0 , tf ] \ . Proof. Using formulas (4.149), we obtain QA + A∗ Q + QBQ + C −1 −1 −1 −1 Fux − Fxu Fuu Q − QFuu Q − Fxu Fuu Fux + Fxx = −QFuu −1 −1 = −QFuu (Q + Fux ) − Fxu Fuu (Q + Fux ) + Fxx −1 = −(Q + Fxu )Fuu (Q + Fux ) + Fxx . Hence, (4.150) is equivalent to (4.162). (4.162) 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 211 Proposition 4.32. If Q satisfies the Riccati equation (4.162), then −1 (Fxx + Q̇)x̄, x̄ + 2(Fxu + Q)ū, x̄ + Fuu ū, ū = Fuu v̄, v̄), (4.163) where v̄ is defined by (4.154). Proof. From (4.154), it follows that −1 −1 Fuu v̄, v̄ = Fuu ((Q + Fux )x̄ + Fuu ū), (Q + Fux )x̄ + Fuu ū) −1 −1 (Q + Fux )x̄, (Q + Fux )x̄ + 2Fuu (Q + Fux )x̄, Fuu ū = Fuu −1 Fuu ū, Fuu ū +Fuu −1 (Q + Fux )x̄, x̄ = (Q + Fxu )Fuu (4.164) + 2(Q + Fux )x̄, ū + Fuu ū, ū . Using (4.162), (4.164), we obtain (4.163). Now, let Q be a symmetric solution of problem (4.150)–(4.152) on [t0 , tf ]. Then, Propositions 4.29–4.32 yield the transformation (4.153) of Q to a perfect square on K. This completes the proof of Theorem 4.28. Now we can easily prove the following sufficient condition for the positive definiteness of the quadratic form on subspace K, which is the main result of this section. Theorem 4.33. If there exists a symmetric solution Q(t) of problem (4.150)–(4.152) on [t0 , tf ], then (z̄) is positive definite on K, i.e., there exists # > 0 such that (z̄) ≥ # γ̄ (z̄) t for all z̄ ∈ K, where γ̄ (z̄) = ξ̄ 2 + t0f ū(t), ū(t) dt. Proof. By Theorem 4.28, formula (4.153) holds for on K. Hence, the form is nonnegative on K. Let us show that is positive on K. Let z̄ ∈ K and (z̄) = 0. Then by (4.153) and (4.154), we have a ξ̄ + q− x̄ − = 0, (Q + Fux )x̄ + Fuu ū = 0 on [t0 , tf ]\. (4.165) Since z̄ ∈ K, we have x̄˙ = ū, and consequently −1 x̄˙ = −Fuu (Q + Fux )x̄ on [t0 , tf ]\. (4.166) Since x̄(t0 ) = 0, we have x̄(t) = 0 on [t0 , t∗ ). Hence x̄ − = 0. Then, by (4.165), we have ξ̄ = 0. Since [x̄] = [ẋ]ξ̄ = 0, we also have x̄ + = 0. Then, by (4.166), x̄(t) = 0 on (t∗ , tf ]. Thus, we obtain that ξ̄ = 0, x̄(t) ≡ 0 on [t0 , tf ], and then ū(t) = 0 a.e. on [t0 , tf ]. Thus z̄ = 0, and therefore is positive on K. But is a Legendre form, and therefore positiveness of on K implies positive definiteness of on K. Remark. Theorem 4.33 was proved for a single point of discontinuity of the control u(·). A similar result could be proved for finitely many points of discontinuity of the control u(·). Now, using Theorem 4.33, we can prove, for s = 1, that the existence of a symmetric solution Q(t) of the problem (4.150)–(4.152) on [t0 , tf ] is not only sufficient (as it was stated by Theorem 4.28) but also necessary for the positive definiteness of on K. This can 212 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals be done in different ways. We choose the following one. Let be positive definite on K, take a small # > 0, and put u(t) = u(t0 ) for t ∈ [t0 − #, t0 ]. Thus, u(t) is continued to the left-hand side of the point t0 by constant value equal to the value at t0 . Now we have w(t) defined on [t0 − #, tf ] with the same single discontinuity point t∗ , and therefore the “continued” quadratic form 1 1 # (z̄) = a ξ̄ 2 − 2[ψ̇]x̄av + 2 2 tf t0 −# Fww (t, w(t))w̄(t), w̄(t) dt is well defined on the subspace K# : x̄˙ = ū a.e. on [t0 − #, tf ], x̄(t0 − #) = x̄(tf ) = 0, [x̄] = [ẋ]ξ̄ . Using the same technique as in [96], one can easily prove that there is # > 0 such that # (z̄) is positive definite on K# (note that condition (SL) is satisfied for # on [t0 − #, tf ], which is important for this proof). Then by Theorem 4.33 applied for [t0 − #, tf ] there exists a solution Q of the Riccati equation (4.150) on (t0 − #, tf ] satisfying inequality (4.151) and jump condition (4.152), and hence we have this solution on the segment [t0 , tf ]. Thus, for s = 1 the following theorem holds. Theorem 4.34. The existence of a symmetric solution Q(t) of problem (4.150)–(4.152) together with condition (SL) is equivalent to the positive definiteness of on K. The formulation of the jump condition (4.152) allows us to solve problem (4.150)– (4.152) in a forward time direction. Certainly it is possible to move in opposite direction. Let us prove the following. Lemma 4.35. Conditions (4.151) and (4.152) imply that b+ := a + q+ [ẋ] > 0, (b+ )[Q] = (q+ )∗ (q+ ), (4.167) (4.168) q+ = [ẋ]∗ Q+ − [ψ̇]. (4.169) where Proof. By Proposition 4.26, jump condition (4.152) is equivalent to [Q](aI − [ẋ](q− )) = (q− )∗ (q− ). This equality implies that a[Q] = (q− )∗ (q− ) + [Q][ẋ](q− ). But [Q][ẋ] = (q+ − q− )∗ . (4.170) a[Q] = (q+ )∗ (q− ). (4.171) Hence Using again (4.170) in (4.171), we obtain a[Q] = (q+ )∗ (q+ − [ẋ]∗ [Q]). It follows from this formula that a[Q] = (q+ )∗ (q+ ) − (q+ )∗ [ẋ]∗ [Q]. Consequently (aI + (q+ )∗ [ẋ]∗ )[Q] = (q+ )∗ (q+ ). (4.172) 4.1. Jacobi-Type Conditions and Riccati Equation for Broken Extremals 213 Let us transform the left-hand side of this equation, using (4.171). Since 1 1 (q+ )∗ (q− ) = (q+ )∗ [ẋ]∗ (q+ )∗ (q− ) a a 1 1 = (q+ )∗ [ẋ]∗ (q+ )∗ (q− ) = (q+ )∗ (q− ) q+ [ẋ] a a = [Q] q+ [ẋ] , (q+ )∗ [ẋ]∗ [Q] = (q+ )∗ [ẋ]∗ we have (aI + (q+ )∗ [ẋ]∗ )[Q] = a[Q] + (q+ )∗ [ẋ]∗ [Q] = a[Q] + q+ [ẋ] [Q] = (a + q+ [ẋ])[Q]. Thus, (4.172) is equivalent to b+ [Q] = (q+ )∗ (q+ ). (4.173) Let us show that b+ > 0. From (4.173) it follows that 2 b+ q+ [Q](q+ )∗ = q+ (q+ )∗ . (4.174) By (4.151) and (4.152), we have q+ [Q](q+ )∗ ≥ 0. (4.175) If q+ = 0, then b+ = a + q+ [ẋ] = a > 0. Assume that q+ = 0. Then q+ (q+ )∗ 2 > 0. (4.176) Conditions (4.174), (4.175), and (4.176) imply that b+ > 0. Lemma 4.35 yields the transformation of on K to a perfect square expressed by q+ and b+ . Lemma 4.36. The following equality holds: ω∗ (ξ̄ , x̄) := a ξ̄ 2 − [ψ̇](x̄ − + x̄ + )ξ̄ − Q− x̄ − , x̄ − + Q+ x̄ + , x̄ k+ = 1 (a ξ̄ + q+ x̄ + )2 . b+ Proof. We have ω∗ (ξ̄ , x̄) = a ξ̄ 2 − [ψ̇](x̄ − + x̄ + )ξ̄ + Q+ x̄ + , x̄ + − Q− x̄ − , x̄ − = a ξ̄ 2 − [ψ̇](2x̄ − + [x̄])ξ̄ + Q+ (x̄ − + [x̄]), x̄ − + [x̄] − Q− x̄ − , x̄ − = a ξ̄ 2 − 2[ψ̇]x̄ − ξ̄ − [ψ̇] [ẋ] ξ̄ 2 + Q+ x̄ − , x̄ − + 2Q+ [x̄] , x̄ + + Q+ [x̄] , [x̄] − Q− x̄ − , x̄ − = (a − [ψ̇][ẋ] + Q+ [ẋ], [ẋ] )ξ̄ 2 + [Q]x̄ − , x̄ − + 2([ẋ]∗ Q+ − [ψ̇])x̄ − ξ̄ = (a + q+ [ẋ])ξ̄ 2 + 2q+ x̄ − ξ̄ + [Q]x̄ − , x̄ − . 214 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Hence ω∗ (ξ̄ , x̄) = b+ ξ̄ 2 + 2q+ x̄ − ξ̄ + [Q] x̄ − , x̄ − . Using (4.173), we obtain 1 2 2 (b ξ̄ + 2(q+ x̄ − )b+ ξ̄ + (q+ , x̄ − )2 ) b+ + 1 1 = (b+ ξ̄ + q+ x̄ − )2 = (a ξ̄ + q+ [x̄] + q+ x̄ − )2 b+ b+ 1 = (a ξ̄ + q+ x̄ + )2 . b+ ω∗ (ξ̄ , x̄) = Thus, we obtain the following. Theorem 4.37. If the assumptions of Theorem 4.28 are satisfied, then the quadratic form has the following transformation to a perfect square on K: tf B(t)v̄(t), v̄(t) dt, (4.177) 2(z̄) = (b+ )−1 (a ξ̄ + q+ x̄ + )2 + t0 where v̄ is defined by (4.154). 4.2 Riccati Equation for Broken Extremal in the General Problem of the Calculus of Variations Now we show that, by analogy with the simplest problem of the calculus of variations, the transformation of the quadratic form to a perfect square may be done in the auxiliary problem connected with the general problem of the calculus of variations (without local equality-type constraints and endpoints inequality-type constraints), and in this way we obtain, for this problem, sufficient optimality conditions for broken extremals in the terms of discontinuous solutions of the corresponding Riccati equation. Consider the problem J(x, u) = J (x0 , xf ) −→ min, K(x0 , xf ) = 0, ẋ = f (t, x, u), (x0 , xf ) ∈ P , (t, x, u) ∈ Q, (4.178) (4.179) (4.180) (4.181) where x0 = x(t0 ), xf = x(tf ), the interval [t0 , tf ] is fixed, P and Q are open sets, x ∈ Rn , u ∈ Rr , J and K are twice continuously differentiable on P , and f is twice continuously differentiable on Q. We set (x0 , xf ) = p, (x, u) = w. The problem is considered in the space W 1,1 ([t0 , tf ], Rn ) × L∞ ([t0 , tf ], Rr ). Let an admissible pair w(·) = (x(·), u(·)) be given. As in Section 2.1.1, we assume that u(·) is piecewise continuous and each point of discontinuity is a Lipschitz point (see Definition 2.1). We denote by = {t1 , . . . , ts } the set of all discontinuity points of the control u(·). For the problem (4.178)–(4.181), let us recall briefly the formulations of the sufficient optimality conditions at the point w(·), given in Chapter 2 for the general problem. Let us introduce the Pontryagin function H (t, x, u, ψ) = ψf (t, x, u), where ψ ∈ (Rn )∗ . Denote the endpoint Lagrange function by l(p, α0 , β) = α0 J (p) + βK(p), where β is a row vector of the same dimension as the vector K. Further, we introduce a collection of Lagrange multipliers λ = (α0 , β, ψ(·)) such that ψ(·) : [t0 , tf ] −→ (Rn )∗ is an absolutely continuous function continuously differentiable on each interval of the set [t0 , tf ]\. 4.2. Riccati Equation in the General Problem 215 We denote by 0 the set of the normed collections λ satisfying the conditions of the local minimum principle for an admissible trajectory w(·) = (x(·), u(·)), |βj | = 1, ψ̇ = −Hx , ψ(t0 ) = −lx0 , ψ(tf ) = lxf , Hu = 0, α0 ≥ 0, α0 + (4.182) where all derivatives are calculated on the trajectory w(·), respectively, at the endpoints (x(t0 ), x(tf )) of this trajectory. The condition 0 = ∅ is necessary for a weak minimum at the point w(·). We set U(t, x) = {u ∈ Rd(u) | (t, x, u) ∈ Q}. Denote by M0 the set of tuples λ ∈ 0 such that for all t ∈ [t0 , tf ]\, the condition u ∈ U(t, x 0 (t)) implies the inequality H (t, x 0 (t), u, ψ(t)) ≥ H (t, x 0 (t), u0 (t), ψ(t)). (4.183) If w0 is a point of Pontryagin minimum, then M0 is nonempty; i.e., the minimum principle holds. Further, denote by M0+ the set of λ ∈ M0 such that (a) (b) H (t, x 0 (t), u, ψ(t)) > H (t, x 0 (t), u0 (t), ψ(t)) for all t ∈ [t0 , tf ]\, u ∈ U(t, x 0 (t)), u = u0 (t) H (tk , x 0 (tk ), u, ψ(tk )) > H λk− = H λk+ for all tk ∈ , u ∈ U(tk , x 0 (tk )), u ∈ / {u0k− , u0k+ }. According to Definition 2.94, w 0 is a bounded strong minimum point if for any compact set C ⊂ Q there exists ε > 0 such that J(w) ≥ J(w 0 ) for all admissible trajectories w such that |x(t0 ) − x 0 (t0 )| < ε, max |x(t) − x 0 (t)| ≤ ε, t∈[t0 ,tf ] (t, w(t)) ∈ C a.e. on [t0 , tf ], where x is the vector composed of the essential components of the vector x. Recall that the component xi of a vector x = (x1 , . . . , xd(x) ) is said to be unessential if the function f is independent of xi and the functions J and K affinely depend on xi0 = xi (t0 ), xif = xi (tf ). Now we shall formulate quadratic sufficient conditions for a strict bounded strong minimum at the point w(·), which follow from the corresponding conditions in Section 2.7.2; see Theorem 2.102. For λ ∈ 0 and tk ∈ , we put (k H )(t) = H (t, x(t), uk+ , ψ(t))− H (t, x(t), uk− , ψ(t)), where uk− = u(tk − 0), uk+ = u(tk + 0), and D k (H ) := − d d (k H )t −0 = − (k H )t +0 . k k dt dt (4.184) The second equality is a consequence of the minimum principle. As we know, the following formula holds: D k (H ) = −Hxk+ Hψk− + Hxk− Hψk+ − [Ht ]k . (4.185) Recall that the conditions [H ]k = 0, D k (H ) ≥ 0, k = 1, . . . , s, also follow from the minimum principle. The same is true for the Legendre condition Huu (t, w(t), ψ(t))v̄, v̄ ≥ 0 a.e. on [t0 , tf ] for all v̄ ∈ Rr . Let us recall the strengthened Legendre (SL) condition for λ ∈ 0 : (a) for any t ∈ [t0 , tf ]\ the quadratic form Huu (t, w(t), ψ(t))v̄, v̄ is positive definite, (b) for any tk ∈ the quadratic forms Huu (tk , x(tk ), uk− , ψ(tk ))v̄, v̄ and Huu (tk , x(tk ), uk+ , ψ(tk ))v̄, v̄ are both positive definite. 216 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals An element λ = (α0 , β, ψ(·)) ∈ 0 is said to be strictly Legendrian if the following conditions are satisfied: (i) [H ]k = 0, k = 1, . . . , s; (ii) D k (H ) > 0, k = 1, . . . , s; (iii) strengthened Legendre (SL) condition. Assume that there exists a strictly Legendrian element λ ∈ 0 such that α0 > 0. In this case we put α0 = 1. Denote by Z2 () the space of triples z̄ = (ξ̄ , x̄(·), ū(·)) such that ξ̄ ∈ Rs , x̄(·) ∈ P W 1,2 ([t0 , tf ], Rn ), ū ∈ L2 ([t0 , tf ], Rr ). Let K denote the subspace of all z̄ = (ξ̄ , x̄(·), ū(·)) ∈ Z2 () such that Kpp̄ = 0, x̄˙ = fw w̄, [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, (4.186) where fw = fw (t, w(t)), Kp = Kp (x(t0 ), x(tf )), [ẋ]k is the jump of x(·) at tk , and [x̄]k is the jump of x̄(·) at tk . We call K the critical subspace. As in (2.13), we define the quadratic form, which corresponds to the element λ ∈ 0 , as 1 1 k k (z̄) = D (H )ξ̄k2 − 2[ψ̇]k x̄av ξ̄k + 2 2 s tf λ k=1 t0 1 Hww w̄, w̄ dt + lpp p̄, p̄ , 2 (4.187) where lpp = lpp (x(t0 ), x(tf ), α0 , β), Hww = Hww (t, w(t), ψ(t)). We set tf ū, ū dt. γ̄ (z̄) = ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) + t0 Theorem 2.102 implies the following assertion. Theorem 4.38. Assume that there exists a strictly Legendrian element λ ∈ M0+ such that α0 > 0 and the quadratic form λ (·) is positive definite on K, i.e., there exists # > 0 such that ∀ z̄ ∈ K. (4.188) λ (z̄) ≥ # γ̄ (z̄) Then, w(·) is a strict bounded strong minimum. Now, let us show that the quadratic form could be transformed into a perfect square if the corresponding Riccati equation has a solution Q(t) defined on [t0 , tf ], satisfying certain jump conditions at each point of the set . Define the Riccati equation along x(t), u(t), and ψ(t) by −1 Q̇ + Qfx + fx∗ Q + Hxx − (Hxu + Qfu )Huu (Hux + fu∗ Q) = 0, t ∈ [t0 , tf ]\, (4.189) for a piecewise continuous function Q(t) is continuously differentiable on each ∗ which interval of the set [t0 , tf ]\. Set qk− = [ẋ]k Qk− − [ψ̇]k and define the conditions bk− := ak − qk− [ẋ]k > 0, k = 1, . . . , s, (4.190) where ak = D k (H ), k = 1, . . . , s. Define the jump conditions for Q at the point tk ∈ as bk− [Q]k = (qk− )∗ (qk− ), k = 1, . . . , s. (4.191) 4.2. Riccati Equation in the General Problem 217 Theorem 4.39. Assume that there exists a symmetric solution Q(t) (piecewise continuous on [t0 , tf ] and continuously differentiable on each interval of the set [t0 , tf ]\) of Riccati equation (4.189) which satisfies at each point tk ∈ conditions (4.190) and jump conditions (4.191). Then the quadratic form λ (z̄) (see (4.187)) has the following transformation into a perfect square on the subspace K (see (4.186)), 2 (z̄) = λ s (bk− ) −1 ak ξ̄k + (qk− )x̄ k− 2 + tf t0 k=1 −1 Huu v̄, v̄ dt + M p̄, p̄ , (4.192) where v̄ = (Hux + fu∗ Q)x̄ + Huu ū, lx0 x0 + Q(t0 ) lx0 xf M= . lxf x0 lxf xf − Q(tf ) (4.193) (4.194) Proof. In perfect analogy with Theorem 4.28 we have to prove statements similar to Propositions 4.29–4.32. They need only small changes. Below we point out these changes. Take a symmetric matrix Q(t) on [t0 , tf ] with piecewise continuous coefficients, which are absolutely continuous on each interval of the set [t0 , tf ]\. Using, for z̄ ∈ K, formula (4.156) and the equalities ˙ x̄ = 2Qfx x̄, x̄ + 2Qfu ū, x̄ = (Qfx + fx∗ Q)x̄, x̄ + Qfu ū, x̄ + (fu∗ Qx̄, ū , 2Qx̄, we obtain (similar to (4.157)) the following zero form on K: 0= s (Qk+ x̄ k+ , x̄ k+ − Qk− x̄ k− , x̄ k− ) k=1 + tf t0 (Q̇ + fx∗ Q + Qfx )x̄, x̄ + fu∗ Qx̄, ū + Qfu ū, x̄ dt − Q(tf )x̄f , x̄f + Q(t0 )x̄0 , x̄0 . (4.195) Adding this zero form (4.195) to the form 2λ (z̄) (see (4.187)) considered for arbitrary z̄ ∈ K, we obtain tf 2λ (z̄) = M p̄, p̄ + (Hxx + Q̇ + Qfx + fx∗ Q)x̄, x̄ t0 + s +(Hxu + Qfu )ū, x̄ + (Hux + fu∗ Q)x̄, ū + Huu ū, ū dt ωk (ξ̄ , x̄), (4.196) k=1 where M is defined by (4.194) and k ωk (ξ̄ , x̄) = ak ξ̄k2 − 2[ψ̇]k x̄av ξ̄k + Qk+ x̄ k+ , x̄ k+ − Qk− x̄ k− , x̄ k− , k = 1, . . . , s. (4.197) According to formula (4.160), ωk (ξ̄ , x̄) = bk− ξ̄k2 + 2qk− x̄ k+ ξ̄k + [Q]k x̄ k+ , x̄ k+ , k = 1, . . . , s. (4.198) 218 Chapter 4. Jacobi-Type Conditions and Riccati Equation for Broken Extremals Further, by Proposition 4.30, ωk = (bk− )−1 (ak ξ̄k + qk− x̄ k− )2 . (4.199) Using the Riccati equation (4.189) and definition (4.193) for v̄, we obtain for the integral part of (4.196), tf t0 (Hxx + Q̇ + Qfx + fx∗ Q)x̄, x̄ + (Hxu + Qfu )ū, x̄ +(Hux + fu∗ Q)x̄, ū + Huu ū, ū dt tf −1 = Huu v̄, v̄ dt, (4.200) t0 namely, tf t0 −1 Huu v̄, v̄ dt = tf t0 = −1 −1 + 2Huu (Hux + fu∗ Q)x̄, Huu ū + Huu Huu ū, Huu ū dt tf t0 = −1 Huu (Hux + fu∗ Q)x̄, (Hux + fu∗ Q)x̄ tf t0 −1 (Hxu + Qfu )Huu (Hux + fu∗ Q)x̄, x̄ +2(Hux + fu∗ Q)x̄, ū + Huu ū, ū dt (Q̇ + Qfx + fx∗ Q + Hxx )x̄, x̄ +(Hux + fu∗ Q)x̄, ū + (Hxu + Qfu )ū, x̄ + Huu ū, ū dt. Equalities (4.196)–(4.200) imply the representation (4.192). Now we can easily prove the following theorem. Theorem 4.40. Assume that (a) there exists a symmetric solution Q(t), piecewise continuous on [t0 , tf ] and continuously differentiable on each interval of the set [t0 , tf ]\, of Riccati equation (4.189) which satisfies at each point tk ∈ conditions (4.190) and jump conditions (4.191). In addition, assume that (b) M p̄, p̄ ≥ 0 for all p̄ ∈ R2n such that Kpp̄ = 0. Also, assume that (c) the conditions Kpp̄ = 0 and M p̄, p̄ = 0 imply that x̄0 = 0 or x̄f = 0. Then the quadratic form λ (z̄) (see (4.187)) is positive definite on the subspace K (see (4.186)); i.e., condition (4.188) holds with some # > 0. Proof. From Theorem 4.39 it follows that λ (·) is nonnegative on K. Let us show that λ (·) is positive on K. Assume that (z̄) = 0 for some z̄ ∈ K. Then by formula (4.192), we have M p̄, p̄ = 0, v̄ = 0, ak ξ̄k = −qk− x̄ k− , k = 1, . . . , s. (4.201) (4.202) (4.203) 4.2. Riccati Equation in the General Problem 219 By assumption (c), we have x̄0 = 0 or x̄f = 0, since Kpp̄ = 0. (i) Let x̄0 = 0. By (4.202) and (4.193), we have −1 ū = −Huu (Hux + fu∗ Q)x̄. (4.204) Using this formula in the equality x̄˙ = fx x̄ + fu ū, we obtain −1 (Hux + fu∗ Q))x̄. x̄˙ = (fx − fu Huu (4.205) Together with the initial condition x̄(t0 ) = 0 this implies that x̄(t) = 0 for all t ∈ [t0 , t1 ). Hence x̄ 1− = 0. Then by (4.203), we obtain ξ̄1 = 0. Then [x̄]1 = 0 by the condition [x̄]1 = [ẋ]1 ξ̄1 . Hence x̄ 1+ = x̄ 1− + [x̄]1 = 0 and then again by (4.205) x̄(t) = 0 for all t ∈ (t1 , t2 ), etc. By induction, we get x̄(t) = 0 on [t0 , tf ], ξ̄ = 0, and then by (4.204), we get ū(t) = 0 a.e. on [t0 , tf ]. Thus z̄ = (ξ̄ , x̄, ū) = 0. Consequently, λ (·) is positive on K. (ii) Consider the case x̄f = 0. Then equation (4.205) and condition x̄(tf ) = 0 imply that x̄(t) = 0 for all t ∈ (ts , tf ]. Hence x̄ s+ = 0. Then [x̄]s = −x s− . Using this condition in (4.203), we obtain as ξ̄s − qs− [x̄]s = 0, or (as − qs− [ẋ]s )ξ̄s = 0, (4.206) because [x̄]s = [ẋ]s ξ̄s . Since bs− = as − qs− [ẋ]s > 0, condition (4.206) implies that ξ̄s = 0. Hence [x̄]s = 0 and then x̄ s− = 0. Then, by virtue of (4.205), x̄(t) = 0 on (ts−1 , ts ), etc. By induction we get x̄(·) = 0, ξ̄ = 0, and then, by (4.204), ū = 0, whence z̄ = 0. Thus, we have proved that is positive on K. It means that is positive definite on K, since is a Legendre form. Notes on SSC, Riccati equations, and sensitivity analysis. For regular controls, several authors have used the Riccati equation approach to verify SSC. Maurer and Pickenhain [73] considered optimal control problems with mixed control-state inequality constraints and derived SSC and the associated Riccati matrix equation on the basis of Klötzler’s duality theory. Similar results were obtained by Zeidan [116, 117]. Extensions of these results to control problems with free final time are to be found in Maurer and Oberle [68]. It is well known that SSC are fundamental for the stability and sensitivity analysis of parametric optimal control problems; cf., Malanowski and Maurer [58, 59], Augustin and Maurer [3], Maurer and Augustin [64, 65], and Maurer and Pesch [71, 72]. SSC also lay firm theoretical grounds to the method of determining neighboring extremals (Bryson and Ho [12] and Pesch [101]) and to real-time control techniques (see Büskens [13] and Büskens and Maurer [14, 15, 16]). Chapter 5 Second-Order Optimality Conditions in Optimal Control Problems Linear in a Part of Controls In this chapter, we derive quadratic optimality conditions for optimal control problems with a vector control variable having two components: a continuous unconstrained control appearing nonlinearly in the control system and a control appearing linearly and belonging to a convex polyhedron. It is assumed that the control components appearing linearly are of bang-bang type. In Section 5.1, we obtain quadratic conditions in the problem with continuous and bang-bang control components on a fixed time interval (that we call the main problem). The case of a nonfixed time interval is considered in Section 5.2. In Section 5.3, we show that, also for the mixed continuous-bang case, there exists a technique to check the positive definiteness of the quadratic form on the critical cone via a discontinuous solution of an associated Riccati equation with appropriate jump conditions at the discontinuity points of the bang-bang control [98]. In Section 5.4, this techniques is applied to an economic control problem in optimal production and maintenance. We show that the numerical solution obtained by Maurer, Kim, and Vossen [67] satisfies the second-order test derived in this chapter, while existing sufficiency results fail to hold. 5.1 Quadratic Optimality Conditions in the Problem on a Fixed Time Interval In this section, we obtain a necessary quadratic condition of a Pontryagin minimum and then show that a strengthening of this condition yields a sufficient condition of a bounded strong minimum. 5.1.1 The Main Problem Let x(t) ∈ Rd(x) denote the state variable, and let u(t) ∈ Rd(u) , v(t) ∈ Rd(v) denote the control variables in the time interval t ∈ [t0 , tf ] with fixed initial time t0 and fixed final time tf . We shall refer to the following optimal control problem (5.1)–(5.4) as the main problem: Minimize J(x(·), u(·), v(·)) = J (x(t0 ), x(tf )) 223 (5.1) 224 Chapter 5. Second-Order Optimality Conditions in Optimal Control subject to the constraints F (x(t0 ), x(tf )) ≤ 0, K(x(t0 ), x(tf )) = 0, (x(t0 ), x(tf )) ∈ P , ẋ(t) = f (t, x(t), u(t), v(t)), u(t) ∈ U , (t, x(t), v(t)) ∈ Q, (5.2) (5.3) where the control variable u appears linearly in the system dynamics, f (t, x, u, v) = a(t, x, v) + B(t, x, v)u. (5.4) Here, F , K, a are column vector functions, B is a d(x) × d(u) matrix function, P ⊂ R2d(x) , Q ⊂ R1+d(x)+d(v) are open sets, and U ⊂ Rd(u) is a convex polyhedron. The functions J , F , K are assumed to be twice continuously differentiable on P , and the functions a, B are twice continuously differentiable on Q. The dimensions of F , K are denoted by d(F ), d(K). By = [t0 , tf ] we denote the interval of control and use the abbreviations x0 = x(t0 ), xf = x(tf ), p = (x0 , xf ). A process T = {(x(t), u(t), v(t)) | t ∈ [t0 , tf ] } is said to be admissible if x(·) is absolutely continuous, u(·), v(·) are measurable bounded on , and the triple of functions (x(t), u(t), v(t)), together with the endpoints p = (x(t0 ), x(tf )), satisfies the constraints (5.2) and (5.3). Thus, the main problem is considered in the space W := W 1,1 ([t0 , tf ], Rd(x) ) × L∞ ([t0 , tf ], Rd(u) ) × L∞ ([t0 , tf ], Rd(v) ). Definition 5.1. An admissible process T pact set C ⊂ Q there exists ε > 0 such T˜ = {(x̃(t), ũ(t), ṽ(t)) | t ∈ [t0 , tf ] } such u(t)| dt < ε, (c) (t, x̃(t), ṽ(t)) ∈ C a.e. on 5.1.2 affords a Pontryagin minimum if for each comthat J(T˜ ) ≥ J(T ) for all admissible processes that (a) max |x̃(t) − x(t)| < ε, (b) |ũ(t) − . First-Order Necessary Optimality Conditions Let T = {(x(t), u(t), v(t) | t ∈ [t0 , tf ] } be a fixed admissible process such that the control u(t) is a piecewise constant function taking all its values in the vertices of polyhedron U , and the control v(t) is a continuous function on the interval = [t0 , tf ]. Denote by = {t1 , . . . , ts }, t0 < t1 < · · · < ts < tf the finite set of all discontinuity points (jump points) of the control u(t). Then ẋ(t) is a piecewise continuous function whose discontinuity points belong to , and hence x(t) is a piecewise smooth function on . We use the notation [u]k = uk+ − uk− to denote the jump of the function u(t) at the point tk ∈ , where uk− = u(tk −), uk+ = u(tk +) are the left- and right-hand values of the control u(t) at tk , respectively. Similarly, we denote by [ẋ]k the jump of the function ẋ(t) at the point tk . In addition, assume that the control v(t) satisfies the following Condition Lθ : There exist constants C > 0 and ε > 0 such that for each point tk ∈ we have |v(t) − v(tk )| ≤ C|t − tk | for all t ∈ (tk − ε, tk + ε). In the literature, one often uses the “Pontryagin minimum principle” instead of the “Pontryagin maximum principle.” In order to pass from one principle to the other, one has to change only the signs of the adjoint variables ψ and ψ0 . It leads to obvious changes in the signs of the Hamiltonian, transversality conditions, Legendre conditions, and quadratic forms. Let us introduce the Pontryagin function (or Hamiltonian) H (t, x, u, v, ψ) = ψf (t, x, u, v) = ψa(t, x, v) + ψB(t, x, v)u, (5.5) 5.1. Quadratic Optimality Conditions in the Problem on a Fixed Time Interval 225 where ψ is a row vector of dimension d(ψ) = d(x), while x, u, f , F , and K are column vectors. The row vector of dimension d(u), φ(t, x, v, ψ) = ψB(t, x, v), (5.6) will be called the switching function for the u-component of the control. Denote by l the endpoint Lagrange function l(p, α0 , α, β) = α0 J (p) + αF (p) + βK(p), where α and β are row vectors with d(α) = d(F ) and d(β) = d(K), and α0 is a number. We introduce a tuple of Lagrange multipliers λ = (α0 , α, β, ψ(·)) such that ψ(·) : → Rd(x) is continuous on and continuously differentiable on each interval of the set \ . In what follows, we will denote first- or second-order partial derivatives by subscripts referring to the variables. Denote by M0 the set of the normalized tuples λ satisfying the minimum principle conditions for the process T : α0 ≥ 0, α ≥ 0, αF (p) = 0, α0 + d(F ) i=1 αi + d(K) |βj | = 1, j =1 ψ̇ = −Hx ∀ t ∈ \ , ψ(t0 ) = −lx0 , ψ(tf ) = lxf , H (t, x(t), u, v, ψ(t)) ≥ H (t, x(t), u(t), v(t), ψ(t)) ∀ t ∈ \ , u ∈ U , v ∈ Rd(v) such that (t, x(t), v) ∈ Q. (5.7) (5.8) (5.9) (5.10) The derivatives lx0 and lxf are taken at the point (p, α0 , α, β), where p = (x(t0 ), x(tf )), and the derivative Hx is evaluated along the trajectory (t, x(t), u(t), v(t), ψ(t)), t ∈ \ . The condition M0 = ∅ constitutes the first-order necessary condition of a Pontryagin minimum for the process T which is called the Pontryagin minimum principle, cf. Pontryagin et al. [103], Hestenes [40], and Milyutin and Osmolovskii [79]. The set M0 is a finite-dimensional compact set and the projector λ → (α0 , α, β) is injective on M0 . In the following, it will be convenient to use the simple abbreviation (t) for indicating all arguments (t, x(t), u(t), v(t), ψ(t)), e.g., H (t) = H (t, x(t), u(t), v(t), ψ(t)), φ(t) = φ(t, x(t), v(t), ψ(t)). Let λ = (α0 , α, β, ψ(·)) ∈ M0 . It is well known that H (t) is a continuous function. In particular, [H ]k = H k+ − H k− = 0 holds for each tk ∈ , where H k− := H (tk − 0) and H k+ := H (tk + 0). We denote by H k the common value of H k− and H k+ . For λ ∈ M0 and tk ∈ consider the function (k H )(t) = H (t, x(t), uk+ , v(tk ), ψ(t)) − H (t, x(t), uk− , v(tk ), ψ(t)) = φ(t, x(t), v(tk ), ψ(t)) [u]k . Proposition 5.2. For each λ ∈ M0 , the following equalities hold: d d (k H )t=t −0 = (k H )t=t +0 , k k dt dt k = 1, . . . , s. (5.11) 226 Chapter 5. Second-Order Optimality Conditions in Optimal Control Consequently, for each λ ∈ M0 , the function (k H )(t) has a derivative at the point tk ∈ . In what follows, we will consider the quantities D k (H ) = − d (k H )(tk ) = −φ̇(tk ±)[u]k , dt k = 1, . . . , s. (5.12) Then the minimum condition (5.10) implies the following property. Proposition 5.3. For each λ ∈ M0 , the following conditions hold: D k (H ) ≥ 0, k = 1, . . . , s. Note that the value D k (H ) also can be written in the form D k (H ) = −Hxk+ Hψk− + Hxk− Hψk+ − [Ht ]k = ψ̇ k+ ẋ k− − ψ̇ k− ẋ k+ + [ψ̇0 ]k , where Hxk− and Hxk+ are the left-hand and right-hand values of the function Hx (t) at tk , respectively, [Ht ]k is the jump of the function Ht (t) := Ht (t, x(t), u(t), v(t), ψ(t)) at tk , etc., and ψ0 (t) = −H (t). 5.1.3 Bounded Strong Minimum As in Section 10.1, we define essential (or main) and unessential (or complementary) state variables in the problem. The state variable xi , i.e., the ith component of the state vector x, is called unessential if the function f does not depend on xi and the functions F , J , and K are affine in xi0 = xi (t0 ) and xi1 = xi (tf ); otherwise the variable xi is called essential. Let x denote the vector of all essential components of state vector x. Definition 5.4. The process T affords a bounded strong minimum if for each compact set C ⊂ Q there exists ε > 0 such that J(T+) ≥ J(T ) for all admissible processes T+ = {(x̃(t), ũ(t), ṽ(t)) | t ∈ [t0 , tf ] } such that (a) |x̃(t0 ) − x(t0 )| < ε, (b) max |x̃(t) − x(t)| < ε, (c) (t, x̃(t), ṽ(t)) ∈ C a.e. on . The strict bounded strong minimum is defined in a similar way, with the nonstrict inequality J(T+) ≥ J(T ) replaced by the strict one and the process T+ required to be different from T . 5.1.4 Critical Cone For a given process T , we introduce the space Z2 () and the critical cone K ⊂ Z2 (). As in Section 2.1.5, we denote by P W 1,2 (, Rd(x) ) the space of piecewise continuous functions x̄(·) : → Rd(x) , which are absolutely continuous on each interval of the set \ and have a square integrable first derivative. For each x̄ ∈ P W 1,2 (, Rd(x) ) and for tk ∈ , we set x̄ k− = x̄(tk −), x̄ k+ = x̄(tk +), [x̄]k = x̄ k+ − x̄ k− . Let z̄ = (ξ̄ , x̄, v̄), where ξ̄ ∈ Rs , x̄ ∈ P W 1,2 (, Rd(x) ), v̄ ∈ L2 (, Rd(v) ). Thus, z̄ ∈ Z2 () := Rs × P W 1,2 (, Rd(x) ) × L2 (, Rd(v) ). 5.1. Quadratic Optimality Conditions in the Problem on a Fixed Time Interval 227 For each z̄, we set x̄0 = x̄(t0 ), x̄f = x̄(tf ), p̄ = (x̄0 , x̄f ). (5.13) The vector p̄ is considered a column vector. Denote by IF (p) = {i ∈ {1, . . . , d(F )} | Fi (p) = 0} the set of indices of all active endpoint inequalities Fi (p) ≤ 0 at the point p = (x(t0 ), x(tf )). Denote by K the set of all z̄ ∈ Z2 () satisfying the following conditions: J (p)p̄ ≤ 0, Fi (p)p̄ ≤ 0 ∀ i ∈ IF (p), ˙ = fx (t)x̄(t) + fv (t)v̄(t), x̄(t) [x̄] = [ẋ] ξ̄k , k k K (p)p̄ = 0, k = 1, . . . , s, (5.14) (5.15) (5.16) where p = (x(t0 ), x(tf )) and [ẋ]k = ẋ(tk + 0) − ẋ(tk − 0). It is obvious that K is a convex cone in the Hilbert space Z2 () with finitely many faces. We call K the critical cone. Note that the variation ū(t) of the bang-bang control u(t) vanishes in the critical cone. 5.1.5 Necessary Quadratic Optimality Conditions Let us introduce a quadratic form on the critical cone K defined by the conditions (5.14)– (5.16). For each λ ∈ M0 and z̄ ∈ K, we set3 (λ, z̄) = s k lpp (p)p̄, p̄ + (D k (H )ξ̄k2 − 2[ψ̇]k x̄av ξ̄k ) k=1 tf Hxx (t)x̄(t), x̄(t) + 2Hxv (t)v̄(t), x̄(t) + t0 + Hvv (t)v̄(t), v̄(t) dt, (5.17) where lpp (p) = lpp (α0 , α, β, p), p = (x(t0 ), x(tf )), 1 k x̄av = (x̄ k− + x̄ k+ ), 2 Hxx (t) = Hxx (t, x(t), u(t), v(t), ψ(t)), etc. Note that the functional (λ, z̄) is linear in λ and quadratic in z̄. The following theorem gives the main second-order necessary condition of optimality. Theorem 5.5. If the process T affords a Pontryagin minimum, then the following Condition A holds: The set M0 is nonempty and maxλ∈M0 (λ, z̄) ≥ 0 for all z̄ ∈ K. We call Condition A the necessary quadratic condition, although it is truly quadratic only if M0 is a singleton. 3 In Part II of the book, we will not use the factor 1/2 in the definition of the quadratic form . 228 Chapter 5. Second-Order Optimality Conditions in Optimal Control 5.1.6 Sufficient Quadratic Optimality Conditions A natural strengthening of the necessary Condition A turns out to be a sufficient optimality condition not only for a Pontryagin minimum, but also for a bounded strong minimum; cf. Definition 5.4. Denote by M0+ the set of all λ ∈ M0 satisfying the following conditions: H (t, x(t), u, v, ψ(t)) > H (t, x(t), u(t), v(t), ψ(t)) for all t ∈ \ , u ∈ U , v ∈ Rd(v) such that (t, x(t), v) ∈ Q and (u, v) = (u(t), v(t)) ; (b) H (tk , x(tk ), u, v, ψ(tk )) > H k for all tk ∈ , u ∈ U , v ∈ Rd(v) such that (tk , x(tk ), v) ∈ Q, (u, v) = (u(tk −), v(tk )), (u, v) = (u(tk +), v(tk )), where H k := H k− = H k+ . Let Arg minũ∈U φ ũ be the set of points u ∈ U where the minimum of the linear function φ ũ is attained. (a) Definition 5.6. For a given admissible process T with a piecewise constant control u(t) and continuous control v(t), we say that u(t) is a strict bang-bang control if the set M0 is nonempty and there exists λ ∈ M0 such that Arg min φ(t)ũ = [u(t−), u(t+)] ∀ t ∈ [t0 , tf ], ũ∈U where [u(t−), u(t+)] denotes the line segment spanned by the vectors u(t−), u(t+). If dim(u) = 1, then the strict bang-bang property is equivalent to φ(t) = 0 ∀ t ∈ \ . It is easy to show that if the set M0+ is nonempty, then u(t) is a strict bang-bang control. Definition 5.7. An element λ ∈ M0 is said to be strictly Legendre if the following conditions are satisfied: (a) For each t ∈ \ the quadratic form Hvv (t, x(t), u(t), v(t), ψ(t))v̄, v̄ is positive definite on Rd(v) ; (b) for each tk ∈ the quadratic form Hvv (tk , x(tk ), u(tk −), v(tk ), ψ(tk ))v̄, v̄ is positive definite on Rd(v) ; (c) for each tk ∈ the quadratic form Hvv (tk , x(tk ), u(tk +), v(tk ), ψ(tk ))v̄, v̄ is positive definite on Rd(v) ; (d) D k (H ) > 0 for all tk ∈ . Denote by Leg+ (M0+ ) the set of all strictly Legendrian elements λ ∈ M0+ , and set γ̄ (z̄) = ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) + tf v̄(t), v̄(t) dt. t0 Theorem 5.8. Let the following Condition B be fulfilled for the process T : (a) The set Leg+ (M0+ ) is nonempty; 5.1. Quadratic Optimality Conditions in the Problem on a Fixed Time Interval 229 there exists a nonempty compact set M ⊂ Leg+ (M0+ ) and a number C > 0 such that maxλ∈M (λ, z̄) ≥ C γ̄ (z̄) for all z̄ ∈ K. Then T is a strict bounded strong minimum. (b) Remark 5.9. If the set Leg+ (M0+ ) is nonempty and K={0}, then Condition (b) is fulfilled automatically. This case can be considered as a first-order sufficient optimality condition for a strict bounded strong minimum. As mentioned in the introduction, the proof of Theorem 5.8 is very similar to the proof of the sufficient quadratic optimality condition for the pure bang-bang case given in Milyutin and Osmolovskii [79, Theorem 12.4, p. 302], and based on the sufficient quadratic optimality condition for broken extremals in the general problem of calculus of variations; see Part I of the present book. The proofs of Theorems 5.5 and 5.8 will be given below. 5.1.7 Proofs of Quadratic Conditions in the Problem on a Fixed Time Interval Problem Z and its convexification with respect to bang-bang control components. Consider the following optimal control problem, which is similar to the main problem (5.1)–(5.4): Minimize J(x(·), u(·), v(·)) = J (x(t0 ), x(tf )) (5.18) subject to the constraints F (x(t0 ), x(tf )) ≤ 0, K(x(t0 ), x(tf )) = 0, (x(t0 ), x(tf )) ∈ P , ẋ = a(t, x, v) + B(t, x, v)u, u ∈ U, (t, x, v) ∈ Q, (5.19) (5.20) where P ⊂ R2d(x) , Q ⊂ R1+d(x)+d(v) are open sets, U ⊂ Rd(u) is a compact set, and = [t0 , tf ] is a fixed time interval. The functions J , F , and K are assumed to be twice continuously differentiable on P , and the functions a and B are twice continuously differentiable on Q. The compact set U is specified by U = {u ∈ Qg | g(u) = 0}, (5.21) where Qg ⊂ Rd(u) is an open set and g : Qg → Rd(g) is a twice continuously differentiable function satisfying the full-rank condition rank gu (u) = d(g) (5.22) for all u ∈ Qg such that g(u) = 0. It follows from (5.22) that d(g) ≤ d(u), but it is possible, in particular, that d(g) = d(u). In this latter case for U we can take any finite set of points in Rd(u) , for example, the set of vertices of a convex polyhedron. For brevity, we will refer to problem (5.18)–(5.20) as the problem Z. Thus, the only difference between problem Z and the main problem (5.1)–(5.4) is that a convex polyhedron U is replaced by an arbitrary compact set U specified by (5.21). The definitions of the Pontryagin minimum, the bounded strong minimum, and the strict bounded strong minimum in the problem Z are the same as in the main problem. 230 Chapter 5. Second-Order Optimality Conditions in Optimal Control We will consider the problem (5.18), (5.19) not only for control system (5.20), but also for its convexification with respect to u: ẋ = a(t, x, v) + B(t, x, v)u, u ∈ co U, (t, x, v) ∈ Q, (5.23) where co U is the convex hull of the compact set U. We will refer to the problem (5.18), (5.19), (5.23) as the problem co Z. We will be interested in the relationships between conditions for a minimum in the problems Z and co Z. Naturally, these conditions concern a process satisfying the constraints of the problem Z. Let w 0 (·) = (x 0 (·), u0 (·), v 0 (·)) be such a process. Then it satisfies the constraints of the problem co Z as well. We assume that w0 satisfy the following conditions: The function u0 (t) is piecewise continuous with the set = {t1 , . . . , ts } of discontinuity points, control v 0 (t) is continuous, and each point tk ∈ is an L-point of the controls u0 (t) and v 0 (t). The latter means that there exist constants C > 0 and ε > 0 such that for each point tk ∈ , we have |u0 (t) − u0k− | ≤ C|t − tk | |u0 (t) − u0k+ | ≤ C|t − tk | |v 0 (t) − v 0 (tk )| ≤ C|t − tk | for for for t ∈ (tk − ε, tk ), t ∈ (tk , tk + ε), t ∈ (tk − ε, tk + ε). We can formulate quadratic conditions for the point w0 in the problem Z. To what extent can they be carried over to the problem co Z? Regarding the necessary quadratic conditions (see Condition A in Theorem 3.9) this is a simple question. If a point w 0 yields a Pontryagin minimum in the problem co Z, this point, a fortiori, affords a Pontryagin minimum in the problem Z. Hence any necessary condition for a Pontryagin minimum at the point w0 in the problem Z is a necessary condition for a Pontryagin minimum at this point in the problem co Z as well. Thus we have the following theorem. Theorem 5.10. Condition A (given by Theorem 3.9) for the point w 0 in the problem Z is a necessary condition for a Pontryagin minimum at this point in the problem co Z. Bounded strong γ1 -sufficiency in problem Z and bounded strong minimum in problem co Z. We now turn to derivation of quadratic sufficient conditions in the problem co Z, which is a more complicated task. As we know, Condition B for the point w 0 in the problem Z ensures a Pontryagin, and even a bounded strong, minimum in this problem. Does it ensure a bounded strong or at least a Pontryagin minimum at the point w 0 in the convexified problem co Z? There are examples where the convexification results in the loss of the minimum. The bounded strong minimum is not stable with respect to the convexification operation. However, some stronger property in the problem Z, which is called bounded strong γ1 sufficiency, ensures a bounded strong minimum in the problem co Z. This property will be defined below. For the problem Z and the point w0 (·) define the violation function 0 + d(F ) (Fi (p))+ + |K(p)| σ (w) = (J (p) − J (p )) + i=1 tf + |ẋ(t) − f (t, x(t), u(t), v(t))| dt, t0 (5.24) 5.1. Quadratic Optimality Conditions in the Problem on a Fixed Time Interval 231 where f (t, x, u, v) = a(t, x, v) + B(t, x, v)u and a + = max{a, 0} for a ∈ R1 . For an arbitrary variation δw = (δx, δu, δv) ∈ W , let γ1 (δw) = ( δu 1 )2 , (5.25) tf t0 |δu(t)| dt. where δu 1 = n Let {w } = {(x n , un , v n )} be a bounded sequence in W . In what follows, the notation n σ (δw ) = o(γ1 (w n − w 0 )) means that there exists a sequence of numbers εn → 0 such that σ (wn ) = εn γ1 (w n − w0 ) (even in the case where γ1 (w n − w0 ) does not tend to zero). Definition 5.11. We say that the bounded strong γ1 -sufficiency holds at the point w0 in the problem Z if there are no compact set C ⊂ Q and sequence {wn } = {(x n , un , v n )} in W such that σ (wn ) = o(γ1 (wn − w 0 )), max |x n (t) − x 0 (t)| → 0, t∈ |x n (t0 ) − x 0 (t0 )| → 0, and for all n, w n = w 0 , (t, x n (t), v n (t)) ∈ C, un (t) ∈ U a.e. on , where x n is composed of the essential components of vector x n . The following proposition holds for the problem Z. Proposition 5.12. A bounded strong γ1 -sufficiency at the point w0 implies a strict bounded strong minimum at this point. Proof. Let w 0 be an admissible point in the problem Z. Assume that w0 is not a point of a strict bounded strong minimum. Then there exist a compact set C ⊂ Q and a sequence of admissible points {wn } = {(x n , un , v n )} such that maxt∈ |x n (t) − x 0 (t)| → 0, |x n (t0 ) − x 0 (t0 )| → 0 and, for all n, J (p n ) − J (p 0 ) < 0, (t, x n (t), v n (t)) ∈ C, un (t) ∈ U a.e. on , wn = w0 . It follows that σ (w n ) = 0 for all n. Hence w0 is not a point of a bounded strong γ1 -sufficiency in the problem Z. A remarkable fact is that the convexification of the constraint u ∈ U turns a bounded strong γ1 -sufficiency into at least a bounded strong minimum. Theorem 5.13. Suppose that for an admissible point w 0 in the problem Z the bounded strong γ1 -sufficiency holds. Then w 0 is a point of the strict bounded strong minimum in the problem co Z. The proof of Theorem 5.13 is based on the following lemma. Lemma 5.14. Let U ⊂ Rd(u) be a bounded set, and let u(t) be a measurable function on such that u(t) ∈ co U a.e. on . Then there exists a sequence un (t) of measurable functions on such that for every n we have un (t) ∈ U , t ∈ , and t t un (τ ) dτ → u(τ ) dτ uniformly in t ∈ . (5.26) t0 t0 232 Chapter 5. Second-Order Optimality Conditions in Optimal Control Moreover, if u0 (t) is a bounded measurable function on such that meas{t ∈ | u(t) = u0 (t)} > 0, then (5.26) implies (5.27) |un (t) − u0 (t)| dt > 0. lim inf n (5.28) This lemma is a consequence of Theorem 16.1 in [79, Appendix, p. 361]. In the proof of Theorem 5.13 we will also use the following theorem, which is similar to Theorem 16.2 in [79, Appendix, p. 366]. Theorem 5.15. Let w ∗ = (x ∗ , u∗ , v ∗ ) be a triple in W satisfying the control system ẋ = a(t, x, v) + B(t, x, v)u, (t, x, v) ∈ Q, x(t0 ) = c0 , where Q, a, and B are the same as in the problem Z, and c0 ∈ Rd(x) . Suppose that there is a sequence un ∈ L∞ ([t0 , tf ], Rd(u) ) such that sup un n and, for each t ∈ [t0 , tf ], t un (τ ) dτ → t0 t ∞ < +∞ u∗ (τ ) dτ (5.29) (n → ∞). (5.30) t0 Then, for all sufficiently large n, the system ẋ = a(t, x, v ∗ (t)) + B(t, x, v ∗ (t))un (t), (t, x, v ∗ (t)) ∈ Q, x(t0 ) = c0 (5.31) has the unique solution x n (t) on [t0 , tf ] and max |x n (t) − x ∗ (t)| → 0 [t0 ,tf ] (n → ∞). (5.32) Proof. Consider the equations ẋ n = a(t, x n , v ∗ ) + B(t, x n , v ∗ )un , ẋ ∗ = a(t, x ∗ , v ∗ ) + B(t, x ∗ , v ∗ )u∗ . Their difference can be written as δ ẋ = δa + (δB)un + B(t, x ∗ , v ∗ )δu, where (5.33) δx = x n − x ∗ , δu = un − u∗ , δa = a(t, x n , v ∗ ) − a(t, x ∗ , v ∗ ) = a(t, x ∗ + δx, v ∗ ) − a(t, x ∗ , v ∗ ), δB = B(t, x n , v ∗ ) − B(t, x ∗ , v ∗ ) = B(t, x ∗ + δx, v ∗ ) − B(t, x ∗ , v ∗ ). We have here δx(t0 ) = 0. (5.34) 5.1. Quadratic Optimality Conditions in the Problem on a Fixed Time Interval 233 It follows from (5.29) and (5.30) that lim sup δu ∞ < +∞, t δu(τ ) dτ → 0 ∀ t ∈ [t0 , tf ]. (5.35) t0 These properties mean that the sequence {δu} converges to zero ∗-weakly (L1 -weakly) in L∞ . Therefore the functions t δy(t) := B(τ , x ∗ (τ ), v ∗ (τ ))δu(τ ) dτ t0 converge to zero pointwise on [t0 , tf ], and hence uniformly on [t0 , tf ] because they possess the common Lipschitz constant. Thus δy Using the relations C → 0. δ ẏ = B(t, x ∗ , v ∗ )δu, (5.36) δy(t0 ) = 0, (5.37) rewrite (5.33) as δ ẋ − δ ẏ = δa + (δB)un . (5.38) Set δz = δx − δy. Then (5.38) can be represented in the form δ ż = δa + (δB)un , where δa = a(t, x ∗ + δy + δz, v ∗ ) − a(t, x ∗ , v ∗ ), Thus δ ż δz(t0 ) δB = B(t, x ∗ + δy + δz, v ∗ ) − B(t, x ∗ , v ∗ ). a(t, x ∗ + δy + δz, v ∗ ) − a(t, x ∗ , v ∗ ) + (B(t, x ∗ + δy + δz, v ∗ ) − B(t, x ∗ , v ∗ )) un , = 0, (t, x ∗ + δy + δz, v ∗ ) ∈ Q. = (5.39) This system is equivalent to the system (5.31) in the following sense: x n solves (5.31) iff x n = x ∗ + δy + δz, (5.40) where δy satisfies the conditions δ ẏ = B(t, x ∗ , v ∗ )(un − u∗ ), δy(t0 ) = 0, (5.41) and δz solves the system (5.39). Hence it suffices to show that for all sufficiently large n the system (5.39) has a unique solution, where δy is determined by (5.41). This is so because for δy ≡ 0 the system (5.39) has a unique solution δz ≡ 0 on [t0 , tf ] and moreover (5.36) holds. From (5.36) and representation (5.39) it follows also that δz C → 0 and hence δx C ≤ δy C + δz C → 0. Therefore (5.32) holds. 234 Chapter 5. Second-Order Optimality Conditions in Optimal Control Proof of Theorem 5.13. Suppose that w 0 is not a strict bounded strong minimum point in the problem co Z. Then there exist a compact set C ⊂ Q and a sequence {w n } = {(x n , un , v n )} such that for all n one has wn = w0 and (x n (t0 ), x n (tf )) ∈ P , (t, x n (t), v n (t)) ∈ C a.e. on , un (t) ∈ co U a.e. on , σ (w n ) = 0, max |x n (t) − x 0 (t)| → 0, |x n (t0 ) − x 0 (t0 )| → 0. t∈ (5.42) (5.43) (5.44) (5.45) We will show that w0 is not a point of the bounded strong γ1 -sufficiency in the problem Z. If un = u0 for infinitely many terms, w0 is not even a strict bounded strong minimum point in the problem Z. Hence, we assume that un = u0 for all n. Apply Lemma 5.14 to each function un . By virtue of this lemma there exists a sequence of measurable functions {unk }∞ k=1 such that, for all k, unk (t) ∈ U ∀ t ∈ , t t nk u (τ ) dτ → un (τ ) dτ uniformly in t ∈ (as k → ∞), t0 (5.47) t0 tf lim inf k→∞ (5.46) |unk (t) − u0 (t)| dt > 0. (5.48) t0 For each k define x nk as the solution of the system ẋ nk = a(t, x nk , v n ) + B(t, x nk , v n )unk , x nk (t0 ) = x n (t0 ). (5.49) According to Theorem 5.15 this system has a solution for all sufficiently large k and x nk − x n C → 0 as k → ∞. It follows from (5.44) and (5.50) that (Fi (pnk ))+ + |K(pnk )| → 0 as k → ∞, (J (p nk ) − J (p0 ))+ + (5.50) (5.51) i where p nk = (x nk (t0 ), x nk (tf )) = (x n (t0 ), x nk (tf )). Combined with (5.49) this implies that σ (wnk ) → 0 as k → ∞, (5.52) where w nk = (x nk , unk , v n ). It follows from (5.48) and (5.52) that there is a number k = k(n) such that tf 2 1 |unk (t) − u0 (t)| dt ∀ k ≥ k(n). (5.53) σ (wnk ) ≤ n t0 By (5.50), k(n) also can be chosen so that x nk(n) − x n C ≤ 1 . n (5.54) 5.1. Quadratic Optimality Conditions in the Problem on a Fixed Time Interval 235 Then by virtue of (5.53), for the sequence {w nk(n) } one has σ (wnk(n) ) = o(γ1 (w nk(n) − w0 )). (5.55) Moreover, we have for this sequence unk(n) (t) ∈ U ∀ t ∈ , x nk(n) − x 0 C ≤ x nk(n) − x n C + x n − x 0 C 1 ≤ + x n − x 0 C → 0 (n → ∞), n |x nk(n) (t0 ) − x 0 (t0 )| = |x n (t0 ) − x 0 (t0 )| → 0. (5.56) (5.57) (5.58) Finally, from (5.42) and (5.54) it follows that there exists a compact set C1 such that C ⊂ C1 ⊂ Q, and for all sufficiently large n we have (x nk(n) (t0 ), x nk(n) (tf )) ∈ P , (t, x nk(n) (t), v n (t)) ∈ C1 a.e. on . (5.59) The existence of a compact set C1 ⊂ Q and a sequence {w nk(n) } in the space W satisfying (5.55)–(5.59) means that the bounded strong γ1 -sufficiency fails at the point w0 in the problem Z. Now we can prove the following theorem. Theorem 5.16. Condition B (given in Definition 3.15) for the point w0 in the problem Z is a sufficient condition for a strict bounded strong minimum at this point in the problem co Z. Proof. Assume that Condition B for the point w0 in the problem Z is satisfied. Then by Theorem 3.16 a bounded strong γ -sufficiency holds in problem Z at the point w 0 . The latter means (see Definition 3.11) that there are no compact set C ⊂ Q and sequence {wn } = {(x n , un , v n )} in W such that σ (w n ) = o(γ (w n − w0 )), max |x n (t) − x 0 (t)| → 0, t∈ |x n (t0 ) − x 0 (t0 )| → 0, and for all n we have wn = w0 , (t, x n (t), v n (t)) ∈ C, g(un (t)) = 0, un (t) ∈ Qg a.e. on , where γ is the higher order defined in Definition 2.17. Note that for each n the conditions g(un (t)) = 0, un (t) ∈ Qg mean that un (t) ∈ U, where U is a compact set. It follows from Proposition 2.98 that for any compact set C ⊂ Q there exists a constant C > 0 such that for any w = (x, u, v) ∈ W satisfying the conditions u(t) ∈ U and (t, x(t), v(t)) ∈ C a.e. on , we have γ1 (w − w0 ) ≤ Cγ (w − w0 ). By virtue of this inequality, a bounded strong γ -sufficiency at the point w 0 in problem Z implies a bounded strong γ1 -sufficiency at w 0 in the same problem. Then by Theorem 5.13, w0 is a point of a strict bounded strong minimum in problem co Z. Proofs of Theorems 5.5 and 5.8. Consider the main problem again: J (p) → min, F (p) ≤ 0, K(p) = 0, p := (x(t0 ), x(tf )) ∈ P , ẋ = a(t, x, v) + B(t, x, v)u, u ∈ U , (t, x, v) ∈ Q, 236 Chapter 5. Second-Order Optimality Conditions in Optimal Control where U is a convex polyhedron. Let U be the set of vertices of U . Consider an admissible process w 0 = (x 0 , u0 , v 0 ) ∈ W in the main problem. Assume that the control u0 is a piecewise constant function on taking all its values in the vertices of U , i.e., u0 (t) ∈ U, t ∈ . (5.60) As usual we denote by = {t1 , . . . , ts } the set of discontinuity points (switching points) of the control u0 . Assume that the control v 0 (t) is a continuous function on , satisfying the Condition Lθ (see Section 5.1.2). By virtue of condition (5.60) the process w0 = (x 0 , u0 , v 0 ) is also admissible in the problem J (p) → min, F (p) ≤ 0, K(p) = 0, p ∈ P , ẋ(t) = a(t, x, v) + B(t, x, v)u, u(t) ∈ U, (t, x(t), v(t)) ∈ Q. (5.61) (5.62) Since U = co U, the main problem can be viewed as a convexification of problem (5.61), (5.62) with respect to u. It is easy to see that, in problem (5.61), (5.62), we can use the results of Sections 3.1.5 and 3.2.3 and formulate both necessary and sufficient optimality conditions for w 0 (see Theorems 3.9 and 3.16, respectively). Indeed, let ui , i = 1, . . . , m, be the vertices of the polyhedron U , i.e., U = {u1 , . . . , um }. Let Q(ui ), i = 1, . . . , m, be disjoint open neighborhoods of the vertices ui ∈ U. Set Qg = m Q(ui ). i=1 Define the function g(u) : Qg → as follows: On each set Q(ui ) ⊂ Qg , i = 1, . . . , m, i let g(u) = u − u . Then g(u) is a function of class C ∞ on Qg , specifying the set of vertices of U , i.e., U = {u ∈ Qg | g(u) = 0}. Moreover, g (u) = I for all u ∈ Qg , where I is the identity matrix of order d(u). Hence, the full-rank condition (3.3) is fulfilled. (This very simple but somewhat unexpected way of using equality constraints g(u) = 0 in the problem with a constraint on the control specified by a polyhedron U is due to Milyutin.) Thus, the problem (5.61), (5.62) can be represented as Rd(u) J (p) → min, F (p) ≤ 0, K(p) = 0, p ∈ P , ẋ(t) = a(t, x, v) + B(t, x, v)u, g(u) = 0, u ∈ Qg , (t, x(t), v(t)) ∈ Q. (5.63) (5.64) (5.65) Let us formulate the quadratic conditions of Sections 3.1.5 and 3.2.3 for the point w 0 in this problem. Put l = α0 J + αF + βK, H = ψf, H̄ = H + νg, where f = a + Bu. The set M0 (cf. (3.12)–(3.15)) consists of tuples λ = (α0 , α, β, ψ, ν) (5.66) α0 ∈ R1 , α ∈ Rd(F ) , β ∈ Rd(K) , ψ ∈ W 1,1 (, Rd(x) ), ν ∈ L∞ (, Rd(u) ), α0 ≥ 0, α ≥ 0, αF (p 0 ) = 0, α0 + |α| + |β| = 1, −ψ̇ = H̄x , ψ(t0 ) = −lx0 , ψ(tf ) = lxf , H (t, x 0 (t), u, v, ψ(t)) ≥ H (t, x 0 (t), u0 (t), v 0 (t), ψ(t)) ∀ t ∈ \ , u ∈ U, v ∈ Rd(v) such that (t, x 0 (t), v) ∈ Q. (5.67) such that 5.2. Quadratic Optimality Conditions on a Variable Time Interval 237 The last inequality implies the conditions of a local minimum principle with respect to u and v: ψfu (t, x 0 (t), u0 (t), v 0 (t)) + ν(t)g (u0 (t)) = 0, ψfv (t, x 0 (t), u0 (t), v 0 (t)) = 0. Since g (u) = I , the first equality uniquely determines the multiplier ν(t). Note that H̄x = Hx . Therefore conditions (5.67) are equivalent to the minimum principle conditions (5.7)–(5.10). More precisely, the linear projection (α0 , α, β, ψ, ν) → (α0 , α, β, ψ) yields a one-to-one correspondence between the elements of the set (5.67) and the elements of the set (5.7)–(5.10). Consider the critical cone K for the point w0 (see Definition 3.5). To the constraint g(u) = 0, u ∈ Qg there corresponds the condition g (u0 )ū = 0. But g = I . Hence ū = 0. This condition implies that the critical cone K can be identified with the set of triples z̄ = (ξ , x̄, v̄) ∈ Z2 () such that conditions (5.14)–(5.16) are fulfilled. The definition of the set M0+ in Section 5.1.6 corresponds to the definition of this set in Section 3.2.3. The same is true for the set Leg+ (M0+ ) (again, due to the equality g = I ). Further, for the point w 0 we write the quadratic form (see formula (3.57)), where we set ū = 0. Since H̄x = Hx and H̄ψ = Hψ , we have [H̄x ]k = [Hx ]k , D k (H̄ ) = D k (H ), k = 1, . . . , s. Moreover, H̄xx = Hxx , H̄xv = Hxv , H̄vv = Hvv . Combined with condition ū = 0 this implies H̄ww w̄, w̄ = Hxx x̄, x̄ + 2Hxv v̄, x̄ + Hvv v̄, v̄ . (5.68) Thus, in view of condition ū = 0 the quadratic form becomes defined by formula (5.17). Now, Theorem 5.5 easily follows from Theorem 5.10, and similarly Theorem 5.8 becomes a simple consequence of Theorem 5.16. 5.2 5.2.1 Quadratic Optimality Conditions in the Problem on a Variable Time Interval Optimal Control Problem on a Variable Time Interval Let x(t) ∈ Rd(x) denote the state variable, and let u(t) ∈ Rd(u) , v(t) ∈ Rd(v) be the two types of control variables in the time interval t ∈ [t0 , tf ] with a nonfixed initial time t0 and final time tf . The following optimal control problem (5.69)–(5.72) will be referred to as the general problem linear in a part of controls: Minimize J(t0 , tf , x(·), u(·), v(·)) = J (t0 , x(t0 ), tf , x(tf )) (5.69) subject to the constraints ẋ(t) = f (t, x(t), u(t), v(t)), u(t) ∈ U , (t, x(t), v(t)) ∈ Q, F (t0 , x(t0 ), tf , x(tf )) ≤ 0, K(t0 , x(t0 ), tf , x(tf )) = 0, (t0 , x(t0 ), tf , x(tf )) ∈ P . (5.70) (5.71) 238 Chapter 5. Second-Order Optimality Conditions in Optimal Control The control variable u appears linearly in the system dynamics, f (t, x, u, v) = a(t, x, v) + B(t, x, v)u, (5.72) whereas the control variable v appears nonlinearly in the dynamics. Here, F , K, and a are column vector functions, B is a d(x) × d(u) matrix function, P ⊂ R2+2d(x) and Q ⊂ R1+d(x)+d(v) are open sets, and U ⊂ Rd(u) is a convex polyhedron. The functions J , F , and K are assumed to be twice continuously differentiable on P and the functions a and B are twice continuously differentiable on Q. The dimensions of F and K are denoted by d(F ) and d(K). By = [t0 , tf ] we denote the interval of control. We shall use the abbreviations x0 = x(t0 ), xf = x(tf ), and p = (t0 , x0 , tf , xf ). A process T = {(x(t), u(t), v(t)) | t ∈ [t0 , tf ] } is said to be admissible, if x(·) is absolutely continuous, u(·), v(·) are measurable bounded on = [t0 , tf ], and the triple of functions (x(t), u(t), v(t)) together with the endpoints p = (t0 , x(t0 ), tf , x(tf )) satisfies the constraints (5.70) and (5.71). Definition 5.17. The process T affords a Pontryagin minimum if there is no sequence of admissible processes T n = {(x n (t), un (t), v n (t)) | t ∈ [t0n , tfn ] }, n = 1, 2, . . . , such that the following properties hold with n = [t0n , tfn ]: (a) (b) (c) (d) J(T n ) < J(T ) for all n and t0n → t0 , tfn → tf for n → ∞; maxn ∩ |x n (t) − x(t)| → 0 for n → ∞; n ∩ |u n (t) − u(t)| dt → 0, n ∩ |v n (t) − v(t)| dt → 0 for n → ∞; there exists a compact set C ⊂ Q (which depends on the choice of the sequence) such that for all sufficiently large n, we have (t, x n (t), v n (t)) ∈ C a.e. on n . For convenience, let us formulate an equivalent definition of the Pontryagin minimum. Definition 5.18. The process T affords a Pontryagin minimum if for each compact set C ⊂ Q there exists ε > 0 such that J(T˜ ) ≥ J(T ) for all admissible processes T˜ = {(x̃(t), ũ(t), ṽ(t)) | t ∈ [t˜0 , t˜f ] } such that (a) |t˜0 − t0 | < ε, |t˜f − tf | < ε; ˜ = [t˜0 , t˜f ]; (b) max ˜ |x̃(t) − x(t)| < ε, where ∩ (c) (d) 5.2.2 |ũ(t) − u(t)| dt ˜ ∩ < ε; |ṽ(t) − v(t)| dt ˜ ∩ < ε; ˜ (t, x̃(t), ṽ(t)) ∈ C a.e. on . First-Order Necessary Optimality Conditions Let T = {(x(t), u(t), v(t)) | t ∈ [t0 , tf ] } be a fixed admissible process such that the control u(t) is a piecewise constant function taking all its values in the vertices of the polyhedron U , and the control v(t) is a Lipschitz continuous function on the interval = [t0 , tf ]. Denote by = {t1 , . . . , ts }, t0 < t1 < · · · < ts < tf the finite set of all discontinuity points (jump points) of the control u(t). Then ẋ(t) is a piecewise continuous function whose points of discontinuity belong to , and hence x(t) is a piecewise smooth function on . Let us formulate a first-order necessary condition for optimality of the process T in the form of the Pontryagin minimum principle. As in Section 5.1.2, we introduce the 5.2. Quadratic Optimality Conditions on a Variable Time Interval 239 Pontryagin function (or Hamiltonian) H (t, x, u, v, ψ) = ψf (t, x, u, v) = ψa(t, x, v) + ψB(t, x, v)u, (5.73) where ψ is a row vector of dimension d(ψ) = d(x), while x, u, f , F , and K are column vectors; the factor of the control u in the Pontryagin function is called the switching function for the u-component φ(t, x, v, ψ) = ψB(t, x, v) (5.74) which is a row vector of dimension d(u). We also introduce the endpoint Lagrange function l(p, α0 , α, β) = α0 J (p) + αF (p) + βK(p), p = (t0 , x0 , tf , xf ), where α and β are row vectors with d(α) = d(F ) and d(β) = d(K), and α0 is a number. We introduce multipliers λ = (α0 , α, β, ψ(·), ψ0 (·)) such that ψ(·) : → d(x) ∗ a tuple of Lagrange , ψ0 (·) : → R1 are continuous on and continuously differentiable on each R interval of the set \ . As usual, we denote first- or second-order partial derivatives by subscripts referring to the variables. Denote by M0 the set of the normalized tuples λ satisfying the minimum principle conditions for the process T : α0 ≥ 0, α ≥ 0, αF (p) = 0, α0 + d(F ) i=1 αi + d(K) |βj | = 1, (5.75) j =1 ψ̇ = −Hx , ψ̇0 = −Ht ∀ t ∈ \ , ψ(t0 ) = −lx0 , ψ(tf ) = lxf , ψ0 (t0 ) = −lt0 , ψ0 (tf ) = ltf , H (t, x(t), u, v, ψ(t)) ≥ H (t, x(t), u(t), v(t), ψ(t)) ∀ t ∈ \ , u ∈ U , v ∈ Rd(v) such that (t, x(t), v) ∈ Q, H (t, x(t), u(t), v(t), ψ(t)) + ψ0 (t) = 0 ∀ t ∈ \ . (5.76) (5.77) (5.78) (5.79) The derivatives lx0 and lxf are taken at the point (p, α0 , α, β), where p = (t0 , x(t0 ), tf , x(tf )), while the derivatives Hx , Ht are evaluated at the point (t, x(t), u(t), v(t), ψ(t)) for t ∈ \ . The condition M0 = ∅ constitutes the first-order necessary condition for a Pontryagin minimum of the process T which is the Pontryagin minimum principle. Theorem 5.19. If the process T affords a Pontryagin minimum, then the set M0 is nonempty. The set M0 is a finite-dimensional compact set and the projector λ → (α0 , α, β) is injective on M0 . Again we use the simple abbreviation (t) for indicating all arguments (t, x(t), u(t), v(t), ψ(t)). Let λ = (α0 , α, β, ψ(·), ψ0 (·)) ∈ M0 . From condition (5.79), it follows that H (t) is a continuous function. In particular, we have [H ]k = H k+ − H k− = 0 for each tk ∈ , where H k− := H (tk , x(tk ), u(tk −), v(tk ), ψ(tk )), H k+ := H (tk , x(tk ), u(tk +), v(tk ), ψ(tk )). 240 Chapter 5. Second-Order Optimality Conditions in Optimal Control We denote by H k the common value of H k− and H k+ . For λ ∈ M0 and tk ∈ we consider the function (k H )(t) = H (t, x(t), uk+ , v(tk ), ψ(t)) − H (t, x(t), uk− , v(tk ), ψ(t)) = φ(t, x(t), v(tk ), ψ(t)) [u]k . For this function, Propositions 5.2 and 5.3 hold, so that for each λ ∈ M0 k = 1, . . . , s, where (5.80) we have D k (H ) ≥ 0, d (k H )(tk ) = −φ̇(tk ±)[u]k dt = −Hxk+ Hψk− + Hxk− Hψk+ − [Ht ]k = ψ̇ k+ ẋ k− − ψ̇ k− ẋ k+ + [ψ0 ]k , D k (H ) := − where Hxk− and Hxk+ are the left- and right-hand values of the function Hx (t) at tk , respectively, [Ht ]k is the jump of the function Ht (t) at tk , etc. 5.2.3 Bounded Strong Minimum As in the case of a fixed time interval we give the following definitions. The state variable xi , i.e., the ith component of the state vector x, is called unessential if the function f does not depend on xi and if the functions F , J , and K are affine in xi0 = xi (t0 ) and xi1 = xi (tf ). We denote by x the vector of all essential components of state vector x. Definition 5.20. We say that the process T affords a bounded strong minimum if there is no sequence of admissible processes T n = {(x n (t), un (t), v n (t)) | t ∈ [t0n , tfn ] }, n = 1, 2, . . ., such that (a) J(T n ) < J(T ); (b) t0n → t0 , tfn → tf , x n (t0 ) → x(t0 ) (n → ∞); (c) maxn ∩ |x n (t) − x(t)| → 0 (n → ∞), where n = [t0n , tfn ]; (d) there exists a compact set C ⊂ Q (which depends on the choice of the sequence) such that for all sufficiently large n we have (t, x n (t), v n (t)) ∈ C a.e. on n . An equivalent definition has the following form. Definition 5.21. The process T affords a bounded strong minimum if for each compact set C ⊂ Q there exists ε > 0 such that J(T˜ ) ≥ J(T ) for all admissible processes T˜ = {(x̃(t), ũ(t), ṽ(t)) | t ∈ [t˜0 , t˜f ] } such that (a) |t˜0 − t0 | < ε, |t˜f − tf | < ε, |x̃(t0 ) − x(t0 )| < ε; ˜ = [t˜0 , t˜f ]; |x̃(t) − x(t)| < ε, where (b) max∩ ˜ ˜ (c) (t, x̃(t), ṽ(t)) ∈ C a.e. on . The strict bounded strong minimum is defined in a similar way, with the nonstrict inequality J(T˜ ) ≥ J(T ) replaced by the strict one and the process T˜ required to be different from T . Below, we shall formulate a quadratic necessary optimality condition of a Pontryagin minimum (Definition 5.17) for given control process T . A strengthening of this quadratic condition yields a quadratic sufficient condition of a bounded strong minimum (Definition 5.20). 5.2. Quadratic Optimality Conditions on a Variable Time Interval 5.2.4 241 Critical Cone For a given process T we introduce the space Z2 () and the critical cone K ⊂ Z2 (). Let z̄ = (t¯0 , t¯f , ξ̄ , x̄, v̄), where t¯0 , t¯f ∈ R1 , ξ̄ ∈ Rs , x̄ ∈ P W 1,2 (, Rd(x) ), v̄ ∈ L2 (, Rd(v) ). Thus, z̄ ∈ Z2 () := R2 × Rs × P W 1,2 (, Rd(x) ) × L2 (, Rd(v) ). For each z̄, we set x̄¯0 = x̄(t0 ) + t¯0 ẋ(t0 ), x̄¯f = x̄(tf ) + t¯f ẋ(tf ), p̄¯ = (t¯0 , x̄¯0 , t¯f , x̄¯f ). (5.81) The vector p̄¯ is considered a column vector. Note that t¯0 = 0, respectively, t¯f = 0, holds for a fixed initial time t0 , respectively, final time tf . Denote by IF (p) = {i ∈ {1, . . . , d(F )} | Fi (p) = 0} the set of indices of all active endpoint inequalities Fi (p) ≤ 0 at the point p = (t0 , x(t0 ), tf , x(tf )). Denote by K the set of all z̄ ∈ Z2 () satisfying the following conditions: J (p)p̄¯ ≤ 0, Fi (p)p̄¯ ≤ 0 ∀ i ∈ IF (p), K (p)p̄¯ = 0, ˙ = fx (t, x(t), u(t), v(t))x̄(t) + fv (t, x(t), u(t), v(t))v̄(t), x̄(t) [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, (5.82) (5.83) (5.84) where p = (t0 , x(t0 ), tf , x(tf )), [ẋ] = ẋ(tk +) − ẋ(tk −). It is obvious that K is a convex cone with finitely many faces in the space Z2 (). The cone K is called the critical cone. 5.2.5 Necessary Quadratic Optimality Conditions Let us introduce a quadratic form on the critical cone K defined by the conditions (5.82)– (5.84). For each λ ∈ M0 and z̄ ∈ K, we set (λ, z̄) = ωe (λ, z̄) + + s k (D k (H )ξ̄k2 − [ψ̇]k x̄av ξ̄k ) k=1 tf t0 Hxx (t)x̄(t), x̄(t) + 2Hxv (t)v̄(t), x̄(t) + Hvv (t)v̄(t), v̄(t) dt, (5.85) where ¯ p̄¯ − 2ψ̇(tf )x̄(tf )t¯f − ψ̇(tf )ẋ(tf ) + ψ̇0 (tf ) t¯f2 ωe (λ, z̄) = lpp p̄, + 2ψ̇(t0 )x̄(t0 )t¯0 + ψ̇(t0 )ẋ(t0 ) + ψ̇0 (t0 ) t¯02 , (5.86) 1 k x̄av = (x̄ k− + x̄ k+ ), 2 Hxx (t) = Hxx (t, x(t), u(t), v(t), ψ(t)), etc. lpp = lpp (p, α0 , α, β, p), p = (t0 , x(t0 ), tf , x(tf )), Note that for a problem on a fixed time interval [t0 , tf ] we have t¯0 = t¯f = 0 and, hence, ¯ p̄¯ . The following theorem gives the main the quadratic form (5.86) reduces to lpp p̄, second-order necessary condition of optimality. 242 Chapter 5. Second-Order Optimality Conditions in Optimal Control Theorem 5.22. If the process T affords a Pontryagin minimum, then the following Condition A holds: The set M0 is nonempty and max (λ, z̄) ≥ 0 ∀ z̄ ∈ K. λ∈M0 We call Condition A the necessary quadratic condition. 5.2.6 Sufficient Quadratic Optimality Conditions A natural strengthening of the necessary Condition A turns out to be a sufficient optimality condition not only for a Pontryagin minimum, but also for a bounded strong minimum; cf. Definition 5.20. Denote by M0+ the set of all λ ∈ M0 satisfying the following conditions: (a) H (t, x(t), u, v, ψ(t)) > H (t, x(t), u(t), v(t), ψ(t)) for all t ∈ \ , u ∈ U , v ∈ Rd(v) such that (t, x(t), v) ∈ Q and (u, v) = (u(t), v(t)); (b) H (tk , x(tk ), u, v, ψ(tk )) > H k for all tk ∈ , u ∈ U , v ∈ Rd(v) such that (tk , x(tk ), v) ∈ Q, (u, v) = (u(tk −), v(tk )), (u, v) = (u(tk +), v(tk )), where H k := H k− = H k+ . Definition 5.23. An element λ ∈ M0 is said to be strictly Legendre if the following conditions are satisfied: (a) For each t ∈ \ the quadratic form Hvv (t, x(t), u(t), v(t), ψ(t))v̄, v̄ is positive definite in Rd(v) ; (b) for each tk ∈ the quadratic form Hvv (tk , x(tk ), u(tk −), v(tk ), ψ(tk ))v̄, v̄ is positive definite in Rd(v) ; (c) for each tk ∈ the quadratic form Hvv (tk , x(tk ), u(tk +), v(tk ), ψ(tk ))v̄, v̄ is positive definite in Rd(v) ; (d) D k (H ) > 0 for all tk ∈ . Denote by Leg+ (M0+ ) the set of all strictly Legendrian elements λ ∈ M0+ and set γ̄ (z̄) = t¯02 + t¯f2 + ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) + tf v̄(t), v̄(t) dt. t0 Theorem 5.24. Let the following Condition B be fulfilled for the process : (a) The set Leg+ (M0+ ) is nonempty; (b) there exists a nonempty compact set M ⊂ Leg+ (M0+ ) and a number C > 0 such that maxλ∈M (λ, z̄) ≥ C γ̄ (z̄) for all z̄ ∈ K. Then T is a strict bounded strong minimum. If the set Leg+ (M0+ ) is nonempty and K = {0}, then (b) is fulfilled automatically. This is a first-order sufficient optimality condition of a strict bounded strong minimum. Let us emphasize that there is no gap between the necessary Condition A and the sufficient Condition B. 5.2. Quadratic Optimality Conditions on a Variable Time Interval 5.2.7 243 Proofs Let us consider the following problem on a variable time interval which is similar to the general problem (5.69)–(5.71): Minimize J(t0 , tf , x(·), u(·), v(·)) = J (t0 , x(t0 ), tf , x(tf )) (5.87) subject to the constraints ẋ(t) = f (t, x(t), u(t), v(t)), u(t) ∈ U, (t, x(t), v(t)) ∈ Q, F (t0 , x(t0 ), tf , x(tf )) ≤ 0, K(t0 , x(t0 ), tf , x(tf )) = 0, (t0 , x(t0 ), tf , x(tf )) ∈ P . (5.88) (5.89) The control variable u appears linearly in the system dynamics, f (t, x, u, v) = a(t, x, v) + B(t, x, v)u, (5.90) whereas the control variable v appears nonlinearly. Here, F , K, and a are column vector functions, B is a d(x) × d(u) matrix function, P ⊂ R2+2d(x) and Q ⊂ R1+d(x)+d(v) are open sets, and U ⊂ Rd(u) is a compact set. The functions J , F , and K are assumed to be twice continuously differentiable on P and the functions a and B are twice continuously differentiable on Q. The dimensions of F and K are denoted by d(F ) and d(K). By = [t0 , tf ] we denote the interval of control. The compact set U is specified by U = {u ∈ Qg | g(u) = 0}, (5.91) where Qg ⊂ Rd(u) is an open set and g : Qg → Rd(g) is a twice continuously differentiable function satisfying the full-rank condition rank gu (u) = d(g) (5.92) for all u ∈ Qg such that g(u) = 0. We refer to the problem (5.87)–(5.89) as the problem A. Along with this problem we treat its convexification co A with respect to u in which the constraint u ∈ U is replaced by the constraint u ∈ co U: ẋ = a(t, x, v) + B(t, x, v)u, u ∈ co U, (t, x, v) ∈ Q. (5.93) Thus co A is the problem (5.87), (5.89), (5.93). Let T = (x(t), u(t), v(t) | t ∈ [t0 , tf ]) be an admissible trajectory in the problem A such that u(t) is a piecewise Lipschitz continuous function on the interval = [t0 , tf ], with the set of discontinuity points = {t1 , . . . , ts }, and the control v(t) is Lipschitz continuous on the same interval. For the trajectory T we deal with the same question as in Section 5.1.7: What is the relationship between quadratic conditions in the problems A and co A? As in Section 5.1.7, for necessary quadratic conditions this question is simple to answer: A Pontryagin minimum in the problem co A implies a Pontryagin minimum in the problem A, and hence Theorem 3.31 implies the following assertion. Theorem 5.25. Condition A for a trajectory T in the problem A is a necessary condition for a Pontryagin minimum at this trajectory in the problem co A. 244 Chapter 5. Second-Order Optimality Conditions in Optimal Control Consider now the same question for sufficient quadratic conditions. It can be solved with the aid of Theorem 5.16 obtained for the problem Z on a fixed time interval, but first we have to make a simple time change. Namely, with the admissible trajectory T in the problem A we associate the trajectory T τ = z(τ ), t(τ ), x(τ ), u(τ ), v(τ ) | τ ∈ [τ0 , τf ] , where τ0 = t0 , τf = tf , t(τ ) ≡ τ , z(τ ) ≡ 1. This is an admissible trajectory in the problem Aτ specified by conditions J(T τ ) := J (t(τ0 ), x(τ0 ), t(τf ), x(τf )) → min (5.94) subject to the constraints F (t(τ0 ), x(τ0 ), t(τf ), x(τf )) ≤ 0, K(t(τ0 ), x(τ0 ), t(τf ), x(τf )) = 0, dx(τ ) = z(τ ) a(t(τ ), x(τ ), v(τ )) + B(t(τ ), x(τ ), v(τ ))u(τ ) , dτ dt(τ ) = z(τ ), dτ dz(τ ) = 0, dτ (5.95) (5.96) (5.97) (5.98) respectively,4 (t(τ0 ), x(τ0 ), t(τf ), x(τf )) ∈ P , (t(τ ), x(τ )) ∈ Q, u(τ ) ∈ U. (5.99) The interval [τ0 , τf ] in the problem Aτ is fixed. Consider also the problem co Aτ differing from Aτ by the constraint u ∈ U replaced with the constraint u ∈ co U. We have the following chain of implications: Condition B for the trajectory T in the problem A. =⇒ Condition B for the trajectory T τ in the problem Aτ. =⇒ A strict bounded strong minimum is attained on the trajectory T τ in the problem co Aτ. =⇒ A strict bounded strong minimum is attained on the trajectory T in the problem co A. The first implication was proved in Section 3.3.4 for problems P and P τ which are more general than A and Aτ , respectively, the second follows from Theorem 5.8, and the third is readily verified. Thus we obtain the following theorem. Theorem 5.26. Condition B for an admissible trajectory T in the problem A is a sufficient condition for a strict strong minimum in the problem co A. Now, recall that the representation of the set of vertices of polyhedron U in the form of equality-type constraint g(u) = 0, u ∈ Qg allowed us to consider the main problem (5.1)– (5.4) as a special case of the problem co Z (5.18), (5.19), (5.23) and thus to obtain both necessary Condition A and sufficient Condition B in the main problem as the consequences of these conditions in problem co Z; more precisely, we have shown that Theorems 5.5 and 4 Note that the function z(τ ) in problem Aτ corresponds to the function v(τ ) in problem P τ . 5.3. Riccati Approach 245 5.8 follow from Theorem 5.10 and 5.16, respectively (see the proofs of Theorems 5.5 and 5.8). Similarly, this representation allows us to consider the general problem on a variable time interval (5.69)–(5.72) as a special case of the problem co A and thus to obtain Theorems 5.22 and 5.24 as the consequences of Theorems 5.25 and 5.26, respectively. 5.3 Riccati Approach The following question suggests itself from a numerical point of view: How does a numerical check of the quadratic sufficient optimality conditions in Theorem 5.8 look? For simplicity, we shall assume that (a) the initial value x(t0 ) is fixed, (b) there are no endpoint constraints of inequality type, and (c) the time interval = [t0 , tf ] is fixed. Thus, we consider the following control problem: Minimize J (x(tf )) under the constraints x(t0 ) = x0 , where K(x(tf )) = 0, ẋ = f (t, x, u, v), u ∈ U, f (t, x, u, v) = a(t, x, v) + B(t, x, v)u, ⊂ Rd(u) is a convex polyhedron, and J , K, a, and B are C 2 -functions. Let w = (x, u, v) be a fixed admissible process satisfying the assumptions of Section 5.1.2 (consequently, the function u(t) is piecewise constant and the function v(t) is continuous). We also assume, for this process, that the set M0 is nonempty and there exists λ ∈ M0 such that α0 > 0; let us fix this element λ. Here we set again = {t1 , t2 , . . . , ts }, where tk denote the discontinuity points of the bang-bang control u(t). Let n = d(x). U 5.3.1 Critical Cone K and Quadratic Form In the considered case, the critical cone is a subspace defined by the relations x̄(t0 ) = 0, K (x(tf ))x̄(tf ) = 0, ˙ = fx (t)x̄(t) + fv (t)v̄(t), [x̄]k = [ẋ]k ξ̄k , x̄(t) k = 1, . . . , s. These relations imply that J (x(tf ))x̄(tf ) = 0 since α0 > 0. The quadratic form is given by (λ, z̄) = lxf xf (x(tf ))x̄f , x̄f + + s k (D k (H )ξ̄k2 − 2[ψ̇]k x̄av ξ̄k ) k=1 tf Hxx (t)x̄(t), x̄(t) + 2Hxv (t)v̄(t), x̄(t) + Hvv (t)v̄(t), v̄(t) dt, t0 where, by definition, x̄f = x̄(tf ). We assume that D k (H ) > 0, k = 1, . . . , s, and the strengthened Legendre (SL) condition with respect to v is satisfied: Hvv (t)v̄, v̄ ≥ cv̄, v̄ ∀ v̄ ∈ Rd(v) , ∀ t ∈ [t0 , tf ] \ (c > 0). 246 Chapter 5. Second-Order Optimality Conditions in Optimal Control 5.3.2 Q-Transformation of on K Let Q(t) be a symmetric n × n matrix on [t0 , tf ] with piecewise continuous entries which are absolutely continuous on each interval of the set [t0 , tf ] \ . For each z̄ ∈ K we obviously have tf tf s d [Qx̄, x̄ ]k , Qx̄, x̄ dt = Qx̄, x̄ − (5.100) dt t0 t0 k=1 where [Qx̄, x̄ ] is the jump of the function Qx̄, x̄ at the point tk ∈ . Using the equation x̄˙ = fx x̄ + fv v̄ and the initial condition x̄(t0 ) = 0, we obtain k −Q(tf )x̄f , x̄f + s [Qx̄, x̄ ]k k=1 tf + (Q̇x̄, x̄ + Q(fx x̄ + fv v̄), x̄ + Qx̄, fx x̄ + fv v̄ dt = 0. t0 Adding this zero term to the form (λ, z̄), we get s k ξ̄k + [Qx̄, x̄ ]k (λ, z̄) = lxf xf − Q(tf ) x̄f , x̄f + D k (H )ξ̄k2 − 2[ψ̇]k x̄av + k=1 tf t0 (Hxx + Q̇ + Qfx + fx∗ Q)x̄, x̄ + (Hxv + Qfv )v̄, x̄ +(Hvx + fv∗ Q)x̄, v̄ + Hvv (t)v̄(t), v̄(t) dt. We call this formula the Q-transformation of on K. 5.3.3 Transformation of on K to Perfect Squares In order to transform the integral term in (λ, z̄) into a perfect square, we assume that Q(t) satisfies the following matrix Riccati equation (cf. equation (4.189)): −1 Q̇ + Qfx + fx∗ Q + Hxx − (Hxv + Qfv )Hvv (Hvx + fv∗ Q) = 0. Then the integral term in can be written as tf −1 Hvv h̄, h̄ dt, where h̄ = (Hvx + fv∗ Q)x̄ + Hvv v̄. t0 As we know, the terms k ξ̄k + [Qx̄, x̄ ]k ωk := D k (H )ξ̄k2 − 2[ψ̇]k x̄av can also be transformed into perfect squares if the matrix Q(t) satisfies a special jump condition at each point tk ∈ . This jump condition was obtained in Chapter 4. Namely, for each k = 1, . . . , s put Qk− = Q(tk −), Qk+ = Q(tk +), [Q]k = Qk+ − Qk− , qk− = ([ẋ]k )∗ Qk− − [ψ̇]k , bk− = D k (H ) − (qk− )[ẋ]k , (5.101) (5.102) 5.3. Riccati Approach 247 where [ẋ]k is a column vector, while qk− , ([ẋ]k )∗ and [ψ̇]k are row vectors, and bk− is a number. We shall assume that bk− > 0, k = 1, . . . , s, (5.103) holds and that Q satisfies the jump conditions [Q]k = (bk− )−1 (qk− )∗ (qk− ), (5.104) where (qk− ) is a row vector, (qk− )∗ is a column vector, and hence (qk− )∗ (qk− ) is a symmetric n × n matrix. Then, as it was shown in Section 4.2, ωk = (bk− )−1 ((bk− )ξ̄k + (qk− )(x̄ k+ ))2 = (bk− )−1 (D k (H )ξ̄k + (qk− )(x̄ k− ))2 . (5.105) Thus, we obtain the following transformation of the quadratic form = (λ, z̄) to perfect squares on the critical cone K: s 2 (bk− )−1 (D k (H )ξ̄k + (qk− )(x̄ k− ) + = lxf xf − Q(tf ) x̄f , x̄f + k=1 tf t0 −1 Hvv h̄, h̄ dt, where h̄ = (Hvx + fv∗ Q)x̄ + Hvv v̄. In addition, let us assume that lxf xf − Q(tf ) x̄f , x̄f ≥ 0 for all x̄f ∈ Rd(x) \ {0} such that Kxf (x(tf ))x̄f = 0. Then, obviously, (λ, z̄) ≥ 0 on K. Now let us show that (λ, z̄) > 0 for each nonzero element z̄ ∈ K. This will imply that (λ, z̄) is positive definite on the critical cone K since (λ, z̄) is a Legendre quadratic form. Assume that (λ, z̄) = 0 for some element z̄ ∈ K. Then, for this element, the following equations hold: x̄(t0 ) = 0, D k (H )ξ̄k + (qk− )(x̄ k− ) = 0, h̄(t) = 0 a.e. in . k = 1, . . . , s, (5.106) (5.107) (5.108) From the last equation, we get −1 (Hvx + fv∗ Q)x̄. v̄ = −Hvv (5.109) Using this formula in the equation x̄˙ = fx x̄ + fv v̄, we see that x̄ is a solution to the linear equation −1 (Hvx + fv∗ Q))x̄. (5.110) x̄˙ = (fx − fv Hvv This equation together with initial condition x̄(t0 ) = 0 implies that x̄(t) = 0 for all t ∈ [t0 , t1 ). Consequently, x̄ 1− = 0, and then, by virtue of (5.107), ξ̄1 = 0. This equality, together with the jump condition [x̄]1 = [ẋ]1 ξ̄1 , implies that [x̄]1 = 0, i.e., x̄ is continuous at t1 . Consequently, x̄ 1+ = 0. From the last condition and equation (5.110) it follows that x̄(t) = 0 for all t ∈ (t1 , t2 ). Repeating this argument, we obtain ξ̄1 = ξ̄2 = · · · = ξ̄s = 0, x̄(t) = 0 for all t ∈ [t0 , tf ]. Then from (5.109) it follows that v̄ = 0. Consequently, we have z̄ = 0 and thus have proved the following theorem; cf. [98]. 248 Chapter 5. Second-Order Optimality Conditions in Optimal Control Theorem 5.27. Assume that there exists a symmetric matrix Q(t), defined on [t0 , tf ], such that (a) Q(t) is piecewise continuous on [t0 , tf ] and continuously differentiable on each interval of the set [t0 , tf ] \ ; (b) Q(t) satisfies the Riccati equation −1 Q̇ + Qfx + fx∗ Q + Hxx − (Hxv + Qfv )Hvv (Hvx + fv∗ Q) = 0 (c) (5.111) on each interval of the set [t0 , tf ] \ ; at each point tk ∈ matrix Q(t) satisfies the jump condition [Q]k = (bk− )−1 (qk− )∗ (qk− ), where qk− = ([ẋ]k )∗ Qk− − [ψ̇]k , bk− = D k (H ) − (qk− )[ẋ]k > 0; (d) lxf xf − Q(tf ) x̄f , x̄f ≥ 0 for all x̄f ∈ Rd(x) \ {0} such that Kxf (x(tf ))x̄f = 0. Then (λ, z̄) is positive definite on the subspace K. In some problems, it is more convenient to integrate the Riccati equation (5.111) backwards from t = tf . A similar proof shows that we can replace condition (c) in Theorem 4.1 by the following condition: (c+) at each point tk ∈ , the matrix Q(t) satisfies the jump condition [Q]k = (bk+ )−1 (qk+ )∗ (qk+ ), where qk+ = ([ẋ]k )∗ Qk+ − [ψ̇]k , 5.4 bk+ = D k (H ) + (qk+ )[ẋ]k > 0. (5.112) Numerical Example: Optimal Control of Production and Maintenance Cho, Abad, and Parlar [22] introduced an optimal control model where a dynamic maintenance problem is incorporated into a production control problem so as to simultaneously compute optimal production and maintenance policies. In this model, the dynamics is linear with respect to both production and maintenance control, whereas the cost functional is quadratic with respect to production control and linear with respect to maintenance control. Hence, the model fits into the type of control problems considered in (5.1)–(5.4). A detailed numerical analysis of solutions for different final times may be found in Maurer, Kim, and Vossen [67]. For a certain range of final times the maintenance control is bang-bang. We will show that the sufficient conditions in Theorems 5.24 and 5.27 are satisfied for the computed solutions. The notation for the state variables is slightly different from that in [22, 67]. The state and control variables and parameters have the following meaning: x1 (t) x2 (t) v(t) m(t) : : : : α(t) : s(t) ρ>0 : : inventory level at time t ∈ [0, tf ] with fixed final time tf > 0, proportion of good units of end items produced at time t ∈ [0, tf ], scheduled production rate (control ), preventive maintenance rate to reduce the proportion of defective units produced (control ), obsolescence rate of the process performance in the absence of maintenance, demand rate, discount rate. 5.4. Numerical Example: Optimal Control of Production and Maintenance 249 The dynamics of the process is given by ẋ1 (t) = x2 (t)v(t) − s(t), x1 (0) = x10 > 0, ẋ2 (t) = −α(t)x2 (t) + (1 − x2 (t))m(t), x2 (0) = x20 > 0, (5.113) with the following bounds on the control variables: 0 ≤ v(t) ≤ vmax , 0 ≤ m(t) ≤ mmax for 0 ≤ t ≤ tf . (5.114) Since all demands must be satisfied, the following state constraint is imposed: x1 (t) ≥ 0 for 0 ≤ t ≤ tf . Computations show that this state constraint is automatically satisfied if we impose the boundary condition (5.115) x1 (tf ) = 0. The optimal control problem then consists in maximizing the total discounted profit plus the salvage value of x2 (tf ), tf J (x1 , x2 , m, v) = [ws − hx1 (t) − rv(t)2 − cm(t)]e−ρt dt (5.116) 0 + b x2 (tf )e−ρtf , under the constraints (5.113)–(5.115). For later computations, the values of constants are chosen as in [22]: α ≡ 2, w = 8, s(t) ≡ 4, r = 2, b = 10, vmax = 3, ρ = 0.1, mmax = 4, h = 1, x10 = 3, c = 2.5, x20 = 1. (5.117) The time horizon tf will be specified below. In the discussion of the minimum principle, we consider the usual Pontryagin function (Hamiltonian), see (5.73), instead of the current value Hamiltonian in [22, 67], H (t, x1 , x2 , ψ1 , ψ2 , m, v) = e−ρt (−ws + hx1 + rv 2 + cm) +ψ1 (x2 v − s) + ψ2 (−αx2 + (1 − x2 )m), (5.118) where ψ1 , ψ2 denote the adjoint variables. The adjoint equations and transversality conditions yield, in view of x1 (tf ) = 0 and the salvage term in the cost functional, ψ1 (tf ) = ν, ψ̇1 = −he−ρt , ψ̇2 = −ψ1 v + ψ2 (α + m), ψ2 (tf ) = −be−ρtf . (5.119) The multiplier ν is not known a priori and will be computed later. We will choose a time horizon for which the control constraint 0 ≤ v(t) ≤ V = 3 will not become active. Hence, the minimum condition in the minimum principle leads to the equation 0 = Hv = 2e−ρt v +ψ1 x2 , which yields the control (5.120) v = −ψ1 x2 eρt /2r. Since the maintenance control enters the Hamiltonian linearly, the control m is determined by the sign of the switching function φm (t) = Hm = e−ρt c + ψ2 (t)(1 − x2 (t)) (5.121) 250 Chapter 5. Second-Order Optimality Conditions in Optimal Control according to ⎧ if φm (t) > 0 ⎨ mmax 0 if φm (t) < 0 m(t) = ⎩ singular if φ (t) ≡ 0 m ⎫ ⎬ . (5.122) for t ∈ Ising ⊂ [0, tf ] ⎭ For the final time tf = 1 which was considered in [22] and [67], the maintenance control contains a singular arc. However, the computations in [67] show that for final times tf ∈ [0.15, 0.98] the maintenance control is bang-bang and has only one switching time: m(t) = 0 mmax = 4 for 0 ≤ t < t1 for t1 ≤ t ≤ tf . (5.123) Let us study the control problem with final time tf = 0.9 in more detail. To compute a solution candidate, we apply nonlinear programming methods to the discretized control problem with a large number N of grid points τi = i · tf /N , i = 0, 1, . . . , N ; cf. [5, 14]. Both the method of Euler and the method of Heun are employed for integrating the differential equation. We use the programming language AMPL developed by Fourer et al. [33] and the interior point optimization code IPOPT of Wächter and Biegler [114]. For N = 5000 grid points, the computed state, control and adjoint functions are displayed in Figure 5.1. The following values for the switching time, functional value, and selected state and adjoint (a) inventory x1 and good items x2 (b) production v and bang-bang maintenance m 3 4 2.5 3.5 3 2 2.5 1.5 2 1.5 1 1 0.5 0.5 0 0 0 0.1 (c) 0.2 0.3 0.4 0.5 adjoint variables 1 0.6 0.7 and 2 0.8 0.9 0 (d) -3 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 maintenance rate m and switching function 0.9 m 4 -4 3 -5 2 -6 1 -7 -8 0 -9 -1 -10 -2 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Figure 5.1. Optimal production and maintenance, final time tf = 0.9. (a) State variables x1 , x2 . (b) Regular production control v and bang-bang maintenance control m. (c) Adjoint variables ψ1 , ψ2 . (d) Maintenance control m with switching function φm . 5.4. Numerical Example: Optimal Control of Production and Maintenance 251 variables are obtained: t1 x1 (t1 ) x1 (tf ) ψ1 (0) ψ1 (t1 ) ψ1 (tf ) = 0.65691, = 0.84924, = 0.0, = −7.8617, = −8.4975, = −8.72313, J = 26.705, x2 (t1 ) = 0.226879, x2 (tf ) = 0.574104, ψ2 (0) = −4.70437, ψ2 (t1 ) = −3.2016, ψ2 (tf ) = −9.13931. (5.124) Now we evaluate the Riccati equation (5.111), Q̇ = −Qfx − fx Q − Hxx + (Hxv + Qfv )(Hvv )−1 (Hvx + fv∗ Q), for the symmetric 2 × 2 matrix Q= Computing the expressions 0 v fx = , 0 −(α + m) Q12 Q22 Q11 Q12 fv = x2 0 (5.125) . , Hxx = 0, Hxv = (0, ψ1 ), the matrix Riccati equation (5.125) yields the following ODE system: Q̇11 = Q211 x22 eρt /2r, (5.126) Q̇12 = Q11 v − Q12 (α + m) + e Q11 x2 (ψ1 + Q12 x2 ), Q̇22 = −2Q12 v + 2Q22 (α + m) + eρt (ψ1 + Q12 x2 )2 /2r. ρt (5.127) (5.128) The equations (5.126) and (5.127) are homogeneous in the variables Q11 and Q12 . Hence, we can try to find a solution to the Riccati system with Q11 (t) = Q12 (t) ≡ 0 on [0, tf ]. Then (5.128) reduces to the linear equation Q̇22 = 2Q22 (α + m) + eρt ψ12 /2r. (5.129) This linear equation always has a solution. The remaining task is to satisfy the jump and boundary conditions in Theorem 5.27 for the matrix Q. Since we shall integrate equation (5.128) backwards, it is more convenient to evaluate the jump (5.112). Moreover, the boundary conditions for Q in Theorem 5.27(d) show that the initial value Q(0) can be chosen arbitrarily, while the terminal condition imposes the sign condition Q22 (tf ) ≤ 0, since x2 (tf ) is free. Therefore, we can choose the terminal value Q22 (tf ) = 0. (5.130) Hence, using the computed values in (5.124), we solve the linear equation (5.129) with terminal condition (5.130). At the switching time t1 , we obtain the value Q22 (t1 ) = −1.5599. Next, we evaluate the jump in the state and adjoint variables and check conditions (5.112). We get ([ẋ]1 )∗ = (0, M(1 − x2 (t1 ))), [ψ̇] = (0, Mψ2 (t1 )), 252 Chapter 5. Second-Order Optimality Conditions in Optimal Control which yield the quantities q1+ = ([ẋ 1 ]T Q1+ − [ψ̇]1 = (0, M(1 − x2 (t1 ))Q22 (t1 +) − Mψ2 (t1 ) = (0, 8.2439), b1+ = D 1 (H ) + (q1+ )[ẋ 1 ] = D 1 (H ) + M 2 ((1 − x2 (t1 ))Q22 (t1 +) − ψ2 (t1 ))ψ2 (t1 ) = 27.028 + 133.55 = 165.58 > 0. Then the jump condition in (5.112), [Q]1 = (b1+ )−1 (q1+ )∗ (q1+ ) = 0 0 0 [Q22 ]1 , reduces to a jump condition for Q22 at t1 . However, we do not need to evaluate this jump condition explicitly because the linear equation (5.129) has a solution regardless of the value Q22 (t1 −). Hence, we conclude from Theorem 5.27 that the numerical solution characterized by (5.124) and displayed in Figure 5.1 provides a strict bounded strong minimum. We may hope to improve on the benefit by choosing a larger time horizon. For the final time tf = 1.1, we get a bang-singular-bang maintenance control m(t) as shown in Figure 5.2. (a) inventory x1 and good items x2 (b) 3 production v and maintenance m 4 2.5 3.5 3 2 2.5 2 1.5 1.5 1 1 0.5 0.5 0 0 0 (c) 0.2 0.4 0.6 adjoint variables 1 0.8 and 0 1 (d) 2 -4 0.2 0.4 0.6 0.8 maintenance m and switching function 1 m 4 -5 -6 3 -7 2 -8 1 -9 -10 0 -11 -1 -12 -13 -2 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 Figure 5.2. Optimal production and maintenance, final time tf = 1.1. (a) State variables x1 , x2 . (b) Regular production control v and bang-singular-bang maintenance control m. (c) Adjoint variables ψ1 , ψ2 . (d) Maintenance control m with switching function φm . 5.4. Numerical Example: Optimal Control of Production and Maintenance 253 The solution characteristics for this solution are given by J = 23.3567, x1 (tf ) = 0, ψ1 (0) = −11.93, ψ1 (tf ) = −12.97, x2 (tf ) = 0.64926, ψ2 (0) = −9.332, ψ2 (tf ) = −8.956. (5.131) Apparently, the larger time horizon tf = 1.1 results in a smaller gain J = 23.357 compared to J = 26.705 for the final time tf = 0.9. We are not aware of any type of sufficient optimality conditions that would apply to the extremal for tf = 1.1, where one control component has a bang-singular-bang structure. Thus one is lead to ask: What is the optimal lifetime of the machine to give maximal gain? This amounts to solving the control problem (5.113)– (5.116) with free final time tf . The solution is very similar to that shown in Figure 5.1. The maintenance control m(t) is bang-bang with one switching time t1 = 0.6523. The optimal final time tf = 0.8633 gives the gain J = 26.833 which slightly improves on J = 26.705 for the final time tf = 0.9. Chapter 6 Second-Order Optimality Conditions for Bang-Bang Control In this chapter, we investigate the pure bang-bang case, where the second-order necessary or sufficient optimality conditions amount to testing the positive (semi)definiteness of a quadratic form on a finite-dimensional critical cone. In Section 6.2, we deduce these conditions from the results obtained in the previous chapter. Although the quadratic conditions turned out to be finite-dimensional, the direct numerical test works only in some special cases. Therefore, in Section 6.3, we study various transformations of the quadratic form and the critical cone which are tailored to different types of control problems in practice. In particular, by a solution to a linear matrix differential equation, the quadratic form can be converted to perfect squares. In Section 6.5, we study second-order optimality conditions for time-optimal control problems with control appearing linearly. In Section 6.6, we show that an approach similar to the above mentioned Riccati equation approach is applicable for such problems. Again, the test requires us to find a solution of a linear matrix differential equation which satisfies certain jump conditions at the switching points. In Section 6.7, we discuss two numerical examples that illustrate the numerical procedure of verifying positive definiteness of the corresponding quadratic forms. In Section 6.8, following [79], we study second-order optimality conditions in a simple, but important class of time-optimal control problems for linear autonomous systems. 6.1 6.1.1 Bang-Bang Control Problems on Nonfixed Time Intervals Optimal Control Problems with Control Appearing Linearly We consider optimal control problems with control appearing linearly. Let x(t) ∈ Rd(x) denote the state variable and u(t) ∈ Rd(u) the control variable in the time interval t ∈ = [t0 , tf ] with a nonfixed initial time t0 and final time tf . We shall refer to the following control problem (6.1)–(6.3) as the basic bang-bang control problem, or briefly, the basic problem: Minimize J := J (t0 , x(t0 ), tf , x(tf )) 255 (6.1) 256 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control subject to the constraints ẋ(t) = f (t, x(t), u(t)), u(t) ∈ U , F (t0 , x(t0 ), tf , x(tf )) ≤ 0, (t0 , x(t0 ), tf , x(tf )) ∈ P , (t, x(t)) ∈ Q, t0 ≤ t ≤ tf , K(t0 , x(t0 ), tf , x(tf )) = 0, (6.2) (6.3) where the control variable appears linearly in the system dynamics, f (t, x, u) = a(t, x) + B(t, x)u. (6.4) Here, F , K, and a are vector functions, B is a d(x) × d(u) matrix function, P ⊂ R2+2d(x) , Q ⊂ R1+d(x) are open sets, and U ⊂ Rd(u) is a convex polyhedron. The functions J , F , and K are assumed to be twice continuously differentiable on P , and the functions a, B are twice continuously differentiable on Q. The dimensions of F and K are denoted by d(F ) and d(K). We shall use the abbreviations x0 = x(t0 ), xf = x(tf ), p = (t0 , x0 , tf , xf ). A trajectory T = (x(t), u(t) | t ∈ [t0 , tf ]) is said to be admissible if x(·) is absolutely continuous, u(·) is measurable bounded, and the pair of functions (x(t), u(t)) on the interval = [t0 , tf ] with the endpoints p = (t0 , x(t0 ), tf , x(tf )) satisfies the constraints (6.2) and (6.3). We set J(T ) := J (t0 , x(t0 ), tf , x(tf )). Obviously, the basic problem (6.1)–(6.3) is a special case of the general problem (5.69)–(5.72) studied in the previous chapter. This special case corresponds to the assumption that the function f does not depend on the variable v, i.e., f = f (t, x, u). Under this assumption it turned out to be possible to obtain certain deeper results than in the general problem. More precisely, we formulate the necessary quadratic Condition A in the problem (6.1)–(6.3), which is a simple consequence of the Condition A in the general problem, whereas the correspondent sufficient quadratic Condition B will be slightly simplified. Let us give the definition of Pontryagin minimum for the basic problem. The trajectory T affords a Pontryagin minimum if there is no sequence of admissible trajectories T n = (x n (t), un (t) | t ∈ [t0n , tfn ]), n = 1, 2, . . . , such that the following properties hold with n = [t0n , tfn ]: (a) (b) (c) J(T n ) < J(T ) for all n and t0n → t0 , tfn → tf for n → ∞; maxn ∩ |x n (t) − x(t)| → 0 for n → ∞; n ∩ |u n (t) − u(t)| dt → 0 for n → ∞. Note that for a fixed time interval , a Pontryagin minimum corresponds to an L1 -local minimum with respect to the control variable. 6.1.2 First-Order Necessary Optimality Conditions Let T = (x(t), u(t) | t ∈ [t0 , tf ]) be a fixed admissible trajectory such that the control u(·) is a piecewise constant function on the interval = [t0 , tf ]. Denote by = {t1 , . . . , ts }, t0 < t1 < · · · < ts < tf the finite set of all discontinuity points ( jump points) of the control u(t). Then ẋ(t) is a piecewise continuous function whose discontinuity points belong to , and hence x(t) is a piecewise smooth function on . We continue to use the notation [u]k = uk+ − uk− for the jump of function u(t) at the point tk ∈ , where uk− = u(tk −), uk+ = u(tk +) are the left- and right-hand values of the control u(t) at tk , respectively. 6.1. Bang-Bang Control Problems on Nonfixed Time Intervals 257 Let us formulate a first-order necessary condition for optimality of the trajectory T , that is, the Pontryagin minimum principle. The Pontryagin function has the form H (t, x, u, ψ) = ψf (t, x, u) = ψa(t, x) + ψB(t, x)u, (6.5) where ψ is a row vector of dimension d(ψ) = d(x), while x, u, f , F , and K are column vectors. The factor of the control u in the Pontryagin function is the switching function φ(t, x, ψ) = Hu (t, x, u, ψ) = ψB(t, x) (6.6) which is a row vector of dimension d(u). The endpoint Lagrange function is l(α0 , α, β, p) = α0 J (p) + αF (p) + βK(p), where α and β are row vectors with d(α) = d(F ) and d(β) = d(K), and α0 is a number. By λ = (α0 , α, β, ψ(·), ψ0 (·)) we denote a tuple of Lagrange multipliers such that ψ(·) : → (Rd(x) )∗ , ψ0 (·) : → R1 are continuous on and continuously differentiable on each interval of the set \ . Let M0 be the set of the normed collections λ satisfying the minimum principle conditions for the trajectory T : α0 ≥ 0, α ≥ 0, αF (p) = 0, α0 + d(F ) αi + d(K) |βj | = 1, (6.7) j =1 i=1 ψ̇ = −Hx , ψ̇0 = −Ht ∀ t ∈ \ , ψ(t0 ) = −lx0 , ψ(tf ) = lxf , ψ0 (t0 ) = −lt0 , ψ0 (tf ) = ltf , min H (t, x(t), u, ψ(t)) = H (t, x(t), u(t), ψ(t)) ∀ t ∈ \ , (6.8) (6.9) (6.10) H (t, x(t), u(t), ψ(t)) + ψ0 (t) = 0 (6.11) u∈U ∀ t ∈ \ . The derivatives lx0 and lxf are taken at the point (α0 , α, β, p), where p = (t0 , x(t0 ), tf , x(tf )), and the derivatives Hx , Ht are evaluated at the point (t, x(t), u(t), ψ(t)). Again, we use the simple abbreviation (t) for indicating all arguments (t, x(t), u(t), ψ(t)), t ∈ \ . Theorem 6.1. If the trajectory T affords a Pontryagin minimum, then the set M0 is nonempty. The set M0 is a finite-dimensional compact set, and the projector λ → (α0 , α, β) is injective on M0 . For each λ ∈ M0 and tk ∈ , let us define the quantity D k (H ). Set (k H )(t) = H (t, x(t), uk+ , ψ(t)) − H (t, x(t), uk− , ψ(t)) = φ(t) [u]k , (6.12) where φ(t) = φ(t, x(t), ψ(t)). For each λ ∈ M0 the following equalities hold: d d (k H )t=t − = (k H )t=t + , k k dt dt k = 1, . . . , s. Consequently, for each λ ∈ M0 the function (k H )(t) has a derivative at the point tk ∈ . Set d D k (H ) = − (k H )(tk ). dt 258 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Then, for each λ ∈ M0 , the minimum condition (6.10) implies the inequality D k (H ) ≥ 0, k = 1, . . . , s. (6.13) As we know, the value D k (H ) can be written in the form D k (H ) = −Hxk+ Hψk− + Hxk− Hψk+ − [Ht ]k = ψ̇ k+ ẋ k− − ψ̇ k− ẋ k+ + [ψ0 ]k , where Hxk− and Hxk+ are the left- and right-hand values of the function Hx (t, x(t), u(t), ψ(t)) at tk , respectively, [Ht ]k is a jump of the function Ht (t) at tk , etc. It also follows from the above representation that we have D k (H ) = −φ̇(tk ±)[u]k , (6.14) where the values on the right-hand side agree for the derivative φ̇(tk +) from the right and the derivative φ̇(tk −) from the left. In the case of a scalar control u, the total derivative φt + φx ẋ + φψ ψ̇ does not contain the control variable explicitly and hence the derivative φ̇(t) is continuous at tk . Proposition 6.2. For any λ ∈ M0 , we have lx0 ẋ(t0 ) + lt0 = 0, lxf ẋ(tf ) + ltf = 0. (6.15) Proof. The equalities (6.15) follow from the equality ψ(t)ẋ(t) + ψ0 (t) = 0 evaluated for t = t0 and t = tf together with the transversality conditions ψ(t0 ) = −lx0 , 6.1.3 ψ0 (t0 ) = −lt0 , ψ(tf ) = lxf , ψ0 (tf ) = ltf . Strong Minimum As already mentioned in Section 2.7, any control problem with a cost functional in intet gral form J = t0f f0 (t, x(t), u(t)) dt can be brought to the canonical form (6.1) by introducing a new state variable y defined by the state equation ẏ = f0 (t, x, u), y(t0 ) = 0. This yields the cost function J = y(tf ). The control variable is assumed to appear linearly in the function f0 , f0 (t, x, u) = a0 (t, x) + B0 (t, x)u. (6.16) It follows that the adjoint variable ψ y associated with the new state variable y is given by ψ y (t) ≡ α0 which yields the Pontryagin function (6.5) in the form H (t, x, ψ, u) = α0 f0 (t, x, u) + ψf (t, x, u) = α0 a0 (t, x) + ψa(t, x) + (α0 B0 (t, x) + ψB(t, x))u. (6.17) Hence, the switching function is given by φ(t, x, ψ) = α0 B0 (t, x) + ψB(t, x), φ(t) = φ(t, x(t), ψ(t)). (6.18) 6.2. Quadratic Necessary and Sufficient Optimality Conditions 259 The component y was called an unessential component in the augmented problem. The general definition was given in Section 5.2.3: the state variable xi is called unessential if the function f does not depend on xi and if the functions F , J , and K are affine in xi0 = xi (t0 ) and xi1 = xi (tf ). Let x denote the vector of all essential components of state vector x. Now we can define a strong minimum in the basic problem. We say that the trajectory T affords a strong minimum if there is no sequence of admissible trajectories T n = (x n (t), un (t) | t ∈ [t0n , tfn ]), n = 1, 2, . . . , such that (a) (b) (c) J(T n ) < J(T ); t0n → t0 , tfn → tf , x n (t0 ) → x(t0 ) (n → ∞); maxn ∩ |x n (t) − x(t)| → 0 (n → ∞), where n = [t0n , tfn ]. The strict strong minimum is defined in a similar way, with the strict inequality (a) replaced by the nonstrict one and the trajectory T n required to be different from T for each n. 6.1.4 Bang-Bang Control For a given extremal trajectory T = { (x(t), u(t)) | t ∈ } with a piecewise constant control u(t) we say that u(t) is a strict bang-bang control if there exists λ = (α0 , α, β, ψ, ψ0 ) ∈ M0 such that (6.19) Arg min φ(t)u = [u(t−), u(t+)], t ∈ [t0 , tf ], u ∈U where [u(t−), u(t+)] denotes the line segment spanned by the vectors u(t−) and u(t+) in Rd(u) and φ(t) := φ(t, x(t), ψ(t)) = ψ(t)B(t, x(t)). Note that [u(t−), u(t+)] is a singleton {u(t)} at each continuity point of the control u(t) with u(t) being a vertex of the polyhedron U . Only at the points tk ∈ does the line segment [uk− , uk+ ] coincide with an edge of the polyhedron. As it was already mentioned in Section 5.1.7, if the control is scalar, d(u) = 1, and U = [umin , umax ], then the strict bang-bang property is equivalent to φ(t) = 0 for all t ∈ \ which yields the control law umin if φ(t) > 0 u(t) = ∀ t ∈ \ . (6.20) umax if φ(t) < 0 For vector-valued control inputs, condition (6.19) imposes further restrictions. For example, if U is the unit cube in Rd(u) , condition (6.19) precludes simultaneous switching of the control components; the case of simultaneous switching was studied in Felgenhauer [31]. This property holds in many examples. The condition (6.19) is indispensable in the sensitivity analysis of optimal bang-bang controls. 6.2 Quadratic Necessary and Sufficient Optimality Conditions In this section, we shall formulate a quadratic necessary optimality condition of a Pontryagin minimum for given bang-bang control. A strengthening of this quadratic condition yields a quadratic sufficient condition for a strong minimum. These quadratic conditions are based on the properties of a quadratic form on the critical cone. 260 6.2.1 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Critical Cone For a given trajectory T , we introduce the space Z() and the critical cone K ⊂ Z(). Denote by P C 1 (, Rd(x) ) the space of piecewise continuous functions x̄(·) : → Rd(x) , continuously differentiable on each interval of the set \ . For each x̄ ∈ P C 1 (, Rd(x) ) and for tk ∈ , we set x̄ k− = x̄(tk −), x̄ k+ = x̄(tk +), [x̄]k = x̄ k+ − x̄ k− . Set z̄ = (t¯0 , t¯f , ξ̄ , x̄), where t¯0 , t¯f ∈ R1 , ξ̄ ∈ Rs , x̄ ∈ P C 1 (, Rd(x) ). Thus, z̄ ∈ Z() := R2 × Rs × P C 1 (, Rd(x) ). For each z̄, we set x̄¯f = x̄(tf ) + t¯f ẋ(tf ), x̄¯0 = x̄(t0 ) + t¯0 ẋ(t0 ), p̄¯ = t¯0 , x̄¯0 , t¯f , x̄¯f . (6.21) The vector p̄¯ is considered as a column vector. Note that t¯0 = 0, respectively, t¯f = 0, for fixed initial time t0 , respectively, final time tf . Let IF (p) = {i ∈ {1, . . . , d(F )} | Fi (p) = 0} be the set of indices of all active endpoint inequalities Fi (p) ≤ 0 at the point p = (t0 , x(t0 ), tf , x(tf )). Denote by K the set of all z̄ ∈ Z() satisfying the following conditions: J (p)p̄¯ ≤ 0, Fi (p)p̄¯ ≤ 0 ∀ i ∈ IF (p), K (p)p̄¯ = 0, ˙ = fx (t, x(t), u(t))x̄(t), [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, x̄(t) (6.22) (6.23) where p = (t0 , x(t0 ), tf , x(tf )). It is obvious that K is a convex finite-dimensional and finitefaced cone in the space Z(). We call it the critical cone. Each element z̄ ∈ K is uniquely defined by numbers t¯0 , t¯f , a vector ξ̄ , and the initial value x̄(t0 ) of the function x̄(t). Proposition 6.3. For any λ ∈ M0 and z̄ ∈ K, we have lx0 x̄(t0 ) + lxf x̄(tf ) = 0. (6.24) Proof. Integrating the equality ψ(x̄˙ − fx x̄) = 0 on [t0 , tf ] and using the adjoint equation t t ψ̇ = −ψfx we obtain t0f dtd (ψ x̄) dt = 0, whence (ψ x̄)|tf0 − sk=1 [ψ x̄]k = 0. From the jump conditions [x̄]k = [ẋ]k ξ̄k and the equality ψ(t)ẋ(t) + ψ0 (t) = 0 it follows that [ψ x̄]k = t ψ(tk )[ẋ]k ξ̄k = [ψ ẋ]k ξ̄k = −[ψ0 ]k ξ̄k = 0 for all k. Then the equation (ψ x̄)|tf0 = 0, together with the transversality conditions ψ(t0 ) = −lx0 and ψ(tf ) = lxf , implies (6.24). Proposition 6.4. For any λ ∈ M0 and z̄ ∈ K we have α0 J (p)p̄¯ + s αi Fi (p)p̄¯ + βK (p)p̄¯ = 0. (6.25) i=1 Proof. For λ ∈ M0 and z̄ ∈ K, we have, by Propositions 6.2 and 6.3, t¯0 (lx0 ẋ(t0 ) + lt0 ) + t¯f (lxf ẋ(tf ) + ltf ) + lx0 x̄(t0 ) + lxf x̄(tf ) = 0. Now using the equalities x̄¯0 = x̄(t0 ) + t¯0 ẋ(t0 ), x̄¯f = x̄(tf ) + t¯f ẋ(tf ), and p̄¯ = t¯0 , x̄¯0 , t¯f , x̄¯f , we get lp p̄¯ = 0 which is equivalent to condition (6.25). 6.2. Quadratic Necessary and Sufficient Optimality Conditions 261 Two important properties of the critical cone follow from Proposition 6.4. Proposition 6.5. For any λ ∈ M0 and z̄ ∈ K, we have α0 J (p)p̄¯ = 0 and αi Fi (p)p̄¯ = 0 for all i ∈ IF (p). Proposition 6.6. Suppose that there exist λ ∈ M0 with α0 > 0. Then adding the equalities αi Fi (p)p̄¯ = 0 for all i ∈ IF (p) to the system (6.22), (6.23) defining K, one can omit the inequality J (p)p̄¯ ≤ 0 in that system without affecting K. Thus, K is defined by condition (6.23) and by the condition p̄¯ ∈ K0 , where K0 is the cone in R2d(x)+2 given by (6.22). But if there exists λ ∈ M0 with α0 > 0, then we can put K0 = {p̄¯ ∈ Rd(x)+2 | Fi (p)p̄¯ ≤ 0, αi Fi (p)p̄¯ = 0 ∀ i ∈ IF (p), K (p)p̄¯ = 0}. (6.26) If, in addition, αi > 0 holds for all i ∈ IF (p), then K0 is a subspace in Rd(x)+2 . An explicit representation of the variations x̄(t) in (6.23) is obtained as follows. For each k = 1, . . . , s, define the vector functions y k (t) as the solutions to the system ẏ = fx (t)y, y(tk ) = [ẋ]k , t ∈ [tk , tf ]. For t < tk we put y k (t) = 0 which yields the jump [y k ]k = [ẋ]k . Moreover, define y 0 (t) as the solution to the system ẏ = fx (t)y, y(t0 ) = x̄(t0 ) =: x̄0 . By the superposition principle for linear ODEs it is obvious that we have x̄(t) = s y k (t)ξ̄k + y 0 (t) k=1 from which we obtain the representation x̄¯f = s y k (tf )ξ̄k + y 0 (tf ) + ẋ(tf )t¯f . (6.27) k=1 Furthermore, denote by x(t; t1 , . . . , ts ) the solution of the state equation (6.2) using the values of the optimal bang-bang control with switching points t1 , . . . , ts . It easily follows from elementary properties of ODEs that the partial derivatives of state trajectories with respect to the switching points are given by ∂x (t; t1 , . . . , ts ) = −y k (t) ∂tk for t ≥ tk , k = 1, . . . , s. (6.28) This gives the following expression for x̄(t): x̄(t) = − s ∂x (t)ξ̄k + y 0 (t). ∂tk (6.29) k=1 In a special case that frequently arises in practice, we can use these formulas to show that K = {0}. This property then yields a first-order sufficient condition in view of Theorem 6.10. 262 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Namely, consider the following problem with an integral cost functional, where the initial time t0 = tˆ0 is fixed, while the final time tf is free and where the initial and final values of the state variables are given: Minimize tf J= f0 (t, x, u)dt (6.30) t0 subject to ẋ = f (t, x, u), x(t0 ) = x̂0 , x(tf ) = x̂f , u(t) ∈ U , (6.31) where f is defined by (6.4), and f0 is defined by (6.16). In the definition of K we then have t¯0 = 0, x̄(t0 ) = 0, x̄¯f = 0. The condition x̄(t0 ) = 0 implies that y 0 (t) ≡ 0, whereas the condition x̄¯f = 0 yields in view of the representation (6.27), s y k (tf )ξ̄k + ẋ(tf )t¯f = 0. k=1 This equation leads to the following statement. Proposition 6.7. In problem (6.30), (6.31), assume that the s + 1 vectors y k (tf ) = − ∂x (tf ) (k = 1, . . . , s), ẋ(tf ) ∂tk are linearly independent. Then the critical cone is K = {0}. We conclude this subsection with a special property of the critical cone for timeoptimal control problems with fixed initial time and state, tf → min, ẋ = f (t, x, u), u ∈ U , t0 = tˆ0 , x(t0 ) = x̂0 , K(x(tf )) = 0, (6.32) where f is defined by (6.4). The following result will be used in Section 6.2.3; cf. Proposition 6.11. Proposition 6.8. Suppose that there exists (ψ0 , ψ) ∈ M0 such that α0 > 0. Then t¯f = 0 holds for each z̄ = (t¯f , ξ̄ , x̄) ∈ K. Proof. For arbitrary λ ∈ M0 and z̄ = (t¯f , ξ̄ , x̄) ∈ K we infer from the proof of Proposition 6.3 that ψ(t)x̄(t) is a constant function on [t0 , tf ]. In view of the relations ψ(tf ) = βKxf (x(tf )), Kxf (x(tf ))x̄¯f = 0, x̄¯f = x̄(tf ) + ẋ(tf )t¯f , we get 0 = (ψ x̄)(t0 ) = (ψ x̄)(tf ) = ψ(tf )(x̄¯f − ẋ(tf )t¯f ) = −ψ(tf )ẋ(tf )t¯f = ψ0 (tf )t¯f . Since ψ0 (tf ) = α0 > 0, this relation yields t¯f = 0. In the case α0 > 0, we note as a consequence that the critical cone is a subspace defined by the conditions x̄˙ = fx (t)x̄, [x̄]k = [ẋ]k ξ̄k (k = 1, . . . , s), (6.33) t¯0 = t¯f = 0, x̄(t0 ) = 0, Kxf (x(tf ))x̄(tf ) = 0. 6.2. Quadratic Necessary and Sufficient Optimality Conditions 6.2.2 263 Quadratic Necessary Optimality Conditions Let us introduce a quadratic form on the critical cone K defined by the conditions (6.22), (6.23). For each λ ∈ M0 and z̄ ∈ K, we set ¯ p̄¯ + (λ, z̄) = Ap̄, s k ξ̄k + D k (H )ξ̄k2 + 2[Hx ]k x̄av k=1 tf Hxx x̄(t), x̄(t) dt, (6.34) t0 where ¯ p̄¯ + 2ψ̇(t0 )x̄¯0 t¯0 + (ψ̇0 (t0 ) − ψ̇(t0 )ẋ(t0 ))t¯02 ¯ p̄¯ = lpp p̄, Ap̄, − 2ψ̇(tf )x̄¯f t¯f − (ψ̇0 (tf ) − ψ̇(tf )ẋ(tf ))t¯f2 , (6.35) lpp = lpp (α0 , α, β, p), p = (t0 , x(t0 ), tf , x(tf )), Hxx = Hxx (t, x(t), u(t), ψ(t)), 1 k x̄av = (x̄ k− + x̄ k+ ). 2 Note that for a problem on a fixed time interval [t0 , tf ] we have t¯0 = t¯f = 0 and, hence, ¯ p̄¯ = lpp p̄, p̄ . The following theorem gives the the quadratic form (6.35) reduces to Ap̄, main second-order necessary condition of optimality. Theorem 6.9. If the trajectory T affords a Pontryagin minimum, then the following Condition A holds: The set M0 is nonempty and maxλ∈M0 (λ, z̄) ≥ 0 for all z̄ ∈ K. 6.2.3 Quadratic Sufficient Optimality Conditions A natural strengthening of the necessary Condition A turns out to be a sufficient optimality condition not only for a Pontryagin minimum, but also for a strong minimum. Theorem 6.10. Let the following Condition B be fulfilled for the trajectory T : (a) there exists λ ∈ M0 such that D k (H ) > 0, k = 1, . . . , s, and condition (6.19) holds (i.e., u(t) is a strict bang-bang control), (b) maxλ∈M0 (λ, z̄) > 0 for all z̄ ∈ K \ {0}. Then T is a strict strong minimum. Note that the condition (b) is automatically fulfilled, if K = {0}, which gives a firstorder sufficient condition for a strong minimum in the problem. A specific situation where K = {0} holds was described in Proposition 6.7. Also note that the condition (b) is automatically fulfilled if there exists λ ∈ M0 such that (λ, z̄) > 0 ∀ z̄ ∈ K \ {0}. (6.36) Example: Resource allocation problem. Let x(t) be the stock of a resource and let the control u(t) be the investment rate. The control problem is to maximize the overall consumption tf 0 x(t)(1 − u(t)) dt 264 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control on a fixed time interval [0, tf ] subject to ẋ(t) = x(t) u(t), x(0) = x0 > 0, 0 ≤ u(t) ≤ 1. The Pontryagin function (6.5) for the equivalent minimization problem and the switching function are given by H = α0 x(u − 1) + ψ xu = −α0 x + φ u, φ(x, ψ) = x(α0 + ψ). We can put α0 = 1, since the terminal state x(tf ) is free. A straightforward discussion of the minimum principle then shows that the optimal control has exactly one switching point at t1 = tf − 1 for tf > 1, 1, 0 ≤ t < t1 u(t) = , 0, t1 ≤ t ≤ tf (x0 et , −e−(t−t1 ) ), 0 ≤ t ≤ t1 . (x(t), ψ(t)) = (x0 et1 , t − tf ), t1 ≤ t ≤ tf The switching function is φ(t) = x(t)(1 + ψ(t)) which yields φ̇(t1 ) = x0 et1 = 0. Here we have k = 1, [u]1 = −1, and thus obtain D 1 (H ) = −φ̇(t1 )[u]1 = φ̇(t1 ) > 0 in view of (6.12) and (6.14). Hence, condition (a) of Theorem 6.10 holds. Checking condition (b) is rather simple, since the quadratic form (6.34) reduces here to (λ, z̄) = D 1 (H )ξ̄12 . This relation follows from Hxx ≡ 0 and [Hx ]1 = (1 + ψ(t1 ))[u]1 = 0 and the fact that the quadratic form (6.35) vanishes. Note that the above control problem cannot be handled in the class of convex optimization problems. This means that the necessary conditions do not automatically imply optimality of the computed solution. We conclude this subsection with the case of a time-optimal control problem (6.32) with a single switching point, i.e., s = 1. Assume that α0 > 0 for a given λ ∈ M0 . Then by Proposition 6.8, we have t¯f = 0, and thus the critical cone is the subspace defined by (6.33). In this case, the quadratic form can be computed explicitly as follows. Denote by y(t), t ∈ [t1 , tf ], the solution to the Cauchy problem ẏ = fx y, y(t1 ) = [ẋ]1 . The following assertion is obvious: If (ξ̄ , x̄) ∈ K, then x̄(t) = 0 for t ∈ [t0 , t1 ) and x̄(t) = y(t)ξ̄ for t ∈ (t1 , tf ]. Therefore, the inequality Kxf (x(tf ))y(tf ) = 0 would imply K = {0}. Consider now the case Kxf (x(tf ))y(tf ) = 0. This condition always holds for time-optimal problems with a scalar function K and α0 > 0. Indeed, the condition dtd (ψy) = 0 implies (ψy)(t) = const in [t1 , tf ], whence (ψy)(tf ) = (ψy)(t1 ) = ψ(t1 )[ẋ]1 = φ(t1 )[u]1 = 0. Using the transversality condition ψ(tf ) = βKxf (x(tf )) and the inequality β = 0 (if β = 0, then ψ(tf ) = 0, and hence ψ(t) = 0 and ψ0 (t) = 0 in [t0 , tf ]), we see that the equality (ψy)(tf ) = 0 implies the equality Kxf (x(tf ))y(tf ) = 0. Observe now that the cone K is a one-dimensional subspace, on which the quadratic form has the representation = ρ ξ̄ 2 , where tf ρ := D 1 (H ) − [ψ̇]1 [ẋ]1 + (y(t))∗ Hxx (t)y(t) dt + (y(tf ))∗ (βK)xf xf y(tf ). (6.37) t1 6.2. Quadratic Necessary and Sufficient Optimality Conditions 265 This gives the following result. Proposition 6.11. Suppose that we have found an extremal for the time-optimal control problem (6.32) that has one switching point and satisfies α0 > 0 and Kxf (x(tf ))y(tf ) = 0. Then the inequality ρ > 0 with ρ defined in (6.37) is equivalent to the positive definiteness of on K. 6.2.4 Proofs of Quadratic Conditions It was already mentioned that problem (6.1)–(6.3) is a special case of the general problem (5.69)–(5.72). It is easy to check that in problem (6.1)–(6.3) we obtain the set M0 and the critical cone K as the special cases of these sets in the general problem. Let us compare the quadratic forms. It suffices to show that the endpoint quadratic ¯ p̄¯ (see (6.35)) can be transformed into the endpoint quadratic form ωe (λ, z̄) in form Ap̄, (5.86) if relations (6.21) hold. Indeed, we have ¯ p̄¯ Ap̄, := = = ¯ p̄¯ + 2ψ̇(t0 )x̄¯0 t¯0 + (ψ̇0 (t0 ) − ψ̇(t0 )ẋ(t0 ))t¯2 lpp p̄, 0 − 2ψ̇(tf )x̄¯f t¯f − (ψ̇0 (tf ) − ψ̇(tf )ẋ(tf ))t¯f2 ¯ p̄¯ + 2ψ̇(t0 )(x̄0 + t¯0 ẋ(t0 ))t¯0 + (ψ̇0 (t0 ) − ψ̇(t0 )ẋ(t0 ))t¯2 lpp p̄, 0 − 2ψ̇(tf )(x̄1 + t¯f ẋ(tf ))t¯f − (ψ̇0 (tf ) − ψ̇(tf )ẋ(tf ))t¯f2 ¯ p̄¯ − 2ψ̇(tf )x̄f t¯f − ψ̇(tf )ẋ(tf ) + ψ̇0 (tf ) t¯2 lpp p̄, f + 2ψ̇(t0 )x̄0 t¯0 + ψ̇(t0 )ẋ(t0 ) + ψ̇0 (t0 ) t¯02 =: ωe (λ, z̄). Thus, Theorem 6.9, which gives necessary quadratic conditions in the problem (6.1)–(6.3), is a consequence of Theorem 5.22. Now let us proceed to sufficient quadratic conditions in the same problem. Here the set M0+ consists of all those elements λ ∈ M0 for which condition (6.19) is fulfilled, and the set Leg+ (M0+ ) consists of all those elements λ ∈ M0+ for which D k (H ) > 0, k = 1, . . . , s. (6.38) Leg+ (M0+ ) = M. Denote for brevity Thus the set M consists of all those elements λ ∈ M0 for which (6.19) and (6.38) are fulfilled. Let us also note that the strict bounded strong minimum in the problem (6.1)–(6.3) is equivalent to the strict strong minimum, since U is a compact set. Thus Theorem 5.24 implies the following result. Theorem 6.12. For a trajectory T in the problem (6.1)–(6.3) let the following Condition B be fulfilled: The set M is nonempty and there exist a nonempty compact M ⊂ M and ε > 0 such that (6.39) max (λ, z̄) ≥ ε γ̄ (z̄) ∀ z̄ ∈ K, λ∈M where γ̄ (z̄) = t¯02 + t¯f2 + ξ̄ , ξ̄ + x̄(t0 ), x̄(t0 ) . Then the trajectory T affords a strict strong minimum in this problem. Remarkably, the fact that the critical cone K in the problem (6.1)–(6.3) is finitedimensional (since each element z̄ = (t¯0 , t¯f , ξ̄ , x̄) is uniquely defined by the parameters 266 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control t¯0 , t¯f , ξ̄ , x̄(t0 )) implies that condition B in Theorem 6.12 is equivalent to the following, generally weaker condition. Condition B0 . The set M is nonempty and max (λ, z̄) > 0 λ∈M0 ∀ z̄ ∈ K \ {0}. (6.40) Lemma 6.13. For a trajectory T in the problem (6.1)–(6.3), Condition B is equivalent to the Condition B0 . Proof. As we pointed out, Condition B always implies Condition B0 . We will show that, in our case, the inverse assertion also holds. Let Condition B0 be fulfilled. Let S1 (K) be the set of elements z̄ ∈ K satisfying the Condition γ̄ (z̄) = 1. Then max (λ, z̄) > 0 λ∈M0 ∀ z̄ ∈ S1 (K). (6.41) Recall that M0 is a finite-dimensional compact set. It is readily verified that the relative interior of the cone con M0 generated by M0 is contained in the cone con M generated by M, i.e., reint(con M0 ) ⊂ con M. Combined with (6.41) this implies that for any element z̄ ∈ S1 (K) there exist λ ∈ M and a neighborhood Oz̄ ⊂ S1 (K) of element z̄ such that (λ, ·) > 0 on Oz̄ . The family of neighborhoods {Oz̄ } forms an open covering of the compact set S1 (K). Select a finite subcovering. To this subcovering there corresponds a finite compact set M = {λ1 , . . . , λr } such that max (λ, z̄) > 0 λ∈M ∀ z̄ ∈ S1 (K), and consequently, due to compactness of the cross-section S1 (K), max (λ, z̄) > ε λ∈M ∀ z̄ ∈ S1 (K), for some ε > 0. Hence Condition B follows. Theorem 6.12 and Lemma 6.13 imply Theorem 6.10, where B = B0 . In what follows, for the problems of the type of basic problem (6.1)–(6.3), by Condition B we will mean Condition B0 . 6.3 Sufficient Conditions for Positive Definiteness of the Quadratic Form on the Critical Cone K Assume that the condition (a) of Theorem 6.10 is fulfilled for the trajectory T . Let λ ∈ M0 be a fixed element (possibly, different from that in condition (a)) and let = (λ, ·) be the quadratic form (6.34) for this element. According to Theorem 6.10, the positive definiteness of on the critical cone K is a sufficient condition for a strict strong minimum. Recall that K is defined by (6.23) and the condition p̄¯ ∈ K0 , where p̄¯ = (t¯0 , x̄¯0 , t¯f , x̄¯f ), x̄¯0 = x̄(t0 ) + t¯0 ẋ(t0 ), x̄¯f = x̄(tf ) + t¯f ẋ(tf ). The cone K0 is defined by (6.26) in the case α0 > 0 and by (6.22) in the general case. 6.3. Sufficient Conditions for Positive Definiteness of the Quadratic Form 267 Now our aim is to find sufficient conditions for the positive definiteness of the quadratic form on the cone K. In what follows, we shall use the ideas and results presented in Chapter 4 (see also [69]). 6.3.1 Q-Transformation of on K Let Q(t) be a symmetric matrix on [t0 , tf ] with piecewise continuous entries which are absolutely continuous on each interval of the set [t0 , tf ] \ . Therefore, Q may have a jump at each point tk ∈ . For z̄ ∈ K, formula (5.100) holds: tf t0 tf s d [Qx̄, x̄ ]k , Qx̄, x̄ dt = Qx̄, x̄ − dt t0 k=1 where [Qx̄, x̄ ]k is the jump of the function Qx̄, x̄ at the point tk ∈ . Using the equation x̄˙ = fx x̄ with fx = fx (t, x(t), u(t)), we obtain s k=1 [Qx̄, x̄ ]k + tf t0 (Q̇ + fx∗ Q + Qfx )x̄, x̄ dt − Qx̄, x̄ (tf ) + Qx̄, x̄ (t0 ) = 0, where the asterisk denotes transposition. Adding this zero form to and using the equality [Hx ]k = −[ψ̇]k , we get ¯ p̄¯ − Qx̄, x̄ (tf ) + Qx̄, x̄ (t0 ) = Ap̄, s k ξ̄k + [Qx̄, x̄ ]k + D k (H )ξ̄k2 − 2[ψ̇]k x̄av + k=1 tf t0 (6.42) (Hxx + Q̇ + fx∗ Q + Qfx )x̄, x̄ dt. We shall call this formula the Q-transformation of on K. In order to eliminate the integral term in , we assume that Q(t) satisfies the following linear matrix differential equation: Q̇ + fx∗ Q + Qfx + Hxx = 0 on [t0 , tf ] \ . (6.43) It is interesting to note that the same equation is obtained from the modified Riccati equation in Maurer and Pickenhain [73, Equation (47)], when all control variables are on the boundary of the control constraints. Using (6.43) the quadratic form (6.42) reduces to = ω0 + s ωk , (6.44) k=1 k ξ̄k + [Qx̄, x̄ ]k , ωk := D k (H )ξ̄k2 − 2[ψ̇]k x̄av ¯ p̄¯ − Qx̄, x̄ (tf ) + Qx̄, x̄ (t0 ). ω0 := Ap̄, k = 1, . . . , s, (6.45) (6.46) 268 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Thus, we have proved the following statement. Proposition 6.14. Let Q(t) satisfy the linear differential equation (6.43) on [t0 , tf ] \ . Then for each z̄ ∈ K the representation (6.44) holds. Now our goal is to derive conditions such that ωk > 0, k = 0, . . . , s, holds on K \ {0}. To this end, as in Section 4.1.6, we shall express ωk via the vector (ξ̄k , x̄ k− ). We use the formula (6.47) x̄ k+ = x̄ k− + [ẋ]k ξ̄k , which implies Qk+ x̄ k+ , x̄ k+ = Qk+ x̄ k− , x̄ k− + 2Qk+ [ẋ]k , x̄ k− ξ̄k + Qk+ [ẋ]k , [ẋ]k ξ̄k2 . Consequently, [Qx̄, x̄ ]k = [Q]k x̄ k− , x̄ k− + 2Qk+ [ẋ]k , x̄ k− ξ̄k + Qk+ [ẋ]k , [ẋ]k ξ̄k2 . k = x̄ k− + 1 [ẋ]k ξ̄ in the definition (6.45) of ω , we Using this relation together with x̄av k k 2 obtain ωk = D k (H ) + ([ẋ]k )∗ Qk+ − [ψ̇]k [ẋ]k ξ̄k2 (6.48) +2 ([ẋ]k )∗ Qk+ − [ψ̇]k x̄ k− ξ̄k + (x̄ k− )∗ [Q]k x̄ k− . Here [ẋ]k and x̄ k− are column vectors while ([ẋ]k )∗ , (x̄ k− )∗ , and [ψ̇]k are row vectors. By putting qk+ = ([ẋ]k )∗ Qk+ − [ψ̇]k , bk+ = D k (H ) + (qk+ )[ẋ]k , (6.49) we get ωk = (bk+ )ξ̄k2 + 2(qk+ )x̄ k− ξ̄k + (x̄ k− )∗ [Q]k x̄ k− . Note that ωk is a quadratic form in the variables (ξ̄k , x̄ k− ) with the matrix bk+ qk+ Mk+ = , (qk+ )∗ [Q]k (6.50) (6.51) where qk+ is a row vector and (qk+ )∗ is a column vector. Similarly, using the relation x̄ k− = x̄ k+ − [ẋ]k ξ̄k , we obtain [Qx̄, x̄ ]k = [Q]k x̄ k+ , x̄ k+ + 2Qk− [ẋ]k , x̄ k+ ξ̄ k − Qk− [ẋ]k , [ẋ]k ξ̄k2 . k = x̄ k+ − 1 [ẋ]k ξ̄ leads to the representation (cf. This formula together with the relation x̄av k 2 formula (4.160)) ωk = (bk− )ξ̄k2 + 2(qk− )x̄ k+ ξ̄k + (x̄ k+ )∗ [Q]k x̄ k+ , where qk− = ([ẋ]k )∗ Qk− − [ψ̇]k , bk− = D k (H ) − (qk− )[ẋ]k . (6.52) (6.53) 6.3. Sufficient Conditions for Positive Definiteness of the Quadratic Form We consider (6.52) as a quadratic form in the variables (ξ̄k , x̄ k+ ) with the matrix qk− bk− . Mk− = (qk− )∗ [Q]k 269 (6.54) Since the right-hand sides of equalities (6.50) and (6.52) are connected by relation (6.47), the following statement obviously holds. Proposition 6.15. For each k = 1, . . . , s, the positive (semi)definiteness of the matrix Mk− is equivalent to the positive (semi)definiteness of the matrix Mk+ . Now we can prove two theorems. Theorem 6.16. Assume that s = 1. Let Q(t) be a solution to the linear differential equation (6.43) on [t0 , tf ] \ which satisfies two conditions: (i) The matrix M1+ is positive semidefinite; (ii) the quadratic form ω0 is positive on the cone K0 \ {0}. Then is positive on K \ {0}. Proof. Take an arbitrary element z̄ ∈ K. Conditions (i) and (ii) imply that ωk ≥ 0 for k = 0, 1, and hence = ω0 + ω1 ≥ 0 for this element. Assume now that = 0. Then ωk = 0 for k = 0, 1. By virtue of (ii) the equality ω0 = 0 implies that t¯0 = t¯f = 0 and x̄(t0 ) = x̄(tf ) = 0. The last two equalities and the equation x̄˙ = fx x̄ show that x̄(t) = 0 in [t0 , t1 ) ∪ (t1 , tf ]. Now using formula (6.45) for ω1 = 0, as well as the conditions D 1 (H ) > 0 and x̄ 1− = x̄ 1+ = 0, we obtain that ξ̄1 = 0. Consequently, we have z̄ = 0 which means that is positive on K \ {0}. Theorem 6.17. Assume that s ≥ 2. Let Q(t) be a solution to the linear differential equation (6.43) on [t0 , tf ] \ which satisfies the following conditions: (a) The matrix Mk+ is positive semidefinite for each k = 1, . . . , s; (b) bk+ := D k (H ) + (qk+ )[ẋ]k > 0 for each k = 1, . . . , s − 1; (c) the quadratic form ω0 is positive on the cone K0 \ {0}. Then is positive on K \ {0}. Proof. Take an arbitrary element z̄ ∈ K. Conditions (a) and (c) imply that ωk ≥ 0 for k = 0, 1, . . . , s, and then ≥ 0 for this element. Assume that = 0. Then ωk = 0 for k = 0, 1, . . . , s. By virtue of (c) the equality ω0 = 0 implies that t¯0 = t¯f = 0 and x̄(t0 ) = x̄(tf ) = 0. The last two equalities and the equation x̄˙ = fx x̄ show that x̄(t) = 0 in [t0 , t1 ) ∪ (ts , tf ] and hence x̄ 1− = x̄ s+ = 0. The conditions ω1 = 0, x̄ 1− = 0, and b1+ > 0 by formula (6.50) (with k = 1) yield ξ̄1 = 0. Then [x̄]1 = 0 and hence x̄ 1+ = 0. The last equality and the equation x̄˙ = fx x̄ show that x̄(t) = 0 in (t1 , t2 ) and hence x̄ 2− = 0. Similarly, the conditions ω2 = 0, x̄ 2− = 0 and b2+ > 0 by formula (6.50) (with k = 2) imply that ξ̄2 = 0 and x̄(t) = 0 in (t2 , t3 ). Therefore, x̄ 3− = 0, etc. Continuing this process we get that x̄ ≡ 0 and ξ̄k = 0 for k = 1, . . . , s −1. Now using formula (6.45) for ωs = 0, as well as the conditions D s (H ) > 0 and x̄ ≡ 0, we obtain that ξ̄s = 0. Consequently, z̄ = 0, and hence is positive on K \ {0}. 270 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Similarly, using representation (6.52) for ωk we can prove the following statement. Theorem 6.18. Let Q(t) be a solution to the linear differential equation (6.43) on [t0 , tf ] \ which satisfies the following conditions: (a ) The matrix Mk− is positive semidefinite for each k = 1, . . . , s; (b ) bk− := D k (H ) − (qk− )[ẋ]k > 0 for each k = 2, . . . , s (if s = 1, then this condition is not required); (c) the quadratic form ω0 is positive on the cone K0 \ {0}. Then is positive on K \ {0}. Remark. Noble and Schättler [80] and Ledzewicz and Schättler [53] use also the linear ODE (6.43) for deriving sufficient conditions. It would be of interest to relate their approach to the results in Theorem 6.18. 6.3.2 Case of Fixed Initial Values t0 and x(t0 ) Consider the problem (6.1)–(6.3) with additional constraints t0 = tˆ0 and x(t0 ) = x̂0 . In this case we have additional equalities in the definition of the critical cone K: t¯0 = 0 and x̄¯0 := x̄(t0 ) + t¯0 ẋ(t0 ) = 0, whence x̄(t0 ) = 0. The last equality and the equation x̄˙ = fx x̄ show that x̄(t) = 0 in [t0 , t1 ), whence x̄ 1− = 0. From definitions (6.46) and (6.35) of ω0 and ¯ p̄¯ , respectively, it follows that for each z̄ ∈ K we have Ap̄, ¯ p̄¯ − Q(tf )(x̄¯f − t¯f ẋ(tf )), (x̄¯f − t¯f ẋ(tf )) , (6.55) ω0 = A1 p̄, where ¯ p̄¯ = ltf tf t¯f2 + 2ltf xf x̄¯f t¯f + lxf xf x̄¯f , x̄¯f A1 p̄, − 2ψ̇(tf )x̄¯f t¯f − (ψ̇0 (tf ) − ψ̇(tf )ẋ(tf ))t¯f2 . (6.56) The equalities t¯0 = 0 and x̄¯0 = 0 hold also for each element p̄¯ of the finite-dimensional and finite-faced cone K0 , given by (6.26) for α0 > 0 and by (6.22) in the general case. Rewriting the terms ω0 , we get the quadratic form in the variables (t¯f , x̃f ) generated by the matrix B11 B12 B := , ∗ B12 B22 where B11 = ltf tf + ψ̇(tf )ẋ(tf ) − ψ̇0 (tf ) − ẋ(tf )∗ Q(tf )ẋ(tf ), B12 = ltf xf − ψ̇(tf ) + ẋ(tf )∗ Q(tf ), B22 = lxf xf − Q(tf ). (6.57) The property x̄(t) = 0 in [t0 , t1 ) for z̄ ∈ K allows us to refine Theorems 6.16 and 6.17. Theorem 6.19. Assume that the initial values t0 = tˆ0 and x(t0 ) = x̂0 are fixed in the problem (6.1)–(6.3), and let s = 1. Let Q(t) be a continuous solution of the linear differential equation (6.43) on [t1 , tf] which satisfies twoconditions: (i) b1 := D 1 (H ) + ([ẋ]1 )∗ Q(t1 ) − [ψ̇]1 [ẋ]1 ≥ 0; (ii) the quadratic form ω0 is positive on the cone K0 \ {0}. Then is positive on K \ {0}. 6.3. Sufficient Conditions for Positive Definiteness of the Quadratic Form 271 Proof. Continue Q(t) arbitrarily as a solution of differential equation (6.43) to the whole interval [t0 , tf ] with possible jump at the point t1 . Note that the value b1 in condition (i) is the same as the value b1+ for the continued solution, and hence b1+ ≥ 0. Let z̄ ∈ K, and hence x̄ 1− = 0. Then by (6.50) with k = 1 the condition b1+ ≥ 0 implies the inequality ω1 ≥ 0. Condition (ii) implies the inequality ω0 ≥ 0. Consequently = ω0 + ω1 ≥ 0. Further arguments are the same as in the proof of Theorem 6.16. Theorem 6.20. Assume that the initial values t0 = tˆ0 and x(t0 ) = x̂0 are fixed in the problem (6.1)–(6.3) and s ≥ 2. Let Q(t) be a solution of the linear differential equation (6.43) on (t1 , tf ] \ which satisfies the following conditions: (a) The matrix Mk+ is positive semidefinite for each k = 2, . . . , s; (b) bk+ := D k (H ) + (qk+ )[ẋ]k > 0 for each k = 1, . . . , s − 1; (c) the quadratic form ω0 is positive on the cone K0 \ {0}. Then is positive on K \ {0}. Proof. Again, without loss of generality, we can consider Q(t) as a discontinuous solution of equation (6.43) on the whole interval [t0 , tf ]. Let z̄ ∈ K. Then by (6.50) with k = 1 the conditions b1+ > 0 and x̄ 1− = 0 imply the inequality ω1 ≥ 0. Further arguments are the same as in the proof of Theorem 6.17. 6.3.3 Q-Transformation of to Perfect Squares As in Section 5.3.3, we shall formulate special jump conditions for the matrix Q at each point tk ∈ . This will make it possible to transform into perfect squares and thus to prove its positive definiteness on K. Proposition 6.21. Suppose that bk+ := D k (H ) + (qk+ )[ẋ]k > 0 (6.58) and that Q satisfies the jump condition at tk , bk+ [Q]k = (qk+ )∗ (qk+ ), (6.59) where (qk+ )∗ is a column vector while qk+ is a row vector (defined as in (6.49)). Then ωk can be written as the perfect square ωk = = 2 (bk+ )−1 (bk+ )ξ̄k + (qk+ )(x̄ k− ) 2 (bk+ )−1 D k (H )ξ̄k + (qk+ )(x̄ k+ ) . (6.60) Proof. These formulas were proved in Section 4.1.6. Theorem 6.22. Let Q(t) satisfy the linear differential equation (6.43) on [t0 , tf ] \ , and let conditions (6.58) and (6.59) hold for each k = 1, . . . , s. Let ω0 be positive on K0 \ {0}. Then is positive on K \ {0}. 272 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Proof. By Proposition 6.21 and formulae (6.50), (6.51) the matrix Mk+ is positive semidefinite for each k = 1, . . . , n. Now using Theorem 6.16 for s = 1 and Theorem 6.17 for s ≥ 2, we obtain that is positive on K \ {0}. Similar assertions hold for the jump conditions that use left-hand values of Q at each point tk ∈ . Suppose that bk− := D k (H ) − (qk− )[ẋ]k > 0 (6.61) and that Q satisfies the jump condition at tk bk− [Q]k = (qk− )∗ (qk− ). (6.62) Then, according to Proposition 4.30, we have ωk = (bk− )−1 (bk− )ξ̄k + (qk− )(x̄ k+ ) 2 = (bk− )−1 D k (H )ξ̄k + (qk− )(x̄ k− ) (6.63) 2 . Using these formulas we deduce the following theorem. Theorem 6.23. Let Q(t) satisfy the linear differential equation (6.43) on [t0 , tf ] \ , and let conditions (6.61) and (6.62) hold for each k = 1, . . . , s. Let ω0 be positive on K0 \ {0}. Then is positive on K \ {0}. 6.4 Example: Minimal Fuel Consumption of a Car The following optimal control problem has been treated by Oberle and Pesch [83] as an exercise of applying the minimum principle. Consider a car whose dynamics (position x1 and velocity x2 ) are subject to friction and gravitational forces. The acceleration u(t) is proportional to the fuel consumption. Thus the control problem is to minimize the total fuel consumption tf J= u(t) dt (6.64) 0 in a time interval [0, tf ] subject to the dynamic constraints, boundary conditions, and the control constraints u c − αg − x22 , mx2 m x1 (0) = 0, x2 (0) = 1, x1 (tf ) = 10, x2 (tf ) = 3, umin ≤ u(t) ≤ umax , 0 ≤ t ≤ tf . ẋ1 = x2 , ẋ2 = (6.65) (6.66) (6.67) The final time tf is unspecified. The following values of the constants will be used in computations: m = 4, α = 1, g = 10, c = 0.4, umin = 100, umax = 140. 6.4. Example: Minimal Fuel Consumption of a Car 273 In view of the integral cost criterion (6.64), we consider the Pontryagin function (Hamiltonian) (cf. (6.17)) in normalized form taking α0 = 1, u c (6.68) H (x1 , x2 , ψ1 , ψ2 , u) = u + ψ1 x2 + ψ2 − αg − x22 . mx2 m The adjoint equations ψ̇ = −Hx are ψ̇1 = 0, ψ̇2 = −ψ1 + ψ2 2c u + x2 . mx22 m (6.69) The transversality condition (6.11) evaluated at the free final time tf yields the additional boundary condition u(tf ) 9c u(tf ) + 3ψ1 (tf ) + ψ2 (tf ) − αg − = 0. (6.70) 3m m The switching function φ(x, ψ) = Du H = 1 + determines the control law u(t) = ψ2 , mx2 umin umax if if φ(t) = φ(x(t), ψ(t)), φ(t) > 0 φ(t) < 0 . Computations give evidence to the fact that the optimal control is bang-bang with one switching time t1 , umin , 0 ≤ t < t1 . u(t) = umax , t1 ≤ t ≤ tf We compute an extremal using the code BNDSCO of Oberle and Grimm [82] or the code NUDOCCCS of Büskens [13]. The solution is displayed in Figure 6.1. Results for the switching time t1 , final time tf , and adjoint variables ψ(t) are 3.924284, t1 = ψ1 (0) = −42.24170, 9.086464, x1 (t1 ) = ψ1 (tf ) = −42.24170, tf = 4.254074, ψ2 (0) = −3.876396, x2 (t1 ) = 2.367329, ψ2 (tf ) = −17.31509. We will show that this trajectory satisfies the assumptions of Proposition 6.7. The critical cone is K = {0}, since the computed vectors ∂x (tf ) = (−0.6326710, −0.7666666)∗ , ∂t1 ẋ(tf ) = (3.0, 0.7666666)∗ are linearly independent. Moreover, we find in view of (6.14) that D 1 (H ) = −φ̇(t1 ) [u]1 = 0.472397 · 40 > 0. Theorem 6.10 then asserts that the computed bang-bang control provides a strict strong minimum. 274 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control (a) position x1 and velocity x2 (b) 10 control u 140 8 130 6 120 4 110 2 100 0 0 0.5 1 (c) 1.5 2 2.5 adjoint variable 3 3.5 4 0 4.5 0.5 1 -2 1.5 2 2.5 3 3.5 4 4.5 3 3.5 4 4.5 switching function (d) 2 0.1 -4 -6 0 -8 -0.1 -10 -0.2 -12 -0.3 -14 -0.4 -16 -0.5 -18 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 0 0.5 1 1.5 2 2.5 Figure 6.1. Minimal fuel consumption of a car. (a) State variables x1 , x2 . (b) Bangbang control u. (c) Adjoint variable ψ2 . (d) Switching function φ. 6.5 Quadratic Optimality Conditions in Time-Optimal Bang-Bang Control Problems In this section, we present the results of our paper [70]. 6.5.1 Statement of the Problem; Pontryagin and Strong Minimum We consider time-optimal control problems with control appearing linearly. Let x(t) ∈ Rd(x) denote the state variable and u(t) ∈ Rd(u) the control variable in the time interval t ∈ = [0, tf ] with a nonfixed final time tf > 0. For simplicity, the initial and terminal states are fixed in the following control problem: Minimize the final time tf (6.71) subject to the constraints on the interval = [0, tf ], dx/dt = ẋ = f (t, x, u) = a(t, x) + B(t, x)u, x(0) = a0 , x(tf ) = a1 , u(t) ∈ U , (t, x(t)) ∈ Q. (6.72) (6.73) (6.74) Here, a0 and a1 are given points in Rd(x) , Q ⊂ R1+d(x) is an open set, and U ⊂ Rd(u) is a convex polyhedron. The functions a, B are twice continuously differentiable on Q, with B being a d(x) × d(u) matrix function. A trajectory or control process T = { (x(t), u(t)) | t ∈ [0, tf ] } is said to be admissible if x(·) is absolutely continuous, u(·) is measurable and essentially bounded, and the pair of functions (x(t), u(t)) satisfies the constraints (6.72)–(6.74) 6.5. Quadratic Conditions in Time-Optimal Bang-Bang Control 275 on the interval = [0, tf ]. Let us define the Pontryagin and the strong minimum in the problem. An admissible trajectory T is said to be a Pontryagin minimum if there is no sequence of admissible trajectories T n = {(x n (t), un (t)) | t ∈ [0, tfn ]}, n = 1, 2, . . . , with (a) tfn < tf for n = 1, 2, . . . ; (b) tfn → tf for n → ∞; (c) maxn |x n (t) − x(t)| → 0 for n → ∞, where n = [0, tfn ]; n (d) n |u (t) − u(t)| dt → 0 for n → ∞. An admissible trajectory T is said to be a strong minimum (respectively, a strict strong minimum) if there is no sequence of admissible trajectories T n , n = 1, 2, . . . such that (a) tfn < tf (tfn ≤ tf , T n = T ) for n = 1, 2, . . .; (b) tfn → tf for n → ∞; (c) maxn |x n (t) − x(t)| → 0 for n → ∞, where n = [0, tfn ]. 6.5.2 Minimum Principle Let T = (x(t), u(t) | t ∈ [0, tf ]) be a fixed admissible trajectory such that the control u(·) is a piecewise constant function on the interval = [0, tf ] with finitely many points of discontinuity. Denote by = {t1 , . . . , ts }, 0 < t1 < · · · < ts < tf , the finite set of all discontinuity points (jump points) of the control u(t). Then ẋ(t) is a piecewise continuous function whose discontinuity points belong to the set and, thus, x(t) is a piecewise smooth function on . We use the notation [u]k = uk+ − uk− for the jump of the function u(t) at the point tk ∈ , where uk− = u(tk −), uk+ = u(tk +) are, respectively, the left- and right-hand values of the control u(t) at tk . Similarly, we denote by [ẋ]k the jump of the function ẋ(t) at the same point. Let us formulate the first-order necessary conditions of optimality for the trajectory T , the Pontryagin minimum principle. To this end we introduce the Pontryagin function or Hamiltonian function H (t, x, u, ψ) = ψf (t, x, u) = ψa(t, x) + ψB(t, x)u, (6.75) where ψ is a row vector of dimension d(x), while x, u, and f are column vectors. The factor of the control u in the Pontryagin function is the switching function φ(t, x, ψ) = ψB(t, x). Consider the pair of functions ψ0 (·) : → R1 , ψ(·) : → Rd(x) , which are continuous on and continuously differentiable on each interval of the set \ . Denote by M0 the set of normed pairs of functions (ψ0 (·), ψ(·)) satisfying the conditions ψ0 (tf ) ≥ 0, |ψ(0)| = 1, (6.76) ψ̇(t) = −Hx (t, x(t), u(t), ψ(t)) ∀ t ∈ \ , ψ̇0 (t) = −Ht (t, x(t), u(t), ψ(t)) ∀ t ∈ \ , min H (t, x(t), u, ψ(t)) = H (t, x(t), u(t), ψ(t)) ∀ t ∈ \ , (6.77) (6.78) (6.79) H (t, x(t), u(t), ψ(t)) + ψ0 (t) = 0 (6.80) u∈U ∀ t ∈ \ . Then the condition M0 = ∅ is equivalent to the Pontryagin minimum principle. This is the first-order necessary condition for a Pontryagin minimum. We assume that this condition is satisfied for the trajectory T . We say in this case that T is an extremal trajectory for the problem. The set M0 is a finite-dimensional compact set, since in (6.76) the initial 276 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control values ψ(0) are assumed to belong to the unit ball of Rd(x) . The case that there exists a multiplier (ψ0 , ψ) ∈ M0 with ψ0 (tf ) > 0 will be called the nondegenerate or normal case. Again we use the simple abbreviation (t) for all arguments (t, x(t), u(t), ψ(t)), e.g., φ(t) = φ(t, x(t), ψ(t)). Let us introduce the quantity D k (H ). For (ψ0 , ψ) ∈ M0 and tk ∈ consider the function (k H )(t) = H (t, x(t), uk+ , ψ(t)) − H (t, x(t), uk− , ψ(t)) = φ(t, x(t), ψ(t))[u]k . This function has a derivative at the point tk ∈ . We set d (k H )(tk ) = −φ̇(tk ±)[u]k . dt D k (H ) = − We know that for each (ψ0 , ψ) ∈ M0 D k (H ) ≥ 0 for k = 1, . . . , s. (6.81) We need the definition of a strict bang-bang control (see Section 6.1.4) to obtain the sufficient conditions in Theorem 6.27. For a given extremal trajectory T = { (x(t), u(t)) | t ∈ } with piecewise constant control u(t) we say that u(t) is a strict bang-bang control if there exists (ψ0 , ψ) ∈ M0 such that Arg min φ(t, x(t), ψ(t))u = [u(t−), u(t+)] ∀ t ∈ [t0 , tf ], u ∈U where (6.82) [u(t−), u(t+)] = {αu(t−) + (1 − α)u(t+) | 0 ≤ α ≤ 1 } denotes the line segment in Rd(u) . As it was mentioned already in Section 6.1.4, if U is the unit cube in Rd(u) , condition (6.82) precludes simultaneous switching of the control components. However, this property holds for all numerical examples in Chapter 8. In order to formulate quadratic optimality condition for a given extremal T with bangbang control u(·) we shall introduce the space Z(), the critical subspace K ⊂ Z(), and the quadratic form defined in Z(). 6.5.3 Critical Subspace As in Section 6.2.1, we denote by P C 1 (, Rn ) the space of piecewise continuous functions x̄(·) : → Rn , that are continuously differentiable on each interval of the set \ . For each x̄ ∈ P C 1 (, Rn ) and for tk ∈ , we use the abbreviation [x̄]k = x̄ k+ − x̄ k− , where x̄ k− = x̄(tk −), x̄ k+ = x̄(tk +). Putting z̄ = (t¯f , ξ̄ , x̄) with t¯f ∈ R1 , ξ̄ ∈ Rs , x̄ ∈ P C 1 (, Rn ), we have z̄ ∈ Z() := R1 × Rs × P C 1 (, Rn ). Denote by K the set of all z̄ ∈ Z() satisfying the following conditions: ˙ = fx (t, x(t), u(t))x̄(t), [x̄]k = [ẋ]k ξ̄k , x̄(t) x̄(0) = 0, x̄(tf ) + ẋ(tf )t¯f = 0. k = 1, . . . , s, (6.83) (6.84) 6.5. Quadratic Conditions in Time-Optimal Bang-Bang Control 277 Then K is a subspace of the space Z() which we call the critical subspace. Each element z̄ ∈ K is uniquely defined by the number t¯f and the vector ξ̄ . Consequently, the subspace K is finite-dimensional. An explicit representation of the variations x̄(t) in (6.83) is obtained as in Section 6.2.1. For each k = 1, . . . , s, define the vector functions y k (t) as the solutions to the system ẏ = fx (t)y, y(tk ) = [ẋ]k , t ∈ [tk , tf ]. (6.85) For t < tk we put y k (t) = 0 which yields the jump [y k ]k = [ẋ]k . Then x̄(t) = s y k (t)ξ̄k (6.86) y k (tf )ξ̄k + ẋ(tf )t¯f . (6.87) k=1 from which we obtain the representation x̄(tf ) + ẋ(tf )t¯f = s k=1 Furthermore, denote by x(t; t1 , . . . , ts ) the solution of the state equation (6.72) using the optimal bang-bang control with switching points t1 , . . . , ts . Then the partial derivatives of state trajectories with respect to the switching points are given by ∂x (t; t1 , . . . , ts ) = −y k (t) ∂tk for t ≥ tk , k = 1, . . . , s. (6.88) ∂x (t) = 0 and y k (t) = 0. This relation holds for all t ∈ [0, tf ]\{tk }, because for t < tk we have ∂t k Hence, (6.86) yields s ∂x x̄(t) = − (t)ξ̄k . (6.89) ∂tk k=1 In the nondegenerate case ψ0 (tf ) > 0, the critical subspace is simplified as follows. Proposition 6.24. If there exists (ψ0 , ψ) ∈ M0 such that ψ0 (tf ) > 0, then t¯f = 0 holds for each z̄ = (t¯f , ξ̄ , x̄) ∈ K. This proposition is a straightforward consequence from Proposition 6.8. In Section 6.5.4, we shall conclude from Theorem 6.27 that the property K = {0} essentially represents a first-order sufficient condition. Since x̄(tf )+ ẋ(tf )t¯f = 0 by (6.84), the representations (6.86) and (6.87) and Proposition 6.24 induce the following conditions for K = {0}. Proposition 6.25. Assume that one of the following conditions is satisfied: ∂x (tf ), k = 1, . . . , s, ẋ(tf ) are linearly independent; (a) The s+1 vectors y k (tf ) = ∂t k ∂x (tf ), (b) there exists (ψ0 , ψ) ∈ M0 with ψ0 (tf ) > 0 and the s vectors y k (tf ) = ∂t k k = 1, . . . , s, are linearly independent; (c) there exists (ψ0 , ψ) ∈ M0 with ψ0 (tf ) > 0, and the bang-bang control has exactly one switching point, i.e., s = 1. Then the critical subspace is K = {0}. 278 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Now we discuss the case of two switching points, i.e., s = 2, to prepare the numerical example in Section 6.5.4. Let us assume that ψ0 (tf ) > 0 (for some (ψ0 , ψ) ∈ M0 ) and [ẋ]1 = 0, [ẋ]2 = 0. By virtue of Proposition 6.24, we have t¯f = 0 and hence x̄(tf ) = 0 for each element z̄ ∈ K. Then the relations (6.84) and (6.86) yield 0 = x̄(tf ) = y 1 (tf )ξ̄1 + y 2 (tf )ξ̄2 . (6.90) The conditions [ẋ]1 = 0 and [ẋ]2 = 0 imply that y 1 (tf ) = 0 and y 2 (tf ) = 0, respectively. Furthermore, assume that K = {0}. Then (6.90) shows that the nonzero vectors y 1 (tf ) and y 2 (tf ) are collinear, i.e., y 2 (tf ) = αy 1 (tf ) (6.91) with some factor α = 0. As a consequence, the relation y 2 (t) = αy 1 (t) is valid for all t ∈ (t2 , tf ]. In particular, we have y 2 (t2 +) = αy 1 (t2 ) and thus [ẋ]2 = αy 1 (t2 ) (6.92) which is equivalent to (6.91). In addition, the equalities (6.90) and (6.91) imply that 1 ξ2 = − ξ1 . α (6.93) We shall use this formula in the next section. 6.5.4 Quadratic Form For (ψ0 , ψ) ∈ M0 and z̄ ∈ K, we define the functional (see formulas (6.34) and (6.35)) tf s k (ψ0 , ψ, z̄) = (D k (H )ξ̄k2 − 2[ψ̇]k x̄av Hxx (t)x̄(t), x̄(t) dt ξ̄k ) + (6.94) 0 k=1 2 ¯ −(ψ̇0 (tf ) − ψ̇(tf )ẋ(tf ))tf . k := 1 (x̄ k− + x̄ k+ ). Now we introduce second-order optimality conditions for where x̄av 2 bang-bang control in the problem (6.71)–(6.74). From Theorem 6.9 we easily deduce the following result. Theorem 6.26. Let a trajectory T affords a Pontryagin minimum. Then the following Condition A holds for the trajectory T : The set M0 is nonempty and max (ψ0 ,ψ)∈M0 (ψ0 , ψ, z̄) ≥ 0 ∀ z̄ ∈ K \ {0}. Similarly, from Theorem 6.10 we obtain the following theorem. Theorem 6.27. Let the following Condition B be fulfilled for the trajectory T : (a) there exists λ ∈ M0 such that D k (H ) > 0, k = 1, . . . , s, and condition (6.82) holds (i.e., u(t) is a strict bang-bang control), (b) max(ψ0 ,ψ)∈M0 (ψ0 , ψ, z̄) > 0 for all z̄ ∈ K \ {0}. Then T is a strict strong minimum. 6.5. Quadratic Conditions in Time-Optimal Bang-Bang Control 279 Remarks. 1. The sufficient Condition B is a natural strengthening of the necessary Condition A. 2. Condition (b) is automatically fulfilled if K = {0} holds (cf. Proposition 6.25) which gives a first-order sufficient condition for a strong minimum. 3. If there exists (ψ0 , ψ) ∈ M0 such that (ψ0 , ψ, z̄) > 0 for all z̄ ∈ K \ {0}, then condition (b) is obviously fulfilled. ≤ ui ≤ umax 4. For boxes U = {u = (u1 , . . . , ud(u) ) ∈ Rd(u) : umin i i , i = 1, . . . , d(u)}, the k condition D (H ) > 0, k = 1, . . . , s, is equivalent to the property φ̇i (tk ) = 0 if tk is a switching point of the ith control component ui (t). Note again that condition (6.82) precludes the simultaneous switching of two or more control components. 5. A further remark concerns the case that the set M0 of Pontryagin multipliers is not a singleton. This case was illustrated in [89] by the following time-optimal control problem for a linear system: ẋ1 = x2 , ẋ2 = x3 , ẋ3 = x4 , ẋ4 = u, |u| ≤ 1, x(0) = a, x(tf ) = b, where x = (x1 , x2 , x3 , x4 ). It was shown in [89] that for some a and b there exists an extremal in this problem with two switching points of the control such that, under appropriate normalization, the set M0 is a segment. For this extremal, the maximum of the quadratic forms is positive on each nonzero element of the critical subspace, and hence the sufficient conditions of Theorem 6.27 are satisfied. But this is not true for any single quadratic form (corresponding to an element of the set M0 ). 6.5.5 Nondegenerate Case Let us assume the nondegenerate or normal case that there exists (ψ0 , ψ) ∈ M0 such that the cost function multiplier ψ0 (tf ) is positive. By virtue of Proposition 6.24 we have in this case that t¯f = 0 for all z̄ ∈ K. Thus the critical subspace K is defined by the conditions ˙ = fx (t)x̄(t), x̄(t) [x̄]k = [ẋ]k ξ̄k (k = 1, . . . , s), x̄(0) = 0, x̄(tf ) = 0. (6.95) In particular, these conditions imply x̄(t) ≡ 0 on [0, t1 ) and (ts , tf ]. Hence we have x̄ 1− = x̄ s+ = 0 for all z̄ ∈ K. Then the quadratic form (6.94) is equal to (ψ, z̄) = s k=1 k D k (H )ξk2 + 2[Hx ]k x̄av ξk + tf Hxx (t)x̄(t), x̄(t) dt. (6.96) 0 This case of a time-optimal (autonomous) control problem was studied by Sarychev [104]. He used a special transformation of the problem and obtained sufficient optimality condition for the transformed problem. It is not easy, but it is possible, to reformulate his results in terms of the original problem. The comparison of both types of conditions reveals that Sarychev used the same critical subspace, but his quadratic form is a lower bound for . Namely, in his quadratic form the positive term D k (H )ξ̄k2 has the factor 14 instead of the factor 1 for the same term in . Therefore, the sufficient Condition B is always fulfilled whenever Sarychev’s condition is fulfilled. However, there is an example of a control problem where the optimal solution satisfies Condition B but does not satisfy Sarychev’s 280 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control condition. Finally, Sarychev proved that his condition is sufficient for an L1 -minimum with respect to the control (which is a Pontryagin minimum in this problem). In fact, it could be proved that his condition is sufficient for a strong minimum. 6.5.6 Cases of One or Two Switching Times of the Control From Theorem 6.27 and Proposition 6.25(c), we immediately deduce sufficient conditions for a bang-bang control with one switching point. The result is used for the example in Section 6.7.1 and is also applicable to the time-optimal control of an image converter discussed by Kim et al. [47]. Theorem 6.28. Let the following conditions be fulfilled for the trajectory T : (a) u(t) is a bang-bang control with one switching point, i.e., s = 1, (b) there exists (ψ0 , ψ) ∈ M0 such that D 1 (H ) > 0 and condition (6.82) holds (i.e., u(t) is a strict bang-bang control), (c) there exists (ψ0 , ψ) ∈ M0 with ψ0 (tf ) > 0. Then T is a strict strong minimum. Now we turn our attention to the case of two switching points where s = 2. Assume the nondegenerate case ψ0 (tf ) > 0 and suppose that [ẋ]1 = 0, [ẋ]1 = 0 and y 2 (tf ) = αy 1 (tf ) as in (6.91). Otherwise, K = {0} holds and, hence, the first-order sufficient condition for a strong minimum is satisfied. For any element z̄ ∈ K, we have t¯f = 0, x̄ 1− = 0, x̄ 2+ = 0. Consequently, 1 1 1 = [x̄]1 = [ẋ]1 ξ̄1 , x̄av 2 2 1 1 1 2 [ẋ]2 ξ̄1 x̄av = x̄ 2− = y 1 (t2 )ξ̄1 = 2 2 2α in view of x̄(t) = y 1 (t)ξ̄1 + y 2 (t)ξ̄2 , y 2 (t2 −) = 0 and (6.92). Using these relations in the quadratic form (6.96) together with (6.93) and the conditions y 2 (t) = 0 for all t < t2 , [Hx ]k = −[ψ̇]k , k = 1, 2, we compute the quadratic form for the element of the critical subspace as t2 1 2 Hxx x̄, x̄ dt ξ̄1 − 2[ψ̇]2 x̄av ξ̄2 + = D 1 (H )ξ̄12 + D 2 (H )ξ̄22 − 2[ψ̇]1 x̄av t1 t2 1 2 1 1 2 2 1 1 2 2 2 2 1 1 = D (H )ξ̄1 + 2 D (H )ξ̄1 − [ψ̇] [ẋ] ξ̄1 + 2 [ψ̇] [ẋ] ξ̄1 + Hxx y , y dt ξ̄12 α α t1 = ρ ξ̄12 , where ρ = D (H ) − [ψ̇] [ẋ] 1 1 1 1 + 2 D 2 (H ) + [ψ̇]2 [ẋ]2 + α t2 Hxx y 1 , y 1 dt. (6.97) t1 Thus, we obtain the following proposition. Proposition 6.29. Assume that ψ0 (tf ) > 0, s = 2, [ẋ]1 = 0, [ẋ]2 = 0, and y 2 (tf ) = αy 1 (tf ) (which is equivalent to (6.91)) with some factor α. Then the condition of the positive definiteness of on K is equivalent to the inequality ρ > 0, where ρ is defined as in (6.97). 6.6. Sufficient Conditions for Time-Optimal Control Problems 6.6 281 Sufficient Conditions for Positive Definiteness of the Quadratic Form on the Critical Subspace K for Time-Optimal Control Problems In this section, we consider the nondegenerate case as in Section 6.5.5 and assume (i) u(t) is a bang-bang control with s > 1 switching points, (ii) there exists (ψ0 , ψ) ∈ M0 such that ψ0 (tf ) > 0 and D k (H ) > 0, k = 1, . . . , s. Under these assumptions the critical subspace K is defined as in (6.95). Let (ψ0 , ψ) ∈ M0 be a fixed element (possibly different from that in assumption (ii)), and denote by = (ψ0 , ψ, ·) the quadratic form for this element. Recall that is given by (6.96). According to Theorem 6.27 the positive definiteness of the quadratic form (6.96) on the subspace K in (6.95) is a sufficient condition for a strict strong minimum of the trajectory. Now our aim is to find conditions that guarantee the positive definiteness of on K. 6.6.1 Q-Transformation of on K Here we shall use the same arguments as in Sections 5.3.2 and 6.3.1. Let Q(t) be a symmetric matrix on [t1 , ts ] with piecewise continuous entries which are absolutely continuous on each interval of the set [t1 , ts ] \ . Therefore, Q may have a jump at each point tk ∈ including t1 , ts , and thus the symmetric matrices Q1− and Qs+ are also defined. For z̄ ∈ K, we obviously have t s + ts s d [Qx̄, x̄ ]k , Qx̄, x̄ dt = Qx̄, x̄ − t1 dt t1 − k=1 where [Qx̄, x̄ ]k is the jump of the function Qx̄, x̄ at the point tk ∈ . Using the conditions x̄˙ = fx (t)x̄ and x̄ 1− = x̄ s+ = 0, we obtain ts s [Qx̄, x̄ ]k + (Q̇ + fx∗ Q + Qfx )x̄, x̄ dt = 0, (6.98) t1 k=1 where the asterisk denotes transposition. Adding this zero form to , we get = s k D k (H )ξ̄k2 − 2[ψ̇]k x̄av ξ̄k + [Qx̄, x̄ ]k k=1 + ts t1 (6.99) (Hxx + Q̇ + fx∗ Q + Qfx )x̄, x̄ dt. We call this formula the Q-transformation of on K. To eliminate the integral term in , we assume that Q(t) satisfies the following linear matrix differential equation: Q̇ + fx∗ Q + Qfx + Hxx = 0 on [t1 , ts ] \ . (6.100) Using (6.100), the quadratic form (6.99) reduces to = s k=1 ωk , k ωk := D k (H )ξk2 − 2[ψ̇]k x̄av ξk + [Qx̄, x̄ ]k . (6.101) 282 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Thus, we have proved the following statement. Proposition 6.30. Let Q(t) satisfy the linear differential equation (6.100) on [t1 , ts ] \ . Then for each z̄ ∈ K the representation (6.101) holds. Now our goal is to derive conditions such that ωk > 0 holds on K \ {0} for k = 1, . . . , s. We shall use the representations of ωk given in Section 6.3.1. According to (6.50), ωk = D k (H ) + (qk+ )[ẋ]k ξ̄k2 + 2(qk+ )x̄ k− ξ̄k + (x̄ k− )∗ [Q]k x̄ k− , (6.102) where qk+ = ([ẋ]k )∗ Qk+ − [ψ̇]k . We immediately see from this representation that one way to enforce ωk > 0 is to impose the following conditions: D k (H ) > 0, qk+ = ([ẋ]k )∗ Qk+ − [ψ̇]k = 0, [Q]k ≥ 0. (6.103) In practice, however, it might be difficult to check this condition since it is necessary to satisfy the d(x) equality constraints qk+ = ([ẋ]k )∗ Qk+ − [ψ̇]k = 0 and the inequality constraints [Q]k ≥ 0. It is more convenient to express ωk as a quadratic form in the variables (ξ̄k , x̄ k− ) with the matrix D k (H ) + (qk+ )[ẋ]k qk+ , (6.104) Mk+ = (qk+ )∗ [Q]k where qk+ is a row vector and (qk+ )∗ is a column vector. Similarly, according to (6.52), the following representation holds: ωk = D k (H ) − (qk− )[ẋ]k ξ̄k2 + 2(qk− )x̄ k+ ξ̄k + (x̄ k+ )∗ [Q]k x̄ k+ , (6.105) where qk− = ([ẋ]k )∗ Qk− − [ψ̇]k . Again, we see that ωk > 0 holds if we require the conditions (6.106) D k (H ) > 0, qk− = ([ẋ]k )∗ Qk− − [ψ̇]k = 0, [Q]k ≥ 0. To find a more general condition for ωk > 0, we consider (6.105) as a quadratic form in the variables (ξ̄k , x̄ k+ ) with the matrix D k (H ) − (qk− )[ẋ]k qk− . (6.107) Mk− = (qk− )∗ [Q]k Since the right-hand sides of equalities (6.102) and (6.105) are connected by the relation x̄ k+ = x̄ k− + [ẋ]k ξ̄k , the following statement obviously holds. Proposition 6.31. For each k = 1, . . . , s, the positive (semi)definiteness of the matrix Mk− is equivalent to the positive (semi)definiteness of the matrix Mk+ . Now we can prove the following theorem. Theorem 6.32. Let Q(t) be a solution of the linear differential equation (6.100) on [t1 , ts ]\ which satisfies the following conditions: 6.6. Sufficient Conditions for Time-Optimal Control Problems 283 (a) the matrix Mk+ is positive semidefinite for each k = 2, . . . , s; (b) bk+ := D k (H ) + (qk+ )[ẋ]k > 0 for each k = 1, . . . , s − 1. Then is positive on K \ {0}. Proof. Take an arbitrary element z̄ = (ξ̄ , x̄) ∈ K. Let us show that ≥ 0 for this element. Condition (a) implies that ωk ≥ 0 for k = 2, . . . , s. Condition (b) for k = 1 together with condition x̄ 1− = 0 implies that ω1 ≥ 0. Consequently, ≥ 0. Assume that = 0. Then ωk = 0, k = 1, . . . , s. The conditions ω1 = 0, x̄ 1− = 0, and b1+ > 0 by formula (6.102) (with k = 1) yield ξ̄1 = 0. Then [x̄]1 = 0 and hence x̄ 1+ = 0. The last equality and the equation x̄˙ = fx (t)x̄ show that x̄(t) = 0 in (t1 , t2 ) and hence x̄ 2− = 0. Similarly, the conditions ω2 = 0, x̄ 2− = 0, and b2+ > 0 by formula (6.102) (with k = 2) imply that ξ̄2 = 0 and x̄(t) = 0 in (t2 , t3 ). Therefore, x̄ 3− = 0 etc. Continuing this process we get x̄ ≡ 0 and ξ̄k = 0 for k = 1, . . . , s − 1. Now using formula (6.101) for ωs = 0, as well as the conditions D s (H ) > 0 and x̄ ≡ 0 we obtain ξ̄s = 0. Consequently, we have z̄ = 0 which means that is positive on K \ {0}. Similarly, using representation (6.105) for ωk we can prove the following statement. Theorem 6.33. Let Q(t) be a solution of the linear differential equation (6.100) on [t1 , ts ]\ which satisfies the following conditions: (a) The matrix Mk− is positive semidefinite for each k = 1, . . . , s − 1; (b) bk− := D k (H ) − (qk− )[ẋ]k > 0 for each k = 2, . . . , s. Then is positive on K \ {0}. 6.6.2 Q-Transformation of to Perfect Squares Here, as in Section 6.3.3, we formulate special jump conditions for the matrix Q at each point tk ∈ , which will make it possible to transform into perfect squares and thus to prove its positive definiteness on K. Suppose that bk− := D k (H ) − (qk− )[ẋ]k > 0 (6.108) and that Q satisfies the jump condition at tk , bk− [Q]k = (qk− )∗ (qk− ), (6.109) where (qk− )∗ is a column vector while qk− is a row vector. Then according to (6.63), ωk = (bk− )−1 (bk− )ξ̄k + (qk− )(x̄ k+ ) 2 = (bk− )−1 D k (H )ξ̄k + (qk− )(x̄ k− ) 2 . (6.110) Theorem 6.34. Let Q(t) satisfy the linear differential equation (6.100) on [t1 , ts ] \ . Let condition (6.108) hold for each k = 1, . . . , s and condition (6.109) hold for each k = 1, . . . , s − 1. Then is positive on K \ {0}. Proof. According to (6.110), the matrix Mk− is positive semidefinite for each k = 1, . . . , s − 1 (cf. (6.105) and (6.107)), and hence both conditions (a) and (b) of Theorem 6.33 are fulfilled. Then by Theorem 6.33, is positive on K \ {0}. 284 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Similar assertions hold for the jump conditions that use right-hand values of Q at each point tk ∈ . Suppose that bk+ := D k (H ) + (qk+ )[ẋ]k > 0 (6.111) and that Q satisfies the jump condition at point tk bk+ [Q]k = (qk+ )∗ (qk+ ). (6.112) Then ωk = (bk+ )−1 (bk+ )ξ̄k + (qk+ )(x̄ k− ) 2 = (bk+ )−1 D k (H )ξ̄k + (qk+ )(x̄ k+ ) 2 . (6.113) Theorem 6.35. Let Q(t) satisfy the linear differential equation (6.100) on [t1 , ts ] \ . Let condition (6.111) hold for each k = 1, . . . , s and condition (6.112) hold for each k = 2, . . . , s. Then is positive on K \ {0}. 6.6.3 Case of Two Switching Times of the Control Let s = 2, i.e., = {t1 , t2 }, and let Q(t) be a symmetric matrix with absolutely continuous entries on [t1 , t2 ]. Put Qk = Q(tk ), qk = ([ẋ]k )∗ Qk − [ψ̇]k , k = 1, 2. Theorem 6.36. Let Q(t) satisfy the linear differential equation (6.100) on (t1 , t2 ) and the following inequalities hold at t1 , t2 : D 1 (H ) + q1 [ẋ]1 > 0, D 2 (H ) − q2 [ẋ]2 > 0. (6.114) Then is positive on K \ {0}. Proof. In the case considered we have Q1+ = Q1 , and q1+ = q1 , b1+ := D 1 (H ) + q1 [ẋ]1 > 0, Define the jumps [Q]1 and [Q]2 Q2− = Q2 , q2− = q2 , b2− := D 2 (H ) − q2 [ẋ]2 > 0. (6.115) by the conditions b1+ [Q]1 = (q1+ )∗ (q1+ ), b2− [Q]2 = (q2− )∗ (q2− ). (6.116) Then [Q]1 and [Q]2 are symmetric matrices. Put Q1− = Q1+ − [Q]1 , Q2+ = Q2− + [Q]2 . Then Q1− and Q2+ are also symmetric matrices. Thus, we obtain a symmetric matrix Q(t) satisfying (6.100) on (t1 , t2 ), the inequalities (6.115), and the jump conditions (6.116). By formulas (6.110) and (6.113) the terms ω1 and ω2 are nonnegative. In view of (6.101), 6.6. Sufficient Conditions for Time-Optimal Control Problems 285 we see that = ω1 + ω2 is nonnegative on K. Suppose that = 0 for some z̄ = (ξ , x̄) ∈ K. Then ωk = 0 for k = 1, 2, and thus formulas (6.110) and (6.113) give b1+ ξ1 + (q1+ )x̄ 1− = 0, b2− ξ2 + (q2− )x̄ 2+ = 0. But x̄ 1− = 0 and x̄ 2+ = 0. Consequently, ξ̄1 = ξ̄2 = 0, and then conditions x̄ 1− = 0 and [x̄]1 = 0 imply that x̄ 1+ = 0. The last equality and the equation x̄˙ = fx (t)x̄ imply that x̄(t) = 0 on (t1 , t2 ). Thus x̄ ≡ 0 and then z̄ = 0. We have proved that is positive on K \ {0}. Control System with a Constant Matrix B 6.6.4 In the case that B(t, x) = B is a constant matrix, the adjoint equation has the form ψ̇ = −ψax , which implies that [ψ̇]k = 0, k = 1, . . . , s. Therefore, qk− = ([ẋ]k )∗ Qk− , qk+ = ([ẋ]k )∗ Qk+ , (qk− )∗ qk− = Qk− [ẋ]k ([ẋ]k )∗ Qk− , (qk+ )∗ qk+ = Qk+ [ẋ]k ([ẋ]k )∗ Qk+ , bk− = D k (H ) − ([ẋ]k )∗ Qk− [ẋ]k , bk+ = D k (H ) + ([ẋ]k )∗ Qk+ [ẋ]k , where D k (H ) = ψ̇(tk )B[u]k , k = 1, . . . , s. In the case of two switching points with s = 2, the conditions (6.114) take the form D 1 (H ) + Q1 [ẋ]1 , [ẋ]1 ) > 0, D 2 (H ) − Q2 [ẋ]2 , [ẋ]2 ) > 0. Now assume, in addition, that u is one-dimensional and that ⎛ ⎞ 0 ⎜ .. ⎟ ⎜ ⎟ U = [−c, c], c > 0. B = βen := ⎜ . ⎟ , β > 0, ⎝ 0 ⎠ β In this case we get and thus [ẋ]k = B[u]k = βen [u]k , k = 1, . . . , s, Qk [ẋ]k , [ẋ]k ) = β 2 Qk en , en |[u]k |2 = 4β 2 c2 Qnn (tk ), where Qnn is the element of matrix ⎛ Q11 ⎜ .. Q=⎝ . Qn1 ⎞ . . . Q1n ⎟ .. .. ⎠. . . . . . Qnn (6.117) 286 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Moreover, in the last case we obviously have D k (H ) = 2βc |ψ̇n (tk )|, k = 1, . . . , s. For s = 2, conditions (6.114) are thus equivalent to the estimates Qnn (t1 ) > − 6.7 |ψ̇n (t1 )| , 2βc Qnn (t2 ) < |ψ̇n (t2 )| . 2βc (6.118) Numerical Examples of Time-Optimal Control Problems 6.7.1 Time-Optimal Control of a Van der Pol Oscillator Consider again the tunnel-diode oscillator displayed in Figure 4.1. In the control problem of the van der Pol oscillator, the state variable x1 represents the voltage, whereas the control u is the voltage at the generator. Time-optimal solutions will be computed in two cases. First, we consider a fixed terminal state x(tf ) = xf . The second case treats the nonlinear terminal constraint x1 (tf )2 + x2 (tf )2 = r 2 for a small r > 0, by which the oscillator is steered only to a neighborhood of the origin. In the first case we consider the control problem of minimizing the final time tf subject to the constraints ẋ1 (t) = x2 (t), ẋ2 (t) = −x1 (t) + x2 (t)(1 − x12 (t)) + u(t), (6.119) x1 (0) = −0.4, x2 (0) = 0.6, (6.120) | u(t) | ≤ 1 x1 (tf ) = 0.6, x2 (tf ) = 0.4, for t ∈ [0, tf ]. (6.121) The Pontryagin function (Hamiltonian) is given by H (x, u, ψ) = ψ1 x2 + ψ2 (−x1 + x2 (1 − x12 ) + u). (6.122) The adjoint equations ψ̇ = −Hx are ψ̇1 = ψ2 (1 + 2x1 x2 ), ψ̇2 = −ψ1 + ψ2 (x12 − 1). (6.123) In view of the free final time we get the additional boundary condition H (tf ) + ψ0 (tf ) = 0.4ψ1 (tf ) + ψ2 (tf )(−0.344 + u(tf )) + 1 = 0. (6.124) The sign of switching function φ(t) = ψ2 (t) determines the optimal control according to 1 if ψ2 (t) < 0 . (6.125) u(t) = −1 if ψ2 (t) > 0 Evaluating the derivatives of the switching function, it can easily be seen that there are no singular arcs with ψ2 (τ ) ≡ 0 holding on a time interval [t1 , t2 ] . Nonlinear programming methods applied to the discretized control problem show that the optimal bang-bang control has two bang-bang arcs, 1 for 0 ≤ t < t1 , (6.126) u(t) = −1 for t1 ≤ t ≤ tf 6.7. Numerical Examples of Time-Optimal Control Problems (a) state variables x1 and x2 time-optimal control u and switching function (b) 1 287 1 0.8 0.5 0.6 0.4 0 0.2 0 -0.5 -0.2 -1 -0.4 0 0.2 0.4 (c) 0.6 0.8 1 0.8 0.7 0.6 0.5 0 0.2 0.4 0.6 adjoint variables 0.8 1 and 1 1.2 2 0.8 0.6 0.4 0.2 0 -0.2 -0.4 -0.6 -0.8 -1 -1.2 0.9 -0.2 0.2 (d) phaseportrait (x1, x2) 1 0.4 -0.4 0 1.2 0.4 0.6 0 0.2 0.4 0.6 0.8 1 1.2 Figure 6.2. Time-optimal solution of the van der Pol oscillator, fixed terminal state (6.120). (a) State variables x1 and x2 (dashed line). (b) Control u and switching function ψ2 (dashed line). (c) Phase portrait (x1 , x2 ). (d) Adjoint variables ψ1 and ψ2 (dashed line). with switching time t1 . This implies the switching condition φ(t1 ) = ψ2 (t1 ) = 0. (6.127) Hence, we must solve the boundary value problem (6.120)–(6.127). Using the code BNDSCO [82] or NUDOCCCS [13, 14], we obtain the extremal solution displayed in Figure 6.2. The optimal final time, the switching point, and some values for the adjoint variables are 1.2540747, tf = ψ1 (0) = −1.0816056, ψ1 (t1 ) = −1.0886321, ψ1 (tf ) = −0.47781383, t1 = 0.158320138, ψ2 (0) = −0.18436798, ψ2 (t1 ) = 0.0, ψ2 (tf ) = 0.60184112. (6.128) Since the bang-bang control has only one switching time, we are in the position to apply Theorem 6.27. For checking the assumptions of this theorem it remains to verify the condition D 1 (H ) = |φ̇(t1 )[u]1 | > 0. Indeed, in view of the adjoint equation (6.123) and the switching condition ψ2 (t1 ) = 0 we get D 1 (H ) = |φ̇(t1 )[u]1 | = 2|ψ1 (t1 )| = 2 · 1.08863205 = 0. Then Theorem 6.27 ensures that the computed solution is a strict strong minimum. Now we treat the second case, where the two boundary conditions (6.120) are replaced by the single nonlinear condition x(tf )2 + x2 (tf )2 = r 2 , r = 0.2. (6.129) 288 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control Imposing this boundary condition, we aim at steering the van der Pol oscillator to a small neighborhood of the origin. The adjoint equation (6.124) remains valid. The transversality condition for the adjoint variable gives ψ1 (tf ) = 2βx1 (tf ), ψ2 (tf ) = 2βx2 (tf ), β ∈ R. (6.130) The boundary condition (6.11) associated with the free final time tf yields ψ1 (tf )x2 (tf ) + ψ2 (tf ) (−x1 (tf ) + x2 (tf )(1 − x1 (tf )2 ) + u(tf )) + 1 = 0. (6.131) Again, the switching function (6.18) is given by φ(t) = Hu (t) = ψ2 (t). The structure of the bang-bang control agrees with that in (6.132), 1 for 0 ≤ t < t1 , (6.132) u(t) = −1 for t1 ≤ t ≤ tf which yields the switching condition φ(t1 ) = ψ2 (t1 ) = 0. (6.133) Using either the boundary value solver BNDSCO of Oberle and Grimm [82] or the direct optimization routine NUDOCCCS of Büskens [13, 14], we obtain the extremal solution depicted in Figure 6.3 and the following values for the switching, final time, state, and adjoint variables: t1 = 0.7139356, ψ1 (0) = 0.9890682, x1 (t1 ) = 1.143759, ψ1 (t1 ) = 1.758128, x1 (tf ) = 0.06985245, ψ1 (tf ) = 0.4581826, β = 3.279646. tf = ψ2 (0) = x2 (t1 ) = ψ2 (t1 ) = x2 (tf ) = ψ2 (tf ) = 2.864192, 0.9945782, −0.5687884, 0.0, −0.1874050, −1.229244, (6.134) There are two alternative ways to check sufficient conditions. We may either use Theorem 6.19 and solve the linear equation (6.43) or evaluate directly the quadratic form in Proposition 6.11. We begin by testing the assumptions of Theorem 6.19 and consider the symmetric 2 × 2 matrix Q11 (t) Q12 (t) . Q(t) = Q12 (t) Q22 (t) The linear equations Q̇ = −Qfx − fx∗ Q − Hxx in (6.100) yield the following ODEs: Q̇11 = 2 Q12 (1 + 2x1 x2 ) + 2ψ2 x2 , Q̇12 = −Q11 − Q12 (1 − x12 ) + Q22 (1 + 2x1 x2 ) + 2ψ2 x1 , (6.135) Q̇22 = −2 (Q12 + Q22 (1 − x12 )). In view of Theorem 6.19 we must find a solution Q(t) only in the interval [t1 , tf ] such that D 1 (H ) + (qk+ )[ẋ]1 > 0, qk+ = ([ẋ]1 )∗ Q(t1 ) − [ψ̇]1 6.7. Numerical Examples of Time-Optimal Control Problems (a) state variables x1 and x2 control u and switching function (b) 1.4 1.2 1 0.8 0.6 0.4 0.2 0 -0.2 -0.4 -0.6 289 1 0.5 0 -0.5 -1 -1.5 -2 0 0.5 (c) 1 1.5 2 2.5 3 0 0.5 (d) phase portrait (x1, x2) 1 1.5 adjoint variables 2 1 and 2.5 3 2.5 3 2 2 1 1.5 0.8 1 0.6 0.5 0.4 0.2 0 0 -0.5 -0.2 -1 -0.4 -1.5 -0.6 -2 0 0.2 0.4 0.6 0.8 1 1.2 1.4 0 0.5 1 1.5 2 Figure 6.3. Time-optimal solution of the van der Pol oscillator, nonlinear boundary condition (6.129). (a) State variables x1 and x2 (dashed line). (b) Control u and switching function ψ2 (dashed line). (c) Phase portrait (x1 , x2 ). (d) Adjoint variables ψ1 and ψ2 (dashed line). holds and the quadratic form ω0 in (6.55)–(6.57) is positive definite on the cone K0 defined in (6.26). Since ψ2 (t1 ) = 0 we get from (6.130), D 1 (H ) = −φ̇(t1 )[u]1 = 2 · ψ1 (t1 ) = 2 · 1.758128 > 0. Furthermore, from [ψ̇]1 = 0 we obtain the condition D 1 (H ) + ([ẋ]1 )∗ Q(t1 )[ẋ]1 = 2 · 1.758128 + 4Q22 (t1 ) > 0, which is satisfied by any initial value Q22 (t1 ) > −0.879064. By Proposition 6.8, we have t¯f = 0 for every element z̄ = (t¯f , ξ , x̄) ∈ K. Therefore, by (6.57) we must check that the matrix B22 = 2β I2 − Q(tf ) is positive definite on the critical cone K0 defined in (6.26), i.e., on the cone K0 = {(t¯f , v1 , v2 ) | t¯f = 0, x1 (tf )v1 + x2 (tf )v2 = 0}. Thus the variations (v1 , v2 ) are related by v2 = −v1 x1 (tf )/x2 (tf ). Evaluating the quadratic form (2β I2 − Q(tf ))(v1 , v2 ), (v1 , v2 ) with v2 = −v1 x1 (tf )/x2 (tf ), we arrive at the test x1 c = 2β 1 + x2 2 2 x1 x1 − Q11 − 2 Q12 + (tf ) > 0. Q22 x2 x2 290 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control A straightforward integration of the ODEs (6.135) using the solution data (6.134) and the initial values Q11 (t1 ) = Q12 (t1 ) = Q22 (t1 ) = 0 gives the numerical results Q11 (tf ) = 0.241897, Q12 (tf ) = −0.706142, Q22 (tf ) = 1.163448, which yield the positive value c = 7.593456 > 0. Thus Theorem 6.19 asserts that the bangbang control characterized by (6.134) provides a strict strong minimum. The alternative test for second-order sufficient conditions (SSC) is based on Proposition 6.11. The variational system ẏ(t) = fx (t)y(t), y(t1 ) = [ẋ]1 , for the variation y = (y1 , y2 ) leads to the variational system y1 (t1 ) = 0, ẏ1 = y2 , ẏ2 = −(1 + 2x1 2x2 )y1 + (1 − x12 )y2 , y2 (t1 ) = 2, for which we compute y1 (tf ) = 4.929925, y2 (tf ) = 1.837486. Note that the relation Kxf (x(tf ))y(tf ) = 2(x1 (tf )y1 (tf ) + x2 (tf )y2 (tf ) = 0 holds. By Proposition 6.11 we have to show that the quantity ρ in (6.37) is positive, ρ = D 1 (H ) − [ψ̇]1 [ẋ]1 + tf t1 (y(t))∗ Hxx (t)y(t) dt + (y(tf ))∗ (βK)xf xf y(tf ) > 0. Using [ψ̇]1 = 0 and (y(tf ))∗ (βK)xf xf y(tf ) = 2β(y1 (tf )2 + y2 (tf )2 ), we finally obtain ρ = D 1 (H ) + 184.550 > 0. 6.7.2 Time-Optimal Control of the Rayleigh Equation In Section 4.1, the Rayleigh problem of controlling oscillations in a tunnel-diode oscillator (Figure 4.1) was considered with a regulator functional. In this section, we treat the timeoptimal case of steering a given initial state to the origin in minimal time. Recall that the state variable x1 (t) = I (t) denotes the electric current. The optimal control problem is to minimize the final time tf subject to the dynamics and control constraints ẋ1 (t) = x2 (t), ẋ2 (t) = −x1 (t) + x2 (t)(1.4 − 0.14x2 (t)2 ) + u(t), (6.136) x1 (0) = x2 (0) = −5, (6.137) | u(t) | ≤ 4 x1 (tf ) = x2 (tf ) = 0, for t ∈ [0, tf ]. (6.138) Note that we have shifted the factor 4 to the control variable in the dynamics (4.134) to the control constraint (6.138). The Pontryagin function (Hamiltonian) (see (6.75)) for this problem is (6.139) H (x, u, ψ) = ψ1 x2 + ψ2 (−x1 + x2 (1.4 − 0.14x22 ) + u). 6.7. Numerical Examples of Time-Optimal Control Problems 291 The transversality condition (6.11) yields, in view of (6.139), H (tf ) + 1 = ψ2 (tf ) u(tf ) + 1 = 0. The switching function φ(x, ψ) = ψ2 determines the optimal control 4 if ψ2 (t) < 0 . u(t) = −4 if ψ2 (t) > 0 (6.140) (6.141) As for the van der Pol oscillator, it is easy to show that there are no singular arcs with ψ2 (t) ≡ 0 holding on a time interval Is ⊂ [0, tf ]. Hence, the optimal control is bang-bang. Applying nonlinear programming methods to the suitably discretized Rayleigh problem, one realizes that the optimal control comprises the following three bang-bang arcs with two switching times t1 , t2 : ⎫ ⎧ ⎪ ⎬ ⎨ 4 for 0 ≤ t < t1 ⎪ −4 for t1 ≤ t < t2 . (6.142) u(t) = ⎪ ⎭ ⎩ 4 for t ≤ t ≤ t ⎪ 2 f This control structure yields two switching conditions ψ2 (t1 ) = 0, ψ2 (t2 ) = 0. (6.143) Thus, we have to solve the multipoint boundary value problem comprising equations (6.136)–(6.143). The codes BNDSCO [82] and NUDOCCS [13, 14] yield the extremal depicted in Figure 6.4. The final time, the switching points, and some values for the adjoint variables are computed as 3.66817339, tf = t1 = 1.12050658, ψ1 (0) = −0.12234128, ψ1 (t1 ) = −0.21521225, 0.84276186, ψ1 (tf ) = t2 = 3.31004698, ψ2 (0) = −0.08265161, ψ1 (t2 ) = 0.89199176, ψ2 (tf ) = −0.25. (6.144) Now we are going to show that the computed control satisfies the assumptions of Theorem 6.19 and thus provides a strict strong minimum. Consider the symmetric 2 × 2 matrix Q11 (t) Q12 (t) Q(t) = , Q12 (t) Q22 (t) The linear equation (6.100), dQ/dt = −Q fx − fx∗ Q − Hxx , leads to the following three equations: Q̇11 = 2Q12 , Q̇12 = −Q11 − Q12 (1.4 − 0.42x22 ) + Q22 , (6.145) Q̇22 = −Q12 − Q22 (1.4 − 0.42x22 ) + 0.84ψ2 x2 . We must find a solution of these equations satisfying the estimates (6.114) at the switching times t1 and t2 . From D k (H ) = | φ̇(tk )[u]k | = 8 |ψ1 (tk )|, k = 1, 2, 292 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control (a) state variables x1 and x2 (b) 6 time-optimal control u and switching function (×4) 4 4 3 2 2 1 0 0 -2 -1 -4 -2 -6 -3 -4 -8 0 0.5 1 (c) 1.5 2 2.5 3 3.5 4 0 0.5 (d) phaseportrait (x1 , x2) 5 1 4 0.8 1 1.5 2 adjoint variables 2.5 1 3 and 3.5 4 3.5 4 2 3 2 0.6 1 0.4 0 0.2 -1 -2 0 -3 -0.2 -4 -0.4 -5 -7 -6 -5 -4 -3 -2 -1 0 1 0 0.5 1 1.5 2 2.5 3 Figure 6.4. Time-optimal control of the Rayleigh equation. (a) State variables x1 and x2 (dashed line). (b) Control u and switching function φ (dashed line). (c) Phase portrait (x1 , x2 ). (d) Adjoint variables ψ1 and ψ2 (dashed line). we get D 1 (H ) = 8 · 0.21521225 = 1.7269800 > 0, D 2 (H ) = 8 · 0.89199176 = 7.1359341 > 0. Furthermore, we have [ẋ]k = [u]k (0, 1)∗ = (0, 8)∗ and thus obtain Qk [ẋ]k , [ẋ]k = 64 Q22 (tk ), for k = 1, 2. The next step is to find a solution for the equations (6.145) in the interval [t1 , t2 ] that satisfies the inequalities D 1 (H ) + Q1 [ẋ]1 , [ẋ]1 = 1.7269800 + 64 Q22 (t1 ) > 0, D 2 (H ) − Q2 [ẋ]2 , [ẋ]2 = 7.1359341 − 64 Q22 (t2 ) > 0 This requires the estimates Q22 (t1 ) > −0.0269841, Q22 (t2 ) < 0.11149897. (6.146) These conditions can be satisfied by choosing, e.g., the following initial values at the switching time t1 : Q11 (t1 ) = 0, Q12 (t1 ) = 0.25, Q22 (t1 ) = 0. Integration yields the value Q22 (t2 ) = −0.1677185 which shows that the estimates (6.146) hold. Note that these estimates do not hold for the choice Q(t1 ) = 0, since this initial value would give Q22 (t2 ) = 0.70592. In summary, Theorem 6.36 asserts that the computed solution provides a strict strong minimum. 6.8. Time-Optimal Control Problems for Linear Systems with Constant Entries 293 6.8 Time-Optimal Control Problems for Linear Systems with Constant Entries In this section, we shall observe some results which were obtained in [77], [79, Part 2, Section 13], and [87]. 6.8.1 Statement of the Problem, Minimum Principle, and Simple Sufficient Optimality Condition Consider the following problem: tf → min, ẋ = Ax + Bu, x(0) = a, x(tf ) = b, u ∈ U, (6.147) where A and B are constant matrices, a and b are fixed vectors in Rd(x) , and U is a convex polyhedron in Rd(u) . A triple (tf , x, u) is said to be admissible if x(t) is a Lipschitz continuous and u(t) is measurable bounded function on the interval = [0, tf ] and the pair (x, u) satisfies on the constraints of the problem (6.147). Definition 6.37. We say that the admissible triple (tf , x, u) affords an almost global minimum if there is no sequence of admissible triples (tf n , x n , un ), n = 1, 2, . . . , such that tfn < tf for n = 1, 2, . . . , and tfn → tf . Proposition 6.38. Suppose that, in the problem (6.147), there exists an u∗ ∈ U such that Aa + Bu∗ = 0 or Ab + Bu∗ = 0 (for example, let b = 0, 0 ∈ U ). Then the almost global minimum is equivalent to the global one. Proof. If, for example, Ab + Bu∗ = 0, u∗ ∈ U , then any pair (x, u) admissible on [0, tf ] can be extended to the right of tf by putting x(t) = b, u(t) = u∗ . Let (tf , x, u) be an admissible triple for which the conditions of the minimum principle are fulfilled: There exists a smooth function ψ : [0, tf ] → Rd(x) such that −ψ̇ = ψA, u(t) ∈ Arg min u ∈U (ψ(t)Bu ), (6.148) ψ ẋ = const ≤ 0, |ψ(0)| = 1. These conditions follow from (6.76)–(6.80), because here H = ψ(Ax + Bu), Ht = 0, hence −ψ ẋ = ψ0 = const =: α0 ≥ 0. Thus M0 can be identified with the set of infinitely differentiable functions ψ(t) on [0, tf ] satisfying conditions (6.148). The condition that M0 is nonempty is necessary for a Pontryagin minimum to hold for the triple (tf , x, u). We will refer to a triple (tf , x, u) with nonempty M0 as an extremal triple. Recall a simple sufficient first-order condition for an almost global minimum obtained in [79, Part 1]. Theorem 6.39 (Milyutin). Suppose there exists ψ ∈ M0 such that α0 := −ψ ẋ > 0. Then (tf , x, u) affords an almost global minimum. 294 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control In what follows, we assume that (tf , x, u) is an extremal triple such that u(t) is a piecewise constant function taking values in the vertices of the polyhedron U . We denote by = {t1 , . . . , ts } the set of discontinuity points of the control u(t). Let ψ ∈ M0 . An important role in formulations of optimality conditions in problem (6.147) is played by the product (ψ̇ ẋ)(t). Proposition 6.40. The product (ψ̇ ẋ)(t) is a monotone nonincreasing step function with discontinuities only at the discontinuity points of the control u(t). We now formulate yet another simple sufficient condition for the almost global minimum, obtained in [79, Part 2]. Theorem 6.41. Suppose that there exists ψ ∈ M0 such that the product (ψ̇ ẋ) fulfills one of the following two conditions: (ψ̇ ẋ)(0) < 0, (ψ̇ ẋ)(tf ) > 0; i.e., (ψ̇ ẋ) strictly retains its sign on [0, tf ]. Then (tf , x, u) affords an almost global minimum. Theorem 6.41 implies a sufficient condition of a geometric nature. Corollary 6.42. Let (tf , x, u) be an extremal triple such that for any point tk ∈ the vectors ẋ k− = ẋ(tk −), ẋ k+ = ẋ(tk +) are different from zero and equally directed (so that ẋ k− = ck ẋ k+ for some ck > 0). Suppose that there exists ψ ∈ M0 such that (ψ̇ ẋ) is not identically zero. Then (tf , x, u) affords an almost global minimum. 6.8.2 Quadratic Optimality Condition For an extremal triple (tf , x, u) in the problem (6.147) satisfying the assumptions of Section 6.8.1 we will write the quadratic necessary Condition A and the quadratic sufficient Condition B using the results in Section 6.5.4. Necessary Condition A. This time we begin by writing the quadratic form as in (6.94). Let us show that it is completely determined by the left and right limits of the step function ψ̇ ẋ at points tk ∈ . Since H = ψ(Ax + Bu) and ψ̇ = −ψA, we have [ψ̇]k = 0, k = 1, . . . , s and Hxx = 0. Moreover, D k (H ) = −ψ̇(tk )B[u]k = −ψ̇(tk )[ẋ]k = −[ψ̇ ẋ]k , k = 1, . . . , s. It follows from the condition Ht = 0 that ψ̇0 = 0. Further, the condition that (ψ̇ ẋ) is constant on (ts , tf ) implies (ψ̇ ẋ)(tf ) = (ψ̇ ẋ)(ts +). Thus according to (6.94), (ψ, z̄) = − s [(ψ̇ ẋ)]k ξ̄k2 + t¯f2 (ψ̇ ẋ)s+ . (6.149) k=1 This formula holds for any ψ ∈ M0 and any z̄ = (t¯f , ξ̄ , x̄) which belongs to the subspace K as in (6.83) and (6.84). 6.8. Time-Optimal Control Problems for Linear Systems with Constant Entries 295 Let us see what is the form of K in the present case. Since x̄(t) satisfies the linear system x̄˙ = Ax̄ on each interval \ , condition x̄(0) = 0 in (6.84) can be replaced by x̄(t1 −) = 0. Since ẋ satisfies the same system, ẍ = Aẋ, condition x̄(tf ) + ẋ(tf )t¯f = 0 in (6.84) can be replaced by x̄(ts +) + ẋ(ts +)t¯f = 0. For brevity, put x̄(t1 −) = x̄ 1− , x̄(ts +) = x̄ s+ , ẋ(ts +) = ẋ s+ . Then by (6.83) and (6.84) the subspace K consists of triples z̄ = (t¯f , ξ̄ , x̄) satisfying the conditions t¯f ∈ R1 , ξ̄ ∈ Rs , x̄(·) ∈ P C ∞ (, Rn ), x̄˙ = Ax̄, [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, x̄ 1− = 0, x̄ s+ + t¯f ẋ s+ = 0, where P C ∞ (, Rn ) is the space of piecewise continuous functions x̄(t) : → Rd(x) infinitely differentiable on each interval of the set \ . This property of x̄ follows from the fact that on each interval of the set \ the function x̄ satisfies the linear system x̄˙ = Ax̄ with constant entries. Consider the cross-section of K specified by condition t¯f = −1. The passage to the cross-section does not weaken the quadratic necessary Condition A because the functional maxψ∈M0 (ψ, z̄) involved in it is homogeneous of degree 2 and nonnegative on any element z̄ ∈ K with t¯f = 0 (since for any ψ ∈ M0 the inequalities D k (H ) = −[ψ̇ ẋ]k ≥ 0, k = 1, . . . , s, hold). Denote by R the cross-section of the subspace K, specified by condition t¯f = −1. We omit the coordinate t¯f in the definition of R. Thus R is a set of pairs (ξ̄ , x̄) such that the following conditions are fulfilled: ξ̄ ∈ Rs , x̄(·) ∈ P C ∞ (, Rn ), x̄˙ = Ax̄, [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, x̄ 1− = 0, x̄ s+ = ẋ s+ . For ψ ∈ M0 , ξ̄ ∈ Rs , let Q(ψ, ξ̄ ) = − s [ψ̇ ẋ]k ξ̄k2 + (ψ̇ ẋ)s+ , k=1 and set Q0 (ξ̄ ) = max Q(ψ, ξ̄ ). ψ∈M0 (6.150) Then Theorem 6.26 implies the following theorem. Theorem 6.43. Let a triple (tf , x, u) afford a Pontryagin minimum in the problem (6.147). Then the set M0 as in (6.148) is nonempty and Q0 (ξ̄ ) ≥ 0 ∀ (ξ̄ , x̄) ∈ R. (6.151) It is clear that the set R, in this necessary condition, can be replaced by its projection under the mapping (ξ̄ , x̄) → ξ̄ . Denote this projection by % and find out what conditions specify it. Conditions x̄˙ = Ax̄ and ẍ = Aẋ imply x̄(t) = eAt c̄(t), ẋ(t) = eAt c(t), 296 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control where c̄(t) and c(t) are step functions whose discontinuity points are contained in . Conditions x̄ 1− = 0, x̄ s+ = ẋ s+ , [x̄]k = [ẋ]k ξ̄k , k = 1, . . . , s, imply c̄1− = 0, c̄s+ = cs+ , Therefore cs+ = c̄s+ = [c̄]k = [c]k ξ̄k , s [c̄]k = k=1 s k = 1, . . . , s. [c]k ξ̄k . k=1 It is easily seen that the conditions s [c]k ξ̄k = cs+ k=1 determine the projection of R under the mapping (ξ̄ , x̄) → ξ̄ , which we denote by %. Since [c]k = e−Atk [ẋ]k ∀ k, cs+ = e−Ats ẋ s+ , the set % is determined by the condition s e−Atk [ẋ]k ξ̄k = e−Ats ẋ s+ , k=1 which after multiplication by eAts from the left takes the final form s eA(ts −tk ) [ẋ]k ξ̄k = ẋ s+ . (6.152) k=1 Hence % is the set of vectors ξ̄ ∈ Rs satisfying the system of algebraic equations (6.152). Thus Theorem 6.43 implies the following theorem. Theorem 6.44. Let a triple (tf , x, u) afford a Pontryagin minimum in the problem (6.147). Then the set M0 as in (6.148) is nonempty and Q0 (ξ̄ ) ≥ 0 ∀ ξ̄ ∈ %. (6.153) It makes sense to use necessary condition (6.153) for investigation of only those extremals (tf , x, u) for which α0 := −ψ ẋ = 0 ∀ ψ ∈ M0 , (6.154) because otherwise, by Theorem 6.39, (tf , x, u) affords an almost global minimum in the problem. Condition (6.154) guarantees that the set % is nonempty (see Theorem 13.7 in [79, p. 322]). Theorem 6.44 implies a simple consequence of a geometric nature. 6.8. Time-Optimal Control Problems for Linear Systems with Constant Entries 297 Corollary 6.45. Suppose that (tf , x, u) affords a Pontryagin minimum in the problem (6.147). Let the vectors ẋ k− and ẋ k+ be different from zero and collinear for some tk ∈ , and let the jump [ψ̇ ẋ]k of the product ψ̇ ẋ at the point tk be different from zero for any ψ ∈ M0 . Then the vectors ẋ k− and ẋ k+ are equally directed. The proof is given in [79]. Sufficient Condition B. Let (tf , x, u) be an admissible triple in the problem (6.147) satisfying the assumptions of Section 6.8.1. As in Section 6.2.4, denote by M the set of functions ψ ∈ M0 satisfying the two conditions −[ψ̇ ẋ]k > 0, k = 1, . . . , s, Arg min ψ(t)Bu = [u(t−), u(t+)] ∀ t ∈ [0, tf ]. u ∈U (6.155) (6.156) Obviously, the interval [u(t−), u(t+)] is a singleton for all t ∈ [0, tf ] \ . Theorem 6.46. For an admissible triple (tf , x, u) satisfying the assumptions of Section 6.8.1, let the set M be nonempty and Q0 (ξ̄ ) > 0 ∀ ξ̄ ∈ %. (6.157) Then (tf , x, u) affords a strict almost global minimum in the problem (6.147). Proof. Condition (6.157) implies that max (λ, z̄) > 0 (6.158) ψ∈M0 for all z̄ ∈ K such that t¯f = 0. Consider an element z̄ ∈ K \ {0} such that t¯f = 0. For this element, ξ̄ = 0, since otherwise z̄ = 0. Take an arbitrary element ψ ∈ M and put q = −ψ̇ ẋ. Then [q]k > 0, k = 1, . . . , s. Hence (λ, z̄) = s [q]k ξ̄k2 > 0. k=1 Thus the inequality (6.158) holds for all z̄ ∈ K \ {0}. Therefore by Theorem 6.27, (tf , x, u) affords a strict strong minimum in the problem (6.147). Moreover, condition M = ∅ in the problem (6.147) implies that the strict strong minimum is equivalent to the strict almost global minimum. The last assertion is a consequence of Proposition 13.4 and Lemma 13.1 in [79]. 6.8.3 Example Consider the problem tf → min, ẋ1 = x2 , ẋ2 = u, |u| ≤ 1, x(0) = a, x(tf ) = b, (6.159) 298 Chapter 6. Second-Order Optimality Conditions for Bang-Bang Control where x = (x1 , x2 ) ∈ R2 , u ∈ R1 . The minimum principle conditions for this problem are as follows: ψ̇1 = 0, ψ̇2 = −ψ1 , u = − sgn ψ2 , (6.160) ψ1 x2 + ψ2 u = −α0 ≤ 0, |ψ(0)| = 1, where ψ = (ψ1 , ψ2 ). This implies that ψ2 (t) is a linear function, and u(t) is a step function having at most one switching and taking values ±1. The system corresponding to u = 1 is ẋ1 = x2 , ẋ2 = 1, and omitting t we have 1 x1 = x22 + C. 2 (6.161) The system corresponding to u = −1 is ẋ1 = x2 , ẋ2 = −1, whence 1 x1 = − x22 + C. 2 (6.162) Through each point x = (x1 , x2 ) of the state plane there pass one curve of the family (6.161) and one curve of the family (6.162). The condition ψ1 x2 + ψ2 u = −α0 ≤ 0 implies that the passage from a curve of the family (6.161) to a curve of the family (6.162) is only possible in the upper half-plane x2 ≥ 0, and the passage from a curve of the family (6.162) to a curve of the family (6.161) in the lower half-plane x2 ≤ 0. The direction of movement is such that along the curves of the family (6.161) the point can go to (+∞, +∞), and along the curves of the family (6.162) to (−∞, −∞). This means that any two points can be joined by the extremals in no more than two ways. By Theorem 6.39 each extremal with α0 := −ψ̇ ẋ = −ψ1 x2 − ψ2 u > 0 yields an almost global minimum. Consider extremals with switching and α0 = 0. These are extremals with x2 = 0 at the switching time and ẋ 1− = (0, u1− ), ẋ 1+ = (0, u1+ ). Since u1− = −u1+ = 0, the vectors ẋ 1− and ẋ 1+ are different from zero and directed in an opposite way. The set M0 consists of a single pair ψ = (ψ1 , ψ2 ), and ψ̇ ẋ = −ψ1 u, [ψ̇ ẋ]1 = −ψ1 (t1 )[u]1 = 0. (6.163) According to Corollary 6.45 this extremal does not yield a Pontryagin minimum because it is necessary for Pontryagin minimum that the collinear vectors ẋ 1− and ẋ 1+ are equally directed. In this case the necessary second-order condition fails. Thus in this problem an extremal with a switching affords an almost global minimum iff α0 > 0 for this extremal; i.e., the point x = (x1 , x2 ) at the switching time does not lie on the x1 axis in the state plane. For an extremal without switchings there always exists ψ ∈ M0 such that α0 > 0. Indeed, for this extremal one can set ψ1 = 0, ψ2 = −u. Then α0 = −ψ2 u = 1, and all conditions of the minimum principle are fulfilled. More interesting examples of investigation of extremals in time-optimal control problems for linear systems with constant entries are given in [79, Part 2, Section 14]. The timeoptimal control in a simplified model of a container crane (ore unloader) was discussed in [63, Section 5]. The optimality of the time-optimal control with three switching times follows from Theorem 6.39. Chapter 7 Bang-Bang Control Problem and Its Induced Optimization Problem We continue our investigation of the pure bang-bang case. As it was mentioned in introduction, second-order sufficient optimality conditions for bang-bang controls had been derived in the literature in two different forms. The first form was discussed in the last chapter. The second one is due to Agrachev, Stefani, and Zezza [1], who first reduced the bang-bang control problem to a finite-dimensional optimization problem and then showed that wellknown sufficient optimality conditions for this optimization problem supplemented by the strict bang-bang property furnish sufficient conditions for the bang-bang control problem. The bang-bang control problem, considered in this chapter, is more general than that in [1]. Following [99, 100], we establish the equivalence of both forms of second-order conditions for this problem. 7.1 Main Results 7.1.1 Induced Optimization Problem Again, let Tˆ = (x̂(t), û(t) | t ∈ [tˆ0 , tˆf ]) be an admissible trajectory for the basic problem (6.1)–(6.3). We denote by V = ex U the set of vertices of the polyhedron U . Assume that ˆ = [tˆ0 , tˆf ] taking values in the set V , û(t) is a bang-bang control in û(t) = uk ∈ V for t ∈ (tˆk−1 , tˆk ), k = 1, . . . , s + 1, ˆ = {tˆ1 , . . . , tˆs } is the set of switching points of the control û(·) with where tˆs+1 = tˆf . Thus, tˆk < tˆk+1 for k = 0, 1, . . . , s. Assume now that the set M0 of multipliers is nonempty for the trajectory Tˆ . Put x̂(tˆ0 ) = x̂0 , θ̂ = (tˆ1 , . . . , tˆs ), ζ̂ = (tˆ0 , tˆf , x̂0 , θ̂ ). Then θ̂ ∈ Rs , ζ̂ ∈ R2 × Rn × Rs , where n = d(x). Take a small neighborhood V of the point ζ̂ in R2 × Rn × Rs , and let ζ = (t0 , tf , x0 , θ ) ∈ V, 299 (7.1) 300 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem where θ = (t1 , . . . , ts ) satisfies t0 < t1 < t2 < · · · < ts < tf . Define the function u(t; θ ) by the condition u(t; θ ) = uk for t ∈ (tk−1 , tk ), k = 1, . . . , s + 1, (7.2) where ts+1 = tf . The values u(tk ; θ), k = 1, . . . , s, may be chosen in U arbitrarily. For definiteness, define them by the condition of continuity of the control from the left: u(tk ; θ) = u(tk −; θ ), k = 1, . . . , s. Let x(t; t0 , x0 , θ) be the solution of the initial value problem (IVP), ẋ = f (t, x, u(t; θ)), t ∈ [t0 , tf ], x(t0 ) = x0 . (7.3) For each ζ ∈ V this solution exists if the neighborhood V of the point ζ̂ is sufficiently small. We obviously have x(t; tˆ0 , x̂0 , θ̂) = x̂(t), ˆ t ∈ , u(t; θ̂) = û(t), ˆ \ . ˆ t ∈ Consider now the following finite-dimensional optimization problem in the space R2 × Rn × Rs of the variables ζ = (t0 , tf , x0 , θ ): F0 (ζ ) := J (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) → min, F (ζ ) := F (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) ≤ 0, G(ζ ) := K(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = 0. (7.4) We call (7.4) the Induced Optimization Problem (IOP) or simply Induced Problem, which represents an extension of the IOP introduced in Agrachev, Stefani, and Zezza [1]. The following assertion is almost obvious. Theorem 7.1. Let the trajectory Tˆ be a Pontryagin local minimum for the basic control problem (6.1)–(6.3). Then the point ζ̂ is a local minimum for the IOP (7.4), and hence it satisfies first- and second-order necessary conditions for this problem. Proof. Assume that ζ̂ is not a local minimum in problem (7.4). Then there exists a sequence of admissible points ζ ν = (t0ν , t1ν , x0ν , θ ν ) in problem (7.4) such that ζ ν → ζ̂ for ν → ∞ and F0 (ζ ν ) < F0 (ζ̂ ). Take the corresponding sequence of admissible trajectories T ν = {x(t; t0ν , x0ν , θ ν ), u(t; θ ν ) | t ∈ [t0ν , tfν ]} in problem (6.1)–(6.3). Then the conditions t0ν → tˆ0 , tfν → tˆf , x0ν → x̂0 , θ ν → θ̂ imply that ˆ ν ∩ |u(t; θ ν ) − û(t)| dt → 0, max |x(t; t0ν , x0ν , θ ν ) − x̂(t)| → 0, ˆ ν ∩ where ν = [t0ν , tfν ]. Moreover, J(T ν ) = F0 (ζ ν ) < F0 (ζ̂ ) = J(Tˆ ). It means that the trajectory Tˆ is not a Pontryagin local minimum for the basic problem (6.1)–(6.3). We shall clarify a relationship between the second-order conditions for the IOP (7.4) at the point ζ̂ and those in the basic bang-bang control problem (6.1)–(6.3) for the trajectory Tˆ . 7.1. Main Results 301 We shall state that there is a one-to-one correspondence between Lagrange multipliers in these problems and a one-to-one correspondence between elements of the critical cones. Moreover, for corresponding Lagrange multipliers, the quadratic forms in these problems take equal values on the corresponding elements of the critical cones. This will allow us to express the necessary and sufficient quadratic optimality conditions for bang-bang control, formulated in Theorems 6.9 and 6.10, in terms of the IOP (7.4). In particular, we thus establish the equivalence between our quadratic sufficient conditions and those due to Agrachev, Stefani, and Zezza [1]. First, for convenience we recall second-order necessary and sufficient conditions for a smooth finite-dimensional optimization problem with inequality- and equality-type constraints (see Section 1.3.1). Consider a problem in Rn , f0 (x) → min, fi (x) ≤ 0 (i = 1, . . . , k), gj (x) = 0 (j = 1, . . . , m), (7.5) where f0 , . . . , fk , g1 , . . . , gm are C 2 -functions in Rn . Let x̂ be an admissible point in this problem. Define, at this point, the set of normalized vectors μ = (α0 , . . . , αk , β1 , . . . , βm ) of Lagrange multipliers μ ∈ Rk+1+m | αi ≥ 0 (i = 0, . . . , k), αi fi (x̂) = 0 (i = 1, . . . , k), 0 = k m αi + |βj | = 1; Lx (μ, x̂) = 0 , i=0 m k j =1 where L(μ, x) = i=0 αi fi (x) + j =1 βj gj (x) is the Lagrange function. Define the set of active indices I = {i ∈ {1, . . . , k} | fi (x̂) = 0} and the critical cone K0 = {x̄ | f0 (x̂)x̄ ≤ 0, fi (x̂)x̄ ≤ 0, i ∈ I , gj (x̂)x̄ = 0, j = 1, . . . , m}. Theorem 7.2. Let x̂ be a local minimum in problem (7.5). Then, at this point, the set 0 is nonempty and the following inequality holds: max Lxx (μ, x̂)x̄, x̄ ≥ 0 ∀ x̄ ∈ K0 . μ∈0 Theorem 7.3. Let the set 0 be nonempty at the point x̂ and let max Lxx (μ, x̂)x̄, x̄ > 0 ∀ x̄ ∈ K0 \ {0}. μ∈0 Then x̂ is a local minimum in problem (7.5). These conditions were obtained by Levitin, Milyutin, and Osmolovskii [54, 55]; cf. also Ben-Tal and Zowe [4]. 7.1.2 Relationship between Second-Order Conditions for the Basic and Induced Optimization Problem Let Tˆ = (x̂(t), û(t) | t ∈ [tˆ0 , tˆf ]) be an admissible trajectory in the basic problem with the properties assumed in Section 7.1.1, and let ζ̂ = (tˆ0 , tˆf , x̂0 , θ̂ ) be the corresponding admissible point in the Induced Problem. 302 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Lagrange multipliers. Let us define the set 0 ⊂ R1+d(F )+d(K) of the triples μ = (α0 , α, β) of normalized Lagrange multipliers at the point ζ̂ for the Induced Problem. The Lagrange function for the Induced Problem is L(μ, ζ ) = L(μ, t0 , tf , x0 , θ ) = α0 J (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) + αF (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) + βK(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = l(μ, t0 , x0 , tf , x(tf ; t0 , x0 , θ )), (7.6) where l = α0 J +αF +βK. By definition, 0 is the set of multipliers μ = (α0 , α, β) such that α0 ≥ 0, α ≥ 0, α0 + |α| + |β| = 1, αF (p̂) = 0, Lζ (μ, ζ̂ ) = 0, (7.7) where p̂ = (tˆ0 , x̂0 , tˆf , x̂f ), x̂0 = x̂(tˆ0 ), x̂f = x̂(tˆf ) = x(tˆf ; tˆ0 , x̂0 , θ̂ ). Now, let us define the corresponding set of normalized Lagrange multipliers for the trajectory Tˆ in the basic problem. Denote by the set of multipliers λ = (α0 , α, β, ψ, ψ0 ) such that α0 ≥ 0, α ≥ 0, α0 + |α| + |β| = 1, αF (p̂) = 0, −ψ̇(t) = ψ(t)fx (t, x̂(t), û(t)), −ψ̇0 (t) = ψ(t)ft (t, x̂(t), û(t)), ψ(tˆ0 ) = −lx0 (μ, p̂), ψ(tˆf ) = lxf (μ, p̂), ψ0 (tˆ0 ) = −lt0 (μ, p̂), ψ0 (tˆf ) = ltf (μ, p̂), ˆ \ , ˆ ψ(t)f (t, x̂(t), û(t)) + ψ0 (t) = 0 ∀ t ∈ (7.8) ˆ = [tˆ0 , tˆf ] and ˆ = {tˆ1 , . . . , tˆs }. where Proposition 7.4. The projector π0 : (α0 , α, β, ψ, ψ0 ) → (α0 , α, β) (7.9) maps one-to-one the set onto the set 0 . Let us define the inverse mapping. Take an arbitrary multiplier μ = (α0 , α, β) ∈ 0 . This tuple defines the gradient lxf (μ, p̂), and hence the system −ψ̇ = ψfx (t, x̂(t), û(t)), ψ(tˆf ) = lxf (μ, p̂) (7.10) defines ψ(t). Define ψ0 (t) by the equality ψ(t)f (t, x̂(t), û(t)) + ψ0 (t) = 0. (7.11) Proposition 7.5. The inverse mapping π0−1 : (α0 , α, β) ∈ 0 → (α0 , α, β, ψ, ψ0 ) ∈ (7.12) is defined by formulas (7.10) and (7.11). We note that M0 ⊂ holds, because the system of conditions (6.7)–(6.9) and (6.11) is equivalent to system (7.8). But it may happen that M0 = , since in the definition of there is no requirement that its elements satisfy minimum condition (6.10). Let us denote MP 0 := π0 (M0 ), where MP stands for Minimum Principle. 7.1. Main Results 303 We say that multipliers μ = (α0 , α, β) and λ = (α0 , α, β, ψ, ψ0 ) correspond to each other if they have the same components α0 , α and β, i.e., π0 (α0 , α, β, ψ, ψ0 ) = (α0 , α, β). Critical cones. We denote by K0 the critical cone at the point ζ̂ in the Induced Problem. Thus, K0 is the set of collections ζ̄ = (t¯0 , t¯f , x̄0 , θ̄ ) such that F0 (ζ̂ )ζ̄ ≤ 0, Fi (ζ̂ )ζ̄ ≤ 0, i ∈ I , G (ζ̂ )ζ̄ = 0, (7.13) where I is the set of indices of the inequality constraints active at the point ζ̂ . Let K be the critical cone for the trajectory Tˆ in the basic problem, i.e., the set of all tuples ˆ satisfying conditions (6.21)–(6.23). z̄ = (t¯0 , t¯f , ξ̄ , x̄) ∈ Z(), Proposition 7.6. The operator π1 : (t¯0 , t¯f , ξ̄ , x̄) → (t¯0 , t¯f , x̄0 , θ̄ ) defined by θ̄ = −ξ̄ , x̄0 = x̄(tˆ0 ) (7.14) is a one-to-one mapping of the critical cone K (for the trajectory Tˆ in the basic problem) onto the critical cone K0 (at the point ζ̂ in the Induced Problem). We say that the elements ζ̄ = (t¯0 , t¯f , x̄0 , θ̄ ) ∈ K0 and z̄ = (t¯0 , t¯f , ξ̄ , x̄) ∈ K correspond to each other if θ̄ = −ξ̄ and x̄0 = x̄(tˆ0 ), i.e., π1 (t¯0 , t¯f , ξ̄ , x̄) = (t¯0 , t¯f , x̄0 , θ̄ ). Now we give explicit formulas for the inverse mapping for π1 . Let V (t) be an n × n ˆ = [tˆ0 , tˆf ] and satisfies matrix-valued function (n = d(x)) which is absolutely continuous in the system V̇ (t) = fx (t, x̂(t), û(t))V (t), V (tˆ0 ) = E, (7.15) where E is the identity matrix. For each k = 1, . . . , s denote by y k (t) the n-dimensional vector function which is equal to zero in [tˆ0 , tˆk ), and in [tˆk , tˆf ] it is the solution to the IVP ẏ k = fx (t, x̂(t), û(t))y k , ˙ k. y k (tˆk ) = −[x̂] (7.16) ˙ k at tˆk . Hence y k is a piecewise continuous function with one jump [y k ]k = −[x̂] Proposition 7.7. The inverse mapping π1−1 : (t¯0 , t¯f , x̄0 , θ̄ ) ∈ K0 → (t¯0 , t¯f , ξ̄ , x̄) ∈ K is given by the formulas ξ̄ = −θ̄, ˙ tˆ0 )t¯0 + x̄(t) = V (t) x̄0 − x̂( s y k (t)t¯k , k=1 where t¯k is the kth component of the vector θ̄. Quadratic forms. For μ ∈ 0 the quadratic form of the IOP is equal to Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ . The main result of this section is the following. (7.17) 304 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Theorem 7.8. Let the Lagrange multipliers μ = (α0 , α, β) ∈ MP and λ = (α0 , α, β, ψ, ψ0 ) ∈ M0 0 correspond to each other, i.e., π0 λ = μ, and let the elements of the critical cones ζ̄ = (t¯0 , t¯f , x̄0 , θ̄) ∈ K0 and z̄ = (t¯0 , t¯f , ξ̄ , x̄) ∈ K correspond to each other, i.e., π1 z̄ = ζ̄ . Then the quadratic forms in the basic and induced problems take equal values: Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ = (λ, z̄). Consequently, max Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ = max (λ, z̄) μ∈MP 0 λ∈M0 for each pair of elements of the critical cones ζ̄ ∈ K0 and z̄ ∈ K such that π1 z̄ = ζ̄ . Theorems 6.9 and 7.8 and Proposition 7.6 imply the following second-order necessary optimality condition for the basic problem. Theorem 7.9. If the trajectory Tˆ affords a Pontryagin minimum in the basic problem, then the following Condition A0 holds: The set M0 is nonempty and max Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ ≥ 0 μ∈MP 0 ∀ ζ̄ ∈ K0 . Theorems 6.10 and 7.8 and Proposition 7.6 imply the following second-order sufficient optimality condition for the basic control problem. Theorem 7.10. Let the following Condition B0 be fulfilled for an admissible trajectory Tˆ in the basic problem: (a) û(t) is a bang-bang control taking values in the set V = ex U , (b) the set M0 is nonempty, and there exists λ ∈ M0 such that D k (H ) > 0, k = 1, . . . , s, and condition (6.19) holds (hence, u(t) is a strict bang-bang control), (c) maxμ∈MP Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ > 0 for all ζ̄ ∈ K0 \ {0}. 0 Then Tˆ is a strict strong minimum. Theorem 7.10 is a generalization of sufficient optimality conditions for bang-bang controls obtained in Agrachev, Stefani, and Zezza [1]. The detailed proofs of the preceding theorems will be given in the following sections. Let us point out that the proofs reveals the useful fact that all elements of the Hessian Lζ ζ (μ, ζ̂ ) can be computed explicitly on the basis of the transition matrix V (tˆf ) in (7.15) and of the first-order variations y k defined by (7.16). We shall need formulas for all first-order partial derivatives of the function x(tf ; t0 , x0 , θ ). We shall make extensive use of the variational system V̇ = fx (t, x(t; t0 , x0 , θ ), u(t; θ)) V , V (t0 ) = E, (7.18) where E is the identity matrix. The solution V (t) is an n × n matrix-valued function (n = d(x)) which is absolutely continuous in = [t0 , tf ]. The solution of (7.18) is denoted by V (t; t0 , x0 , θ). Along the reference trajectory x̂(t), û(t), i.e., for ζ = ζ̂ , we shall use the notation V (t) for simplicity. 7.2. First-Order Derivatives of x(tf ; t0 , x0 , θ ) 7.2 305 First-Order Derivatives of x(tf ; t0 , x0 , θ ) with Respect to t0 , tf , x0 , and θ. Lagrange Multipliers and Critical Cones Let x(t; t0 , x0 , θ) be the solution of the IVP (7.3) and put g(ζ ) = g(t0 , tf , x0 , θ ) := x(tf ; t0 , x0 , θ ). (7.19) Under our assumptions, the operator g : V → Rn is well defined and C 2 -smooth if the neighborhood V of the point ζ̂ is sufficiently small. In this section, we shall derive the firstorder partial derivatives of g(t0 , tf , x0 , θ ) with respect to t0 , tf , x0 , and θ at the point ζ̂ . We shall use well-known results from the theory of ODEs about differentiation of solutions to ODEs with respect to parameters and initial values. In what follows, it will be convenient to drop those arguments in x(t; t0 , x0 , θ ), u(t, θ ), V (t; t0 , x0 , θ ), etc., that are kept fixed. 7.2.1 Derivative ∂x/∂x0 Let us fix θ and t0 . The following result is well known in the theory of ODEs. Proposition 7.11. We have ∂x(t; x0 ) = V (t; x0 ), ∂x0 where the matrix-valued function V (t; x0 ) is the solution to the IVP (7.18), i.e., V̇ = fx (t, x(t), u(t))V , where x(t) = x(t; x0 ), V̇ = ∂V ∂t V |t=t0 = E, (7.20) (7.21) . Consequently, we have gx0 (ζ̂ ) := ∂x(tˆf ; tˆ0 , x̂0 , θ̂ ) = V (tˆf ), ∂x0 (7.22) where V (t) satisfies IVP (7.21) along the trajectory (x̂(t), û(t)), t ∈ [tˆ0 , tˆf ]. 7.2.2 Derivatives ∂x/∂t0 and ∂x/∂tf Fix x0 and θ and put w(t; t0 ) = ∂x(t; t0 ) . ∂t0 Proposition 7.12. The vector function w(t; t0 ) is the solution to the IVP ẇ = fx (t, x(t), u(t))w, w|t=t0 = −ẋ(t0 ), (7.23) where x(t) = x(t; t0 ), ẇ = ∂w ∂t . Therefore, we have w(t; t0 ) = −V (t; t0 )ẋ(t0 ), where the matrix-valued function V (t; t0 ) is the solution to the IVP (7.18). 306 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Hence, we obtain gt0 (ζ̂ ) := ∂x(tˆf ; tˆ0 , x̂0 , θ̂ ) ˙ tˆ0 ). = −V (tˆf )x̂( ∂t0 (7.24) ∂x(tˆf ; tˆ0 , x̂0 , θ̂ ) ˙ = x̂(tˆf ). ∂tf (7.25) Obviously, we have gtf (ζ̂ ) := 7.2.3 Derivatives ∂x/∂tk Fix t0 and x0 . Take some k and fix tj for all j = k. Put y k (t; tk ) = ∂x(t; tk ) ∂tk and denote by ẏ k the derivative of y k with respect to t. Proposition 7.13. For t ≥ tk the function y k (t; tk ) is the solution to the IVP ẏ k = fx (t, x(t; tk ), u(t; tk )) y k , y k |t=tk = −[f ]k , (7.26) where [f ]k = f (tk , x(tk ; tk ), uk+1 )−f (tk , x(tk ; tk ), uk ) is the jump of the function f (t, x(t; tk ), u(t; tk )) at the point tk . For t < tk , we have y k (t; tk ) = 0. Thus, [y]k = −[f ]k , where [y]k = y(tk +; tk ) − y(tk −; tk ) is the jump of the function y k (t; tk ) at the point tk . Proof. Let us sketch how to obtain the representation (7.26). For t ≥ tk the trajectory x(t; tk ) satisfies the integral equation t x(t; tk ) = x(tk −; tk ) + f (h, x(h; tk ), u(h, tk )) dh. tk + By differentiating this equation with respect to tk , we obtain t fx (h, x(h; tk ), u(h, tk ))y k (h; tk ) dh, y k (t; tk ) = ẋ(tk −; tk ) − ẋ(tk +; tk ) + tk + from which we get y k |t=tk = −[f ]k and the variational equation in (7.26). In particular, we obtain gtk (ζ̂ ) := 7.2.4 ∂x(tˆf ; tˆ0 , x̂0 , θ̂ ) = y k (tˆf ). ∂tk (7.27) Comparison of Lagrange Multipliers Here, we prove Propositions 7.4 and 7.5. Consider the Lagrangian (7.6) with a multiplier μ = (α0 , α, β) ∈ 0 , where 0 is the set (7.7) of normalized Lagrange multipliers at the point ζ̂ in the Induced Problem (7.4). Define the absolutely continuous function ψ(t) and the function ψ0 (t) by equation (7.10) and (7.11), respectively. We will show that the function 7.2. First-Order Derivatives of x(tf ; t0 , x0 , θ ) 307 ψ0 (t) is absolutely continuous and the collection λ = (α0 , α, β, ψ, ψ0 ) satisfies all conditions in (7.8) and hence belongs to the set . The conditions α0 ≥ 0, α ≥ 0, α0 + |α| + |β| = 1, αF (p̂) = 0 in the definitions of 0 and are identical. Hence, we must analyze the equation Lζ (μ, ζ̂ ) = 0 in the definition of 0 which is equivalent to the system Lt0 (μ, ζ̂ ) = lt0 (p̂) + lxf (p̂)gt0 (ζ̂ ) = 0, Ltf (μ, ζ̂ ) = ltf (p̂) + lxf (p̂)gtf (ζ̂ ) = 0, Lx0 (μ, ζ̂ ) = lx0 (p̂) + lxf (p̂)gx0 (ζ̂ ) = 0, Ltk (μ, ζ̂ ) = lxf (p̂)gtk (ζ̂ ) = 0, k = 1, . . . , s. Using the equality lxf (p̂) = ψ(tˆf ) and formulas (7.24), (7.25), (7.22), (7.27) for the derivatives of g with respect to t0 , tf , x0 , tk , at the point ζ̂ , we get ˙ tˆ0 ) = 0, Lt0 (μ, ζ̂ ) = lt0 (p̂) − ψ(tˆf )V (tˆf )x̂( ˙ tˆf ) = 0, Lt (μ, ζ̂ ) = lt (p̂) + ψ(tˆf )x̂( (7.29) Lx0 (μ, ζ̂ ) = lx0 (p̂) + ψ(tˆf )V (tˆf ) = 0, (7.30) Ltk (μ, ζ̂ ) = ψ(tˆf )y (tˆf ) = 0, (7.31) f f k k = 1, . . . , s. (7.28) Analysis of (7.28). The n × n matrix-value function V (t) satisfies the equation V̇ = fx V , V (tˆ0 ) = E with fx = fx (t, x̂(t), û(t)). Then, "(t) := V −1 (t) is the solution to the adjoint equation ˙ = "fx , −" "(tˆ0 ) = E. Consequently, ψ(tˆf ) = ψ(tˆ0 )"(tˆf ) = ψ(tˆ0 )V −1 (tˆf ). Using these relations in (7.28), we get ˙ tˆ0 ) = 0. lt0 (p̂) − ψ(tˆ0 )x̂( ˙ tˆ0 ) = −ψ0 (tˆ0 ). Hence, (7.28) is equivalent to the By virtue of (7.11), we have ψ(tˆ0 )x̂( transversality condition for ψ0 at the point tˆ0 : lt0 (p̂) + ψ0 (tˆ0 ) = 0. ˙ tˆf ) = −ψ0 (tˆf ) holds, (7.29) is equivalent to the Analysis of (7.29). Since ψ(tˆf )x̂( transversality condition for ψ0 at the point tˆf : ltf (p̂) − ψ0 (tˆf ) = 0. Analysis of (7.30). Since ψ(tˆf ) = ψ(tˆ0 )V −1 (tˆf ), equality (7.30) is equivalent to the transversality condition for ψ at the point tˆ0 : lx0 (p̂) + ψ(tˆ0 ) = 0. 308 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Analysis of (7.31). We need the following result. Proposition 7.14. Let the absolutely continuous function y be a solution to the system ẏ = fx y on an interval , and let the absolutely continuous function ψ be a solution to the adjoint system −ψ̇ = ψfx on the same interval, where fx = fx (t, x̂(t), û(t)). Then ψ(t)y(t) ≡ const on . Proof. We have d dt (ψy) = ψ̇y + ψ ẏ = −ψfx y + ψfx y = 0. It follows from this proposition and (7.26) that for k = 1, . . . , s, ˙ k = −[ψ x̂] ˙ k = [ψ0 ]k . ψ(tˆf )y k (tˆf ) = ψ(tˆk )y k (tˆk + 0) = ψ(tˆk )[y k ]k = −ψ(tˆk )[x̂] Therefore, (7.31) is equivalent to the conditions [ψ0 ]k = 0, k = 1, . . . , s, which means that ψ0 is continuous at each point tˆk , k = 1, . . . , s, and hence absolutely ˙ k that ˆ = [tˆ0 , tˆf ]. Moreover, it follows from 0 = [ψ0 ]k = −ψ(tˆk )[x̂] continuous on φ(tˆk )[û]k = 0, (7.32) where φ(t) = ψ(t)B(t, x̂(t)) denotes the switching function. Finally, differentiating (7.11) with respect to t, we get −ψfx x̂˙ + ψft + ψfx x̂˙ + ψ̇0 = 0, i.e., − ψ̇0 = ψft . Thus, we have proved that λ = (α0 , α, β, ψ, ψ0 ) ∈ . Conversely, if (α0 , α, β, ψ) ∈ , then one can show similarly that (α0 , α, β) ∈ 0 . Moreover, it is obvious that the projector (7.9) is injective on 0 , because ψ and ψ0 are defined uniquely by condition (7.10) and (7.11), respectively. 7.2.5 Comparison of the Critical Cones Take an element ζ̄ = (t¯0 , t¯f , x̄0 , θ̄) of the critical cone K0 (see (7.13)) at the point ζ̂ in the Induced Problem: F0 (ζ̂ )ζ̄ ≤ 0, Fi (ζ̂ )ζ̄ ≤ 0, i ∈ I , G (ζ̂ )ζ̄ = 0. Define ξ̄ and x̄ by formulas (7.17), ξ̄ = −θ̄ , ˙ tˆ0 )t¯0 + x̄(t) = V (t) x̄0 − x̂( s y k (t)t¯k , k=1 where t¯k is the kth component of the vector θ̄ , and put z̄ = (t¯0 , t¯f , ξ̄ , x̄). We shall show that z̄ is an element of the critical cone K (Equations (6.22) and (6.23)) for the trajectory 7.2. First-Order Derivatives of x(tf ; t0 , x0 , θ ) 309 Tˆ = {(x̂(t), û(t) | t ∈ [tˆ0 , tˆf ] } in the basic problem. Consider the first inequality F0 (ζ̂ )ζ̄ ≤ 0, where F0 (ζ ) := J (t0 , x0 , tf , x(tf ; t0 , x0 , θ )). We obviously have F0 (ζ̂ )ζ̄ = (Jt0 (p̂) + Jxf (p̂)gt0 (ζ̂ ))t¯0 + (Jtf (p̂) + Jxf (p̂)gtf (ζ̂ ))t¯f + (Jx0 (p̂) + Jxf (p̂)gx0 (ζ̂ ))x̄0 + s Jxf (p̂)gtk (ζ̂ )θ̄k , k=1 where θ̄k = t¯k is the kth component of the vector θ̄. Using formulas (7.24), (7.25), (7.22), (7.27) for the derivatives of g with respect to t0 , tf , x0 , tk , at the point ζ̂ , we get ˙ tˆ0 ))t¯0 + (Jt (p̂) + Jx (p̂)x̂( ˙ tˆf ))t¯f F0 (ζ̂ )ζ̄ = (Jt0 (p̂) − Jxf (p̂)V (tf )x̂( f f + (Jx0 (p̂) + Jxf (p̂)V (tˆf ))x̄0 + s Jxf (p̂)y k (tˆf )θ̄k . k=1 Hence, the inequality F0 (ζ̂ )ζ̄ ≤ 0 is equivalent to the inequality Jt0 (p̂)t¯0 + Jtf (p̂)t¯f + Jx0 (p̂)x̄0 s k ˙ ˙ + Jx (p̂) V (tf )(x̄0 − x̂(tˆ0 )t¯0 ) + y (tˆf )θ̄k + x̂(tˆf )t¯f ≤ 0. f k=1 It follows from the definition (7.17) of x̄ that ˙ tˆ0 )t¯0 = x̄0 , x̄¯0 := x̄(tˆ0 ) + x̂( (7.33) since V (tˆ0 ) = E, and y k (tˆ0 ) = 0, k = 1, . . . , s. Moreover, using the same definition, we get ˙ tˆf )t¯f = V (tˆf )(x̄0 − x̂( ˙ tˆ0 )t¯0 ) + x̄¯f := x̄(tˆf ) + x̂( s ˙ tˆf )t¯f . y k (tˆf )t¯k + x̂( (7.34) k=1 Thus, the inequality F0 (ζ̂ )ζ̄ ≤ 0 is equivalent to the inequality Jt0 (p̂)t¯0 + Jtf (p̂)t¯f + Jx0 (p̂)x̄¯0 + Jxf (p̂)x̄¯f ≤ 0, or briefly, J (p̂)p̄¯ ≤ 0, where p̄¯ = (t¯0 , x̄¯0 , t¯f , x̄¯f ); see equation (6.21). Similarly, the inequalities Fi (ζ̂ )ζ̄ ≤ 0 for all i ∈ I and the equality G (ζ̂ )ζ̄ = 0 in the definition of K0 are equivalent to the inequalities (respectively, equalities) Fi (p̂)p̄¯ ≤ 0, i ∈ I, K (p̂)p̄¯ = 0, in the definition of K; cf. (6.22). Since V̇ = fx (t, x̂(t), û(t))V and ẏ k = fx (t, x̂(t), û(t))y k , k = 1, . . . , s, it follows from definition (7.17) that x̄ is a solution to the same linear system x̄˙ = fx (t, x̂(t), û(t))x̄. 310 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Finally, recall from (7.26) that for each k = 1, . . . , s the function y k (t) is piecewise ˙ k at the point tˆk and is absolutely continuous on continuous with only one jump [y k ]k = −[x̂] each of the half-open intervals [tˆ0 , tˆk ) and (tˆk , tˆf ]. Moreover, the function V (t) is absolutely continuous in [tˆ0 , tˆf ]. Hence, x̄(t) is a piecewise continuous function which is absolutely ˆ and satisfies the jump conditions continuous on each interval of the set [tˆ0 , tˆf ] \ ˙ k ξ̄k , [x̄]k = [x̂] ξ̄k = −t¯k , k = 1, . . . , s. Thus, we have proved that z̄ = (t¯0 , t¯f , ξ̄ , x̄) is an element of the critical cone K. Similarly, one can show that if z̄ = (t¯0 , t¯f , ξ̄ , x̄) ∈ K, then putting x̄0 = x̄(tˆ0 ) and θ̄ = −ξ̄ , we obtain the element ζ̄ = (t¯0 , t¯f , x̄0 , θ̄) of the critical cone K0 . 7.3 Second-Order Derivatives of x(tf ; t0 , x0 , θ ) with Respect to t0 , tf , x0 , and θ In this section we shall give formulas for all second-order partial derivatives of the functions x(t; t0 , x0 , θ) g(ζ ) = g(t0 , tf , x0 , θ ) := x(tf ; t0 , x0 , θ ) and at the point ζ̂ . We are not sure whether all of them are known; therefore we shall also sketch the proofs. Here x(t; t0 , x0 , θ) is the solution to IVP (7.3). Denote by gk (ζ ) := xk (tf ; t0 , x0 , θ ) the kth component of the function g. 7.3.1 Derivatives (gk )x0 x0 Let x(t; x0 ) be the solution to the IVP (7.3) with fixed t0 and θ , and let xk (t; x0 ) be its kth component. For k = 1, . . . , n, we define the n × n matrix W k (t; x0 ) := ∂ 2 xk (t; x0 ) ∂x0 ∂x0 with entries wijk (t; x0 ) = ∂ 2 xk (t; x0 ) , ∂x0i ∂x0j where x0i is the ith component of the column vector x0 ∈ Rn . Proposition 7.15. The matrix-valued functions W k (t; x0 ), k = 1, . . . , n, satisfy the IVPs Ẇ k = V ∗fkxx V + n fkxr W r , W k |t=t0 = O, k = 1, . . . , n, (7.35) r=1 where Ẇ k = ∂W k ∂t , O is the zero matrix, fk is the kth component of the vector function f , and fkxr = ∂fk (t, x(t; x0 ), u(t)) , ∂xr fkxx = ∂ 2 fk (t, x(t; x0 ), u(t)) ∂x∂x are its partial derivatives at the point (t, x(t; x0 ), u(t)) for t ∈ [t0 , tf ]. Proof. For notational convenience, we use the function ϕ(t, x) := f (t, x, u(t)). By Propo∂xi (t;x0 ) 0) sition 7.11, the matrix-valued function V (t; x0 ) = ∂x(t;x ∂x0 with entries vij (t; x0 ) = ∂x0j is 7.3. Second-Order Derivatives of x(tf ; t0 , x0 , θ ) with Respect to t0 , tf , x0 , and θ 311 the solution to the IVP (7.18). Consequently, its entries satisfy the equations ∂ ẋk (t; x0 ) ∂xr (t; x0 ) = ϕkxr (t, x(t; x0 )) ∂x0i ∂x0i r ∂xk (t0 ; x0 ) = eki , ∂x0i k, i = 1, . . . , n, where eki are the elements of the identity matrix E. By differentiating these equations with respect to x0j , we get ∂xr (t; x0 ) ∂ 2 ẋk (t; x0 ) ϕkxr (t, x(t; x0 )) x = 0j ∂x0i ∂x0j ∂x0i r + ϕkxr (t, x(t; x0 )) r ∂ 2 xk (t0 ; x0 ) = 0, ∂x0i ∂x0j ∂ 2 xr (t; x0 ) , ∂x0i ∂x0j k, i, j = 1, . . . , n. (7.36) (7.37) Transforming the first sum in the right-hand side of (7.36), we get ϕkxr (t, x(t; x0 )) r = r x0j ∂xr (t; x0 ) ∂x0i ϕkxr xs (t, x(t; x0 )) s = V ∗ ϕkxx (t, x(t; x0 ))V ij , ∂xs (t; x0 ) ∂xr (t; x0 ) · ∂x0j ∂x0i k, i, j = 1, . . . , n, where (A)ij denotes the element aij of a matrix A, and A∗ denotes the transposed matrix. Thus, (7.36) and (7.37) imply (7.35). It follows from Proposition 7.15 that (gk )x0 x0 (ζ̂ ) := ∂ 2 xk (tˆf ; tˆ0 , x̂0 , θ̂ ) = W k (tˆf ), ∂x0 ∂x0 k = 1, . . . , n, (7.38) where the matrix-valued functions W k (t), k = 1, . . . , n, satisfy the IVPs (7.35) along the reference trajectory (x̂(t), û(t)). 7.3.2 Mixed Derivatives gx0 tk Let s = 1 for notational convenience, and thus θ = t1 . Fix t0 and consider the functions ∂x(t; x0 , θ ) ∂x(t; x0 , θ ) , , y(t; x0 , θ ) = ∂x0 ∂θ ∂V (t; x0 , θ ) ∂ 2 x(t; x0 , θ ) R(t; x0 , θ) = = , ∂θ ∂x0 ∂θ ∂V (t; x0 , θ ) ∂R(t; x0 , θ ) , Ṙ(t; x0 , θ ) = . V̇ (t; x0 , θ) = ∂t ∂t V (t; x0 , θ) = 312 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Then V , V̇ and R, Ṙ are n × n matrix-valued functions, and y is a vector function of dimension n. Proposition 7.16. For t ≥ θ, the function R(t; x0 , θ ) is the solution to the IVP Ṙ = (y ∗fxx )V + fx R, R(θ; x0 , θ ) = −[fx ]V (θ ; x0 , θ ), (7.39) where fx and fxx are taken along the trajectory (t, x(t; x0 , θ ), u(t, θ )), t ∈ [t0 , tf ]. Here, by definition, (y ∗fxx ) is an n × n matrix with entries ∗ (y fxx )k,j n ∂ 2 fk = yi ∂xi ∂xj (7.40) i=1 in the kth row and j th column, and [fx ] = fx (θ , x(θ ; x0 , θ ), u2 ) − fx (θ , x(θ; x0 , θ ), u1 ) is the jump of the function fx (·, x(·; x0 , θ ), u(·, θ )) at the point θ . For t < θ, we have R(t; x0 , θ) = 0. Proof. According to Proposition 7.11 the matrix-valued function V is the solution to the system V̇ (t; x0 , θ) = fx (t, x(t; x0 , θ ), u(t; θ))V (t; x0 , θ ). (7.41) By differentiating this equality with respect to θ , we get the equation ∂xi ∂V ∂ V̇ = + fx , (fx V )xi ∂θ ∂θ ∂θ i which is equivalent to Ṙ = (fx V )xi yi + fx R. (7.42) i Upon defining A= (fx V )xi yi , i the element in the rth row and sth column of the matrix A is equal to ⎛ ⎞ ⎝ ((fx V )rs )xi yi = frxj vj s ⎠ yi ars = i = i = i yi frxi xj vj s = j ∗ y fxx v rj j s j x j ∗ = (y fxx )V i yi frxi xj vj s i rs , j where vj s is the element in the j th row and sth column of the matrix V . Hence we have A = (y ∗fxx )V and see that (7.42) is equivalent to (7.39). The initial condition in (7.39), 7.3. Second-Order Derivatives of x(tf ; t0 , x0 , θ ) with Respect to t0 , tf , x0 , and θ 313 which is similar to the initial condition (7.26) in Proposition 7.13, follows from (7.41) (see the proof of Proposition 7.13). The condition R(t; x0 , θ ) = 0 for t < θ is obvious. Proposition 7.16 yields gx0 tk (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) = R k (tˆf ), ∂x0 ∂tk (7.43) where the matrix-valued function R k (t) satisfies the IVP Ṙ k (t) = y k (t)∗ fxx (t, x̂(t), û(t)) V (t) + fx (t, x̂(t), û(t))R k (t), t ∈ [tˆk , tˆf ], R k (tˆk ) = −[fx ]k V (tˆk ). (7.44) Here, V (t) is the solution to the IVP (7.18), y k (t) is the solution to the IVP (7.26) (for t0 = tˆ0 , x0 = x̂0 , θ = θ̂), and [fx ]k = f (tˆk , x̂(tˆk ), û(tˆk +)) − f (tˆk , x̂(tˆk ), û(tˆk −)), k = 1, . . . , s. 7.3.3 Derivatives gtk tk Again, for simplicity let s = 1. Fix t0 and x0 and put ∂x(t; θ ) , ∂θ ∂y(t; θ ) , ẏ(t; θ ) = ∂t y(t; θ) = ∂y(t; θ ) ∂ 2 x(t; θ) = , ∂θ ∂θ 2 ∂z(t; θ) ż(t; θ ) = . ∂t z(t; θ ) = Then y, ẏ and z, ż are vector functions of dimension n. Proposition 7.17. For t ≥ θ the function z(t; θ ) is the solution to the system ż = fx z + y ∗fxx y (7.45) with the initial condition at the point t = θ , z(θ; θ ) + ẏ(θ +; θ ) = −[ft ] − [fx ](ẋ(θ +; θ ) + y(θ ; θ)). (7.46) In (7.45), fx and fxx are taken along the trajectory (t, x(t; θ ), u(t; θ )), t ∈ [t0 , tf ], and y ∗fxx y is a vector with elements (y ∗fxx y)k = y ∗fkxx y = n ∂ 2 fk yi yj , ∂xi ∂xj k = 1, . . . , n. i,j =1 In (7.46), the expressions [ft ] = ft (θ, x(θ ; θ ), u2 ) − ft (θ , x(θ; θ ), u1 ), [fx ] = fx (θ , x(θ; θ ), u2 ) − fx (θ , x(θ; θ ), u1 ) are the jumps of the derivatives ft (t, x(t; θ), u(t; θ )) and fx (t, x(t; θ), u(t; θ )) at the point θ (u2 = u(θ +; θ), u1 = u(θ−; θ )). For t < θ , we have z(t; θ) = 0. 314 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Proof. By Proposition 7.13, for t ≥ θ the function y(t; θ ) is the solution to the IVP ẏ(t; θ ) = fx (t, x(t; θ ), u(t; θ))y(t; θ), y(θ ; θ ) = −(f (θ, x(θ ; θ ), u2 ) − f (θ , x(θ; θ ), u1 )). By differentiating these equalities with respect to θ at the points θ and θ +, we obtain (7.45) and (7.46). For t < θ we have y = 0 and hence z = 0. For the solution x(t; t0 , x0 , θ ) to the IVP (7.3) with an arbitrary s, it follows from Proposition 7.17 that gtk tk (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) = zkk (tˆf ), ∂tk ∂tk k = 1, . . . , s, (7.47) where for t ≥ tˆk the vector function zkk (t) satisfies the equation żkk (t) = fx (t, x̂(t), û(t))zkk (t) + y k (t)∗ fxx (t, x̂(t), û(t))y k (t) (7.48) with the initial condition at the point t = tˆk : ˙ tˆk +) + y k (tˆk )). zkk (tˆk ) + ẏ k (tˆk +) = −[ft ]k − [fx ]k (x̂( (7.49) Here, for t ≥ tˆk , the function y k (t) is the solution to the IVP (7.26), and y k (t) = 0 for t < tˆk , k = 1, . . . , s. Furthermore, by definition, [ft ]k = ft (tˆk , x̂(tˆk ), û(tˆk +)) − ft (tˆk , x̂(tˆk ), û(tˆk −)) and [fx ]k = fx (tˆk , x̂(tˆk ), û(tˆk +)) − fx (tˆk , x̂(tˆk ), û(tˆk −)) are the jumps of the derivatives ft (t, x̂(t), û(t)) and fx (t, x̂(t), û(t)) at the point tˆk . For t < tˆk we put zkk (t) = 0, k = 1, . . . , s. 7.3.4 Mixed Derivatives gtk tj For simplicity, let s = 2, θ = (t1 , t2 ), and t0 < t1 < t2 < tf . Fix x0 and t0 and put ∂x(t; θ ) ∂y 1 (t; θ ) ∂ 2 x(t; θ ) , k = 1, 2, z12 (t; θ ) = = , ∂tk ∂t2 ∂t1 ∂t2 ∂y k (t; θ ) ∂z12 (t; θ ) , k = 1, 2, ż12 (t; θ ) = . ẏ k (t; θ ) = ∂t ∂t y k (t; θ ) = Then y k , ẏ k , k = 1, 2, and z12 , ż12 are vector functions of dimension n. Proposition 7.18. For t ≥ t2 the function z12 (t; θ ) is the solution to the system ż12 = fx z12 + (y 1 )∗ fxx y 2 (7.50) with the initial condition at the point t = t2 , z12 (t2 ; θ ) = −[ẏ 1 ]2 . (7.51) 7.3. Second-Order Derivatives of x(tf ; t0 , x0 , θ ) with Respect to t0 , tf , x0 , and θ 315 In (7.50), fx and fxx are taken along the trajectory (t, x(t; θ ), u(t; θ )), t ∈ [t0 , tf ], and (y 1 )∗ fxx y 2 is a vector with elements ((y 1 )∗ fxx y 2 )k = (y 1 )∗ fkxx y 2 = n ∂ 2 fk 1 2 y y , ∂xi ∂xj i j k = 1, . . . , n. i,j =1 In (7.51) we have [ẏ 1 ]2 = [fx ]2 y 1 (t2 ; θ ), where [fx ]2 = fx (t2 , x(t2 ; θ), u3 ) − fx (t2 , x(t2 ; θ ), u2 ). For t < t2 we have z12 (t; θ ) = 0. Proof. By Proposition 7.13, for t ≥ t1 the function y 1 (t; θ ) is a solution to the equation ẏ 1 (t; θ ) = fx (t, x(t; θ ), u(t; θ))y 1 (t; θ ), where y 1 (t; θ ) = 0 for t < t1 . Differentiating this equation with respect to t2 , we see that for 1 t ≥ t2 , the function z12 (t; θ ) = ∂y ∂t(t;θ) is a solution to system (7.50). The initial condition 2 (7.51) is similar to the initial condition (7.26) in Proposition 7.13. For t < t2 , we obviously have z12 (t; θ ) = 0. For the solution x(t; t0 , x0 , θ ) of IVP (7.3) and for tk < tj (k, j = 1, . . . , s), it follows from Proposition 7.18 that gtk tj (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) = zkj (tˆf ), ∂tk ∂tj (7.52) where for t ≥ tˆj the vector function zkj (t) is the solution to the equation żkj (t) = fx (t, x̂(t), û(t))zkj (t) + y k (t)∗ fxx (t, x̂(t), û(t))y j (t) (7.53) satisfying the initial condition zkj (tˆj ) = −[ẏ k ]j = −[fx ]j y k (tˆj ). (7.54) Here, for t ≥ tˆk , the function y k (t) is the solution to the IVP (7.26), while = 0 holds for t < tˆk , k = 1, . . . , s. By definition, [ẏ k ]j = ẏ k (tˆj +) − ẏ k (tˆj −) and [fx ]j = fx (tˆj , x̂(tˆj ), û(tˆj +)) − fx (tˆj , x̂(tˆj ), û(tˆj −)) are the jumps of the derivatives ẏ k (t) and fx (t, x̂(t), û(t)), respectively, at the point tˆj . For t < tˆj we put zkj (t) = 0. y k (t) 7.3.5 Derivatives gt0 t0 , gt0 tf , and gtf tf Here, we fix x0 and θ and study the functions w(t; t0 ) = ∂x(t; t0 ) , ∂t0 q(t; t0 ) = ∂w(t; t0 ) ∂ 2 x(t; t0 ) = , ∂t0 ∂t02 ẇ(t; t0 ) = ∂w(t; t0 ) , ∂t q̇(t; t0 ) = ∂q(t; t0 ) , ∂t ẍ(t; t0 ) = ∂ 2 x(t; t0 , ) . ∂t 2 316 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Proposition 7.19. The function q(t; t0 ) is the solution to the system q̇ = fx q + w ∗fxx w, t ∈ [t0 , tf ] (7.55) satisfying the initial condition at the point t = t0 , ẍ(t0 ; t0 ) + 2ẇ(t0 ; t0 ) + q(t0 ; t0 ) = 0. (7.56) In (7.55), fx and fxx are taken along the trajectory (t, x(t; t0 ), u(t)), t ∈ [t0 , tf ], and w ∗fxx w is a vector with elements (w ∗fxx w)k = w∗fkxx w = n ∂ 2 fk wi wj , ∂xi ∂xj k = 1, . . . , n. i,j =1 Proof. By Proposition 7.12 we have ẇ(t; t0 ) = fx (t, x(t; t0 ))w(t; t0 ), ẋ(t0 ; t0 ) + w(t0 ; t0 ) = 0. Differentiating these equalities with respect to t0 , we obtain (7.55) and (7.56). From Proposition 7.19 it follows that gt0 t0 (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) ∂t02 = q(tˆf ), (7.57) where the vector function q(t) is the solution to the equation q̇(t) = fx (t, x̂(t), û(t))q(t) + w ∗ (t)fxx (t, x̂(t), û(t))w(t) (7.58) satisfying the initial condition ¨ tˆ0 ) + 2ẇ(tˆ0 ) + q(tˆ0 ) = 0. x̂( (7.59) ˙ tˆ0 ) in view of Proposition 7.12, V̇ = fx V , and V (tˆ0 ) = E, we obtain Since w(t) = −V (t)x̂( ˙ tˆ0 ) = −fx (tˆ0 , x̂(tˆ0 ), û(tˆ0 ))x̂( ˙ tˆ0 ). ẇ(tˆ0 ) = −V̇ (tˆ0 )x̂( Thus, the initial condition (7.59) is equivalent to ¨ tˆ0 ) − 2fx (tˆ0 , x̂(tˆ0 ), û(tˆ0 ))x̂( ˙ tˆ0 ) + q(tˆ0 ) = 0. x̂( (7.60) From (7.24) it follows that ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) ∂t0 ∂tf ˙ tˆ0 ) = −fx (tˆf , x̂(tˆf ), û(tˆf ))V (tˆf )x̂( ˙ tˆ0 ). = −V̇ (tˆf )x̂( gt0 tf (ζ̂ ) := (7.61) Formula (7.25) implies that gtf tf (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) ∂tf2 ¨ tˆf ). = x̂( (7.62) 7.3. Second-Order Derivatives of x(tf ; t0 , x0 , θ ) with Respect to t0 , tf , x0 , and θ 7.3.6 317 Derivatives gx0 tf and gtk tf Formula (7.22) implies that gx0 tf (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) = V̇ (tˆf ), ∂x0 ∂tf (7.63) where V (t) is the solution to the IVP (7.18). From (7.27) it follows that gtk tf (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) = ẏ k (tˆf ), ∂tk ∂tf k = 1, . . . , s, (7.64) where y k (t) is the solution to the IVP (7.26). 7.3.7 Derivative gx0 t0 Let us fix θ and consider ∂x(t; t0 , x0 ) ∂V (t; t0 , x0 ) ∂ 2 x(t; t0 , x0 ) , S(t; t0 , x0 ) = = , ∂x0 ∂t0 ∂x0 ∂t0 ∂V (t; t0 , x0 ) ∂S(t; t0 , x0 ) , Ṡ(t; t0 , x0 ) = . V̇ (t; t0 , x0 ) = ∂t ∂t V (t; t0 , x0 ) = Proposition 7.20. The elements sij (t; t0 , x0 ) of the matrix S(t; t0 , x0 ) satisfy the system ṡij = −ej∗ V ∗ (fi )xx V ẋ(t0 ) + fix Sej , i, j = 1, . . . , n, (7.65) and the matrix S itself satisfies the initial condition at the point t = t0 , S(t0 ; t0 , x0 ) + V̇ (t0 ; t0 , x0 ) = 0. (7.66) In (7.65), the derivatives fx and fxx are taken along the trajectory (t, x(t; t0 , x0 ), u(t)), t ∈ [t0 , tf ], ej is the j th column of the identity matrix E, and, by definition, ẋ(t0 ) = ẋ(t0 ; t0 , x0 ). Proof. By Proposition 7.11, V̇ (t; t0 , x0 ) = fx (t, x(t; t0 , x0 ), u(t))V (t; t0 , x0 ), V (t0 ; t0 , x0 ) = E. (7.67) The first equality in (7.67) is equivalent to v̇ij (t; t0 , x0 ) = fix (t, x(t; t0 , x0 ), u(t))V (t, t0 )ej , i, j = 1, . . . , n. By differentiating these equalities with respect to t0 and using Proposition 7.12, we obtain (7.65). Differentiating the second equality in (7.67) with respect to t0 yields (7.66). Proposition 7.20 implies that gx0 t0 (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) = S(tˆf ), ∂x0 ∂t0 (7.68) 318 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem where the elements sij (t) of the matrix S(t) satisfy the system ˙ tˆ0 ) + fix (t, x̂(t), û(t))S(t)ej , ṡij (t) = ej∗ V ∗ (t)(fi )xx (t, x̂(t), û(t))V (t)x̂( i, j = 1, . . . , n. (7.69) Here, V (t) is the solution to the IVP (7.18), and the matrix S(t) itself satisfies the initial condition at the point t = tˆ0 , S(tˆ0 ) + V̇ (tˆ0 ) = 0. (7.70) . 7.3.8 Derivative gtk t0 Consider again the case s = 1 with θ = t1 and define ∂x(t; t0 , θ ) , ∂θ ∂y(t; t0 , θ ) , ẏ(t; t0 , θ) = ∂t ∂x(t; t0 , θ ) ẋ(t; t0 , θ) = , ∂t y(t; t0 , θ) = ∂y(t; t0 , θ ) ∂ 2 x(t; t0 , θ ) , = ∂t0 ∂t0 ∂θ ∂r(t, t0 , θ ) ṙ(t; t0 , θ ) = , ∂t ∂x(t; t0 , θ ) . V (t; t0 , θ ) = ∂x0 r(t; t0 , θ ) = Proposition 7.21. For t ≥ θ , the function r(t; t0 , θ ) is the solution to the IVP ṙ = fx r − y ∗fxx V ẋ(t0 ), r|t=θ = [fx ]V (θ)ẋ(t0 ), (7.71) where y ∗fxx V ẋ(t0 ) is the vector with elements (y ∗fxx V ẋ(t0 ))i = y ∗fixx V ẋ(t0 ), i = 1, . . . , n, V (θ ) = V (θ; t0 , θ), and [fx ] = fx (θ , x(θ; t0 , θ ), u2 ) − fx (θ , x(θ; t0 , θ ), u1 ) is the jump of the derivative fx (t, x(t; t0 , θ ), u(t; θ )) at the point θ . The derivatives fx and fxx are taken along the trajectory (t, x(t; t0 , θ ), u(t; θ )), t ∈ [θ , tf ]. For t < θ we have r(t; t0 , θ) = 0. Then the jump of the function r(t; t0 , θ ) at the point t = θ is given by [r] = [fx ]V (θ )ẋ(t0 ). Proof. By Proposition 7.13 we have y(t; t0 , θ ) = 0 for t < θ and hence r(t; t0 , θ ) = 0 for t < θ. According to the same proposition, for t ≥ θ the function y(t; t0 , θ ) satisfies the equation ẏ(t; t0 , θ) = fx (t, x(t; t0 , θ ), u(t; θ))y(t; t0 , θ ). Differentiating this equation with respect to t0 , we get ṙ = fx r + y ∗fxx ∂x . ∂t0 According to Proposition 7.12, ∂x(t; t0 , θ ) = −V (t; t0 , θ )ẋ(t0 ), ∂t0 7.4. Quadratic Form for the Induced Optimization Problem 319 where ẋ(t0 ) = ẋ(t0 ; t0 , θ). This yields ṙ = fx r − y ∗fxx V ẋ(t0 ). By Proposition 7.13, the following initial condition holds at the point t = θ : y(θ ; t0 , θ) = −(f (θ, x(θ ; t0 , θ ), u2 ) − f (θ , x(θ; t0 , θ ), u1 )). Differentiating this condition with respect to t0 , we get r|t=θ = −[fx ] ∂x |t=θ = [fx ]V (θ)ẋ(t0 ), ∂t0 where V (θ ) = V (θ ; t0 , θ). It follows from Proposition 7.21 that for each k = 1, . . . , s, gtk t0 (ζ̂ ) := ∂ 2 x(tˆf ; tˆ0 , x̂0 , θ̂ ) = r k (tˆf ), ∂tk ∂t0 (7.72) where the function r k (t) is the solution to the system ˙ tˆ0 ) ṙ k (t) = fx (t, x̂(t), û(t))r k (t) − (y k (t))∗ fxx (t, x̂(t), û(t))V (t)x̂( (7.73) and satisfies the initial condition at the point t = tˆk , ˙ tˆ0 ). r k (tˆk ) = [fx ]k V (tˆk )x̂( (7.74) Here V (t) is the solution to the IVP (7.18) and y k (t) is the solution to the IVP (7.26). The ˙ tˆ0 ) has components vector (y k )∗ fxx V x̂( ˙ tˆ0 ))j = (y k )∗ fj xx V x̂( ˙ tˆ0 ), ((y k )∗ fxx V x̂( 7.4 j = 1, . . . , n. Explicit Representation of the Quadratic Form for the Induced Optimization Problem Let the Lagrange multipliers μ = (α0 , α, β) ∈ 0 and λ = (α0 , α, β, ψ, ψ0 ) ∈ correspond to each other, i.e., let π0 λ = μ hold; see Proposition 7.4. For any ζ̄ = (t¯0 , t¯f , x̄0 , θ̄ ) ∈ R2+n+s , let us find an explicit representation for the quadratic form Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ . By definition, Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ = Lx0 x0 x̄0 , x̄0 + 2 s k=1 + 2Lx0 tf x̄0 t¯f + 2 s Lx0 tk x̄0 t¯k + s Ltk tj t¯k t¯j k,j =1 Ltk tf t¯k t¯f + Ltf tf t¯f2 k=1 + 2Lx0 t0 x̄0 t¯0 + 2 s k=1 Lt0 tk t¯0 t¯k + 2Lt0 tf t¯0 t¯f + Lt0 t0 t¯02 . (7.75) 320 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem All derivatives in formula (7.75) are taken at the point (μ, ζ̂ ). Now we shall calculate these derivatives. Recall the definition (7.6) of the Lagrangian, L(μ, ζ ) = L(μ, t0 , tf , x0 , θ ) = l(μ, t0 , x0 , tf , x(tf ; t0 , x0 , θ )). (7.76) Note that all functions V , W k , y k , zkj , S, R k , q, w, r k , introduced in Sections 7.1.2 and 7.3, depend now on t, t0 , x0 , and θ . For simplicity, we put V (t) = V (t; tˆ0 , x̂0 , θ̂ ), etc. 7.4.1 Derivative Lx0 x0 Using Proposition 7.11, we get ∂ l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) x̄0 = lx0 (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))x̄0 ∂x0 + lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))V (tf ; t0 , x0 , θ )x̄0 . (7.77) Let us find the derivative of this function with respect to x0 . We have ∂ ∂x0 lx0 (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))x̄0 = x̄0∗ lx0 x0 (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) + x̄0∗ lx0 xf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))V (tf ; t0 , x0 , θ ), (7.78) and ∂ ∂x0 lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))V (tf ; t0 , x0 , θ ))x̄0 = x̄0∗ V ∗ (tf ; t0 , x0 , θ ) lxf x0 (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) + lxf xf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))V (tf ; t0 , x0 , θ ) +lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) ∂ V (tf ; t0 , x0 , θ )x̄0 . ∂x0 (7.79) From (7.77)–(7.79) and the transversality condition lxf (p̂) = ψ(tˆf ) it follows that at the point ζ̂ , we have Lx0 x0 x̄0 , x̄0 = x̄0∗ lx0 x0 (p̂)x̄0 + 2x̄0∗ lx0 xf (p̂)V (tˆf )x̄0 + x̄0∗ V ∗ (tˆf )lxf xf (p̂)V (tˆf )x̄0 ∂ + ψ(tf ) (V (tf ; t0 , x0 , θ )x̄0 )x̄0 . (7.80) ∂x ζ =ζ̂ 0 Let us calculate the last term in this formula. Proposition 7.22. The following equality holds: ∂ ∗ k ψ(tf ) V (tf ; t0 , x0 , θ)x̄0 x̄0 = x̄0 ψk (tf )W (tf ; t0 , x0 , θ ) x̄0 . ∂x0 k (7.81) 7.4. Quadratic Form for the Induced Optimization Problem 321 Proof. For brevity, put ψ(tf ) = ψ, V (tf ; t0 , x0 , θ ) = V , W (tf ; t0 , x0 , θ ) = W . Then we have ∂ ∂ ∂x ∂x ∂ (V x̄0 ) x̄0 = ψ x̄0 x̄0 = ψ x̄0i x̄0 ψ ∂x0 ∂x0 ∂x0 ∂x0 ∂x0i i ∂ 2x ∂ 2 xk = ψ x̄0i x̄0j = ψk x̄0i x̄0j ∂x0i ∂x0j ∂x0i ∂x0j j i k j i ∂ 2 xk ∗ k x̄0i x̄0j = x̄0 = ψk ψk (tf )W x̄0 . ∂x0i ∂x0j i j k k Proposition 7.23. For ζ = ζ̂ , the following equality holds: d k ψk W = V ∗ Hxx V , dt (7.82) k where H = ψf (t, x, u), Hxx = Hxx (t, x̂(t), ψ(t), û(t)). Proof. According to Proposition 7.15, we have fkxr W r , Ẇ k = V ∗fkxx V + k = 1, . . . , n. (7.83) r Using these equations together with the adjoint equation −ψ̇ = ψfx , we obtain d k ψk W ψk Ẇ k ψ̇k W k + = dt k k k k ∗ r = − ψfxk W + ψk V fkxx V + fkxr W r k k ψfxk W k + V ∗ (ψk fkxx ) V + ψk fkxr W r = − k k k r r ∗ = − ψfxr W + V ψk fkxx V + ψk fkxr W r r r k k r ∗ = − ψfxr W + V (ψfxx )V + ψfxr W r = V ∗ Hxx V . r r Now we can prove the following assertion. Proposition 7.24. The following formula holds: ∂ ψ(tf ) V (tf ; t0 , x0 , θ )x̄0 x̄0 ∂x0 ζ̂ tˆf (V (t)x̄0 )∗ Hxx (t, x̂(t), û(t), ψ(t))V (t)x̄0 dt. = tˆ0 (7.84) 322 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Proof. Using Propositions 7.22 and 7.23 and the initial conditions W k (tˆ0 ) = 0 for k = 1. . . . , n, we get ∂ V (tf ; t0 , x0 , θ )x̄0 x̄0 ψ(tf ) ∂x0 ζ̂ ˆ tf ∗ k ∗ k = x̄0 ψk (tˆf )W (tˆf ) x̄0 = x̄0 ψk (t)W (t) x̄0 = tˆf tˆ0 = k tˆf tˆ0 x̄0∗ d ψk W k x̄0 dt = dt k k tˆf tˆ0 tˆ0 x̄0∗ V ∗ Hxx V x̄0 dt (V x̄0 )∗ Hxx (V x̄0 ) dt. In view of formulas (7.80) and (7.84), we obtain Lx0 x0 x̄0 , x̄0 = x̄0∗ lx0 x0 (p̂)x̄0 +2x̄0∗ lx0 xf (p̂)V (tˆf )x̄0 + (V (tˆf )x̄0 )∗ lxf xf (p̂)V (tˆf )x̄0 tˆf (V (t)x̄0 )∗ Hxx (t, x̂(t), ψ(t), û(t))V (t)x̄0 dt. + tˆ0 7.4.2 (7.85) Derivative Lx0 tk Differentiating (7.77) with respect to tk and using Propositions 7.13 and 7.16, we get ∂2 l(t0 , x0 , tf , x(tf ; t0 , x0 , θ ))x̄0 ∂x0 ∂tk ∂ = lx (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))x̄0 ∂tk 0 ∂ lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))V (tf ; t0 , x0 , θ )x̄0 + ∂tk ∂x(tf ; t0 , x0 , θ ) = x̄0∗ lx0 xf (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) ∂t k ∂ + lx (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) V (tf ; t0 , x0 , θ )x̄0 ∂tk f ∂V (tf ; t0 , x0 , θ ) + lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) x̄0 ∂tk = x̄0∗ lx0 xf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))y k (tf ; t0 , x0 , θ ) + (V (tf ; t0 , x0 , θ )x̄0 )∗ lxf xf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))y k (tf ; t0 , x0 , θ ) + lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))R k (tf ; t0 , x0 , θ ))x̄0 . (7.86) Hence at the point ζ = ζ̂ , we have ∗ Lx0 tk x̄0 t¯k = x̄0∗ lx0 xf (p̂)y k (tˆf )t¯k + V (tˆf )x̄0 lxf xf (p̂)y k (tˆf )t¯k + ψ(tˆf )R k (tˆf )x̄0 t¯k . Let us transform the last term. (7.87) 7.4. Quadratic Form for the Induced Optimization Problem 323 Proposition 7.25. The following formula holds: ψ(tˆf )R k (tˆf )x̄0 t¯k = −[Hx ]k V (tˆk )x̄0 t¯k + tˆf tˆk Hxx y k t¯k , V x̄0 dt. (7.88) Proof. Using equation (7.44) and the adjoint equation −ψ̇ = ψfx , we get for t ∈ [tˆk , tˆf ], d (ψR k ) dt = ψ̇R k + ψ Ṙ k = −ψfx R k + ψ ((y k )∗ fxx )V + fx R k ψj (y k )∗ fj xx V = (y k )∗ ψj fj xx V = (y k )∗ Hxx V , = j j where Hxx is taken along the trajectory (t, x̂(t), ψ(t), û(t)). Consequently, tˆf k ˆ k ˆ ˆ ˆ ψ(tf )R (tf ) = ψ(tk )R (tk ) + (y k )∗ Hxx V dt. tˆk Using the initial condition (7.44) for Rk at tˆk , we get ψ(tˆf )R k (tˆf ) = −ψ(tˆk )[fx ]k V (tˆk ) + tˆf tˆk (y k )∗ Hxx V dt. Hence, ψ(tˆf )R k (tˆf )x̄0 t¯k = −[Hx ]k V (tˆk )x̄0 t¯k + tˆf tˆk Hxx y k t¯k , V x̄0 dt. Formulas (7.87) and (7.88) and the condition y k (t) = 0 for t < tˆk imply the equality Lx0 tk x̄0 t¯k x̄0∗ lx0 xf (p̂)y k (tˆf )t¯k + (V (tˆf )x̄0 )∗ lxf xf (p̂)y k (tˆf )t¯k = − [Hx ]k V (tˆk )x̄0 t¯k + 7.4.3 tˆf Hxx y k t¯k , V x̄0 tˆ0 dt. Derivative Ltk tk Using the notation ∂x ∂tk = y k from Proposition 7.13, we get ∂ l(t0 , x0 , tf , x(tf ; t0 , x0 , θ)) = lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))y k (tf ; t0 , x0 , θ )). ∂tk Now, using the notation ∂y k ∂tk (7.89) = zkk as in Proposition 7.17, we obtain ∂2 l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) ∂tk2 ∂ = lx (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) y k (tf ; t0 , x0 , θ ) ∂tk f ∂ + lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) y k (tf ; t0 , x0 , θ ) ∂tk = lxf xf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))y k (tf ; t0 , x0 , θ ), y k (tf ; t0 , x0 , θ ) + lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))zkk (tf ; t0 , x0 , θ ), (7.90) 324 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem and thus, Ltk tk = = ∂2 l(t , x , t , x(t ; t , x , θ )) 0 0 f f 0 0 ∂tk2 ζ =ζ̂ lxf xf (p̂)y k (tˆf ), y k (tˆf ) + lxf (p̂)zkk (tˆf ). Let us rewrite the last term in this formula. The transversality condition lxf = ψ(tˆf ) implies lxf (p̂)zkk (tˆf ) = ψ(tˆf )zkk (tˆf ) = tˆf tˆk d (ψzkk ) dt + ψ(tˆk )zkk (tˆk ). dt (7.91) By formula (7.48), we have żkk = fx zkk + (y k )∗ fxx y k , t ≥ tˆk . Using this equation together with the adjoint equation −ψ̇ = ψfx , we get d ψj ((y k )∗ fj xx y k ) = (y k )∗ Hxx y k , (ψzkk ) = ψ̇zkk + ψ żkk = −ψfx zkk + ψfx zkk + dt j (7.92) and thus lxf (p̂)zkk (tˆf ) = tˆf tˆk (y k )∗ Hxx y k dt + ψ(tˆk )zkk (tˆk ). (7.93) We shall transform the last term in (7.93) using the relations (k H )(t) = H (t, x̂(t), ûk+ , ψ(t)) − H (t, x̂(t), ûk− , ψ(t)), d ˙ tˆk +) − ψ̇(tˆk +)[Hψ ]k D k (H ) = − (k H )t=t + = −[Ht ]k − [Hx ]k x̂( k dt (7.94) (see Section 5.2.2). Proposition 7.26. The following equality holds: ψ(tˆk )zkk (tˆk ) = D k (H ) − [Hx ]k [y k ]k . (7.95) Proof. Multiplying the initial condition (7.49) for zkk at the point t = tˆk by ψ(tˆk ), we get ˙ tˆk +) + y k (tˆk ) . ψ(tˆk )zkk (tˆk ) + ψ(tˆk )ẏ k (tˆk +) = −ψ(tˆk )[ft ]k − ψ(tˆk )[fx ]k x̂( (7.96) Here, we obviously have the relations ψ(tˆk )[ft ]k = [Ht ]k , ψ(tˆk )[fx ]k = [Hx ]k , and y k (tˆk ) = [y k ]k . Moreover, equation (7.26) for y k together with the adjoint equation −ψ̇ = ψfx implies that ψ ẏ k = ψfx y k = −ψ̇y k . Hence, in view of the initial condition (7.26) for y k , we find ψ(tˆk )ẏ k (tˆk +) = −ψ̇(tˆk +)y k (tˆk ) = ψ̇(tˆk +)[f ]k = ψ̇(tˆk +)[Hψ ]k . Thus, (7.96) and (7.94) imply (7.95). 7.4. Quadratic Form for the Induced Optimization Problem 325 From the relations (7.91), (7.93), and (7.95) and the equality y k (t) = 0 for t < tˆk , it follows that tˆf (y k t¯k )∗ Hxx y k t¯k dt Ltk tk t¯k2 = lxf xf (p̂)y k (tˆf )t¯k , y k (tˆf )t¯k + tˆ0 +D 7.4.4 k (H )t¯k2 − [Hx ]k [y k ]k t¯k2 , k = 1, . . . , s. (7.97) Derivative Ltk tj Note that Ltk tj = Ltj tk for all k, j . Therefore, s Ltk tj t¯k t¯j = s k,j =1 Ltk tk t¯k2 + 2 Ltk tj t¯k t¯j . (7.98) k<j k=1 Let us calculate Ltk tj for k < j . Differentiating (7.89) with respect to tj , we get ∂2 l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) ∂tk ∂tj ∂ = lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) y k (tf ; t0 , x0 , θ ) ∂tj ∂ k y (tf ; t0 , x0 , θ ) + lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ )) ∂tj = lxf xf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))y k (tf ; t0 , x0 , θ ), y j (tf ; t0 , x0 , θ ) + lxf (t0 , x0 , tf , x(tf ; t0 , x0 , θ ))zkj (tf ; t0 , x0 , θ ). (7.99) Thus, Ltk tj = ∂2 l(t0 , x0 , tf , x(tf ; t0 , x0 , θ ))|ζ =ζ̂ ∂tk ∂tj = lxf xf (p̂)y k (tˆf ), y j (tˆf ) + lxf (p̂)zkj (tˆf ). We can rewrite the last term in this formula as lxf (p̂)zkj (tˆf ) = ψ(tˆf )zkj (tˆf ) = tˆf (7.100) d (ψzkj ) dt + ψ(tˆj )zkj (tˆj ). dt tˆj By formula (7.53), żkk = fx zkk + (y k )∗ fxx y j for t ≥ tˆj . Similarly to (7.92), we get d kj k ∗ j dt (ψz ) = (y ) Hxx y and thus obtain lxf (p̂)z (tˆf ) = kj tˆf tˆj (y k )∗ Hxx y j dt + ψ(tˆj )zkj (tˆj ). Since y j (t) = 0 for t < tˆj , we have tˆf k ∗ j (y ) Hxx y dt = tˆj tˆf tˆ0 (y k )∗ Hxx y j dt. (7.101) (7.102) 326 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Using the initial condition (7.54) for zkj at the point θ̂ j , we get ψ(tˆj )zkj (tˆj ) = −ψ(tˆj )[fx ]j y k (tˆj ) = −[Hx ]j y k (tˆj ). (7.103) Formulas (7.100)–(7.103) imply the following representation for all k < j : Ltk tj t¯k t¯j = lxf xf (p̂)y k (tˆf )t¯k , y j (tˆf )t¯j tˆf (y k t¯k )∗ Hxx y j t¯j dt − [Hx ]j y k (tˆj )t¯k t¯j . + tˆ0 7.4.5 (7.104) Derivative Lx0 tf Using Proposition 7.11, we get ∂2 ∂ l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = {lx + lxf V }|t=tf ∂x0 ∂tf ∂tf 0 ∂lxf = (lx0 tf + lx0 xf ẋ + V + lxf V̇ )|t=tf ∂tf = (lx0 tf + lx0 xf ẋ + (lxf tf + lxf xf ẋ)V + lxf fx V )|t=tf . Again, we transform the last term in this formula at the point ζ = ζ̂ . Using the adjoint equation −ψ̇ = ψfx and the transversality condition ψ(tˆf ) = lxf , we get lxf fx V |t=tˆf = ψfx V |t=tˆf = −ψ̇(tˆf )V (tˆf ). Consequently, ˙ tˆf )t¯f , x̄0 + lx t V (tˆf )x̄0 t¯f Lx0 tf x̄0 t¯f = lx0 tf x̄0 t¯f + lx0 xf x̂( f f ˙ + lxf xf x̂(tˆf )t¯f , V (tˆf )x̄0 − ψ̇(tˆf )V (tˆf )x̄0 t¯f . 7.4.6 (7.105) Derivative Ltk tf Using the notation ∂x ∂tk = y k and Proposition 7.13, we get ∂ ∂2 l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = {lx y k }|t=tf ∂tk ∂tf ∂tf f = {(lxf xf ẋ + lxf tf )y k + lxf ẏ k }|t=tf = {lxf xf ẋy k + lxf tf y k + lxf fx y k }|t=tf . We evaluate the last term in this formula at the point ζ = ζ̂ using the adjoint equation −ψ̇ = ψfx and the transversality condition ψ(tˆf ) = lxf : lxf fx y k |t=tˆf = ψfx y k |t=tˆf = −ψ̇(tˆf )y k (tˆf ). Therefore, ˙ tˆf )t¯f , y k (tˆf )t¯k + lx t y k (tˆf )t¯k t¯f − ψ̇(tˆf )y k (tˆf )t¯k t¯f . Ltk tf t¯k t¯f = lxf xf x̂( f f (7.106) 7.4. Quadratic Form for the Induced Optimization Problem 7.4.7 327 Derivative Ltf tf We have ∂2 ∂ l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = {lt + lxf ẋ}|t=tf 2 ∂tf f ∂tf = {(ltf tf + ltf xf ẋ) + (lxf tf + lxf xf ẋ)ẋ + lxf ẍ} , t=tf which gives ˙ tˆf ) + lx x x̂( ˙ tˆf ), x̂( ˙ tˆf ) + ψ(tˆf )x̂( ¨ tˆf ). Ltf tf = ltf tf + 2ltf xf x̂( f f (7.107) Let us transform the last term. Equation (6.11) in the definition of M0 is equivalent to the relation ψ x̂˙ + ψ0 = 0. Differentiating this equation with respect to t, we get ψ̇ x̂˙ + ψ x̂¨ + ψ̇0 = 0. (7.108) Hence, formula (7.107) implies the equality Ltf tf t¯f2 7.4.8 = ˙ tˆf )t¯2 + lx x x̂( ˙ tˆf )t¯f , x̂( ˙ tˆf )t¯f − (ψ̇(tˆf )x̂( ˙ tˆf ) + ψ̇0 (tˆf ))t¯2 . ltf tf t¯f2 + 2ltf xf x̂( f f f f Derivative Lx0 t0 In view of the relation ∂x ∂x0 = V , we obtain ∂ ∂2 l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = {lx + lxf V }|t=tf ∂x0 ∂t0 ∂t0 0 ∂x ∂x ∂V V + lxf = lx0 t0 + lx0 xf + lxf t0 + lxf xf . ∂t0 ∂t0 ∂t0 t=tf Now, using the transversality condition lxf = ψ(tˆf ), formula (7.24), and the notation ∂V ∂t0 = S, we get ˙ tˆ0 ) + lx t V (tˆf ) − x̂( ˙ tˆ0 )∗ V (tˆf )∗ lx x V (tˆf ) + ψ(tˆf )S(tˆf ). Lx0 t0 = lx0 t0 − lx0 xf V (tˆf )x̂( f 0 f f The transformation of the last term in this formula proceeds as follows. Using the adjoint equation for ψ and the system (7.69) for S, we obtain the equation d (ψS) = dt = ˙ tˆ0 )∗ V ∗ ψ̇S + ψ Ṡ = −ψfx S + ψfx S − x̂( ˙ tˆ0 )∗ V ∗ Hxx V , −x̂( which yields ψ(tˆf )S(tˆf ) = − tˆf tˆ0 ψi fixx V i ˙ tˆ0 )∗ V ∗ Hxx V dt + ψ(tˆ0 )S(tˆ0 ). x̂( (7.109) 328 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Using now the initial condition (7.70) for S at the point t = tˆ0 and the equation V̇ = fx V , we get (7.110) (ψS)|tˆ0 = −(ψ V̇ )|tˆ0 = −(ψfx V )|tˆ0 = (ψ̇V )|tˆ0 = ψ̇(tˆ0 ), since V (tˆ0 ) = E. Formulas (7.109) and (7.110) then imply the equality ˙ tˆ0 ) + lx t V (tˆf ) − x̂( ˙ tˆ0 )∗ V (tˆf )∗ lx x V (tˆf ) Lx0 t0 = lx0 t0 − lx0 xf V (tˆf )x̂( f 0 f f tˆf ˙ tˆ0 )∗ V ∗ Hxx V dt + ψ̇(tˆ0 ). x̂( − (7.111) ˙ tˆ0 )t¯0 , x̄0 + lx t V (tˆf )x̄0 t¯0 Lx0 t0 x̄0 t¯0 = lx0 t0 x̄0 t¯0 − lx0 xf V (tˆf )x̂( f 0 ˙ −lxf xf V (tˆf )x̄0 , V (tˆf )x̂(tˆ0 )t¯0 + ψ̇(tˆ0 )x̄0 t¯0 tˆf ˙ tˆ0 )t¯0 dt. − Hxx V x̄0 , V x̂( (7.112) tˆ0 Therefore, tˆ0 7.4.9 Derivative Ltk t0 Using the notation ∂x ∂t0 = w, ∂x ∂tk = y k , and ∂yk ∂t0 = r k , we obtain ∂2 ∂ l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = {lx y k }|t=tf ∂tk ∂t0 ∂t0 f ∂x ∂y k k k ∗ = lxf t0 y + (y ) lxf xf + lxf ∂t0 ∂t0 t=tf = {lxf t0 y k + (y k )∗ lxf xf w + lxf r k }|t=tf . ˙ tˆ0 ). Using this condition together According to condition (7.24) we have w|t=tˆf = −V (tˆf )x̂( with the transversality condition lxf = ψ(tˆf ), we find ˙ tˆ0 ) + ψ(tˆf )r k (tˆf ). Ltk t0 = lxf t0 y k (tˆf ) − (y k (tˆf ))∗ lxf xf V (tˆf )x̂( (7.113) Let us transform the last term in this formula. Using the adjoint equation for ψ and the system (7.73) for r k , we get for t ≥ tˆk : d ˙ tˆ0 ) (ψr k ) = ψ̇r k + ψ ṙ k = −ψfx r k + ψfx r k − (yi )∗ ψj fj xx V x̂( dt j ˙ tˆ0 ). = −(yk ) Hxx V x̂( ∗ It follows that ψ(tˆf )r k (tˆf ) = − tˆf tˆk ˙ tˆ0 ) dt + ψ(tˆk )r k (tˆk ). (yk )∗ Hxx V x̂( (7.114) 7.4. Quadratic Form for the Induced Optimization Problem 329 The initial condition (7.74) for r k at the point tˆk then yields ˙ tˆ0 ) = [Hx ]k V (tˆk )x̂( ˙ tˆ0 ). ψ(tˆk )r k (tˆk ) = ψ(tˆk )[fx ]k V (tˆk )x̂( (7.115) Formulas (7.113)–(7.115) and the condition y k (t) = 0 for t < tˆk then imply the equality ˙ tˆ0 ) Ltk t0 = lxf t0 y k (tˆf ) − (y k (tˆf ))∗ lxf xf V (tˆf )x̂( tˆf ˙ tˆ0 ) − ˙ tˆ0 ) dt. (yk )∗ Hxx V x̂( + [Hx ]k V (tˆk )x̂( (7.116) ˙ tˆ0 )t¯0 Ltk t0 t¯k t¯0 = lxf t0 y k (tˆf )t¯k t¯0 − (y k (tˆf )t¯k )∗ lxf xf V (tˆf )x̂( tˆf ˙ tˆ0 )t¯0 t¯k − ˙ tˆ0 )t¯0 dt. + [Hx ]k V (tˆk )x̂( (y k t¯k )∗ Hxx V x̂( (7.117) tˆ0 Hence, tˆ0 7.4.10 Derivative Ltf t0 We have ∂2 ∂ l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = {lt + lxf ẋ}|t=tf ∂tf ∂t0 ∂t0 f ∂x ∂ ∂ = ltf t0 + ltf xf + lx ẋ + lxf ẋ . ∂t0 ∂t0 f ∂t0 t=tf Using the equalities ∂x ∂ lx = lxf t0 + lxf xf , ∂t0 f ∂t0 ∂x = −V ẋ(t0 ), ∂t0 we get ∂2 l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = ltf t0 − ltf xf V ẋ(t0 ) + lxf t0 ẋ(tf ) ∂tf ∂t0 − (V (tf )ẋ(t0 ))∗ lxf xf ẋ(tf ) + lxf ∂ ẋ|t=tf . ∂t0 Let us calculate the last term. Differentiating the equation ẋ(t; t0 , x0 , θ ) = f (t, x(t; t0 , x0 , θ ), u(t; θ )) with respect to t0 , we get ∂ ∂ ẋ = fx x = −fx V ẋ(t0 ). ∂t0 ∂t0 Consequently, at the point ζ = ζ̂ , we obtain ∂ ∂ ˙ tˆ0 )}| ˆ = ψ̇(tˆf )V (tˆf )x̂( ˙ tˆ0 ). ẋ = ψ ẋ = {−ψfx V x̂( lxf t=tf ∂t0 t=tˆf ∂t0 t=tˆf (7.118) 330 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Using this equality in (7.118), we get at the point ζ = ζ̂ , ˙ tˆf ) ˙ tˆ0 ) + lx t x̂( Ltf t0 = ltf t0 − ltf xf V (tˆf )x̂( f 0 ˙ tˆ0 ), x̂( ˙ tˆf ) + ψ̇(tˆf )V (tˆf )x̂( ˙ tˆ0 ), − lx x V (tˆf )x̂( f f (7.119) which yields ˙ tˆ0 )t¯0 )t¯f + lx t (x̂( ˙ tˆf )t¯f )t¯0 Ltf t0 t¯f t¯0 = ltf t0 t¯f t¯0 − ltf xf (V (tˆf )x̂( f 0 ˙ tˆ0 )t¯0 , x̂( ˙ tˆf )t¯f + ψ̇(tˆf )(V (tˆf )x̂( ˙ tˆ0 )t¯0 )t¯f . − lx x V (tˆf )x̂( f f 7.4.11 (7.120) Derivative Lt0 t0 We have ∂ ∂2 ∂x lt + lxf l(t0 , x0 , tf , x(tf ; t0 , x0 , θ )) = ∂t0 0 ∂t0 t=tf ∂t02 ∂x ∂x ∂x ∂ 2 x = lt0 t0 + lt0 xf + lxf t0 + lxf xf + lxf 2 ∂t0 ∂t0 ∂t0 ∂t0 t=tf ( ' 2 ∂x ∂x ∂x ∂ x + lxf 2 = lt0 t0 + 2lt0 xf + lxf xf , ∂t0 ∂t0 ∂t0 ∂t0 t=tf = lt0 t0 + 2lt0 xf w + lxf xf w, w + lxf q |t=tf , where w= ∂x , ∂t0 (7.121) ∂w ∂ 2 x = 2. ∂t0 ∂t0 q= The transversality condition lxf = ψ(tˆf ) yields Lt0 t0 = lt0 t0 + 2lt0 xf w(tˆf ) + lxf xf w(tˆf ), w(tˆf ) + ψ(tˆf )q(tˆf ). (7.122) Let us transform the last term using the adjoint equation for ψ and the system (7.58) for q: d (ψq) = ψ̇q + ψ q̇ = −ψfx q + ψfx q + ψj (w ∗fj xx w) = w ∗ Hxx w. dt j ˙ tˆ0 ), we obtain Also, using the equality w = −V x̂( ψ(tˆf )q(tˆf ) = ψ(tˆ0 )q(tˆ0 ) + = ψ(tˆ0 )q(tˆ0 ) + tˆf tˆ0 tˆf tˆ0 w∗ Hxx w dt ˙ tˆ0 ), V x̂( ˙ tˆ0 ) dt. Hxx V x̂( (7.123) 7.4. Quadratic Form for the Induced Optimization Problem 331 The initial condition (7.59) for q then implies ¨ tˆ0 ) − 2ψ(tˆ0 )ẇ(tˆ0 ). ψ(tˆ0 )q(tˆ0 ) = −ψ(tˆ0 )x̂( (7.124) From the equation ẇ = fx w (see Proposition 7.12), the adjoint equation −ψ̇ = ψfx , and ˙ tˆ0 ), it follows that the formula w = −V x̂( ˙ tˆ0 ). −ψ ẇ = −ψfx w = ψ̇w = −ψ̇V x̂( Since V (tˆ0 ) = E, we obtain ˙ tˆ0 ). ψ(tˆ0 )ẇ(tˆ0 ) = ψ̇(tˆ0 )x̂( (7.125) Moreover, by formula (7.108) we have −ψ x̂¨ = ψ̇ x̂˙ + ψ̇0 . (7.126) ˙ tˆ0 ). ψ(tˆ0 )q(tˆ0 ) = ψ̇0 (tˆ0 ) − ψ̇(tˆ0 )x̂( (7.127) Formulas (7.124)–(7.126) imply Combining formulas (7.122), (7.123), and (7.127), we obtain ˙ tˆ0 ) + lx x V (tˆf )x̂( ˙ tˆ0 ), V (tˆf )x̂( ˙ tˆ0 ) Lt0 t0 = lt0 t0 − 2lt0 xf V (tˆf )x̂( f f tˆf ˙ tˆ0 ) + ˙ tˆ0 ), V (tˆf )x̂( ˙ tˆ0 ) dt. Hxx V (tˆf )x̂( + ψ̇0 (tˆ0 ) − ψ̇(tˆ0 )x̂( tˆ0 (7.128) Thus we have found the representation ˙ tˆ0 )t¯2 + lx x V (tˆf )x̂( ˙ tˆ0 )t¯0 , V (tˆf )x̂( ˙ tˆ0 )t¯0 Lt0 t0 t¯02 = lt0 t0 t¯02 − 2lt0 xf V (tˆf )x̂( f f 0 tˆf ˙ tˆ0 )t¯2 + ˙ tˆ0 )t¯0 , V (tˆf )x̂( ˙ tˆ0 )t¯0 dt. Hxx V (tˆf )x̂( + ψ̇0 (tˆ0 )t¯02 − ψ̇(tˆ0 )x̂( 0 tˆ0 7.4.12 (7.129) Representation of the Quadratic Form Lζ ζ ζ̄ , ζ̄ Combining all results and formulas in the preceding sections, we have proved the following theorem. Theorem 7.27. Let the Lagrange multipliers μ = (α0 , α, β) ∈ 0 and λ = (α0 , α, β, ψ, ψ0 ) ∈ correspond to each other, i.e., let π0 λ = μ hold; see Proposition 7.4. Then, for any ζ̄ = (t¯0 , t¯f , x̄0 , θ̄) ∈ R2+n+s , formulas (7.75), (7.85), (7.89), (7.97), (7.98), (7.104)–(7.106), (7.109), (7.112), (7.117), (7.120), and (7.129) hold, where the matrix V (t) is the solution to the IVP (7.18) and the function y k is the solution to the IVP (7.26) for each k = 1, . . . , s. 332 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Thus we have obtained the following explicit and massive representation of the quadratic form in the IOP: Lζ ζ ζ̄ , ζ̄ = Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ (7.130) = Lx0 x0 x̄0 , x̄0 + 2 + 2Lx0 tf x̄0 t¯f + 2 s k=1 s Lx0 tk x̄0 t¯k + s Ltk tk t¯k2 + 2 s Ltk tj t¯k t¯j k<j k=1 Ltk tf t¯k t¯f + Ltf tf t¯f2 k=1 + 2Lx0 t0 x̄0 t¯0 + 2 s Lt0 tk t¯0 t¯k + 2Lt0 tf t¯0 t¯f + Lt0 t0 t¯02 k=1 = x̄0∗ lx0 x0 x̄0 + 2x̄0∗ lx0 xf V (tˆf )x̄0 + (V (tˆf )x̄0 )∗ lxf xf V (tˆf )x̄0 tˆf (V x̄0 )∗ Hxx V x̄0 dt + tˆ0 s s + 2x̄0∗ lx0 xf y k (tˆf )t¯k + 2(V (tˆf )x̄0 )∗ lxf xf y k (tˆf )t¯k k=1 k=1 s s tˆf − 2[Hx ]k V (tˆk )x̄0 t¯k + 2Hxx y k t¯k , V x̄0 dt ˆ0 t k=1 k=1 + s lxf xf y k (tˆf )t¯k , y k (tˆf )t¯k + s k=1 + s D k (H )t¯k2 − k=1 + ˆ k=1 t0 ˆ k<j t0 (y k t¯k )∗ Hxx y k t¯k dt s [Hx ]k [y k ]k t¯k2 + 2lxf xf y k (tˆf )t¯k , y j (tˆf )t¯j k<j k=1 tˆf tˆf 2(y k t¯k )∗ Hxx y j t¯j dt − 2[Hx ]j y k (tˆj )t¯k t¯j k<j ˙ tˆf )t¯f , x̄0 + 2lx t V (tˆf )x̄0 t¯f + 2lx0 tf x̄0 t¯f + 2lx0 xf x̂( f f ˙ tˆf )t¯f , V (tˆf )x̄0 − 2ψ̇(tˆf )V (tˆf )x̄0 t¯f + 2lxf xf x̂( + s ˙ tˆf )t¯f , y k (tˆf )t¯k + 2lxf xf x̂( k=1 − s s 2lxf tf y k (tˆf )t¯k t¯f k=1 2ψ̇(tˆf )y k (tˆf )t¯k t¯f k=1 ˙ tˆf )t¯2 + lx x x̂( ˙ tˆf )t¯f , x̂( ˙ tˆf )t¯f + ltf tf t¯f2 + 2ltf xf x̂( f f f ˙ tˆf ) + ψ̇0 (tˆf ))t¯2 − (ψ̇(tˆf )x̂( f ˙ tˆ0 )t¯0 , x̄0 + 2lx t V (tˆf )x̄0 t¯0 + 2lx0 t0 x̄0 t¯0 − 2lx0 xf V (tˆf )x̂( f 0 7.5. Equivalence of the Quadratic Forms 333 ˙ tˆ0 )t¯0 + 2ψ̇(tˆ0 )x̄0 t¯0 − 2lxf xf V (tˆf )x̄0 , V (tˆf )x̂( tˆf ˙ tˆ0 )t¯0 dt − 2Hxx V x̄0 , V x̂( tˆ0 + s 2lxf t0 y k (tˆf )t¯k t¯0 − k=1 + s s ˙ tˆ0 )t¯0 2(y k (tˆf )t¯k )∗ lxf xf V (tˆf )x̂( k=1 ˙ tˆ0 )t¯0 t¯k − 2[Hx ] V (tˆk )x̂( k k=1 s tˆf ˆ k=1 t0 ˙ tˆ0 )t¯0 dt 2(y k t¯k )∗ Hxx V x̂( ˙ tˆ0 )t¯0 )t¯f + 2lx t (x̂( ˙ tˆf )t¯f )t¯0 + 2ltf t0 t¯f t¯0 − 2ltf xf (V (tˆf )x̂( f 0 ˙ tˆ0 )t¯0 , x̂( ˙ tˆf )t¯f + 2ψ̇(tˆf )(V (tˆf )x̂( ˙ tˆ0 )t¯0 )t¯f − 2lxf xf V (tˆf )x̂( ˙ tˆ0 )t¯2 + lx x V (tˆf )x̂( ˙ tˆ0 )t¯0 , V (tˆf )x̂( ˙ tˆ0 )t¯0 + lt0 t0 t¯02 − 2lt0 xf V (tˆf )x̂( f f 0 tˆf ˙ tˆ0 )t¯2 + ˙ tˆ0 )t¯0 , V x̂( ˙ tˆ0 )t¯0 dt. Hxx V x̂( + ψ̇0 (tˆ0 )t¯02 − ψ̇(tˆ0 )x̂( 0 tˆ0 Again, we wish to emphasize that this explicit representation involves only first-order variations y k and V of the trajectories x(t; t0 , x0 , θ ). 7.5 Equivalence of the Quadratic Forms in the Basic and Induced Optimization Problem In this section, we shall prove Theorem 7.8 which is the main result of this chapter. Let the Lagrange multipliers μ = (α0 , α, β) ∈ 0 and λ = (α0 , α, β, ψ, ψ0 ) ∈ correspond to each other, and take any ζ̄ = (t¯0 , t¯f , x̄0 , θ̄ ) ∈ R2+n+s . Consider the representation (7.130) of the quadratic form Lζ ζ ζ̄ , ζ̄ , which is far from revealing the equivalence of the quadratic forms for the basic control problem and the IOP. However, we show now that by a careful regrouping of the terms in (7.130) we shall arrive at the desired equivalence. The quadratic form (7.130) contains terms of the following types. Type (a): Positive terms with coefficients D k (H ) multiplied by the variation of the switching time t¯k squared, s D k (H )t¯k2 . (7.131) a := k=1 Type (b): Mixed terms with [Hx b := − s 2[Hx ]k V (tˆk )x̄0 t¯k − k=1 ]k connected with the variation t¯k , s [Hx ]k [y k ]k t¯k2 k=1 − k<j 2[Hx ]j y k (tˆj )t¯k t¯j + s k=1 ˙ tˆ0 )t¯0 t¯k . (7.132) 2[Hx ]k V (tˆk )x̂( 334 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Since [Hx ]j y k (tˆj )t¯k t¯j = k<j s k−1 [Hx ]k y j (tˆk )t¯k t¯j = [Hx ]k y j (tˆk )t¯k t¯j , k=1 j =1 j <k we get from (7.132), b=− s k=1 k−1 1 ˙ tˆ0 )t¯0 t¯k . 2[Hx ]k V (tˆk )x̄0 + [y k ]k t¯k + y j (tˆk )t¯j − V (tˆk )x̂( 2 (7.133) j =1 According to (7.17) put x̄(t) = V (t)x̄0 + s ˙ tˆ0 )t¯0 . y k (t)t¯k − V (t)x̂( (7.134) k=1 Then we have x̄(tˆk −) = V (tˆk )x̄0 + k−1 ˙ tˆ0 )t¯0 , y j (tˆk )t¯j − V (tˆk )x̂( j =1 since y j (tˆk −) = y j (tˆk ) = 0 for j > k and y k (tˆk −) = 0. Moreover, the jump of x̄(t) at the point tˆk is equal to the jump of y k (t)t¯k at the same point, i.e., [x̄]k = [y k ]k t¯k . Therefore, 1 ˙ tˆ0 )t¯0 y j (tˆk )t¯j − V (tˆk )x̂( V (tˆk )x̄0 + [y k ]k t¯k + 2 k−1 j =1 1 1 k = x̄(tˆk −) + [x̄]k = (x̄(tˆk −) + x̄(tˆk +)) = x̄av . 2 2 Thus, we get b=− s k ¯ tk . 2[Hx ]k x̄av (7.135) k=1 Type (c): Integral terms c := tˆf ∗ (V x̄0 ) Hxx V x̄0 dt + tˆ0 + s ˆ k=1 t0 − tˆf tˆ0 + tˆf tˆf tˆ0 s tˆf ˆ k=1 t0 (y k t¯k )∗ Hxx y k t¯k dt + 2Hxx y k t¯k , V x̄0 dt tˆf 2(y k t¯k )∗ Hxx y j t¯j dt ˆ k<j t0 s tˆf ˙ tˆ0 )t¯0 dt − 2Hxx V x̄0 , V x̂( ˆ k=1 t0 ˙ tˆ0 )t¯0 , V x̂( ˙ tˆ0 )t¯0 dt. Hxx V x̂( ˙ tˆ0 )t¯0 dt 2(y k t¯k )∗ Hxx V x̂( (7.136) 7.5. Equivalence of the Quadratic Forms 335 Obviously, this sum can be transformed into a perfect square: ( tˆf ' s s ˙ tˆ0 )t¯0 ), V x̄0 + ˙ tˆ0 )t¯0 dt c= Hxx (V x̄0 + y k t¯k − V x̂( y k t¯k − V x̂( tˆ0 = k=1 tˆf tˆ0 k=1 Hxx x̄, x̄ dt. (7.137) Type (d): Endpoints terms. We shall divide them into several groups. Group (d1): This group contains the terms with second-order derivatives of the endpoint Lagrangian l with respect to t0 , x0 , and tf : d1 := x̄0∗ lx0 x0 x̄0 + 2lx0 tf x̄0 t¯f + ltf tf t¯f2 + 2lx0 t0 x̄0 t¯0 + 2ltf t0 t¯f t¯0 + lt0 t0 t¯02 . (7.138) Group (d2): We collect the terms with lt0 xf : d2 := 2lxf t0 V (tˆf )x̄0 t¯0 + s 2lxf t0 y k (tˆf )t¯k t¯0 k=1 ˙ tˆf )t¯f t¯0 − 2lt x V (tˆf )x̂( ˙ tˆ0 )t¯2 + 2lxf t0 x̂( 0 f 0 s k ˙ tˆf )t¯f − V (tˆf )x̂( ˙ tˆ0 )t¯0 t¯0 , = 2lxf t0 V (tˆf )x̄0 + y (tˆf )t¯k + x̂( k=1 = 2lxf t0 x̄¯f t¯0 , (7.139) where, in view of (7.34), x̄¯f := V (tˆf )x̄0 + s ˙ tˆf )t¯f − V (tˆf )x̂( ˙ tˆ0 )t¯0 = x̄(tˆf ) + x̂( ˙ tˆf )t¯f . y k (tˆf )t¯k + x̂( (7.140) k=1 Group (d3): Consider the terms with lx0 xf : d3 := 2x̄0∗ lx0 xf V (tˆf )x̄0 + s 2x̄0∗ lx0 xf y k (tˆf )t¯k k=1 ˙ tˆf )t¯f , x̄0 − 2lx x V (tˆf )x̂( ˙ tˆ0 )t¯0 , x̄0 + 2lx0 xf x̂( 0 f ' ( s ˙ tˆf )t¯f − V (tˆf )x̂( ˙ tˆ0 )t¯0 , x̄0 y k (tˆf )t¯k + x̂( = 2 lx0 xf V (tˆf )x̄0 + k=1 = 2lx0 xf x̄¯f , x̄0 . (7.141) Group (d4): This group contains all terms with ltf xf : d4 := 2lxf tf V (tˆf )x̄0 t¯f + s 2lxf tf y k (tˆf )t¯k t¯f k=1 ˙ tˆ0 )t¯0 )t¯f ˙ tˆf )t¯2 − 2lt x (V (tˆf )x̂( + 2ltf xf x̂( f f f s k ˙ ˙ = 2lxf tf V (tˆf )x̄0 + y (tˆf )t¯k + x̂(tˆf )t¯f − V (tˆf )x̂(tˆ0 )t¯0 t¯f k=1 = 2lxf tf x̄¯f t¯f . (7.142) 336 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Group (d5): We collect all terms containing lxf xf : d5 := (V (tˆf )x̄0 )∗ lxf xf V (tˆf )x̄0 + s 2(V (tˆf )x̄0 )∗ lxf xf y k (tˆf )t¯k k=1 + s lxf xf y k (tˆf )t¯k , y k (tˆf )t¯k + 2lxf xf y k (tˆf )t¯k , y j (tˆf )t¯j k<j k=1 ˙ tˆf )t¯f , V (tˆf )x̄0 + + 2lxf xf x̂( s ˙ tˆf )t¯f , y k (tˆf )t¯k 2lxf xf x̂( k=1 ˙ tˆf )t¯f − 2lx x V (tˆf )x̄0 , V (tˆf )x̂( ˙ tˆ0 )t¯0 ˙ tˆf )t¯f , x̂( + lxf xf x̂( f f − s ˙ tˆ0 )t¯0 − 2lx x V (tˆf )x̂( ˙ tˆ0 )t¯0 , x̂( ˙ tˆf )t¯f 2(y k (tˆf )t¯k )∗ lxf xf V (tˆf )x̂( f f k=1 ˙ tˆ0 )t¯0 , V (tˆf )x̂( ˙ tˆ0 )t¯0 . + lxf xf V (tˆf )x̂( (7.143) One can easily check that this sum can be transformed into the perfect square ' s ˙ tˆf )t¯f − V (tˆf )x̂( ˙ tˆ0 )t¯0 , y k (tˆf )t¯k + x̂( d5 := lxf xf V (tˆf )x̄0 + k=1 V (tˆf )x̄0 + s ˙ tˆf )t¯f − V (tˆf )x̂( ˙ tˆ0 )t¯0 y (tˆf )t¯k + x̂( ( k k=1 = lxf xf x̄¯f , x̄¯f . (7.144) Group (d6): Terms with ψ̇(tˆ0 ) and ψ̇0 (tˆ0 ): ˙ tˆ0 ))t¯2 . d6 := 2ψ̇(tˆ0 )x̄0 t¯0 + (ψ̇0 (tˆ0 ) − ψ̇(tˆ0 )x̂( 0 (7.145) Group (d7): Terms with ψ̇(tˆf ) and ψ̇0 (tˆf ): d7 := −2ψ̇(tˆf )V (tˆf )x̄0 t¯f − s 2ψ̇(tˆf )y k (tˆf )t¯k t¯f k=1 ˙ tˆf ) + ψ̇0 (tˆf ))t¯2 + 2ψ̇(tˆf )(V (tˆf )x̂( ˙ tˆ0 )t¯0 )t¯f − (ψ̇(tˆf )x̂( f s ˙ tˆ0 )t¯0 ) t¯f = −2ψ̇(tˆf ) V (tˆf )x̄0 + y k (tˆf )t¯k − V (tˆf )x̂( k=1 ˙ tˆf ) + ψ̇0 (tˆf ))t¯2 − (ψ̇(tˆf )x̂( f ˙ tˆf ) + ψ̇0 (tˆf ))t¯2 . = −2ψ̇(tˆf )x̄(tˆf )t¯f − (ψ̇(tˆf )x̂( f (7.146) ˙ tˆf )t¯f in (7.146), we obtain Using the equality x̄¯f = x̄(tˆf ) + x̂( ˙ tˆf ))t¯2 . d7 = −2ψ̇(tˆf )x̄¯f t¯f − (ψ̇0 (tˆf ) − ψ̇(tˆf )x̂( f (7.147) 7.5. Equivalence of the Quadratic Forms 337 This completes the whole list of all terms in the quadratic form associated with the IOP. Hence, we have Lζ ζ ζ̄ , ζ̄ = a + b + c + d, d= 7 di . i=1 We thus have found the following representation of this quadratic form; see formulas (7.131) for a, (7.135) for b, and (7.137) for c: Lζ ζ ζ̄ , ζ̄ = s k=1 D (H )t¯k2 − k s k ¯ tk + 2[Hx ]k x̄av k=1 tˆf tˆ0 Hxx x̄, x̄ + d, (7.148) where according to formulas (7.138), (7.139), (7.141), (7.142), (7.144), (7.145), (7.147) for d1 , . . . , d7 , respectively, d = lx0 x0 x̄0 , x̄0 + 2lx0 tf x̄0 t¯f + ltf tf t¯f2 + 2lx0 t0 x̄0 t¯0 + 2ltf t0 t¯f t¯0 + lt0 t0 t¯02 + 2lxf t0 x̄¯f t¯0 ˙ tˆ0 ))t¯2 + 2lx0 xf x̄¯f , x̄0 + 2lxf tf x̄¯f t¯f + lxf xf x̄¯f , x̄¯f + 2ψ̇(tˆ0 )x̄0 t¯0 + (ψ̇0 (tˆ0 ) − ψ̇(tˆ0 )x̂( 0 ˙ tˆf ))t¯2 . − 2ψ̇(tˆf )x̄¯f t¯f − (ψ̇0 (tˆf ) − ψ̇(tˆf )x̂( f (7.149) In (7.148) and (7.149) the function x̄(t) and the vector x̄¯f are defined by (7.134) and (7.140), respectively. Note that in (7.149), lx0 x0 x̄0 , x̄0 + 2lx0 tf x̄0 t¯f + ltf tf t¯f2 + 2lx0 t0 x̄0 t¯0 + 2ltf t0 t¯f t¯0 + lt0 t0 t¯02 + 2lxf t0 x̄¯f t¯0 ¯ p̄¯ , + 2lx0 xf x̄¯f , x̄0 + 2lxf tf x̄¯f t¯f + lxf xf x̄¯f , x̄¯f = lpp p̄, where, by definition, (7.150) p̄¯ = (t¯0 , x̄0 , t¯f , x̄¯f ). (7.151) ˙ tˆ0 ))t¯2 ¯ p̄¯ + 2ψ̇(tˆ0 )x̄0 t¯0 + (ψ̇0 (tˆ0 ) − ψ̇(tˆ0 )x̂( d = lpp p̄, 0 2 ˙ ¯ − 2ψ̇(tˆf )x̄1 t¯f − (ψ̇0 (tˆf ) − ψ̇(tˆf )x̂(tˆf ))t¯f . (7.152) Finally, we get Thus, we have proved the following result. Theorem 7.28. Let the Lagrange multipliers μ = (α0 , α, β) ∈ 0 and λ = (α0 , α, β, ψ, ψ0 ) ∈ correspond to each other, i.e., let π0 λ = μ hold. Then for any ζ̄ = (t¯0 , t¯f , x̄0 , θ̄ ) ∈ R2+n+s the quadratic form Lζ ζ ζ̄ , ζ̄ has the representation (7.148)–(7.152), where the vector function x̄(t) and the vector x̄¯f are defined by (7.134) and (7.140). The matrix-valued function V (t) is the solution to the IVP (7.18), and, for each k = 1, . . . , s, the vector function y k is the solution to the IVP (7.26). 338 Chapter 7. Bang-Bang Control Problem and Its Induced Optimization Problem Finally, we have arrived at the main result of this section. Theorem 7.29. Let λ = (α0 , α, β, ψ, ψ0 ) ∈ and ζ̄ = (t¯0 , t¯f , x̄0 , θ̄ ) ∈ R2+n+s . Put μ = (α0 , α, β), i.e., let π0 λ = μ ∈ 0 hold; see Proposition 7.4. Define the function x̄(t) by formula (7.134). Put ξ̄ = −θ̄ and z̄ = (t¯0 , t¯f , ξ̄ , x̄), which means π1 z̄ = ζ̄ ; see Propositions 7.6 and 7.7. Then the following equality holds: Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ = (λ, z̄), (7.153) where (λ, z̄) is defined by formulas (6.34) and (6.35). Proof. By Theorem 7.28, the equalities (7.148)–(7.152) hold. In view of equation (6.21), put ˙ tˆ0 ) = V (tˆ0 )x̄0 + x̄¯0 = x̄(tˆ0 ) + t¯0 x̂( s ˙ tˆ0 ). ˙ tˆ0 )t¯0 + t¯0 x̂( y k (tˆ0 )t¯k − V (tˆ0 )x̂( k=1 Since y k (tˆ0 ) = 0 for k = 1, . . . , s and V (tˆ0 ) = E, it follows that x̄¯0 = x̄0 . Consequently, the ¯ which was defined by (6.21) as (t¯0 , x̄¯0 , t¯f , x̄¯f ), coincides with the vector p̄, ¯ defined vector p̄, in this section by (7.151). Hence, the endpoint quadratic form d in (7.152) and the endpoint ¯ p̄¯ in (6.35) take equal values, d = Ap̄, ¯ p̄¯ . Moreover, the integral quadratic form Ap̄, tˆ terms tˆf Hxx x̄, x̄ dt in the representation (7.148) of the form Lζ ζ ζ̄ , ζ̄ and those in the 0 representation (6.34) of the form coincide, and s k=1 k D k (H )ξ̄k2 + 2[Hx ]k x̄av ξ̄k = s k=1 D k (H )t¯k2 − s k ¯ tk , 2[Hx ]k x̄av k=1 because ξ̄k = −t¯k , k = 1, . . . , s. Thus, the representation (7.148) of the form Lζ ζ ζ̄ , ζ̄ implies the equality (7.153) of both forms. Theorem 7.8, which is the main result of this chapter, then follows from Theorem 7.29. Remark. Theorems 7.8 and 7.29 pave the way for the sensitivity analysis of parametric bang-bang control problems. Since the IOP is finite-dimensional, the SSC imply that we may take advantage of the well-known sensitivity results by Fiacco [32] to obtain solution differentiability for the IOP. The strict bang-bang property then ensures solution differentiability for the parametric bang-bang control problems; cf. Kim and Maurer [48] and Felgenhauer [31]. Chapter 8 Numerical Methods for Solving the Induced Optimization Problem and Applications 8.1 The Arc-Parametrization Method 8.1.1 Brief Survey on Numerical Methods for Optimal Control Problems It has become customary to divide numerical methods for solving optimal control problems into two main classes: indirect methods and direct methods. Indirect methods take into account the full set of necessary optimality conditions of the Minimum Principle which gives rise to a multipoint boundary value problem (MPBVP) for state and adjoint variables. Shooting methods provide a powerful approach to solving an MPBVP; cf. articles from the German School [18, 19, 61, 64, 65, 81, 101, 102, 106] and French School [9, 10, 60]. E.g., the code BNDSCO by Oberle and Grimm [82] represents a very efficient implementation of shooting methods. Indirect methods produce highly accurate solutions but suffer from the drawback that the control structure, i.e., the sequence and number of bang-bang and singular arcs or constrained arcs, must be known a priori. Moreover, shooting methods need rather accurate initial estimates of the adjoint variables which in general are tedious to obtain. In direct methods, the optimal control problem is transcribed into a nonlinear programming problem (NLP) by suitable discretization techniques; cf. Betts [5], Bock and Plitt [7], and Bueskens [13, 14]. Direct methods dispense with adjoint variables which are computed a posteriori from Lagrange multipliers of the NLP. These methods have shown to be very robust and are mostly capable of determining the correct control structure without assuming an a priori knowledge of the structure. In all examples and applications in this chapter, optimal bang-bang and singular controls were computed in two steps. First, direct methods are applied to determine the correct optimal structure and obtain good estimates for the switching times. In a second step, the Induced Optimization Problem (IOP) in (7.4) of Section 7.1.1 is solved by using the arcparametrization method developed in [44, 45, 66]. To prove optimality of the computed extremal, we then verify that the second-order sufficient conditions (SSC) in Theorem 7.10 hold. The next section presents the arc-parametrization method by which the arc-lengths of bang-bang arcs are optimized instead of the switching times. Section 8.1.3 describes 339 340 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem an extension of the arc-parametrization method to control problems, where the control is piecewise defined by feedback functions. Numerical examples for bang-bang controls are presented in Sections 8.2–8.6, while Section 8.7 exhibits a bang-singular control for a van der Pol oscillator with a regulator functional. 8.1.2 The Arc-Parametrization Method for Solving the Induced Optimization Problem This section is based on the article of Maurer et al. [66] but uses a slightly different terminology. We consider the optimal control problem (6.1)–(6.3). To simplify the exposition we assume that the initial time t0 is fixed and that the mixed boundary conditions are given as equality constraints. Hence, we study the following optimal control problem: J (x(t0 ), tf , x(tf )) Minimize (8.1) subject to the constraints ẋ(t) = f (t, x(t), u(t)), K(x(t0 ), tf , x(tf )) = 0. u(t) ∈ U , t0 ≤ t ≤ tf , (8.2) (8.3) The control variable appears linearly in the system dynamics, f (t, x, u) = a(t, x) + B(t, x)u. (8.4) The control set U ⊂ Rd(u) is a convex polyhedron with V = ex U denoting the set of vertices. ˆ = [t0 , tˆf ] Let us recall the IOP in (7.4). Assume that û(t) is a bang-bang control in k with switching points tˆk ∈ (t0 , tˆf ) and values u in the set V = ex U , û(t) = uk ∈ V for t ∈ (tˆk−1 , tˆk ), k = 1, . . . , s + 1, ˆ = {tˆ1 , . . . , tˆs } is the set of switching points of the control where tˆ0 = t0 , tˆs+1 = tˆf . Thus, û(·) with tˆk < tˆk+1 for k = 0, 1, . . . , s. Put x̂(t0 ) = x̂0 ∈ Rn , θ̂ = (tˆ1 , . . . , tˆs ) ∈ Rs , ζ̂ = (x̂0 , θ̂ , tˆf ) ∈ Rn × Rs+1 . (8.5) For convenience, the sequence of components of the vector ζ̂ in definition (7.1) has been modified. Take a small neighborhood V of the point ζ̂ and let ζ = (x0 , θ, ts+1 ) ∈ V, θ = (t1 , . . . , ts ), ts+1 = tf , where the switching times satisfy t0 < t1 < t2 < · · · < ts < ts+1 = tf . Define the function u(t; θ) by the condition u(t; θ) = uk for t ∈ (tk−1 , tk ), k = 1, . . . , s + 1. (8.6) The values u(tk ; θ ), k = 1, . . . , s, may be chosen in U arbitrarily. For definiteness, define them by the condition of continuity of the control from the left, u(tk ; θ ) = u(tk −; θ ) for k = 1, . . . , s, and let u(0; θ) = u1 . Denote by x(t; x0 , θ ) the absolutely continuous solution of the Initial Value Problem (IVP) ẋ = f (t, x, u(t; θ )), t ∈ [t0 , tf ], x(t0 ) = x0 . (8.7) 8.1. The Arc-Parametrization Method 341 For each ζ ∈ V this solution exists in a sufficiently small neighborhood V of the point ζ̂ . Obviously, we have x(t; x̂0 , θ̂) = x̂(t), ˆ t ∈ , u(t; θ̂ ) = û(t), ˆ \ . ˆ t ∈ Consider now the following Induced Optimization Problem (IOP) in the space Rn × Rs+1 of variables ζ = (x0 , θ, ts+1 ): F0 (ζ ) := J (x0 , ts+1 , x(ts+1 ; x0 , θ )) → min, G(ζ ) := K(x0 , ts+1 , x(ts+1 ; x0 , θ )) = 0. (8.8) To get uniqueness of the Lagrange multiplier for the equality constraint we assume that the following regularity (normality) condition holds: rank ( Gζ (ζ̂ ) ) = d(K). (8.9) The Lagrange function (7.6) in normalized form is given by L(β, ζ ) = F0 (ζ ) + β G(ζ ), β ∈ Rd(K) . (8.10) The critical cone K0 in (7.13) then reduces to the subspace K0 = { ζ̄ ∈ Rn × Rs+1 | Gζ (ζ̂ )ζ̄ = 0 }. (8.11) Assuming regularity (8.9), the SSC for the IOP reduce to the condition that there exist β̂ ∈ Rd(K) with Lζ (β̂, ζ̂ ) = (F0 )ζ (ζ̂ ) + β̂ Gζ (ζ̂ ) = 0, Lζ ζ (β̂, ζ̂ )ζ̄ , ζ̄ > 0 ∀ ζ̄ ∈ K0 \ {0}. (8.12) (8.13) From a numerical point of view it is not convenient to optimize the switching times tk (k = 1, . . . , s) and terminal time ts+1 = tf directly. Instead, as suggested in [44, 45, 66] one computes the arc durations or arc lengths τk := tk − tk−1 , k = 1, . . . , s, s + 1, (8.14) of the bang-bang arcs. Hence, the final time tf can be expressed by the arc lengths as tf = t0 + s+1 τk . (8.15) k=1 Next, we replace the optimization variable ζ = (x0 , t1 , . . . , ts , ts+1 ) by the optimization variable (8.16) z := (x0 , τ1 , . . . , τs , τs+1 ) ∈ Rn × Rs+1 , τk := tk − tk−1 . The variables ζ and z are related by the following linear transformation involving the regular (n + s + 1) × (n + s + 1) matrix R: In 0 In 0 , , ζ = R −1 z, R −1 = z = Rζ, R = 0 S 0 S −1 ⎛ ⎞ ⎛ ⎞ 1 0 ... 0 1 0 ... 0 (8.17) ⎜ ⎜ . ⎟ . ⎟ . .. ⎜ 1 ⎜ −1 . .. ⎟ 1 . . .. ⎟ 1 −1 ⎜ ⎟ ⎜ ⎟ S =⎜ . S=⎜ ⎟. ⎟, .. .. ⎝ .. . . . . . . 0 ⎠ ⎝ . . 0 ⎠ 1 ... 1 1 0 −1 1 342 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem Denoting the solution to the equations (8.7) by x(t; z), the IOP (8.8) obviously is equivalent to the following IOP with tf defined by (8.15): F0 (z) G(z) := J (x0 , tf , x(tf ; x0 , τ1 , . . . , τs )) → min, (8.18) := K(x0 , tf , x(tf ; x0 , τ1 , . . . , τs )) = 0. This approach is called the arc-parametrization method. The Lagrangian for this problem is given in normal form by L(ρ, z) = F0 (z) + ρ G(z). (8.19) It is easy to see that the Lagrange multiplier ρ agrees with the multiplier β in the Lagrangian (8.10). Furthermore, the SSC for the optimization problems (8.8), respectively, (8.18), are equivalent. This immediately follows from the fact that the Jacobian and the Hessian for both optimization problems are related through Kζ = Kz R, Lζ = Lz R, Lζ ζ = R ∗ Lzz R. Thus we can express the positive definiteness condition (8.13) evaluated for the variable z as Lzz (β̂, ẑ)z̄, z̄ > 0 ∀ z̄ ∈ (Rn × Rs+1 ) \ {0}, Gẑ (ẑ)z̄ = 0. (8.20) This condition is equivalent to the property that the so-called reduced Hessian is positive definite. Let N be the (nz × (nz − d(K))) matrix, nz = n + s + 1, with full column rank nz − d(K), whose columns span the kernel of Gz (ẑ). Then condition (8.20) is reformulated as N ∗ Lzz (β̂, ẑ) N > 0 (positive definite). (8.21) Rn+s+1 is based on a The computational method for determining the optimal vector ẑ ∈ multiprocess approach proposed in [44, 45, 66]. The time interval [tk−1 , tk ] is mapped to k the fixed interval Ik = [ k−1 s+1 , s+1 ] by the linear transformation 4 3 k−1 k t = ak + bk r, ak = tk − kτk , bk = (s + 1)τk , r ∈ Ik = , , (8.22) s +1 s +1 where r denotes the running time. Identifying x(r) ∼ = x(ak + bk · r) = x(t) in the relevant intervals, we obtain the transformed dynamic system dx dx dt (8.23) = · = (s + 1) τk f (ak + bk r, x(r), uk ) for r ∈ Ik . dr dt dr By way of concatenation of the solutions on the intervals Ik , we obtain an absolutely continuous solution x(r) = x(r; τ1 , . . . , τs ) for r ∈ [0, 1]. Thus we are confronted with the task of solving the IOP s+1 tf = t0 + τk , F0 (z) := J (x0 , tf , x(1; x0 , τ1 , . . . , τs )) → min, (8.24) G(z) := K(x0 , tf , x(1; x0 , τ1 , . . . , τs )) = 0. k=1 This approach can be conveniently implemented using the routine NUDOCCCS developed by Büskens [13]. In this way, we can also take advantage of the fact that NUDOCCCS provides the Jacobian of the equality constraints and the Hessian of the Lagrangian, which are needed in the check of the second-order condition (8.13), respectively, the positive definiteness of the reduced Hessian (8.21). Moreover, this code allows for the computation of parametric sensitivity derivatives; cf. [48, 74, 23]. 8.1. The Arc-Parametrization Method 8.1.3 343 Extension of the Arc-Parametrization Method to Piecewise Feedback Control The purpose of this section is to extend the IOPs (8.18) and (8.24) to the situation in which the control is piecewise defined by feedback functions and not only by a constant vector uk ∈ V = ex U . We consider the control problem (8.1)–(8.3) with an arbitrary dynamical system ẋ(t) = f (t, x(t), u(t)), t0 ≤ t ≤ tf , (8.25) and fixed initial time t0 . Instead of considering a bang-bang control u(t) with s switching times t0 = t0 < t1 < · · · < ts < ts+1 = tf and constant values u(t) = uk t ∈ (tk−1 , tk ), for we assume that there exist continuous functions uk : D → Rd(u) , where D ⊂ R × Rn is open, with u(t) = uk (t, x(t)) for t ∈ (tk−1 , tk ). (8.26) Such functions are provided, e.g., by singular controls in feedback form or boundary controls in the presence of state constraints, or may be simply viewed as suitable feedback approximations of an optimal control. The vector of switching times is denoted by θ = (t1 , . . . , ts ). Let x(t; x0 , θ ) be the absolutely continuous solution of the piecewise defined equations ẋ(t) = f (t, x(t), uk (t, x(t))) for t ∈ (tk−1 , tk ) (k = 1, . . . , s + 1) (8.27) with given initial value x(t0 ; x0 , θ) = x0 . The IOP with the optimization variable ζ = (x0 , θ, tf ) = (x0 , t1 , . . . , ts , ts+1 ) ∈ Rn × Rs+1 agrees with (8.8): F0 (ζ ) := J (x0 , ts+1 , x(ts+1 ; x0 , θ )) → min, G(ζ ) := K(x0 , ts+1 , x(ts+1 ; x0 , θ )) = 0. (8.28) The arc-parametrization method consists of optimizing x0 and the arc lengths τk := tk − tk−1 , k = 1, . . . , s + 1. Invoking the linear time transformation (8.22) for mapping the time interval [tk−1 , tk ] to the k fixed interval Ik = [ k−1 s+1 , s+1 ], t = ak + bk r, ak = tk − kτk , bk = (s + 1)τk , r ∈ Ik , the dynamic system is piecewise defined by dx/dr = (s + 1)τi f (ak + bk r, x(r), uk (ak + bk r, x(r)) for r ∈ Ik . (8.29) Therefore, the implementation of the arc-parametrization method using the routine NUDOCCCS [13] requires only a minor modification of the dynamic system. Applications of this method to bang-singular controls may be found in [50, 51] and to state constrained problems in [74]. An example for a bang-singular control will be given in Section 8.7. 344 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem 8.2 Time-Optimal Control of the Rayleigh Equation Revisited We revisit the problem of time-optimal control of the Rayleigh equation from Section 6.4 and show that the sufficient conditions in Theorem 7.10 can be easily verified numerically on the basis of the IOP (8.8) or (8.18). The control problem is to minimize the final time tf subject to ẋ1 (t) = x2 (t), ẋ2 (t) = −x1 (t) + x2 (t)(1.4 − 0.14x2 (t)2 ) + u(t), x1 (0) = x2 (0) = −5, x1 (tf ) = x2 (tf ) = 0, | u(t) | ≤ 4 for t ∈ [0, tf ]. (8.30) (8.31) (8.32) The Pontryagin function (Hamiltonian) H (x, ψ, u) = ψ1 x2 + ψ2 (−x1 + x2 (1.4 − 0.14x22 ) + u) (8.33) yields the adjoint equations ψ̇1 = ψ2 , ψ̇2 = ψ1 + ψ2 (1.4 − 0.42x22 ). (8.34) The transversality conditions (6.9) gives the relation The switching function ψ2 (tf ) u(tf ) + 1 = 0. (8.35) φ(t) = Hu (t) = ψ2 (t) (8.36) determines the optimal control via the minimum condition as 4 if ψ2 (t) < 0 . u(t) = −4 if ψ2 (t) > 0 (8.37) In section 6.4 it was found that the optimal control is composed of three bang-bang arcs, ⎫ ⎧ ⎪ ⎪ ⎬ ⎨ 4 for 0 ≤ t < t1 −4 for t1 ≤ t < t2 . (8.38) u(t) = ⎪ ⎭ ⎩ 4 for t ≤ t ≤ t = t ⎪ 2 3 f This implies the two switching conditions ψ2 (t1 ) = 0 and ψ2 (t2 ) = 0. Hence, the optimization vector for the IOP (8.18) is given by z = (τ1 , τ2 , τ3 ), τ1 = t1 , τ2 = t2 − t1 , τ3 = tf − t2 . The code NUDOCCCS gives the following numerical results for the arc lengths, switching times, and adjoint variables: τ1 = t1 = 1.12051, 0.35813, τ3 = ψ1 (0) = −0.122341, ψ1 (t1 ) = −0.215212, ψ1 (tf ) = 0.842762, τ2 = tf = ψ2 (0) = ψ1 (t2 ) = ψ2 (tf ) = 2.18954, t2 3.66817, −0.082652, 0.891992, −0.25, β = 3.31005, (8.39) = ψ(tf ). 8.2. Time-Optimal Control of the Rayleigh Equation Revisited (a) time-optimal control u and switching function ( x 4 ) (b) 345 state variables x1 and x2 6 4 3 4 2 2 1 0 0 -2 -1 -2 -4 -3 -6 -4 -8 0 0.5 1 1.5 2 2.5 3 3.5 4 0 0.5 1 1.5 2 2.5 3 3.5 4 Figure 8.1. Time-optimal control of the Rayleigh equation with boundary conditions (8.31). (a) Bang-bang control and scaled switching function (×4), (b) State variables x1 , and x2 . The corresponding time-optimal bang-bang control with two switches and the state variables are shown in Figure 8.1. We have already shown in Section 6.4 that the control u in (8.38) enjoys the strict bang-bang property and that the estimates D k (H ) > 0, k = 1, 2, in (6.114) are satisfied. For the terminal conditions (8.31), the Jacobian is the (2 × 3) matrix −4.53176 −3.44715 0.0 Gz (ẑ) = −11.2768 −7.62049 4.0 which is of rank 2. The Hessian of the Lagrangian is the (3 × 3) matrix ⎛ ⎞ −10.3713 −8.35359 −6.68969 Lzz (β̂, ẑ) = ⎝ −8.35359 −5.75137 −4.61687 ⎠ . −6.68969 −4.61687 1.97104 Note that this Hessian is not positive definite. However, the projected Hessian (8.21) is the positive number N ∗ L̃zz (ẑ, β̂)N = 0.515518, which shows that the second-order test (8.20) holds. Hence, the extremal characterized by the data (8.39) provides a strict strong minimum. Now we consider a modified control problem, where the two terminal conditions x1 (tf ) = x2 (tf ) = 0 are substituted by the scalar terminal condition x1 (tf )2 + x2 (tf )2 = 0.25. (8.40) The Hamiltonian (8.33) and the adjoint equations (8.34) remain the same. The transversality condition (6.9) yields λi (tf ) = 2 β xi (tf ) (i = 1, 2), β ∈ R. (8.41) The transversality condition for the free final time is H (t) + 1 ≡ 0. It turns out that the control is bang-bang with only one switching point t1 in contrast to the control structure (8.38), 4 for 0 ≤ t < t1 . (8.42) u(t) = −4 for t1 ≤ t ≤ tf 346 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem (a) time-optimal control u and switching function (x4) (b) state variables x1 and x2 6 4 4 3 2 2 1 0 0 -2 -1 -2 -4 -3 -6 -4 -8 0 0.5 1 1.5 2 2.5 3 0 0.5 1 1.5 2 2.5 3 Figure 8.2. Time-optimal control of the Rayleigh equation with boundary condition (8.40). (a) Bang-bang control u and scaled switching function φ (dashed line). (b) State variables x1 , x2 . Hence, the optimization vector is z = (τ1 , τ2 ), τ1 = t1 , τ2 = tf − t1 . Using the code NUDOCCCS, we obtain the following numerical results for the arc lengths and adjoint variables: t1 = ψ1 (0) = ψ1 (t1 ) = x1 (tf ) = ψ1 (tf ) = 1.27149, −0.117316, −0.213831, 0.426176, 0.448201, τ2 = 1.69227, ψ2 (0) = −0.0813638, ψ2 (t1 ) = 0.0, x2 (tf ) = 0.261484, ψ2 (tf ) = 0.274997, tf = 2.96377, (8.43) β = 0.525839. Figure 8.2 displays the time-optimal bang-bang control with only one switch and the two state variables. The relations φ̇(t1 ) = ψ̇2 (t1 ) = ψ1 (t1 ) and [u]1 = 8 yield D 1 (H ) = −8 · φ̇(t1 ) = 8 · 0.213831 > 0. For the scalar terminal condition (8.40), the Jacobian is the nonzero row vector Gz (ẑ) = (−1.90175, −1.90173), while the Hessian of the Lagrangian is the positive definite (2 × 2) matrix 28.1299 19.0384 . Lzz (β̂, ẑ) = 19.0384 14.0048 Hence, the second-order conditions (8.20), respectively, the second-order conditions in Theorem 7.10, hold, which shows that the extremal characterized by the data (8.43) furnishes a strict strong minimum. 8.3 Time-Optimal Control of a Two-Link Robot The control of two-link robots has been the subject of various articles; cf., e.g., [21, 35, 37, 81]. In these papers, optimal control policies are determined solely on the basis of first-order 8.3. Time-Optimal Control of a Two-Link Robot 347 x2 P C q2 Q q1 O x1 lower arm OP , and angles q1 and q2 . Figure 8.3. Two-link robot [67]: upper arm OQ, necessary conditions, since sufficient conditions were not available. In this section, we show that SSC hold for both types of robots considered in [21, 37, 81]. First, we study the robot model considered in Chernousko et al. [21]. Göllmann [37] has shown that the optimal control candidate presented in [21] is not optimal, since the sign conditions of the switching functions do not comply with the Minimum Principle. Figure 8.3 represents the two-link robot schematically. The state variables are the angles q1 and q2 . The parameters I1 and I2 are the moments of inertia of the upper arm OQ and the lower arm QP with respect to the points O and Q. Further, let m2 be the mass of the lower arm, L1 = |OQ| the length of the upper arm, and L1 = |QC| the distance between the second link Q and the center of gravity C of the lower arm. With the abbreviations A R1 D = I1 + m2 L21 + I2 + 2m2 L1 L cos q2 , B = u1 + m2 L1 L(2q̇1 + q̇2 )q̇2 sin q2 , R2 = I2 , = I2 + m2 L1 L cos q2 , = u2 − m2 L1 Lq̇12 sin q2 , = AD − B 2 , (8.44) the dynamics of the two-link robot can be described by the ODE system 1 (DR1 − BR2 ), (8.45) 1 (AR2 − BR1 ), q̇2 = ω2 , ω̇2 = where ω1 and ω2 are the angular velocities. The torques u1 and u2 in the two links represent the two control variables. The control problem consists of steering the robot from a given initial position to a terminal position in minimal final time tf , q̇1 q1 (0) q1 (tf ) = 0, = −0.44, = ω1 , ω̇1 = q2 (0) = 0, ω1 (0) = 0, ω2 (0) q2 (tf ) = 1.83, ω1 (tf ) = 0, ω2 (tf ) = 0, = 0. (8.46) Both control components are bounded by |u1 (t)| ≤ 2, |u2 (t)| ≤ 1, t ∈ [0, tf ]. (8.47) 348 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem The Pontryagin function (Hamiltonian) is ψ3 ψ4 (DR1 (u1 ) − BR2 (u2 )) + (AR2 (u2 ) − BR1 (u1 )) . (8.48) H = ψ1 ω1 + ψ2 ω2 + The adjoint equations are rather complicated and are not given here explicitly. The switching functions are φ1 (t) = Hu1 (t) = ψ4 ψ3 D− B, φ2 (t) = Hu2 (t) = ψ3 ψ4 A− B. (8.49) For the parameter values L1 = 1, L = 0.5, m2 = 10, I1 = I2 = 10 , 3 Göllmann [37] has found the following control structure with four bang-bang arcs: ⎧ ⎫ (−2, 1), 0 ≤ t < t1 ⎪ ⎪ ⎨ ⎬ (2, 1), t1 ≤ t < t2 , 0 < t1 < t2 < t3 < tf . (8.50) u(t) = (u1 (t), u2 (t)) = t2 ≤ t < t3 ⎪ ⎪ ⎩ (2, −1), ⎭ (−2, −1), t3 ≤ t ≤ tf This control structure differs substantially from that in Chernousko et al. [21] which violates the switching conditions. Obviously, the bang-bang control (8.50) satisfies the assumption that only one control component switches at a time. Since the initial point (q1 (0), q2 (0), ω1 (0), ω2 (0)) is specified, the optimization variable in the IOP (8.18) is z = (τ1 , τ2 , τ3 , τ4 ), τ1 = t1 , τ2 = t2 − t1 , τ3 = t3 − t2 , τ4 = tf − t3 . Using the code NUDOCCCS, we compute the following arc durations and switching times: t1 = 0.7677893, τ2 = 0.3358820, τ3 = 1.2626739, t3 = 2.3663452, tf = 3.1971119. t2 = 1.1036713, τ4 = 0.8307667, (8.51) Numerical values for the adjoint functions are also provided by the code NUDOCCCS, e.g., the initial values ψ1 (0) ψ3 (0) = = −1.56972, −2.90537, ψ2 (0) ψ4 (0) = = −0.917955, −1.45440. (8.52) Figure 8.4 shows that the switching functions φ1 and φ2 comply with the minimum condition and that the strict bang-bang property (6.19) and the inequalities (6.14) are satisfied: φ1 (t) = 0 for t = t1 , t3 , φ2 (t) = 0 for φ̇1 (t1 ) < 0, φ̇1 (t3 ) > 0, φ̇2 (t2 ) > 0. For the terminal conditions (8.46) we compute the Jacobian ⎛ −0.751043 0.0351060 0.258904 −0.204170 ⎜ 3.76119 1.84929 Gz (ẑ) = ⎝ −0.326347 0.0770047 0.212722 1.26849 0.445447 −0.487447 t = t2 , ⎞ 0 0 ⎟ . −0.107819 ⎠ −0.233634 8.3. Time-Optimal Control of a Two-Link Robot (a) control u1 and switching function (b) 1 2 349 control u2 and switching function 2 1 1.5 1 0.5 0.5 0 0 -0.5 -1 -0.5 -1.5 -2 -1 0 0.5 (c) 1 1.5 2 2.5 angle q1 and velocity 3 3.5 0 0.5 (d) 1 1 1.5 2 angle q2 and velocity 2.5 3 3.5 3 3.5 2 2 0.1 1.8 0 1.6 1.4 -0.1 1.2 1 -0.2 0.8 -0.3 0.6 0.4 -0.4 0.2 0 -0.5 0 0.5 1 1.5 2 2.5 3 0 3.5 0.5 1 1.5 2 2.5 Figure 8.4. Control of the two-link robot (8.44)–(8.47). (a) Control u1 and scaled switching function φ1 (dashed line). (b) Control u2 and scaled switching function φ2 (dashed line). (c) Angle q1 and velocity ω1 . (d) Angle q2 and velocity ω2 . This square matrix has full-rank in view of det Gz (ẑ) = 0.0766524 = 0, which means that the positive definiteness condition (8.13) trivially holds. We have thus verified first-order sufficient conditions showing that the extremal solution given by (8.50)– (8.52) provides a strict strong minimum. In the model treated above, some parameters such as the mass of the upper arm and the mass of a load at the end of the lower arm appear implicitly in the system equations. The mass m1 of the upper arm is included in the moment of inertia I2 , and the mass M of a load in the point P can be added to the mass m2 , where the point C and therefore the length L have to be adjusted. The length L2 of the lower arm is incorporated in the parameter L. The second robot model that we are going to discuss is taken from Geering et al. [35] and Oberle [81]. Here, every physical parameter enters the system equation explicitly. The dynamic system is as follows: 1 q̇1 = ω1 , (AI22 − BI12 cos q2 ), ω̇1 = (8.53) 1 ω̇2 = (BI11 − AI12 cos q2 ), q̇2 = ω2 − ω1 , where we have used the abbreviations A = I12 ω22 sin q2 + u1 − u2 , B = −I12 ω12 sin q2 + u2 , = 2 cos2 q , I11 I22 − I12 2 = I1 + (m2 + M)L21 , = I2 + I3 + ML22 . I12 = m2 LL1 + ML1 L2 , I11 I22 (8.54) 350 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem I3 denotes the moment of inertia of the load with respect to the point P , and ω2 is now the angular velocity of the angle q1 + q2 . For simplicity, we set I3 = 0. Again, the torques u1 and u2 in the two links are used as control variables by which the robot is steered from a given initial position to a nonfixed end position in minimal final time tf , ! (x1 (tf ) − x1 (0))2 + (x2 (tf ) − x2 (0))2 = r, q1 (0) = 0, q2 (tf ) = 0, q2 (0) = 0, (8.55) ω1 (0) = 0, ω1 (tf ) = 0, ω2 (0) = 0, ω2 (tf ) = 0, where (x1 (t), x2 (t)) are the Cartesian coordinates of the point P , x1 (t) x2 (t) = = L1 cos q1 (t) + L2 cos(q1 (t) + q2 (t)), L1 sin q1 (t) + L2 sin(q1 (t) + q2 (t)). (8.56) The initial point (x1 (0), x2 (0)) = (2, 0) is fixed. Both control components are bounded, |u1 (t)| ≤ 1, |u2 (t)| ≤ 1, t ∈ [0, tf ]. (8.57) The Hamilton–Pontryagin function is given by ψ3 (A(u1 , u2 )I22 − B(u2 )I12 cos q2 ) ψ1 ω1 + ψ2 (ω2 − ω1 ) + ψ4 (B(u2 )I11 − A(u1 , u2 )I12 cos q2 ) . + The switching functions are computed as H = φ 1 = H u1 = 1 (ψ3 I22 − ψ4 I12 cos q2 ) , φ2 = Hu2 = 1 (ψ3 (−I22 − I12 cos q2 ) + ψ4 (I11 + I12 cos q2 )) . (8.58) (8.59) For the parameter values 1 I1 = I2 = , r = 3, 3 we will show that the optimal control has the following structure with five bang-bang arcs: ⎫ ⎧ (−1, 1) for 0 ≤ t < t1 ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎨ (−1, −1) for t1 ≤ t < t2 ⎬ (1, −1) for t2 ≤ t < t3 , u(t) = (u1 (t), u2 (t)) = (8.60) ⎪ ⎪ ⎪ ⎪ (1, 1) for t ≤ t < t ⎪ ⎪ 3 4 ⎭ ⎩ (−1, 1) for t4 ≤ t ≤ tf L1 = L2 = 1, L = 0.5, m1 = m2 = M = 1, where 0 = t0 < t1 < t2 < t3 < t4 < t5 = tf . Since the initial point (q1 (0), q2 (0), ω1 (0), ω2 (0)) is specified, the optimization variable in the optimization problem (8.8), respectively, (8.18), is z = (τ1 , τ2 , τ3 , τ4 , τ5 ), τk = tk − tk−1 , k = 1, . . . , 5. The code NUDOCCCS yields the arc durations and switching times t1 τ3 t4 = = = 0.546174, 1.03867, 3.70439, τ2 t3 τ5 = = = 1.21351, 2.79835, 0.185023, t2 τ4 tf = = = 1.75968, 0.906039, 3.889409, (8.61) 8.3. Time-Optimal Control of a Two-Link Robot (a) control u1 (b) 1 1 0.5 0.5 0 0 -0.5 -0.5 -1 -1 0 0.5 (c) 1 1.5 2 351 2.5 angle q1 and velocity 3 3.5 4 control u2 0 0.5 (d) 1 1 1 1.5 2 2.5 angle q2 and velocity 3 3.5 4 3 3.5 4 2 3 2.5 0.5 2 0 1.5 -0.5 1 -1 0.5 0 -1.5 -0.5 -2 -1 -2.5 -1.5 0 0.5 1 1.5 2 2.5 3 3.5 4 0 0.5 1 1.5 2 2.5 Figure 8.5. Control of the two-link robot (8.53)–(8.57). (a) Control u1 . (b) Control u2 . (c) Angle q1 and velocity ω1 . (d) Angle q2 and velocity ω2 [17]. as well as the initial values of the adjoint variables, ψ1 (0) ψ3 (0) = = 0.184172, 1.482636, ψ2 (0) ψ4 (0) = = −0.011125, 0.997367. (8.62) The two bang-bang control components as well as the four state variables are shown in Figure 8.5. The strict bang-bang property (6.19) and the inequalities (6.14) hold in view of φ1 (t) = 0 for t = t2 , t4 , φ2 (t) = 0 for t = t1 , t3 , φ̇1 (t2 ) < 0, φ̇1 (t4 ) > 0, φ̇2 (t1 ) > 0, φ̇2 (t3 ) < 0. For the terminal conditions in (8.55), the Jacobian in the optimization problem is computed as the (4 × 5) matrix ⎛ ⎞ −10.8575 −12.7462 −5.88332 −1.14995 0 0.199280 −2.71051 −1.45055 −1.91476 −4.83871 ⎟ ⎜ , Gz (ẑ) = ⎝ −0.622556 3.31422 2.31545 2.94349 6.19355 ⎠ 9.36085 3.03934 0.484459 0.0405811 0 which has full rank. The Hessian of the Lagrangian is given by ⎛ 71.1424 90.7613 42.1301 8.49889 112.544 51.3129 10.7691 ⎜ 90.7613 ⎜ 51.3129 23.9633 5.12403 Lzz (β̂, ẑ) = ⎜ 42.1301 ⎝ 8.49889 10.7691 5.12403 1.49988 −0.0518216 0.149854 0.138604 0.170781 −0.0518216 0.149854 0.138604 0.170781 0.297359 ⎞ ⎟ ⎟ ⎟. ⎠ 352 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem (a) control u1 (b) 1 1 0.5 0.5 0 0 -0.5 -0.5 -1 -1 0 0.5 (c) 1 1.5 2 2.5 angle q1 and velocity 3 3.5 4 control u2 0 0.5 1 1.5 2 2.5 angle q2 and velocity (d) 1 3 3.5 4 3 3.5 4 2 1.5 1 1 0.5 0.5 0 0 -0.5 -0.5 -1 -1 -1.5 -2 -1.5 -2.5 -2 -3 0 0.5 1 1.5 2 2.5 3 3.5 4 0 0.5 1 1.5 2 2.5 Figure 8.6. Control of the two-link robot (8.53)–(8.57): Second solution. (a) Control u1 . (b) Control u2 . (c) Angle q1 and velocity ω1 . (d) Angle q2 and velocity ω2 . This yields the projected Hessian (8.21) as the positive number N ∗ Lzz (β̂, ẑ)N = 0.326929. Hence, all conditions in Theorem 7.10 are satisfied, and thus the extremal (8.60)–(8.62) yields a strict strong minimum. It is interesting to note that there exists a second local minimum with the same terminal time tf = 3.88941. Though the control has also five bang-bang arcs, the control structure is substantially different from that in (8.60), ⎫ ⎧ (1, −1), 0 ≤ t < t1 ⎪ ⎪ ⎪ ⎪ ⎪ t 1 ≤ t < t2 ⎪ ⎬ ⎨ (−1, −1), (−1, 1), t 2 ≤ t < t3 , (8.63) u(t) = (u1 (t), u2 (t)) = ⎪ ⎪ ⎪ ⎪ (1, 1), t ≤ t < t ⎪ ⎪ 3 4 ⎭ ⎩ (1, −1), t4 ≤ t ≤ tf where 0 < t1 < t2 < t3 < t4 < t5 = tf . Figure 8.6 displays the second time-optimal solution: the bang-bang controls and state variables. The code NUDOCCCS determines the switching times t1 = 0.1850163, t2 = 1.091075, t3 = 2.129721, (8.64) t4 = 3.343237, tf = 3.889409. for which the strict bang-bang property (6.19) and the inequalities (6.14) hold, i.e., D k (H ) > 0 for k = 1, 2, 3, 4. Moreover, computations show that rank ( Gz (ẑ) ) = 4 and the projected Hessian of the Lagrangian (8.21) is the positive number N ∗ Lzz (β̂, ẑ)N = 0.326929. 8.4. Time-Optimal Control of a Single Mode Semiconductor Laser 353 It is remarkable that this value is identical to the value of the projected Hessian for the first local minimum. Therefore, also for the second solution we have verified that all conditions in Theorem 7.10 hold, and thus the extremal (8.63), (8.64) is a strict strong minimum. The phenomenon of multiple local solutions all with the same minimal time tf has also been observed by Betts [5, Example 6.8 (Reorientation of a rigid body)]. 8.4 Time-Optimal Control of a Single Mode Semiconductor Laser In [46] we studied the optimal control for two classes of laser, a so-called class B laser and a semiconductor laser. In this section, we present only the semiconductor laser, whose dynamical model has been derived in [20, 29]. The primary goal is to control the transition between the initial stationary state, the switch-off state, and the terminal stationary state, the switch-on state. The control variable is the electric current injected into the laser by which the laser output power is modified. In response to an abrupt switch from a low value of the current (initial state) to a high value (terminal state), the system responds with damped oscillations (cf. Figure 8.8) which are a great nuisance in several laser applications. We will show that by injecting an appropriate bang-bang current (control) the oscillations can be completely removed while simultaneously shortening the transition time. Semiconductor lasers are a case where the removal of the oscillations can be particularly beneficial, since in telecommunications one would like to be able to obtain the fastest possible response to the driving, with the cleanest and most direct transition, to maximize the data transmission rate and efficiency. We consider the dynamics of a standard single-mode laser model [20], reduced to single-mode form [29]. In this model, S(t) represents the normalized photon number and N(t) the carrier density, I (t) is the injected current (control) that is used to steer the transition between different laser power levels. Ṡ = Ṅ = S + G(N , S)S + βBN (N + P0 ), tp I (t) − R(N ) − G(N , S)S. q − (8.65) The process is considered in a time interval t ∈ [0, tf ] with terminal time tf > 0. The gain function G(N , S) and recombination function R(N ) are given by G(N , S) = Gp (N − Ntr )(1 − #S), R(N ) = AN + BN (N + P0 ) + CN (N + P0 )2 . (8.66) The parameters have the following meaning: tp , cavity lifetime of the photon; , cavity confinement factor; β, coefficient that weights the (average) amount of spontaneous emission coupled into the lasing mode; B, incoherent band-to-band recombination coefficient; P0 , carrier number without injection; q, carrier charge; Gp , gain term; #, gain compression factor; Ntr , number of carriers at transparency. All parameter values are given in Table 8.1. The following bounds are imposed for the injected current: Imin ≤ I (t) ≤ Imax ∀ t ∈ [0, tf ], (8.67) 354 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem Table 8.1. List of parameters from [29]. The time unit is a picosecond [ps] = [10−12 s]. tp 2.072 × 10−12 s Gp 2.628 × 104 s−1 0.3 # 9.6 × 10−8 Ntr 7.8 × 107 β 1.735 × 10−4 P0 1.5 × 107 q 1.60219 × 10−19 C A 1 × 108 s−1 B 2.788 s−1 C 7.3 × 10−9 s−1 I0 20.5 mA Imin 2.0 mA Imax 67.5 mA I∞ 42.5 mA where 0 ≤ Imin < Imax . To define appropriate initial and terminal values for S(t) and N(t), we choose two values I0 and I∞ with Imin < I0 < I∞ < Imax . Then inserting the constant control functions I (t) ≡ I0 and I (t) ≡ I∞ into the dynamics (8.65), one can show that there exist two asymptotically stable stationary points (S0 , N0 ) and (Sf , Nf ) with Ṡ = Ṅ = 0 such that (S(t), N(t)) → (S0 , N0 ) for t → ∞ and I (t) ≡ I0 , (S(t), N (t)) → (Sf , Nf ) for t → ∞ and I (t) ≡ I∞ . Hence, we shall impose the following initial and terminal conditions for the control process (8.65): (8.68) S(0) = S0 , N(0) = N0 and S(tf ) = Sf , N (tf ) = Nf . When controlling the process by the function I (t), one goal is to determine a control function by which the terminal stationary point is reached in a finite time tf > 0. But we can set a higher goal by considering the following time-optimal control problem: Minimize the final time tf (8.69) subject to the dynamic constraints and boundary conditions (8.65)–(8.68). For computation, we shall use the nominal parameters from [29] (see Table 8.1). For these parameters, the stationary points, respectively, initial and terminal values, in (8.68) are computed in normalized units as S0 = 0.6119512914 × 105 , Sf = 3.4063069073 × 105 , N0 = 1.3955581328 × 108 , Nf = 1.4128116637 × 108 . The Hamilton–Pontryagin function is given by S H (S, N, ψS , ψN , I ) = ψS − + G(N , S)S + βBN (N + P0 ) tp I − R(N ) − G(N , S)S , + ψN q where the adjoint variables (ψS , ψN ) satisfy the adjoint equations 1 − G(N , S) + Gp (N − Ntr )#S ψ̇S = −HS = ψS tp + ψN G(N , S) − Gp (N − Ntr )#S , ψ̇N = −HN = −ψS (Gp (1 − #S)S + βB(2N + P0 )) + ψN ( A + B(2N + P0 ) + C(3N 2 + 4N P0 + P02 ) + Gp (1 − #S)S). (8.70) (8.71) (8.72) 8.4. Time-Optimal Control of a Single Mode Semiconductor Laser 355 The switching function becomes φ(t) = HI (t) = ψN (t)/q. (8.73) The minimization of the Hamiltonian (8.71) yields the following characterization of bangbang controls: Imin if ψN (t) > 0, (8.74) I (t) = Imax if ψN (t) < 0. In this problem, a singular control satisfying the condition ψN (t) ≡ 0 for all t ∈ [t1 , t2 ] ⊂ [0, tf ], t1 < t2 , cannot be excluded a priori. However, the direct optimization approach yields the following bang-bang control with only one switching time t1 , Imax if 0 ≤ t < t1 , I (t) = (8.75) Imin if t1 ≤ t ≤ tf . In view of the control law (8.74), we get the switching condition ψN (t1 ) = 0. Moreover, since the final time tf is free and the control problem is autonomous, we obtain the additional boundary condition for a normal trajectory, H (S(tf ), N (tf ), ψS (tf ), ψN (tf ), I (tf )) + 1 = 0. (8.76) The optimization variable in the IOP (8.18) is z = (τ1 , τ2 ), τ1 = t1 , τ2 = tf − t1 . (8.77) It is noteworthy that the IOP (8.18) reduces to solving an implicitly defined nonlinear equation: determine two variables τ1 , τ2 such that the two boundary conditions S(tf ) = Sf and N (tf ) = Nf in (8.70) are satisfied. Thus solving the IOP is equivalent to applying a Newton-type method to the system of equations. We obtain the following switching time, terminal time, and initial values of adjoint variables: 29.52274, t1 = ψS (0) = −21.6227, ψS (tf ) = −4.6956, tf = 56.89444 ps, ψN (0) = −340.892, ψN (tf ) = 395.60. (8.78) The corresponding control and (normalized) state functions as well as adjoint variables are shown in Figure 8.7. Note that the constant control I (t) ≡ I∞ has to be applied for t ≥ tf in order to fix the system at the terminal stationary point (Sf , Nf ). Since the bang-bang control I (t) has only one switch, Proposition 6.25 asserts that the computed extremal furnishes a strict strong minimum. The computed trajectory is normal, because the adjoint variables satisfy the necessary condition (6.7)–(6.11) with α0 = 1. Moreover, the graph of ψN (t) in Figure 8.7 shows that the strict bang-bang property and D 1 (H ) > 0 in (6.14) hold in view of φ(t) < 0 ∀ 0 ≤ t < t1 , φ(t) > 0 ∀ t1 < t ≤ tf , φ̇(t1 ) > 0. These conditions provide first-order sufficient conditions. Alternatively, we can use Theorem 7.10 for proving optimality. The critical cone is the zero element, since the computed 2 × 2 Jacobian matrix 0.199855 −0.000155599 Gz (ẑ) = 0.0 −0.00252779 356 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem (a) (b) 1.48 1.47 1.46 1.45 1.44 1.43 1.42 1.41 1.4 1.39 Normalized photon number S 3.5 3 2.5 2 1.5 1 0.5 0 10 (c) 20 30 40 50 0 60 Injected electric current (control) I Normalized carrier density N (d) 10 20 30 adjoint variables 70 400 60 300 50 200 40 100 40 S and 50 60 N 0 30 -100 20 -200 10 -300 0 -400 -20 0 20 40 60 80 0 10 20 30 40 50 60 Figure 8.7. Time-optimal control of a semiconductor laser. (a) Normalized photon density S(t) × 10−5 . (b) Normalized photon density N (t) × 10−8 . (c) Electric current (control) I (t) with I (t) = I0 = 20.5 for t < 0 and I (t) = I∞ = 42.5 for t > tf . (d) Adjoint variables ψS (t), ψN (t). 6 5 4 3 2 1 0 0 100 200 300 400 500 600 Figure 8.8. Normalized photon number S(t) for I (t) ≡ 42.5 mA and optimal I (t) [46]. is regular. Comparing the optimal control approach with the topological phase-space technique proposed in [29], we recognize that the control structure (8.75) constitutes a translation of the latter technique into rigorous mathematical terms. The comparison in Figure 8.8 between the uncontrolled and optimally controlled laser shows the strength of the optimal control approach: the damped oscillations have been completely eliminated and a substantial shortening of the transient time has been achieved. Indeed, it is surprising that such a dramatic improvement is caused by the simple control strategy (8.75) adopted here. 8.5. Optimal Control of a Batch-Reactor 357 Feed of B Cooling Figure 8.9. Schematics of a batch-reaction with two control variables. It is worth stressing the improvement that the optimal control approach has brought to the problem. The disappearance of the damped oscillations allows the laser to attain its final state in a finite time rather than asymptotically. In practical terms, one can set a threshold value, δ, around the asymptotic level, S∞ , and consider the state attained once S∞ − δ < S(t) < S∞ + δ holds. The parameter δ can be determined by the amount of noise present in the system, which we do not consider in our analysis. Visually, this operation corresponds to saying that the damped oscillation and asymptotic state are indistinguishable below a certain level of detail, e.g., if t > 500 ps as in Figure 8.8. With this convention, one can give a quantitative estimate of the amount of improvement introduced by the control function: the asymptotic state is attained at t ≈ 50 ps, even before the first crossing of the same level which occurs at t ≈ 70 ps (dashed line in Figure 8.8). The improvement is of the order of a factor 10! 8.5 Optimal Control of a Batch-Reactor The following optimal control problem for a batch-reactor is taken from [17, 110, 66]. Consider a chemical reaction A+B → C and its side reaction B +C → D which are assumed to be strongly exothermic. Thus, direct mixing of the entire necessary amounts of the reactants must be avoided. The reactant A is charged in the reactor vessel, which is fitted with a cooling jacket to remove the generated heat of the reaction, while the reactant B is added. These reactions result in the product C and the undesired byproduct D. The vector of state variables is denoted by x = (MA , MB , MC , MD , H ) ∈ R5 , where Mi (t) [mol] and Ci (t) [mol/m3 ] stand for the molar holdups and the molar concentrations of the components i = A, B, C, D, respectively. H (t) [MJ] denotes the total energy 358 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem holdup, TR (t) [K] the reactor temperature and V (t) [m3 ] the volume of liquid in the system. The two-dimensional control vector is given by u = (FB , Q) ∈ R2 , where FB (t) [mol/s] controls the feed rate of the component B while Q(t) [kW] controls the cooling load. The objective is to determine a control u that maximizes the molar holdup of the component C. Hence, the performance index is J1 (x(tf )) = −MC (tf ) (8.79) which has to be minimized subject to the dynamical equations = −V · r1 , ṀB = FB − V · (r1 + r2 ), = V · (r1 − r2 ), ṀD = V · r2 , = FB · hf − Q − V · (r1 · H1 + r2 · H2 ). ṀA ṀC Ḣ (8.80) Here, rj denote the reaction rates and kj the corresponding Arrhenius rate constants of both reactions (j = 1, 2): A+B → C : C+B → D : r1 r2 = = where functions are defined by S = k1 · CA · CB k2 · CB · CC Mi · α i , with with = Ci = = = W = i=A,B,C,D TR k1 k2 A1 · e−E1 /TR , A2 · e−E2 /TR , (8.81) Mi · βi , i=A,B,C,D ! 1 · −S + (W · Tref + S)2 + 2 · W · H , W Mi Mi ( i = A, B, C, D ), V = . V ρi (8.82) i=A,B,C,D The reference temperature for the enthalpy calculations is Tref = 298 K and the specific molar enthalpy of the reactor feed stream is hf = 20 kJ/mol. Initial values are given for all state variables, MA (0) = 9000, Mi (0) = 0 (i = B, C, D), H (0) = 152509.97, (8.83) while there is only one terminal constraint TR (tf ) = 300 (8.84) with TR defined as in (8.82). The control vector u = (FB , Q) appears linearly in the control system (8.80) and is bounded by 0 ≤ FB (t) ≤ 10 and 0 ≤ Q(t) ≤ 1000 ∀ t ∈ [0, tf ]. The reaction and component data appearing in (8.80)–(8.82) are given in Table 8.2. (8.85) 8.5. Optimal Control of a Batch-Reactor 359 Table 8.2. Reaction and component data. Notation Reactions j =1 j =2 % 3 & Aj m 0.008 mol·s Ej [K] 3000 ·p % & kJ Hj −100 mol Notation i=A % & mol ρi m3 11250 Meaning 0.002 2400 Preexponential Arrhenius constants Activation energies −75 Enthalpies Components i=B i=C i=D Meaning 16000 10000 10400 Molar density of pure component i % & kJ 0.1723 0.2 0.16 0.155 Coefficient of the linear (αi ) % mol·K & kJ βi 0.000474 0.0005 0.00055 0.000323 and quadratic (βi ) term in mol·K2 the pure component specific enthalpy expression αi Calculations show that for increasing tf the switching structure gets more and more complex. However, the total profit of MC (tf ) is nearly constant if tf is greater than a certain value, tf ≈ 1600. For these values one obtains singular controls. We choose the final time tf = 1450 and will show that the optimal control has the following bang-bang structure with 0 < t1 < t2 < tf : ⎫ ⎧ for 0 ≤ t < t1 ⎬ ⎨ (10, 0) (10, 1000) for t1 ≤ t < t2 . (8.86) u(t) = (FB (t), Q(t)) = ⎩ (0, 1000) for t ≤ t ≤ t ⎭ 2 f Since the initial point x(0) is specified and the final time tf is fixed, the optimization variable in the IOP (8.18) is given by z = (τ1 , τ2 ), τ1 = t1 , τ2 = t2 − t1 . Then the arc-length of the terminal bang-bang arc is τ3 = 1450 − τ1 − τ2 . The code NUDOCCCS yields the following arc-lengths and switching times: J1 (x(tf )) = 3555.292, t1 = 433.698, t2 = 767.273, τ2 = 333.575, τ3 = 1450 − t2 = 682.727. (8.87) We note that for this control the state constraint TR (t) ≤ 520 imposed in [17] does not become active. The adjoint equations are rather complicated and are not given here explicitly. The code NUDOCCCS also provides the adjoint functions, e.g., the initial values ψMA (0) ψMC (0) ψH (0) = = = −0.0299034, −2.83475, 0.00192489. ψMB (0) ψMD (0) = = 0.0433083, −0.10494943, (8.88) 360 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem control FB and scaled switching function (a) 1 control Q and scaled switching function (b) 10 1000 8 800 6 600 4 400 2 200 2 0 0 -200 -2 -400 -4 0 200 400 600 800 1000 1200 0 1400 200 molar concentration MA 9000 400 600 800 1000 1200 1400 molar concentration MB 400 350 8000 300 7000 250 6000 200 150 5000 100 4000 50 0 3000 0 200 400 600 800 1000 1200 1400 0 200 400 molar concentrations MC and MD 4000 700000 3500 600000 3000 600 800 1000 1200 1400 total energy holdup H 500000 2500 400000 2000 300000 1500 1000 200000 500 100000 0 0 0 200 400 600 800 1000 1200 1400 0 200 400 600 800 1000 1200 1400 Figure 8.10. Control of a batch reactor with functional (8.79). Top row: Control u = (FB , Q) and scaled switching functions. Middle row: Molar concentrations MA and MB . Bottom row: Molar concentrations (MC , MD ) and energy holdup H . Figure 8.10 (top row) clearly shows that the strict bang-bang property holds with φ̇2 (t1 ) < 0, φ̇1 (t2 ) > 0. The Jacobian of the scalar terminal condition (8.84) is computed as Gz (ẑ) = (0.764966, 0.396419), while the Hessian of the Lagrangian is the positive matrix 0.00704858 0.00555929 . Lzz (β̂, ẑ) = 0.00555929 0.00742375 Thus, the second-order condition (8.20) is satisfied and Theorem 7.10 tells us that the solution (8.86)–(8.88) provides a strict strong minimum. 8.6. Optimal Production and Maintenance with L1 -Functional 361 Let us now change the cost functional (8.79) and maximize the average gain of the component C in time, i.e., minimize J2 (tf , x(tf )) = − MC (tf ) , tf where the final time tf is free. We will show that the bang-bang control (10, 1000) for 0 ≤ t < t1 u(t) = (FB (t), Q(t)) = (0, 1000) for t1 ≤ t ≤ tf (8.89) (8.90) with only one switching point 0 < t1 < tf of the control u1 (t) = FB (t) is optimal. Since the initial point x(0) is specified, the optimization variable in the IOP (8.18) is z = (τ1 , τ2 ), τ1 = t1 , τ2 = tf − t1 . Using the code NUDOCCCS, we obtain the switching and terminal times J2 (tf , x(tf )) = 3.877103, τ2 = 171.399, t1 = 285.519, tf = 456.918. (8.91) and initial values of the adjoint variables ψMA (0) ψMC (0) ψH (0) = = = −0.9932 · 10−4 , ψMB (0) −0.108578 · 10−2 , ψMD (0) 0.298 · 10−5 . = = −0.61036 · 10−3 , 0.9495 · 10−4 , (8.92) We may conclude from Figure 8.11 that φ̇1 (t1 ) > 0 holds, while the switching function for the control u2 (t) = Q(t) = 1000 satisfies φ2 (t) < 0 on [0, tf ]. The Jacobian for the scalar terminal condition (8.84) is Gz (ẑ) = (0.054340, −0.31479), while the Hessian of the Lagrangian is the positive definite matrix 0.37947 −0.11213 Lzz (β̂, ẑ) = · 10−4 . −0.11213 0.37748 Hence the SSC (8.20) hold and, again, Theorem 7.10 asserts that the solution (8.90)– (8.92) is a strict strong minimum. 8.6 Optimal Production and Maintenance with L1 -Functional In section 5.4, we studied the optimal control model presented in Cho, Abad, and Parlar [22], where optimal production and maintenance policies are determined simultaneously. The cost functional was quadratic with respect to the production control u, which enhances a continuous control. In this section, we consider the case when the production control enters the cost functional linearly. In this model, x = x1 denotes the inventory level, y = x2 362 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem (a) control FB and scaled switching function control Q and scaled switching function (b) 1 10 1000 8 800 6 600 4 2 400 2 200 0 0 -2 -200 -4 -400 -6 0 100 200 300 400 0 500 100 molar concentration MA 200 300 400 500 400 500 molar concentration MB 9000 900 800 8500 700 600 8000 500 400 7500 300 200 7000 100 0 6500 0 100 200 300 400 500 0 100 molar concentrations MC and MD 200 300 total energy holdup H 1800 160000 1600 140000 1400 120000 1200 100000 1000 800 80000 600 60000 400 40000 200 20000 0 0 0 100 200 300 400 500 0 100 200 300 400 500 Figure 8.11. Control of a batch reactor with functional (8.89). Top row: Control u = (FB , Q) and scaled switching functions. Middle row: Molar concentrations MA , MB . Bottom row: Molar concentrations (MC , MD ) and energy holdup H . the proportion of good units of end items produced, v is the scheduled production rate (control), and m denotes the preventive maintenance rate (control). The parameters are α = 2 obsolescence rate, s = 4 demand rate, and ρ = 0.1 discount rate. The control bounds and weights in the cost functional will be specified below. The dynamics of the process is given by ẋ(t) = y(t)v(t) − 4, x(0) = 3, x(tf ) = 0, (8.93) ẏ(t) = −2y(t) + (1 − y(t))m(t), y(0) = 1, with the following bounds on the control variables: 0 ≤ v(t) ≤ 3, 0 ≤ m(t) ≤ 4 for 0 ≤ t ≤ tf . (8.94) Recall also that the terminal condition x(tf ) = 0 implies the nonnegativity condition x(t) ≥ 0 for all t ∈ [0, tf ]. Now we choose the following L1 -functional: Maximize the 8.6. Optimal Production and Maintenance with L1 -Functional 363 total discounted profit and salvage value of y(tf ), J (x, y, m, v) = tf [8s − x(t) − 4 v(t) − 2.5 m(t)]e−0.1·t dt 0 + 10 y(tf )e−0.1·tf (8.95) ( tf = 1 ) under the constraints (8.93) and (8.94). Though the objective has to be maximized, we discuss the necessary conditions on the basis of the Minimum Principle which has been used throughout the book. Again, we use the standard Hamilton–Pontryagin function instead of the current value Hamiltonian: H (t, x, y, ψ1 , ψ2 , m, v) = e−0.1·t (−32 + x + 4v + 2.5 m) + ψ1 (yv − 4) + ψ2 (−2y + (1 − y)m), (8.96) where ψ1 , ψ2 denote the adjoint variables. The adjoint equations and transversality condition yield, in view of the terminal constraint x1 (tf ) = 0 and the salvage term in the cost functional, ψ1 (tf ) = ν, ψ̇1 = −e−0.1 , (8.97) ψ̇2 = −ψ1 v + ψ2 (2 + m), ψ2 (tf ) = −10e−0.1 , where ν ∈ R is an unknown multiplier. The switching functions are given by φm = Hm = 2.5 e−0.1·t + ψ2 (1 − y), φv = Hv = 4 e−0.1·t + ψ1 y, which determines the controls according to 4 if φm (t) < 0 3 if φv (t) < 0 , m(t) = . v(t) = 0 if φv (t) > 0 0 if φm (t) > 0 (8.98) Singular controls for which either φv (t) = 0 or φm (t) = 0 holds on an interval I ⊂ [0, tf ] cannot be excluded a priori. However, for the data and the final time tf = 1 chosen here, the application of direct optimization methods [5, 13, 14] reveals the following control structure with four bang-bang arcs: ⎧ (3, 0) for 0 ≤ t < t1 ⎪ ⎨ (0, 0) for t1 ≤ t < t2 (v(t), m(t)) = ⎪ ⎩ (0, 4) for t2 ≤ t < t3 (3, 4) for t3 ≤ t ≤ tf = 1 ⎫ ⎪ ⎬ ⎪ ⎭ . (8.99) The optimization vector for the IOP (8.18) is z = (τ1 , τ2 , τ3 ), τ1 = t1 , τ2 = t2 − t1 , τ3 = t3 − t2 . Therefore, the terminal arc-length is given by τ4 = tf −(τ1 +τ2 +τ3 ). The code NUDOCCCS yields the following numerical results for the arc-lengths and adjoint variables: t1 τ3 J = = = 0.346533, 0.114494, 25.7969, τ2 = t3 = ψ(tf ) = 0.380525, 0.841552, −0.833792, t2 = tf = ψ2 (tf ) = 0.727058, 1.0, −0.904837. (8.100) 364 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem (a) stock x and good items y (b) production v and maintenance m 3 4 3.5 3 2.5 2 1.5 1 0.5 0 2.5 2 1.5 1 0.5 0 0 (c) 0.2 0.4 0.6 0.8 1 production control v and switching function v 0 (d) 3 4 2 3 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 maintenance m and switching function m 2 0 1 -1 0 -2 -1 -3 -4 -2 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Figure 8.12. Optimal production and maintenance with L1 -functional (8.95). (a) State variables x and y. (b) Control variables v and m. (c), (d) Control variables and switching functions. Figure 8.12 clearly indicates that the strict bang-bang property (6.19) holds, since in particular we have φ̇v (t1 ) > 0, φ̇v (t3 ) < 0, φ̇m (t2 ) < 0. For the scalar terminal condition x(tf ) = 0, the Jacobian is the nonzero row vector Gz (ẑ) = (−0.319377, −1.81950, −1.35638). The Hessian of the Lagrangian is computed as the (3 × 3) matrix ⎛ 41.6187 Lzz (β̂, ẑ) = ⎝ 21.0442 −3.43687 21.0442 21.0442 −3.43687 ⎞ −3.43687 −3.43687 ⎠ , 34.4731 from which the reduced Hessian (8.20) is obtained as the positive definite (2 × 2) matrix ∗ N Lzz (β̂, ẑ)N = 20.4789 6.61585 6.61585 49.6602 . Hence the second-order test (8.20) holds, which ensures that the control (8.99) with the data (8.100) yields a strict strong minimum. 8.7. Van der Pol Oscillator with Bang-Singular Control 365 8.7 Van der Pol Oscillator with Bang-Singular Control The following example with a bang-singular control is taken from Vossen [111, 112]. The optimal control is a concatenation of two bang-bang arcs and one terminal singular arc. The singular control is given by a feedback expression which allows us to optimize switching times directly using the arc-parametrization method presented in Section 8.1.3; cf. also [111, 112]. We consider again the dynamic model of a Van der Pol oscillator introduced in Section 6.7.1. The aim is to minimize the regulator functional 1 J (x, u) = 2 tf (x1 (t)2 + x2 (t)2 ) dt (tf = 4) (8.101) 0 subject to the dynamics, boundary conditions, and control constraints ẋ1 (t) = x2 (t), x1 (0) = 0, ẋ2 (t) = −x1 (t) + x2 (t)(1 − x1 (t)2 ) + u(t), x2 (0) = 1, | u(t) | ≤ 1 for t ∈ [0, tf ]. (8.102) The Hamilton–Pontryagin function H (x, u, ψ) = ψ1 x2 + ψ2 (−x1 + x2 (1 − x12 ) + u) (8.103) yields the adjoint equations and transversality conditions ψ̇1 = −x1 + ψ2 (1 + 2x1 x2 ), ψ(tf ) = 0, ψ̇2 = −x2 − ψ − ψ2 (1 − x12 ), ψ(tf ) = 0. (8.104) φ(t) = Hu (t) = ψ2 (t) (8.105) The switching function determines the bang-bang controls by u(t) = 1 if ψ2 (t) < 0 −1 if ψ2 (t) > 0 . (8.106) A singular control on an interval I ⊂ [0, tf ] can be computed from the relations φ = ψ2 = 0, φ̇ = ψ̇2 = −x2 − ψ1 = 0, φ̈ = 2x1 − x2 (1 − x12 ) − u = 0, which give the feedback expression u = using (x) = 2x1 − x2 (1 − x12 ). (8.107) It follows that the order of a singular arc is q = 1; cf. the definition of the order in [49]. Moreover, the strict Generalized Legendre Condition holds: (−1)q 3 4 ∂ d2 ∂ φ̈ = 1 > 0. H u =− ∂u dt 2 ∂u 366 Chapter 8. Numerical Methods for Solving the Induced Optimization Problem optimal control u state variables x1 , x2 1 1 0.8 0.6 0.5 0.4 0 0.2 0 -0.5 -0.2 -0.4 -1 -0.6 0 0.5 1 1.5 2 2.5 3 3.5 4 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 Figure 8.13. Control of the van der Pol oscillator with regulator functional. (a) Bang-singular control u. (b) State variables x1 and x2 . The application of direct optimization methods yields the following control structure with two bang-bang arcs and a terminal singular arc: ⎫ ⎧ for 0 ≤ t < t1 ⎬ ⎨ −1 1 for t1 ≤ t < t2 . (8.108) u(t) = ⎭ ⎩ 2 2x1 (t) − x2 (t)(1 − x1 (t) ) for t2 ≤ t ≤ tf = 4 Hence, the feedback functions in (8.26) are given by u1 (x) = −1, u2 (x) = 1, u3 (x) = 2x1 − x2 (1 − x12 ), which are inserted into the dynamic equations (8.102). Therefore, the optimization vector for the IOP (8.28) is given by z = (τ1 , τ2 ), τ 1 = t1 , τ2 = t2 − t1 . This yields the terminal arc length τ3 = 4 − τ1 − τ2 . The code NUDOCCCS provides the following numerical results for the arc-lengths, switching times, and adjoint variables: t1 = 1.36674, J (x, u) = 0.757618, τ2 ψ1 (0) = 0.109404, t2 = −0.495815, ψ2 (0) = 2.46078, = 2.58862. (8.109) The corresponding control and state variables are shown in Figure 8.13. The code NUDOCCCS furnishes the Hessian of the Lagrangian as the positive definite matrix 194.93 −9.9707 Lzz (ẑ) = . −9.9707 0.56653 At this stage we have only found a strict local minimum of the IOP (8.28). Recently, SSC for a class of bang-singular controls were obtained by Aronna et al. [2]. Combining these new results with the Riccati approach in Vossen [111] for testing the positive definiteness of quadratic forms, one can now verify the SSC for the extremal solutions (8.108) and (8.109). Here, the following property of the switching function is important: φ(t) > 0 for 0 ≤ t < t1 , φ̇(t1 ) < 0, φ(t) < 0 for t1 < t < t2 , φ(t2 ) = φ̇(t2 ) = 0, φ̈(t2− ) < 0, φ(t) = 0 for t2 < t < tf . Bibliography [1] A. A. Agrachev, G. Stefani, and P. Zezza, Strong optimality for a bang-bang trajectory, SIAM Journal on Control and Optimization, 41, pp. 991–1014 (2002). [5, 299, 300, 301, 304] [2] M. S. Aronna, J. F. Bonnans, A.V. Dmitruk, and P. A. Lotito, Quadratic conditions for bang-singular extremals, to appear in Numerical Algebra, Control and Optimization, 2, pp. 511–546 (2012). [366] [3] D. Augustin and H. Maurer, Sensitivity analysis and real-time control of a container crane under state constraints, In: Online Optimization of Large Scale Systems (M. Grötschel, S. O. Krumke, J. Rambau, eds.), pp. 69–82, Springer, Berlin, 2001. [219] [4] A. Ben-Tal and J. Zowe, Second-order optimality conditions for the L1 minimization problem, Applied Mathematics and Optimization, 13, pp. 45–48 (1985). [301] [5] J. T. Betts, Practical Methods for Optimal Control and Estimation Using Nonlinear Programming, 2nd ed., Advances in Design and Control 19, SIAM, Philadelphia, 2010. [250, 339, 353, 363] [6] G. A. Bliss, Lectures on the Calculus of Variations, University of Chicago Press, Chicago, 1946. [24] [7] H. G. Bock and K. J. Plitt, A multiple shooting algorithm for direct solution of optimal control problems, In: Proceedings of the 9th IFAC World Congress, Budapest, pp. 243–247, 1984. [339] [8] J. F. Bonnans and N. P. Osmolovskii, Characterization of a local quadratic growth of the Hamiltonian for control constrained optimal control problems, Dynamics of Continuous, Discrete and Impulsive Systems, Series B: Applications and Algorithms, 19, pp. 1–16 (2012). [177] [9] B. Bonnard, J.-B. Caillau, and E. Trélat, Second order optimality conditions in the smooth case and applications in optimal control, ESAIM Control, Optimization and Calculus of Variations, 13, pp. 207–236 (2007). [339] [10] B. Bonnard, J.-B. Caillau, and E. Trélat, Computation of conjugate times in smooth optimal control: The COTCOT algorithm, In: Proceedings of the 44th 367 368 Bibliography IEEE Conference on Decision and Control and European Control Conference 2005, pp. 929–933, Sevilla, December 2005. [339] [11] A. Bressan, A high order test for optimality of bang-bang controls, SIAM Journal on Control and Optimization, 23, pp. 38–48 (1985). [12] A. E. Bryson andY. C. Ho, Applied Optimal Control, Revised Printing, Hemisphere Publishing Corporation, New York, 1975. [219] [13] C. Büskens, Optimierungsmethoden und Sensitivitätsanalyse für optimale Steuerprozesse mit Steuer- und Zustands-Beschränkungen, Dissertation Institut für Numerische Mathematik, Universität Münster, Germany (1998). [6, 219, 273, 287, 288, 291, 339, 342, 343, 363] [14] C. Büskens and H. Maurer, SQP-methods for solving optimal control problems with control and state constraints: Adjoint variables, sensitivity analysis and realtime control, Journal of Computational and Applied Mathematics, 120, pp. 85–108 (2000). [6, 219, 250, 287, 288, 291, 339, 363] [15] C. Büskens and H. Maurer, Sensitivity analysis and real-time optimization of parametric nonlinear programming problems, In: Online Optimization of Large Scale Systems (M. Grötschel, S. O. Krumke, J. Rambau, eds.), pp. 3–16, SpringerVerlag, Berlin, 2001. [219] [16] C. Büskens and H. Maurer, Sensitivity analysis and real-time control of parametric optimal control problems using nonlinear programming methods, In: Online Optimization of Large Scale Systems (M. Grötschel, S. O. Krumke, J. Rambau, eds.), pp. 57–68, Springer-Verlag, Berlin, 2001. [219] [17] C. Büskens, H. J. Pesch, and S. Winderl, Real-time solutions of bang-bang and singular optimal control problems, In: Online Optimization of Large Scale Systems (M. Grötschel et al., eds.), pp. 129–142, Springer-Verlag, Berlin, 2001. [xii, 351, 357, 359] [18] R. Bulirsch, Die Mehrzielmethode zur numerischen Lösung von nichtlinearen Randwertproblemen und Aufgaben der optimalen Steuerung, Report of the CarlCranz-Gesellschaft, Oberpfaffenhofen, Germany, 1971. [339] [19] R. Bulirsch, F. Montrone, and H. J. Pesch, Abort landing in the presence of windshear as a minimax optimal control problem. Part 2: Multiple shooting and homotopy, Journal of Optimization Theory and Applications, 70, pp. 223–254 (1991). [339] [20] D. M. Byrne, Accurate simulation of multifrequency semiconductor laser dynamics under gigabit-per-second modulation, Journal of Lightwave Technology, 10, pp. 1086–1096 (1992). [353] [21] F. L. Chernousko, L. D. Akulenko, and N. N. Bolotnik, Time-optimal control for robotic manipulators, Optimal Control Applications & Methods, 10, pp. 293–311 (1989). [346, 347, 348] Bibliography 369 [22] D. I. Cho, P. L. Abad, and M. Parlar, Optimal production and maintenance decisions when a system experiences age-dependent deterioration, Optimal Control Applications & Methods, 14, pp. 153–167 (1993). [4, 248, 249, 250, 361] [23] B. Christiansen, H. Maurer, O. Zirn, Optimal control of servo actuators with flexible load and Coulombic friction, European Journal of Control, 17, pp. 1–11 (2011). [342] [24] B. Christiansen, H. Maurer, O. Zirn, Optimal control of machine tool manipulators, In: Recent Advances in Optimization and Its Applications in Engineering: The 14th Belgian-French-German Conference on Optimization, Leuven, September 2009 (M. Diehl, F. Glineur, E. Jarlebring, W. Michels, eds.), pp. 451–460, Springer, Berlin, 2010. [25] A. V. Dmitruk, Euler–Jacobi equation in the calculus of variations, Mat. Zametki, 20, pp. 847–858 (1976). [4, 191, 197] [26] A. V. Dmitruk, Jacobi-type conditions for the Bolza problem with inequalities, Mat. Zametki, 35, pp. 813–827 (1984). [4, 191, 193] [27] A. V. Dmitruk, Quadratic conditions for the Pontryagin minimum in a controllinear optimal control problem. I. Decoding theorem, Izv. Akad. Nauk SSSR Ser. Mat., 50, pp. 284–312 (1986). [68] [28] A. V. Dmitruk, A. A. Milyutin, and N. P. Osmolovskii, Lyusternik’s theorem and extremum theory, Usp. Mat. Nauk, 35, pp. 11–46 (1980); English translation: Russian Math. Surveys, 35, pp. 6, 11–51 (1980). [132] [29] N. Dokhane and G. L. Lippi, Chirp reduction in semiconductor lasers through injection current patterning, Applied Physics Letters, 78, pp. 3938–3940 (2001). [353, 354, 356] [30] A. Ya. Dubovitskii and A. A. Milyutin, Theory of the maximum principle, In: Methods of Extremal Problem Theory in Economics [in Russian], Nauka, Moscow, pp. 6–47 (1981). [166, 167] [31] U. Felgenhauer, On stability of bang-bang type controls, SIAM Journal on Control and Optimization, 41, pp. 1843–1867 (2003). [259, 338] [32] A.V. Fiacco, Introduction to Sensitivity and Stability Analysis in Nonlinear Programming, Mathematics in Science and Engineering 165, Academic Press, New York, 1983. [338] [33] R., Fourer, D. M. Gay, and B. W. Kernighan, AMPL: A Modeling Language for Mathematical Programming, Duxbury Press, Brooks-Cole Publishing Company, 1993. [250] [34] R. Gabasov and F. Kirillova, The Qualitative Theory of Optimal Processes, Marcel Dekker Inc., New York, Basel, 1976. 370 Bibliography [35] H. P. Geering, L. Guzella, S. A. R. Hepner, and Ch. Onder, Time-optimal motions of robots in assembly tasks, IEEE Transactions on Automatic Control, AC-31, pp. 512–518 (1986). [346, 349] [36] E. G. Gilbert and D. S. Bernstein, Second-order necessary conditions of optimal control: Accessory-problem results without normality conditions, Journal of Optimization Theory Applications, 41, pp. 75–106 (1983). [37] L. Göllmann, Numerische Berechnung zeitoptimaler Trajektorien für zweigliedrige Roboterarme, Diploma thesis, Institut für Numerische Mathematik, Universität Münster, 1991. [346, 347, 348] [38] B. S. Goh, Necessary conditions for singular extremals involving multiple control variables, SIAM Journal on Control, 4, pp. 716–731 (1966). [39] R. Henrion, La Théorie de la Variation Seconde et ses Applications en Commande Optimale, Academie Royal de Belgique, Bruxelles: Palais des Academies (1979). [1, 32] [40] M. Hestens, Calculus of Variations and Optimal Control Theory, John Wiley, New York, 1966. [225] [41] A. J. Hoffman, On approximate solutions of systems of linear inequalities, Journal of Research of the National Bureau of Standards, 49, pp. 263–265 (1952). [16] [42] A. D. Ioffe and V. M. Tikhomirov, Theory of Extremal Problems [in Russian], Nauka, Moscow (1974). Also published in German as Theorie der Extremalaufgaben, VEB Deutscher Verlag der Wissenschaften, Berlin, 1979. [191] [43] D. H. Jacobson and D.Q. Mayne, Differential Dynamic Programming, American Elsevier Publishing Company Inc., New York, 1970. [44] Y. Kaya and J. L. Noakes, Computations and time-optimal controls, Optimal Control Applications and Methods, 17, pp. 171–185 (1996). [6, 339, 341, 342] [45] C. Y. Kaya and J. L. Noakes, Computational method for time-optimal switching control, Journal of Optimization Theory and Applications, 117, pp. 69–92 (2003). [6, 339, 341, 342] [46] J.-H. R. Kim, G. L. Lippi, and H. Maurer, Minimizing the transition time in lasers by optimal control methods. Single mode semiconductor lasers with homogeneous transverse profile, Physica D (Nonlinear Phenomena) 191, pp. 238–260 (2004). [xii, 353, 356] [47] J.-H. R. Kim, H. Maurer,Yu. A. Astrov, M. Bode, and H.-G. Purwins, High-speed switch-on of a semiconductor gas discharge image converter using optimal control methods, Journal of Computational Physics, 170, pp. 395–414 (2001). [280] [48] J.-H. R. Kim and H. Maurer, Sensitivity analysis of optimal control problems with bang-bang controls, In: Proceedings of the 42nd IEEE Conference on Decision and Control, Maui, Dec. 9–12, 2003, IEEE Control Society, Washington, DC, pp. 3281– 3286, 2003. [338, 342] Bibliography 371 [49] A. J. Krener, The high order maximum principle, and its application to singular extremals, SIAM Journal on Control and Optimization, 15, pp. 256–293 (1977). [365] [50] U. Ledzewicz, H. Maurer, and H. Schättler, Optimal and suboptimal protocols for a mathematical model for tumor anti-angiogenesis in combination with chemotherapy, Mathematical Biosciences and Engineering, 8, pp. 307–328 (2011). [343] [51] U. Ledzewicz, H. Maurer, and H. Schättler, On optimal delivery of combination therapy for tumors, Mathematical Biosciences, 22, pp. 13–26 (2009). [343] [52] U. Ledzewicz, J. Marriott, H. Maurer, and H. Schättler, The scheduling of angiogenic inhibitors minimizing tumor volume, Journal of Medical Informatics and Technologies, 12, pp. 23–28 (2008). [53] U. Ledzewicz and H. Schättler, Optimal bang-bang controls for a 2-compartment model in cancer chemotherapy, Journal of Optimization Theory and Applications, 114, pp. 609–637 (2002). [270] [54] E. S. Levitin, A. A. Milyutin, and N. P. Osmolovskii, On local minimum conditions in problems with constraints, In: Mathematical Economics and Functional Analysis [in Russian], Nauka, Moscow, pp. 139–202 (1974). [3, 301] [55] E. S. Levitin, A. A. Milyutin, and N. P. Osmolovskii, Higher-order local minimum conditions in problems with constraints, Uspehi Mat. Nauk, 33, pp. 85–148 (1978). [3, 21, 174, 301] [56] E. S. Levitin, A. A. Milyutin, and N. P. Osmolovskii, Theory of higher-order conditions in smooth constrained extremal problems, In: Theoretical and Applied Optimal Control Problems [in Russian], Nauka, Novosibirsk, pp. 4–40, 246 (1985). [3] [57] K. Malanowski, C. Büskens, and H. Maurer, Convergence of approximations to nonlinear optimal control problems, In: Mathematical Programming with Data Perturbations (A.V. Fiacco, ed.), Lecture Notes in Pure and Applied Mathematics, Vol. 195, pp. 253–284, Marcel-Dekker, Inc., New York, 1997. [58] K. Malanowski and H. Maurer, Sensitivity analysis for parametric control problems with control-state constraints, Computational Optimization and Applications, 5, pp. 253–283 (1996). [219] [59] K. Malanowski and H. Maurer, Sensitivity analysis for state constrained optimal control problems, Discrete and Continuous Dynamical Systems, 4, pp. 241–272 (1998). [219] [60] P. Martinon and J. Gergaud, Using switching detection and variational equations for the shooting method, Optimal Control Applications and Methods, 28, pp. 95–116 (2007). [339] [61] H. Maurer, Numerical solution of singular control problems using multiple shooting techniques, Journal of Optimization Theory and Applications, 18, pp. 235–257 (1976). [339] 372 Bibliography [62] H. Maurer, First and second order sufficient optimality conditions in mathematical programming and optimal control, Mathematical Programming Study, 14, pp. 163–177 (1981). [25] [63] H. Maurer, Differential stability in optimal control problems, Applied Mathematics and Optimization, 5, pp. 283–295 (1979). [298] [64] H. Maurer and D. Augustin, Second order sufficient conditions and sensitivity analysis for the controlled Rayleigh problem, In: Parametric Optimization and Related Topics IV (J. Guddat et al., eds.) Lang, Frankfurt am Main, pp. 245–259 (1997). [219, 339] [65] H. Maurer and D. Augustin, Sensitivity analysis and real-time control of parametric optimal control problems using boundary value methods, In: Online Optimization of Large Scale Systems (M. Grötschel et al., eds.), pp. 17–55, Springer Verlag, Berlin, 2001. [203, 219, 339] [66] H. Maurer, C. Büskens, J.-H. R. Kim, and Y. Kaya, Optimization methods for the verification of second-order sufficient conditions for bang-bang controls, Optimal Control Methods and Applications, 26, pp. 129–156 (2005). [6, 339, 340, 341, 342, 357] [67] H. Maurer, J.-H. R. Kim, and G. Vossen, On a state-constrained control problem in optimal production and maintenance, In: Optimal Control and Dynamic Games, Applications in Finance, Management Science and Economics (C. Deissenberg, R. F. Hartl, eds.), pp. 289–308, Springer Verlag, 2005. [xii, 5, 223, 248, 249, 250, 347] [68] H. Maurer and H. J. Oberle, Second order sufficient conditions for optimal control problems with free final time: The Riccati approach, SIAM Journal on Control and Optimization, 41, pp. 380–403 (2002). [219] [69] H. Maurer and N. P. Osmolovskii, Second order optimality conditions for bangbang control problems, Control and Cybernetics, 32, pp. 555–584 (2003). [267] [70] H. Maurer and N. P. Osmolovskii, Second order sufficient conditions for time-optimal bang-bang control, SIAM Journal on Control and Optimization, 42, pp. 2239–2263 (2004). [274] [71] H. Maurer and H. J. Pesch, Solution differentiability for nonlinear parametric control problems, SIAM Journal on Control and Optimization, 32, pp. 1542–1554 (1994). [205, 219] [72] H. Maurer and H. J. Pesch, Solution differentiability for nonlinear parametric control problems with control-state constraints, Journal of Optimization Theory and Applications, 86, pp. 285–309 (1995). [219] [73] H. Maurer and S. Pickenhain, Second order sufficient conditions for optimal control problems with mixed control-state constraints, Journal of Optimization Theory and Applications, 86 (1995), pp. 649–667 (1995). [219, 267] Bibliography 373 [74] H. Maurer and G. Vossen, Sufficient conditions and sensitivity analysis for bang-bang control problems with state constraints, In: Proceedings of the 23rd IFIP Conference on System Modeling and Optimization, Cracow, Poland (A. Korytowski, M. Szymkat, eds.), pp. 82–99, Springer Verlag, Berlin, 2009. [342, 343] [75] H. Maurer and J. Zowe, First and second-order necessary and sufficient conditions for infinite-dimensional programming problems, Mathematical Programming, 16, pp. 98–110 (1979). [25] [76] A. A. Milyutin, Maximum Principle in the General Optimal Control Problem [in Russian], Fizmatlit, Moscow (2001). [179] [77] A. A. Milyutin, A. E. Ilyutovich, N. P. Osmolovskii, and S. V. Chukanov, Optimal Control in Linear Systems [in Russian], Nauka, Moscow (1993). [293] [78] A. A. Milyutin and N. P. Osmolovskii, (1) Higher-order minimum conditions on a set of sequences in the abstract problem with inequality-type constraints; (2) Higher-order minimum conditions on a set of sequences in the abstract problem with inequality- and equality-type constraints; (3) Higher-order minimum conditions with respect to a subsystem in the abstract minimization problems on a set of sequences, In: Optimality of Control Dynamical Systems [in Russian], No. 14, AllUnion Institute of System Studies, Moscow, pp. 68–95 (1990); English translation: Computational Mathematics and Modeling, 4, pp. 393–400 (1993). [9, 16] [79] A. A. Milyutin and N. P. Osmolovskii, Calculus of Variations and Optimal Control, Translations of Mathematical Monographs, Vol. 180, American Mathematical Society, Providence, 1998. [1, 2, 3, 4, 5, 30, 32, 71, 102, 103, 104, 105, 106, 107, 108, 109, 115, 127, 128, 130, 131, 207, 225, 229, 232, 255, 293, 294, 296, 297, 298] [80] J. Noble and H. Schättler, Sufficient conditions for relative minima of broken extremals in optimal control theory, Journal of Mathematical Analysis and Applications, 269, pp. 98–128 (2002). [270] [81] H. J. Oberle, Numerical computation of singular control functions for a two-link robot arm, In: Optimal Control, Lecture Notes in Control and Information Sciences, 95, pp. 244–253 (1987). [339, 346, 347, 349] [82] H. J. Oberle and W. Grimm, BNDSCO—A program for the numerical solution of optimal control problems, Institute for Flight Systems Dynamics, DLR, Oberpfaffenhofen, Germany, Internal Report No. 515–89/22, 1989. [204, 273, 287, 288, 291, 339] [83] H. J. Oberle and H. J. Pesch, Private communication, 2000. [272] [84] N. P. Osmolovskii, Second-order weak local minimum conditions in an optimal control problem (necessity, sufficiency), Dokl. Akad. Nauk SSSR, 225, pp. 259–262 (1975); English translation: Soviet Math. Dokl., 15, pp. 1480–1484 (1975). [25] [85] N. P. Osmolovskii, High-order necessary and sufficient conditions for Pontryagin and bounded strong minima in the optimal control problems, Dokl. Akad. Nauk SSSR Ser. Cybernetics and Control Theory, 303, pp. 1052–1056 (1988); English translation: Soviet Physics Doklady, 33, pp. 883–885 (1988). [1, 2] 374 Bibliography [86] N. P., Osmolovskii, Theory of higher-order conditions in optimal control, Doctoral Thesis, MISI (Moscow Civil Engineering Institute), Moscow, (1988). [1, 68, 173, 175] [87] N. P. Osmolovskii, Second-order conditions in a time-optimal control problem for linear systems, In: System Modelling and Optimization, Notes in Control and Information Sciences, 143, pp. 368–376 (1989). [293] [88] N. P. Osmolovskii, Quadratic conditions for nonsingular extremals in optimal control (Theory), Russian Journal of Mathematical Physics, 2, pp. 487–516 (1995). [2] [89] N. P. Osmolovskii, Quadratic conditions for nonsingular extremals in optimal control (examples), Russian Journal of Mathematical Physics, 5, pp. 487–516 (1998). [279] [90] N. P. Osmolovskii, Second-order conditions for broken extremals, In: Calculus of Variations and Optimal Control, Boca Raton, FL, pp. 198–216 (2000). [2] [91] N. P. Osmolovskii, Second-order sufficient conditions for an extremum in optimal control, Control and Cybernetics, 31, pp. 803–831 (2002). [92] N. P. Osmolovskii, Quadratic optimality conditions for broken extremals in the general problem of calculus of variations, Journal of Mathematical Sciences, 123, pp. 3987–4122 (2004). [27, 68, 127] [93] N. P. Osmolovskii, Second order conditions in optimal control problems with mixed equality-type constraints on a variable time interval, Control and Cybernetics, 38, pp. 1535–1556 (2009). [150] [94] N. P. Osmolovskii, Sufficient quadratic conditions of extremum for discontinuous controls in optimal control problem with mixed constraints, Journal of Mathematical Sciences, 173, pp. 1–106 (2011). [4, 175] [95] N. P. Osmolovskii, Necessary quadratic conditions of extremum for discontinuous controls in optimal control problem with mixed constraints, Journal of Mathematical Sciences, 183, pp. 435–576 (2012). [4, 68, 173, 175] [96] N. P. Osmolovskii and F. Lempio, Jacobi conditions and the Riccati equation for a broken extremal, Journal of Mathematical Sciences, 100, pp. 2572–2592 (2000). [2, 4, 212] [97] N. P. Osmolovskii and F. Lempio, Transformation of quadratic forms to perfect squares for broken extremals, Journal of Set-Valued Analysis, 10, pp. 209–223 (2002). [2, 4] [98] N. P. Osmolovskii and H. Maurer, Second order sufficient optimality conditions for a control problem with continuous and bang-bang control components: Riccati approach, In: Proceedings of the 23rd IFIP Conference on System Modeling and Optimization, Cracow, Poland (A. Korytowski, M. Szymkat, eds.), pp. 411–429, Springer Verlag, Berlin, 2009. [223, 247] Bibliography 375 [99] N. P. Osmolovskii and H. Maurer, Equivalence of second order optimality conditions for bang-bang control problems. Part 1: Main results, Control and Cybernetics, 34, pp. 927–950 (2005). [299] [100] N. P. Osmolovskii and H. Maurer, Equivalence of second order optimality conditions for bang-bang control problems. Part 2: Proofs, variational derivatives and representations, Control and Cybernetics, 36, pp. 5–45 (2007). [299] [101] H. J. Pesch, Real-time computation of feedback controls for constrained optimal control problems, Part 1: Neighbouring extremals, Optimal Control Applications and Methods, 10, pp. 129–145 (1989). [219, 339] [102] H. J. Pesch, Real-time computation of feedback controls for constrained optimal control problems, Part 2: A correction method based on multiple shooting, Optimal Control Applications and Methods, 10, pp. 147–171 (1989). [339] [103] L. S. Pontryagin, V. G. Boltyanski, R. V. Gramkrelidze, and E. F. Miscenko, The Mathematical Theory of Optimal Processes [in Russian], Fitzmatgiz, Moscow; English translation: Pergamon Press, New York, 1964. [225] [104] A. V. Sarychev, First- and second-order sufficient optimality conditions for bangbang controls, SIAM Journal on Control Optimization, 35, pp. 315–340 (1997). [279] [105] È. È. Shnol , On the degeneration in the simplest problem of the calculus of variations, Mat. Zametki, 24 (1978). [106] J. Stoer and R. Bulirsch, Introduction to Numerical Analysis, 2nd ed., SpringerVerlag, Berlin, 1992. [206, 339] [107] H. J. Sussmann, The structure of time-optimal trajectories for single-input systems in the plane: The C ∞ nonsingular case, SIAM Journal on Control and Optimization, 25, pp. 433–465 (1987). [1] [108] H. J. Sussmann, The structure of time-optimal trajectories for single-input systems in the plane: The general real analytic case, SIAM Journal on Control and Optimization, 25, pp. 868–904 (1987). [109] M. G. Tagiev, Necessary and sufficient strong extremum condition in a degenerate problem of the calculus of variations, Uspehi. Mat. Nauk, 34, pp. 211–212 (1979). [110] V. S. Vassiliadis, R. W. Sargent, and C. C. Pantelides, Solution of a class of multistage dynamic optimization problems. 2. Problems with path constraints, Industrial & Engineering Chemistry Research, 33, pp. 2123–2133 (1994). [357] [111] G. Vossen, Numerische Lösungsmethoden, hinreichende Optimalitätsbedingungen und Sensitivitätsanalyse für optimale bang-bang und singuläre Steuerungen, Dissertation, Institut für Numerische und Angewandte Mathematik, Universität Münster, Münster, Germany, 2005. [6, 365, 366] [112] G. Vossen, Switching time optimization for bang-bang and singular controls, Journal of Optimization Theory and Applications, 144, pp. 409–429 (2010). [6, 365] 376 Bibliography [113] G. Vossen and H. Maurer, On L1 -minimization in optimal control and applications to robotics, Optimal Control Applications and Methods, 27, pp. 301–321 (2006). [6] [114] A. Wächter and L. T. Biegler, On the implementation of an interior-point filter line-search algorithm for large-scale nonlinear programming, Mathematical Programming, 106, pp. 25–57 (2006); cf. IPOPT home page (C. Laird and A. Wächter): https://projects.coin-or.org/Ipopt. [250] [115] J. Warga, A second-order condition that strengthens Pontryagin’s maximum principle, Journal of Differential Equations, 28, pp. 284–307 (1979). [116] V. Zeidan, Sufficiency criteria via focal points and via coupled points, SIAM Journal of Control Optimization, 30, pp. 82–98 (1992). [219] [117] V. Zeidan, The Riccati equation for optimal control problems with mixed statecontrol constraints: Necessity and sufficiency, SIAM Journal on Control and Optimization, 32, pp. 1297–1321 (1994). [219] Index basic bang-bang control problem, 255 basic constant, 3, 13, 38 basic problem, 255, 299 batch-reactor, 357 bounded strong γ -sufficiency, 120, 139 bounded strong γ1 -sufficiency, 231 bounded strong minimum, 117, 154, 173, 180, 185, 226, 240 bounded strong sequence on Q, 173 B, 33, 99, 142, 149, 154, 155, 228, 242, 263, 265, 278, 297 B(), 94, 174, 181 Bco (), 94 Bτ , 156 BV , 149 B0 , 266, 304 ℵγ , 21 cone Ct , 177 Ct (), 176 C , 68 conjugate point, 188, 206 constant CK , 74 Cγ (m, g ), 11 Cγ (ω0 , K), 25 Cγ (0 , K), 22 Cγ (0 , σ ), 22 Cγ (0 , σ γ ), 22 Cγ̄ (C ; S4 ), 89 Cγ̄ (C ; S5 ), 92 Cγ (C ; o(√γ ) ), 76 √ Cγ (C ; loc o( γ ) ), 79 √ Cγ (M ; o( γ ) ), 101 Cγ (0 , σ γ ), 13, 22, 38 ˜ 1 ; S1 ), 84 Cγ ( C ˜ 1 ; loc√ , 83 Cγ ( C o( γ ) C-growth condition for H , 173 canonical problem, 27 canonical representation of a sequence, 41, 42 closure [· ]2 , 73 condition A, 33, 115, 137, 153, 173, 180, 227, 242, 263, 278, 294 Aco , 74 Aτ , 156 A0 , 304 Cγ1 (2C ; S2 ), 85 Cγ2 (2C ; S2 ), 87 Cγ2 (3C ; S3 ), 88 Cγ (σ , ), 12 convexification of control system, 230 critical cone, 22, 25, 31, 132, 153, 170, 180, 227, 241, 260, 301 K0 , 261 K0 in the induced problem, 303 K τ , 158 absorbing set of sequences, 9 abstract set of sequences , 9 adjacent conditions, 13 admissible function (t, u), 51 admissible pair, 28, 164, 184 admissible pair (variation) with respect to Qtu , 52 admissible process, 224, 238, 274 admissible trajectory, 256 almost global minimum, 293 arc lengths, 341 arc-parametrization method, 342, 343 augmented Hamiltonian, 166 augmented Pontryagin function, 150, 166, 178 auxiliary minimization problem, 186 auxiliary problem V , 138 377 378 K V , 147 K ζ , 110 KZ , 68 critical subspace, 186, 216, 277 decoding of the basic constant, 3 element b, 58 b̄, 63 ¯ 65 b̄, endpoint Lagrange function, 24, 150, 166, 178, 225, 239, 257 equivalent normalization, 13 essential component, 116, 139, 154, 173, 180, 226, 240 Euler equation, 184 external quadratic form, 172 extremal, 184 extremal trajectory, 275 family Ord(w 0 ), 174 fine linear approximation, 11 full-rank condition, 127 function (H ), 185 δH [t, v], 175 δu∗ , 41 δu∗k , 41 δuV , 77 δuV , 77 δv, 42 δvk+ , 42 δvk− , 42 δvk , 42 δwV , 77 δwloc , 77 δ¯u f , 122 δ̄u H λ , 122 (k H ), 240, 257, 276 (k H λ )(t), 32 (k H̄ ), 152, 168 g(ζ ), 305 H , 30, 130 H , 130 H̄ , 130 H τ , 157 H̄ τ , 157 H V , 142 Index l, 30, 130 l τ , 157 l V , 142 , 22, 25 0 , 22 ϕ(t, x), 310 "(t), 307 q(t; t0 ), 315 Q, 295 Q0 , 295 r(t; t0 , θ ), 318 ρ(t, v), 176 u(t; θ ), 300, 340 w(t; t0 ), 315 x(t; t0 , x0 , θ ), 300 x(t; x0 , θ ), 340 χi , 176 χ V , 77 χ ∗ , 41 χk∗ , 41 ∗ , 40 χk+ ∗ , 40 χk− y(t; t0 , θ ), 318 y k (t; tk ), 306 zkj (t; θ ) (k < j ), 314 zkk (t), 314 functional γ1 , 231 γ , 52 γ̄ , 33, 89, 142, 155, 174, 180, 187, 228, 242 γ V , 139, 149 γ1 , 58, 85 γ2 , 59, 63, 86, 87 0 , 32 C , 89 m, 10 m+ , 10 , 12, 37 , 58 C , 76 0 , 12, 38 1C , 82 1λ , 50 2C , 85 2λ , 58, 85 3C , 87 3λ , 63, 87 Index ˜ 1 , 83 C ˜ 1λ , 82 σ , 10, 100 379 local sequence, 39 Lyusternik condition, 11 Lyusternik normalization, 14 γ -conditions, 12 γ -minimum, 175 γ -necessity on , 11 γ -sufficiency on , 12 γ -sufficiency with respect to the rigid subsystem, 175 general problem linear in a part of controls, 237 general problem of the calculus of variations, 2, 27 generalized strengthened Legendre condition, 177 main problem, 223 matrix Riccati equation, 189, 246 matrix-valued function R(t; x0 , θ ), 311 S(t; t0 , x0 ), 317 V (t), 304 W k (t; x0 ), 310 minimal fuel consumption of a car, 272 minimum condition, 167 of strictness C, 173 minimum on a set of sequences, 9, 28, 127, 165 Minimum Principle, 151, 179 of strictness C, 72 Hamiltonian, 166, 224, 239, 275 higher order, 3, 11, 21 Hoffman lemma, 16, 23 needle-shaped variation, 54 neighborhood V of u0 , 39 Induced Optimization Problem (IOP), 300, 341, 342 Initial Value Problem (IVP), 300, 340 integral minimum principle, 167 internal quadratic form, 172 Jacobi conditions, 188 jump condition, 190, 216, 246 jump of a function, 2, 184 L-point, 28, 165 Lagrange function, 12, 37 for the IOP, 302 laser, 353 Legendre condition, 186 strengthened, 155, 186 Legendre element, 170 in nonstrict sense, 33 Legendre form, 191 Legendre–Clebsch condition, 169 limit cycle, 203 linear independence assumption, 165 Lipschitz point, 28 Lipschitz-continuous function, 178 local minimum principle, 24, 166 local quadratic growth of the Hamiltonian, 176 k neighborhood Vk of u0 , 39 nondegenerate case, 276, 279 nonlinear programming, 339 normal case, 276, 279 normalization condition, 12 operator π1 , 303 optimal production and maintenance policies, 248 optimality conditions, 9 order, 11 order γ , 174 order function, 51, 180 -continuous mapping, 10 0 -extension, 9 passage of quadratic form through zero, 191 point of L-discontinuity, 28 Pontryagin function, 24, 150, 166, 178, 224, 239, 257, 275 γ -sufficiency, 100 minimum, 29, 128, 151, 165, 184, 224, 238, 256, 275 minimum principle, 30, 167, 185, 225, 239, 257, 275 sequence on Q, 165 380 problem A, 243 Aτ , 244 co A, 243 co Z, 230 P , 156 P τ , 156 V , 138 Z, 229 ZN , 104 problem on a variable time interval, 150, 178 problem with a local equality, 127 production and maintenance, 361 projection, 128 projector π0 , 302 Q-transformation of on K, 246, 267, 281 quadratic conditions, 1 quadratic form, 22, 32, 227, 278 ¯ p̄¯ , 263 Ap̄, Lζ ζ (μ, ζ̂ )ζ̄ , ζ̄ , 319 ω, 153, 188 ωe , 153, 241 ωλ , 25, 172 λ , 172 ω , 153, 186, 227, 241, 263 # , 212 λ , 136, 172 τ , 159 U U λ , 137 V V λ , 148 ζ , 113 quadratic form in the IOP, 303 quantity quantity D k (H ), 226, 257, 276 D k (H λ ), 32 D k (H̄ λ ), 168 D k (H̄ ), 135, 152, 179 Rayleigh problem, 203, 344 time-optimal, 344 time-optimal control of the, 290 reduced Hessian, 342 regularity condition, 341 resource allocation problem, 263 Riccati equation, 216 Index second variation of Lagrange functional, 50 set A, 71 A0 , 68 AC , 68 co 0 , 30 D(), 103 F , 104 Fζ , 104 G, 128 , 103, 302 (), 104 0 , 12, 24, 30, 36, 166, 215, 301 0 for the IOP, 302 0 , 31, 121, 167 ζ 0 , 106 MP 0 , 302 ζ , 103 Leg+ (0 ), 96 Leg+ (M0+ ), 33, 96, 141, 155, 177, 228, 242 Leg+ (M), 170 M0 , 30, 130, 151, 167, 179, 215, 225, 236, 239, 257, 275 M(C), 94, 173, 181 M0co , 30 M co (C), 72 M(C ; co 0 ), 71 M0+ , 33, 96, 141, 177, 215, 228, 242 M0V , 142 M0τ , 157 ζ M0 , 108 M ∗ , 41 Mk∗ , 41 ∗ , 40 Mk+ ∗ , 40 Mk− M V , 77 N0 , 104 Qtu , 39 R, 294 , 28, 165 U, 229 U(t, x), 30, 151, 167, 179 u0 , 39 u0 (tk−1 , tk ), 39 Index 381 k u0 , 39 V ∗ , 39 ∗ , 39 Vk+ ∗ , 39 Vk− Vk∗ , 39 V 0 , 39 %, 296 set of active indices, 11, 24, 31, 165, 170, 227, 241, 301 set of indices I , 56 I in the IOP, 303 I ∗ , 34 S set of sequences , 117 , 29 , 54 ∗ , 41 ∗uk+ , 41 ∗uk− , 40 ∗uk , 41 ∗u , 41 + , 10 + g , 10 0 , 29, 34 S , 116, 173 √ loc o( γ ) , 76 1 loc σ γ , 57 loc , 39 loc u , 39 loc σ γ , 54 0 , 9 g , 10 σ γ , 38 σ , 22 σ γ , 13 o(√γ ) , 76 S 2 , 65 S 3 , 65 S 4 , 66 S1 , 84 S2 , 85 S3 , 88 S4 , 89 S5 , 91 simplest problem of the calculus of variations, 1, 183 singular control, 365 smooth problem, 11 smooth system, 11 space L1 (, Rd(u) ), 29 L2 (, Rd(u) ), 31 L∞ ([t0 , tf ], Rr ), 24 L∞ (, Rd(u) ), 28 P C 1 (, Rd(x) ), 260 P C 1 (, Rn ), 276 P W 1,2 ([t0 , tf ], Rm ), 185 P W 1,2 ([t0 , tf ], Rd(x) ), 170 P W 1,1 (, Rd(x) ), 60 P W 1,2 (, Rd(x) ), 30 Rn∗ , 24 W , 24, 28, 164, 184, 224 W 1,1 ([t0 , tf ], Rn ), 24 W 1,∞ (, (Rd(x) )∗ ), 30 W 1,1 (, Rd(x) ), 28 Z(), 67, 89 Z2 (), 30, 170, 186, 216, 226 Z(), 260 Z2 (), 180, 241 standard normalization, 12, 13 strict bang-bang control, 228, 259, 276 strict bounded strong minimum, 139, 154, 173, 180, 226, 240 strict minimum on a set of sequences, 10, 29, 127, 165 strict minimum principle, 33, 154, 185 strict order, 11, 21 strict Pontryagin minimum, 128 strict strong minimum, 154, 259, 275 strictly -differentiable mapping, 10 strictly Legendre element, 33, 141, 228, 242 strong minimum, 116, 154, 173, 259, 275 subspace K# , 212 K(τ ), 191 K0 , 188 L0 , 68 L̃2 ([τ , tf ], Rm ), 195 T , 92 support of a function, 40 switching function, 225, 257, 275 for the u-component, 239 system S, 10 382 -conjugate point, 4, 198 table, 197 time-optimal bang-bang control problem, 274 time-optimal control problem with a single switching point, 264 time-optimal control problems for linear system with constant entries, 293 trajectory T τ , 244 transformation to perfect square, 209, 217, 246, 271, 283 tuple of Lagrange multipliers, 30, 225, 239, 257, 301 two-link robot, 346 unessential component, 116, 139, 154, 173, 180, 226, 240 value ρ, 264 Van der Pol oscillator, 365 time-optimal control of a, 286 variable ζ of the IOP, 300 variational problem, 205 ¯ 241 vector p̄, violation function, 10, 38, 100, 139, 174, 230 weak minimum, 29, 165 Weierstrass-Erdmann conditions, 184 Index